query
stringlengths
7
3.85k
document
stringlengths
11
430k
metadata
dict
negatives
sequencelengths
0
101
negative_scores
sequencelengths
0
101
document_score
stringlengths
3
10
document_rank
stringclasses
102 values
Code gets the status code for the patch hyperflex server models moid default response
func (o *PatchHyperflexServerModelsMoidDefault) Code() int { return o._statusCode }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *PatchManagementEntitiesMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchHyperflexSoftwareVersionPoliciesMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *ObmsPatchByIDDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *IbmsPatchByIDDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchEquipmentIoCardsMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchGraphicsControllersMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetHyperflexServerModelsMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchEquipmentIoExpandersMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchMachineConfigurationDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchApplianceDiagSettingsMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchOpsNoteByIDDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchClusterV2Default) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchAddonDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchLicenseCustomerOpsMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchVnicEthNetworkPoliciesMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchRoleDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *ListClusterPatchPatchesDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchApplianceUpgradePoliciesMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchServiceAccountTokenDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchRecipientGroupByIDDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchActivityDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchDeviceGroupPropertyByNameDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchProductsByIDProductOptionsByIDDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *ReplaceServerDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchReportGroupByIDDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *UpdateAPIDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *EmsConfigModifyDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PostHyperflexHxdpVersionsMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *WebModifyDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchComponentHubDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *SvmPeerModifyDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchServiceAccountTokenOK) Code() int {\n\treturn 200\n}", "func (o *DcimDeviceTypesPartialUpdateDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *SnapmirrorRelationshipModifyDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *UpdateAntivirusServerDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *AddAPIDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchFirmwareServerConfigurationUtilityDistributablesMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchServiceAccountTokenNotFound) Code() int {\n\treturn 404\n}", "func (o *PatchServiceAccountTokenNoContent) Code() int {\n\treturn 204\n}", "func (o *UpdateMTOServiceItemStatusOK) Code() int {\n\treturn 200\n}", "func (o *PostHyperflexClusterProfilesMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetNiaapiApicSweolsMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *UpdateAppDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *UpdateAppDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *CreateServerSwitchingRuleDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PutLTENetworkIDGatewaysGatewayIDMagmadDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetNiaapiDcnmHweolsMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PostClustersMulticlusterConfigDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *EmailUpdatePostDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PutExtensionsUUIDDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PostDomainServiceNameGlueRecordHostUpdateDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PostApplianceUpgradesMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *UpdateMTOServiceItemStatusInternalServerError) Code() int {\n\treturn 500\n}", "func (o *GetNiaapiDcnmSweolsMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *UnixUserSettingsModifyDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PostHclCompatibilityStatusesDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *DeleteApplianceRestoresMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetHyperflexServerFirmwareVersionsDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *DcimRackRolesBulkPartialUpdateDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *AddPMMAgentDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *UpdateDashboardGroupByIDDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *UpdateAgentDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *FormatScoreServiceUpdateDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PostEquipmentIoExpandersMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *ReplaceStickRuleDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetServerConfigImportsMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *EmailTemplatePutDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetComputeRackUnitsMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetVersionDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *UpdateJobJobDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *NvmeSubsystemMapCreateDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *SchemaObjectsUpdateOK) Code() int {\n\treturn 200\n}", "func (o *ReplaceTCPResponseRuleDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *UpdateRecipientGroupByIDDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *UpdateHostGroupsDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *UpdateDeviceTagsDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *CreateAntivirusServerDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetManagementControllersMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *UpdateProjectSettingsDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetModelRegistryDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *UpdateMTOServiceItemStatusUnprocessableEntity) Code() int {\n\treturn 422\n}", "func (o *GetLibcVersionDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *ModifyClusterAppConfigV5Default) Code() int {\n\treturn o._statusCode\n}", "func (o *UpdateHookDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetHyperflexSysConfigPoliciesMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *DescribeServerOK) Code() int {\n\treturn 200\n}", "func (o *UpdateMTOPostCounselingInformationOK) Code() int {\n\treturn 200\n}", "func (o *SchemaObjectsUpdateUnprocessableEntity) Code() int {\n\treturn 422\n}", "func (r *request) StatusCode() int {\n\treturn 400\n}", "func (o *UpdateMTOPostCounselingInformationInternalServerError) Code() int {\n\treturn 500\n}", "func (o *UpdateIngredientDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *DeleteFirmwareUpgradesMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PostPciLinksMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *ForgotPasswordVerifyPutDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetVersionOK) Code() int {\n\treturn 200\n}", "func (o *SchemaObjectsUpdateInternalServerError) Code() int {\n\treturn 500\n}", "func (o *SMSInboundAutomationPostDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *EmsDestinationModifyDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *UpdateFirewallPoliciesInternalServerError) Code() int {\n\treturn 500\n}", "func (o *GetWorkflowWorkflowInfosMoidDefault) Code() int {\n\treturn o._statusCode\n}" ]
[ "0.8179358", "0.7980112", "0.79247344", "0.79240793", "0.7822035", "0.7821128", "0.76780146", "0.7638689", "0.7638088", "0.7634656", "0.7634247", "0.7608735", "0.7581057", "0.75798213", "0.7543436", "0.7512752", "0.74794847", "0.74466586", "0.74387866", "0.74146473", "0.73588693", "0.7351825", "0.73496234", "0.7336531", "0.7328002", "0.7323749", "0.7307719", "0.7299648", "0.72927076", "0.72796106", "0.72772324", "0.72641945", "0.72174317", "0.7157076", "0.71099615", "0.7083572", "0.70715225", "0.70707047", "0.7069142", "0.70646864", "0.7040117", "0.7025691", "0.7020111", "0.7020111", "0.7019455", "0.70136786", "0.70060253", "0.69942796", "0.697803", "0.6975061", "0.69727427", "0.6960383", "0.6958887", "0.69514126", "0.6937485", "0.69207495", "0.6917974", "0.6915712", "0.6905517", "0.68930405", "0.68886775", "0.6887596", "0.688006", "0.68794674", "0.68645173", "0.68627745", "0.6858258", "0.68477196", "0.68472344", "0.6844106", "0.683968", "0.68366253", "0.6832333", "0.6831047", "0.68264896", "0.68224996", "0.68133616", "0.6809574", "0.68078196", "0.6803872", "0.67987865", "0.67830616", "0.67747045", "0.67722255", "0.67663324", "0.67654634", "0.6761491", "0.67604315", "0.6760398", "0.67590034", "0.67565906", "0.6750895", "0.6749637", "0.67459095", "0.6738893", "0.6731478", "0.67313856", "0.6729001", "0.67250794", "0.67190206" ]
0.8719866
0
NewDiscounts returns a Discounts object.
func NewDiscounts() *Discounts { o := Discounts{} o.Args = NewArgs() o.Args.Resource = EndpointDiscounts return &o }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (d *DiscountControllerStruct) NewDiscount(c *gin.Context) {\n\tvar discountInfo domain.DiscountInfo\n\tif err := c.Bind(&discountInfo); err != nil {\n\t\terrorsHandler.GinErrorResponseHandler(c, err)\n\t\treturn\n\t}\n\n\tdiscountNumber := discountInfo.Number\n\tphoneNumber := discountInfo.PhoneNumber\n\n\tresult, err := logic.NewCreateDiscount(c).CreateNewDiscount(phoneNumber, discountNumber)\n\tif err != nil {\n\t\terrorsHandler.GinErrorResponseHandler(c, err)\n\t\treturn\n\t}\n\n\tc.JSON(http.StatusCreated,\n\t\tresult,\n\t)\n\treturn\n}", "func (o *Discounts) NewList() *Args {\n\to.Args.Verb = \"GET\"\n\to.Args.RequiresAuth = true\n\treturn o.Args\n}", "func New() *Disco {\n\treturn NewWithCredentialsSource(nil)\n}", "func NewDicountService(repo repository.Repository, timeout time.Duration) pb.DiscountServiceHandler {\n\treturn &discountService{repo, timeout}\n}", "func (o *AccountDashboardStatistic) SetDiscounts(v []AccountDashboardStatisticDiscount) {\n\to.Discounts = &v\n}", "func newDiscountMutation(c config, op Op, opts ...discountOption) *DiscountMutation {\n\tm := &DiscountMutation{\n\t\tconfig: c,\n\t\top: op,\n\t\ttyp: TypeDiscount,\n\t\tclearedFields: make(map[string]struct{}),\n\t}\n\tfor _, opt := range opts {\n\t\topt(m)\n\t}\n\treturn m\n}", "func NewDollar(amount int64) *DollarImpl {\n\treturn &DollarImpl{\n\t\tAmount: amount,\n\t}\n}", "func (o *AccountDashboardStatistic) GetDiscounts() []AccountDashboardStatisticDiscount {\n\tif o == nil || o.Discounts == nil {\n\t\tvar ret []AccountDashboardStatisticDiscount\n\t\treturn ret\n\t}\n\treturn *o.Discounts\n}", "func NewDiscrete(weights ...float64) (*Discrete, error) {\n\n\ttotal := 0.0\n\tfor _, v := range weights {\n\t\ttotal += v\n\t}\n\n\tL := len(weights)\n\tfreq := make([]float64, L)\n\tfreqcumul := make([]float64, L)\n\n\tfor j, v := range weights {\n\t\tfreq[j] = v / total\n\t}\n\n\tprec := 0.0\n\tfor j, v := range weights {\n\t\tfreq[j] = v / total\n\t\tprec += freq[j]\n\t\tfreqcumul[j] = prec\n\t}\n\n\tdd := &Discrete{\n\t\tWeights: &weights,\n\t\tTotal: total,\n\t\tFreq: &freq,\n\t\tFreqCumul: &freqcumul,\n\t}\n\n\treturn dd, nil\n}", "func New() *Counter {\n\treturn new(Counter)\n}", "func New(shuffleTheDeck bool) Deck {\n\tvar newDeck Deck\n\tfor i := 0; i < numSuits; i++ {\n\t\tfor j := 0; j < numCardsPerSuit; j++ {\n\t\t\tnewDeck.AddCard(buildCard(i, j), false)\n\t\t}\n\t}\n\n\tif shuffleTheDeck {\n\t\tnewDeck.Shuffle()\n\t}\n\n\treturn newDeck\n}", "func New() *deck.Deck {\n\tn := *d\n\treturn &n\n}", "func New(ctx context.Context) (*CrosDisks, error) {\n\tconn, obj, err := dbusutil.Connect(ctx, dbusName, dbusPath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &CrosDisks{conn, obj}, nil\n}", "func GenerateDiscountController(discounter service.Discounter) echo.HandlerFunc {\n\treturn func(c echo.Context) error {\n\n\t\tdiscounter.Clear()\n\t\tif err := c.Bind(discounter); err != nil {\n\t\t\treturn echo.NewHTTPError(http.StatusUnprocessableEntity, \"bind value error\")\n\t\t}\n\n\t\tvalidateInstance := c.Get(\"validate\").(*validator.Validate)\n\t\terr := validateInstance.Struct(discounter)\n\t\tif err != nil {\n\t\t\treturn echo.NewHTTPError(http.StatusUnprocessableEntity, \"validate value error\")\n\t\t}\n\n\t\tdiscountCode, err := discounter.GenerateDiscount(c)\n\t\tif err != nil {\n\t\t\treturn echo.NewHTTPError(http.StatusInternalServerError, \"can not generate discount code\")\n\t\t}\n\n\t\treturn c.JSON(http.StatusCreated, echo.Map{\n\t\t\t\"code\": discountCode,\n\t\t})\n\t}\n}", "func NewCounter() *Counter {\n\treturn &Counter{\n\t\tclient: redis2.RedisClient,\n\t}\n}", "func NewDeposit(ctx *middleware.Context, handler DepositHandler) *Deposit {\n\treturn &Deposit{Context: ctx, Handler: handler}\n}", "func (c *CPU) NewDisassembler() *rcs.Disassembler {\n\tdasm := rcs.NewDisassembler(c.mem, Reader, Formatter())\n\treturn dasm\n}", "func New() *CreatePaymentService {\n\tconn := c.Mysql()\n\n\tuserRepository := *u.Repo(conn)\n\tproductRepository := *p.Repo(conn)\n\ttokenRepository := *t.Repo(conn)\n\tpaymentRepository := *payment.Repo(conn)\n\tdiscountRepository := *d.Repo(conn)\n\temailService := e.Service{}\n\n\treturn &CreatePaymentService{\n\t\tuserRepository,\n\t\tproductRepository,\n\t\ttokenRepository,\n\t\tpaymentRepository,\n\t\tdiscountRepository,\n\t\temailService,\n\t}\n}", "func NewDice(sides int, rand randomock.Randomizer) *Dice {\n\treturn &Dice{sides: sides, rand: rand}\n}", "func NewBudget(name string) *Budget {\n\treturn &Budget{\n\t\tid: fmt.Sprintf(\"BUD-%s-%d\", name, time.Now().Unix()),\n\t\tname: name,\n\t\tmonths: make(map[string]float64),\n\t}\n}", "func New(params *stripe.CouponParams) (*stripe.Coupon, error) {\n\treturn getC().New(params)\n}", "func New(ctx context.Context, namespace string, options ...statsd.Option) context.Context {\n\treturn SetClient(ctx, CreateClient(namespace, options...))\n}", "func New(service *dynamodb.DynamoDB) *Client {\n\treturn &Client{\n\t\tdynamoDBClient: service,\n\t}\n}", "func NewDex(creator Name, staking Coins, description string) *Dex {\n\treturn &Dex{\n\t\tCreator: creator,\n\t\tStaking: staking,\n\t\tDescription: description,\n\t}\n}", "func (c Client) New(params *stripe.TreasuryCreditReversalParams) (*stripe.TreasuryCreditReversal, error) {\n\tcreditreversal := &stripe.TreasuryCreditReversal{}\n\terr := c.B.Call(\n\t\thttp.MethodPost,\n\t\t\"/v1/treasury/credit_reversals\",\n\t\tc.Key,\n\t\tparams,\n\t\tcreditreversal,\n\t)\n\treturn creditreversal, err\n}", "func NewDNSDiscover(targetproject, targetgroup, targetapp, host string, interval time.Duration, crpcport, cgrpcport, webport int) (DI, error) {\n\ttargetfullname, e := name.MakeFullName(targetproject, targetgroup, targetapp)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\tif interval < time.Second {\n\t\tinterval = time.Second * 10\n\t}\n\td := &DnsD{\n\t\ttarget: targetfullname,\n\t\thost: host,\n\t\tcrpcport: crpcport,\n\t\tcgrpcport: cgrpcport,\n\t\twebport: webport,\n\t\tinterval: interval,\n\t\tnotices: make(map[chan *struct{}]*struct{}, 1),\n\t\ttriger: make(chan *struct{}, 1),\n\t}\n\td.triger <- nil\n\td.status = 1\n\tgo d.run()\n\treturn d, nil\n}", "func New(client *http.Client) (*Service, error) {\n\tif client == nil {\n\t\treturn nil, errors.New(\"client is nil\")\n\t}\n\ts := &Service{client: client, BasePath: basePath}\n\ts.Purchases = NewPurchasesService(s)\n\treturn s, nil\n}", "func New(cfg config.Discord) (*Discord, error) {\n\tdg, err := discordgo.New(fmt.Sprintf(\"Bot %s\", cfg.Token))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err = dg.Open(); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Discord{\n\t\tsession: dg,\n\t\ttoken: cfg.Token,\n\t\tchannels: cfg.Channels,\n\t}, nil\n}", "func New(params *CouponParams) (*Coupon, error) {\n\treturn getC().New(params)\n}", "func NewStatsdNew(addr string, options ...statsd.Option) *statsd.Client {\n\tclient, err := statsd.New(addr, options...)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn client\n}", "func NewDiscord(args ...interface{}) *Discord {\n\treturn &Discord{\n\t\targs: args,\n\t\tmessageChan: make(chan Message, 200),\n\t}\n}", "func NewDiscord(args ...interface{}) *Discord {\n\treturn &Discord{\n\t\targs: args,\n\t\tmessageChan: make(chan Message, 200),\n\t}\n}", "func Test_employeeDiscounts(t *testing.T) {\n\trow, err := row()\n\tassert.NoError(t, err)\n\tassert.NotNil(t, row)\n\n\tvar d = coletores.Discount{}\n\tassert.NoError(t, employeeDiscounts(row, &d))\n\tassert.Equal(t, *expectedEmployee().Discounts, d)\n}", "func NewUSD() *USD {\n\tm := USD(0)\n\treturn &m\n}", "func NewDisassemblerBTC() Disassembler { return disassemblerBTC{} }", "func NewCounter() *Counter {\n\treturn &Counter{\n\t\tval: 0,\n\t\tincreaseBy: 0,\n\t}\n}", "func NewList() *List {\n newObj := &List {\n counters : make(map[string]Counter),\n }\n\n return newObj\n}", "func NewCounter() *Counter {\n\treturn &Counter{\n\t\tcount: map[string]int{},\n\t}\n}", "func NewCounter() *Counter {\n\treturn &Counter{}\n}", "func NewCounter() *Counter {\n\treturn &Counter{}\n}", "func NewClient() *Client {\n\tbase := apirequest.New().Client(&http.Client{}).Base(discogsAPI)\n\treturn &Client{\n\t\tapi: base,\n\t\tcurrency: \"USD\",\n\n\t\tArtist: newArtistService(base.New()),\n\t\tLabel: newLabelService(base.New()),\n\t\tMaster: newMasterService(base.New()),\n\t\tSearch: newSearchService(base.New()),\n\t}\n}", "func New(config Config) api.BookService {\n\treturn &service{\n\t\tclient: config.Datastore,\n\t}\n}", "func NewDiscrete(values ...Value) *Discrete {\n\tif len(values) == 0 {\n\t\tpanic(\"diviner.NewDiscrete: no values passed\")\n\t}\n\tkind := values[0].Kind()\n\tfor _, v := range values {\n\t\tif v.Kind() != kind {\n\t\t\tpanic(fmt.Sprintf(\"diviner.NewDiscrete: mixed kinds: %s and %s\", v.Kind(), kind))\n\t\t}\n\t}\n\treturn &Discrete{values, kind}\n}", "func New() *Scraper {\n\tseedURL, _ := env.GetCrawlerVars(env.SeedURL)\n\treturn &Scraper{\n\t\tlock: &sync.RWMutex{},\n\t\tvisitsCount: 0,\n\t\tseedURL: seedURL.(string),\n\t\trequests: make(scrapingRequests, 0),\n\t\tacquiredProducts: make(item.Items, 0),\n\t}\n}", "func (c *Correlation) New(d Dataset, ms []Metric) Insight {\n\treturn &Correlation{dt: d, ms: ms}\n}", "func New(key, secret, symbol, currency string, priority int, fee, availShort, availFunds float64) *Client {\n\treturn &Client{\n\t\tkey: key,\n\t\tsecret: secret,\n\t\tsymbol: symbol,\n\t\tcurrency: currency,\n\t\tpriority: priority,\n\t\tfee: fee,\n\t\tavailShort: availShort,\n\t\tavailFunds: availFunds,\n\t\tcurrencyCode: 0,\n\t\tname: fmt.Sprintf(\"Bitfinex(%s)\", currency),\n\t\tbaseURL: \"https://api.bitfinex.com\",\n\t\tdone: make(chan bool, 1),\n\t}\n}", "func NewCounter(httpClient *http.Client,\n\terrorMessage string,\n\tlogger *log.Logger,\n\toutputPhrase string) *Сounter {\n\treturn &Сounter{\n\t\thttpClient: httpClient,\n\t\terrorMessage: errorMessage,\n\t\tlogger: logger,\n\t\toutputPhrase: outputPhrase,\n\t}\n}", "func NewUserDisco() *UserDisco {\n\tthis := UserDisco{}\n\tvar status UserStatusEnum = USERSTATUSENUM_ACTIVE\n\tthis.Status = &status\n\tvar role UserRoleEnum = USERROLEENUM_USER\n\tthis.Role = &role\n\tvar mode UserModeEnum = USERMODEENUM_SEAL\n\tthis.Mode = &mode\n\treturn &this\n}", "func newDemos(c *DemocontrollerV1alpha1Client, namespace string) *demos {\n\treturn &demos{\n\t\tclient: c.RESTClient(),\n\t\tns: namespace,\n\t}\n}", "func NewCounts() *Counts {\n\treturn &Counts{\n\t\tCounts: map[string]int{},\n\t}\n}", "func New(repo repo.Repo,\n\tprod producer.Producer,\n\tprom prommetrics.PromMetrics,\n\tchunks int,\n) *FeedbackService {\n\treturn &FeedbackService{\n\t\trepo: repo,\n\t\tprod: prod,\n\t\tprom: prom,\n\t\tchunks: chunks}\n}", "func InsertDiscount(c *gin.Context) {\n\trequestBody, _ := ioutil.ReadAll(c.Request.Body)\n\tvar requestObj Discount\n\tjson.Unmarshal(requestBody, &requestObj)\n\n\trequestPoint := &requestObj\n\n\tappEngine := appengine.NewContext(c.Request)\n\tkey := datastore.NewIncompleteKey(appEngine, DiscountKey, SandboxPromotionsKey(appEngine, DiscountKey))\n\n\tif _, err := datastore.Put(appEngine, key, requestPoint); err != nil {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\n\t\t\t\"status_code\": http.StatusBadRequest,\n\t\t\t\"status_message\": err.Error(),\n\t\t})\n\t} else {\n\t\tc.JSON(http.StatusCreated, gin.H{\n\t\t\t\"status_code\": http.StatusCreated,\n\t\t\t\"status_message\": \"Discount created.\",\n\t\t})\n\t}\n\n}", "func New() handler.MetricHandler {\n\treturn &collectdMetricsHandler{}\n}", "func New() *Clac {\n\tc := &Clac{keepHist: true}\n\tc.Reset()\n\treturn c\n}", "func NewCounter() *StandardCounter {\n\treturn &StandardCounter{0}\n}", "func NewDraw(participants, ticketsSold uint32, prize sdk.Coins, endTime time.Time) Draw {\n\treturn Draw{\n\t\tParticipants: participants,\n\t\tTicketsSold: ticketsSold,\n\t\tPrize: prize,\n\t\tEndTime: endTime,\n\t}\n}", "func New(middleware []Middleware) DealService {\n\tvar svc DealService = NewBasicDealService()\n\tfor _, m := range middleware {\n\t\tsvc = m(svc)\n\t}\n\treturn svc\n}", "func New(conn redis.Conn, logger log.Logger) (s CountService) {\n\tvar svc CountService\n\t{\n\t\tsvc = &stubCountService{conn: conn, logger: logger}\n\t\tsvc = LoggingMiddleware(logger)(svc)\n\t}\n\treturn svc\n}", "func New() *Classifier {\n\treturn &Classifier{\n\t\tNewFrequencyStorage(),\n\t}\n}", "func NewCounter(name string, desc string) *Counter {\n\treturn &Counter{name: name, desc: desc, value: 0}\n}", "func (c *DynamoComponent) New(ctx context.Context, conf *DynamoConfig) (domain.Store, error) {\n\tclient, err := c.DynamoDB.New(ctx, conf.AWS.DynamoDBConfig)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &DynamoDB{\n\t\tClient: client,\n\t\tTableName: conf.TableName,\n\t\tPartitionKeyName: conf.PartitionKey,\n\t\tTTLKeyName: conf.TTLKey,\n\t}, nil\n}", "func New(address string, config *Config) (*Consul, error) {\n\tclient, err := api.NewClient(&api.Config{Address: address})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlog.Printf(\"Creating a Consul (@%v) backed service registrator with configuration %+v\\n\", address, config)\n\treturn &Consul{\n\t\tagent: client.Agent(),\n\t\tconfig: config,\n\t\tderegisters: make(map[string]deregisterFunc),\n\t}, nil\n}", "func NewCounter(isNil bool) Counter {\n\tif isNil {\n\t\treturn NilCounter{}\n\t}\n\treturn &StandardCounter{0}\n}", "func newCommandDisconn(cli *client.Client) command {\n\treturn &commandDisconn{\n\t\tcli: cli,\n\t}\n}", "func NewReceipt() *Receipt {\n return &Receipt{}\n}", "func New() Scorecard {\n\treturn make(Scorecard)\n}", "func (c *ChargeClient) New(ctx context.Context, idempotencyKey string, paymentID string, params *ChargeParams, clientInfo *ClientInfo) (*Charge, error) {\n\tcharge := &Charge{}\n\n\theaders := map[string]string{headerIdempotencyKey: idempotencyKey}\n\n\tif clientInfo != nil {\n\t\theaders[headerClientIPAddress] = clientInfo.IPAddress\n\t\theaders[headerClientUserAgent] = clientInfo.UserAgent\n\t}\n\n\tif err := c.Caller.Call(ctx, \"POST\", c.chargesPath(paymentID), headers, params, charge); err != nil {\n\t\treturn nil, err\n\t}\n\treturn charge, nil\n}", "func NewCounter(name string) *Counter {\n\treturn &Counter{name: name}\n}", "func NewDisque(servers []string, cycle int) *Disque {\n\treturn &Disque{\n\t\tservers: servers,\n\t\tcycle: cycle,\n\t\tnodes: make(map[string]string),\n\t\tstats: make(map[string]int),\n\t}\n}", "func NewCounter() *Counter {\n\tc := new(Counter)\n\tc.Reset()\n\treturn c\n}", "func NewK8sDiscover(ctx context.Context, clientset kubernetes.Interface, cfg *option.Conf) Discover {\n\tctx, cancel := context.WithCancel(ctx)\n\treturn &k8sDiscover{\n\t\tctx: ctx,\n\t\tcancel: cancel,\n\t\tclientset: clientset,\n\t\tcfg: cfg,\n\t\tprojects: make(map[string]CallbackUpdate),\n\t}\n}", "func NewDeck(cards ...spanish.Card) *Deck {\n\treturn &Deck{\n\t\tcards: cards,\n\t}\n}", "func New() datastore.Datastore {\n\tctx := context.Background()\n\n\tc, err := fc.New(ctx)\n\tif err != nil {\n\t\tlog.Fatalln(err)\n\t}\n\treturn &client{\n\t\tfirestoreClient: c,\n\t\tctx: ctx,\n\t}\n}", "func New(c service.GapidClient) service.Service {\n\treturn &client{c, func() error { return nil }}\n}", "func NewScrape(cfg *domain.Config) *Scrape {\n\treturn &Scrape{\n\t\tcfg: cfg,\n\t}\n}", "func New(computeAPI ComputeAPI, dnsAPI DNSAPI, removalPredicate IPAddressRemovalPredicate) *Collector {\n\treturn &Collector{computeAPI, dnsAPI, removalPredicate}\n}", "func NewCounter() Counter {\n\tif !Enabled {\n\t\treturn NilCounter{}\n\t}\n\treturn &StandardCounter{0}\n}", "func NewCounter(prefix, name string, value int64, rate float32, tags map[string]string) *Counter {\n\treturn &Counter{\n\t\tPrefix: prefix,\n\t\tName: name,\n\t\tValue: value,\n\t\tRate: rate,\n\t\tTags: tags,\n\t}\n}", "func (cfg Configuration) NewDownsampler(\n\topts DownsamplerOptions,\n) (Downsampler, error) {\n\tagg, err := cfg.newAggregator(opts)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &downsampler{\n\t\topts: opts,\n\t\tagg: agg,\n\t}, nil\n}", "func New() (Scraper, error) {\n\treturn &scraper{}, nil\n}", "func New(namespace, name string, conf fixture.DMConfig) *Ops {\n\treturn &Ops{\n\t\tcli: tests.TestClient.Cli,\n\t\tdm: newDM(namespace, name, conf),\n\t}\n}", "func New(s *store.Store) *Service {\n\treturn &Service{\n\t\tus: client.New(s, &models.Stats{}),\n\t}\n}", "func ListDiscountController(discounter service.Discounter) echo.HandlerFunc {\n\treturn func(c echo.Context) error {\n\n\t\tbearer := c.Request().Header.Get(\"Authorization\")\n\n\t\ttoken, err := service.ValidateToken(bearer[7:])\n\t\tif err != nil {\n\t\t\treturn echo.NewHTTPError(http.StatusForbidden, \"token invalid\")\n\t\t}\n\n\t\tdiscounter.SetParkID(int(token.ID))\n\n\t\tinformation, err := discounter.ListDiscount(c)\n\t\tif err != nil {\n\t\t\treturn echo.NewHTTPError(http.StatusUnprocessableEntity, \"can not list discount\")\n\t\t}\n\n\t\treturn c.JSON(http.StatusOK, information)\n\t}\n}", "func New(db *sql.DB, deliverOrderChan chan CreateReq) Service {\n\treturn Service{\n\t\tdb: db,\n\t\tdeliverOrderChan: deliverOrderChan,\n\t}\n}", "func NewCounter(name string, options ...Option) Counter {\n\treturn newCounter(name, options...)\n}", "func (d *Driver) DiscoverNew(r *sdk.DiscoveryNotification) error {\n\treturn nil\n}", "func New() *Metrics {\n\treturn &Metrics{\n\t\tSectionCounts: make(map[string]int),\n\t}\n}", "func New(db datastore.Datastore) *aggregator {\n\treturn &aggregator{\n\t\tdatabase: db,\n\t}\n}", "func NewService(client cbr.Client, cache cache.Cache) RatesService {\n\treturn ratesService{\n\t\tclient,\n\t\tcache,\n\t}\n}", "func New(m int64, c string) *Money {\n\treturn &Money{m, c}\n}", "func NewDecCoins(coins ...DecCoin) DecCoins {\n\t// remove zeroes\n\tnewCoins := removeZeroDecCoins(coins)\n\tif len(newCoins) == 0 {\n\t\treturn DecCoins{}\n\t}\n\n\tnewCoins.Sort()\n\n\t// detect duplicate Denoms\n\tif dupIndex := findDup(newCoins); dupIndex != -1 {\n\t\tpanic(fmt.Errorf(\"find duplicate denom: %s\", newCoins[dupIndex]))\n\t}\n\n\tif !newCoins.IsValid() {\n\t\tpanic(fmt.Errorf(\"invalid dec coin set: %s\", newCoins))\n\t}\n\n\treturn newCoins\n}", "func New(db *sqlx.DB) promo.Repository {\n\treturn &promoRepository{db}\n}", "func newFreqCount() freqCount {\n\treturn make(freqCount, 3000000)\n}", "func NewOffice365ServicesUserCounts()(*Office365ServicesUserCounts) {\n m := &Office365ServicesUserCounts{\n Entity: *NewEntity(),\n }\n return m\n}", "func New(c *conf.Config, rpcdaos *service.RPCDaos) *Service {\n\ts := &Service{\n\t\tc: c,\n\t\tdm: danmu.New(c),\n\t\tacc: rpcdaos.Acc,\n\t\tarc: rpcdaos.Arc,\n\t\tsub: rpcdaos.Sub,\n\t\telec: elec.New(c),\n\t}\n\treturn s\n}", "func New(d dao.Dao) (s *Service, cf func(), err error) {\n\tstoreRpc,err := storeApi.NewClient(microservice.MicroService)\n\tif err != nil{\n\t\tlogger.Warnf(\"client error:%v\", err)\n\t}\n\ts = &Service{\n\t\tdao: d,\n\t\tstoreRpc:storeRpc,\n\t}\n\tcf = s.Close\n\treturn\n}", "func New(capacity int, client dnsutil.Client) *Cache {\n\treturn NewWithBackend(capacity, client, nil)\n}", "func New(ctx context.Context, config Config) *Client {\n\tclient := &Client{\n\t\tid: generateClientID(\"C\"),\n\t\tvalues: make(map[string]interface{}),\n\t\tevents: make(chan *Event, 64),\n\t\tsends: make(chan string, 64),\n\t\tcapEnabled: make(map[string]bool),\n\t\tcapData: make(map[string]string),\n\t\tconfig: config.WithDefaults(),\n\t\tstatus: &Status{id: generateClientID(\"T\")},\n\t}\n\n\tclient.ctx, client.cancel = context.WithCancel(ctx)\n\n\t_ = client.AddTarget(client.status)\n\n\tgo client.handleEventLoop()\n\tgo client.handleSendLoop()\n\n\tclient.EmitNonBlocking(NewEvent(\"client\", \"create\"))\n\n\treturn client\n}", "func NewCounter() Counter {\n\treturn NewSetCounter(0)\n}", "func NewDisciplineClient(c config) *DisciplineClient {\n\treturn &DisciplineClient{config: c}\n}" ]
[ "0.69762826", "0.6378792", "0.5970773", "0.5854628", "0.57058483", "0.5504026", "0.544882", "0.54114485", "0.52210313", "0.51922053", "0.51892406", "0.518502", "0.5098182", "0.50534093", "0.50098646", "0.50096697", "0.49883297", "0.4977273", "0.49675134", "0.49519634", "0.49436542", "0.4924841", "0.4923247", "0.49046308", "0.48928508", "0.48493823", "0.48374167", "0.48332068", "0.4832592", "0.4820597", "0.48199314", "0.48199314", "0.48194537", "0.48163873", "0.48005646", "0.47944042", "0.47911388", "0.47808406", "0.4779672", "0.4779672", "0.47778773", "0.4775312", "0.47506747", "0.47484198", "0.47441986", "0.47424167", "0.47407904", "0.4736382", "0.4730911", "0.47287315", "0.4723702", "0.47199544", "0.4712988", "0.47127652", "0.47117323", "0.47015992", "0.4696567", "0.46899727", "0.4688105", "0.46857342", "0.468438", "0.46810347", "0.46740165", "0.46737173", "0.46586314", "0.46380198", "0.46376905", "0.46303824", "0.46291783", "0.4623889", "0.46167716", "0.4611105", "0.46092173", "0.46050426", "0.46027103", "0.459573", "0.458499", "0.4584334", "0.45814377", "0.45761696", "0.45747322", "0.45719752", "0.45692572", "0.45607713", "0.455947", "0.45576325", "0.45506933", "0.4549094", "0.454257", "0.4535041", "0.45324478", "0.45302674", "0.45271462", "0.4516633", "0.45158798", "0.45154393", "0.45137557", "0.45127484", "0.45086095", "0.45009294" ]
0.86059636
0
NewList returns the args required for a Discounts GET request.
func (o *Discounts) NewList() *Args { o.Args.Verb = "GET" o.Args.RequiresAuth = true return o.Args }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (g *Goods) ListNew(c Context) {\n\t// TODO\n\tc.String(http.StatusOK, \"get new goods list\")\n}", "func NewDiscounts() *Discounts {\n\to := Discounts{}\n\to.Args = NewArgs()\n\to.Args.Resource = EndpointDiscounts\n\treturn &o\n}", "func (o *DailyBurstUsage) NewList() *Args {\n\to.Args.Verb = \"GET\"\n\to.Args.RequiresAuth = true\n\treturn o.Args\n}", "func (o *Transactions) NewList() *Args {\n\to.Args.Verb = \"GET\"\n\to.Args.RequiresAuth = true\n\treturn o.Args\n}", "func NewList() *List {\n newObj := &List {\n counters : make(map[string]Counter),\n }\n\n return newObj\n}", "func NewListCommand(parent common.Registerer, globals *config.Data) *ListCommand {\n\tvar c ListCommand\n\tc.Globals = globals\n\tc.manifest.File.SetOutput(c.Globals.Output)\n\tc.manifest.File.Read(manifest.Filename)\n\tc.CmdClause = parent.Command(\"list\", \"List Syslog endpoints on a Fastly service version\")\n\tc.CmdClause.Flag(\"service-id\", \"Service ID\").Short('s').StringVar(&c.manifest.Flag.ServiceID)\n\tc.CmdClause.Flag(\"version\", \"Number of service version\").Required().IntVar(&c.Input.ServiceVersion)\n\treturn &c\n}", "func (r *Replenish) List(args *Replenish) (*ReplenishList, error) {\n\taction := \"sales.replenish.list.get\"\n\tdata, err := json.Marshal(args)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbody, err := mafengwo.NewDeals().Fetch(action, data)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresult := ReplenishList{}\n\terr = json.Unmarshal(body, &result)\n\treturn &result, err\n}", "func ListDiscountController(discounter service.Discounter) echo.HandlerFunc {\n\treturn func(c echo.Context) error {\n\n\t\tbearer := c.Request().Header.Get(\"Authorization\")\n\n\t\ttoken, err := service.ValidateToken(bearer[7:])\n\t\tif err != nil {\n\t\t\treturn echo.NewHTTPError(http.StatusForbidden, \"token invalid\")\n\t\t}\n\n\t\tdiscounter.SetParkID(int(token.ID))\n\n\t\tinformation, err := discounter.ListDiscount(c)\n\t\tif err != nil {\n\t\t\treturn echo.NewHTTPError(http.StatusUnprocessableEntity, \"can not list discount\")\n\t\t}\n\n\t\treturn c.JSON(http.StatusOK, information)\n\t}\n}", "func NewList(client *secretsapi.Client, p listPrimeable) *List {\n\treturn &List{\n\t\tsecretsClient: client,\n\t\tout: p.Output(),\n\t\tproj: p.Project(),\n\t}\n}", "func newListFromUIDs(uids []string) *CSPList {\n\treturn NewListBuilder().WithUIDs(uids...).List()\n}", "func NewList(args ...interface{}) *List {\n\tl := List{}\n\tfor _, data := range args {\n\t\tl.PushBack(data)\n\t}\n\treturn &l\n}", "func (ctl *SaleCounterProductController) GetList() {\n\tviewType := ctl.Input().Get(\"view\")\n\tif viewType == \"\" || viewType == \"table\" {\n\t\tctl.Data[\"ViewType\"] = \"table\"\n\t}\n\tctl.PageAction = utils.MsgList\n\tctl.Data[\"tableId\"] = \"table-sale-counter-product\"\n\tctl.Layout = \"base/base_list_view.html\"\n\tctl.TplName = \"sale/sale_counter_product_list_search.html\"\n}", "func newListFormulaArg(l []formulaArg) formulaArg {\n\treturn formulaArg{Type: ArgList, List: l}\n}", "func newList() *List {\n\tl := &List{\n\t\tch: make(chan sh.QData),\n\t}\n\treturn l\n}", "func (client *CloudServicesClient) listAllCreateRequest(ctx context.Context, options *CloudServicesClientListAllOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/providers/Microsoft.Compute/cloudServices\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2022-09-04\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func (c *ChargeClient) GetList(ctx context.Context, paymentID string) ([]Charge, error) {\n\tvar charges []Charge\n\tif err := c.Caller.Call(ctx, \"GET\", c.chargesPath(paymentID), nil, nil, &charges); err != nil {\n\t\treturn nil, err\n\t}\n\treturn charges, nil\n}", "func (rm *resourceManager) newListRequestPayload(\n\tr *resource,\n) (*svcsdk.DescribeVpnConnectionsInput, error) {\n\tres := &svcsdk.DescribeVpnConnectionsInput{}\n\n\tif r.ko.Spec.DryRun != nil {\n\t\tres.SetDryRun(*r.ko.Spec.DryRun)\n\t}\n\n\treturn res, nil\n}", "func NewListCommand() cli.Command {\n\treturn newListCommand(defaultEnv, newClients)\n}", "func newListCmd(clientFn func() (*fic.ServiceClient, error), out io.Writer) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"list\",\n\t\tShort: \"List ports\",\n\t\tExample: \"fic ports list\",\n\t\tArgs: cobra.NoArgs,\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tclient, err := clientFn()\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"creating FIC client: %w\", err)\n\t\t\t}\n\n\t\t\tpages, err := ports.List(client, nil).AllPages()\n\t\t\tif err != nil {\n\t\t\t\tvar e *json.UnmarshalTypeError\n\t\t\t\tif errors.As(err, &e) {\n\t\t\t\t\treturn fmt.Errorf(\"extracting ports from API response: %w\", err)\n\t\t\t\t}\n\t\t\t\treturn fmt.Errorf(\"calling List ports API: %w\", err)\n\t\t\t}\n\n\t\t\tps, _ := ports.ExtractPorts(pages)\n\n\t\t\tt := utils.NewTabby(out)\n\t\t\tt.AddHeader(\"id\", \"name\", \"operationStatus\", \"isActivated\", \"vlanRanges\", \"tenantID\", \"switchName\",\n\t\t\t\t\"portType\", \"location\", \"area\")\n\t\t\tfor _, p := range ps {\n\t\t\t\tt.AddLine(p.ID, p.Name, p.OperationStatus, p.IsActivated, p.VLANRanges, p.TenantID, p.SwitchName,\n\t\t\t\t\tp.PortType, p.Location, p.Area)\n\t\t\t}\n\t\t\tt.Print()\n\n\t\t\treturn nil\n\t\t},\n\t}\n\n\treturn cmd\n}", "func NewListCommand() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"list\",\n\t\tShort: \"List the proxies\",\n\t\tRun: listCommandFunc,\n\t}\n\n\treturn cmd\n}", "func newCoverageList(name string) *CoverageList {\n\treturn &CoverageList{\n\t\tCoverage: &Coverage{Name: name},\n\t\tGroup: []Coverage{},\n\t}\n}", "func (ctl *SaleCounterProductController) PostList() {\n\tquery := make(map[string]interface{})\n\texclude := make(map[string]interface{})\n\tcond := make(map[string]map[string]interface{})\n\tcondAnd := make(map[string]interface{})\n\tcondOr := make(map[string]interface{})\n\tfilterMap := make(map[string]interface{})\n\tfields := make([]string, 0, 0)\n\tsortby := make([]string, 0, 1)\n\torder := make([]string, 0, 1)\n\tif ID, err := ctl.GetInt64(\"Id\"); err == nil {\n\t\tquery[\"Id\"] = ID\n\t}\n\n\tfilter := ctl.GetString(\"filter\")\n\tif filter != \"\" {\n\t\tjson.Unmarshal([]byte(filter), &filterMap)\n\t}\n\n\tif len(condAnd) > 0 {\n\t\tcond[\"and\"] = condAnd\n\t}\n\tif len(condOr) > 0 {\n\t\tcond[\"or\"] = condOr\n\t}\n\toffset, _ := ctl.GetInt64(\"offset\")\n\tlimit, _ := ctl.GetInt64(\"limit\")\n\torderStr := ctl.GetString(\"order\")\n\tsortStr := ctl.GetString(\"sort\")\n\tif orderStr != \"\" && sortStr != \"\" {\n\t\tsortby = append(sortby, sortStr)\n\t\torder = append(order, orderStr)\n\t} else {\n\t\tsortby = append(sortby, \"Id\")\n\t\torder = append(order, \"desc\")\n\n\t}\n\tif result, err := ctl.counterProductList(query, exclude, cond, fields, sortby, order, offset, limit); err == nil {\n\t\tctl.Data[\"json\"] = result\n\t}\n\tctl.ServeJSON()\n\n}", "func NewList(g ...Getter) *List {\n\tlist := &List{\n\t\tlist: g,\n\t}\n\tlist.GetProxy = NewGetProxy(list) // self\n\treturn list\n}", "func NewListCmd(f *cmdutil.Factory) *ListCmd {\n\tccmd := &ListCmd{\n\t\tfactory: f,\n\t}\n\tcmd := &cobra.Command{\n\t\tUse: \"list\",\n\t\tShort: \"Get configuration collection\",\n\t\tLong: `Get a collection of configuration (managedObjects) based on filter parameters`,\n\t\tExample: heredoc.Doc(`\n$ c8y configuration list\nGet a list of configuration files\n `),\n\t\tPreRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\treturn nil\n\t\t},\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\thandler := c8yquerycmd.NewInventoryQueryRunner(\n\t\t\t\tcmd,\n\t\t\t\targs,\n\t\t\t\tccmd.factory,\n\t\t\t\tflags.WithC8YQueryFixedString(\"(type eq 'c8y_ConfigurationDump')\"),\n\t\t\t\tflags.WithC8YQueryFormat(\"name\", \"(name eq '%s')\"),\n\t\t\t\tflags.WithC8YQueryFormat(\"deviceType\", \"(c8y_Filter.type eq '%s')\"),\n\t\t\t\tflags.WithC8YQueryFormat(\"description\", \"(description eq '%s')\"),\n\t\t\t)\n\t\t\treturn handler()\n\t\t},\n\t}\n\n\tcmd.SilenceUsage = true\n\n\tcmd.Flags().String(\"name\", \"\", \"Configuration name filter\")\n\tcmd.Flags().String(\"description\", \"\", \"Configuration description filter\")\n\tcmd.Flags().String(\"deviceType\", \"\", \"Configuration device type filter\")\n\n\tcompletion.WithOptions(\n\t\tcmd,\n\t)\n\n\tflags.WithOptions(\n\t\tcmd,\n\t\tflags.WithCommonCumulocityQueryOptions(),\n\t\tflags.WithExtendedPipelineSupport(\"query\", \"query\", false, \"c8y_DeviceQueryString\"),\n\t\tflags.WithCollectionProperty(\"managedObjects\"),\n\t)\n\n\t// Required flags\n\n\tccmd.SubCommand = subcommand.NewSubCommand(cmd)\n\n\treturn ccmd\n}", "func GetClientListCmd() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"list\",\n\t\tShort: \"List price entries\",\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\tlogger := initLogger()\n\n\t\t\t// parse inputs\n\t\t\tpageSkip, pageLimit := parseIntFlag(logger, flagPageSkip, cmd.Flags()), parseIntFlag(logger, flagPageLimit, cmd.Flags())\n\t\t\tsortByName, sortByPrice, sortByTimestamp := parseSortFlag(flagSortByName, cmd.Flags()), parseSortFlag(flagSortByPrice, cmd.Flags()), parseSortFlag(flagSortByTimestamp, cmd.Flags())\n\n\t\t\t// create gRPC client\n\t\t\tconn, err := createGRPCClientConnection(logger)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Fatalf(err.Error())\n\t\t\t}\n\t\t\tdefer conn.Close()\n\n\t\t\tclient := v1.NewPriceEntryReaderClient(conn)\n\n\t\t\t// request\n\t\t\trequestCtx, requestCancel := context.WithTimeout(context.Background(), 1*time.Second)\n\t\t\tdefer requestCancel()\n\t\t\tresp, err := client.List(requestCtx, &v1.ListRequest{\n\t\t\t\tPagination: &v1.PaginationParams{\n\t\t\t\t\tSkip: uint32(pageSkip),\n\t\t\t\t\tLimit: uint32(pageLimit),\n\t\t\t\t},\n\t\t\t\tSortByName: sortByName,\n\t\t\t\tSortByPrice: sortByPrice,\n\t\t\t\tSortByTimestamp: sortByTimestamp,\n\t\t\t})\n\t\t\tif err != nil {\n\t\t\t\tlogger.Fatalf(\"request failed: %v\", err)\n\t\t\t}\n\n\t\t\t// print result\n\t\t\tif len(resp.Entries) == 0 {\n\t\t\t\tlogger.Infof(\"no entries found\")\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tfor _, entry := range resp.Entries {\n\t\t\t\tlogger.Infof(\"%s\\t->\\t%s\\t->\\t%s\",\n\t\t\t\t\tentry.ProductName,\n\t\t\t\t\tstrconv.FormatInt(int64(entry.Price), 10),\n\t\t\t\t\ttime.Unix(entry.Timestamp, 0).Format(time.RFC3339),\n\t\t\t\t)\n\t\t\t}\n\t\t},\n\t}\n\tcmd.Flags().Int(flagPageSkip, 0, \"(optional) pagination param: skip\")\n\tcmd.Flags().Int(flagPageLimit, 50, \"(optional) pagination param: limit\")\n\tcmd.Flags().String(flagSortByName, \"\", \"(optional) sort param: by product name (ASC/DESC)\")\n\tcmd.Flags().String(flagSortByPrice, \"\", \"(optional) sort param: by price (ASC/DESC)\")\n\tcmd.Flags().String(flagSortByTimestamp, \"\", \"(optional) sort param: by timestamp (ASC/DESC)\")\n\n\treturn cmd\n}", "func NewListOpts(validator ValidatorFctType) ListOpts {\n var values []string\n return *NewListOptsRef(&values, validator)\n}", "func (client *ChargesClient) listCreateRequest(ctx context.Context, scope string, options *ChargesClientListOptions) (*policy.Request, error) {\n\turlPath := \"/{scope}/providers/Microsoft.Consumption/charges\"\n\turlPath = strings.ReplaceAll(urlPath, \"{scope}\", scope)\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-10-01\")\n\tif options != nil && options.StartDate != nil {\n\t\treqQP.Set(\"startDate\", *options.StartDate)\n\t}\n\tif options != nil && options.EndDate != nil {\n\t\treqQP.Set(\"endDate\", *options.EndDate)\n\t}\n\tif options != nil && options.Filter != nil {\n\t\treqQP.Set(\"$filter\", *options.Filter)\n\t}\n\tif options != nil && options.Apply != nil {\n\t\treqQP.Set(\"$apply\", *options.Apply)\n\t}\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func (client *ManagedClustersClient) listCreateRequest(ctx context.Context, options *ManagedClustersClientListOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/providers/Microsoft.ContainerService/managedClusters\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2020-11-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func NewList(args ...interface{}) *List {\n\tl := &List{}\n\tfor _, v := range args {\n\t\tl.PushBack(v)\n\t}\n\treturn l\n}", "func NewList()(*List) {\n m := &List{\n BaseItem: *NewBaseItem(),\n }\n odataTypeValue := \"#microsoft.graph.list\";\n m.SetOdataType(&odataTypeValue);\n return m\n}", "func NewList(list uint32, mode uint32) {\n\tsyscall.Syscall(gpNewList, 2, uintptr(list), uintptr(mode), 0)\n}", "func (w *ClusterDynamicClient) List(opts metav1.ListOptions) (*unstructured.UnstructuredList, error) {\n\treturn w.dClient.Resource(w.resource).Namespace(w.namespace).List(w.ctx, opts)\n}", "func (s *podDisruptionBudgetLister) List(selector labels.Selector) (ret []*policyv1.PodDisruptionBudget, err error) {\n\terr = kcpcache.ListAllByCluster(s.indexer, s.clusterName, selector, func(i interface{}) {\n\t\tret = append(ret, i.(*policyv1.PodDisruptionBudget))\n\t})\n\treturn ret, err\n}", "func NewListCmd(f *cmdutil.Factory) *ListCmd {\n\tccmd := &ListCmd{\n\t\tfactory: f,\n\t}\n\tcmd := &cobra.Command{\n\t\tUse: \"list\",\n\t\tShort: \"Get managed object collection\",\n\t\tLong: `Get a collection of managedObjects based on filter parameters`,\n\t\tExample: heredoc.Doc(`\n$ c8y inventory list\nGet a list of managed objects\n\n$ c8y inventory list --ids 1111,2222\nGet a list of managed objects by ids\n\n$ echo 'myType' | c8y inventory list\nSearch by type using pipeline. piped input will be mapped to type parameter\n\n$ c8y inventory get --id 1234 | c8y inventory list\nGet managed objects which have the same type as the managed object id=1234. piped input will be mapped to type parameter\n `),\n\t\tPreRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\treturn nil\n\t\t},\n\t\tRunE: ccmd.RunE,\n\t}\n\n\tcmd.SilenceUsage = true\n\n\tcmd.Flags().StringSlice(\"ids\", []string{\"\"}, \"List of ids.\")\n\tcmd.Flags().String(\"type\", \"\", \"ManagedObject type. (accepts pipeline)\")\n\tcmd.Flags().String(\"fragmentType\", \"\", \"ManagedObject fragment type.\")\n\tcmd.Flags().String(\"text\", \"\", \"managed objects containing a text value starting with the given text (placeholder {text}). Text value is any alphanumeric string starting with a latin letter (A-Z or a-z).\")\n\tcmd.Flags().Bool(\"withParents\", false, \"include a flat list of all parents and grandparents of the given object\")\n\tcmd.Flags().Bool(\"skipChildrenNames\", false, \"Don't include the child devices names in the response. This can improve the API response because the names don't need to be retrieved\")\n\n\tcompletion.WithOptions(\n\t\tcmd,\n\t)\n\n\tflags.WithOptions(\n\t\tcmd,\n\n\t\tflags.WithExtendedPipelineSupport(\"type\", \"type\", false, \"type\"),\n\t\tflags.WithCollectionProperty(\"managedObjects\"),\n\t)\n\n\t// Required flags\n\n\tccmd.SubCommand = subcommand.NewSubCommand(cmd)\n\n\treturn ccmd\n}", "func newList(ctx TransactionContextInterface) *list {\n\t stateList := new(ledgerapi.StateList)\n\t stateList.Ctx = ctx\n\t stateList.Class = \"Asset\"\n\t stateList.Deserialize = func(bytes []byte, state ledgerapi.StateInterface) error {\n\t\t return Deserialize(bytes, state.(*Asset))\n\t }\n \n\t list := new(list)\n\t list.stateList = stateList\n \n\t return list\n }", "func (s *podDisruptionBudgetNamespaceLister) List(selector labels.Selector) (ret []*policyv1.PodDisruptionBudget, err error) {\n\terr = kcpcache.ListAllByClusterAndNamespace(s.indexer, s.clusterName, s.namespace, selector, func(i interface{}) {\n\t\tret = append(ret, i.(*policyv1.PodDisruptionBudget))\n\t})\n\treturn ret, err\n}", "func (client *ApplyUpdatesClient) listCreateRequest(ctx context.Context, options *ApplyUpdatesClientListOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/providers/Microsoft.Maintenance/applyUpdates\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2023-04-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func (client *ManagedInstancesClient) listCreateRequest(ctx context.Context, options *ManagedInstancesClientListOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/providers/Microsoft.Sql/managedInstances\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\tif options != nil && options.Expand != nil {\n\t\treqQP.Set(\"$expand\", *options.Expand)\n\t}\n\treqQP.Set(\"api-version\", \"2022-08-01-preview\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func NewList() *List {\n\tl := List{\n\t\tpostings: make(map[uint64]*Posting),\n\t}\n\treturn &l\n}", "func (client *ClustersClient) listCreateRequest(ctx context.Context, options *ClustersListOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/providers/Microsoft.OperationalInsights/clusters\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-06-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func (h *Handle) NewDBList() IDBList {\n\treturn &DBList{nil, *h}\n}", "func (client *CapacitiesClient) listCreateRequest(ctx context.Context, options *CapacitiesClientListOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/providers/Microsoft.PowerBIDedicated/capacities\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-01-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func createList(arg string) []string {\n\tvar retObject = []string{arg}\n\treturn retObject\n}", "func GetList(c Context) {\n\tres, err := db.SelectAllItems()\n\tif err != nil {\n\t\tc.JSON(http.StatusInternalServerError, err)\n\t\treturn\n\t}\n\tc.JSON(http.StatusOK, res)\n}", "func (p *PostgresDb) GetAllDefaultersNew() ([]*model.Defaulter, error) {\n\tdefaulters := []*model.Defaulter{}\n\n\tall, err := p.GetAllCustomers()\n\tif err != nil {\n\t\tlog.Println(\"Error getting all customers: \", err)\n\t\treturn defaulters, err\n\t}\n\n\tpotentials := []*model.Defaulter{}\n\tfor _, c := range all {\n\t\tdue, err := p.GetDueAmount(c.ID)\n\t\tif err != nil {\n\t\t\tlog.Println(\"Error getting due amount for customer:\", err)\n\t\t\treturn defaulters, err\n\t\t}\n\t\tif due > float64(c.CreditLimit) {\n\t\t\tc.DueAmount = due\n\t\t\tlastCreditDate, err := p.getLastCreditDateByCreditor(*c)\n\t\t\tif err != nil {\n\t\t\t\treturn defaulters, err\n\t\t\t}\n\t\t\tlastPayDate, err := p.getLastPayDateByCreditor(*c)\n\t\t\tif err != nil {\n\t\t\t\treturn defaulters, err\n\t\t\t}\n\t\t\td := model.Defaulter{FullName: c.FullName, SearchName: c.SearchName,\n\t\t\t\tDeliveryRoute: c.DeliveryRoute, DueAmount: c.DueAmount,\n\t\t\t\tDueFrom: lastCreditDate,\n\t\t\t\tLastPaidOn: lastPayDate,\n\t\t\t\tDefaultCause: \"Total due amount exceeds credit limit.\",\n\t\t\t}\n\t\t\tpotentials = append(potentials, &d)\n\t\t}\n\t}\n\n\tquery := `\n\t\tSELECT DISTINCT ON (customer_id)\n\t\t\tcredit.customer_id, \n\t\t\tcredit.date, \n\t\t\tcredit.amount,\n\t\t\tcustomer.pay_cycle, \n\t\t\tcustomer.full_name, \n\t\t\tcustomer.search_name,\n\t\t\tcustomer.delivery_route\n\t\tFROM credit\n\t\tINNER JOIN customer on credit.customer_id = customer.id;\n\t`\n\n\trows, err := p.dbConn.Query(query)\n\tif err != nil {\n\t\tif err == sql.ErrNoRows {\n\t\t\tlog.Printf(\"no payments found for customer id: %v \\n\")\n\t\t}\n\t\treturn defaulters, err\n\t}\n\tdefer rows.Close()\n\n\tfor rows.Next() {\n\t\tvar d model.Defaulter\n\t\tif err := rows.Scan(&d.ID, &d.DueFrom, &d.LatestCredit, &d.PayCycle, &d.FullName, &d.SearchName, &d.DeliveryRoute); err != nil {\n\t\t\tlog.Panicln(\"error scanning payment row:\", err)\n\t\t}\n\t\tdate, err := time.Parse(time.RFC3339, d.DueFrom)\n\t\tif err != nil {\n\t\t\tlog.Println(\"error parsing date: \", err)\n\t\t}\n\t\td.DueFrom = date.Format(\"02-01-2006\")\n\n\t\tpotentials = append(potentials, &d)\n\t}\n\n\tquery = `\n\t\tSELECT \n COALESCE(SUM (amount), 0)\n FROM\n payment\n WHERE TRUE\n AND customer_id=$1\n AND amount IN (\n SELECT \n amount\n FROM\n payment\n WHERE date > $2\n )\n GROUP BY\n customer_id \n\t`\n\tnow := time.Now()\n\tweekAgo := now.AddDate(0, 0, -7)\n\tweekAgoStr := weekAgo.Format(\"2006-01-02\")\n\tfor _, potential := range potentials {\n\t\tvar paymentsInCycle float64\n\t\terr := p.dbConn.Get(&paymentsInCycle, query, potential.ID, weekAgoStr)\n\t\tif err != nil {\n\t\t\tlog.Println(\"Error getting sum payments in pay cycle:\", err)\n\t\t}\n\t\tpotential.PaymentInCycle = paymentsInCycle\n\t\tif potential.PaymentInCycle < potential.LatestCredit {\n\t\t\tpotential.DefaultCause = \"Credit not repaid within payment cycle\"\n\t\t\tdue, err := p.GetDueAmount(potential.ID)\n\t\t\tif err != nil {\n\t\t\t\tlog.Println(\"Error getting due amount for customer:\", err)\n\t\t\t\treturn defaulters, err\n\t\t\t}\n\t\t\tpotential.DueAmount = due\n\t\t\tdefaulters = append(defaulters, potential)\n\t\t}\n\t}\n\n\treturn defaulters, nil\n}", "func NewBehatGetList(variables templateUtils.TemplateVariables) Template {\n\trawTemplate, err := template.New(\"behat_get_list\").Parse(BehatGetListTemplate)\n\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\treturn New(resource.New(geography.BehatDir+string(variables.Entity)+\"/crud/\", \"get_list.feature\"),\n\t\trawTemplate, variables)\n}", "func NewListCommand(banzaiCli cli.Cli) *cobra.Command {\n\toptions := listOptions{}\n\n\tcmd := &cobra.Command{\n\t\tUse: \"list\",\n\t\tShort: \"List secrets\",\n\t\tArgs: cobra.NoArgs,\n\t\tAliases: []string{\"l\", \"ls\"},\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\toptions.format, _ = cmd.Flags().GetString(\"output\")\n\t\t\trunList(banzaiCli, options)\n\t\t},\n\t}\n\n\tflags := cmd.Flags()\n\n\tflags.StringVarP(&options.secretType, \"type\", \"t\", \"\", \"Filter list to the given type\")\n\n\treturn cmd\n}", "func newIDList(p *idElementPool) *idList {\n\tl := &idList{Pool: p}\n\treturn l.Init()\n}", "func NewList() List {\n\treturn List{}\n}", "func (a *CloudCostBudgetsApiService) ListBudgets(ctx context.Context, accountIdentifier string) (ResponseDtoListBudget, *http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Get\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue ResponseDtoListBudget\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/ccm/api/budgets\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\tlocalVarQueryParams.Add(\"accountIdentifier\", parameterToString(accountIdentifier, \"\"))\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tif ctx != nil {\n\t\t// API Key Authentication\n\t\tif auth, ok := ctx.Value(ContextAPIKey).(APIKey); ok {\n\t\t\tvar key string\n\t\t\tif auth.Prefix != \"\" {\n\t\t\t\tkey = auth.Prefix + \" \" + auth.Key\n\t\t\t} else {\n\t\t\t\tkey = auth.Key\n\t\t\t}\n\t\t\tlocalVarHeaderParams[\"x-api-key\"] = key\n\n\t\t}\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t}\n\n\tif localVarHttpResponse.StatusCode < 300 {\n\t\t// If we succeed, return the data, otherwise pass on to decode error.\n\t\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\t\tif err == nil {\n\t\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t\t}\n\t}\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\tif localVarHttpResponse.StatusCode == 400 {\n\t\t\tvar v Failure\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t}\n\t\tif localVarHttpResponse.StatusCode == 500 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t}\n\t\tif localVarHttpResponse.StatusCode == 0 {\n\t\t\tvar v ResponseDtoListBudget\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t}\n\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHttpResponse, nil\n}", "func (f *DMAClient) NewListWatch() *cache.ListWatch {\n\treturn cache.NewListWatchFromClient(f.client, f.plural, f.namespace, fields.Everything())\n}", "func getFactListHandler(cliCtx context.CLIContext, storeName string) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tres, _, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/getFactList\", storeName), nil)\n\t\tif err != nil {\n\t\t\trest.WriteErrorResponse(w, http.StatusNotFound, err.Error())\n\t\t\treturn\n\t\t}\n\n\t\trest.PostProcessResponse(w, cliCtx, res)\n\t}\n}", "func New(values ...interface{}) *List {\n\tlist := &List{}\n\tif len(values) > 0 {\n\t\tlist.Add(values)\n\t}\n\treturn list\n}", "func (s *DealService) List(ctx context.Context) (*DealsResponse, *Response, error) {\n\treq, err := s.client.NewRequest(http.MethodGet, \"/deals\", nil, nil)\n\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tvar record *DealsResponse\n\n\tresp, err := s.client.Do(ctx, req, &record)\n\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\n\treturn record, resp, nil\n}", "func ListCommand() cli.Command {\n\treturn cli.Command{\n\t\tName: \"list\",\n\t\tAliases: []string{\"l\"},\n\t\tUsage: \"List all invoices.\",\n\t\tDescription: \"Gets all records of database and lists them.\",\n\t\tAction: listAction,\n\t\tFlags: []cli.Flag{\n\t\t\tcli.StringFlag{\n\t\t\t\tName: \"format,f\",\n\t\t\t\tUsage: `list format; \n\tformats: (defualt is \"simple\")\n\t\t\t\"simple\" (or \"s\") \"brief\" (or \"b\"), \"pretty(or \"p\")\"`,\n\t\t\t},\n\t\t},\n\t}\n}", "func NewList(capacity int) *List {\n\tif capacity < 0 {\n\t\tpanic(\"negative capacity\")\n\t}\n\n\tpids := make([]ID, capacity)\n\tfor i := 0; i < capacity; i++ {\n\t\tpids[i] = ID(i)\n\t}\n\n\treturn &List{\n\t\tentities: make([]*Entity, capacity+1), // + 1 due to the 'highest' variable\n\n\t\tlowest: -1,\n\t\thighest: -1,\n\n\t\tavailableIDs: pids,\n\t}\n}", "func New(values ...interface{}) *List {\n\tlist := &List{}\n\tif len(values) > 0 {\n\t\tlist.Add(values...)\n\t}\n\treturn list\n}", "func CreateList(request *api.Request, title string) *CreateListRequest {\n\trequest.Add(\"title\", title)\n\n\treturn &CreateListRequest{\n\t\trequest: request,\n\t}\n}", "func NewList(kubeClient kubernetes.Interface, appConfig config.Config, items ...v1.Ingress) *List {\n\treturn &List{\n\t\tkubeClient: kubeClient,\n\t\tappConfig: appConfig,\n\t\titems: items,\n\t}\n}", "func (rm *resourceManager) newListRequestPayload(\n\tr *resource,\n) (*svcsdk.DescribeReservedInstancesListingsInput, error) {\n\tres := &svcsdk.DescribeReservedInstancesListingsInput{}\n\n\tif r.ko.Spec.ReservedInstancesID != nil {\n\t\tres.SetReservedInstancesId(*r.ko.Spec.ReservedInstancesID)\n\t}\n\n\treturn res, nil\n}", "func runList(cmd *cobra.Command, args []string) error {\n\tverb := \"GET\"\n\turl := \"/v1/query\"\n\n\tresp, err := web.Request(cmd, verb, url, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tcmd.Printf(\"\\n%s\\n\\n\", resp)\n\treturn nil\n}", "func newDiscountMutation(c config, op Op, opts ...discountOption) *DiscountMutation {\n\tm := &DiscountMutation{\n\t\tconfig: c,\n\t\top: op,\n\t\ttyp: TypeDiscount,\n\t\tclearedFields: make(map[string]struct{}),\n\t}\n\tfor _, opt := range opts {\n\t\topt(m)\n\t}\n\treturn m\n}", "func NewListOpts() []func(*ListOptions) {\n\treturn make([]func(*ListOptions), 0)\n}", "func (fkw *FakeClientWrapper) List(ctx context.Context, list runtime.Object, opts ...k8sCl.ListOption) error {\n\tif fkw.shouldPatchNS(list) {\n\t\topts = fkw.removeNSFromListOptions(opts)\n\t}\n\treturn fkw.client.List(ctx, list, opts...)\n}", "func (client *KustoOperationsClient) listCreateRequest(ctx context.Context, options *KustoOperationsClientListOptions) (*policy.Request, error) {\n\turlPath := \"/providers/Microsoft.Synapse/kustooperations\"\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-06-01-preview\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func (client *RecommendationsClient) listCreateRequest(ctx context.Context, options *RecommendationsClientListOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/providers/Microsoft.Web/recommendations\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\tif options != nil && options.Featured != nil {\n\t\treqQP.Set(\"featured\", strconv.FormatBool(*options.Featured))\n\t}\n\treqQP.Set(\"api-version\", \"2021-03-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\tunencodedParams := []string{req.Raw().URL.RawQuery}\n\tif options != nil && options.Filter != nil {\n\t\tunencodedParams = append(unencodedParams, \"$filter=\"+*options.Filter)\n\t}\n\treq.Raw().URL.RawQuery = strings.Join(unencodedParams, \"&\")\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func NewList() List {\n\tl := List{}\n\tl.Set = make(map[string]int)\n\treturn l\n}", "func newListServicesCmd(options *edgeOptions) *cobra.Command {\n\tvar asIdentity string\n\tvar configTypes []string\n\tvar roleFilters []string\n\tvar roleSemantic string\n\n\tcmd := &cobra.Command{\n\t\tUse: \"services <filter>?\",\n\t\tShort: \"lists services managed by the Ziti Edge Controller\",\n\t\tLong: \"lists services managed by the Ziti Edge Controller\",\n\t\tArgs: cobra.MaximumNArgs(1),\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\toptions.Cmd = cmd\n\t\t\toptions.Args = args\n\t\t\terr := runListServices(asIdentity, configTypes, roleFilters, roleSemantic, options)\n\t\t\tcmdhelper.CheckErr(err)\n\t\t},\n\t\tSuggestFor: []string{},\n\t}\n\n\t// allow interspersing positional args and flags\n\tcmd.Flags().SetInterspersed(true)\n\tcmd.Flags().StringVar(&asIdentity, \"as-identity\", \"\", \"Allow admins to see services as they would be seen by a different identity\")\n\tcmd.Flags().StringSliceVar(&configTypes, \"config-types\", nil, \"Override which config types to view on services\")\n\tcmd.Flags().StringSliceVar(&roleFilters, \"role-filters\", nil, \"Allow filtering by roles\")\n\tcmd.Flags().StringVar(&roleSemantic, \"role-semantic\", \"\", \"Specify which roles semantic to use \")\n\toptions.AddCommonFlags(cmd)\n\n\treturn cmd\n}", "func NewListRequest() *todopb.ListRequest {\n\tmessage := &todopb.ListRequest{}\n\treturn message\n}", "func GetCmdGetFactList(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"get-factlist\",\n\t\tShort: \"get-factlist\",\n\t\tArgs: cobra.ExactArgs(0),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\tres, _, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/getFactList\", queryRoute), nil)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Printf(\"Error: %s\", err)\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\tvar out types.QueryResFactList\n\t\t\tcdc.MustUnmarshalJSON(res, &out)\n\t\t\treturn cliCtx.PrintOutput(out)\n\t\t},\n\t}\n}", "func (g *Goods) List(c Context) {\n\t// TODO\n\tc.String(http.StatusOK, \"get goods list\")\n}", "func (s *democrdLister) List(selector labels.Selector) (ret []*v1.Democrd, err error) {\n\terr = cache.ListAll(s.indexer, selector, func(m interface{}) {\n\t\tret = append(ret, m.(*v1.Democrd))\n\t})\n\treturn ret, err\n}", "func NewListResourcesCommand() *cobra.Command {\n\t// Use Discovery's function for listing resources\n\tcmd := discovery.NewQueryDiscoveryCommand()\n\t// Change use for consistency (cl resource list vs cl resource query)\n\tcmd.Use = \"list\"\n\n\treturn cmd\n}", "func NewListCommand(c Command, run RunListFunc, subCommands SubCommands, mod ...CommandModifier) *cobra.Command {\n\treturn newCommand(c, run, subCommands, mod...)\n}", "func NewDmnList(src interface{}) (*DmnList, error) {\n\tthis := new(DmnList)\n\terr := this.Load(src)\n\treturn this, err\n}", "func NewCmdPriceListsCreate() *cobra.Command {\n\tcfgs, curCfg, err := configmgr.GetCurrentConfig()\n\tif err != nil {\n\t\tfmt.Fprintf(os.Stderr, \"%+v\\n\", err)\n\t\tos.Exit(1)\n\t}\n\tvar cmd = &cobra.Command{\n\t\tUse: \"create\",\n\t\tShort: \"Create a price list\",\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\tcurrent := cfgs.Configurations[curCfg]\n\t\t\tclient := eclient.New(current.Endpoint)\n\t\t\tif err := client.SetToken(&current); err != nil {\n\t\t\t\tfmt.Fprintf(os.Stderr, \"%+v\\n\", err)\n\t\t\t\tos.Exit(1)\n\t\t\t}\n\n\t\t\treq, err := promptCreatePriceList()\n\t\t\tif err != nil {\n\t\t\t\tfmt.Fprintf(os.Stderr, \"%v\\n\", err)\n\t\t\t\tos.Exit(1)\n\t\t\t}\n\n\t\t\t// attempt to create the price list\n\t\t\tctx := context.Background()\n\t\t\tpriceList, err := client.CreatePriceList(ctx, req)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Printf(\"%+v\\n\", err)\n\t\t\t\tos.Exit(1)\n\t\t\t}\n\n\t\t\tshowPriceList(priceList)\n\t\t},\n\t}\n\treturn cmd\n}", "func NewListCommand() cli.Command {\n\treturn NewListCommandWithEnv(commoncli.DefaultEnv)\n}", "func (client *CloudServicesClient) listCreateRequest(ctx context.Context, resourceGroupName string, options *CloudServicesClientListOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Compute/cloudServices\"\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2022-09-04\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func (client *ReplicationvCentersClient) listCreateRequest(ctx context.Context, options *ReplicationvCentersClientListOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.RecoveryServices/vaults/{resourceName}/replicationvCenters\"\n\tif client.resourceName == \"\" {\n\t\treturn nil, errors.New(\"parameter client.resourceName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceName}\", url.PathEscape(client.resourceName))\n\tif client.resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter client.resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(client.resourceGroupName))\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-11-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func DisConfList(c *gin.Context) {\n\tsearchDir := disconfDir + \"\" + c.Query(\"path\")\n\tfileList := []string{}\n\terr := filepath.Walk(searchDir, func(path string, f os.FileInfo, err error) error {\n\t\tfileList = append(fileList, path)\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\tlogrus.Error(\"error list file.\")\n\t\tc.AbortWithError(http.StatusBadRequest, err)\n\t\treturn\n\t}\n\tc.JSON(http.StatusOK, struct {\n\t\tFileList []string\n\t}{fileList})\n}", "func (client *ChildAvailabilityStatusesClient) listCreateRequest(ctx context.Context, resourceURI string, options *ChildAvailabilityStatusesClientListOptions) (*policy.Request, error) {\n\turlPath := \"/{resourceUri}/providers/Microsoft.ResourceHealth/childAvailabilityStatuses\"\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceUri}\", resourceURI)\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2017-07-01\")\n\tif options != nil && options.Filter != nil {\n\t\treqQP.Set(\"$filter\", *options.Filter)\n\t}\n\tif options != nil && options.Expand != nil {\n\t\treqQP.Set(\"$expand\", *options.Expand)\n\t}\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func (client *AccountsClient) listCreateRequest(ctx context.Context, options *AccountsClientListOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/providers/Microsoft.CognitiveServices/accounts\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2023-05-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func List(c *gophercloud.ServiceClient) pagination.Pager {\n\treturn common.List(c)\n}", "func NewAccountList(client ioctl.Client) *cobra.Command {\n\tshort, _ := client.SelectTranslation(_listCmdShorts)\n\n\treturn &cobra.Command{\n\t\tUse: \"list\",\n\t\tShort: short,\n\t\tArgs: cobra.ExactArgs(0),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcmd.SilenceUsage = true\n\t\t\tlistmessage := listMessage{}\n\t\t\taliases := client.AliasMap()\n\n\t\t\tif client.IsCryptoSm2() {\n\t\t\t\tsm2Accounts, err := listSm2Account(client)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn errors.Wrap(err, \"failed to get sm2 accounts\")\n\t\t\t\t}\n\t\t\t\tfor _, addr := range sm2Accounts {\n\t\t\t\t\tlistmessage.Accounts = append(listmessage.Accounts, account{\n\t\t\t\t\t\tAddress: addr,\n\t\t\t\t\t\tAlias: aliases[addr],\n\t\t\t\t\t})\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tks := client.NewKeyStore()\n\t\t\t\tfor _, v := range ks.Accounts() {\n\t\t\t\t\taddr, err := address.FromBytes(v.Address.Bytes())\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn errors.Wrap(err, \"failed to convert bytes into address\")\n\t\t\t\t\t}\n\t\t\t\t\tlistmessage.Accounts = append(listmessage.Accounts, account{\n\t\t\t\t\t\tAddress: addr.String(),\n\t\t\t\t\t\tAlias: aliases[addr.String()],\n\t\t\t\t\t})\n\t\t\t\t}\n\t\t\t}\n\t\t\tcmd.Println(listmessage.String())\n\t\t\treturn nil\n\t\t},\n\t}\n}", "func NewListNamespaceDescriptors(ctx context.Context, opts ...func(Call) error) (*ListNamespaceDescriptors, error) {\n\ttn := &ListNamespaceDescriptors{\n\t\tbase: base{\n\t\t\tctx: ctx,\n\t\t\tresultch: make(chan RPCResult, 1),\n\t\t},\n\t}\n\tif err := applyOptions(tn, opts...); err != nil {\n\t\treturn nil, err\n\t}\n\treturn tn, nil\n}", "func rcList(ctx context.Context, in Params) (out Params, err error) {\n\tout = make(Params)\n\tout[\"commands\"] = Calls.List()\n\treturn out, nil\n}", "func (d *DiscountControllerStruct) NewDiscount(c *gin.Context) {\n\tvar discountInfo domain.DiscountInfo\n\tif err := c.Bind(&discountInfo); err != nil {\n\t\terrorsHandler.GinErrorResponseHandler(c, err)\n\t\treturn\n\t}\n\n\tdiscountNumber := discountInfo.Number\n\tphoneNumber := discountInfo.PhoneNumber\n\n\tresult, err := logic.NewCreateDiscount(c).CreateNewDiscount(phoneNumber, discountNumber)\n\tif err != nil {\n\t\terrorsHandler.GinErrorResponseHandler(c, err)\n\t\treturn\n\t}\n\n\tc.JSON(http.StatusCreated,\n\t\tresult,\n\t)\n\treturn\n}", "func (a *Service) List(query *Usage) (*Usages, error) {\n\n\tusages, err := a.dataSvc.List(query)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn usages, nil\n}", "func (s *CouponsService) List(ctx context.Context, productID int) ([]*Coupon, *http.Response, error) {\n\tu := \"2.0/product/list_coupons\"\n\n\toptions := &CouponsOptions{\n\t\tProductID: productID,\n\t}\n\n\treq, err := s.client.NewRequest(\"POST\", u, options)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tcouponsResponse := new(CouponsResponse)\n\tresponse, err := s.client.Do(ctx, req, couponsResponse)\n\tif err != nil {\n\t\treturn nil, response, err\n\t}\n\n\treturn couponsResponse.Response, response, nil\n}", "func (c *SubresourceClient) List(namespace string, labels map[string]string) (result []metav1.Object, e error) {\n\tif c.Error != \"\" {\n\t\te = fmt.Errorf(c.Error)\n\t} else {\n\t\tresult = []metav1.Object{c.Subresource}\n\t}\n\treturn\n}", "func listCommandFunc(cmd *cobra.Command, args []string) {\n\turl := fmt.Sprintf(\"http://%s%s\", Global.Endpoints, proxy.APIProxies)\n\tcli := &http.Client{}\n\trsp, err := cli.Get(url)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t} else {\n\t\tdefer rsp.Body.Close()\n\t\tdata, err := ioutil.ReadAll(rsp.Body)\n\t\tif err != nil {\n\t\t\tfmt.Println(err)\n\t\t} else {\n\t\t\tfmt.Printf(\"%s\\n\", data)\n\t\t}\n\t}\n}", "func CreateDescribePortConnsListRequest() (request *DescribePortConnsListRequest) {\n\trequest = &DescribePortConnsListRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"ddoscoo\", \"2020-01-01\", \"DescribePortConnsList\", \"\", \"\")\n\trequest.Method = requests.POST\n\treturn\n}", "func (c *Client) List(_ string, action func(key string)) ([]string, error) {\n\n\treturn nil, errors.New(\"To be implemented\")\n}", "func NewQueryForbiddenInfoListRequestWithoutParam() *QueryForbiddenInfoListRequest {\n\n return &QueryForbiddenInfoListRequest{\n JDCloudRequest: core.JDCloudRequest{\n URL: \"/forbiddenInfo:query\",\n Method: \"POST\",\n Header: nil,\n Version: \"v1\",\n },\n }\n}", "func (client *CertificateOrdersClient) listCreateRequest(ctx context.Context, options *CertificateOrdersClientListOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/providers/Microsoft.CertificateRegistration/certificateOrders\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2022-09-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func getCategoriesList(lister listing.Service) http.HandlerFunc {\n\treturn func(rw http.ResponseWriter, r *http.Request) {\n\t\trw.Header().Set(\"Content-Type\", \"application/json\")\n\t\tvar list *listing.CategoriesResponse\n\n\t\tlimit := r.FormValue(\"limit\")\n\t\ttop := r.FormValue(\"top\")\n\n\t\tl, err := strconv.ParseInt(limit, 10, 2)\n\t\tif err != nil {\n\t\t\tl = 100\n\t\t}\n\t\tif top == \"true\" {\n\t\t\tlist, err = lister.TopCategories(r.Context(), int(l))\n\t\t} else {\n\t\t\tlist, err = lister.Categories(r.Context())\n\t\t}\n\n\t\tif err != nil {\n\t\t\t// server problem while answering the request\n\t\t\tr := constructError(http.StatusInternalServerError,\n\t\t\t\t\"unable to process request\",\n\t\t\t\t\"A problem occurs while processing the request\")\n\n\t\t\trw.WriteHeader(http.StatusInternalServerError)\n\t\t\tjson.NewEncoder(rw).Encode(r)\n\t\t\treturn\n\t\t}\n\n\t\trw.WriteHeader(http.StatusOK)\n\t\terr = json.NewEncoder(rw).Encode(list)\n\t\tif err != nil {\n\t\t\tfmt.Println(err)\n\t\t}\n\t}\n}", "func NewList(list uint32, mode uint32) {\n\tC.glowNewList(gpNewList, (C.GLuint)(list), (C.GLenum)(mode))\n}", "func (client *RecommendationsClient) listCreateRequest(ctx context.Context, options *RecommendationsListOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/providers/Microsoft.Web/recommendations\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\tif options != nil && options.Featured != nil {\n\t\treqQP.Set(\"featured\", strconv.FormatBool(*options.Featured))\n\t}\n\treqQP.Set(\"api-version\", \"2021-02-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\tunencodedParams := []string{req.Raw().URL.RawQuery}\n\tif options != nil && options.Filter != nil {\n\t\tunencodedParams = append(unencodedParams, \"$filter=\"+*options.Filter)\n\t}\n\treq.Raw().URL.RawQuery = strings.Join(unencodedParams, \"&\")\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func NewListCommand(commonOpts *common.Options) (cmd *cobra.Command) {\n\tcmd = &cobra.Command{\n\t\tUse: \"list\",\n\t\tShort: \"List all images of applications\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\treturn fmt.Errorf(\"not support yet\")\n\t\t},\n\t}\n\treturn\n}", "func (a *API) CreateList(list ListCreate) (*CreateListResponse, error) {\n\t// root namespace has id 1\n\tif list.NamespaceID == 0 {\n\t\treturn nil, fmt.Errorf(\"no namespace set\")\n\t}\n\n\t// no need to check UnSubscribtionMode since 0 is a valid option\n\tep, err := url.ParseRequestURI(a.endPoint.String() + \"/api/list\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// use the default configured system setting if nothing is set\n\tif list.SendConfiguration == 0 {\n\t\tlist.SendConfiguration = 1\n\t}\n\n\t// we add all data by reflecting the list\n\tdata := createData(list)\n\n\treq, err := http.NewRequest(http.MethodPost, ep.String(), strings.NewReader(data.Encode()))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq.Header.Add(\"Content-Type\", \"application/x-www-form-urlencoded\")\n\treq.Header.Add(\"Content-Length\", strconv.Itoa(len(data.Encode())))\n\n\tres, err := a.Request(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar resp CreateListResponse\n\terr = json.Unmarshal(res, &resp)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &resp, nil\n}" ]
[ "0.6457452", "0.6289995", "0.62045825", "0.58108985", "0.55878896", "0.54547507", "0.5450152", "0.54490876", "0.54153097", "0.53629786", "0.5332164", "0.5329603", "0.5305065", "0.5301289", "0.52555454", "0.5253076", "0.5235621", "0.52275664", "0.51807666", "0.51492125", "0.51309824", "0.51168925", "0.5109699", "0.5098235", "0.50893855", "0.5086864", "0.5085393", "0.5070191", "0.50561583", "0.502389", "0.50183713", "0.5017797", "0.5010876", "0.49889827", "0.49876788", "0.49776447", "0.49775162", "0.4974886", "0.49673772", "0.49419224", "0.49330807", "0.49264857", "0.49133414", "0.49119055", "0.49055317", "0.4903359", "0.48975754", "0.48878312", "0.4871738", "0.4867711", "0.48665035", "0.4864524", "0.4863501", "0.48634925", "0.4862723", "0.4859457", "0.48503855", "0.48480874", "0.48476654", "0.48412895", "0.48405853", "0.4837701", "0.48343435", "0.4833362", "0.4830381", "0.482874", "0.48219937", "0.48215166", "0.4815348", "0.48126662", "0.4811577", "0.4810742", "0.4807201", "0.48051265", "0.47999096", "0.47992554", "0.479684", "0.47964656", "0.47947913", "0.47917733", "0.47917646", "0.47871628", "0.47859755", "0.47757638", "0.4773594", "0.4770849", "0.47613457", "0.47569698", "0.4747045", "0.47412208", "0.47262454", "0.47207084", "0.47181025", "0.4716959", "0.47149023", "0.47142166", "0.4711043", "0.47076386", "0.47065216", "0.47064257" ]
0.81620395
0
Update syncs the struct instance changes into the database
func (u *User) Update() error { var prevUser User err := db.Todo.Get(&prevUser, `Select * from users where id = $1 or email = $2`, u.ID, u.Email) if err != nil { return errors.New("No user with specified ID to update") } if u.Password != nil && *u.Password != "" { if len(*u.Password) < 6 { return errors.New("Password needs to be at least four characters long") } hash, err := bcrypt.GenerateFromPassword([]byte(*u.Password), bcrypt.DefaultCost) if err != nil { fmt.Printf("Here2") return err } encryptedPassword := string(hash) u.EncryptedPassword = &encryptedPassword } _, err = db.Todo.Exec( `UPDATE users SET email=$2, first_name=$3, last_name=$4, encrypted_password=$5 WHERE id=$1`, prevUser.ID, u.Email, u.FirstName, u.LastName, u.EncryptedPassword) if err != nil { fmt.Printf("Here1") return err } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *Store) Update(w http.ResponseWriter, r *http.Request) {\n\t// We don't set up the: \"defer dphttp.DrainBody(r)\" here, as the body is fully read in function unmarshalInstance() below\n\t// and a call to DrainBody() puts this error: \"invalid Read on closed Body\" into the logs - to no good effect\n\t// because there is no more body to be read - so instead we just set up the usual Close() on the Body.\n\tdefer func() {\n\t\tif bodyCloseErr := r.Body.Close(); bodyCloseErr != nil {\n\t\t\tlog.Error(r.Context(), \"could not close response body\", bodyCloseErr)\n\t\t}\n\t}()\n\n\tctx := r.Context()\n\tvars := mux.Vars(r)\n\tinstanceID := vars[\"instance_id\"]\n\teTag := getIfMatch(r)\n\n\tlogData := log.Data{\"instance_id\": instanceID}\n\n\tinstance, err := unmarshalInstance(ctx, r.Body, false)\n\tif err != nil {\n\t\tlog.Error(ctx, \"update instance: failed unmarshalling json to model\", err, logData)\n\t\thandleInstanceErr(ctx, taskError{error: err, status: 400}, w, logData)\n\t\treturn\n\t}\n\n\tif err = validateInstanceUpdate(instance); err != nil {\n\t\thandleInstanceErr(ctx, taskError{error: err, status: 400}, w, logData)\n\t\treturn\n\t}\n\n\t// acquire instance lock so that the dp-graph call to AddVersionDetailsToInstance and the mongoDB update are atomic\n\tlockID, err := s.AcquireInstanceLock(ctx, instanceID)\n\tif err != nil {\n\t\tlog.Error(ctx, \"update instance: failed to lock the instance\", err, logData)\n\t\thandleInstanceErr(ctx, taskError{error: err, status: http.StatusInternalServerError}, w, logData)\n\t\treturn\n\t}\n\tdefer s.UnlockInstance(ctx, lockID)\n\n\t// Get the current document\n\tcurrentInstance, err := s.GetInstance(ctx, instanceID, eTag)\n\tif err != nil {\n\t\tlog.Error(ctx, \"update instance: store.GetInstance returned error\", err, logData)\n\t\thandleInstanceErr(ctx, err, w, logData)\n\t\treturn\n\t}\n\n\tlogData[\"current_state\"] = currentInstance.State\n\tlogData[\"requested_state\"] = instance.State\n\tif instance.State != \"\" && instance.State != currentInstance.State {\n\t\tif err = validateInstanceStateUpdate(instance, currentInstance); err != nil {\n\t\t\tlog.Error(ctx, \"update instance: instance state invalid\", err, logData)\n\t\t\thandleInstanceErr(ctx, err, w, logData)\n\t\t\treturn\n\t\t}\n\t}\n\n\tdatasetID := currentInstance.Links.Dataset.ID\n\n\t// edition confirmation is a one time process - cannot be edited for an instance once done\n\tif instance.State == models.EditionConfirmedState && instance.Version == 0 {\n\t\tif instance.Edition == \"\" {\n\t\t\tinstance.Edition = currentInstance.Edition\n\t\t}\n\n\t\tedition := instance.Edition\n\t\teditionLogData := log.Data{\"instance_id\": instanceID, \"dataset_id\": datasetID, \"edition\": edition}\n\n\t\teditionDoc, editionConfirmErr := s.confirmEdition(ctx, datasetID, edition, instanceID)\n\t\tif editionConfirmErr != nil {\n\t\t\tlog.Error(ctx, \"update instance: store.getEdition returned an error\", editionConfirmErr, editionLogData)\n\t\t\thandleInstanceErr(ctx, editionConfirmErr, w, logData)\n\t\t\treturn\n\t\t}\n\n\t\t// update instance with confirmed edition details\n\t\tinstance.Links = currentInstance.Links\n\t\tinstance.Links.Edition = &models.LinkObject{\n\t\t\tID: instance.Edition,\n\t\t\tHRef: editionDoc.Next.Links.Self.HRef,\n\t\t}\n\n\t\tinstance.Links.Version = editionDoc.Next.Links.LatestVersion\n\t\tinstance.Version, editionConfirmErr = strconv.Atoi(editionDoc.Next.Links.LatestVersion.ID)\n\t\tif editionConfirmErr != nil {\n\t\t\tlog.Error(ctx, \"update instance: failed to convert edition latestVersion id to instance.version int\", editionConfirmErr, editionLogData)\n\t\t\thandleInstanceErr(ctx, editionConfirmErr, w, logData)\n\t\t\treturn\n\t\t}\n\n\t\t// update dp-graph instance node (only for non-cantabular types)\n\t\tif currentInstance.Type == models.CantabularBlob.String() || currentInstance.Type == models.CantabularTable.String() || currentInstance.Type == models.CantabularFlexibleTable.String() || currentInstance.Type == models.CantabularMultivariateTable.String() {\n\t\t\teditionLogData[\"instance_type\"] = instance.Type\n\t\t\tlog.Info(ctx, \"skipping dp-graph instance update because it is not required by instance type\", editionLogData)\n\t\t} else {\n\t\t\tif versionErr := s.AddVersionDetailsToInstance(ctx, currentInstance.InstanceID, datasetID, edition, instance.Version); versionErr != nil {\n\t\t\t\tlog.Error(ctx, \"update instance: datastore.AddVersionDetailsToInstance returned an error\", versionErr, editionLogData)\n\t\t\t\thandleInstanceErr(ctx, versionErr, w, logData)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t\tlog.Info(ctx, \"update instance: added version details to instance\", editionLogData)\n\t}\n\n\t// Set the current mongo timestamp on instance document\n\tinstance.UniqueTimestamp = currentInstance.UniqueTimestamp\n\tnewETag, err := s.UpdateInstance(ctx, currentInstance, instance, eTag)\n\tif err != nil {\n\t\tlog.Error(ctx, \"update instance: store.UpdateInstance returned an error\", err, logData)\n\t\thandleInstanceErr(ctx, err, w, logData)\n\t\treturn\n\t}\n\n\tif instance, err = s.GetInstance(ctx, instanceID, newETag); err != nil {\n\t\tlog.Error(ctx, \"update instance: store.GetInstance for response returned an error\", err, logData)\n\t\thandleInstanceErr(ctx, err, w, logData)\n\t\treturn\n\t}\n\n\tb, err := json.Marshal(instance)\n\tif err != nil {\n\t\tlog.Error(ctx, \"add instance: failed to marshal instance to json\", err, logData)\n\t\thandleInstanceErr(ctx, err, w, logData)\n\t\treturn\n\t}\n\n\tsetJSONContentType(w)\n\tdpresponse.SetETag(w, newETag)\n\tw.WriteHeader(http.StatusOK)\n\twriteBody(ctx, w, b, logData)\n\n\tlog.Info(ctx, \"update instance: request successful\", logData)\n}", "func (d *Demo) UpdateStruct(g *gom.Gom) {\n\ttoolkit.Println(\"===== Update With Struct =====\")\n\thero := models.NewHero(\"Wonderwoman\", \"Gal Gadot\", 34)\n\n\tvar err error\n\tif d.useParams {\n\t\t_, err = g.Set(&gom.SetParams{\n\t\t\tTableName: \"hero\",\n\t\t\tFilter: gom.Eq(\"RealName\", \"Scarlett Johansson\"),\n\t\t\tTimeout: 10,\n\t\t}).Cmd().Update(hero)\n\t} else {\n\t\t_, err = g.Set(nil).Table(\"hero\").Timeout(10).Filter(gom.Eq(\"RealName\", \"Scarlett Johansson\")).Cmd().Update(hero)\n\t}\n\n\tif err != nil {\n\t\ttoolkit.Println(err.Error())\n\t\treturn\n\t}\n}", "func (s *Specification) Update() {\n\tInitDB()\n\n\tm, err := s.MarshalBinary()\n\n\tif err != nil {\n\t\tlog.Println(\"err\", err)\n\t}\n\n\tif err := db.Redis.Set(fmt.Sprintf(\"specification_%s\", string(s.ID)), m, 0).Err(); err != nil {\n\t\tfmt.Printf(\"Unable to store example struct into redis due to: %s \\n\", err)\n\t}\n}", "func Update(ctx context.Context, vo ProjectVO.Instance) {\n\tds.Update(ctx, vo)\n\tcache.InvalidateCache(ctx)\n}", "func (u SysDBUpdater) Update() error {\n\treturn u.db.Updates(u.fields).Error\n}", "func (r *Entity) Update() (result sql.Result, err error) {\n\treturn Model.Data(r).Where(gdb.GetWhereConditionOfStruct(r)).Update()\n}", "func (r *Entity) Update() (result sql.Result, err error) {\n\treturn Model.Data(r).Where(gdb.GetWhereConditionOfStruct(r)).Update()\n}", "func (r *Entity) Update() (result sql.Result, err error) {\n\treturn Model.Data(r).Where(gdb.GetWhereConditionOfStruct(r)).Update()\n}", "func (r *Entity) Update() (result sql.Result, err error) {\n\treturn Model.Data(r).Where(gdb.GetWhereConditionOfStruct(r)).Update()\n}", "func (r *Entity) Update() (result sql.Result, err error) {\n\treturn Model.Data(r).Where(gdb.GetWhereConditionOfStruct(r)).Update()\n}", "func (s *GrpcServer) Update(ctx context.Context, in *Empty) (*Empty, error) {\n\treturn &Empty{}, s.driver.Update()\n}", "func (w *Wrapper) Update(obj interface{}) (err error) {\n\t//grab a read lock\n\tw.connLock.RLock()\n\tdefer w.connLock.RUnlock()\n\n\t//make sure we have a connection\n\tif w.connection == nil {\n\t\treturn fmt.Errorf(\"No connection. Bootstrap a connection first.\")\n\t}\n\n\t//if we have a pointer, just grab the actual value\n\tprv := reflect.ValueOf(obj)\n\trv := reflect.Indirect(prv)\n\n\t//make sure it wasn't a nil pointer\n\tif !rv.IsValid() {\n\t\treturn fmt.Errorf(\"Invalid object passed in to Update\")\n\t}\n\n\t//make sure we were passed a struct\n\tif rv.Kind() != reflect.Struct {\n\t\treturn fmt.Errorf(\"%T is not a struct or pointer to struct\", obj)\n\t}\n\n\t//get the type\n\ttv := rv.Type()\n\n\t//grab the id of the object\n\tid, err := getId(rv)\n\tif err != nil {\n\t\treturn\n\t}\n\n\t//if we're making a new object, we have to have a pointer\n\t//so that we can set the value later after the insert.\n\tif id == 0 && prv.Kind() != reflect.Ptr {\n\t\treturn fmt.Errorf(\"Must pass in a pointer to the object if it is new\")\n\t}\n\n\t//begin a transaction\n\ttx, err := w.begin()\n\tif err != nil {\n\t\treturn\n\t}\n\n\t//set up our transaction logic\n\tdefer func() {\n\t\tif err == nil {\n\t\t\terr = tx.Commit()\n\t\t} else {\n\t\t\t//dont care about the error rolling back\n\t\t\ttx.Rollback()\n\t\t}\n\t}()\n\n\t//record if we made a new object\n\tvar newObject bool\n\n\t//insert the object. by the end we should have a nonzero id.\n\tif id != 0 {\n\t\t//lets check if the id exists and has the correct type\n\t\trow := tx.QueryRow(`SELECT object_type FROM objects WHERE\n\t\t\t object_id = $1\n\t\t\tAND object_deleted = false`, id)\n\n\t\t//grab the type\n\t\tvar resultType string\n\t\tif err = row.Scan(&resultType); err != nil {\n\t\t\treturn\n\t\t}\n\n\t\t//make sure the types match\n\t\tswitch resultType {\n\t\tcase tv.Name():\n\t\tcase \"\":\n\t\t\treturn fmt.Errorf(\"No object with id %d exists.\", id)\n\t\tdefault:\n\t\t\treturn fmt.Errorf(\"Cannot update an object of a different type: %s into %s\", tv.Name(), resultType)\n\t\t}\n\t\t//we now have a valid object\n\t} else {\n\t\t//do an insert and update the id\n\t\tvar result sql.Result\n\t\tresult, err = tx.Exec(`INSERT INTO objects (object_type) VALUES ($1)`, tv.Name())\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\n\t\t//try to grab the last insert id\n\t\tvar tid int64\n\t\ttid, err = result.LastInsertId()\n\t\tif err != nil {\n\t\t\t//lets try a little hack for postgres (not all drivers support LastInsertId)\n\t\t\trow := tx.QueryRow(`SELECT currval('objects_object_id_seq')`)\n\n\t\t\tif err = row.Scan(&tid); err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t\t//set our id\n\t\tid = int(tid)\n\n\t\t//update the object's id\n\t\trv.FieldByName(\"ID\").SetInt(tid)\n\n\t\t//record that we had a new object\n\t\tnewObject = true\n\t}\n\n\t//if we don't have an id by now, theres a problem\n\tif id == 0 {\n\t\treturn fmt.Errorf(\"Unable to get the ID of the last inserted object. Aborted.\")\n\t}\n\n\t//if we have a new object just insert all the fields (fast path)\n\tif newObject {\n\n\t\t//loop over the fields\n\t\tfor i := 0; i < rv.NumField(); i++ {\n\t\t\tfld, name := rv.Field(i), tv.Field(i).Name\n\t\t\tif name == \"ID\" {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t//check to see if the field is a Serializer\n\t\t\t//little wacky but you should be able to figure it out\n\t\t\tdata := fld.Addr().Interface()\n\t\t\tif ser, ok := data.(Serializer); ok {\n\t\t\t\tdata = ser.Serialize()\n\t\t\t}\n\n\t\t\t//insert the data\n\t\t\t_, err = tx.Exec(`INSERT INTO attributes\n\t\t\t\t(object_id, attribute_key, attribute_value)\n\t\t\t\tVALUES ($1, $2, $3)`, id, name, data)\n\n\t\t\t//check the error\n\t\t\tif err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t\t//object has been inserted\n\t\treturn\n\t}\n\n\t//now we have an actual update and the id of the object, so lets grab the\n\t//attributes\n\trows, err := tx.Query(`SELECT attribute_key, attribute_value\n\t\tFROM attributes WHERE\n\t\t object_id = $1\n\t\tAND attribute_archived = false\n\t\tAND attribute_preview = $2`, id, w.preview)\n\tif err != nil {\n\t\treturn\n\t}\n\tdefer rows.Close()\n\n\t//loop over the results\n\tvar (\n\t\tattrs = map[string][]byte{}\n\t\tkey string\n\t\tvalue []byte\n\t)\n\tfor rows.Next() {\n\t\t//scan in the attribute\n\t\tif err = rows.Scan(&key, &value); err != nil {\n\t\t\treturn\n\t\t}\n\n\t\t//store a copy of the data\n\t\tvcopy := make([]byte, len(value))\n\t\tcopy(vcopy, value)\n\t\tattrs[key] = vcopy\n\t}\n\n\t//return any errors\n\tif err = rows.Err(); err != nil {\n\t\treturn\n\t}\n\n\t//now lets look at our object and update the attributes in the database\n\t//a list of things to be archived\n\tarchive := []string{}\n\tfor i := 0; i < rv.NumField(); i++ {\n\t\tfld, name := rv.Field(i), tv.Field(i).Name\n\t\tif name == \"ID\" {\n\t\t\tcontinue\n\t\t}\n\n\t\t//if we have a value in the database and it differs\n\t\tif dbval, ex := attrs[name]; ex && differs(fld, dbval) {\n\t\t\t//then we need to archive\n\t\t\tarchive = append(archive, name)\n\t\t}\n\n\t\t//remove the value from the database\n\t\t//this way after the loop we have a set of values that are no longer\n\t\t//active for the type, so we can archive them.\n\t\tdelete(attrs, name)\n\t}\n\n\t//append the values that didn't get looked at from the struct to the list\n\t//to be archived\n\tfor name := range attrs {\n\t\tarchive = append(archive, name)\n\t}\n\n\t//archive the old attributes\n\t_, err = tx.Exec(fmt.Sprintf(`UPDATE attributes\n\t\tSET attribute_archived = true WHERE\n\t\tattribute_key IN ('%s')`, strings.Join(archive, `','`)))\n\tif err != nil {\n\t\treturn\n\t}\n\n\t//time to insert\n\tfor _, name := range archive {\n\t\tfld := rv.FieldByName(name)\n\n\t\t//check to see if the field is a Serializer\n\t\t//little wacky but you should be able to figure it out\n\t\tdata := fld.Interface()\n\t\tif ser, ok := data.(Serializer); ok {\n\t\t\tdata = ser.Serialize()\n\t\t}\n\n\t\t//insert it in\n\t\t_, err = tx.Exec(`INSERT INTO attributes\n\t\t\t(object_id, attribute_key, attribute_value)\n\t\t\tVALUES ($1, $2, $3)`, id, name, data)\n\n\t\t//return any errors\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\n\t//we're done!\n\treturn\n}", "func (p *Polling) Update(data Polling) {\n\tdb.Model(p).Updates(data)\n}", "func (s *Store) Update(c *gin.Context) {\n\n}", "func (m *MonkeyWrench) UpdateStruct(table string, sourceData interface{}) error {\n\treturn m.applyStructMutations(table, []interface{}{sourceData}, spanner.UpdateStruct)\n}", "func (db *DB) Update(model interface{}) error {\n\treturn db.inner.Update(model)\n}", "func (w *Wrapper) Update(data interface{}) (err error) {\n\tw.query = w.buildUpdate(data)\n\t_, err = w.executeQuery()\n\treturn\n}", "func (ch *ClickHouse) Update(object map[string]interface{}) error {\n\treturn ch.SyncStore(nil, []map[string]interface{}{object}, \"\", true)\n}", "func (c *peer) Update(obj *api.Peer) error {\n\tobj.UpdatedAt = time.Now().UTC().Format(time.RFC3339)\n\tjsonData, err := json.Marshal(obj)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn c.store.Set(path.Join(c.prefix, obj.UID), jsonData)\n}", "func (c *Connection) Update(model interface{}) error {\n\tsm := &Model{Value: model}\n\treturn sm.iterate(func(m *Model) error {\n\t\tvar err error\n\t\tm.touchUpdatedAt()\n\t\tif err = c.Dialect.Update(c.DB, m); err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t})\n}", "func (peer *Peer) Update(db *sql.DB) {\n\tpeer.Delete(db)\n\tpeer.Add(db)\n}", "func (gi *Sensor) Update(db *pg.DB) error {\r\n\tlog.Printf(\"===>sensorItem.Update()\")\r\n\r\n\t_, updateErr := db.Model(gi).\r\n\t\tWhere(\"id = ?0\", gi.ID).Update()\r\n\tif updateErr != nil {\r\n\t\tlog.Printf(\"Error while updating item in sensorItem.Update()\\n\")\r\n\t\tlog.Printf(\"Reason %v\\n\", updateErr)\r\n\t\treturn updateErr\r\n\t}\r\n\tlog.Printf(\"Product %s updated successfully in table\", gi.Sensorname)\r\n\treturn nil\r\n}", "func (s structModel)Update(table string)(query string, values[]interface{}, err error){\n\tif s.err != nil{\n\t\treturn \"\", nil, s.err\n\t}\n\tvar arrQuery []string\n\tquery = \"UPDATE \" + table + \" SET\"\n\tlistValues := make([]interface{}, 0)\n\tfor i, _ := range s.value{\n\t\tarrQuery = append(arrQuery, \" \" + s.key[i] + \"= $\" + strconv.Itoa(i+1))\n\t\tlistValues = append(listValues, s.value[i])\n\t}\n\tquery = query + strings.Join(arrQuery, \",\")\n\treturn query, listValues, nil\n}", "func (f *FakeInstance) Update(_ context.Context, _ string, _ *govultr.InstanceUpdateReq) (*govultr.Instance, *http.Response, error) {\n\tpanic(\"implement me\")\n}", "func (pmt *Payment) Update(db *bolt.DB) error {\n\treturn pmt.Create(db)\n}", "func (o *Vote) Update(exec boil.Executor, whitelist ...string) error {\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\to.UpdatedAt = currTime\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn err\n\t}\n\tkey := makeCacheKey(whitelist, nil)\n\tvoteUpdateCacheMut.RLock()\n\tcache, cached := voteUpdateCache[key]\n\tvoteUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(\n\t\t\tvoteColumns,\n\t\t\tvotePrimaryKeyColumns,\n\t\t\twhitelist,\n\t\t)\n\n\t\tif len(whitelist) == 0 {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"models: unable to update vote, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `vote` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, votePrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(voteType, voteMapping, append(wl, votePrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update vote row\")\n\t}\n\n\tif !cached {\n\t\tvoteUpdateCacheMut.Lock()\n\t\tvoteUpdateCache[key] = cache\n\t\tvoteUpdateCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpdateHooks(exec)\n}", "func (h *Handler) Update(c echo.Context) error {\n\trec := models.Rec{}\n\terr := c.Bind(&rec)\n\tif err != nil {\n\t\treturn c.JSON(http.StatusInternalServerError, err)\n\t}\n\n\terr = h.Store.Update(&rec)\n\tif err != nil {\n\t\treturn c.JSON(http.StatusInternalServerError, err)\n\t}\n\n\treturn c.JSON(http.StatusOK, rec)\n}", "func (db *InMemDatabase) Update(obj contrail.IObject, refs UIDList) error {\n\tuid := parseUID(obj.GetUuid())\n\tdata, ok := db.objectData[uid]\n\tif !ok {\n\t\treturn fmt.Errorf(\"Object %s: not in database\", obj.GetUuid())\n\t}\n\tdb.updateBackReferences(obj, refs, data.refs)\n\tdata.refs = refs\n\treturn nil\n}", "func (self boardModel) Update(db Database) BoardModel {\n // ignore error\n perpage, _ := db.GetThreadsPerPage(self.board)\n // refetch all on this page\n model := db.GetGroupForPage(self.prefix, self.frontend, self.board, self.page, int(perpage))\n var threads []ThreadModel\n for _, th := range model.Threads() {\n threads = append(threads, th.Update(db))\n }\n return boardModel{\n frontend: self.frontend,\n prefix: self.prefix,\n board: self.board,\n page: self.page,\n pages: self.pages,\n threads: threads,\n }\n}", "func (p *pollData) Update(poll Poll) (err error) {\n\n\tpollTags := GetStructTags(\"non-null\", \"db\", poll)\n\tpollFields := func(tags []string) string {\n\t\tvar temp []string\n\t\tfor _, tag := range tags {\n\t\t\ttemp = append(temp, fmt.Sprintf(`%s = :%s`, tag, tag))\n\t\t}\n\t\treturn strings.Join(temp, \", \")\n\t}(pollTags)\n\tquery := fmt.Sprintf(`Update polls SET %s WHERE id = :id`, pollFields)\n\tif _, err := rrsql.DB.NamedExec(query, poll); err != nil {\n\t\tlog.Fatal(err)\n\t\treturn err\n\t}\n\treturn nil\n}", "func (m *Model) updated() {\n\tif m.IsTimestamps() {\n\t\tnow := time.Now()\n\t\tm.UpdatedAt = &now\n\t\tm.applyTimestamps()\n\t}\n}", "func (r *Entity) Update() (result sql.Result, err error) {\n\twhere, args, err := gdb.GetWhereConditionOfStruct(r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn Model.Data(r).Where(where, args).Update()\n}", "func (m *Module) Update() (*Module, error) {\n\tif err := database.BackendDB.DB.Save(m).Error; err != nil {\n\t\tlog.Print(err)\n\t\treturn nil, err\n\t}\n\treturn m, nil\n}", "func (r *wagerRepositoryImpl) Update(wager *entity.Wager) error {\n\treturn r.db.Save(wager).Error\n}", "func (dao *Wager) Update(tx *gorm.DB, mod *models.Wager) error {\n\terr := tx.Save(mod).Error\n\tif err != nil {\n\t\treturn errs.Wrap(errs.ErrMySQLUpdate, err.Error(), \"tx.Save\")\n\t}\n\treturn nil\n}", "func (blt Bolt) Update(execute dbtx.Execute) error {\n\treturn blt.db.Update(func(tx *b.Tx) error {\n\t\treturn execute(tx.Bucket(blt.Bucket))\n\t})\n}", "func (c *Company) Update() {\n\tinitdb.DbInstance.Save(&c)\n\tlog.Println(\"Updated -> \", c)\n}", "func (self *DB) EmitUpdate(obj interface{}) (err error) {\n\tvalue := reflect.ValueOf(obj).Elem()\n\treturn self.emit(reflect.TypeOf(value.Interface()), &value, &value)\n}", "func (database *Database) Update(object interface{}) error {\n\tif err := database.DB.Model(object).Updates(object).Error; err != nil {\n\t\treturn fmt.Errorf(\"Failed to update '%s': %s\", reflect.TypeOf(object), err)\n\t}\n\treturn nil\n}", "func (u GroupUpdater) Update() error {\n\treturn u.db.Updates(u.fields).Error\n}", "func (r *resourceFrameworkShare) Update(ctx context.Context, req resource.UpdateRequest, resp *resource.UpdateResponse) {\n}", "func (c *Class) Update() error {\n\tvar err error\n\n\tif c.ID.Valid() == false {\n\t\treturn errmsg.ErrInvalidID\n\t}\n\n\tif err = c.save(); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (d *Dataset) Update(a *config.AppContext) error {\n\t/*\n\t * Then we will update\n\t */\n\n\t//updating the model\n\td.UserID = a.Session.User.ID\n\terr := a.Db.Model(d).Updates(map[string]interface{}{\n\t\t\"name\": d.Name,\n\t\t\"description\": d.Description,\n\t\t\"table_created\": d.TableCreated,\n\t\t\"datastore_id\": d.DatastoreID,\n\t}).Error\n\n\treturn err\n}", "func (t *DbService) Update(request *UpdateRequest) (*UpdateResponse, error) {\n\trsp := &UpdateResponse{}\n\treturn rsp, t.client.Call(\"db\", \"Update\", request, rsp)\n}", "func (s *stepsrvc) Update(ctx context.Context, p *step.StoredListOfSteps) (err error) {\n\ts.logger.Print(\"step.update\")\n\n\terr = s.db.UpdateOneStep(*p)\n\n\treturn err\n}", "func (w *wireguardServerConfig) Update(obj *api.WireguardServerConfig) error {\n\tobj.UpdatedAt = time.Now().UTC().Format(time.RFC3339)\n\tjsonData, err := json.Marshal(obj)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// if err := w.store.CreateBucketIfNotExists(obj.UID); err != nil {\n\t// \treturn err\n\t// }\n\treturn w.store.Set(path.Join(w.prefix, obj.UID), jsonData)\n}", "func (h *Hotel) Update(a *config.AppContext) error {\n\treturn a.Db.Model(h).Updates(map[string]interface{}{\n\t\t\"name\": h.Name,\n\t\t\"Description\": h.Description,\n\t\t\"GoogleLocation\": h.GoogleLocation,\n\t\t\"Address\": h.Address,\n\t\t\"Contact\": h.Contact,\n\t}).Error\n}", "func updateObject(object *remember.DataObject, connection *sql.DB) error {\n\tstatement, err := connection.Prepare(UPDATE)\n\tif nil != err {\n\t\treturn err\n\t}\n\tdefer statement.Close()\n\n\tobject.UpdatedAt = time.Now()\n\n\t_, err = statement.Exec(\n\t\tobject.Title,\n\t\tobject.GroupId,\n\t\tobject.Payload,\n\t\tobject.UpdatedAt.Unix(),\n\t\tobject.ID,\n\t)\n\n\treturn err\n}", "func (w WaitListStatusRepository) Update() (interface{}, error) {\n\tvar sqlStm = `SELECT update_waitlist_status( $1, $2, $3 )`\n\n\ttx, err := w.DB.Begin()\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"%s\", err)\n\t}\n\n\tdefer tx.Rollback()\n\n\tstmt, err := tx.Prepare(sqlStm)\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"%s\", err)\n\t}\n\n\tdefer stmt.Close()\n\n\tif _, err = stmt.Exec(\n\t\tw.Model.ID,\n\t\tw.Model.Description,\n\t\tw.Model.Value,\n\t); err != nil {\n\t\treturn nil, fmt.Errorf(\"%s\", err)\n\t}\n\n\tif err = tx.Commit(); err != nil {\n\t\treturn nil, fmt.Errorf(\"%s\", err)\n\t}\n\n\treturn w.Model, nil\n}", "func (s *PollVoteStore) Update(record *PollVote, cols ...kallax.SchemaField) (updated int64, err error) {\n\trecord.CreatedAt = record.CreatedAt.Truncate(time.Microsecond)\n\trecord.UpdatedAt = record.UpdatedAt.Truncate(time.Microsecond)\n\n\trecord.SetSaving(true)\n\tdefer record.SetSaving(false)\n\n\tif err := record.BeforeSave(); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn s.Store.Update(Schema.PollVote.BaseSchema, record, cols...)\n}", "func (v *Status) Update() error {\n\treturn nil\n}", "func (hg *Hg) Update() error {\n\treturn nil\n}", "func Update(db DB, table string, src interface{}) error {\n\treturn UpdateContext(context.Background(), db, table, src)\n}", "func (i *Instance) Update() error {\n\tvar err error\n\n\tif i.ID.Valid() == false {\n\t\treturn errmsg.ErrInvalidID\n\t}\n\n\tif err = i.save(); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (a *apiEndpoint) Update() error {\n\ta.e.mutex.RLock()\n\ta.SetEntry(\"server\", api.String(a.e.poolEntry.Desc))\n\ta.SetEntry(\"status\", api.String(a.e.status.String()))\n\tif a.e.lastErr != nil {\n\t\ta.SetEntry(\"last_error\", api.String(a.e.lastErr.Error()))\n\t\ta.SetEntry(\"last_error_time\", api.String(a.e.lastErrTime.Format(time.RFC3339)))\n\t} else {\n\t\ta.SetEntry(\"last_error\", api.Null)\n\t\ta.SetEntry(\"last_error_time\", api.Null)\n\t}\n\ta.SetEntry(\"pendingPayloads\", api.Number(a.e.NumPending()))\n\ta.SetEntry(\"publishedLines\", api.Number(a.e.LineCount()))\n\ta.SetEntry(\"averageLatency\", api.Float(a.e.AverageLatency()/time.Millisecond))\n\ta.e.mutex.RUnlock()\n\n\treturn nil\n}", "func (t *table) Update(router.Route) error {\n\treturn nil\n}", "func (u *UserModel) Update() error {\n\treturn DB.Self.Save(u).Error\n}", "func (s *HelloSystem) Update(ctx core.UpdateCtx) {}", "func (d *Database) Update(db DB, table string, src interface{}) error {\n\treturn d.UpdateContext(context.Background(), db, table, src)\n}", "func (wu *WsubUpdate) Update(db XODB) error {\n\tvar err error\n\n\t// if doesn't exist, bail\n\tif !wu._exists {\n\t\treturn errors.New(\"update failed: does not exist\")\n\t}\n\n\t// if deleted, bail\n\tif wu._deleted {\n\t\treturn errors.New(\"update failed: marked for deletion\")\n\t}\n\n\t// sql query\n\tconst sqlstr = `UPDATE jlabo.wsub_updates SET ` +\n\t\t`update_site_id = ?, extension_id = ?, name = ?, description = ?, element = ?, type = ?, folder = ?, client_id = ?, version = ?, data = ?, detailsurl = ?, infourl = ?, extra_query = ?` +\n\t\t` WHERE update_id = ?`\n\n\t// run query\n\tXOLog(sqlstr, wu.UpdateSiteID, wu.ExtensionID, wu.Name, wu.Description, wu.Element, wu.Type, wu.Folder, wu.ClientID, wu.Version, wu.Data, wu.Detailsurl, wu.Infourl, wu.ExtraQuery, wu.UpdateID)\n\t_, err = db.Exec(sqlstr, wu.UpdateSiteID, wu.ExtensionID, wu.Name, wu.Description, wu.Element, wu.Type, wu.Folder, wu.ClientID, wu.Version, wu.Data, wu.Detailsurl, wu.Infourl, wu.ExtraQuery, wu.UpdateID)\n\treturn err\n}", "func (r *reflectorStore) Update(obj interface{}) error {\n\treturn r.Add(obj)\n}", "func (v *Vessel) Update() (*Vessel, error) {\n\tdb := adaptors.DBConnector()\n\tdefer db.Close()\n\n\t_, notFoundErr := v.FetchOne()\n\n\tif notFoundErr != nil {\n\t\treturn v, notFoundErr\n\t}\n\n\terr := db.Table(\"vessels\").Where(\"id = ?\", v.ID).Updates(&v).Error\n\treturn v, err\n}", "func (o *Peer) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tpeerUpdateCacheMut.RLock()\n\tcache, cached := peerUpdateCache[key]\n\tpeerUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tpeerAllColumns,\n\t\t\tpeerPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"model: unable to update peers, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `peers` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, peerPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(peerType, peerMapping, append(wl, peerPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model: unable to update peers row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model: failed to get rows affected by update for peers\")\n\t}\n\n\tif !cached {\n\t\tpeerUpdateCacheMut.Lock()\n\t\tpeerUpdateCache[key] = cache\n\t\tpeerUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (u *usecase) Update() error {\n\t// Time execution\n\tstart := time.Now()\n\n\t// Creating context with timeout duration process\n\tctx, cancel := context.WithTimeout(context.Background(), 20*time.Minute)\n\tdefer cancel()\n\n\t// Get all archieve from scrapper repository\n\tarchieves, err := u.scrapperRepo.GetAllArchieve()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Create array (slice) of archieve object\n\tvar newArchieves []*model.Archieve\n\n\t// Check if archieve is exist in DB\n\tfor _, archieve := range archieves {\n\t\t// Get archieve from DB by archieve code\n\t\t_, err := u.mysqlRepo.GetArchieveByCode(ctx, archieve.Code)\n\n\t\t// if archieve not exist then add to newArchieve array (slice)\n\t\tif err == model.ErrDataNotFound {\n\t\t\t// Add archieve\n\t\t\tnewArchieves = append(newArchieves, archieve)\n\t\t\tlog.Printf(\"New archieve: %v\", archieve.Code)\n\t\t} else if err != nil && err != model.ErrDataNotFound {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Counter new journal from archieves\n\tvar totalNewJournal int\n\n\t// Get new archieves journals\n\tfor _, newArchieve := range newArchieves {\n\t\t// Get all journal from scrapper repository based on archieve\n\t\tjournals, err := u.scrapperRepo.GetAllJournalByArchieveObject(newArchieve)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// Append Journals into archieve object\n\t\tnewArchieve.Journals = journals\n\t\ttotalNewJournal += len(newArchieve.Journals)\n\t}\n\n\t// Check if there's new archieve then saved new archieve into DB\n\tif len(newArchieves) > 0 {\n\t\t// Insert new archieves into DB\n\t\tif err := u.mysqlRepo.BatchArchieves(ctx, newArchieves); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tlog.Printf(\"Added %v archieve and %v journal (%v)m\", len(newArchieves), totalNewJournal, time.Since(start).Minutes())\n\n\t// if there's no update then do nothing or finish pull data from archieve scrapper\n\treturn nil\n}", "func (s *permisoService) Update(id string, p *model.Permiso) error {\n\n\t// p.UpdatedAt = time.Now()\n\n\t// return s.service.Update(id, p)\n\n\treturn nil\n}", "func (r *DarwinTimetable) Update(f func(*bolt.Tx) error) error {\n\treturn r.db.Update(f)\n}", "func (r *RadioStation) Update() error {\n\treturn nil\n}", "func (r *Reportwallettb) Update() *ReportwallettbUpdateOne {\n\treturn (&ReportwallettbClient{config: r.config}).UpdateOne(r)\n}", "func (api *bucketAPI) SyncUpdate(obj *objstore.Bucket) error {\n\tif api.ct.objResolver != nil {\n\t\tlog.Fatal(\"Cannot use Sync update when object resolver is enabled on ctkit\")\n\t}\n\tnewObj := obj\n\tvar writeErr error\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\tnewObj, writeErr = apicl.ObjstoreV1().Bucket().Update(context.Background(), obj)\n\t}\n\n\tif writeErr == nil {\n\t\tapi.ct.handleBucketEvent(&kvstore.WatchEvent{Object: newObj, Type: kvstore.Updated})\n\t}\n\n\treturn writeErr\n}", "func update(updateStruct *Update) (bool, error) {\n\terr := updateStruct.Collection.UpdateId(updateStruct.Id, updateStruct.Data)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t\treturn false, err\n\t}\n\treturn true, nil\n}", "func (api *objectAPI) SyncUpdate(obj *objstore.Object) error {\n\tif api.ct.objResolver != nil {\n\t\tlog.Fatal(\"Cannot use Sync update when object resolver is enabled on ctkit\")\n\t}\n\tnewObj := obj\n\tvar writeErr error\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\tnewObj, writeErr = apicl.ObjstoreV1().Object().Update(context.Background(), obj)\n\t}\n\n\tif writeErr == nil {\n\t\tapi.ct.handleObjectEvent(&kvstore.WatchEvent{Object: newObj, Type: kvstore.Updated})\n\t}\n\n\treturn writeErr\n}", "func (a *Anime) Update() error {\n\tdata, err := json.Marshal(a)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn database.DB.Set(a.Key(), data, -1).Err()\n}", "func (table *ClassTable) Update(strId string, updates Class) (updated Class, err error) {\n\tdata, err := table.connection.Update(strId, CLASS_TABLE, updates)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = utilities.FillStruct(data, &updated)\n\treturn\n}", "func (c *Collection) Update(entry interface{}) error {\n\tkeyComponents, err := c.generateKey(entry)\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to update in collection. %s\", err.Error())\n\t}\n\n\tkey, err := c.formatKey(keyComponents)\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to update in collection. %s\", err.Error())\n\t}\n\n\texists, err := c.exists(key)\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to update in collection. %s\", err.Error())\n\t}\n\n\tif !exists {\n\t\treturn fmt.Errorf(\"Failed to update in collection. Key already exists\")\n\t}\n\n\tbytes, err := c.Serializer.ToBytes(entry)\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to update in collection. %s\", err.Error())\n\t}\n\n\tif c.Name != WorldStateIdentifier {\n\t\terr = c.Stub.PutPrivateData(c.Name, key, bytes)\n\t} else {\n\t\terr = c.Stub.PutState(key, bytes)\n\t}\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to update in collection. %s\", err.Error())\n\t}\n\n\treturn nil\n}", "func Update(db DB, table string, src interface{}) error {\n\treturn Default.Update(db, table, src)\n}", "func (o *PaymentObject) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tpaymentObjectUpdateCacheMut.RLock()\n\tcache, cached := paymentObjectUpdateCache[key]\n\tpaymentObjectUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tpaymentObjectAllColumns,\n\t\t\tpaymentObjectPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update payment_objects, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `payment_objects` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, paymentObjectPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(paymentObjectType, paymentObjectMapping, append(wl, paymentObjectPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update payment_objects row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for payment_objects\")\n\t}\n\n\tif !cached {\n\t\tpaymentObjectUpdateCacheMut.Lock()\n\t\tpaymentObjectUpdateCache[key] = cache\n\t\tpaymentObjectUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (tablePreeCheck *TablePreeCheck) UpdateModel() {\n\tdb := db.GetPostgresDb()\n\tdb.Save(tablePreeCheck)\n}", "func (factory ObjectFactoryImpl) UpdateWithTime(updatedInstance entity.Object, instance interface{}) entity.Object {\n\treflect.ValueOf(&updatedInstance).Elem().FieldByName(\"UpdatedAt\").SetString(generator.CurrentDateTime())\n\treflect.ValueOf(&updatedInstance).Elem().FieldByName(\"Extra\").Set(reflect.ValueOf(&instance))\n\treturn updatedInstance\n}", "func (m *File) Update(attr string, value interface{}) error {\n\treturn UnscopedDb().Model(m).UpdateColumn(attr, value).Error\n}", "func (t *Track) Update(i interface{}) error {\n\n\tinfo := i.(Track)\n\n\tif err := db.Model(&t).Updates(info).Error; err != nil {\n\t\tlogrus.Warn(err)\n\t\treturn err\n\t}\n\n\tlogrus.Infof(\"icii updated track #%x with new information\", t.ID)\n\n\treturn nil\n\n}", "func (u *CardRepository) Update(model *models.Card, data map[string]interface{}) error {\n\tquery := u.CardTable().Model(model).Updates(data)\n\tif err := query.Error; err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (m *MovieModel) Update(movie *Movie) error {\n\n\t/* potential to use uuid here\n\tUPDATE movies\n\tSET title = $1, year = $2, runtime = $3, genres = $4, version = uuid_generate_v4()\n\tWHERE id = $5 AND\n\t**/\n\t// Add version = $6, so we can stop race conditions\n\tquery := `\n\tUPDATE movies\n\tSET title = $1, year = $2, runtime = $3, genres = $4, version = version + 1\n\tWHERE id = $5 AND version = $6\n\tRETURNING version \n\t`\n\n\t// create the arg slice contaninig the values for the placeholder params.\n\targs := []interface{}{\n\t\tmovie.Title,\n\t\tmovie.Year,\n\t\tmovie.Runtime,\n\t\tpq.Array(movie.Genres),\n\t\tmovie.ID,\n\t\tmovie.Version, // Add the expected movie version\n\t}\n\n\tctx, cancel := context.WithTimeout(context.Background(), 3*time.Second)\n\tdefer cancel()\n\n\t// If no matching row could be found (version has been changed)\n\terr := m.DB.QueryRowContext(ctx, query, args...).Scan(&movie.Version)\n\tif err != nil {\n\t\tswitch {\n\t\tcase errors.Is(err, sql.ErrNoRows):\n\t\t\treturn ErrEditConflict\n\t\tdefault:\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (inst *Instance) fieldUpdate(t time.Time) {\n\tinst.lifePool.Update(t)\n}", "func (o *Status) Update() {\n o.Time = time.Now()\n}", "func (r *ItemsRepository) update(i *Item) error {\n\tif query := r.databaseHandler.DB().Save(&i); query.Error != nil {\n\t\treturn query.Error\n\t}\n\treturn nil\n}", "func (l *lead) Update() error {\n\tif err := l.Validate(); err != nil {\n\t\treturn err\n\t}\n\tl.UpdatedAt = bson.Now()\n\terr := config.LeadsCollection.UpdateId(l.ID, l)\n\treturn err\n}", "func (u *updater) Update(ctx context.Context, at int64, payload []byte) error {\n\terr := u.Put(ctx, &index{u.next}, at, payload)\n\tif err != nil {\n\t\treturn err\n\t}\n\tu.next++\n\treturn nil\n}", "func (u *updater) Update(ctx context.Context, at int64, payload []byte) error {\n\terr := u.Put(ctx, &index{u.next}, at, payload)\n\tif err != nil {\n\t\treturn err\n\t}\n\tu.next++\n\treturn nil\n}", "func (c client) Update(q Query) error {\n\treturn c.exec(q)\n}", "func (o *sampleUpdateHandler) Update(rw http.ResponseWriter, req *http.Request) {\n\to.UpdateHandler.Update(rw, req)\n}", "func Update(config TomlConfig, metaData toml.MetaData, db *DB) error {\n\t// Load data\n\tlog.Println(\"load data from database\")\n\tdataSet, err := db.ToDataSet()\n\tif err != nil {\n\t\treturn err\n\t}\n\t// Update popularity\n\tif err = UpdatePopularity(dataSet, db); err != nil {\n\t\treturn err\n\t}\n\t// Update popular items\n\tif err = UpdatePopItem(config.Recommend.CacheSize, db); err != nil {\n\t\treturn err\n\t}\n\t// Update latest items\n\tif err = UpdateLatest(config.Recommend.CacheSize, db); err != nil {\n\t\treturn err\n\t}\n\t// Generate recommends\n\tparams := config.Params.ToParams(metaData)\n\tif err = UpdateRecommends(config.Recommend.Model, params, config.Recommend.CacheSize, config.Recommend.FitJobs, config.Recommend.Once,\n\t\tdataSet, db); err != nil {\n\t\treturn err\n\t}\n\t// Generate neighbors\n\tif err = UpdateNeighbors(config.Recommend.Similarity, config.Recommend.CacheSize, dataSet, db); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (s *Dao) Update(ctx context.Context, kv *model.KVDoc) error {\n\tkeyKv := key.KV(kv.Domain, kv.Project, kv.ID)\n\tresp, err := etcdadpt.Get(ctx, keyKv)\n\tif err != nil {\n\t\topenlog.Error(err.Error())\n\t\treturn err\n\t}\n\tif resp == nil {\n\t\treturn datasource.ErrRecordNotExists\n\t}\n\n\tvar old model.KVDoc\n\terr = json.Unmarshal(resp.Value, &old)\n\tif err != nil {\n\t\topenlog.Error(err.Error())\n\t\treturn err\n\t}\n\told.LabelFormat = kv.LabelFormat\n\told.Value = kv.Value\n\told.Status = kv.Status\n\told.Checker = kv.Checker\n\told.UpdateTime = kv.UpdateTime\n\told.UpdateRevision = kv.UpdateRevision\n\n\tbytes, err := json.Marshal(old)\n\tif err != nil {\n\t\topenlog.Error(err.Error())\n\t\treturn err\n\t}\n\terr = etcdadpt.PutBytes(ctx, keyKv, bytes)\n\tif err != nil {\n\t\topenlog.Error(err.Error())\n\t\treturn err\n\t}\n\treturn nil\n}", "func UpdatePerson(db *sql.DB) {}", "func Update(db gorp.SqlExecutor, i interface{}) error {\n\treturn Mapper.Update(db, i)\n}", "func (bt *Tree) Update(fields []sql.Field, resolution time.Duration, truncateBefore time.Time, key []byte, vals encoding.TSParams, metadata bytemap.ByteMap) int {\n\tbytesAdded, newNode := bt.doUpdate(fields, resolution, truncateBefore, key, vals, metadata)\n\tbt.bytes += bytesAdded\n\tif newNode {\n\t\tbt.length++\n\t}\n\treturn bytesAdded\n}", "func (r *sampleRepository) Update(sample *model.Sample) error {\n\treturn r.DB.Save(sample).Error\n}", "func (contact *Contact) Update() error {\n\treturn DB().Save(contact).Error\n}", "func memberUpdate(db *sql.DB, client *chef.Client) {\n\t// Get a unique list of all the users\n\tusers, err := db.Query(\"SELECT user_name FROM org_groups;\")\n\tif err != nil {\n\t\tpanic(err.Error()) // proper error handling instead of panic in your app\n\t}\n\tvar members []string\n\tfor users.Next() {\n\t\tvar name string\n\t\terr = users.Scan(&name)\n\t\tif err != nil {\n\t\t\tpanic(err.Error()) // proper error handling instead of panic in your app\n\t\t}\n\t\tmembers = append(members, name)\n\t}\n\tmembers = co.Unique(members)\n\tusers.Close()\n\tstmtInsMember, err := db.Prepare(\"INSERT INTO members (user_name, email, display_name) VALUES( ?, ?, ? )\")\n\tif err != nil {\n\t\tpanic(err.Error()) // proper error handling instead of panic in your app\n\t}\n // transaction - begin, delete existing users, add replacement data\n\ttx, err := db.Begin()\n\t_, err = db.Query(\"DELETE FROM members;\")\n\tfor _, member := range members {\n\t\t// Extract information for each user\n\t\tmemberInfo := getMember(client, member)\n\t\t// Update the data base with a new set of user records\n\t\t_, err = stmtInsMember.Exec(member, memberInfo.Email, memberInfo.DisplayName)\n\t\tif err != nil {\n\t\t\tpanic(err.Error()) // proper error handling instead of panic in your app\n\t\t}\n\t}\n stmtInsMember.Close()\n\ttx.Commit()\n}", "func (s *System) Update(dt *world.Clock) error {\n\treturn s.save()\n}", "func (g *GameDBModel) Update(game Game) error {\n\terr := database.C(COLLECTION).UpdateId(game.ID, &game)\n\treturn err\n}", "func (a *paymentUsecase) Update(c context.Context, ar *models.Payment) (*models.Payment, error) {\n\tctx, cancel := context.WithTimeout(c, a.contextTimeout)\n\tdefer cancel()\n\n\tar.UpdatedAt = time.Now()\n\treturn a.repo.Update(ctx, ar)\n}" ]
[ "0.6663857", "0.64484656", "0.64176106", "0.6386076", "0.63846666", "0.63183945", "0.63183945", "0.63183945", "0.63183945", "0.63183945", "0.6280699", "0.6221633", "0.62214124", "0.6215622", "0.62033606", "0.6172274", "0.60822904", "0.6041722", "0.60226613", "0.6006191", "0.5973311", "0.594424", "0.5925653", "0.5913681", "0.59054744", "0.59006655", "0.58831435", "0.5865293", "0.585237", "0.5834893", "0.58343923", "0.58320564", "0.5819343", "0.58095765", "0.5806885", "0.580434", "0.57992995", "0.578961", "0.5777545", "0.5767551", "0.5765028", "0.5762621", "0.57521117", "0.5747183", "0.5733086", "0.5725354", "0.5723133", "0.57098615", "0.5707844", "0.5707807", "0.57078016", "0.5700633", "0.5690091", "0.5686742", "0.56739163", "0.5673069", "0.56708574", "0.56694186", "0.56617266", "0.56565106", "0.56437844", "0.56291825", "0.56101584", "0.56028455", "0.56024384", "0.56018156", "0.55962646", "0.5596241", "0.55872285", "0.5581596", "0.55801505", "0.55781966", "0.55739975", "0.55723584", "0.557142", "0.55706203", "0.5557911", "0.55575955", "0.55518204", "0.5548625", "0.55453146", "0.5540036", "0.55332077", "0.5533083", "0.55199516", "0.55109614", "0.5500758", "0.5500758", "0.5498367", "0.5490768", "0.5490634", "0.5488148", "0.54871565", "0.5487098", "0.5484795", "0.5483465", "0.54785866", "0.54785776", "0.5478565", "0.54749846", "0.54651105" ]
0.0
-1
Update syncs the struct instance changes into the database
func (u *User) Get() error { // could potentially get other ways as well, keeping this simple if u.Email != "" { err := db.Todo.Get(u, `Select * from users where email = $1`, u.Email) if err != nil { return errors.New("No user with specified email to get") } } else if u.ID != "" { err := db.Todo.Get(u, `Select * from users where id = $1`, u.ID) if err != nil { return errors.New("No user with specified email to get") } } else { return errors.New("Need email or id for user to get") } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *Store) Update(w http.ResponseWriter, r *http.Request) {\n\t// We don't set up the: \"defer dphttp.DrainBody(r)\" here, as the body is fully read in function unmarshalInstance() below\n\t// and a call to DrainBody() puts this error: \"invalid Read on closed Body\" into the logs - to no good effect\n\t// because there is no more body to be read - so instead we just set up the usual Close() on the Body.\n\tdefer func() {\n\t\tif bodyCloseErr := r.Body.Close(); bodyCloseErr != nil {\n\t\t\tlog.Error(r.Context(), \"could not close response body\", bodyCloseErr)\n\t\t}\n\t}()\n\n\tctx := r.Context()\n\tvars := mux.Vars(r)\n\tinstanceID := vars[\"instance_id\"]\n\teTag := getIfMatch(r)\n\n\tlogData := log.Data{\"instance_id\": instanceID}\n\n\tinstance, err := unmarshalInstance(ctx, r.Body, false)\n\tif err != nil {\n\t\tlog.Error(ctx, \"update instance: failed unmarshalling json to model\", err, logData)\n\t\thandleInstanceErr(ctx, taskError{error: err, status: 400}, w, logData)\n\t\treturn\n\t}\n\n\tif err = validateInstanceUpdate(instance); err != nil {\n\t\thandleInstanceErr(ctx, taskError{error: err, status: 400}, w, logData)\n\t\treturn\n\t}\n\n\t// acquire instance lock so that the dp-graph call to AddVersionDetailsToInstance and the mongoDB update are atomic\n\tlockID, err := s.AcquireInstanceLock(ctx, instanceID)\n\tif err != nil {\n\t\tlog.Error(ctx, \"update instance: failed to lock the instance\", err, logData)\n\t\thandleInstanceErr(ctx, taskError{error: err, status: http.StatusInternalServerError}, w, logData)\n\t\treturn\n\t}\n\tdefer s.UnlockInstance(ctx, lockID)\n\n\t// Get the current document\n\tcurrentInstance, err := s.GetInstance(ctx, instanceID, eTag)\n\tif err != nil {\n\t\tlog.Error(ctx, \"update instance: store.GetInstance returned error\", err, logData)\n\t\thandleInstanceErr(ctx, err, w, logData)\n\t\treturn\n\t}\n\n\tlogData[\"current_state\"] = currentInstance.State\n\tlogData[\"requested_state\"] = instance.State\n\tif instance.State != \"\" && instance.State != currentInstance.State {\n\t\tif err = validateInstanceStateUpdate(instance, currentInstance); err != nil {\n\t\t\tlog.Error(ctx, \"update instance: instance state invalid\", err, logData)\n\t\t\thandleInstanceErr(ctx, err, w, logData)\n\t\t\treturn\n\t\t}\n\t}\n\n\tdatasetID := currentInstance.Links.Dataset.ID\n\n\t// edition confirmation is a one time process - cannot be edited for an instance once done\n\tif instance.State == models.EditionConfirmedState && instance.Version == 0 {\n\t\tif instance.Edition == \"\" {\n\t\t\tinstance.Edition = currentInstance.Edition\n\t\t}\n\n\t\tedition := instance.Edition\n\t\teditionLogData := log.Data{\"instance_id\": instanceID, \"dataset_id\": datasetID, \"edition\": edition}\n\n\t\teditionDoc, editionConfirmErr := s.confirmEdition(ctx, datasetID, edition, instanceID)\n\t\tif editionConfirmErr != nil {\n\t\t\tlog.Error(ctx, \"update instance: store.getEdition returned an error\", editionConfirmErr, editionLogData)\n\t\t\thandleInstanceErr(ctx, editionConfirmErr, w, logData)\n\t\t\treturn\n\t\t}\n\n\t\t// update instance with confirmed edition details\n\t\tinstance.Links = currentInstance.Links\n\t\tinstance.Links.Edition = &models.LinkObject{\n\t\t\tID: instance.Edition,\n\t\t\tHRef: editionDoc.Next.Links.Self.HRef,\n\t\t}\n\n\t\tinstance.Links.Version = editionDoc.Next.Links.LatestVersion\n\t\tinstance.Version, editionConfirmErr = strconv.Atoi(editionDoc.Next.Links.LatestVersion.ID)\n\t\tif editionConfirmErr != nil {\n\t\t\tlog.Error(ctx, \"update instance: failed to convert edition latestVersion id to instance.version int\", editionConfirmErr, editionLogData)\n\t\t\thandleInstanceErr(ctx, editionConfirmErr, w, logData)\n\t\t\treturn\n\t\t}\n\n\t\t// update dp-graph instance node (only for non-cantabular types)\n\t\tif currentInstance.Type == models.CantabularBlob.String() || currentInstance.Type == models.CantabularTable.String() || currentInstance.Type == models.CantabularFlexibleTable.String() || currentInstance.Type == models.CantabularMultivariateTable.String() {\n\t\t\teditionLogData[\"instance_type\"] = instance.Type\n\t\t\tlog.Info(ctx, \"skipping dp-graph instance update because it is not required by instance type\", editionLogData)\n\t\t} else {\n\t\t\tif versionErr := s.AddVersionDetailsToInstance(ctx, currentInstance.InstanceID, datasetID, edition, instance.Version); versionErr != nil {\n\t\t\t\tlog.Error(ctx, \"update instance: datastore.AddVersionDetailsToInstance returned an error\", versionErr, editionLogData)\n\t\t\t\thandleInstanceErr(ctx, versionErr, w, logData)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t\tlog.Info(ctx, \"update instance: added version details to instance\", editionLogData)\n\t}\n\n\t// Set the current mongo timestamp on instance document\n\tinstance.UniqueTimestamp = currentInstance.UniqueTimestamp\n\tnewETag, err := s.UpdateInstance(ctx, currentInstance, instance, eTag)\n\tif err != nil {\n\t\tlog.Error(ctx, \"update instance: store.UpdateInstance returned an error\", err, logData)\n\t\thandleInstanceErr(ctx, err, w, logData)\n\t\treturn\n\t}\n\n\tif instance, err = s.GetInstance(ctx, instanceID, newETag); err != nil {\n\t\tlog.Error(ctx, \"update instance: store.GetInstance for response returned an error\", err, logData)\n\t\thandleInstanceErr(ctx, err, w, logData)\n\t\treturn\n\t}\n\n\tb, err := json.Marshal(instance)\n\tif err != nil {\n\t\tlog.Error(ctx, \"add instance: failed to marshal instance to json\", err, logData)\n\t\thandleInstanceErr(ctx, err, w, logData)\n\t\treturn\n\t}\n\n\tsetJSONContentType(w)\n\tdpresponse.SetETag(w, newETag)\n\tw.WriteHeader(http.StatusOK)\n\twriteBody(ctx, w, b, logData)\n\n\tlog.Info(ctx, \"update instance: request successful\", logData)\n}", "func (d *Demo) UpdateStruct(g *gom.Gom) {\n\ttoolkit.Println(\"===== Update With Struct =====\")\n\thero := models.NewHero(\"Wonderwoman\", \"Gal Gadot\", 34)\n\n\tvar err error\n\tif d.useParams {\n\t\t_, err = g.Set(&gom.SetParams{\n\t\t\tTableName: \"hero\",\n\t\t\tFilter: gom.Eq(\"RealName\", \"Scarlett Johansson\"),\n\t\t\tTimeout: 10,\n\t\t}).Cmd().Update(hero)\n\t} else {\n\t\t_, err = g.Set(nil).Table(\"hero\").Timeout(10).Filter(gom.Eq(\"RealName\", \"Scarlett Johansson\")).Cmd().Update(hero)\n\t}\n\n\tif err != nil {\n\t\ttoolkit.Println(err.Error())\n\t\treturn\n\t}\n}", "func (s *Specification) Update() {\n\tInitDB()\n\n\tm, err := s.MarshalBinary()\n\n\tif err != nil {\n\t\tlog.Println(\"err\", err)\n\t}\n\n\tif err := db.Redis.Set(fmt.Sprintf(\"specification_%s\", string(s.ID)), m, 0).Err(); err != nil {\n\t\tfmt.Printf(\"Unable to store example struct into redis due to: %s \\n\", err)\n\t}\n}", "func Update(ctx context.Context, vo ProjectVO.Instance) {\n\tds.Update(ctx, vo)\n\tcache.InvalidateCache(ctx)\n}", "func (u SysDBUpdater) Update() error {\n\treturn u.db.Updates(u.fields).Error\n}", "func (r *Entity) Update() (result sql.Result, err error) {\n\treturn Model.Data(r).Where(gdb.GetWhereConditionOfStruct(r)).Update()\n}", "func (r *Entity) Update() (result sql.Result, err error) {\n\treturn Model.Data(r).Where(gdb.GetWhereConditionOfStruct(r)).Update()\n}", "func (r *Entity) Update() (result sql.Result, err error) {\n\treturn Model.Data(r).Where(gdb.GetWhereConditionOfStruct(r)).Update()\n}", "func (r *Entity) Update() (result sql.Result, err error) {\n\treturn Model.Data(r).Where(gdb.GetWhereConditionOfStruct(r)).Update()\n}", "func (r *Entity) Update() (result sql.Result, err error) {\n\treturn Model.Data(r).Where(gdb.GetWhereConditionOfStruct(r)).Update()\n}", "func (s *GrpcServer) Update(ctx context.Context, in *Empty) (*Empty, error) {\n\treturn &Empty{}, s.driver.Update()\n}", "func (w *Wrapper) Update(obj interface{}) (err error) {\n\t//grab a read lock\n\tw.connLock.RLock()\n\tdefer w.connLock.RUnlock()\n\n\t//make sure we have a connection\n\tif w.connection == nil {\n\t\treturn fmt.Errorf(\"No connection. Bootstrap a connection first.\")\n\t}\n\n\t//if we have a pointer, just grab the actual value\n\tprv := reflect.ValueOf(obj)\n\trv := reflect.Indirect(prv)\n\n\t//make sure it wasn't a nil pointer\n\tif !rv.IsValid() {\n\t\treturn fmt.Errorf(\"Invalid object passed in to Update\")\n\t}\n\n\t//make sure we were passed a struct\n\tif rv.Kind() != reflect.Struct {\n\t\treturn fmt.Errorf(\"%T is not a struct or pointer to struct\", obj)\n\t}\n\n\t//get the type\n\ttv := rv.Type()\n\n\t//grab the id of the object\n\tid, err := getId(rv)\n\tif err != nil {\n\t\treturn\n\t}\n\n\t//if we're making a new object, we have to have a pointer\n\t//so that we can set the value later after the insert.\n\tif id == 0 && prv.Kind() != reflect.Ptr {\n\t\treturn fmt.Errorf(\"Must pass in a pointer to the object if it is new\")\n\t}\n\n\t//begin a transaction\n\ttx, err := w.begin()\n\tif err != nil {\n\t\treturn\n\t}\n\n\t//set up our transaction logic\n\tdefer func() {\n\t\tif err == nil {\n\t\t\terr = tx.Commit()\n\t\t} else {\n\t\t\t//dont care about the error rolling back\n\t\t\ttx.Rollback()\n\t\t}\n\t}()\n\n\t//record if we made a new object\n\tvar newObject bool\n\n\t//insert the object. by the end we should have a nonzero id.\n\tif id != 0 {\n\t\t//lets check if the id exists and has the correct type\n\t\trow := tx.QueryRow(`SELECT object_type FROM objects WHERE\n\t\t\t object_id = $1\n\t\t\tAND object_deleted = false`, id)\n\n\t\t//grab the type\n\t\tvar resultType string\n\t\tif err = row.Scan(&resultType); err != nil {\n\t\t\treturn\n\t\t}\n\n\t\t//make sure the types match\n\t\tswitch resultType {\n\t\tcase tv.Name():\n\t\tcase \"\":\n\t\t\treturn fmt.Errorf(\"No object with id %d exists.\", id)\n\t\tdefault:\n\t\t\treturn fmt.Errorf(\"Cannot update an object of a different type: %s into %s\", tv.Name(), resultType)\n\t\t}\n\t\t//we now have a valid object\n\t} else {\n\t\t//do an insert and update the id\n\t\tvar result sql.Result\n\t\tresult, err = tx.Exec(`INSERT INTO objects (object_type) VALUES ($1)`, tv.Name())\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\n\t\t//try to grab the last insert id\n\t\tvar tid int64\n\t\ttid, err = result.LastInsertId()\n\t\tif err != nil {\n\t\t\t//lets try a little hack for postgres (not all drivers support LastInsertId)\n\t\t\trow := tx.QueryRow(`SELECT currval('objects_object_id_seq')`)\n\n\t\t\tif err = row.Scan(&tid); err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t\t//set our id\n\t\tid = int(tid)\n\n\t\t//update the object's id\n\t\trv.FieldByName(\"ID\").SetInt(tid)\n\n\t\t//record that we had a new object\n\t\tnewObject = true\n\t}\n\n\t//if we don't have an id by now, theres a problem\n\tif id == 0 {\n\t\treturn fmt.Errorf(\"Unable to get the ID of the last inserted object. Aborted.\")\n\t}\n\n\t//if we have a new object just insert all the fields (fast path)\n\tif newObject {\n\n\t\t//loop over the fields\n\t\tfor i := 0; i < rv.NumField(); i++ {\n\t\t\tfld, name := rv.Field(i), tv.Field(i).Name\n\t\t\tif name == \"ID\" {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t//check to see if the field is a Serializer\n\t\t\t//little wacky but you should be able to figure it out\n\t\t\tdata := fld.Addr().Interface()\n\t\t\tif ser, ok := data.(Serializer); ok {\n\t\t\t\tdata = ser.Serialize()\n\t\t\t}\n\n\t\t\t//insert the data\n\t\t\t_, err = tx.Exec(`INSERT INTO attributes\n\t\t\t\t(object_id, attribute_key, attribute_value)\n\t\t\t\tVALUES ($1, $2, $3)`, id, name, data)\n\n\t\t\t//check the error\n\t\t\tif err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t\t//object has been inserted\n\t\treturn\n\t}\n\n\t//now we have an actual update and the id of the object, so lets grab the\n\t//attributes\n\trows, err := tx.Query(`SELECT attribute_key, attribute_value\n\t\tFROM attributes WHERE\n\t\t object_id = $1\n\t\tAND attribute_archived = false\n\t\tAND attribute_preview = $2`, id, w.preview)\n\tif err != nil {\n\t\treturn\n\t}\n\tdefer rows.Close()\n\n\t//loop over the results\n\tvar (\n\t\tattrs = map[string][]byte{}\n\t\tkey string\n\t\tvalue []byte\n\t)\n\tfor rows.Next() {\n\t\t//scan in the attribute\n\t\tif err = rows.Scan(&key, &value); err != nil {\n\t\t\treturn\n\t\t}\n\n\t\t//store a copy of the data\n\t\tvcopy := make([]byte, len(value))\n\t\tcopy(vcopy, value)\n\t\tattrs[key] = vcopy\n\t}\n\n\t//return any errors\n\tif err = rows.Err(); err != nil {\n\t\treturn\n\t}\n\n\t//now lets look at our object and update the attributes in the database\n\t//a list of things to be archived\n\tarchive := []string{}\n\tfor i := 0; i < rv.NumField(); i++ {\n\t\tfld, name := rv.Field(i), tv.Field(i).Name\n\t\tif name == \"ID\" {\n\t\t\tcontinue\n\t\t}\n\n\t\t//if we have a value in the database and it differs\n\t\tif dbval, ex := attrs[name]; ex && differs(fld, dbval) {\n\t\t\t//then we need to archive\n\t\t\tarchive = append(archive, name)\n\t\t}\n\n\t\t//remove the value from the database\n\t\t//this way after the loop we have a set of values that are no longer\n\t\t//active for the type, so we can archive them.\n\t\tdelete(attrs, name)\n\t}\n\n\t//append the values that didn't get looked at from the struct to the list\n\t//to be archived\n\tfor name := range attrs {\n\t\tarchive = append(archive, name)\n\t}\n\n\t//archive the old attributes\n\t_, err = tx.Exec(fmt.Sprintf(`UPDATE attributes\n\t\tSET attribute_archived = true WHERE\n\t\tattribute_key IN ('%s')`, strings.Join(archive, `','`)))\n\tif err != nil {\n\t\treturn\n\t}\n\n\t//time to insert\n\tfor _, name := range archive {\n\t\tfld := rv.FieldByName(name)\n\n\t\t//check to see if the field is a Serializer\n\t\t//little wacky but you should be able to figure it out\n\t\tdata := fld.Interface()\n\t\tif ser, ok := data.(Serializer); ok {\n\t\t\tdata = ser.Serialize()\n\t\t}\n\n\t\t//insert it in\n\t\t_, err = tx.Exec(`INSERT INTO attributes\n\t\t\t(object_id, attribute_key, attribute_value)\n\t\t\tVALUES ($1, $2, $3)`, id, name, data)\n\n\t\t//return any errors\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\n\t//we're done!\n\treturn\n}", "func (p *Polling) Update(data Polling) {\n\tdb.Model(p).Updates(data)\n}", "func (s *Store) Update(c *gin.Context) {\n\n}", "func (m *MonkeyWrench) UpdateStruct(table string, sourceData interface{}) error {\n\treturn m.applyStructMutations(table, []interface{}{sourceData}, spanner.UpdateStruct)\n}", "func (db *DB) Update(model interface{}) error {\n\treturn db.inner.Update(model)\n}", "func (w *Wrapper) Update(data interface{}) (err error) {\n\tw.query = w.buildUpdate(data)\n\t_, err = w.executeQuery()\n\treturn\n}", "func (ch *ClickHouse) Update(object map[string]interface{}) error {\n\treturn ch.SyncStore(nil, []map[string]interface{}{object}, \"\", true)\n}", "func (c *peer) Update(obj *api.Peer) error {\n\tobj.UpdatedAt = time.Now().UTC().Format(time.RFC3339)\n\tjsonData, err := json.Marshal(obj)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn c.store.Set(path.Join(c.prefix, obj.UID), jsonData)\n}", "func (c *Connection) Update(model interface{}) error {\n\tsm := &Model{Value: model}\n\treturn sm.iterate(func(m *Model) error {\n\t\tvar err error\n\t\tm.touchUpdatedAt()\n\t\tif err = c.Dialect.Update(c.DB, m); err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t})\n}", "func (peer *Peer) Update(db *sql.DB) {\n\tpeer.Delete(db)\n\tpeer.Add(db)\n}", "func (gi *Sensor) Update(db *pg.DB) error {\r\n\tlog.Printf(\"===>sensorItem.Update()\")\r\n\r\n\t_, updateErr := db.Model(gi).\r\n\t\tWhere(\"id = ?0\", gi.ID).Update()\r\n\tif updateErr != nil {\r\n\t\tlog.Printf(\"Error while updating item in sensorItem.Update()\\n\")\r\n\t\tlog.Printf(\"Reason %v\\n\", updateErr)\r\n\t\treturn updateErr\r\n\t}\r\n\tlog.Printf(\"Product %s updated successfully in table\", gi.Sensorname)\r\n\treturn nil\r\n}", "func (s structModel)Update(table string)(query string, values[]interface{}, err error){\n\tif s.err != nil{\n\t\treturn \"\", nil, s.err\n\t}\n\tvar arrQuery []string\n\tquery = \"UPDATE \" + table + \" SET\"\n\tlistValues := make([]interface{}, 0)\n\tfor i, _ := range s.value{\n\t\tarrQuery = append(arrQuery, \" \" + s.key[i] + \"= $\" + strconv.Itoa(i+1))\n\t\tlistValues = append(listValues, s.value[i])\n\t}\n\tquery = query + strings.Join(arrQuery, \",\")\n\treturn query, listValues, nil\n}", "func (f *FakeInstance) Update(_ context.Context, _ string, _ *govultr.InstanceUpdateReq) (*govultr.Instance, *http.Response, error) {\n\tpanic(\"implement me\")\n}", "func (pmt *Payment) Update(db *bolt.DB) error {\n\treturn pmt.Create(db)\n}", "func (o *Vote) Update(exec boil.Executor, whitelist ...string) error {\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\to.UpdatedAt = currTime\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn err\n\t}\n\tkey := makeCacheKey(whitelist, nil)\n\tvoteUpdateCacheMut.RLock()\n\tcache, cached := voteUpdateCache[key]\n\tvoteUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(\n\t\t\tvoteColumns,\n\t\t\tvotePrimaryKeyColumns,\n\t\t\twhitelist,\n\t\t)\n\n\t\tif len(whitelist) == 0 {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"models: unable to update vote, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `vote` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, votePrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(voteType, voteMapping, append(wl, votePrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update vote row\")\n\t}\n\n\tif !cached {\n\t\tvoteUpdateCacheMut.Lock()\n\t\tvoteUpdateCache[key] = cache\n\t\tvoteUpdateCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpdateHooks(exec)\n}", "func (h *Handler) Update(c echo.Context) error {\n\trec := models.Rec{}\n\terr := c.Bind(&rec)\n\tif err != nil {\n\t\treturn c.JSON(http.StatusInternalServerError, err)\n\t}\n\n\terr = h.Store.Update(&rec)\n\tif err != nil {\n\t\treturn c.JSON(http.StatusInternalServerError, err)\n\t}\n\n\treturn c.JSON(http.StatusOK, rec)\n}", "func (db *InMemDatabase) Update(obj contrail.IObject, refs UIDList) error {\n\tuid := parseUID(obj.GetUuid())\n\tdata, ok := db.objectData[uid]\n\tif !ok {\n\t\treturn fmt.Errorf(\"Object %s: not in database\", obj.GetUuid())\n\t}\n\tdb.updateBackReferences(obj, refs, data.refs)\n\tdata.refs = refs\n\treturn nil\n}", "func (self boardModel) Update(db Database) BoardModel {\n // ignore error\n perpage, _ := db.GetThreadsPerPage(self.board)\n // refetch all on this page\n model := db.GetGroupForPage(self.prefix, self.frontend, self.board, self.page, int(perpage))\n var threads []ThreadModel\n for _, th := range model.Threads() {\n threads = append(threads, th.Update(db))\n }\n return boardModel{\n frontend: self.frontend,\n prefix: self.prefix,\n board: self.board,\n page: self.page,\n pages: self.pages,\n threads: threads,\n }\n}", "func (p *pollData) Update(poll Poll) (err error) {\n\n\tpollTags := GetStructTags(\"non-null\", \"db\", poll)\n\tpollFields := func(tags []string) string {\n\t\tvar temp []string\n\t\tfor _, tag := range tags {\n\t\t\ttemp = append(temp, fmt.Sprintf(`%s = :%s`, tag, tag))\n\t\t}\n\t\treturn strings.Join(temp, \", \")\n\t}(pollTags)\n\tquery := fmt.Sprintf(`Update polls SET %s WHERE id = :id`, pollFields)\n\tif _, err := rrsql.DB.NamedExec(query, poll); err != nil {\n\t\tlog.Fatal(err)\n\t\treturn err\n\t}\n\treturn nil\n}", "func (m *Model) updated() {\n\tif m.IsTimestamps() {\n\t\tnow := time.Now()\n\t\tm.UpdatedAt = &now\n\t\tm.applyTimestamps()\n\t}\n}", "func (r *Entity) Update() (result sql.Result, err error) {\n\twhere, args, err := gdb.GetWhereConditionOfStruct(r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn Model.Data(r).Where(where, args).Update()\n}", "func (m *Module) Update() (*Module, error) {\n\tif err := database.BackendDB.DB.Save(m).Error; err != nil {\n\t\tlog.Print(err)\n\t\treturn nil, err\n\t}\n\treturn m, nil\n}", "func (r *wagerRepositoryImpl) Update(wager *entity.Wager) error {\n\treturn r.db.Save(wager).Error\n}", "func (dao *Wager) Update(tx *gorm.DB, mod *models.Wager) error {\n\terr := tx.Save(mod).Error\n\tif err != nil {\n\t\treturn errs.Wrap(errs.ErrMySQLUpdate, err.Error(), \"tx.Save\")\n\t}\n\treturn nil\n}", "func (blt Bolt) Update(execute dbtx.Execute) error {\n\treturn blt.db.Update(func(tx *b.Tx) error {\n\t\treturn execute(tx.Bucket(blt.Bucket))\n\t})\n}", "func (c *Company) Update() {\n\tinitdb.DbInstance.Save(&c)\n\tlog.Println(\"Updated -> \", c)\n}", "func (self *DB) EmitUpdate(obj interface{}) (err error) {\n\tvalue := reflect.ValueOf(obj).Elem()\n\treturn self.emit(reflect.TypeOf(value.Interface()), &value, &value)\n}", "func (database *Database) Update(object interface{}) error {\n\tif err := database.DB.Model(object).Updates(object).Error; err != nil {\n\t\treturn fmt.Errorf(\"Failed to update '%s': %s\", reflect.TypeOf(object), err)\n\t}\n\treturn nil\n}", "func (u GroupUpdater) Update() error {\n\treturn u.db.Updates(u.fields).Error\n}", "func (r *resourceFrameworkShare) Update(ctx context.Context, req resource.UpdateRequest, resp *resource.UpdateResponse) {\n}", "func (c *Class) Update() error {\n\tvar err error\n\n\tif c.ID.Valid() == false {\n\t\treturn errmsg.ErrInvalidID\n\t}\n\n\tif err = c.save(); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (d *Dataset) Update(a *config.AppContext) error {\n\t/*\n\t * Then we will update\n\t */\n\n\t//updating the model\n\td.UserID = a.Session.User.ID\n\terr := a.Db.Model(d).Updates(map[string]interface{}{\n\t\t\"name\": d.Name,\n\t\t\"description\": d.Description,\n\t\t\"table_created\": d.TableCreated,\n\t\t\"datastore_id\": d.DatastoreID,\n\t}).Error\n\n\treturn err\n}", "func (t *DbService) Update(request *UpdateRequest) (*UpdateResponse, error) {\n\trsp := &UpdateResponse{}\n\treturn rsp, t.client.Call(\"db\", \"Update\", request, rsp)\n}", "func (s *stepsrvc) Update(ctx context.Context, p *step.StoredListOfSteps) (err error) {\n\ts.logger.Print(\"step.update\")\n\n\terr = s.db.UpdateOneStep(*p)\n\n\treturn err\n}", "func (w *wireguardServerConfig) Update(obj *api.WireguardServerConfig) error {\n\tobj.UpdatedAt = time.Now().UTC().Format(time.RFC3339)\n\tjsonData, err := json.Marshal(obj)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// if err := w.store.CreateBucketIfNotExists(obj.UID); err != nil {\n\t// \treturn err\n\t// }\n\treturn w.store.Set(path.Join(w.prefix, obj.UID), jsonData)\n}", "func (h *Hotel) Update(a *config.AppContext) error {\n\treturn a.Db.Model(h).Updates(map[string]interface{}{\n\t\t\"name\": h.Name,\n\t\t\"Description\": h.Description,\n\t\t\"GoogleLocation\": h.GoogleLocation,\n\t\t\"Address\": h.Address,\n\t\t\"Contact\": h.Contact,\n\t}).Error\n}", "func updateObject(object *remember.DataObject, connection *sql.DB) error {\n\tstatement, err := connection.Prepare(UPDATE)\n\tif nil != err {\n\t\treturn err\n\t}\n\tdefer statement.Close()\n\n\tobject.UpdatedAt = time.Now()\n\n\t_, err = statement.Exec(\n\t\tobject.Title,\n\t\tobject.GroupId,\n\t\tobject.Payload,\n\t\tobject.UpdatedAt.Unix(),\n\t\tobject.ID,\n\t)\n\n\treturn err\n}", "func (w WaitListStatusRepository) Update() (interface{}, error) {\n\tvar sqlStm = `SELECT update_waitlist_status( $1, $2, $3 )`\n\n\ttx, err := w.DB.Begin()\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"%s\", err)\n\t}\n\n\tdefer tx.Rollback()\n\n\tstmt, err := tx.Prepare(sqlStm)\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"%s\", err)\n\t}\n\n\tdefer stmt.Close()\n\n\tif _, err = stmt.Exec(\n\t\tw.Model.ID,\n\t\tw.Model.Description,\n\t\tw.Model.Value,\n\t); err != nil {\n\t\treturn nil, fmt.Errorf(\"%s\", err)\n\t}\n\n\tif err = tx.Commit(); err != nil {\n\t\treturn nil, fmt.Errorf(\"%s\", err)\n\t}\n\n\treturn w.Model, nil\n}", "func (s *PollVoteStore) Update(record *PollVote, cols ...kallax.SchemaField) (updated int64, err error) {\n\trecord.CreatedAt = record.CreatedAt.Truncate(time.Microsecond)\n\trecord.UpdatedAt = record.UpdatedAt.Truncate(time.Microsecond)\n\n\trecord.SetSaving(true)\n\tdefer record.SetSaving(false)\n\n\tif err := record.BeforeSave(); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn s.Store.Update(Schema.PollVote.BaseSchema, record, cols...)\n}", "func (v *Status) Update() error {\n\treturn nil\n}", "func (hg *Hg) Update() error {\n\treturn nil\n}", "func Update(db DB, table string, src interface{}) error {\n\treturn UpdateContext(context.Background(), db, table, src)\n}", "func (i *Instance) Update() error {\n\tvar err error\n\n\tif i.ID.Valid() == false {\n\t\treturn errmsg.ErrInvalidID\n\t}\n\n\tif err = i.save(); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (a *apiEndpoint) Update() error {\n\ta.e.mutex.RLock()\n\ta.SetEntry(\"server\", api.String(a.e.poolEntry.Desc))\n\ta.SetEntry(\"status\", api.String(a.e.status.String()))\n\tif a.e.lastErr != nil {\n\t\ta.SetEntry(\"last_error\", api.String(a.e.lastErr.Error()))\n\t\ta.SetEntry(\"last_error_time\", api.String(a.e.lastErrTime.Format(time.RFC3339)))\n\t} else {\n\t\ta.SetEntry(\"last_error\", api.Null)\n\t\ta.SetEntry(\"last_error_time\", api.Null)\n\t}\n\ta.SetEntry(\"pendingPayloads\", api.Number(a.e.NumPending()))\n\ta.SetEntry(\"publishedLines\", api.Number(a.e.LineCount()))\n\ta.SetEntry(\"averageLatency\", api.Float(a.e.AverageLatency()/time.Millisecond))\n\ta.e.mutex.RUnlock()\n\n\treturn nil\n}", "func (t *table) Update(router.Route) error {\n\treturn nil\n}", "func (u *UserModel) Update() error {\n\treturn DB.Self.Save(u).Error\n}", "func (s *HelloSystem) Update(ctx core.UpdateCtx) {}", "func (d *Database) Update(db DB, table string, src interface{}) error {\n\treturn d.UpdateContext(context.Background(), db, table, src)\n}", "func (wu *WsubUpdate) Update(db XODB) error {\n\tvar err error\n\n\t// if doesn't exist, bail\n\tif !wu._exists {\n\t\treturn errors.New(\"update failed: does not exist\")\n\t}\n\n\t// if deleted, bail\n\tif wu._deleted {\n\t\treturn errors.New(\"update failed: marked for deletion\")\n\t}\n\n\t// sql query\n\tconst sqlstr = `UPDATE jlabo.wsub_updates SET ` +\n\t\t`update_site_id = ?, extension_id = ?, name = ?, description = ?, element = ?, type = ?, folder = ?, client_id = ?, version = ?, data = ?, detailsurl = ?, infourl = ?, extra_query = ?` +\n\t\t` WHERE update_id = ?`\n\n\t// run query\n\tXOLog(sqlstr, wu.UpdateSiteID, wu.ExtensionID, wu.Name, wu.Description, wu.Element, wu.Type, wu.Folder, wu.ClientID, wu.Version, wu.Data, wu.Detailsurl, wu.Infourl, wu.ExtraQuery, wu.UpdateID)\n\t_, err = db.Exec(sqlstr, wu.UpdateSiteID, wu.ExtensionID, wu.Name, wu.Description, wu.Element, wu.Type, wu.Folder, wu.ClientID, wu.Version, wu.Data, wu.Detailsurl, wu.Infourl, wu.ExtraQuery, wu.UpdateID)\n\treturn err\n}", "func (r *reflectorStore) Update(obj interface{}) error {\n\treturn r.Add(obj)\n}", "func (v *Vessel) Update() (*Vessel, error) {\n\tdb := adaptors.DBConnector()\n\tdefer db.Close()\n\n\t_, notFoundErr := v.FetchOne()\n\n\tif notFoundErr != nil {\n\t\treturn v, notFoundErr\n\t}\n\n\terr := db.Table(\"vessels\").Where(\"id = ?\", v.ID).Updates(&v).Error\n\treturn v, err\n}", "func (o *Peer) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tpeerUpdateCacheMut.RLock()\n\tcache, cached := peerUpdateCache[key]\n\tpeerUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tpeerAllColumns,\n\t\t\tpeerPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"model: unable to update peers, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `peers` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, peerPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(peerType, peerMapping, append(wl, peerPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model: unable to update peers row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model: failed to get rows affected by update for peers\")\n\t}\n\n\tif !cached {\n\t\tpeerUpdateCacheMut.Lock()\n\t\tpeerUpdateCache[key] = cache\n\t\tpeerUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (u *usecase) Update() error {\n\t// Time execution\n\tstart := time.Now()\n\n\t// Creating context with timeout duration process\n\tctx, cancel := context.WithTimeout(context.Background(), 20*time.Minute)\n\tdefer cancel()\n\n\t// Get all archieve from scrapper repository\n\tarchieves, err := u.scrapperRepo.GetAllArchieve()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Create array (slice) of archieve object\n\tvar newArchieves []*model.Archieve\n\n\t// Check if archieve is exist in DB\n\tfor _, archieve := range archieves {\n\t\t// Get archieve from DB by archieve code\n\t\t_, err := u.mysqlRepo.GetArchieveByCode(ctx, archieve.Code)\n\n\t\t// if archieve not exist then add to newArchieve array (slice)\n\t\tif err == model.ErrDataNotFound {\n\t\t\t// Add archieve\n\t\t\tnewArchieves = append(newArchieves, archieve)\n\t\t\tlog.Printf(\"New archieve: %v\", archieve.Code)\n\t\t} else if err != nil && err != model.ErrDataNotFound {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Counter new journal from archieves\n\tvar totalNewJournal int\n\n\t// Get new archieves journals\n\tfor _, newArchieve := range newArchieves {\n\t\t// Get all journal from scrapper repository based on archieve\n\t\tjournals, err := u.scrapperRepo.GetAllJournalByArchieveObject(newArchieve)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// Append Journals into archieve object\n\t\tnewArchieve.Journals = journals\n\t\ttotalNewJournal += len(newArchieve.Journals)\n\t}\n\n\t// Check if there's new archieve then saved new archieve into DB\n\tif len(newArchieves) > 0 {\n\t\t// Insert new archieves into DB\n\t\tif err := u.mysqlRepo.BatchArchieves(ctx, newArchieves); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tlog.Printf(\"Added %v archieve and %v journal (%v)m\", len(newArchieves), totalNewJournal, time.Since(start).Minutes())\n\n\t// if there's no update then do nothing or finish pull data from archieve scrapper\n\treturn nil\n}", "func (s *permisoService) Update(id string, p *model.Permiso) error {\n\n\t// p.UpdatedAt = time.Now()\n\n\t// return s.service.Update(id, p)\n\n\treturn nil\n}", "func (r *DarwinTimetable) Update(f func(*bolt.Tx) error) error {\n\treturn r.db.Update(f)\n}", "func (r *RadioStation) Update() error {\n\treturn nil\n}", "func (r *Reportwallettb) Update() *ReportwallettbUpdateOne {\n\treturn (&ReportwallettbClient{config: r.config}).UpdateOne(r)\n}", "func (api *bucketAPI) SyncUpdate(obj *objstore.Bucket) error {\n\tif api.ct.objResolver != nil {\n\t\tlog.Fatal(\"Cannot use Sync update when object resolver is enabled on ctkit\")\n\t}\n\tnewObj := obj\n\tvar writeErr error\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\tnewObj, writeErr = apicl.ObjstoreV1().Bucket().Update(context.Background(), obj)\n\t}\n\n\tif writeErr == nil {\n\t\tapi.ct.handleBucketEvent(&kvstore.WatchEvent{Object: newObj, Type: kvstore.Updated})\n\t}\n\n\treturn writeErr\n}", "func update(updateStruct *Update) (bool, error) {\n\terr := updateStruct.Collection.UpdateId(updateStruct.Id, updateStruct.Data)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t\treturn false, err\n\t}\n\treturn true, nil\n}", "func (api *objectAPI) SyncUpdate(obj *objstore.Object) error {\n\tif api.ct.objResolver != nil {\n\t\tlog.Fatal(\"Cannot use Sync update when object resolver is enabled on ctkit\")\n\t}\n\tnewObj := obj\n\tvar writeErr error\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\tnewObj, writeErr = apicl.ObjstoreV1().Object().Update(context.Background(), obj)\n\t}\n\n\tif writeErr == nil {\n\t\tapi.ct.handleObjectEvent(&kvstore.WatchEvent{Object: newObj, Type: kvstore.Updated})\n\t}\n\n\treturn writeErr\n}", "func (a *Anime) Update() error {\n\tdata, err := json.Marshal(a)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn database.DB.Set(a.Key(), data, -1).Err()\n}", "func (table *ClassTable) Update(strId string, updates Class) (updated Class, err error) {\n\tdata, err := table.connection.Update(strId, CLASS_TABLE, updates)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = utilities.FillStruct(data, &updated)\n\treturn\n}", "func (c *Collection) Update(entry interface{}) error {\n\tkeyComponents, err := c.generateKey(entry)\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to update in collection. %s\", err.Error())\n\t}\n\n\tkey, err := c.formatKey(keyComponents)\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to update in collection. %s\", err.Error())\n\t}\n\n\texists, err := c.exists(key)\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to update in collection. %s\", err.Error())\n\t}\n\n\tif !exists {\n\t\treturn fmt.Errorf(\"Failed to update in collection. Key already exists\")\n\t}\n\n\tbytes, err := c.Serializer.ToBytes(entry)\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to update in collection. %s\", err.Error())\n\t}\n\n\tif c.Name != WorldStateIdentifier {\n\t\terr = c.Stub.PutPrivateData(c.Name, key, bytes)\n\t} else {\n\t\terr = c.Stub.PutState(key, bytes)\n\t}\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to update in collection. %s\", err.Error())\n\t}\n\n\treturn nil\n}", "func Update(db DB, table string, src interface{}) error {\n\treturn Default.Update(db, table, src)\n}", "func (o *PaymentObject) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tpaymentObjectUpdateCacheMut.RLock()\n\tcache, cached := paymentObjectUpdateCache[key]\n\tpaymentObjectUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tpaymentObjectAllColumns,\n\t\t\tpaymentObjectPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update payment_objects, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `payment_objects` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, paymentObjectPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(paymentObjectType, paymentObjectMapping, append(wl, paymentObjectPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update payment_objects row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for payment_objects\")\n\t}\n\n\tif !cached {\n\t\tpaymentObjectUpdateCacheMut.Lock()\n\t\tpaymentObjectUpdateCache[key] = cache\n\t\tpaymentObjectUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (tablePreeCheck *TablePreeCheck) UpdateModel() {\n\tdb := db.GetPostgresDb()\n\tdb.Save(tablePreeCheck)\n}", "func (factory ObjectFactoryImpl) UpdateWithTime(updatedInstance entity.Object, instance interface{}) entity.Object {\n\treflect.ValueOf(&updatedInstance).Elem().FieldByName(\"UpdatedAt\").SetString(generator.CurrentDateTime())\n\treflect.ValueOf(&updatedInstance).Elem().FieldByName(\"Extra\").Set(reflect.ValueOf(&instance))\n\treturn updatedInstance\n}", "func (m *File) Update(attr string, value interface{}) error {\n\treturn UnscopedDb().Model(m).UpdateColumn(attr, value).Error\n}", "func (t *Track) Update(i interface{}) error {\n\n\tinfo := i.(Track)\n\n\tif err := db.Model(&t).Updates(info).Error; err != nil {\n\t\tlogrus.Warn(err)\n\t\treturn err\n\t}\n\n\tlogrus.Infof(\"icii updated track #%x with new information\", t.ID)\n\n\treturn nil\n\n}", "func (u *CardRepository) Update(model *models.Card, data map[string]interface{}) error {\n\tquery := u.CardTable().Model(model).Updates(data)\n\tif err := query.Error; err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (m *MovieModel) Update(movie *Movie) error {\n\n\t/* potential to use uuid here\n\tUPDATE movies\n\tSET title = $1, year = $2, runtime = $3, genres = $4, version = uuid_generate_v4()\n\tWHERE id = $5 AND\n\t**/\n\t// Add version = $6, so we can stop race conditions\n\tquery := `\n\tUPDATE movies\n\tSET title = $1, year = $2, runtime = $3, genres = $4, version = version + 1\n\tWHERE id = $5 AND version = $6\n\tRETURNING version \n\t`\n\n\t// create the arg slice contaninig the values for the placeholder params.\n\targs := []interface{}{\n\t\tmovie.Title,\n\t\tmovie.Year,\n\t\tmovie.Runtime,\n\t\tpq.Array(movie.Genres),\n\t\tmovie.ID,\n\t\tmovie.Version, // Add the expected movie version\n\t}\n\n\tctx, cancel := context.WithTimeout(context.Background(), 3*time.Second)\n\tdefer cancel()\n\n\t// If no matching row could be found (version has been changed)\n\terr := m.DB.QueryRowContext(ctx, query, args...).Scan(&movie.Version)\n\tif err != nil {\n\t\tswitch {\n\t\tcase errors.Is(err, sql.ErrNoRows):\n\t\t\treturn ErrEditConflict\n\t\tdefault:\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (inst *Instance) fieldUpdate(t time.Time) {\n\tinst.lifePool.Update(t)\n}", "func (o *Status) Update() {\n o.Time = time.Now()\n}", "func (r *ItemsRepository) update(i *Item) error {\n\tif query := r.databaseHandler.DB().Save(&i); query.Error != nil {\n\t\treturn query.Error\n\t}\n\treturn nil\n}", "func (l *lead) Update() error {\n\tif err := l.Validate(); err != nil {\n\t\treturn err\n\t}\n\tl.UpdatedAt = bson.Now()\n\terr := config.LeadsCollection.UpdateId(l.ID, l)\n\treturn err\n}", "func (u *updater) Update(ctx context.Context, at int64, payload []byte) error {\n\terr := u.Put(ctx, &index{u.next}, at, payload)\n\tif err != nil {\n\t\treturn err\n\t}\n\tu.next++\n\treturn nil\n}", "func (u *updater) Update(ctx context.Context, at int64, payload []byte) error {\n\terr := u.Put(ctx, &index{u.next}, at, payload)\n\tif err != nil {\n\t\treturn err\n\t}\n\tu.next++\n\treturn nil\n}", "func (c client) Update(q Query) error {\n\treturn c.exec(q)\n}", "func (o *sampleUpdateHandler) Update(rw http.ResponseWriter, req *http.Request) {\n\to.UpdateHandler.Update(rw, req)\n}", "func Update(config TomlConfig, metaData toml.MetaData, db *DB) error {\n\t// Load data\n\tlog.Println(\"load data from database\")\n\tdataSet, err := db.ToDataSet()\n\tif err != nil {\n\t\treturn err\n\t}\n\t// Update popularity\n\tif err = UpdatePopularity(dataSet, db); err != nil {\n\t\treturn err\n\t}\n\t// Update popular items\n\tif err = UpdatePopItem(config.Recommend.CacheSize, db); err != nil {\n\t\treturn err\n\t}\n\t// Update latest items\n\tif err = UpdateLatest(config.Recommend.CacheSize, db); err != nil {\n\t\treturn err\n\t}\n\t// Generate recommends\n\tparams := config.Params.ToParams(metaData)\n\tif err = UpdateRecommends(config.Recommend.Model, params, config.Recommend.CacheSize, config.Recommend.FitJobs, config.Recommend.Once,\n\t\tdataSet, db); err != nil {\n\t\treturn err\n\t}\n\t// Generate neighbors\n\tif err = UpdateNeighbors(config.Recommend.Similarity, config.Recommend.CacheSize, dataSet, db); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (s *Dao) Update(ctx context.Context, kv *model.KVDoc) error {\n\tkeyKv := key.KV(kv.Domain, kv.Project, kv.ID)\n\tresp, err := etcdadpt.Get(ctx, keyKv)\n\tif err != nil {\n\t\topenlog.Error(err.Error())\n\t\treturn err\n\t}\n\tif resp == nil {\n\t\treturn datasource.ErrRecordNotExists\n\t}\n\n\tvar old model.KVDoc\n\terr = json.Unmarshal(resp.Value, &old)\n\tif err != nil {\n\t\topenlog.Error(err.Error())\n\t\treturn err\n\t}\n\told.LabelFormat = kv.LabelFormat\n\told.Value = kv.Value\n\told.Status = kv.Status\n\told.Checker = kv.Checker\n\told.UpdateTime = kv.UpdateTime\n\told.UpdateRevision = kv.UpdateRevision\n\n\tbytes, err := json.Marshal(old)\n\tif err != nil {\n\t\topenlog.Error(err.Error())\n\t\treturn err\n\t}\n\terr = etcdadpt.PutBytes(ctx, keyKv, bytes)\n\tif err != nil {\n\t\topenlog.Error(err.Error())\n\t\treturn err\n\t}\n\treturn nil\n}", "func UpdatePerson(db *sql.DB) {}", "func Update(db gorp.SqlExecutor, i interface{}) error {\n\treturn Mapper.Update(db, i)\n}", "func (bt *Tree) Update(fields []sql.Field, resolution time.Duration, truncateBefore time.Time, key []byte, vals encoding.TSParams, metadata bytemap.ByteMap) int {\n\tbytesAdded, newNode := bt.doUpdate(fields, resolution, truncateBefore, key, vals, metadata)\n\tbt.bytes += bytesAdded\n\tif newNode {\n\t\tbt.length++\n\t}\n\treturn bytesAdded\n}", "func (r *sampleRepository) Update(sample *model.Sample) error {\n\treturn r.DB.Save(sample).Error\n}", "func (contact *Contact) Update() error {\n\treturn DB().Save(contact).Error\n}", "func memberUpdate(db *sql.DB, client *chef.Client) {\n\t// Get a unique list of all the users\n\tusers, err := db.Query(\"SELECT user_name FROM org_groups;\")\n\tif err != nil {\n\t\tpanic(err.Error()) // proper error handling instead of panic in your app\n\t}\n\tvar members []string\n\tfor users.Next() {\n\t\tvar name string\n\t\terr = users.Scan(&name)\n\t\tif err != nil {\n\t\t\tpanic(err.Error()) // proper error handling instead of panic in your app\n\t\t}\n\t\tmembers = append(members, name)\n\t}\n\tmembers = co.Unique(members)\n\tusers.Close()\n\tstmtInsMember, err := db.Prepare(\"INSERT INTO members (user_name, email, display_name) VALUES( ?, ?, ? )\")\n\tif err != nil {\n\t\tpanic(err.Error()) // proper error handling instead of panic in your app\n\t}\n // transaction - begin, delete existing users, add replacement data\n\ttx, err := db.Begin()\n\t_, err = db.Query(\"DELETE FROM members;\")\n\tfor _, member := range members {\n\t\t// Extract information for each user\n\t\tmemberInfo := getMember(client, member)\n\t\t// Update the data base with a new set of user records\n\t\t_, err = stmtInsMember.Exec(member, memberInfo.Email, memberInfo.DisplayName)\n\t\tif err != nil {\n\t\t\tpanic(err.Error()) // proper error handling instead of panic in your app\n\t\t}\n\t}\n stmtInsMember.Close()\n\ttx.Commit()\n}", "func (s *System) Update(dt *world.Clock) error {\n\treturn s.save()\n}", "func (g *GameDBModel) Update(game Game) error {\n\terr := database.C(COLLECTION).UpdateId(game.ID, &game)\n\treturn err\n}", "func (a *paymentUsecase) Update(c context.Context, ar *models.Payment) (*models.Payment, error) {\n\tctx, cancel := context.WithTimeout(c, a.contextTimeout)\n\tdefer cancel()\n\n\tar.UpdatedAt = time.Now()\n\treturn a.repo.Update(ctx, ar)\n}" ]
[ "0.6663857", "0.64484656", "0.64176106", "0.6386076", "0.63846666", "0.63183945", "0.63183945", "0.63183945", "0.63183945", "0.63183945", "0.6280699", "0.6221633", "0.62214124", "0.6215622", "0.62033606", "0.6172274", "0.60822904", "0.6041722", "0.60226613", "0.6006191", "0.5973311", "0.594424", "0.5925653", "0.5913681", "0.59054744", "0.59006655", "0.58831435", "0.5865293", "0.585237", "0.5834893", "0.58343923", "0.58320564", "0.5819343", "0.58095765", "0.5806885", "0.580434", "0.57992995", "0.578961", "0.5777545", "0.5767551", "0.5765028", "0.5762621", "0.57521117", "0.5747183", "0.5733086", "0.5725354", "0.5723133", "0.57098615", "0.5707844", "0.5707807", "0.57078016", "0.5700633", "0.5690091", "0.5686742", "0.56739163", "0.5673069", "0.56708574", "0.56694186", "0.56617266", "0.56565106", "0.56437844", "0.56291825", "0.56101584", "0.56028455", "0.56024384", "0.56018156", "0.55962646", "0.5596241", "0.55872285", "0.5581596", "0.55801505", "0.55781966", "0.55739975", "0.55723584", "0.557142", "0.55706203", "0.5557911", "0.55575955", "0.55518204", "0.5548625", "0.55453146", "0.5540036", "0.55332077", "0.5533083", "0.55199516", "0.55109614", "0.5500758", "0.5500758", "0.5498367", "0.5490768", "0.5490634", "0.5488148", "0.54871565", "0.5487098", "0.5484795", "0.5483465", "0.54785866", "0.54785776", "0.5478565", "0.54749846", "0.54651105" ]
0.0
-1
Delete the struct user from the database
func (u *User) Delete() error { _, err := db.Todo.Exec( `DELETE FROM users where id = $1 or email = $2`, u.ID, u.Email) if err != nil { return errors.New("No user with specified ID or email to delete") } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (user *User) Delete() {\n\tdb := common.GetDatabase()\n\n\tdb.Delete(&user)\n}", "func DeleteUser(c *gin.Context) {}", "func DeleteUser(userid int64) error {\n _, err := model.Database.Exec(\"DELETE FROM users WHERE userid = ? AND isadmin = ?\", userid, false)\n if err != nil {\n return err\n }\n return nil\n}", "func (o *AuthUser) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthUser provided for delete\")\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), authUserPrimaryKeyMapping)\n\tsql := \"DELETE FROM `auth_user` WHERE `id`=?\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete from auth_user\")\n\t}\n\n\treturn nil\n}", "func Delete(id bson.ObjectId) error {\n\tdb, err := storm.Open(dbPath)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer db.Close()\n\n\tu := new(User)\n\terr = db.One(\"ID\", id, u)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// instead of returning the object, return the result of DeleteStruct(u)\n\treturn db.DeleteStruct(u)\n}", "func (user *User) Delete() error {\n\treturn db.Delete(user).Error\n}", "func (ug *userGorm) Delete(id uint) error{\n\tuser :=User{Model:gorm.Model{ID:id}}\n\treturn ug.db.Delete(&user).Error\n}", "func (model *ModelUser) DeleteUser(v ModelUser) error {\n\tcommon.Logger(\"info\", \"Initialize Get Database in PostgreSQL\", \"Modul User : DeleteUser\")\n\tdb := common.GetPostgreSQLDB()\n\ttx := db.Begin()\n\n\tcommon.Logger(\"info\", \"Prepare Delete Data at Database in PostgreSQL\", \"Modul User : DeleteUser\")\n\terr := db.Model(model).Delete(v).Error\n\ttx.Commit()\n\n\tcommon.Logger(\"info\", \"Finnished Update Data at Database in PostgreSQL\", \"Modul User : DeleteUser\")\n\treturn err\n}", "func (r User) Delete() error {\n\terr := db.C(\"user\").Remove(&r)\n\treturn err\n}", "func (u *User) Delete() error {\n\tconn, err := db.Connect()\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer conn.Close()\n\n\tsql := `DELETE FROM users WHERE id = $1`\n\tconn.QueryRow(sql, u.ID)\n\n\treturn nil\n}", "func (*userDAO) DeleteUser() error {\n\treturn DBInstance.Delete(&User{}).Error\n}", "func DeleteUser(db *pg.DB, pk int64) error {\n\n\tuser := models.User{ID: pk}\n\n\tfmt.Println(\"Deleting User...\")\n\n\terr := db.Delete(&user)\n\n\treturn err\n}", "func (ug *userGorm) Delete(id uint) error {\n\tuser := User{Model: gorm.Model{ID: id}}\n\treturn ug.db.Delete(&user).Error\n\n}", "func (u *User) Delete() error {\n\tvar s Storage\n\tdb := s.Init()\n\tdefer db.Close()\n\n\treturn db.Update(func(tx *bolt.Tx) error {\n\t\tbkt := tx.Bucket([]byte(\"USERS\"))\n\t\tid := u.ID.String()\n\t\terr := bkt.Delete([]byte(id))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t})\n}", "func DeleteUser(c *gin.Context) {\n\tid := c.Param(\"id\")\n\tlog.Printf(\"DeleteUser in db %v\", id)\n\tvar user models.User\n\tdb := db.GetDB()\n\n\tif err := db.Where(\"id = ?\", id).First(&user).Error; err != nil {\n\t\tc.AbortWithStatus(http.StatusNotFound)\n\t\tlog.Println(\"Failed to DeleteUser in db\")\n\t\treturn\n\t}\n\n\tdb.Delete(&user)\n}", "func DeleteUser(c *gin.Context) {\n\t// Get model if exist\n\tvar user models.User\n\tif err := models.DB.Where(\"id = ?\", c.Param(\"id\")).First(&user).Error; err != nil {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"error\": \"Record not found!\"})\n\t\treturn\n\t}\n\n\tmodels.DB.Delete(&user)\n\n\tc.JSON(http.StatusOK, gin.H{\"data\": true})\n}", "func deleteUser(w http.ResponseWriter, r *http.Request) {\r\n\tparams := mux.Vars(r)\r\n\tstmt, err := db.Prepare(\"DELETE FROM users WHERE id = ?\")\r\n\tif err != nil {\r\n\t\tpanic(err.Error())\r\n\t}\r\n\t_, err = stmt.Exec(params[\"id\"])\r\n\tif err != nil {\r\n\t\tpanic(err.Error())\r\n\t}\r\n\tfmt.Fprintf(w, \"User with id = %s was deleted\", params[\"id\"])\r\n}", "func deleteUser(c *gin.Context) {\n\tvar user user\n\tuserID := c.Param(\"id\")\n\n\tdb.First(&user, userID)\n\n\tif user.Id == 0 {\n\t\tc.JSON(http.StatusNotFound, gin.H{\"status\": http.StatusNotFound, \"message\": \"No user found!\"})\n\t\treturn\n\t}\n\n\tdb.Delete(&user)\n\tc.JSON(http.StatusOK, gin.H{\"status\": http.StatusOK, \"message\": \"User deleted successfully!\"})\n}", "func (ug *userDbHandle) Delete(id uint) error {\n\tuser := User{Model: gorm.Model{ID: id}}\n\treturn ug.db.Delete(&user).Error\n}", "func (ug *userGorm) Delete(id uint) error {\n\tuser := User{Model: gorm.Model{ID: id}}\n\treturn ug.db.Delete(&user).Error\n}", "func (ug *userGorm) Delete(id uint) error {\n\tuser := User{Model: gorm.Model{ID: id}}\n\treturn ug.db.Delete(&user).Error\n}", "func DeleteUser(id int) (err error) {\n\to := orm.NewOrm()\n\tv := User{Id: id}\n\t// ascertain id exists in the database\n\tif err = o.Read(&v); err == nil {\n\t\tvar num int64\n\t\tif num, err = o.Delete(&User{Id: id}); err == nil {\n\t\t\tfmt.Println(\"Number of records deleted in database:\", num)\n\t\t}\n\t}\n\treturn\n}", "func DeleteSecUser(id int64) (err error) {\n\to := orm.NewOrm()\n\tv := SecUser{Id: id}\n\t// ascertain id exists in the database\n\tif err = o.Read(&v); err == nil {\n\t\tvar num int64\n\t\tif num, err = o.Delete(&SecUser{Id: id}); err == nil {\n\t\t\tfmt.Println(\"Number of records deleted in database:\", num)\n\t\t}\n\t}\n\treturn\n}", "func (u *User) Delete(tx *sql.Tx) error {\n\tif u.ID == \"\" {\n\t\treturn errors.New(`user ID is not valid`)\n\t}\n\tlog.Printf(\"db.User.Delete %s\", u.ID)\n\n\tstmt := bytes.Buffer{}\n\tstmt.WriteString(`DELETE FROM `)\n\tstmt.WriteString(userTable)\n\tstmt.WriteString(` WHERE id = ?`)\n\tlog.Printf(\"SQL QUERY: %s: with values %s\", stmt.String(), u.ID)\n\n\t_, err := tx.Exec(stmt.String(), u.ID)\n\n\treturn err\n}", "func Delete() error {\n\tuser := &Users{}\n\ti, err := engine.Id(1).Delete(user)\n\tif err == nil {\n\t\treturn nil\n\t} else if i <= 0 {\n\t\treturn errors.New(\"删除失败\")\n\t}\n\n\treturn nil\n}", "func DeleteUser(id int) {\n\tvar i int\n\ti = GetIndexOfUser(id)\n\tDeleteUserFromDatabase(i)\n}", "func DeleteUser(dbmap *gorp.DbMap, id string) error {\n\tvar u User\n\terr := dbmap.SelectOne(&u, \"SELECT * FROM user WHERE object_id = ?\", id)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttx, err := dbmap.Begin()\n\tif err != nil {\n\t\treturn nil\n\t}\n\n\t_, err = tx.Exec(\"DELETE FROM user_session WHERE user_id = ?;\", u.PK)\n\tif err != nil {\n\t\ttx.Rollback()\n\t\treturn err\n\t}\n\t_, err = tx.Exec(\"DELETE FROM user_role WHERE user_id = ?;\", u.PK)\n\tif err != nil {\n\t\ttx.Rollback()\n\t\treturn err\n\t}\n\t_, err = tx.Exec(\"DELETE FROM domain_user WHERE user_id = ?;\", u.PK)\n\tif err != nil {\n\t\ttx.Rollback()\n\t\treturn err\n\t}\n\t_, err = tx.Delete(&u)\n\tif err != nil {\n\t\ttx.Rollback()\n\t\treturn err\n\t}\n\treturn tx.Commit()\n}", "func (self Users) Delete() {\n\tsqlStatement := `DELETE FROM users WHERE id = $1`\n\t_, err := self.DB.Exec(sqlStatement, self.Id)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func (uv *userValidator) Delete(id uint) error {\n\tvar user User\n\tuser.ID = id\n\terr := runUserValFuncs(&user, uv.idGreaterThan(0))\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn uv.UserDB.Delete(id)\n}", "func (uv *userValidator) Delete(id uint) error {\n\tvar user User\n\tuser.ID = id\n\terr := runUserValFuncs(&user, uv.idGreaterThan(0))\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn uv.UserDB.Delete(id)\n}", "func (c *UserRepoImpl) Delete(id int) (*model.User, error) {\n\n\tuser := new(model.User)\n\n\tif err := c.db.Table(\"user\").First(&user, id).Error; err != nil {\n\t\treturn nil, errors.New(\"id is doesnt exists\")\n\t}\n\n\tif err := c.db.Table(\"user\").Where(\"user_id = ?\", id).Delete(&model.User{}).Error; err != nil {\n\t\treturn nil, errors.New(\"delete courier data: error\")\n\t}\n\n\treturn nil, nil\n}", "func (uv *userValidator) Delete(id uint) error{\n\tvar user User\n\tuser.ID = id\n\terr := runUserValidatorFunction(&user, uv.idGreaterThan(0))\n\tif err != nil{\n\t\treturn err\n\t}\n\treturn uv.UserDB.Delete(id)\n}", "func (user *User) Delete() *errors.RestErr {\n\t//prepare and execute the delete query\n\tstmt, err := usersdb.Client.Prepare(queryDeleteUser)\n\tif err != nil {\n\t\treturn errors.NewInternalServerError(err.Error())\n\t}\n\tdefer stmt.Close()\n\n\t//\n\tif _, err = stmt.Exec(user.ID); err != nil {\n\t\treturn errors.ParseError(err)\n\t}\n\n\treturn nil\n\n}", "func (u *User) Delete() *errorsutils.RestErr {\n\tstmt, err := usersdb.Client.Prepare(queryDeleteUser)\n\tif err != nil {\n\t\tlogger.Error(\"error when trying to prepare delete user statement\", err)\n\t\treturn errorsutils.NewInternalServerError(\"database error\", errors.New(\"database error\"))\n\t}\n\tdefer stmt.Close()\n\n\tif _, err = stmt.Exec(u.ID); err != nil {\n\t\tlogger.Error(\"error when trying to delete user\", err)\n\t\treturn errorsutils.NewInternalServerError(\"database error\", errors.New(\"database error\"))\n\t}\n\n\treturn nil\n}", "func DeleteUser(c *gin.Context) {\n\tuuid := c.Params.ByName(\"uuid\")\n\tvar user models.User\n\tdb := db.GetDB()\n\tif uuid != \"\" {\n\n\t\tjwtClaims := jwt.ExtractClaims(c)\n\t\tauthUserAccessLevel := jwtClaims[\"access_level\"].(float64)\n\t\tauthUserUUID := jwtClaims[\"uuid\"].(string)\n\t\tif authUserAccessLevel != 1 {\n\t\t\tif authUserUUID != uuid {\n\t\t\t\tc.AbortWithStatusJSON(http.StatusUnauthorized, gin.H{\n\t\t\t\t\t\"error\": \"Sorry but you can't delete user, ONLY admins can\",\n\t\t\t\t})\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\t// DELETE FROM users WHERE uuid= user.uuid\n\t\t// exemple : UPDATE users SET deleted_at=date.now WHERE uuid = user.uuid;\n\t\tif err := db.Where(\"uuid = ?\", uuid).Delete(&user).Error; err != nil {\n\t\t\t// error handling...\n\t\t\tc.AbortWithStatusJSON(http.StatusBadRequest, gin.H{\n\t\t\t\t\"error\": err.Error(),\n\t\t\t})\n\t\t\treturn\n\t\t}\n\n\t\t// Display JSON result\n\t\t// c.JSON(200, gin.H{\"success\": \"User #\" + uuid + \" deleted\"})\n\t\tc.JSON(200, gin.H{\"success\": \"User successfully deleted\"})\n\t} else {\n\t\t// Display JSON error\n\t\tc.JSON(404, gin.H{\"error\": \"User not found\"})\n\t}\n\n}", "func db_delete_user(username string) {\n file_path := path.Join(\"db/users\", strings.ToLower(username) + \".json\")\n\n err := os.Remove(file_path)\n \n if err != nil {\n fmt.Println(err.Error())\n return\n }\n fmt.Println(\"User Removed: \", username)\n}", "func (m *Manager) Delete(ctx context.Context, tx *sql.Tx, user v0.User) error {\n\tresult, err := tx.ExecContext(ctx, `\n\t\t\t\tUPDATE users \n\t\t\t\tSET deleted_at = ?, \n\t\t\t\tdeleted = TRUE \n\t\t\t\tWHERE primary_public_key = ?;`,\n\t\ttime.Now(),\n\t\tuser.PrimaryPublicKey,\n\t)\n\tif err != nil {\n\t\treturn err\n\t}\n\taffected, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn err\n\t}\n\tif affected == 0 {\n\t\treturn ErrUserDoesNotExists\n\t}\n\treturn nil\n}", "func DeleteUser(props neoism.Props) error {\n\n\t// return error if user is not found in the database\n\tif u, _ := FindUser(props); u == nil {\n\t\treturn errors.New(\"user not found\")\n\t}\n\n\tdb := vantaadb.Connect()\n\tcq := neoism.CypherQuery{\n\t\tStatement: `MATCH (u:User)\n OPTIONAL MATCH (s:Session)-[r]->(u)\n WHERE ` + vantaadb.PropString(\"u\", props) + `DELETE u, s, r`,\n\t\tParameters: props,\n\t}\n\tif err := db.Cypher(&cq); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func deleteUser(res http.ResponseWriter, req *http.Request, p httprouter.Params) {\n\t_, err := db.Exec(`\n\t\tDELETE FROM accounts\n\t\tWHERE username = $1;`, p.ByName(\"username\"),\n\t)\n\tif err != nil {\n\t\tlog.Println(\"deleteUser:\", err)\n\t}\n\n\twriteJSON(res, 200, jsMap{\"status\": \"OK\"})\n}", "func DeleteUser(w http.ResponseWriter, r *http.Request) {\r\n\tdefer r.Body.Close()\r\n\tuser := r.Context().Value(\"user\").(string)\r\n\r\n\tif err := dao.DBConn.RemoveUserByEmail(user); err != nil {\r\n\t\tlog.Println(err)\r\n\t\tu.RespondWithError(w, http.StatusBadRequest, \"User doesn't exist or has already been deleted\")\r\n\t\treturn\r\n\t}\r\n\r\n\tif err := dao.DBConn.RemoveUserExpenses(user); err != nil {\r\n\t\tlog.Println(err)\r\n\t\tu.RespondWithError(w, http.StatusBadRequest, \"User doesn't exist or has already been deleted\")\r\n\t\treturn\r\n\t}\r\n\r\n\tu.RespondWithJSON(w, http.StatusOK, \"User deleted\")\r\n}", "func (m *User) Delete() (err error) {\n\t_, err = orm.NewOrm().Delete(m)\n\treturn\n}", "func (u *User) Delete(db XODB) error {\n\tvar err error\n\n\t// if doesn't exist, bail\n\tif !u._exists {\n\t\treturn nil\n\t}\n\n\t// if deleted, bail\n\tif u._deleted {\n\t\treturn nil\n\t}\n\n\t// sql query\n\tconst sqlstr = `DELETE FROM test_database.users WHERE user_id = ?`\n\n\t// run query\n\tXOLog(sqlstr, u.UserID)\n\t_, err = db.Exec(sqlstr, u.UserID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set deleted\n\tu._deleted = true\n\n\treturn nil\n}", "func (s *Server) deleteUser(request *restful.Request, response *restful.Response) {\n\t// Authorize\n\tif !s.auth(request, response) {\n\t\treturn\n\t}\n\t// get user-id and put into temp\n\tuserId := request.PathParameter(\"user-id\")\n\tif err := s.dataStore.DeleteUser(userId); err != nil {\n\t\tinternalServerError(response, err)\n\t\treturn\n\t}\n\tok(response, Success{RowAffected: 1})\n}", "func (uv *userValidator) Delete(id uint) error {\r\n\tvar user User\r\n\tuser.ID = id\r\n\terr := runUserValFns(&user, uv.idGreaterThan(0))\r\n\tif err != nil {\r\n\t\treturn err\r\n\t}\r\n\treturn uv.UserDB.Delete(id)\r\n}", "func DeleteUser(id int) error {\n\tuser := User{ID: id}\n\t_, err := db.Model(&user).WherePK().Delete()\n\treturn err\n}", "func (u *UserModel) Delete() error {\n\treturn database.DB().Delete(u).Error\n}", "func (uv *userValidator) Delete(id uint) error {\n\tvar user User\n\tuser.ID = id\n\terr := runUserValFns(&user, uv.idGreaterThan(0))\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn uv.UserDB.Delete(id)\n}", "func (uv *userValidator) Delete(id uint) error {\n\tvar user User\n\tuser.ID = id\n\terr := runUserValFns(&user, uv.idGreaterThan(0))\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn uv.UserDB.Delete(id)\n}", "func (uv *userValidator) Delete(id uint) error {\n\tvar user User\n\tuser.ID = id\n\terr := runUserValFns(&user, uv.idGreaterThan(0))\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn uv.UserDB.Delete(id)\n}", "func DeleteUser(w http.ResponseWriter, req *http.Request) {\n\tvars := mux.Vars(req)\n\tid := vars[\"id\"]\n\n\tif err := db.Remove(id); err != nil {\n\t\thandleError(err, \"Failed to remove User: %v\", w)\n\t\treturn\n\t}\n\n\tw.Write([]byte(\"OK\"))\n}", "func DeleteUser(person *Person, id string) (err error) {\n\tConfig.DB.Where(\"id = ?\", id).Delete(person)\n\treturn nil\n}", "func DeleteUser(user *User) {\n\tDb.Delete(&user)\n}", "func DeleteUser(c *gin.Context) {\n\tnID := c.Param(\"user_id\")\n\tdb := dbConn()\n\tstatement, _ := db.Prepare(\"CALL delete_user(?)\")\n\tstatement.Exec(nID)\n\tdefer db.Close()\n}", "func Delete() error {\n\ti, err := orm.SetTable(\"tb_user\").SetPK(\"uid\").Where(\"name=$1 and uid>$2\", \"viney\", 3).DeleteRow()\n\tif err == nil {\n\t\tfmt.Println(i)\n\t\treturn nil\n\t}\n\treturn err\n}", "func (s *Sqlite) DeleteUser(user gogios.User) error {\n\tdb, err := s.openConnection()\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer db.Close()\n\n\tdb.Where(\"Username = ?\", user.Username).Update(\"password\", \"\")\n\tdb.Where(\"Username = ?\", user.Username).Delete(&user)\n\n\treturn nil\n}", "func DeleteUser(user *models.User, id string) (err error) {\n\tconfig.DB.Where(\"id = ?\", id).Delete(user)\n\treturn nil\n}", "func DeleteUser(c *gin.Context) {\n\tuserID := c.Param(\"userID\")\n\tuser := &userModel.User{ID: userID}\n\n\terr := dbConnect.Delete(user)\n\tif err != nil {\n\t\tlog.Printf(\"Error while deleting a single user, Reason: %v\\n\", err)\n\t\tc.JSON(http.StatusInternalServerError, gin.H{\n\t\t\t\"status\": http.StatusInternalServerError,\n\t\t\t\"message\": \"Something went wrong\",\n\t\t})\n\t\treturn\n\t}\n\n\tc.JSON(http.StatusOK, gin.H{\n\t\t\"status\": http.StatusOK,\n\t\t\"message\": \"User deleted successfully\",\n\t})\n\treturn\n}", "func DeleteUser(ctx iris.Context) {\n\tvar (\n\t\tuser model.User\n\t\tresult iris.Map\n\t)\n\tid := ctx.Params().Get(\"id\") // get id by params\n\tdb := config.GetDatabaseConnection()\n\tdefer db.Close()\n\terr := db.First(&user, id).Error\n\tif err != nil {\n\t\tresult = iris.Map{\n\t\t\t\"error\": \"true\",\n\t\t\t\"status\": iris.StatusBadRequest,\n\t\t\t\"message\": \"User not found\",\n\t\t\t\"result\": nil,\n\t\t}\n\t}\n\n\terr = db.Where(\"id = ?\", id).Delete(&user, id).Error\n\tif err != nil {\n\t\tresult = iris.Map{\n\t\t\t\"error\": \"true\",\n\t\t\t\"status\": iris.StatusBadRequest,\n\t\t\t\"message\": \"Failed Delete user\",\n\t\t\t\"result\": err.Error(),\n\t\t}\n\t} else {\n\t\tresult = iris.Map{\n\t\t\t\"error\": \"false\",\n\t\t\t\"status\": iris.StatusOK,\n\t\t\t\"message\": \"Failed Delete user\",\n\t\t\t\"result\": nil,\n\t\t}\n\t}\n\tctx.JSON(result)\n\treturn\n}", "func (s *Server) deleteUser(request *restful.Request, response *restful.Response) {\n\t// Authorize\n\tif !s.auth(request, response) {\n\t\treturn\n\t}\n\t// get user-id and put into temp\n\tuserId := request.PathParameter(\"user-id\")\n\tif err := s.DataStore.DeleteUser(userId); err != nil {\n\t\tinternalServerError(response, err)\n\t\treturn\n\t}\n\tok(response, Success{RowAffected: 1})\n}", "func (g *Game) deleteUser(user *User) {\n\tdelete(g.Players, (*user).Name)\n\tlog.Println(\"Deleted user \" + (*user).Name)\n}", "func deleteUser(userID int) error {\n\tdb, err := sql.Open(\"mysql\", DB_USER_NAME+\":\"+DB_PASSWORD+\"@unix(/var/run/mysql/mysql.sock)/\"+DB_NAME)\n\tif err != nil {\n\t\treturn errors.New(\"No connection\")\n\t}\n\n\tres, err := db.Exec(\"delete from Users where UserID=? and not exists(select 1 from StudentCourses where Student=? limit 1)\", userID, userID)\n\n\tif err != nil {\n\t\treturn errors.New(\"User is currently enrolled in a class. Please remove the student from the class before deleting the user.\")\n\t}\n\trowsAffected, err := res.RowsAffected()\n\n\tif rowsAffected != 1 {\n\t\treturn errors.New(\"Query didn't match any users.\")\n\t}\n\n\treturn nil\n}", "func (userRepository UserRepository) Delete(userId uint64) error {\n\tstatement, err := userRepository.db.Prepare(\n\t\t\"delete from users where id = ?\",\n\t)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer statement.Close()\n\n\tif _, err = statement.Exec(userId); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (u *User) Delete(userId string) (err error) {\n\tdb := G.DB()\n\t_, err = db.Exec(db.AR().Update(Table_User_Name, map[string]interface{}{\n\t\t\"is_delete\": User_Delete_False,\n\t\t\"update_time\": time.Now().Unix(),\n\t}, map[string]interface{}{\n\t\t\"user_id\": userId,\n\t}))\n\tif err != nil {\n\t\treturn\n\t}\n\treturn\n}", "func Delete(c *gin.Context) {\n\tuserID, err := getUserID(c.Param(\"user_id\"))\n\tif err != nil {\n\t\tc.JSON(err.Status, err)\n\t\treturn\n\t}\n\n\tif err := services.UserServ.DeleteUser(userID); err != nil {\n\t\tc.JSON(err.Status, err)\n\t\treturn\n\t}\n\tc.JSON(http.StatusOK, map[string]string{\"status\": \"deleted\"})\n}", "func (c *UserStore) Delete(ctx context.Context, u *chronograf.User) error {\n\treturn c.Ctrl.DeleteUser(ctx, u.Name)\n}", "func (u *User) Delete() error {\n\treturn DB.DeleteUser(u)\n}", "func DeleteUserDB(name string) error {\n\tsession, err := mgo.Dial(local)\n\tif err != nil {\n\t\tfmt.Println(\"Problema conexao\")\n\t\treturn err\n\t}\n\tdefer session.Close()\n\tc := session.DB(db).C(collection)\n\terr = c.Remove(bson.M{\"nome\": name})\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (us *UserStorage) DeleteUser(id string) error {\n\treturn nil\n}", "func (s *UserStore) Delete(record *User) error {\n\treturn s.Store.Delete(Schema.User.BaseSchema, record)\n}", "func (us *UserStorage) DeleteUser(id string) error {\n\tif !bson.IsObjectIdHex(id) {\n\t\treturn model.ErrorWrongDataFormat\n\t}\n\ts := us.db.Session(UsersCollection)\n\tdefer s.Close()\n\n\terr := s.C.RemoveId(bson.ObjectIdHex(id))\n\treturn err\n}", "func (uv *userValidator) Delete(id uint) error {\n\tu := User{Model: gorm.Model{ID: id}}\n\n\tif err := runUserValFuncs(&u, uv.isGreaterThan(0)); err != nil {\n\t\treturn err\n\t}\n\n\treturn uv.UserDB.Delete(id)\n}", "func DeleteUser(c *gin.Context) {\n\tvar user Models.User\n\tid := c.Params.ByName(\"id\")\n\terr := Models.DeleteUser(&user, id)\n\tif err != nil {\n\t\tc.AbortWithStatus(http.StatusNotFound)\n\t} else { \n\t\tc.JSON(http.StatusOK, gin.H{\"id\":\"is deleted\"})\n\t}\n}", "func (uv *userValidator) Delete(id uint) error {\n\tif id == 0 {\n\t\treturn ErrIDInvalid\n\t}\n\treturn uv.UserDB.Delete(id)\n}", "func (mss *MySQLStore) Delete(id int64) error {\n\tinsq := \"delete from users where user_id=?\"\n\t_, err := mss.Client.Exec(insq, id)\n\tif err != nil {\n\t\treturn ErrUserNotFound\n\t}\n\treturn nil\n}", "func (db Database) DeleteUser(username string) error {\n\treturn errors.New(\"I'm not implemented yet\")\n}", "func DeleteUser(pid int) {\n\tstmt, err := database.Prepare(`DELETE FROM USER_INFO WHERE pid = ?;`)\n\tdefer stmt.Close()\n\tif err != nil {\n\t\tprintln(err.Error())\n\t}\n\n\t_, err = stmt.Exec(pid)\n\tif err != nil {\n\t\tprintln(err.Error())\n\t}\n}", "func DeleteUser(db *gorm.DB, w http.ResponseWriter, r *http.Request) {\n\tvars := mux.Vars(r)\n\n\tid := vars[\"id\"]\n\tuser := getUserByID(db, id, w, r)\n\tif user == nil {\n\t\treturn\n\t}\n\tif err := db.Delete(&user).Error; err != nil {\n\t\tRespondError(w, http.StatusInternalServerError, \"\")\n\t\treturn\n\t}\n\tRespondJSON(w, http.StatusNoContent, nil)\n}", "func (u *User) Delete(c echo.Context, userid string) error {\n\treturn u.udb.Delete(u.db, u.ce, userid)\n}", "func (ua *UserAuth) Delete(ctx context.Context, db XODB) error {\n\tvar err error\n\n\t// if doesn't exist, bail\n\tif !ua._exists {\n\t\treturn nil\n\t}\n\n\t// if deleted, bail\n\tif ua._deleted {\n\t\treturn nil\n\t}\n\n\t// sql query\n\tconst sqlstr = `DELETE FROM user_auths WHERE user_id = ?`\n\n\t// run query\n\tXOLog(sqlstr, ua.UserID)\n\t_, err = db.ExecContext(ctx, sqlstr, ua.UserID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set deleted\n\tua._deleted = true\n\n\treturn nil\n}", "func UserDelete(w http.ResponseWriter, r *http.Request) {\n\n\t// Init output\n\toutput := []byte(\"\")\n\n\t// Add content type header to the response\n\tcontentType := \"application/json\"\n\tcharset := \"utf-8\"\n\tw.Header().Add(\"Content-Type\", fmt.Sprintf(\"%s; charset=%s\", contentType, charset))\n\n\t// Grab context references\n\trefStr := gorillaContext.Get(r, \"str\").(stores.Store)\n\t// Grab url path variables\n\turlVars := mux.Vars(r)\n\turlUser := urlVars[\"user\"]\n\n\tuserUUID := auth.GetUUIDByName(urlUser, refStr)\n\n\terr := auth.RemoveUser(userUUID, refStr)\n\tif err != nil {\n\t\tif err.Error() == \"not found\" {\n\t\t\terr := APIErrorNotFound(\"User\")\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\t\terr := APIErrGenericInternal(err.Error())\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// Write empty response if anything ok\n\trespondOK(w, output)\n\n}", "func (ur *UserRepository) Delete(ctx context.Context, id uint) error {\n\tq := `\n\tDELETE FROM users WHERE id=$1;\n\t`\n\n\tstmt, err := ur.Data.DB.PrepareContext(ctx, q)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdefer stmt.Close()\n\n\t_, err = stmt.ExecContext(ctx, id)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (ur *UserRepository) Delete(ctx context.Context, id uint) error {\n\tq := `\n\tDELETE FROM users WHERE id=$1;\n\t`\n\n\tstmt, err := ur.Data.DB.PrepareContext(ctx, q)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdefer stmt.Close()\n\n\t_, err = stmt.ExecContext(ctx, id)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func Delete(c *gin.Context) {\n\tuserId, idErr := getUserID(c.Param(\"user_id\"))\n\tif idErr != nil {\n\t\tc.JSON(idErr.Status, idErr)\n\t\treturn\n\t}\n\n\tif err := services.UsersService.DeleteUser(userId); err != nil {\n\t\tc.JSON(err.Status, err)\n\t\treturn\n\t}\n\tc.JSON(http.StatusOK, map[string]string{\"status\": \"deleted\"})\n}", "func (c *SQLiteConn) AuthUserDelete(username string) error {\n\t// NOOP\n\treturn nil\n}", "func DeleteUser(w http.ResponseWriter, r *http.Request) {\n\tfmt.Println(\"Delete user endpoint hit\")\n\t\n\tvars := mux.Vars(r)\n\n\tid := vars[\"id\"]\n\n\tvar user models.User\n\n\tmessage := user.Destroy(id)\n\n json.NewEncoder(w).Encode(message)\n}", "func (wu *WxUser) Delete(ctx context.Context, key ...interface{}) error {\n\tvar err error\n\tvar dbConn *sql.DB\n\n\t// if deleted, bail\n\tif wu._deleted {\n\t\treturn nil\n\t}\n\n\ttx, err := components.M.GetConnFromCtx(ctx)\n\tif err != nil {\n\t\tdbConn, err = components.M.GetMasterConn()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\ttableName, err := GetWxUserTableName(key...)\n\tif err != nil {\n\t\treturn err\n\t}\n\t//1\n\n\t// sql query with composite primary key\n\tsqlstr := `UPDATE ` + tableName + ` SET is_del = 1 WHERE id = ?`\n\n\t// run query\n\tutils.GetTraceLog(ctx).Debug(\"DB\", zap.String(\"SQL\", fmt.Sprint(sqlstr, wu.ID)))\n\tif tx != nil {\n\t\t_, err = tx.Exec(sqlstr, wu.ID)\n\t} else {\n\t\t_, err = dbConn.Exec(sqlstr, wu.ID)\n\t}\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set deleted\n\twu._deleted = true\n\n\treturn nil\n}", "func DeleteUser(user *entity.User, id string, client *statsd.Client) (err error) {\n\tt := client.NewTiming()\n\tif config.DB.Where(\"id = ?\", id).First(&user); user.ID == \"\" {\n\t\treturn errors.New(\"the user doesn't exist!!!\")\n\t}\n\tconfig.DB.Where(\"id = ?\", id).Delete(&user)\n\tt.Send(\"delete_user.query_time\")\n\treturn nil\n}", "func DeleteUser(c *gin.Context) {\n\tvar json db.UserDeleteForm\n\tif err := c.ShouldBind(&json); err != nil {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\n\t\t\t\"msg\": \"Form doens't bind.\",\n\t\t\t\"err\": err.Error(),\n\t\t})\n\t\treturn\n\t}\n\tsession := sessions.Default(c)\n\tuserID := session.Get(\"userID\")\n\tvar user db.Users\n\tif err := db.DB.Where(userID).\n\t\tFirst(&user).Error; gorm.IsRecordNotFoundError(err) {\n\t\t// User not found\n\t\tc.JSON(http.StatusOK, gin.H{\n\t\t\t\"msg\": \"User not found in database.\",\n\t\t\t\"err\": err,\n\t\t})\n\t\treturn\n\t}\n\tif checkPasswordHash(json.Password, user.Password) {\n\t\tsession.Clear()\n\t\tsession.Save()\n\t\t// Soft delete user\n\t\tdb.DB.Where(userID).Delete(&user)\n\t\tc.JSON(http.StatusOK, gin.H{\n\t\t\t\"msg\": user.Username,\n\t\t\t\"err\": \"\",\n\t\t})\n\t} else {\n\t\tc.JSON(http.StatusUnauthorized, gin.H{\n\t\t\t\"msg\": fmt.Sprintf(\"Check password hash failed for user %s\", user.Username),\n\t\t\t\"err\": user.Username,\n\t\t})\n\t}\n}", "func (u *User) Destroy() error {\n\n\tdb, err := sql.Open(u.db.Driver(), u.db.Name())\n\tif err != nil {\n\t\tfmt.Println(\"Destroy 1:\", err)\n\t\treturn err\n\t}\n\tdefer db.Close()\n\n\tstmt, err := db.Prepare(deleteUserById)\n\tif err != nil {\n\t\tfmt.Println(\"Destroy 2:\", err)\n\t\treturn err\n\t}\n\tdefer stmt.Close()\n\n\t_, err = stmt.Exec(u.id)\n\tif err != nil {\n\t\tfmt.Println(\"Destroy 3:\", err)\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (app *App) deleteUser(w http.ResponseWriter, r *http.Request) {\n\tvars := mux.Vars(r)\n\tid, err := strconv.Atoi(vars[\"id\"])\n\tif err != nil {\n\t\trespondWithError(w, http.StatusInternalServerError, err.Error())\n\t\treturn\n\t}\n\n\tuser := &users.User{ID: int64(id)}\n\terr = user.DeleteUser(app.Db)\n\tif err != nil {\n\t\trespondWithError(w, http.StatusNotFound, err.Error())\n\t\treturn\n\t}\n\n\trespondWithJSON(w, http.StatusOK, map[string]string{\"message\": \"User deleted successfully\"})\n}", "func DeleteUser(\n\tctx context.Context,\n\ttx *sql.Tx,\n\trequest *models.DeleteUserRequest) error {\n\tdeleteQuery := deleteUserQuery\n\tselectQuery := \"select count(uuid) from user where uuid = ?\"\n\tvar err error\n\tvar count int\n\tuuid := request.ID\n\tauth := common.GetAuthCTX(ctx)\n\tif auth.IsAdmin() {\n\t\trow := tx.QueryRowContext(ctx, selectQuery, uuid)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"not found\")\n\t\t}\n\t\trow.Scan(&count)\n\t\tif count == 0 {\n\t\t\treturn errors.New(\"Not found\")\n\t\t}\n\t\t_, err = tx.ExecContext(ctx, deleteQuery, uuid)\n\t} else {\n\t\tdeleteQuery += \" and owner = ?\"\n\t\tselectQuery += \" and owner = ?\"\n\t\trow := tx.QueryRowContext(ctx, selectQuery, uuid, auth.ProjectID())\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"not found\")\n\t\t}\n\t\trow.Scan(&count)\n\t\tif count == 0 {\n\t\t\treturn errors.New(\"Not found\")\n\t\t}\n\t\t_, err = tx.ExecContext(ctx, deleteQuery, uuid, auth.ProjectID())\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"delete failed\")\n\t}\n\n\terr = common.DeleteMetaData(tx, uuid)\n\tlog.WithFields(log.Fields{\n\t\t\"uuid\": uuid,\n\t}).Debug(\"deleted\")\n\treturn err\n}", "func DeleteUser(c *gin.Context) {\n\tvar user models.User\n\tid := c.Params.ByName(\"id\")\n\terr := models.DeleteUser(&user, id)\n\tif err != nil {\n\t\tc.AbortWithStatus(http.StatusNotFound)\n\t} else {\n\t\tc.JSON(http.StatusOK, gin.H{\"id\" + id: \"is deleted\"})\n\t}\n}", "func RemoveUser(userID string) (err error) {\n\n err = checkInit()\n if err != nil {\n return\n }\n\n err = createError(032)\n\n if _, ok := data[\"users\"].(map[string]interface{})[userID]; ok {\n\n delete(data[\"users\"].(map[string]interface{}), userID)\n err = saveDatabase(data)\n\n return\n }\n\n return\n}", "func Delete(id int) error {\n\tstmt, err := db.Prepare(\"DELETE FROM user where id ?\")\n\tdefer stmt.Close()\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t\treturn err\n\t}\n\t_, err = stmt.Exec(id)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t\treturn err\n\t}\n\treturn nil\n}", "func (d *DB) DeleteUser(id uint) error {\n\tvar user User\n\tuser.Model.ID = id\n\terr := d.db.Delete(&user).Error\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func DelUser(uid int64) (user UserOut, err error) {\n\t// var err error\n\tusers, err := GetUser(uid, 0, 0)\n\tif err != nil {\n\t\treturn\n\t}\n\tif len(users) == 0 {\n\t\terr = errors.New(\"User not found.\")\n\t\treturn\n\t}\n\n\tq := `UPDATE user SET deleted=1 WHERE uid = %d`\n\tresult, err := Exec(fmt.Sprintf(q, users[0].Uid))\n\tif err != nil {\n\t\treturn\n\t}\n\n\taffected, err := result.RowsAffected()\n\tlog.Println(affected)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tuser = users[0]\n\n\treturn\n}", "func (u *UserTest) Delete() error {\n\treturn nil\n}", "func (d *database) deleteUser(publicKey string) (err error) {\n\ttx, err := d.db.Begin()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"deleteUser\")\n\t}\n\tquery := \"DELETE FROM letters WHERE sender == ?;\"\n\tlogger.Log.Debug(query)\n\tstmt, err := tx.Prepare(query)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"deleteUser\")\n\t}\n\tdefer stmt.Close()\n\n\t_, err = stmt.Exec(publicKey)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"deleteUser\")\n\t}\n\n\terr = tx.Commit()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"deleteUser\")\n\t}\n\treturn\n}", "func _delete(context echo.Context, user *User) error {\n\tdeleteErr := Remove(user.Key)\n\tif deleteErr != nil {\n\t\tlog.Printf(\"Cannot delete user %v\", deleteErr)\n\t\treturn context.JSON(http.StatusInternalServerError, errors.New(\"Cannot delete user with ID: \"+user.ID))\n\t}\n\treturn context.NoContent(http.StatusNoContent)\n}", "func DeleteUser(id int, db *gorm.DB) (err error) {\n\tvar user User\n\tif err = db.First(&user, id).Error; err != nil {\n\t\terr = ErrUserNotFound\n\t} else {\n\t\tdb.Delete(&user)\n\t}\n\treturn\n}", "func (a *Server) DeleteUser(w http.ResponseWriter, r *http.Request) {\n\tfmt.Println(\"delete a user\")\n}" ]
[ "0.7445775", "0.73911726", "0.7235373", "0.7189712", "0.7177783", "0.7176776", "0.7167774", "0.71673125", "0.7158333", "0.71264553", "0.7119575", "0.71035236", "0.71017534", "0.70990646", "0.70971656", "0.7089264", "0.7067626", "0.70577", "0.7051368", "0.70268077", "0.70268077", "0.6984143", "0.6983752", "0.697737", "0.69615823", "0.69613737", "0.69476247", "0.6942175", "0.6938423", "0.6938423", "0.69358194", "0.6935095", "0.69275635", "0.6921525", "0.6914482", "0.6895195", "0.68943286", "0.688829", "0.6887", "0.68798065", "0.68751174", "0.6873828", "0.6873299", "0.6873096", "0.68589216", "0.68495345", "0.68479407", "0.68479407", "0.68479407", "0.6845181", "0.6831825", "0.6810688", "0.680769", "0.6803002", "0.68024385", "0.68023646", "0.6787167", "0.6787093", "0.67791206", "0.6778762", "0.67655873", "0.67533773", "0.6749922", "0.6742492", "0.67314416", "0.6728377", "0.67215395", "0.671513", "0.67138034", "0.6712194", "0.6707773", "0.6707017", "0.670029", "0.6682155", "0.66761553", "0.6675783", "0.6675437", "0.66734505", "0.6671787", "0.6667789", "0.666458", "0.666458", "0.66546947", "0.66506207", "0.66474235", "0.66423714", "0.6639149", "0.66305906", "0.6624221", "0.6623362", "0.6623163", "0.6620557", "0.66067666", "0.66055745", "0.65995556", "0.65988636", "0.658311", "0.6578893", "0.6575208", "0.6572882", "0.65701634" ]
0.0
-1
Authenticate returns true if the provided password matches the one stored in the database
func (u *User) Authenticate(password string) error { if u.EncryptedPassword == nil { return errors.New("OAuth user cannot be authenticated with pasword") } err := bcrypt.CompareHashAndPassword( []byte(*u.EncryptedPassword), []byte(password)) if err != nil { return errors.New("invalid password") } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func authenticate(username string, password string) (bool, error) {\n\tfoundPwd := users[username]\n\tif foundPwd == \"\" {\n\t\treturn false, fmt.Errorf(\"user %s not found\", username)\n\t}\n\n\tif foundPwd != password {\n\t\treturn false, errors.New(\"password does not match\")\n\t}\n\n\treturn true, nil\n}", "func Authenticate(usrPassword string) bool {\n\treturn blockchainhelpers.HashStr(usrPassword) != model.PasswordHash\n}", "func (u *User) Authenticate(pw string) bool {\n\terr := bcrypt.CompareHashAndPassword([]byte(u.PasswordHash), []byte(pw))\n\treturn err == nil\n}", "func authenticate(userID string, password string) bool {\n\tsession, err := mgo.Dial(\"127.0.0.1\") //TODO use connection pool\n\tif err != nil {\n\t\t//Raise error\n\t\tpanic(err)\n\t}\n\t//close afterwards\n\tdefer session.Close()\n\n\t//read from DB\n\tc := session.DB(\"wellnomics\").C(\"authentication\")\n\tcount, err := c.Find(bson.M{\"name\": userID, \"password\": password}).Count()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\t//return true if record exists, false otherwise\n\treturn count > 0\n}", "func (sa *staticAuthenticator) Authenticate(user string, pass string) bool {\n\tpassword, found := sa.credentials[user]\n\tif !found {\n\t\treturn false\n\t}\n\treturn pass == password\n}", "func (store *Authenticator) Authenticate(user, pass string) (bool, error) {\n\th, err := store.GetPass(user)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\terr = bcrypt.CompareHashAndPassword(h, []byte(pass))\n\tif err == nil {\n\t\treturn true, nil\n\t}\n\treturn false, InvalidPassErr\n}", "func (self *CassandraMetaStore) Authenticate(user, pass string) (bool, error) {\n\tmu, err := self.findUser(user)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\treturn meta.CheckPass([]byte(mu.Password), []byte(pass))\n}", "func authenticateUser(username string, password string) bool {\n\tpasswordHash := md5Hash(password)\n\tvar user User\n\terr := mysql_client.QueryRow(\"SELECT username, password FROM User WHERE username=?\", username).Scan(&user.Username, &user.Password)\n\tif err == sql.ErrNoRows {\n\t\treturn false\n\t} else {\n\t\tif user.Username == username && user.Password == passwordHash {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (m *mysqlDBRepo) Authenticate(email, testPassword string) (int, string, error) {\n\t// request last longer 3 second so discard write record into db\n\tctx, cancel := context.WithTimeout(context.Background(), 3*time.Second)\n\tdefer cancel()\n\n\tvar id int\n\tvar hashedPassword string\n\n\terr := m.DB.QueryRowContext(ctx, `select id, password from users where email = ?`, email).Scan(\n\t\t&id,\n\t\t&hashedPassword,\n\t)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn id, \"\", err\n\t}\n\n\terr = bcrypt.CompareHashAndPassword([]byte(hashedPassword), []byte(testPassword))\n\tif err == bcrypt.ErrMismatchedHashAndPassword {\n\t\treturn 0, \"\", errors.New(\"incorrect password\")\n\t} else if err != nil {\n\t\treturn 0, \"\", err\n\t}\n\n\treturn id, hashedPassword, nil\n}", "func Authenticate(username, password string, db *sql.DB) (bool, error) {\n\tvar userName string\n\trow := db.QueryRow(\"select email from users where email=$1 and password=$2;\", strings.TrimSpace(username), password)\n\terr := row.Scan(&userName)\n\tif err != nil || userName == \"\" {\n\t\tlog.Println(err)\n\t\treturn false, err\n\t}\n\treturn true, nil\n\n}", "func authentication(user, pass string) bool {\n\tvar username, password string\n\tvar tipo int\n\tdb_mu.Lock()\n\tquery2, err := db.Query(\"SELECT username, password, type FROM admin WHERE username = ?\", user)\n\tdb_mu.Unlock()\n\tif err != nil {\n\t\tError.Println(err)\n\t\treturn false\n\t}\n\tfor query2.Next() {\n\t\terr = query2.Scan(&username, &password, &tipo)\n\t\tif err != nil {\n\t\t\tError.Println(err)\n\t\t\treturn false\n\t\t}\n\t}\n\tquery2.Close()\n\n\tif user == username && pass == password && tipo == 0 {\n\t\treturn true\n\t} else {\n\t\treturn false\n\t}\n}", "func Authenticate(userCol *mgo.Collection, username, password string) (bool, error) {\n\tvar auth Authentication\n\terr := userCol.Find(bson.M{\n\t\t\"username\": username,\n\t}).Select(bson.M{\n\t\t\"password\": 1,\n\t\t\"salt\": 1,\n\t}).One(&auth)\n\tif err != nil {\n\t\tif err == mgo.ErrNotFound {\n\t\t\treturn false, errors.NewNotFound(\"user\", username)\n\t\t}\n\t\treturn false, err\n\t}\n\tif auth.Password != encryptPassword(auth.Salt, password) {\n\t\treturn false, nil\n\t}\n\treturn true, nil\n}", "func (c Client) Authenticate(ctx context.Context, creds api.Creds) (string, error) {\n\tvar dbCreds api.Creds\n\tif err := c.db.GetContext(ctx, &dbCreds, \"select email, password from users where email = $1\", creds.Email); err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif dbCreds == creds {\n\t\treturn \"\", errors.New(\"passwords don't match\")\n\t}\n\n\treturn \"\", nil\n}", "func (m *postgresDBRepo) Authenticate(email, testPassword string) (int, string, error) {\n\tctx, cancel := context.WithTimeout(context.Background(), 3*time.Second)\n\tdefer cancel()\n\n\tvar (\n\t\tuserID int\n\t\thashedPassword string\n\t)\n\n\tquery := `select id, password from users where email = $1`\n\trow := m.DB.QueryRowContext(ctx, query, email)\n\terr := row.Scan(&userID, &hashedPassword)\n\tif err != nil {\n\t\treturn userID, \"\", err\n\t}\n\n\terr = bcrypt.CompareHashAndPassword([]byte(hashedPassword), []byte(testPassword))\n\tif err == bcrypt.ErrMismatchedHashAndPassword {\n\t\treturn 0, \"\", errors.New(\"incorrect password\")\n\t} else if err != nil {\n\t\treturn 0, \"\", err\n\t}\n\n\treturn userID, hashedPassword, nil\n}", "func (u User) Authenticate(password string) error {\n\tpass := []byte(u.PrivateID + \":\" + password)\n\treturn bcrypt.CompareHashAndPassword([]byte(u.Hash), pass)\n}", "func authenticate(uname string, psw string) bool {\n\tuser_map_lock.Lock()\n\tdefer user_map_lock.Unlock()\n\tif _, is_exist := user_map[uname]; is_exist && user_map[uname] == psw {\n\t\treturn true\n\t} else {\n\t\treturn false\n\t}\n}", "func authenticateByPassword(usr *User) error {\n\tvar passwordHash string\n\terr := db.QueryRow(\"select id, password from users where username = $1\", usr.Username).Scan(&usr.id, &passwordHash)\n\tif err != nil {\n\t\tif err == sql.ErrNoRows {\n\t\t\treturn ErrUserNotFound\n\t\t}\n\t\treturn err\n\t}\n\terr = scrypt.CompareHashAndPassword([]byte(passwordHash), []byte(usr.password))\n\tif err != nil {\n\t\treturn ErrInvalidPassword\n\t}\n\treturn addSession(usr)\n}", "func (handler *SimpleHandler) Authenticate(login, password string) bool {\n\tif _, ok := handler.Users[login]; !ok {\n\t\treturn false\n\t}\n\n\treturn handler.Hash(password) == handler.Users[login]\n}", "func Authenticate(db *sql.DB, name string, pass string) error {\n\tq := \"select salt, hash from user where name = $1;\"\n\tsalt := []byte{}\n\tdbHashedPass := []byte{}\n\tif err := db.QueryRow(q, name).Scan(&salt, &dbHashedPass); err != nil {\n\t\treturn fmt.Errorf(\"querying: %v\", err)\n\t}\n\thashedPass, err := scrypt.Key([]byte(pass), salt, ScryptN, ScryptR, ScryptP, ScryptBytes)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"hash failure: %v\", err)\n\t}\n\tif !bytes.Equal(hashedPass, dbHashedPass) {\n\t\treturn fmt.Errorf(\"bad password\")\n\t}\n\treturn nil\n}", "func (c *V1) Authenticate(ctx context.Context, username, cleartextPassword string) bool {\n\tif c.EnsureInit() != nil {\n\t\treturn false\n\t}\n\n\tp, ok := c.users[username]\n\tif !ok {\n\t\treturn false\n\t}\n\tmatch, err := p.check(ctx, c.bcryptLimiter, cleartextPassword)\n\treturn err == nil && match\n}", "func (h *File) Authenticate(u string, p string) bool {\n\th.mutex.Lock()\n\tdefer h.mutex.Unlock()\n\n\terr := bcrypt.CompareHashAndPassword([]byte(h.users[u]), []byte(p))\n\treturn err == nil\n}", "func TestAuthenticate(t *testing.T) {\n\tuser, _ := testUser.ToUser()\n\t//Correct password\n\terr := user.Authenticate(\"password\")\n\tif err != nil {\n\t\tt.Errorf(\"authentication failed, expected to pass\")\n\t}\n\t//Incorrect password\n\terr2 := user.Authenticate(\"incorrect\")\n\tif err2 == nil {\n\t\tt.Errorf(\"authentication passed, expected to fail\")\n\t}\n}", "func Authenticate(db *sqlx.DB, email string, testPassword string) (user User, err error) {\n\t// Find the hashed password for the given email\n\tvar rows *sql.Rows\n\trows, err = db.Query(findUserIdAndPasswordByEmail, email)\n\tif err != nil {\n\t\treturn\n\t}\n\tfor rows.Next() {\n\t\tvar userId int\n\t\tvar hashedPassword sql.NullString\n\t\terr = rows.Scan(&userId, &hashedPassword)\n\t\tif err != nil {\n\t\t\tlog.Println(\"Error in Scan\", err)\n\t\t} else if userId == 0 {\n\t\t\t//fmt.Println(\"userId == 0. user not found\")\n\t\t} else if hashedPassword.Valid {\n\t\t\tif misc.ComparePassword(testPassword, hashedPassword.String) {\n\t\t\t\t// We found a match so find the user\n\t\t\t\tuser, err = UserFindById(db, userId)\n\t\t\t} else {\n\t\t\t\tlog.Println(\"Password mismatched for email\", email)\n\t\t\t}\n\t\t}\n\t\trows.Close()\n\t}\n\n\t// return the user and error\n\treturn\n}", "func AuthenticateUser(email, password string) (bool, models.Account) {\n\t//cari document dengan username dan password yg diberikan\n\n\tvar user models.Account\n\tsqlStatement := \"SELECT id,email,name,password,role FROM account WHERE email=?\"\n\n\terr = db.QueryRow(sqlStatement, email).\n\t\tScan(&user.ID, &user.Email, &user.Name, &user.Password, &user.Role)\n\tif err == sql.ErrNoRows {\n\t\treturn false, user\n\t}\n\n\t// check_match := bcrypt.CompareHashAndPassword([]byte(user.Password), []byte(password))\n\t// log.Println(user.Password)\n\t// if check_match != nil {\n\tif password != user.Password {\n\t\t//LOGIN FAILED, PASSWORD SALAH\n\t\tlog.Println(\"Password atau email salah\")\n\t\treturn false, user\n\t}\n\t//LOGIN SUCCESS\n\t// log.Println(\"Password atau email salah\")\n\treturn true, user\n}", "func (admin *Admin) Authenticate(password string) error {\n\terr := bcrypt.CompareHashAndPassword(admin.PassHash, []byte(password))\n\tif err != nil {\n\t\treturn fmt.Errorf(\"invalid password: %v\", err)\n\t}\n\treturn nil\n}", "func (r *Registrar) IsPasswordAuthentic(username, password string) bool {\n\tcheckSum := adler32.Checksum([]byte(password))\n\ttrueCheckSum := r.users[username].passwordCheckSum\n\treturn checkSum == trueCheckSum\n}", "func (u *UserModel) Authenticate(email, password string) (int, error) {\n\tstmt := `select id, hashed_password from users where email = $1 and active = true`\n\trow := u.DB.QueryRow(stmt, email)\n\n\tvar id int\n\tvar hashedPassword []byte\n\terr := row.Scan(&id, &hashedPassword)\n\tif err != nil {\n\t\tif errors.Is(err, sql.ErrNoRows) {\n\t\t\treturn 0, models.ErrInvalidCredentials\n\t\t}\n\t\treturn 0, err\n\t}\n\n\terr = bcrypt.CompareHashAndPassword(hashedPassword, []byte(password))\n\tif err != nil {\n\t\tif errors.Is(err, bcrypt.ErrMismatchedHashAndPassword) {\n\t\t\treturn 0, models.ErrInvalidCredentials\n\t\t}\n\t\treturn 0, err\n\t}\n\n\treturn id, nil\n}", "func authenticateUser(username, password string) (bool, uint64) {\n\tuser := models.User{Name: username, Password: password}\n\n\terr := database.DB.Where(&user).First(&user)\n\tif err.Error != nil {\n\t\treturn false, 0\n\t}\n\n\treturn true, user.ID\n}", "func matchPassword(u *models.User, password string) bool {\n\treturn utils.Encrypt(password, u.Salt, u.PasswordVersion) == u.Password\n}", "func (model *UserModel) Authenticate(email, password string) (int, error) {\n\tvar id int\n\tvar hashedPassword []byte\n\tstmt := `SELECT id, hashed_password FROM users WHERE email = ? AND active = TRUE`\n\terr := model.DB.QueryRow(stmt, email).Scan(&id, &hashedPassword)\n\tif err != nil {\n\t\tif errors.Is(err, sql.ErrNoRows) {\n\t\t\treturn 0, models.ErrInvalidCredentials\n\t\t}\n\t\treturn 0, err\n\t}\n\n\terr = bcrypt.CompareHashAndPassword(hashedPassword, []byte(password))\n\tif err != nil {\n\t\tif errors.Is(err, bcrypt.ErrMismatchedHashAndPassword) {\n\t\t\treturn 0, models.ErrInvalidCredentials\n\t\t}\n\t\treturn 0, err\n\t}\n\treturn id, nil\n}", "func (m *UserModel) Authenticate(email, password string) (int, error) {\n\tvar id int\n\tvar hashedPassword []byte\n\n\t// Get the user from the db\n\tstmt := `SELECT id, password FROM users WHERE email = ? AND active = TRUE`\n\trow := m.DB.QueryRow(stmt, email)\n\terr := row.Scan(&id, &hashedPassword)\n\tif err != nil {\n\t\tif errors.Is(err, sql.ErrNoRows) {\n\t\t\treturn 0, models.ErrInvalidCredentials\n\t\t}\n\n\t\treturn 0, err\n\t}\n\n\t// Check the password\n\terr = bcrypt.CompareHashAndPassword(hashedPassword, []byte(password))\n\tif err != nil {\n\t\tif errors.Is(err, bcrypt.ErrMismatchedHashAndPassword) {\n\t\t\treturn 0, models.ErrInvalidCredentials\n\t\t}\n\n\t\treturn 0, err\n\t}\n\n\treturn id, nil\n}", "func MiddleWareAuth(w http.ResponseWriter, r *http.Request) (bool, int) {\n\tusername := r.FormValue(\"username\")\n\tuserpass := r.FormValue(\"userpass\")\n\tvar dbPass string\n\tvar dbSalt string\n\tvar DbUID int\n\n\tuer := database.QueryRow(\"select user_password ,user_salt,user_id from users where user_nickname = ?\",\n\t\tusername).Scan(&dbPass, &dbSalt, DbUID)\n\n\tif uer != nil {\n\n\t}\n\texpectedPassword := password.GenerateHash(dbSalt, userpass)\n\n\tif dbPass == expectedPassword {\n\t\treturn true, DbUID\n\t}\n\treturn false, 0\n\n}", "func Authenticate(udb *userdb.DB, user, passwd string) bool {\n\tdefer func(start time.Time) {\n\t\telapsed := time.Since(start)\n\t\tdelay := AuthenticateTime - elapsed\n\t\tif delay > 0 {\n\t\t\tmaxDelta := int64(float64(delay) * 0.2)\n\t\t\tdelay += time.Duration(rand.Int63n(maxDelta))\n\t\t\ttime.Sleep(delay)\n\t\t}\n\t}(time.Now())\n\n\t// Note that the database CAN be nil, to simplify callers.\n\tif udb == nil {\n\t\treturn false\n\t}\n\n\treturn udb.Authenticate(user, passwd)\n}", "func (u *UserModel) Authenticate(email, password string) (int, error) {\n\t// Retrieve the id and hashed password assocaited with the given email.\n\t// If no matching email exists, or the user is not active, we return the\n\t// ErrInvalidCredentials error.\n\tvar id int\n\tvar hashedPw []byte\n\tstmt := `SELECT id, hashed_password FROM users WHERE email = ? AND active = TRUE`\n\trow := u.DB.QueryRow(stmt, email)\n\terr := row.Scan(&id, &hashedPw)\n\tif err != nil {\n\t\tif errors.Is(err, sql.ErrNoRows) {\n\t\t\treturn 0, models.ErrInvalidCredentials\n\t\t} else {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\t// Check whether the hashed password and plain-text password provided match.\n\t// If they don't, we return the ErrInvalidCredentials error.\n\n\terr = bcrypt.CompareHashAndPassword(hashedPw, []byte(password))\n\tif err != nil {\n\t\tif errors.Is(err, bcrypt.ErrMismatchedHashAndPassword) {\n\t\t\treturn 0, models.ErrInvalidCredentials\n\t\t} else {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\t// Otherwise, the password is correct, so return the userID.\n\treturn id, nil\n}", "func (u *User) Authenticate(email, password string) (bool, error) {\n\n\tlog.WithFields(log.Fields{\n\t\t\"user_email\": email,\n\t}).Debug(\"Authenticate a user\")\n\n\tuser, err := u.GetUserByEmail(email)\n\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\tok := util.CheckPasswordHash(password, user.PasswordHash)\n\n\tif !ok {\n\t\treturn false, nil\n\t}\n\n\treturn true, nil\n}", "func authentication_admin(user, pass string) bool {\n\tvar username, password string\n\tvar tipo int\n\tdb_mu.Lock()\n\tquery2, err := db.Query(\"SELECT username, password, type FROM admin WHERE username = ?\", user)\n\tdb_mu.Unlock()\n\tif err != nil {\n\t\tError.Println(err)\n\t\treturn false\n\t}\n\tfor query2.Next() {\n\t\terr = query2.Scan(&username, &password, &tipo)\n\t\tif err != nil {\n\t\t\tError.Println(err)\n\t\t\treturn false\n\t\t}\n\t}\n\tquery2.Close()\n\n\tif user == username && pass == password && tipo == 1 {\n\t\treturn true\n\t} else {\n\t\treturn false\n\t}\n}", "func (u *UserFunctions) Authenticate(email, password string) (primitive.ObjectID, error) {\n\t//Authenticate user before login by retrieving the user id and hashed password from database\n\t//Hash the password entered and compare it to the one retrieved from database\n\n\tnewUser := u.CLIENT.Database(\"queue\")\n\tuserCollection := newUser.Collection(\"user\")\n\n\tvar output models.User\n\tfilter := bson.M{\"email\": email}\n\n\terr := userCollection.FindOne(context.TODO(), filter).Decode(&output)\n\tif err != nil {\n\t\t// ErrNoDocuments means that the filter did not match any documents in the collection\n\t\tif err == mongo.ErrNoDocuments {\n\t\t\treturn output.ID, errors.New(\"filter did not match any documents in the collection\")\n\t\t}\n\t\tlog.Fatal(err)\n\t}\n\t// Check whether the hashed password and plain-text password provided match\n\terr = bcrypt.CompareHashAndPassword(output.Password, []byte(password))\n\tif err != nil {\n\t\tif errors.Is(err, bcrypt.ErrMismatchedHashAndPassword) {\n\t\t\treturn output.ID, models.ErrInvalidCredentials\n\t\t} else {\n\t\t\treturn output.ID, err\n\t\t}\n\t}\n\tfmt.Println(\"Authenticated\")\n\treturn output.ID, nil\n}", "func (u User) Authenticate(cleartext string, hasher fosite.Hasher) error {\n\treturn hasher.Compare(context.TODO(), u.GetHashedSecret(), []byte(cleartext))\n}", "func (c *SQLiteConn) authenticate(username, password string) int {\n\t// NOOP\n\treturn 0\n}", "func (s *UserStoreFile) Authenticate(email, password string) (*app.User, error) {\r\n\tfoundUser, err := s.ByEmail(email)\r\n\tif err != nil {\r\n\t\treturn nil, err\r\n\t}\r\n\r\n\terr = bcrypt.CompareHashAndPassword([]byte(foundUser.PasswordHash), []byte(password+s.UserPwPepper))\r\n\tswitch err {\r\n\tcase nil:\r\n\t\treturn foundUser, nil\r\n\t// case bcrypt.ErrMismatchedHashAndPassword:\r\n\t// \treturn nil, app.ErrInvalidPassword\r\n\tdefault:\r\n\t\treturn nil, err\r\n\t}\r\n}", "func PassMatch(username, password string) bool {\n\torm := get_DBFront()\n\tvar user User\n\terr := orm.SetTable(\"user\").Where(\"username=?\", username).Find(&user)\n\tif !check_err(err) {\n\t\tLog(Log_Struct{\"error\", \"DB_Error_Line_265\", err})\n\t\treturn false\n\t}\n\n\treturn match_pass(password, user.Password)\n}", "func checkPassword(db *sql.DB, w http.ResponseWriter, creds s.Credentials) bool {\n\n\ttxt, err := db.Begin()\n\tif err != nil {\n\t\t// Error handling\n\t\tfmt.Println(err)\n\t}\n\n\t//Getting password\n\tresult, err := txt.Query(\"SELECT password FROM sample.tb WHERE username=\" + creds.Username)\n\tif err != nil {\n\t\t// Error handling\n\t\tfmt.Println(err)\n\t}\n\tvar passwordSQL string\n\tfor result.Next() {\n\t\tif err := result.Scan(&passwordSQL); err != nil {\n\t\t\tw.WriteHeader(http.StatusOK)\n\t\t}\n\t\treturn hash.ComparePasswords(passwordSQL, []byte(creds.Password))\n\t}\n\tdb.Close()\n\treturn false\n}", "func (a *Admin) Verify(username, password string) bool {\n\tfound := a.FindByUsername(username)\n\tif !found {\n\t\treturn false\n\t}\n\n\tif a.Password != tools.EncodeMD5(password) {\n\t\treturn false\n\t}\n\treturn true\n}", "func (u *User) Auth(password string) bool {\n\treturn common.CheckPasswordHash(password, u.Password)\n}", "func (a *Authenticator) Authenticate(username, password string) bool {\n\tvar l *ldap.Conn\n\tvar err error\n\n\tl, err = a.dial()\n\n\tif err != nil {\n\t\tlog.Error().Msgf(\"failed to connect to ldap server: %s\", err.Error())\n\t\treturn false\n\t}\n\n\tdefer l.Close()\n\n\tif a.cfg.UseStartTLS {\n\t\t// Reconnect with TLS\n\t\terr = l.StartTLS(a.getTLSConfig())\n\t\tif err != nil {\n\t\t\tlog.Error().Msgf(\"failed to connect to ldap server with StartTLS: %s\", err.Error())\n\t\t\treturn false\n\t\t}\n\t}\n\n\t// First bind with a read only user\n\terr = l.Bind(a.cfg.BindUser, a.cfg.BindPass)\n\tif err != nil {\n\t\tlog.Error().Msgf(\"failed to connect to ldap server with specified bind credentials: %s\", err.Error())\n\t\treturn false\n\t}\n\n\tif a.cfg.LoginNameAttribute == \"\" {\n\t\ta.cfg.LoginNameAttribute = \"sAMAccountName\"\n\t}\n\n\tif a.cfg.ObjectClass == \"\" {\n\t\ta.cfg.ObjectClass = \"organizationalPerson\"\n\t}\n\n\tsearchRequest := ldap.NewSearchRequest(\n\t\ta.cfg.BaseDN, // The base dn to search\n\t\tldap.ScopeWholeSubtree, ldap.NeverDerefAliases, 0, 0, false,\n\t\tfmt.Sprintf(\"(&(objectClass=%s)(%s=%s))\", a.cfg.ObjectClass, a.cfg.LoginNameAttribute, username),\n\t\t[]string{\"dn\", a.cfg.LoginNameAttribute}, // A list attributes to retrieve\n\t\tnil,\n\t)\n\n\tsr, err := l.Search(searchRequest)\n\tif err != nil {\n\t\tlog.Error().Msgf(\"failed to find any results: %s\", err.Error())\n\t\treturn false\n\t}\n\n\tfor _, entry := range sr.Entries {\n\t\tif entry.GetAttributeValue(a.cfg.LoginNameAttribute) != username {\n\t\t\tcontinue\n\t\t}\n\n\t\t// Disconnect and then reconnect as the user to validate the password.\n\t\tl.Close()\n\t\tl, err = a.dial()\n\t\tif err != nil {\n\t\t\tlog.Error().Msgf(\"failed to unbind: %s\", err.Error())\n\t\t\treturn false\n\t\t}\n\n\t\t// Try and bind as the other user.\n\t\terr = l.Bind(entry.DN, password)\n\t\tif err != nil {\n\t\t\tlog.Error().Msgf(\"failed to bind with specified username and password: %s\", err.Error())\n\t\t\treturn false\n\t\t}\n\t\treturn true\n\t}\n\n\treturn a.allowedUsers[username] == password\n}", "func (pg *PostgresqlDb)AuthenticateUser(username string, password string) (int, *domain.ErrHandler) {\n if pg.conn == nil {\n return -1, &domain.ErrHandler{7, \"func (pg PostgresqlDb)\", \"AuthenticateUser(username string, password string)\", \"\"}\n }\n recordset := pg.conn.QueryRow(\"select id from users where username=$1 and password=crypt($2,password);\", username, password)\n\n var id int\n var err domain.ErrHandler\n switch err := recordset.Scan(&id); err {\n case sql.ErrNoRows:\n return -1, &domain.ErrHandler{13, \"func (pg PostgresqlDb)\", \"AuthenticateUser(username string, password string)\", \"\"}\n case nil:\n return id, nil\n }\n return -1, &domain.ErrHandler{1, \"func (pg PostgresqlDb)\", \"AuthenticateUser(username string, password string)\", err.Error()}\n}", "func (c *Conn) Authenticate(ctx context.Context, email string) (string, error) {\n\tvar password string\n\n\terr := c.db.GetContext(ctx, &password, queryAuthAuthenticate, email)\n\tif err != nil && err != sql.ErrNoRows {\n\t\treturn \"\", err\n\t}\n\n\treturn password, nil\n}", "func (a AccountManager) Authenticate(username, password string) (success bool, err error) {\n\tsuccess = true\n\treturn\n}", "func (u *user) CheckPassword(password string) bool {\n return u.passwordHash == generatePasswordHash(password)\n}", "func (dal *UserDAL) AuthenticateUser(email, password string) (bool, int, error) {\n\tvar c int\n\terr := dal.db.Get(&c, \"SELECT COUNT(1) FROM userdata WHERE email = $1 AND password = $2\", email, password)\n\tif err != nil {\n\t\treturn false, -1, err\n\t}\n\n\tif c != 1 {\n\t\treturn false, -1, nil\n\t}\n\n\tvar id int\n\terr = dal.db.Get(&id,\n\t\t`SELECT userid FROM userdata WHERE email = $1 AND password = $2`, email, password)\n\tif err != nil {\n\t\treturn false, -1, err\n\t}\n\n\treturn true, id, nil\n}", "func (us *userService) Authenticate(email,password string) (*User, error){\n\tfoundUser, err := us.ByEmail(email)\n\tif err != nil{\n\t\treturn nil, err\n\t}\n\terr = bcrypt.CompareHashAndPassword([]byte(foundUser.PasswordHash),[]byte(password + us.pepper))\n\tif err != nil{\n\t\tswitch err {\n\t\t\tcase bcrypt.ErrMismatchedHashAndPassword: \n\t\t\t\treturn nil, ErrPasswordIncorrect\n\t\t\tdefault:\n\t\t\t\treturn nil, err\n\t\t}\n\t} \n\n\treturn foundUser, nil\n}", "func (otp *OTPConfig) Authenticate(password string) (bool, error) {\n\tcode, err := strconv.Atoi(password)\n\tif err != nil {\n\t\treturn false, ErrInvalidCode\n\t}\n\n\tswitch {\n\n\t// TOTP code\n\tcase len(password) == 6:\n\t\tt0 := int(time.Now().UTC().Unix() / 30)\n\t\tminT := t0 - (otp.WindowSize / 2)\n\t\tmaxT := t0 + (otp.WindowSize / 2)\n\t\tfor t := minT; t <= maxT; t++ {\n\t\t\tif cc, _ := ComputeCode(otp.Secret, int64(t)); cc == password {\n\t\t\t\t// check \"UsedCodes\"\n\t\t\t\tfor i := range otp.UsedCodes {\n\t\t\t\t\tif otp.UsedCodes[i] == t {\n\t\t\t\t\t\treturn false, nil\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\totp.UsedCodes = append(otp.UsedCodes, t)\n\t\t\t\totp.GC()\n\t\t\t\t// code OK\n\t\t\t\treturn true, nil\n\t\t\t}\n\t\t}\n\t\treturn false, nil\n\n\t// Scratch code\n\tcase len(password) == 8 && password[0] >= '1':\n\t\tfor i := range otp.ScratchCodes {\n\t\t\tif code == otp.ScratchCodes[i] {\n\t\t\t\totp.ScratchCodes = append(otp.ScratchCodes[:i], otp.ScratchCodes[i+1:]...)\n\t\t\t\treturn true, nil\n\t\t\t}\n\t\t}\n\t\treturn false, nil\n\n\tdefault:\n\t\treturn false, ErrInvalidCode\n\t}\n}", "func (z *User) MatchPassword(password string) bool {\n\thashedPassword, err := hex.DecodeString(z.PasswordHash)\n\tif err != nil {\n\t\treturn false\n\t}\n\treturn bcrypt.CompareHashAndPassword(hashedPassword, []byte(password)) == nil\n}", "func Authenticate(username, password string) (*User, error) {\n\tuser, found := FindUserByUsername(username)\n\tif !found {\n\t\treturn nil, fmt.Errorf(\"Invalid username/password\")\n\t}\n\tresult := user.VerifyPassword(password)\n\tif result == nil {\n\t\tuser.LastLogin = time.Now()\n\t\tif err := user.Save(); err != nil {\n\t\t\tresult = err\n\t\t}\n\t}\n\treturn user, result\n}", "func (keystoneAuthenticator *KeystoneAuthenticator) AuthenticatePassword(username string, password string) (user.Info, bool, error) {\n\topts := gophercloud.AuthOptions{\n\t\tIdentityEndpoint: keystoneAuthenticator.authURL,\n\t\tUsername: username,\n\t\tPassword: password,\n\t}\n\n\t_, err := keystoneAuthenticator.AuthenticatedClient(opts)\n\tif err != nil {\n\t\tglog.Info(\"Failed: Starting openstack authenticate client:\" + err.Error())\n\t\treturn nil, false, errors.New(\"Failed to authenticate\")\n\t}\n\n\treturn &user.DefaultInfo{Name: username}, true, nil\n}", "func ComparePassword(password string, encryptedPassword string) bool{\n\tpasswordBytes := []byte(password)\n\tdatabasePassword := []byte(encryptedPassword)\n\terr := bcrypt.CompareHashAndPassword(databasePassword, passwordBytes)\n\tif err != nil {\n\t\treturn false\n\t}\n\treturn true\n}", "func (a BasicAuthenticator) Auth(r *http.Request) bool {\n\t// Retrieve Authorization header\n\tauth := r.Header.Get(\"Authorization\")\n\n\t// No header provided\n\tif auth == \"\" {\n\t\treturn false\n\t}\n\n\t// Ensure format is valid\n\tbasic := strings.Split(auth, \" \")\n\tif basic[0] != \"Basic\" {\n\t\treturn false\n\t}\n\n\t// Decode base64'd user:password pair\n\tbuf, err := base64.URLEncoding.DecodeString(basic[1])\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\treturn false\n\t}\n\n\t// Split into username/password\n\tcredentials := strings.Split(string(buf), \":\")\n\n\t// Load user by username, verify user exists\n\tuser := new(data.UserRecord).Load(credentials[0], \"username\")\n\tif user == (data.UserRecord{}) {\n\t\treturn false\n\t}\n\n\t// Load user's API key\n\tkey := new(data.APIKey).Load(user.ID, \"user_id\")\n\tif key == (data.APIKey{}) {\n\t\treturn false\n\t}\n\n\t// Hash input password\n\tsha := sha1.New()\n\tif _, err = sha.Write([]byte(credentials[1] + key.Salt)); err != nil {\n\t\tlog.Println(err.Error())\n\t\treturn false\n\t}\n\n\thash := fmt.Sprintf(\"%x\", sha.Sum(nil))\n\n\t// Verify hashes match, using timing-attack resistant method\n\t// If function returns 1, hashes match\n\treturn subtle.ConstantTimeCompare([]byte(hash), []byte(key.Key)) == 1\n}", "func (authentication *Authentication) Check(hashedPassword, password string) bool {\n\treturn PasswordCheck(hashedPassword, password)\n}", "func FundPassMatch(username, password string) bool {\n\torm := get_DBFront()\n\tvar user User\n\terr := orm.SetTable(\"user\").Where(\"username=?\", username).Find(&user)\n\tif !check_err(err) {\n\t\tLog(Log_Struct{\"error\", \"DB_Error_Line_280\", err})\n\t\treturn false\n\t}\n\n\treturn match_pass(password, user.Fundpassword)\n}", "func (u *User) Authorize(password string) bool {\n\treturn u.password == sha256.Sum256([]byte(password))\n}", "func TestAuthenticate_Success(t *testing.T) {\n\tdb, mock, err := sqlmock.New()\n\tif err != nil {\n\t\tt.Fatalf(\"an error '%s' was not expected when opening a stub database connection\", err)\n\t}\n\tdefer db.Close()\n\n\tuser := models.User{\n\t\tID: 1,\n\t\tEmail: \"[email protected]\",\n\t\tPassword: \"personia\",\n\t}\n\n\trows := sqlmock.NewRows([]string{\"id\", \"email\"}).AddRow(user.ID, user.Email)\n\tmock.ExpectQuery(regexp.QuoteMeta(constants.LoginDetailsSelectQuery)).WithArgs(user.Email, user.Password).WillReturnRows(rows)\n\n\tloginRepository := NewLoginRepository(db)\n\n\tloginModel := &models.Login{\n\t\tEmail: \"[email protected]\",\n\t\tPassword: \"personia\",\n\t}\n\n\tcntx := context.Background()\n\tdbuser, err := loginRepository.Authenticate(cntx, loginModel)\n\tassert.Nil(t, err)\n\tassert.Equal(t, user.ID, dbuser.ID)\n\tassert.Equal(t, user.Email, dbuser.Email)\n}", "func VerifyPassword(db *gorm.DB, userID hide.ID, password string) bool {\n\tvar user model.User\n\tif err := db.Model(&model.User{}).Where(userID).First(&user).Error; err != nil {\n\t\treturn false\n\t}\n\n\tstart := time.Now()\n\n\terr := argonpass.Verify(password, user.Password)\n\n\telapsed := time.Since(start)\n\tlog.Printf(\"Password hash verify took %s\", elapsed)\n\n\treturn err == nil\n}", "func (us *userService) Authenticate(email, password string) (*User, error) {\r\n\tfoundUser, err := us.ByEmail(email)\r\n\tif err != nil {\r\n\t\treturn nil, err\r\n\t}\r\n\terr = bcrypt.CompareHashAndPassword(\r\n\t\t[]byte(foundUser.PasswordHash),\r\n\t\t[]byte(password+us.pepper))\r\n\tswitch err {\r\n\tcase nil:\r\n\t\treturn foundUser, nil\r\n\tcase bcrypt.ErrMismatchedHashAndPassword:\r\n\t\treturn nil, ErrPasswordIncorrect\r\n\tdefault:\r\n\t\treturn nil, err\r\n\t}\r\n\r\n}", "func (user User) Authenticate(c appengine.Context) (User, error) {\n\n\tpotential_user, err := getUserFromUsername(c, user.Username)\n\tif err != nil {\n\t\treturn user, err\n\t}\n\n\tif potential_user != (User{}) {\n\t\treturn potential_user, nil\n\t}\n\n\t// Check password match\n\terr = bcrypt.CompareHashAndPassword([]byte(potential_user.Password), []byte(user.Password))\n\tif err != nil {\n\t\treturn user, errors.New(\"Username / Password Invalid\")\n\t}\n\treturn potential_user, nil\n}", "func (a Admin) Login(user,passwd string) (error, bool) {\n if user == a.Name && passwd == a.Pass {\n return nil, true\n } else {\n return errors.New(\"Wrong login or password\"), false\n }\n}", "func (u *MockUserRecord) VerifyPassword(password string) error { return nil }", "func authenticateWithCorrectPassword(ctx, ctxForCleanUp context.Context, testUser string, r *hwsecremote.CmdRunnerRemote, helper *hwsecremote.CmdHelperRemote, userParam pinWeaverWithAuthAPIParam) error {\n\tcryptohomeHelper := helper.CryptohomeClient()\n\n\t// Authenticate a new auth session via the new password auth factor and mount the user.\n\t_, authSessionID, err := cryptohomeHelper.StartAuthSession(ctx, testUser, false /*ephemeral*/, uda.AuthIntent_AUTH_INTENT_DECRYPT)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to start auth session for password authentication\")\n\t}\n\tdefer cryptohomeHelper.InvalidateAuthSession(ctxForCleanUp, authSessionID)\n\n\t// Authenticate with correct password.\n\tif userParam.useAuthFactor {\n\t\treply, err := cryptohomeHelper.AuthenticateAuthFactor(ctx, authSessionID, passwordAuthFactorLabel, passwordAuthFactorSecret)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"failed to authenticate auth factor\")\n\t\t}\n\t\tif !reply.Authenticated {\n\t\t\treturn errors.New(\"AuthSession not authenticated despite successful reply\")\n\t\t}\n\t\tif err := cryptohomecommon.ExpectAuthIntents(reply.AuthorizedFor, []uda.AuthIntent{\n\t\t\tuda.AuthIntent_AUTH_INTENT_DECRYPT,\n\t\t\tuda.AuthIntent_AUTH_INTENT_VERIFY_ONLY,\n\t\t}); err != nil {\n\t\t\treturn errors.Wrap(err, \"unexpected AuthSession authorized intents\")\n\t\t}\n\t} else {\n\t\terr = cryptohomeHelper.AuthenticateAuthSession(ctx, passwordAuthFactorSecret, passwordAuthFactorLabel, authSessionID, false /*kiosk_mount*/)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"failed to authenticate AuthSession\")\n\t\t}\n\t}\n\n\treturn nil\n}", "func ComparePassword(password, hash string) (bool, error) {\r\n\r\n\tparts := strings.Split(hash, \"$\")\r\n\r\n\tc := &PasswordConfig{}\r\n\t_, err := fmt.Sscanf(parts[3], \"m=%d,t=%d,p=%d\", &c.memory, &c.time, &c.threads)\r\n\tif err != nil {\r\n\t\treturn false, err\r\n\t}\r\n\r\n\tsalt, err := base64.RawStdEncoding.DecodeString(parts[4])\r\n\tif err != nil {\r\n\t\treturn false, err\r\n\t}\r\n\r\n\tdecodedHash, err := base64.RawStdEncoding.DecodeString(parts[5])\r\n\tif err != nil {\r\n\t\treturn false, err\r\n\t}\r\n\tc.keyLen = uint32(len(decodedHash))\r\n\r\n\tcomparisonHash := argon2.IDKey([]byte(password), salt, c.time, c.memory, c.threads, c.keyLen)\r\n\r\n\treturn (subtle.ConstantTimeCompare(decodedHash, comparisonHash) == 1), nil\r\n}", "func (c *SQLiteConn) Authenticate(username, password string) error {\n\t// NOOP\n\treturn nil\n}", "func checkPassword(mode bool, pass, login string) error {\r\n\tif !mode {\r\n\t\tvar validPassTmpl = regexp.MustCompile(`[A-Z]`)\r\n\t\tok := validPassTmpl.MatchString(pass)\r\n\t\tif !ok {\r\n\t\t\treturn errors.New(\"password must have A-Z\")\r\n\t\t}\r\n\t\tvalidPassTmpl = regexp.MustCompile(`[a-z]`)\r\n\t\tok = validPassTmpl.MatchString(pass)\r\n\t\tif !ok {\r\n\t\t\treturn errors.New(\"password must have a-z(small)\")\r\n\t\t}\r\n\t\tvalidPassTmpl = regexp.MustCompile(`[0-9]`)\r\n\t\tok = validPassTmpl.MatchString(pass)\r\n\t\tif !ok {\r\n\t\t\treturn errors.New(\"password must have 0-9\")\r\n\t\t}\r\n\t\tif len(pass) < 8 {\r\n\t\t\treturn errors.New(\"password must have at least 8 chars\")\r\n\t\t}\r\n\t} else {\r\n\t\tfilter := bson.M{\"email\": login}\r\n\t\tres, e := db.GetOneByFilter(db.GetUsersColl(), filter)\r\n\t\tif res != nil {\r\n\t\t\te = bcrypt.CompareHashAndPassword([]byte(res[\"password\"].(string)), []byte(pass))\r\n\t\t\tif e != nil {\r\n\t\t\t\treturn errors.New(\"wrong password\")\r\n\t\t\t}\r\n\t\t} else {\r\n\t\t\treturn e\r\n\t\t}\r\n\t}\r\n\treturn nil\r\n}", "func (authCredentials *AuthCredentials) Authenticate(database *sql.DB) {\n\thasUsername := len(authCredentials.Username) > 0\n\thasEmail := len(authCredentials.Email) > 0\n\thasPassword := len(authCredentials.Password) > 0\n\tif (!hasUsername && !hasEmail) || !hasPassword {\n\t\tpanic(&ModelError{\n\t\t\tCode: AuthErrorMissingParams,\n\t\t\tMessage: \"either 'username' or 'email', and 'password', should be specified\",\n\t\t})\n\t}\n\tif hasUsername {\n\t\tlogger.Infof(\"[auth] authenticating username '%s'\", authCredentials.Username)\n\t} else if hasEmail {\n\t\tlogger.Infof(\"[auth] authenticating email '%s'\", authCredentials.Email)\n\t}\n\tauthCredentials.authenticate(database)\n}", "func Authenticate(rw http.ResponseWriter, req *http.Request) {\n\tvar (\n\t\tdata AuthenticatePostData\n\t\tuser schema.User\n\t\tuserID int\n\t)\n\n\tbody, err := ioutil.ReadAll(io.LimitReader(req.Body, 1048576))\n\tif err != nil {\n\t\thelper.CreateResponse(rw, req, 500, nil, err)\n\t\treturn\n\t}\n\n\terr = req.Body.Close()\n\tif err != nil {\n\t\treturn\n\t}\n\n\tif err = json.Unmarshal(body, &data); err != nil {\n\t\thelper.CreateResponse(rw, req, 500, nil, err)\n\t\treturn\n\t}\n\n\tdatabase, err := db.Connection()\n\tif err != nil {\n\t\thelper.CreateResponse(rw, req, 500, nil, err)\n\t\treturn\n\t}\n\n\trows, err := database.Query(\"SELECT * FROM users WHERE email=$1\", data.Email)\n\tif err != nil {\n\t\thelper.CreateResponse(rw, req, 500, nil, err)\n\t\treturn\n\t}\n\tdefer rows.Close()\n\n\tfor rows.Next() {\n\t\terr := rows.Scan(&userID, &user.FirstName, &user.LastName, &user.Email, &user.PasswordHash, &user.PasswordSalt, &user.Disabled)\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t}\n\n\terr = database.Close()\n\tif err != nil {\n\t\thelper.CreateResponse(rw, req, 500, nil, err)\n\t\treturn\n\t}\n\n\tif err = helper.Compare(data.Password, user); err != nil {\n\t\thelper.CreateResponse(rw, req, 500, nil, err)\n\t\treturn\n\t}\n\n\thelper.CreateResponse(rw, req, 500, user, err)\n}", "func (s *Store) Authenticate(username, password string) (ui *UserInfo, err error) {\n\terr = s.read(func(data *Data) error {\n\t\t// Find user.\n\t\tu := data.User(username)\n\t\tif u == nil {\n\t\t\treturn ErrUserNotFound\n\t\t}\n\n\t\t// Compare password with user hash.\n\t\tif err := bcrypt.CompareHashAndPassword([]byte(u.Hash), []byte(password)); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tui = u\n\t\treturn nil\n\t})\n\treturn\n}", "func (us *UserService) Authenticate(email, password string) (*User, error) {\n\tfoundUser, err := us.ByEmail(email)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\terr = bcrypt.CompareHashAndPassword(\n\t\t[]byte(foundUser.PasswordHash),\n\t\t[]byte(password+userPwPepper))\n\tswitch err {\n\tcase nil:\n\t\treturn foundUser, nil\n\tcase bcrypt.ErrMismatchedHashAndPassword:\n\t\treturn nil, ErrInvalidPassword\n\tdefault:\n\t\treturn nil, err\n\t}\n}", "func (us *userService) Authenticate(email, password string) (*User, error) {\n\tfoundUser, err := us.ByEmail(email)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = bcrypt.CompareHashAndPassword([]byte(foundUser.PasswordHash), []byte(password+userPwPepper))\n\n\tif err != nil {\n\t\tswitch err {\n\t\tcase bcrypt.ErrMismatchedHashAndPassword:\n\t\t\treturn nil, ErrPasswordIncorrect\n\t\tdefault:\n\t\t\treturn nil, err\n\t\t}\n\n\t}\n\n\treturn foundUser, nil\n\n}", "func Authenticate(reqUser User) (bool, string) {\n\tuser, found := userList[reqUser.Username]\n\tif found == false || (user.Password != reqUser.Password) {\n\t\treturn false, \"\"\n\t}\n\n\tuser.update()\n\n\tif user.online {\n\t\treturn true, user.token\n\t}\n\n\ttoken := uuid.TimeUUID().String()\n\tuser.token = token\n\tuser.online = true\n\n\ttokenList[token] = user.Username\n\n\treturn true, token\n}", "func (a *Auth) ComparePassword(pw []byte) error {\n\treturn bcrypt.CompareHashAndPassword(a.PasswordHash, a.addSaltToPassword(pw))\n}", "func (u *userService) Authenticate(email string, password string) (*User, error) {\n\t//You can call methods of the interface implementation directly\n\t//because no reciever function on the UserService struct has overwritten the\n\t//UserDB implementation of ByEmail\n\tfoundUser, err := u.ByEmail(email)\n\t//fmt.Println(\"This was the found user : \", foundUser)\n\n\tif foundUser == nil {\n\t\treturn foundUser, ErrNotFound\n\t}\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\terr = bcrypt.CompareHashAndPassword([]byte(foundUser.PasswordHash),\n\t\t[]byte(password+userPwPepper))\n\tswitch err {\n\tcase nil:\n\t\treturn foundUser, nil\n\tdefault:\n\t\treturn nil, ErrEmailInvalid\n\t}\n}", "func ComparePassword(hashedPassword string, password string) bool {\n\terr := bcrypt.CompareHashAndPassword([]byte(hashedPassword), []byte(password))\n\tif err != nil {\n\t\treturn false\n\t}\n\treturn true\n}", "func (us *userService) Authenticate(email, password string) (*User, error) {\n\tu, err := us.ByEmail(email)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\thpwBytes := []byte(u.PasswordHash)\n\tpwByes := []byte(password + utils.GetPepper())\n\terr = bcrypt.CompareHashAndPassword(hpwBytes, pwByes)\n\tif err != nil {\n\t\tif err == bcrypt.ErrMismatchedHashAndPassword {\n\t\t\treturn nil, ErrPasswordInccorect\n\t\t}\n\n\t\treturn nil, err\n\t}\n\n\treturn u, nil\n}", "func ComparePassword(password string, hash string) bool {\n\tbyteHash := []byte(hash)\n\tbytePassword := []byte(password)\n\terr := bcrypt.CompareHashAndPassword(byteHash, bytePassword)\n\treturn err == nil\n}", "func ComparePassword(savedPassword string, password string) bool {\n\tif err := bcrypt.CompareHashAndPassword([]byte(savedPassword), []byte(password)); err != nil {\n\t\tfmt.Println(err)\n\t\treturn false\n\t}\n\n\treturn true\n}", "func IsPwdMatch(user *entities.User, password string) bool {\n\treturn utils.Sha(password) == user.Password\n}", "func (us *userService) Authenticate(email, password string) (*User, error) {\n\tuser, err := us.GetByEmail(email)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = bcrypt.CompareHashAndPassword(\n\t\t[]byte(user.PasswordHash),\n\t\t[]byte(password+us.pepper),\n\t)\n\n\tif err != nil {\n\t\treturn nil, ErrInvalidEmailOrPassword\n\t}\n\n\treturn user, nil\n}", "func VerifyPassword(rawPwd, encodedPwd string) bool {\n\n\t// for discuz accounts\n\tif len(encodedPwd) == 39 {\n\t\tsalt := encodedPwd[:6]\n\t\tencoded := encodedPwd[7:]\n\t\treturn encoded == utils.EncodeMd5(utils.EncodeMd5(rawPwd)+salt)\n\t}\n\n\t// split\n\tvar salt, encoded string\n\tif len(encodedPwd) > 11 {\n\t\tsalt = encodedPwd[:10]\n\t\tencoded = encodedPwd[11:]\n\t}\n\n\treturn utils.EncodePassword(rawPwd, salt) == encoded\n}", "func (e EncryptionHandler) PasswordCompare(c echo.Context) error {\n\tbody, err := ioutil.ReadAll(c.Request().Body)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tlogrus.Infof(\"body %v\", string(body))\n\tvar auth models.Auth\n\terr = json.Unmarshal(body, &auth)\n\tif err != nil {\n\t\tlogrus.Errorf(\"unable to unmarshal json %v\", err)\n\t\treturn err\n\t}\n\n\terr = bcrypt.CompareHashAndPassword([]byte(auth.SaltedPassword), []byte(auth.Password))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tc.JSON(200, gin.H{\"message\": \"salted password matched unencrypted\"})\n\treturn nil\n}", "func CheckAuth(username, password string) bool {\n\ttrx := db.Begin()\n\tdefer trx.Commit()\n\n\thash := md5.New()\n\tio.WriteString(hash, password)\n\tpassword = fmt.Sprintf(\"%x\", hash.Sum(nil))\t//\tfor safety, don't just save the plain text\n\tauth := Auth{}\n\ttrx.Set(\"gorm:query_option\", \"FOR UPDATE\").\n\t\tWhere(\"user_name = ? AND password = ?\", username, password).\n\t\tFirst(&auth)\n\tif auth.ID > 0 {\n\t\treturn true\n\t}\n\treturn false\n}", "func Login(e *Engine, username string, password string) (bool, error) {\n\tres, _, err := e.RawSelect(Filter(\"autoscope_users\",\n\t\tmap[string]interface{}{\"username\": username}))\n\tif err != nil { return false, err }\n\tuser, err := GetRow(res)\n\tif err != nil { return false, err }\n\n\t//CompareHashAndPassword returns nil on success\n\tsalted := password + strconv.FormatInt(user[\"salt\"].(int64), 10)\n\thashErr := bcrypt.CompareHashAndPassword([]byte(user[\"passhash\"].(string)),\n\t\t[]byte(salted))\n\treturn hashErr == nil, hashErr\n}", "func (s service) authenticate(ctx context.Context, username, password string) (*user.User, error) {\n\tlogger := s.logger.With(ctx, \"user\", username)\n\n\tuser := s.userService.NewEntity()\n\tuser.Name = username\n\n\tuser, err := s.userService.First(ctx, user)\n\tif err != nil {\n\t\treturn user, errorshandler.BadRequest(\"User not found\")\n\t}\n\n\tif comparePassword([]byte(user.Passhash), []byte(password)) {\n\t\tlogger.Infof(\"authentication successful\")\n\t\treturn user, nil\n\t}\n\n\tlogger.Infof(\"authentication failed\")\n\treturn user, errorshandler.Unauthorized(\"\")\n}", "func (c *Client) authenticate() error {\n\t// Send the packet.\n\thead, _, err := c.sendPacket(packetTypeAuth, []byte(c.password))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// If the credentials were bad, throw error.\n\tif head.RequestID == requestIDBadLogin {\n\t\treturn errors.New(\"Bad auth, could not authenticate.\")\n\t}\n\n\treturn nil\n}", "func (s *Passhash) Match(hash, password string) bool {\n\treturn passhash.MatchString(hash, password)\n}", "func IsAuthenticate(s string) bool {\n\tswitch s {\n\tcase\n\t\t\"all\",\n\t\t\"authenticate\":\n\t\treturn true\n\t}\n\treturn false\n}", "func (us *userService) Authenticate(email, password string) (*User, error) {\n\tfoundUser, err := us.ByEmail(email)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = bcrypt.CompareHashAndPassword([]byte(foundUser.PasswordHash), []byte(password+userPwPepper))\n\tswitch err {\n\tcase nil:\n\t\treturn foundUser, nil\n\tcase bcrypt.ErrMismatchedHashAndPassword:\n\t\treturn nil, ErrPasswordIncorrect\n\tdefault:\n\t\treturn nil, err\n\t}\n}", "func comparePassword(hash, pw []byte) bool {\n\treturn bytes.Equal(hash, hashPassword(pw, hash[:saltSize]))\n}", "func Authenticate(st db.Storer, email, pass string) (*User, error) {\n\tusr, err := GetUserWithEmail(st, email)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif usr.Deleted {\n\t\treturn nil, ErrUserDeleted\n\t}\n\terr = bcrypt.CompareHashAndPassword([]byte(usr.HashedPass), []byte(pass))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ttok, err := token.New(usr.UserID)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tusr.Token = tok\n\treturn usr, nil\n}", "func ComparePassword(hash string, password string) bool {\n\n\tif len(password) == 0 || len(hash) == 0 {\n\t\treturn false\n\t}\n\n\terr := bcrypt.CompareHashAndPassword([]byte(hash), []byte(password))\n\treturn err == nil\n}", "func (us *userService) Authenticate(email, password string) (*User, error) {\n\tfoundUser, err := us.ByEmail(email)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\terr = bcrypt.CompareHashAndPassword(\n\t\t[]byte(foundUser.PasswordHash),\n\t\t[]byte(password+userPwPepper))\n\tswitch err {\n\tcase nil:\n\t\treturn foundUser, nil\n\tcase bcrypt.ErrMismatchedHashAndPassword:\n\t\treturn nil, ErrPasswordIncorrect\n\tdefault:\n\t\treturn nil, err\n\t}\n}", "func (s service) authenticate(ctx context.Context, username, password string) (\n\tIdentity, error) {\n\n\tuser, err := s.userSvc.Get(ctx, username)\n\tif err != nil {\n\t\treturn nil, errUserNotFound\n\t}\n\tif !s.sec.HashMatchesPassword(user.Password, password) {\n\t\treturn nil, errWrongPassword\n\t}\n\tif !user.Confirmed {\n\t\treturn nil, errUserNotConfirmed\n\t}\n\treturn entity.User{Username: username}, nil\n}", "func (u *User) ComparePassword(password string) bool {\n\treturn bcrypt.CompareHashAndPassword([]byte(u.EncryptedPassword), []byte(password)) == nil\n}", "func (c client) Auth(username, password string) error {\n\t// establish connection\n\tconn, err := connect(c.Host)\n\tif err != nil {\n\t\tWriteLogFile(err)\n\t\treturn err\n\t}\n\tdefer conn.Close()\n\n\t// perform initial read-only bind\n\tif err = conn.Bind(c.ROUser.Name, c.ROUser.Password); err != nil {\n\t\tWriteLogFile(err)\n\t\treturn err\n\t}\n\n\t// find the user attempting to login\n\tresults, err := conn.Search(ldap.NewSearchRequest(\n\t\tc.BaseDN, ldap.ScopeWholeSubtree,\n\t\tldap.NeverDerefAliases,\n\t\t0, 0, false, fmt.Sprintf(\"(%v=%v)\", c.Filter, username),\n\t\t[]string{}, nil,\n\t))\n\tif err != nil {\n\t\tWriteLogFile(err)\n\t\treturn err\n\t}\n\tif len(results.Entries) < 1 {\n\t\treturn errors.New(\"not found\")\n\t}\n\n\t// attempt auth\n\tlog.Println(results.Entries)\n\t//role = strings.Join(results.Entries[0].Attributes[3].Values, \"\")\n\treturn conn.Bind(results.Entries[0].DN, password)\n}" ]
[ "0.73781884", "0.7255777", "0.7220583", "0.71641976", "0.71482325", "0.7145474", "0.7134889", "0.7122546", "0.70890397", "0.7088894", "0.7088624", "0.70870215", "0.702402", "0.7004266", "0.6986257", "0.69798476", "0.69560224", "0.69157857", "0.6909883", "0.6903701", "0.6870277", "0.68240976", "0.6766762", "0.6760372", "0.6756286", "0.670577", "0.666685", "0.6659717", "0.664221", "0.663011", "0.6625684", "0.6587197", "0.6572169", "0.65470695", "0.6544128", "0.652793", "0.6512962", "0.64531106", "0.6450248", "0.64449066", "0.64245397", "0.64127064", "0.64038455", "0.63956845", "0.6374165", "0.63736", "0.6350879", "0.63273", "0.63234204", "0.62934643", "0.62721574", "0.6267367", "0.6262891", "0.62578213", "0.62557393", "0.6246338", "0.623657", "0.62362593", "0.622586", "0.6185134", "0.6180766", "0.61630356", "0.61543447", "0.614899", "0.613312", "0.6127531", "0.6111082", "0.6091253", "0.6083658", "0.60826695", "0.6075993", "0.60673445", "0.6066753", "0.6042919", "0.6037969", "0.60346323", "0.60338223", "0.6025491", "0.6023235", "0.6019686", "0.6016012", "0.60113275", "0.60049915", "0.60022664", "0.5990886", "0.59857374", "0.5978607", "0.59627414", "0.5962195", "0.5960446", "0.59576106", "0.5956001", "0.59501433", "0.5932803", "0.59250647", "0.5924063", "0.59219927", "0.5919475", "0.59007955", "0.5900095" ]
0.624377
56
Get takes name of the tZCronJob, and returns the corresponding tZCronJob object, and an error if there is any.
func (c *FakeTZCronJobs) Get(name string, options v1.GetOptions) (result *v1alpha1.TZCronJob, err error) { obj, err := c.Fake. Invokes(testing.NewGetAction(tzcronjobsResource, c.ns, name), &v1alpha1.TZCronJob{}) if obj == nil { return nil, err } return obj.(*v1alpha1.TZCronJob), err }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func GetJob(name string) (Job, error) {\n\tj, ok := jobs[name]\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"Job(%v) not found\", name)\n\t}\n\treturn j, nil\n}", "func (c *scheduledJobs) Get(name string) (result *batch.ScheduledJob, err error) {\n\tresult = &batch.ScheduledJob{}\n\terr = c.r.Get().Namespace(c.ns).Resource(\"scheduledjobs\").Name(name).Do().Into(result)\n\treturn\n}", "func GetJob(ctx *gin.Context) {\n\tlog := logger.RuntimeLog\n\tzoneName := ctx.Param(\"zone\")\n\tnamespace := ctx.Param(\"ns\")\n\tname := ctx.Param(\"name\")\n\n\t// fetch k8s-client handler by zoneName\n\tkclient, err := GetClientByAzCode(zoneName)\n\tif err != nil {\n\t\tlog.WithError(err)\n\t\tSendResponse(ctx, errno.ErrTokenInvalid, nil)\n\t}\n\n\tstartAt := time.Now()\n\tjbs, err := kclient.BatchV1().Jobs(namespace).Get(name, metav1.GetOptions{})\n\tif err != nil {\n\t\tSendResponse(ctx, err, \"failed to get job info.\")\n\t\treturn\n\t}\n\tlogger.MetricsEmit(\n\t\tJOB_CONST.K8S_LOG_Method_GetJob,\n\t\tutil.GetReqID(ctx),\n\t\tfloat32(time.Since(startAt)/time.Millisecond),\n\t\terr == err,\n\t)\n\n\tSendResponse(ctx, errno.OK, jbs)\n}", "func GetJob(ns, name string) (*v1.Job, error) {\n\treturn kubecli.BatchV1().Jobs(ns).Get(name, metav1.GetOptions{})\n}", "func (c *Clientset) GetJob(namespace, name string) (*batch_v1.Job, error) {\n\treturn c.kubernetes.BatchV1().Jobs(namespace).Get(name, metav1.GetOptions{})\n}", "func GetJob(ctx context.Context, p string) (j *Job, err error) {\n\tt := utils.FromTaskContext(ctx)\n\n\tj = &Job{}\n\n\tcontent, err := contexts.DB.Get(constants.FormatJobKey(t, p), nil)\n\tif err == leveldb.ErrNotFound {\n\t\treturn nil, nil\n\t}\n\tif err != nil {\n\t\treturn\n\t}\n\n\terr = msgpack.Unmarshal(content, j)\n\tif err != nil {\n\t\tlogrus.Panicf(\"Msgpack unmarshal failed for %v.\", err)\n\t}\n\treturn\n}", "func (w *Worker) GetJob(namespace, id string) (*Job, error) {\n\tw.shutdownLock.RLock()\n\tdefer w.shutdownLock.RLock()\n\tif w.shutdown {\n\t\treturn nil, ErrShutdown\n\t}\n\tw.jobsLock.RLock()\n\tdefer w.jobsLock.RUnlock()\n\treturn w.jobs[namespace][id], nil\n}", "func getCronJobByJob(clientset *kubernetes.Clientset, namespace, name string) (*Owner, error) {\n\tjob, err := clientset.BatchV1().Jobs(namespace).Get(name, meta_v1.GetOptions{})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\towner := &Owner{}\n\tfor _, ref := range job.OwnerReferences {\n\t\tif strings.EqualFold(ref.Kind, \"CronJob\") {\n\t\t\towner.Kind = ref.Kind\n\t\t\towner.Name = ref.Name\n\t\t\towner.Controller = ref.Controller\n\t\t\towner.ApiVersion = ref.APIVersion\n\t\t\treturn owner, nil\n\t\t}\n\t}\n\treturn owner, fmt.Errorf(\"Job %s/%s has no controller of cronjob\", namespace, name)\n}", "func GetJob(client kubernetes.Interface, options GetJobOptions) (*v1.Job, error) {\n\tkey := fmt.Sprintf(\"%s-%s\", options.Name, options.Namespace)\n\tif job, found := jobs[key]; found {\n\t\treturn job, nil\n\t}\n\tresource, err := client.BatchV1().Jobs(options.Namespace).Get(options.Name, meta_v1.GetOptions{})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tjobs[key] = resource\n\n\treturn resource, nil\n}", "func (r *TransferJobsService) Get(jobName string) *TransferJobsGetCall {\n\tc := &TransferJobsGetCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.jobName = jobName\n\treturn c\n}", "func (s *Scheduler) GetJob(jid ffs.JobID) (ffs.Job, error) {\n\tj, err := s.js.Get(jid)\n\tif err != nil {\n\t\tif err == jstore.ErrNotFound {\n\t\t\treturn ffs.Job{}, ErrNotFound\n\t\t}\n\t\treturn ffs.Job{}, fmt.Errorf(\"get Job from store: %s\", err)\n\t}\n\treturn j, nil\n}", "func (cl *Client) GetJob(jobID string) (*schedulerpb.Job, error) {\n\tif jobID == \"\" {\n\t\treturn nil, errors.New(\"No job ID provided\")\n\t}\n\treq := &schedulerpb.GetJobRequest{\n\t\tName: removeColon(jobID),\n\t}\n\tresp, err := cl.CloudSchedulerClient.GetJob(cl.ctx, req)\n\tif err != nil {\n\t\tif strings.Contains(fmt.Sprintf(\"%s\", err), \"or the resource may not exist\") {\n\t\t\treturn nil, nil\n\t\t}\n\t\tif strings.Contains(fmt.Sprintf(\"%s\", err), \"Job not found\") {\n\t\t\treturn nil, nil\n\t\t}\n\t\treturn nil, errors.Wrap(err, \"GetJob->cl.CloudSchedulerClient.GetJob\")\n\t}\n\treturn resp, nil\n}", "func (d DB) Get(id string) (*job.Job, error) {\n\ttemplate := `select to_jsonb(j.job) from (select * from %[1]s where job ->> 'id' = $1) as j;`\n\tquery := fmt.Sprintf(template, TABLE_NAME)\n\tvar r sql.NullString\n\terr := d.conn.QueryRow(query, id).Scan(&r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresult := &job.Job{}\n\tif r.Valid {\n\t\terr = json.Unmarshal([]byte(r.String), &result)\n\t}\n\treturn result, err\n}", "func (m *JobManager) Get(path, body string) (*Job, error) {\n\treturn m.table.Get(path, body)\n}", "func (n *Namespace) GetJob(backend Backend, jobID string) (*Job, error) {\n\n\t// ensure the namespace exists\n\terr := n.checkOrCreateNamespace(backend)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// get the stored JSON\n\tjson, _, err := backend.ReadKey(getJobPath(n.namespace, jobID))\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"problem retrieving job %v\", err)\n\t}\n\n\t// load into a Job\n\tjob := &Job{}\n\terr = job.Deserialize(json)\n\treturn job, err\n}", "func GetJob(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *JobState, opts ...pulumi.ResourceOption) (*Job, error) {\n\tvar resource Job\n\terr := ctx.ReadResource(\"aws-native:databrew:Job\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (s *inMemoryJobStore) Get(ctx context.Context, name string) (*v1.JobStatus, error) {\n\ts.mu.RLock()\n\tjob, ok := s.jobs[name]\n\ts.mu.RUnlock()\n\n\tif !ok {\n\t\treturn nil, ErrNotFound\n\t}\n\n\treturn &job, nil\n}", "func (c *CloudSchedulerClient) GetJob(ctx context.Context, req *schedulerpb.GetJobRequest, opts ...gax.CallOption) (*schedulerpb.Job, error) {\n\treturn c.internalClient.GetJob(ctx, req, opts...)\n}", "func (q *priorityLocalQueue) Get(ctx context.Context, name string) (amboy.Job, bool) {\n\treturn q.storage.Get(name)\n}", "func (m *queueManager) getJob(ctx context.Context, info amboy.JobInfo) (amboy.Job, error) {\n\tif info.Retry.Retryable {\n\t\tvar j amboy.Job\n\t\tvar err error\n\t\tisRetryable := amboy.WithRetryableQueue(m.queue, func(rq amboy.RetryableQueue) {\n\t\t\tj, err = rq.GetAttempt(ctx, info.ID, info.Retry.CurrentAttempt)\n\t\t})\n\n\t\t// If the queue is retryable, return the result immediately. Otherwise,\n\t\t// if it's not a retryable queue, then a retryable job is treated no\n\t\t// differently from a non-retryable job (i.e. it's not retried).\n\t\tif isRetryable {\n\t\t\treturn j, err\n\t\t}\n\t}\n\n\tj, ok := m.queue.Get(ctx, info.ID)\n\tif !ok {\n\t\treturn j, errors.New(\"could not find job\")\n\t}\n\treturn j, nil\n\n}", "func (r *JobsService) Get(name string) *JobsGetCall {\n\tc := &JobsGetCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.name = name\n\treturn c\n}", "func (cl *Client) GetJob() (*MapReduceJob, error) {\n\tresp, err := cl.client.Get(cl.baseurl)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif resp.StatusCode != http.StatusOK {\n\t\treturn nil, fmt.Errorf(\"Got status: %s\", resp.Status)\n\t}\n\tdefer resp.Body.Close()\n\tdecoder := json.NewDecoder(resp.Body)\n\tjb := &MapReduceJob{}\n\terr = decoder.Decode(jb)\n\treturn jb, err\n}", "func (c *controller) GetJob(namespace, jobName string) (*apibatchv1.Job, error) {\n\treturn c.k8sBatchClient.Jobs(namespace).Get(jobName, metav1.GetOptions{})\n}", "func (c *aITrainingJobs) Get(name string, options metav1.GetOptions) (result *v1.AITrainingJob, err error) {\n\tresult = &v1.AITrainingJob{}\n\terr = c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"aitrainingjobs\").\n\t\tName(name).\n\t\tVersionedParams(&options, scheme.ParameterCodec).\n\t\tDo().\n\t\tInto(result)\n\treturn\n}", "func GetJob(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *JobState, opts ...pulumi.ResourceOption) (*Job, error) {\n\tvar resource Job\n\terr := ctx.ReadResource(\"google-native:dataflow/v1b3:Job\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func GetJob(params job.GetJobParams) middleware.Responder {\n\tjobRepo, err := repository.GetCassandraJobberRepository()\n\tif err != nil {\n\t\tnewErr := fmt.Errorf(\"Unable to access jobs repository: %v\", err)\n\t\tse := createServiceError(newErr, http.StatusInternalServerError)\n\t\treturn job.NewGetJobDefault(0).WithPayload(&se)\n\t}\n\n\tj, err := jobRepo.GetJob(params.ID)\n\tif err != nil {\n\t\tnewErr := fmt.Errorf(\"Unable to get job: %v\", err)\n\t\tse := createServiceError(newErr, http.StatusInternalServerError)\n\t\treturn job.NewGetJobDefault(0).WithPayload(&se)\n\t}\n\n\tif j == nil {\n\t\tnewErr := fmt.Errorf(\"Unable to locate job: %s\", params.ID)\n\t\tse := createServiceError(newErr, http.StatusNotFound)\n\t\treturn job.NewGetJobDefault(0).WithPayload(&se)\n\t}\n\n\tapiJob := j.ToAPI()\n\treturn job.NewGetJobOK().WithPayload(&apiJob)\n}", "func (d *defaultJobRepository) GetJob(ctxIn context.Context, jobID string) (*Job, error) {\n _, span := trace.StartSpan(ctxIn, \"(*defaultJobRepository).GetJob\")\n defer span.End()\n\n job := &Job{ID: jobID}\n err := d.storageService.DB().\n Model(job).\n Where(\"id = ?\", jobID).\n Where(\"audit_deleted_timestamp is null\").\n Select()\n\n if err == pg.ErrNoRows {\n return nil, nil\n } else if err != nil {\n return nil, fmt.Errorf(\"get job with id %s failed: %s\", jobID, err)\n }\n\n return job, err\n}", "func (m *MemoryStorage) Get(uuid string) (Job, error) {\n\trawJob, ok := m.jobs.Load(uuid)\n\tif !ok {\n\t\treturn Job{}, ErrJobNotFound\n\t}\n\n\t// cast to pointer as loadResult has a pointer receiver\n\tjob := rawJob.(*Job)\n\n\tres, _ := m.results.Load(uuid)\n\n\tvar errString string\n\n\trawErr, ok := m.errors.Load(uuid)\n\tif ok {\n\t\terrString = rawErr.(string)\n\t}\n\n\tjob.loadResult(res, errString)\n\n\treturn *job, nil\n}", "func (s *deploymentServer) findCronjob(ctx context.Context, ns, name, tier string) (*apibatch.CronJob, error) {\n\t// log.Println(\"find cronjob \" + ns + \" : \" + name + \".\" + tier)\n\tlog.Println(\"find cronjob \" + ns + \" : \" + name)\n\tbatchAPI := s.clientset.BatchV1beta1()\n\tapiJobs := batchAPI.CronJobs(ns)\n\n\t// cronjob, err := apiJobs.Get(ctx, name+\".\"+tier, metav1.GetOptions{})\n\tcronjob, err := apiJobs.Get(ctx, name, metav1.GetOptions{})\n\tif err != nil {\n\t\tswitch t := err.(type) {\n\t\tcase *errors.StatusError:\n\t\t\t{\n\t\t\t\tstatusCode := t.Status().Code\n\t\t\t\tif statusCode == 404 {\n\t\t\t\t\treturn nil, nil\n\t\t\t\t}\n\t\t\t\treturn nil, fmt.Errorf(\"could not get cronjob `%s`, got error '%v' with status %d\", name, err, statusCode)\n\t\t\t}\n\t\t}\n\t\treturn nil, fmt.Errorf(\"could not get cronjob `%s`, got error '%v'\", name, err)\n\t}\n\treturn cronjob, nil\n}", "func (engine *JobEngine) GetJobByName(name string) (types.Job, error) {\n\tengine.jobsMapSync.Lock()\n\tdefer engine.jobsMapSync.Unlock()\n\n\tjob, ok := engine.jobs[name]\n\tif !ok {\n\t\treturn types.Job{}, errors.New(\"Job \" + name + \" not found\")\n\t}\n\treturn job.Copy(), nil\n}", "func (m *Mongo) GetJob(ctx context.Context, id string) (*models.Job, error) {\n\tvar job models.Job\n\tif err := m.connection.Collection(m.ActualCollectionName(config.ImportsCollection)).FindOne(ctx, bson.M{\"id\": id}, &job); err != nil {\n\t\tif errors.Is(err, mongodriver.ErrNoDocumentFound) {\n\t\t\treturn nil, apierrors.ErrJobNotFound\n\t\t}\n\t\treturn nil, err\n\t}\n\n\treturn &job, nil\n}", "func (s *ContinuousRestoreJobsServiceOp) Get(ctx context.Context, groupID, clusterID, jobID string) (*atlas.ContinuousJob, *Response, error) {\n\tif clusterID == \"\" {\n\t\treturn nil, nil, atlas.NewArgError(\"clusterID\", \"must be set\")\n\t}\n\tif groupID == \"\" {\n\t\treturn nil, nil, atlas.NewArgError(\"groupID\", \"must be set\")\n\t}\n\tif jobID == \"\" {\n\t\treturn nil, nil, atlas.NewArgError(\"jobID\", \"must be set\")\n\t}\n\tdefaultPath := fmt.Sprintf(continuousRestoreJobsPath, groupID, clusterID)\n\n\tpath := fmt.Sprintf(\"%s/%s\", defaultPath, jobID)\n\n\treq, err := s.Client.NewRequest(ctx, http.MethodGet, path, nil)\n\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\troot := new(atlas.ContinuousJob)\n\tresp, err := s.Client.Do(ctx, req, root)\n\n\treturn root, resp, err\n}", "func (c *JobClient) Get(ctx context.Context, id int) (*Job, error) {\n\treturn c.Query().Where(job.ID(id)).Only(ctx)\n}", "func (c *cloudSchedulerRESTClient) GetJob(ctx context.Context, req *schedulerpb.GetJobRequest, opts ...gax.CallOption) (*schedulerpb.Job, error) {\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v1/%v\", req.GetName())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"name\", url.QueryEscape(req.GetName()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\topts = append((*c.CallOptions).GetJob[0:len((*c.CallOptions).GetJob):len((*c.CallOptions).GetJob)], opts...)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tresp := &schedulerpb.Job{}\n\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"GET\", baseUrl.String(), nil)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}, opts...)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\treturn resp, nil\n}", "func (c *Jobs) GetJob(id int) (out *JobOut, err error) {\n\tbody, err := c.call(\"GET\", fmt.Sprintf(\"/jobs/%d\", id), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer body.Close()\n\n\terr = json.NewDecoder(body).Decode(&out)\n\treturn\n}", "func GetJob(conf *entity.JenkinsConfig, jobName string) (job gojenkins.Job, err error) {\n\tdebug(\"[GetJob]conf\", *conf)\n\tdebug(\"[GetJob]jobName\", jobName)\n\n\tauth := &gojenkins.Auth{\n\t\tUsername: conf.User,\n\t\tApiToken: conf.Token,\n\t}\n\tjenkins := gojenkins.NewJenkins(auth, conf.URL)\n\tjob, err = jenkins.GetJob(jobName)\n\n\tif err != nil {\n\t\tFatalf(\"error exec command: %s\", err)\n\t\treturn\n\t}\n\n\treturn\n}", "func (r *ProjectsPatchJobsService) Get(name string) *ProjectsPatchJobsGetCall {\n\tc := &ProjectsPatchJobsGetCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.name = name\n\treturn c\n}", "func (m *DataHistoryManager) GetByNickname(nickname string, fullDetails bool) (*DataHistoryJob, error) {\n\tif m == nil {\n\t\treturn nil, ErrNilSubsystem\n\t}\n\tif atomic.LoadInt32(&m.started) == 0 {\n\t\treturn nil, ErrSubSystemNotStarted\n\t}\n\tif fullDetails {\n\t\tdbJ, err := m.jobDB.GetJobAndAllResults(nickname)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"job %s could not load job from database: %w\", nickname, err)\n\t\t}\n\t\tresult, err := m.convertDBModelToJob(dbJ)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"could not convert model with nickname %s %w\", nickname, err)\n\t\t}\n\t\treturn result, nil\n\t}\n\tm.m.Lock()\n\tfor i := range m.jobs {\n\t\tif strings.EqualFold(m.jobs[i].Nickname, nickname) {\n\t\t\tcpy := m.jobs[i]\n\t\t\tm.m.Unlock()\n\t\t\treturn cpy, nil\n\t\t}\n\t}\n\tm.m.Unlock()\n\t// now try the database\n\tj, err := m.jobDB.GetByNickName(nickname)\n\tif err != nil {\n\t\tif errors.Is(err, sql.ErrNoRows) {\n\t\t\t// no need to display normal sql err to user\n\t\t\treturn nil, errJobNotFound\n\t\t}\n\t\treturn nil, fmt.Errorf(\"job %s %w, %s\", nickname, errJobNotFound, err)\n\t}\n\tjob, err := m.convertDBModelToJob(j)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn job, nil\n}", "func GetJob(w http.ResponseWriter, r *http.Request) {\n\tresponse := services.LoadJob(r)\n\n\trender.Status(r, response.Code)\n\trender.JSON(w, r, response)\n}", "func (j *jobCache) Get(guid string) *model.Job {\n\tj.m.RLock()\n\tdefer j.m.RUnlock()\n\n\tjob, ok := j.store[guid]\n\tif ok {\n\t\treturn job\n\t}\n\treturn nil\n}", "func (schematics *SchematicsV1) GetJob(getJobOptions *GetJobOptions) (result *Job, response *core.DetailedResponse, err error) {\n\treturn schematics.GetJobWithContext(context.Background(), getJobOptions)\n}", "func (c *Client) GetJob(jobID string) (*Job, error) {\n\tclient := &http.Client{}\n\turl := fmt.Sprintf(\"http://%s/api/v1/jobs/%s\", c.options.ServerAddr, jobID)\n\treq, err := http.NewRequest(\"GET\", url, nil)\n\tif err != nil {\n\t\tlogrus.WithError(err)\n\t\treturn nil, err\n\t}\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\tlogrus.WithError(err)\n\t\treturn nil, err\n\t}\n\tdefer resp.Body.Close()\n\n\tbody, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\tlogrus.WithError(err)\n\t\treturn nil, err\n\t}\n\tjob := &Job{}\n\tif err := json.Unmarshal(body, job); err != nil {\n\t\tlogrus.WithError(err)\n\t\treturn nil, err\n\t}\n\treturn job, nil\n}", "func (self *TaskBolt) GetJob(ctx context.Context, job *ga4gh_task_exec.JobId) (*ga4gh_task_exec.Job, error) {\n\tlog.Printf(\"Getting Task Info\")\n\tch := make(chan *ga4gh_task_exec.Task, 1)\n\tself.db.View(func(tx *bolt.Tx) error {\n\t\tb := tx.Bucket(TASK_BUCKET)\n\t\tv := b.Get([]byte(job.Value))\n\t\tout := ga4gh_task_exec.Task{}\n\t\tproto.Unmarshal(v, &out)\n\t\tch <- &out\n\t\treturn nil\n\t})\n\ta := <-ch\n\tif a == nil {\n\t\treturn nil, fmt.Errorf(\"Job Not Found\")\n\t}\n\tb, err := self.getTaskJob(a)\n\treturn b, err\n}", "func (agent *Agent) GetJob() *Job {\n\tagent.rwLock.RLock()\n\tdefer agent.rwLock.RUnlock()\n\tif agent.currentJob == nil {\n\t\treturn nil\n\t}\n\treturn agent.currentJob.Copy()\n}", "func (b *BQService) GetJob(jobID string) *Job {\n\tj, err := b.jobsService.Get(b.projectID, jobID).Do()\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tjob := b.parseJob(j)\n\treturn job\n}", "func (m *DataHistoryManager) GetByNickname(nickname string, fullDetails bool) (*DataHistoryJob, error) {\n\tif m == nil {\n\t\treturn nil, ErrNilSubsystem\n\t}\n\tif atomic.LoadInt32(&m.started) == 0 {\n\t\treturn nil, ErrSubSystemNotStarted\n\t}\n\tif fullDetails {\n\t\tdbJ, err := m.jobDB.GetJobAndAllResults(nickname)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"job %s could not load job from database: %w\", nickname, err)\n\t\t}\n\t\tresult, err := m.convertDBModelToJob(dbJ)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"could not convert model with nickname %s %w\", nickname, err)\n\t\t}\n\t\treturn result, nil\n\t}\n\tj, err := m.jobDB.GetByNickName(nickname)\n\tif err != nil {\n\t\tif err == sql.ErrNoRows {\n\t\t\t// no need to display normal sql err to user\n\t\t\treturn nil, errJobNotFound\n\t\t}\n\t\treturn nil, fmt.Errorf(\"job %s %w, %s\", nickname, errJobNotFound, err)\n\t}\n\tjob, err := m.convertDBModelToJob(j)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn job, nil\n}", "func (r *Reconciler) getJob(ctx context.Context, perfJob *perfv1alpha1.PerfJob, ls labels.Selector) (*v1.Job, error) {\n\tlist := &v1.JobList{}\n\topts := &client.ListOptions{\n\t\tNamespace: perfJob.Namespace,\n\t\tLabelSelector: ls,\n\t\t// Set Raw because if we need to get more than one page, then we will put the continue token\n\t\t// into opts.Raw.Continue.\n\t\tRaw: &metav1.ListOptions{},\n\t}\n\n\terr := r.List(ctx, opts, list)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tfor _, i := range list.Items {\n\t\tif metav1.IsControlledBy(&i, perfJob) {\n\t\t\treturn &i, nil\n\t\t}\n\t}\n\n\treturn nil, k8serrors.NewNotFound(schema.GroupResource{}, \"\")\n}", "func (c *JobsClient) GetJob(ctx context.Context, req *runpb.GetJobRequest, opts ...gax.CallOption) (*runpb.Job, error) {\n\treturn c.internalClient.GetJob(ctx, req, opts...)\n}", "func (t *JobImpl) GetJob() *gojenkins.Job {\n\treturn t.Job\n}", "func (c *jobsRESTClient) GetJob(ctx context.Context, req *runpb.GetJobRequest, opts ...gax.CallOption) (*runpb.Job, error) {\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v2/%v\", req.GetName())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"name\", url.QueryEscape(req.GetName()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\topts = append((*c.CallOptions).GetJob[0:len((*c.CallOptions).GetJob):len((*c.CallOptions).GetJob)], opts...)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tresp := &runpb.Job{}\n\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"GET\", baseUrl.String(), nil)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}, opts...)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\treturn resp, nil\n}", "func (db jobDatabase) GetJobByID(id int) (mgm.Job, error) {\n\tcon, err := db.mysql.GetConnection()\n\tif err != nil {\n\t\treturn mgm.Job{}, err\n\t}\n\tdefer con.Close()\n\n\tj := mgm.Job{}\n\terr = con.QueryRow(\"SELECT * FROM jobs WHERE id=?\", id).Scan(&j.ID, &j.Timestamp, &j.Type, &j.User, &j.Data)\n\tif err != nil {\n\t\treturn mgm.Job{}, err\n\t}\n\n\treturn j, nil\n}", "func GetJob(cmd *cobra.Command, args []string) error {\n\n\tclient, err := auth.GetClient()\n\tif err != nil {\n\t\treturn err\n\t}\n\t// Parse all flags\n\n\tvar sid string\n\terr = flags.ParseFlag(cmd.Flags(), \"sid\", &sid)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"sid\": ` + err.Error())\n\t}\n\n\t// Silence Usage\n\tcmd.SilenceUsage = true\n\n\tresp, err := client.SearchService.GetJob(sid)\n\tif err != nil {\n\t\treturn err\n\t}\n\tjsonx.Pprint(cmd, resp)\n\treturn nil\n}", "func (d *InMemoryJobDB) GetJobById(_ context.Context, id string) (*types.Job, error) {\n\td.jobsMtx.RLock()\n\tdefer d.jobsMtx.RUnlock()\n\tif job := d.jobs[id]; job != nil {\n\t\treturn job.Copy(), nil\n\t}\n\treturn nil, nil\n}", "func (s elasticDLJobNamespaceLister) Get(name string) (*v1alpha1.ElasticDLJob, error) {\n\tobj, exists, err := s.indexer.GetByKey(s.namespace + \"/\" + name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif !exists {\n\t\treturn nil, errors.NewNotFound(v1alpha1.Resource(\"elasticdljob\"), name)\n\t}\n\treturn obj.(*v1alpha1.ElasticDLJob), nil\n}", "func (m *MemoryStorage) Get(checkID string) (Job, error) {\n\tm.mutex.RLock()\n\tdefer m.mutex.RUnlock()\n\n\tjob, ok := m.checks[checkID]\n\tif !ok {\n\t\treturn Job{}, errors.New(\"error retrieving unexisting job\")\n\t}\n\n\treturn job, nil\n}", "func (taskBolt *TaskBolt) GetJob(ctx context.Context, id *ga4gh_task_exec.JobID) (*ga4gh_task_exec.Job, error) {\n\n\tvar job *ga4gh_task_exec.Job\n\terr := taskBolt.db.View(func(tx *bolt.Tx) error {\n\t\tjob = getJob(tx, id.Value)\n\t\tloadJobLogs(tx, job)\n\t\treturn nil\n\t})\n\treturn job, err\n}", "func (jm *JobManager) Job(jobName string) (job Job) {\n\tjm.Lock()\n\tif jobMeta, hasJob := jm.jobs[jobName]; hasJob {\n\t\tjob = jobMeta.Job\n\t}\n\tjm.Unlock()\n\treturn\n}", "func (sched *StdScheduler) GetScheduledJob(key int) (*ScheduledJob, error) {\n\tsched.Lock()\n\tdefer sched.Unlock()\n\n\tfor _, item := range *sched.Queue {\n\t\tif item.Job.Key() == key {\n\t\t\treturn &ScheduledJob{\n\t\t\t\titem.Job,\n\t\t\t\titem.Trigger.Description(),\n\t\t\t\titem.priority,\n\t\t\t}, nil\n\t\t}\n\t}\n\n\treturn nil, errors.New(\"No Job with the given Key found\")\n}", "func (r *ProjectsLocationsJobsService) Get(name string) *ProjectsLocationsJobsGetCall {\n\tc := &ProjectsLocationsJobsGetCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.name = name\n\treturn c\n}", "func (r *ExactMatchLookupJobRequest) Get(ctx context.Context) (resObj *ExactMatchLookupJob, err error) {\n\tvar query string\n\tif r.query != nil {\n\t\tquery = \"?\" + r.query.Encode()\n\t}\n\terr = r.JSONRequest(ctx, \"GET\", query, nil, &resObj)\n\treturn\n}", "func (client JobClient) Get(ctx context.Context, resourceGroupName string, automationAccountName string, jobName string, clientRequestID string) (result Job, err error) {\n\tif tracing.IsEnabled() {\n\t\tctx = tracing.StartSpan(ctx, fqdn+\"/JobClient.Get\")\n\t\tdefer func() {\n\t\t\tsc := -1\n\t\t\tif result.Response.Response != nil {\n\t\t\t\tsc = result.Response.Response.StatusCode\n\t\t\t}\n\t\t\ttracing.EndSpan(ctx, sc, err)\n\t\t}()\n\t}\n\tif err := validation.Validate([]validation.Validation{\n\t\t{TargetValue: resourceGroupName,\n\t\t\tConstraints: []validation.Constraint{{Target: \"resourceGroupName\", Name: validation.MaxLength, Rule: 90, Chain: nil},\n\t\t\t\t{Target: \"resourceGroupName\", Name: validation.MinLength, Rule: 1, Chain: nil},\n\t\t\t\t{Target: \"resourceGroupName\", Name: validation.Pattern, Rule: `^[-\\w\\._]+$`, Chain: nil}}}}); err != nil {\n\t\treturn result, validation.NewError(\"automation.JobClient\", \"Get\", err.Error())\n\t}\n\n\treq, err := client.GetPreparer(ctx, resourceGroupName, automationAccountName, jobName, clientRequestID)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"automation.JobClient\", \"Get\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresp, err := client.GetSender(req)\n\tif err != nil {\n\t\tresult.Response = autorest.Response{Response: resp}\n\t\terr = autorest.NewErrorWithError(err, \"automation.JobClient\", \"Get\", resp, \"Failure sending request\")\n\t\treturn\n\t}\n\n\tresult, err = client.GetResponder(resp)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"automation.JobClient\", \"Get\", resp, \"Failure responding to request\")\n\t\treturn\n\t}\n\n\treturn\n}", "func (c *Connection) GetJob(count int, timeout time.Duration, queues ...string) (Job, error){\n arguments := redis.Args{}.\n Add(\"TIMEOUT\").Add(int64(timeout.Seconds() * 1000)).\n Add(\"COUNT\").Add(count).\n Add(\"FROM\").AddFlat(queues)\n\n values, err := redis.Values(c.Do(\"GETJOB\", arguments...)); if err != nil {\n return Job{}, err\n }\n\n for _, value := range values {\n jobData, err := redis.Values(value, nil); if err != nil {\n return Job{}, err\n }\n\n if len(jobData) < 3 {\n return Job{}, errors.New(\"Malformed job fetched from Disque\")\n }\n\n return Job{\n Queue: string(jobData[0].([]byte)),\n ID: string(jobData[1].([]byte)),\n Payload: jobData[2].([]byte),\n }, nil\n }\n\n return Job{}, errors.New(\"timeout reached\")\n}", "func (y *Yeelight) CronGet(t string) string {\n\tcmd := `{\"id\":13,\"method\":\"cron_get\",\"params\":[` + t + `]}`\n\treturn y.request(cmd)\n}", "func (c *CISImpl) GetTryJob(ctx context.Context, id string) (ci.TryJob, error) {\n\treturn ci.TryJob{\n\t\tSystemID: id,\n\t\tSystem: c.system,\n\t\tDisplayName: id,\n\t\tUpdated: time.Now(),\n\t}, nil\n}", "func (client JobClient) Get(ctx context.Context, resourceGroupName string, accountName string, jobName string) (result JobResourceDescription, err error) {\n if tracing.IsEnabled() {\n ctx = tracing.StartSpan(ctx, fqdn + \"/JobClient.Get\")\n defer func() {\n sc := -1\n if result.Response.Response != nil {\n sc = result.Response.Response.StatusCode\n }\n tracing.EndSpan(ctx, sc, err)\n }()\n }\n req, err := client.GetPreparer(ctx, resourceGroupName, accountName, jobName)\n if err != nil {\n err = autorest.NewErrorWithError(err, \"microsoftazuremanagementaisupercomputer.JobClient\", \"Get\", nil , \"Failure preparing request\")\n return\n }\n\n resp, err := client.GetSender(req)\n if err != nil {\n result.Response = autorest.Response{Response: resp}\n err = autorest.NewErrorWithError(err, \"microsoftazuremanagementaisupercomputer.JobClient\", \"Get\", resp, \"Failure sending request\")\n return\n }\n\n result, err = client.GetResponder(resp)\n if err != nil {\n err = autorest.NewErrorWithError(err, \"microsoftazuremanagementaisupercomputer.JobClient\", \"Get\", resp, \"Failure responding to request\")\n }\n\n return\n}", "func (q *Queue) GetJobByID(id []byte) (*Job, error) {\n\tvar job *Job\n\terr := q.db.View(func(tx *nutsdb.Tx) error {\n\t\te, err := tx.Get(jobsBucketName, id)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tjob, err = DecodeJob(e.Value)\n\t\treturn err\n\t})\n\treturn job, err\n}", "func (tt *TensorFlowJobTrainer) getTrainingJobFromCache(name, ns string) (TrainingJob, error) {\n\n\tvar (\n\t\ttfjob tfv1.TFJob\n\t)\n\n\t// 1. Find the batch job\n\tfor _, item := range allTfjobs {\n\t\tif tt.isTensorFlowJob(name, ns, item) {\n\t\t\ttfjob = item\n\t\t\tbreak\n\t\t}\n\t}\n\ttfjob.Status.Conditions = makeJobStatusSortedByTime(tfjob.Status.Conditions)\n\t// 2. Find the pods, and determine the pod of the job\n\tpods, chiefPod := getPodsOfTFJob(name, tt, tfjob, allPods)\n\n\treturn &TensorFlowJob{\n\t\tBasicJobInfo: &BasicJobInfo{\n\t\t\tresources: podResources(pods),\n\t\t\tname: name,\n\t\t},\n\t\ttfjob: tfjob,\n\t\tchiefPod: chiefPod,\n\t\tpods: pods,\n\t\ttrainerType: tt.Type(),\n\t}, nil\n}", "func (d Dispatcher) JobName(id string) (string, error) {\n\tj, err := d.GetBC().FindJob(id)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn j.GetName(), nil\n}", "func (cli *Client) GetTranscodingJob(jobId string) (*api.GetTranscodingJobResponse, error) {\n\treturn api.GetTranscodingJob(cli, jobId)\n}", "func (o *Job) GetName() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.Name\n}", "func GetCronJobs(contextName string, namespace string) []batchv1beta1.CronJob {\n\n\tctxReceiver, ok := contextReceivers[contextName]\n\tif !ok {\n\t\treturn nil\n\t}\n\n\tnsReceiver, ok := ctxReceiver.namespaceReceivers[namespace]\n\tif !ok {\n\t\treturn nil\n\t}\n\n\treceiver := nsReceiver.cronJobEventReceiver\n\tif receiver == nil {\n\t\treturn nil\n\t}\n\n\treturn receiver.getCronJobs()\n}", "func GetJob(process string, pid string, tempLocation string) ([]string, error) {\n\treturn []string{\"Minidump\", process, pid, tempLocation}, nil\n}", "func (c *JobClient) GetX(ctx context.Context, id int) *Job {\n\tobj, err := c.Get(ctx, id)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn obj\n}", "func (r *ProjectsLocationsMigrationJobsService) Get(name string) *ProjectsLocationsMigrationJobsGetCall {\n\tc := &ProjectsLocationsMigrationJobsGetCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.name = name\n\treturn c\n}", "func (c *Controller) GetJob(jobID string) (models.JobStats, error) {\n\tif utils.IsEmptyStr(jobID) {\n\t\treturn models.JobStats{}, errors.New(\"empty job ID\")\n\t}\n\n\treturn c.backendPool.GetJobStats(jobID)\n}", "func Get(name string) (Driver, error) {\n\tif d, ok := schedulers[name]; ok {\n\t\treturn d, nil\n\t}\n\treturn nil, &errors.ErrNotFound{\n\t\tID: name,\n\t\tType: \"Scheduler\",\n\t}\n}", "func (j *JobCacheClient) GetJobCache(job string, dedup bool) (*JobCache, error) {\n\tlogrus.Infof(\"Getting %s build results from GCS\", job)\n\n\ttempJSON := j.gcpClient.CopyJobCache(job)\n\tif tempJSON == \"\" {\n\t\treturn nil, nil\n\t}\n\n\tif !util.Exists(tempJSON) {\n\t\t// If there's no file up on job doesn't exist: Skip it.\n\t\tlogrus.Infof(\"Skipping non existing job: %s\", job)\n\t\treturn nil, nil\n\t}\n\tdefer os.RemoveAll(tempJSON)\n\n\t// Additional select on .version is because we have so many empty versions\n\t// for now 2 passes. First pass sorts by buildnumber, second builds the\n\t// dictionary.\n\tout, err := command.New(\"jq\", \"-r\",\n\t\t`.[] | `+\n\t\t\t`select(.result == \"SUCCESS\") | `+\n\t\t\t`select(.version != null) | `+\n\t\t\t`[.version,.buildnumber] | \"\\(.[0]|rtrimstr(\"\\n\")) \\(.[1])\"`,\n\t\ttempJSON,\n\t).Pipe(\"sort\", \"-rn\", \"-k2,2\").RunSilentSuccessOutput()\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"filtering job cache\")\n\t}\n\n\tlastVersion := \"\"\n\tres := &JobCache{Name: job}\n\tscanner := bufio.NewScanner(strings.NewReader(out.OutputTrimNL()))\n\tfor scanner.Scan() {\n\t\tsplit := strings.Split(scanner.Text(), \" \")\n\t\tif len(split) != 2 {\n\t\t\treturn nil, errors.Wrapf(err,\n\t\t\t\t\"unexpected string in job results cache %s: %s\",\n\t\t\t\ttempJSON, scanner.Text(),\n\t\t\t)\n\t\t}\n\n\t\tversion := split[0]\n\t\tbuildNumber := split[1]\n\n\t\tif dedup && version == lastVersion {\n\t\t\tcontinue\n\t\t}\n\t\tlastVersion = version\n\n\t\tif buildNumber != \"\" && version != \"\" {\n\t\t\tres.BuildNumbers = append(res.BuildNumbers, buildNumber)\n\t\t\tres.Versions = append(res.Versions, version)\n\t\t}\n\t}\n\n\treturn res, nil\n}", "func (r JobResource) GetName() string {\n\treturn \"jobs\"\n}", "func (r JobResource) GetName() string {\n\treturn \"jobs\"\n}", "func (m *DataHistoryManager) GetByID(id uuid.UUID) (*DataHistoryJob, error) {\n\tif m == nil {\n\t\treturn nil, ErrNilSubsystem\n\t}\n\tif atomic.LoadInt32(&m.started) == 0 {\n\t\treturn nil, ErrSubSystemNotStarted\n\t}\n\tif id == uuid.Nil {\n\t\treturn nil, errEmptyID\n\t}\n\tm.m.Lock()\n\tfor i := range m.jobs {\n\t\tif m.jobs[i].ID == id {\n\t\t\tcpy := *m.jobs[i]\n\t\t\tm.m.Unlock()\n\t\t\treturn &cpy, nil\n\t\t}\n\t}\n\tm.m.Unlock()\n\tdbJ, err := m.jobDB.GetByID(id.String())\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"%w with id %s %s\", errJobNotFound, id, err)\n\t}\n\tresult, err := m.convertDBModelToJob(dbJ)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"could not convert model with id %s %w\", id, err)\n\t}\n\treturn result, nil\n}", "func GetJobsByCronjob(client kubernetes.Interface, options GetJobOptions) (*v1.JobList, error) {\n\n\t//sel := strings.Join([]string{\"items.metadata.ownerReferences.name\", options.Name}, \"=\")\n\t//log.Printf(\"sel: %v\", sel)\n\t//resource, err := client.BatchV1().Jobs(options.Namespace).List(meta_v1.ListOptions{LabelSelector: sel})\n\n\tresource, err := client.BatchV1().Jobs(options.Namespace).List(meta_v1.ListOptions{})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn resource, nil\n}", "func (c *Check) Job() string {\n\tjob, ok := (*c)[\"job\"]\n\tif !ok {\n\t\treturn \"\"\n\t}\n\treturn job.(string)\n}", "func (self JobController) Get(allPods []v1.Pod, allEvents []v1.Event) ResourceOwner {\n\tmatchingPods := common.FilterPodsForJob(batch.Job(self), allPods)\n\tpodInfo := common.GetPodInfo(self.Status.Active, self.Spec.Completions, matchingPods)\n\tpodInfo.Warnings = event.GetPodsEventWarnings(allEvents, matchingPods)\n\n\treturn ResourceOwner{\n\t\tTypeMeta: api.NewTypeMeta(api.ResourceKindJob),\n\t\tObjectMeta: api.NewObjectMeta(self.ObjectMeta),\n\t\tPods: podInfo,\n\t\tContainerImages: common.GetContainerImages(&self.Spec.Template.Spec),\n\t\tInitContainerImages: common.GetInitContainerImages(&self.Spec.Template.Spec),\n\t}\n}", "func GetJobFromQueue() (*QueueResponse, error) {\n\tclient := &http.Client{}\n\tresponseModel := QueueResponse{}\n\n\tres, err := client.Get(Host + \"/queues/\" + Queue + \"/next\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\trespBody := res.Body\n\tdefer respBody.Close()\n\n\tif err := json.NewDecoder(respBody).Decode(&responseModel); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &responseModel, nil\n}", "func FindJob(nomad *NomadServer, name string) (*Job, error) {\n\tjobs, _, _ := Jobs(nomad)\n\tfor _, job := range jobs {\n\t\tif job.Name == name {\n\t\t\treturn &job, nil\n\t\t}\n\t}\n\treturn &Job{}, errors.New(\"job not found\")\n}", "func (d Dispatcher) Job(id string) (string, error) {\n\tj, err := d.GetBC().FindJob(id)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tjBytes, err := helpers.Serialize(j)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn string(jBytes), nil\n}", "func (m *DataHistoryManager) GetByID(id uuid.UUID) (*DataHistoryJob, error) {\n\tif m == nil {\n\t\treturn nil, ErrNilSubsystem\n\t}\n\tif atomic.LoadInt32(&m.started) == 0 {\n\t\treturn nil, ErrSubSystemNotStarted\n\t}\n\tif id == uuid.Nil {\n\t\treturn nil, errEmptyID\n\t}\n\tdbJ, err := m.jobDB.GetByID(id.String())\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"%w with id %s %s\", errJobNotFound, id, err)\n\t}\n\tresult, err := m.convertDBModelToJob(dbJ)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"could not convert model with id %s %w\", id, err)\n\t}\n\treturn result, nil\n}", "func GetJobFromRadixJob(job *v1.RadixJob, jobDeployments []*deploymentModels.DeploymentSummary) *Job {\n\tsteps := GetJobStepsFromRadixJob(job)\n\n\tcreated := radixutils.FormatTime(&job.CreationTimestamp)\n\tif job.Status.Created != nil {\n\t\t// Use this instead, because in a migration this may be more correct\n\t\t// as migrated jobs will have the same creation timestamp in the new cluster\n\t\tcreated = radixutils.FormatTime(job.Status.Created)\n\t}\n\n\tvar jobComponents []*deploymentModels.ComponentSummary\n\tif len(jobDeployments) > 0 {\n\t\tjobComponents = jobDeployments[0].Components\n\t}\n\n\tjobModel := Job{\n\t\tName: job.GetName(),\n\t\tBranch: job.Spec.Build.Branch,\n\t\tCommitID: job.Spec.Build.CommitID,\n\t\tCreated: created,\n\t\tStarted: radixutils.FormatTime(job.Status.Started),\n\t\tEnded: radixutils.FormatTime(job.Status.Ended),\n\t\tStatus: GetStatusFromRadixJobStatus(job.Status, job.Spec.Stop),\n\t\tPipeline: string(job.Spec.PipeLineType),\n\t\tSteps: steps,\n\t\tDeployments: jobDeployments,\n\t\tComponents: jobComponents,\n\t\tTriggeredBy: job.Spec.TriggeredBy,\n\t}\n\tif job.Spec.PipeLineType == v1.Promote {\n\t\tjobModel.PromotedFromEnvironment = job.Spec.Promote.FromEnvironment\n\t\tjobModel.PromotedToEnvironment = job.Spec.Promote.ToEnvironment\n\t\tjobModel.PromotedDeploymentName = job.Spec.Promote.DeploymentName\n\t}\n\n\treturn &jobModel\n}", "func (c *TestController) lookupJob(id TestID) *control.Job {\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\treturn c.running[id]\n}", "func NewGetJobOK() *GetJobOK {\n\treturn &GetJobOK{}\n}", "func NewGetJobNotFound() *GetJobNotFound {\n\treturn &GetJobNotFound{}\n}", "func (j *Job) Job() JobFunc {\n\treturn j.job\n}", "func GetMigrationJob(client lib.Client, id int) (mj brain.MigrationJob, err error) {\n\tr, err := client.BuildRequest(\"GET\", lib.BrainEndpoint, \"/admin/migration_jobs/%s\", strconv.Itoa(id))\n\tif err != nil {\n\t\treturn\n\t}\n\n\t_, _, err = r.Run(nil, &mj)\n\treturn\n}", "func (m *SynchronizationJob) GetSchedule()(SynchronizationScheduleable) {\n val, err := m.GetBackingStore().Get(\"schedule\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(SynchronizationScheduleable)\n }\n return nil\n}", "func (s *DataStore) GetRecurringJob(name string) (*longhorn.RecurringJob, error) {\n\tresult, err := s.getRecurringJob(name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.DeepCopy(), nil\n}", "func (r *DeviceManagementExportJobRequest) Get(ctx context.Context) (resObj *DeviceManagementExportJob, err error) {\n\tvar query string\n\tif r.query != nil {\n\t\tquery = \"?\" + r.query.Encode()\n\t}\n\terr = r.JSONRequest(ctx, \"GET\", query, nil, &resObj)\n\treturn\n}", "func GetDataLoadJob(client client.Client, name, namespace string) (*batchv1.Job, error) {\n\tkey := types.NamespacedName{\n\t\tNamespace: namespace,\n\t\tName: name,\n\t}\n\tvar job batchv1.Job\n\tif err := client.Get(context.TODO(), key, &job); err != nil {\n\t\treturn nil, err\n\t}\n\treturn &job, nil\n}", "func (o DatasetImportJobOutput) JobName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v DatasetImportJob) *string { return v.JobName }).(pulumi.StringPtrOutput)\n}", "func (db *PgDB) JobByID(jID model.JobID) (*model.Job, error) {\n\tvar j model.Job\n\tif err := db.query(`\nSELECT *\nFROM jobs\nWHERE job_id = $1\n`, &j, jID); err != nil {\n\t\treturn nil, errors.Wrap(err, \"querying job\")\n\t}\n\treturn &j, nil\n}", "func (j *JobWorker) GetJobStatus(id string) (Job, error) {\n\tif job, ok := j.load(id); ok {\n\t\t// make a shallow copy of the job object\n\t\tjobCopy := *job\n\t\tjobCopy.Output = jobCopy.outputBuffer.String()\n\t\treturn jobCopy, nil\n\t} else {\n\t\treturn Job{}, fmt.Errorf(\"job %v: not found\", id)\n\t}\n}" ]
[ "0.72829413", "0.7240856", "0.6836308", "0.6597606", "0.6550326", "0.653562", "0.64546186", "0.64341086", "0.6410725", "0.6383163", "0.6351446", "0.6336248", "0.632086", "0.6307784", "0.6307345", "0.63056725", "0.6291917", "0.6289772", "0.62587214", "0.6249659", "0.6248043", "0.6234525", "0.6222283", "0.62113", "0.6174809", "0.61600006", "0.61427444", "0.61354816", "0.61219525", "0.6118036", "0.6116823", "0.60992235", "0.60941386", "0.60903025", "0.60654247", "0.6030773", "0.6010049", "0.60065335", "0.59774345", "0.5976747", "0.5936808", "0.5929819", "0.59267145", "0.5899833", "0.5880881", "0.587582", "0.5821414", "0.5809666", "0.5809065", "0.58014786", "0.57961285", "0.5775661", "0.5756959", "0.57515085", "0.5740062", "0.5730969", "0.56913537", "0.56727064", "0.5667679", "0.5607048", "0.55948097", "0.5590716", "0.55905443", "0.5586069", "0.55859804", "0.55613", "0.55606925", "0.554822", "0.55324095", "0.5525946", "0.55218", "0.5514918", "0.54715455", "0.546753", "0.5444807", "0.5444788", "0.5422586", "0.54038644", "0.54038644", "0.5392199", "0.53886235", "0.5379136", "0.5371935", "0.53664637", "0.5360956", "0.535243", "0.5341577", "0.5337654", "0.5336421", "0.52918833", "0.5265591", "0.52463675", "0.52435744", "0.5224277", "0.5221262", "0.5214405", "0.5212291", "0.52089965", "0.52002907", "0.5198968" ]
0.6742163
3
List takes label and field selectors, and returns the list of TZCronJobs that match those selectors.
func (c *FakeTZCronJobs) List(opts v1.ListOptions) (result *v1alpha1.TZCronJobList, err error) { obj, err := c.Fake. Invokes(testing.NewListAction(tzcronjobsResource, tzcronjobsKind, c.ns, opts), &v1alpha1.TZCronJobList{}) if obj == nil { return nil, err } label, _, _ := testing.ExtractFromListOptions(opts) if label == nil { label = labels.Everything() } list := &v1alpha1.TZCronJobList{ListMeta: obj.(*v1alpha1.TZCronJobList).ListMeta} for _, item := range obj.(*v1alpha1.TZCronJobList).Items { if label.Matches(labels.Set(item.Labels)) { list.Items = append(list.Items, item) } } return list, err }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func List(ctx context.Context, c client.Client, namespace string, selector map[string]string) ([]batchv1.CronJob, error) {\n\tlist := &batchv1.CronJobList{}\n\topts := []client.ListOption{\n\t\tclient.InNamespace(namespace),\n\t\tclient.MatchingLabels(selector),\n\t}\n\tif err := c.List(ctx, list, opts...); err != nil {\n\t\treturn nil, kverrors.Wrap(err, \"failed to list cronjobs\",\n\t\t\t\"namespace\", namespace,\n\t\t)\n\t}\n\n\treturn list.Items, nil\n}", "func (c *scheduledJobs) List(opts api.ListOptions) (result *batch.ScheduledJobList, err error) {\n\tresult = &batch.ScheduledJobList{}\n\terr = c.r.Get().Namespace(c.ns).Resource(\"scheduledjobs\").VersionedParams(&opts, api.ParameterCodec).Do().Into(result)\n\treturn\n}", "func ListJobs(cmd *cobra.Command, args []string) error {\n\n\tclient, err := auth.GetClient()\n\tif err != nil {\n\t\treturn err\n\t}\n\t// Parse all flags\n\n\tvar countDefault int32\n\tcount := &countDefault\n\terr = flags.ParseFlag(cmd.Flags(), \"count\", &count)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"count\": ` + err.Error())\n\t}\n\tvar filter string\n\terr = flags.ParseFlag(cmd.Flags(), \"filter\", &filter)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"filter\": ` + err.Error())\n\t}\n\tvar status model.SearchStatus\n\terr = flags.ParseFlag(cmd.Flags(), \"status\", &status)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"status\": ` + err.Error())\n\t}\n\t// Form query params\n\tgenerated_query := model.ListJobsQueryParams{}\n\tgenerated_query.Count = count\n\tgenerated_query.Filter = filter\n\tgenerated_query.Status = status\n\n\t// Silence Usage\n\tcmd.SilenceUsage = true\n\n\tresp, err := client.SearchService.ListJobs(&generated_query)\n\tif err != nil {\n\t\treturn err\n\t}\n\tjsonx.Pprint(cmd, resp)\n\treturn nil\n}", "func List(ctx context.Context) (err error) {\n\tif t.Status == constants.TaskStatusCreated {\n\t\t_, err = model.CreateJob(ctx, \"/\")\n\t\tif err != nil {\n\t\t\tlogrus.Panic(err)\n\t\t}\n\n\t\tt.Status = constants.TaskStatusRunning\n\t\terr = t.Save(ctx)\n\t\tif err != nil {\n\t\t\tlogrus.Panic(err)\n\t\t}\n\t}\n\n\t// Traverse already running but not finished object.\n\tp := \"\"\n\tfor {\n\t\to, err := model.NextObject(ctx, p)\n\t\tif err != nil {\n\t\t\tlogrus.Panic(err)\n\t\t}\n\t\tif o == nil {\n\t\t\tbreak\n\t\t}\n\n\t\toc <- o\n\t\tp = o.Key\n\t}\n\n\t// Traverse already running but not finished job.\n\tp = \"\"\n\tfor {\n\t\tj, err := model.NextJob(ctx, p)\n\t\tif err != nil {\n\t\t\tlogrus.Panic(err)\n\t\t}\n\t\tif j == nil {\n\t\t\tbreak\n\t\t}\n\n\t\tjwg.Add(1)\n\t\tjc <- j\n\t\tp = j.Path\n\t}\n\n\treturn\n}", "func listJobs(c *config) []string {\n\turl := fmt.Sprintf(jobsURL, c.Jenkins.Url)\n\tcode, b := jenkinsGet(url, c.Jenkins.User, c.Jenkins.Password, c.Jenkins.Verify)\n\n\tif code != 200 {\n\t\tlog.Fatalf(\"List jobs: response code: %d\", code)\n\t}\n\n\tjobs := Jobs{}\n\tres := []string{}\n\n\te := json.Unmarshal(b, &jobs)\n\tlogFatal(\"List jobs: json\", e)\n\n\tfor _, j := range jobs.Jobs {\n\t\tres = append(res, j.Name)\n\t}\n\n\treturn res\n}", "func (dt *Tracker) ListJobs() ([]string, error) {\n\treturn dt.processTracker.ListJobs()\n}", "func (c *aITrainingJobs) List(opts metav1.ListOptions) (result *v1.AITrainingJobList, err error) {\n\tvar timeout time.Duration\n\tif opts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*opts.TimeoutSeconds) * time.Second\n\t}\n\tresult = &v1.AITrainingJobList{}\n\terr = c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"aitrainingjobs\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tDo().\n\t\tInto(result)\n\treturn\n}", "func (s *elasticDLJobLister) List(selector labels.Selector) (ret []*v1alpha1.ElasticDLJob, err error) {\n\terr = cache.ListAll(s.indexer, selector, func(m interface{}) {\n\t\tret = append(ret, m.(*v1alpha1.ElasticDLJob))\n\t})\n\treturn ret, err\n}", "func (taskBolt *TaskBolt) ListJobs(ctx context.Context, in *ga4gh_task_exec.JobListRequest) (*ga4gh_task_exec.JobListResponse, error) {\n\tlog.Debug(\"ListJobs called\")\n\n\tjobs := make([]*ga4gh_task_exec.JobDesc, 0, 10)\n\n\ttaskBolt.db.View(func(tx *bolt.Tx) error {\n\t\ttaskopB := tx.Bucket(TaskBucket)\n\t\tc := taskopB.Cursor()\n\n\t\tfor k, v := c.First(); k != nil; k, v = c.Next() {\n\t\t\tjobID := string(k)\n\t\t\tjobState := getJobState(tx, jobID)\n\n\t\t\ttask := &ga4gh_task_exec.Task{}\n\t\t\tproto.Unmarshal(v, task)\n\n\t\t\tjob := &ga4gh_task_exec.JobDesc{\n\t\t\t\tJobID: jobID,\n\t\t\t\tState: jobState,\n\t\t\t\tTask: &ga4gh_task_exec.TaskDesc{\n\t\t\t\t\tName: task.Name,\n\t\t\t\t\tProjectID: task.ProjectID,\n\t\t\t\t\tDescription: task.Description,\n\t\t\t\t},\n\t\t\t}\n\t\t\tjobs = append(jobs, job)\n\t\t}\n\t\treturn nil\n\t})\n\n\tout := ga4gh_task_exec.JobListResponse{\n\t\tJobs: jobs,\n\t}\n\n\treturn &out, nil\n}", "func GetJobsByCronjob(client kubernetes.Interface, options GetJobOptions) (*v1.JobList, error) {\n\n\t//sel := strings.Join([]string{\"items.metadata.ownerReferences.name\", options.Name}, \"=\")\n\t//log.Printf(\"sel: %v\", sel)\n\t//resource, err := client.BatchV1().Jobs(options.Namespace).List(meta_v1.ListOptions{LabelSelector: sel})\n\n\tresource, err := client.BatchV1().Jobs(options.Namespace).List(meta_v1.ListOptions{})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn resource, nil\n}", "func (dtm *DfgetTaskManager) List(ctx context.Context, filter map[string]string) (dfgetTaskList []*types.DfGetTask, err error) {\n\treturn nil, nil\n}", "func (cli *Client) CronList(ctx context.Context, opts types.CronListOptions) ([]types.Cron, error) {\n\tvar crons = []types.Cron{}\n\tquery := url.Values{}\n\n\tif opts.Filters.Len() > 0 {\n\t\tfilterJSON, err := filters.ToParamWithVersion(cli.version, opts.Filters)\n\t\tif err != nil {\n\t\t\treturn crons, err\n\t\t}\n\t\tquery.Set(\"filters\", filterJSON)\n\t}\n\tresp, err := cli.get(ctx, \"/crons\", query, nil)\n\tif err != nil {\n\t\treturn crons, err\n\t}\n\n\terr = json.NewDecoder(resp.body).Decode(&crons)\n\tensureReaderClosed(resp)\n\treturn crons, err\n}", "func (r *TransferJobsService) List() *TransferJobsListCall {\n\tc := &TransferJobsListCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\treturn c\n}", "func (j *StorageJobs) List(ctx context.Context, config ListConfig) (*adminPb.ListStorageJobsResponse, error) {\n\tsel := userPb.StorageJobsSelector_STORAGE_JOBS_SELECTOR_UNSPECIFIED\n\tswitch config.Select {\n\tcase All:\n\t\tsel = userPb.StorageJobsSelector_STORAGE_JOBS_SELECTOR_ALL\n\tcase Queued:\n\t\tsel = userPb.StorageJobsSelector_STORAGE_JOBS_SELECTOR_QUEUED\n\tcase Executing:\n\t\tsel = userPb.StorageJobsSelector_STORAGE_JOBS_SELECTOR_EXECUTING\n\tcase Final:\n\t\tsel = userPb.StorageJobsSelector_STORAGE_JOBS_SELECTOR_FINAL\n\t}\n\treq := &adminPb.ListStorageJobsRequest{\n\t\tCidFilter: config.CidFilter,\n\t\tLimit: config.Limit,\n\t\tAscending: config.Ascending,\n\t\tNextPageToken: config.NextPageToken,\n\t\tSelector: sel,\n\t}\n\treturn j.client.ListStorageJobs(ctx, req)\n}", "func listWorker(ctx context.Context) {\n\tdefer utils.Recover()\n\n\tfor j := range jc {\n\t\tlogrus.Infof(\"Start listing job %s.\", j.Path)\n\n\t\terr := listJob(ctx, j)\n\t\tif err != nil {\n\t\t\tcontinue\n\t\t}\n\n\t\tlogrus.Infof(\"Job %s listed.\", j.Path)\n\t}\n}", "func (r *ProjectsPatchJobsService) List(parent string) *ProjectsPatchJobsListCall {\n\tc := &ProjectsPatchJobsListCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.parent = parent\n\treturn c\n}", "func (c *ci) ListJobs(refreshResult, buildResult bool) *format.ListResponse {\n\n\tjs := make([]*format.Job, 0, len(c.jobs))\n\tfor _, j := range c.jobs {\n\t\tjs = append(js, j.Status(refreshResult, buildResult))\n\t}\n\n\treturn &format.ListResponse{\n\t\tJobs: js,\n\t}\n}", "func (c *ci) ListJobs(refreshResult, buildResult bool) *format.ListResponse {\n\n\t//make and fill the slice\n\tjs := make([]*format.Job, 0, len(c.jobs))\n\tfor _, j := range c.jobs {\n\t\tjs = append(js, j.Status(refreshResult, buildResult))\n\t}\n\n\treturn &format.ListResponse{\n\t\tJobs: js,\n\t}\n}", "func ListJobs(cmd CmdInterface) {\n\tqueue, err := store.FindQueueBySlug(cmd.Parts[1], cmd.User.CurrentGroup, false)\n\tif err != nil {\n\t\tReturnError(cmd, err.Error())\n\t\treturn\n\t}\n\n\trawJSON, _ := json.Marshal(queue.Jobs)\n\tReturnString(cmd, string(rawJSON))\n}", "func List(ctx *cli.Context) error {\n\targs := &listArgs{}\n\targs.Parse(ctx)\n\n\tmanager := yata.NewTaskManager()\n\ttasks, err := manager.GetAll()\n\thandleError(err)\n\n\tif args.showTags {\n\t\treturn displayTags(tasks)\n\t}\n\n\ttasks = yata.FilterTasks(tasks, func(t yata.Task) bool {\n\t\treturn (args.tag == \"\" || sliceContains(t.Tags, args.tag)) &&\n\t\t\t(args.description == \"\" || strings.Contains(t.Description, args.description)) &&\n\t\t\t(args.all || !t.Completed)\n\t})\n\n\tsortTasks(args.sort, &tasks)\n\n\tfor _, v := range tasks {\n\t\tstringer := yata.NewTaskStringer(v, taskStringer(args.format))\n\t\tswitch v.Priority {\n\t\tcase yata.LowPriority:\n\t\t\tyata.PrintlnColor(\"cyan+h\", stringer.String())\n\t\tcase yata.HighPriority:\n\t\t\tyata.PrintlnColor(\"red+h\", stringer.String())\n\t\tdefault:\n\t\t\tyata.Println(stringer.String())\n\t\t}\n\t}\n\n\treturn nil\n}", "func (k *KubernetesScheduler) ListPods(selector labels.Selector) (*api.PodList, error) {\n\tlog.V(2).Infof(\"List pods for '%v'\\n\", selector)\n\n\tk.RLock()\n\tdefer k.RUnlock()\n\n\tvar result []api.Pod\n\tfor _, task := range k.runningTasks {\n\t\tpod := task.Pod\n\n\t\tvar l labels.Set = pod.Labels\n\t\tif selector.Matches(l) || selector.Empty() {\n\t\t\tresult = append(result, *pod)\n\t\t}\n\t}\n\n\t// TODO(nnielsen): Refactor tasks append for the three lists.\n\tfor _, task := range k.pendingTasks {\n\t\tpod := task.Pod\n\n\t\tvar l labels.Set = pod.Labels\n\t\tif selector.Matches(l) || selector.Empty() {\n\t\t\tresult = append(result, *pod)\n\t\t}\n\t}\n\n\t// TODO(nnielsen): Wire up check in finished tasks\n\n\tmatches := &api.PodList{Items: result}\n\tlog.V(2).Infof(\"Returning pods: '%v'\\n\", matches)\n\n\treturn matches, nil\n}", "func (c *Controller) ListJobs(filter weles.JobFilter, sorter weles.JobSorter,\n\tpaginator weles.JobPagination) ([]weles.JobInfo, weles.ListInfo, error) {\n\treturn c.jobs.List(filter, sorter, paginator)\n}", "func (c *FakeImagePullJobs) List(ctx context.Context, opts v1.ListOptions) (result *v1alpha1.ImagePullJobList, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewListAction(imagepulljobsResource, imagepulljobsKind, c.ns, opts), &v1alpha1.ImagePullJobList{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\n\tlabel, _, _ := testing.ExtractFromListOptions(opts)\n\tif label == nil {\n\t\tlabel = labels.Everything()\n\t}\n\tlist := &v1alpha1.ImagePullJobList{ListMeta: obj.(*v1alpha1.ImagePullJobList).ListMeta}\n\tfor _, item := range obj.(*v1alpha1.ImagePullJobList).Items {\n\t\tif label.Matches(labels.Set(item.Labels)) {\n\t\t\tlist.Items = append(list.Items, item)\n\t\t}\n\t}\n\treturn list, err\n}", "func (s *ContinuousRestoreJobsServiceOp) List(ctx context.Context, groupID, clusterID string, opts *ListOptions) (*atlas.ContinuousJobs, *Response, error) {\n\tif clusterID == \"\" {\n\t\treturn nil, nil, atlas.NewArgError(\"clusterID\", \"must be set\")\n\t}\n\tif groupID == \"\" {\n\t\treturn nil, nil, atlas.NewArgError(\"groupID\", \"must be set\")\n\t}\n\n\tpath := fmt.Sprintf(continuousRestoreJobsPath, groupID, clusterID)\n\n\tpath, err := setQueryParams(path, opts)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treq, err := s.Client.NewRequest(ctx, http.MethodGet, path, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\troot := new(atlas.ContinuousJobs)\n\tresp, err := s.Client.Do(ctx, req, root)\n\n\treturn root, resp, err\n}", "func (t *DRMAATracker) ListJobs() ([]string, error) {\n\t// need to get the job list from the internal DB\n\tt.Lock()\n\tdefer t.Unlock()\n\treturn t.store.GetJobIDs(), nil\n}", "func (c *ClientGoUtils) ListPodsByJob(name string) (*corev1.PodList, error) {\n\tjob, err := c.ClientSet.BatchV1().Jobs(c.namespace).Get(c.ctx, name, metav1.GetOptions{})\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"\")\n\t}\n\tset := labels.Set(job.Spec.Selector.MatchLabels)\n\tpods, err := c.ClientSet.CoreV1().Pods(c.namespace).List(\n\t\tc.ctx, metav1.ListOptions{LabelSelector: set.AsSelector().String()},\n\t)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"\")\n\t}\n\treturn pods, nil\n}", "func (r *ProjectsLocationsJobsService) List(parent string) *ProjectsLocationsJobsListCall {\n\tc := &ProjectsLocationsJobsListCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.parent = parent\n\treturn c\n}", "func (r *JobsService) List() *JobsListCall {\n\tc := &JobsListCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\treturn c\n}", "func (a *apiServer) ListJob(request *pps.ListJobRequest, resp pps.API_ListJobServer) (retErr error) {\n\tfilterJob, err := newMessageFilterFunc(request.GetJqFilter(), request.GetProjects())\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"error creating message filter function\")\n\t}\n\n\tctx := resp.Context()\n\tpipeline := request.GetPipeline()\n\tif pipeline != nil {\n\t\tensurePipelineProject(pipeline)\n\t\t// If 'pipeline is set, check that caller has access to the pipeline's\n\t\t// output repo; currently, that's all that's required for ListJob.\n\t\t//\n\t\t// If 'pipeline' isn't set, then we don't return an error (otherwise, a\n\t\t// caller without access to a single pipeline's output repo couldn't run\n\t\t// `pachctl list job` at all) and instead silently skip jobs where the user\n\t\t// doesn't have access to the job's output repo.\n\t\tif err := a.env.AuthServer.CheckRepoIsAuthorized(ctx, &pfs.Repo{Type: pfs.UserRepoType, Project: pipeline.Project, Name: pipeline.Name}, auth.Permission_PIPELINE_LIST_JOB); err != nil && !auth.IsErrNotActivated(err) {\n\t\t\treturn errors.EnsureStack(err)\n\t\t}\n\t}\n\n\tnumber := request.Number\n\t// If number is not set, return all jobs that match the query\n\tif number == 0 {\n\t\tnumber = math.MaxInt64\n\t}\n\t// pipelineVersions holds the versions of pipelines that we're interested in\n\tpipelineVersions := make(map[string]bool)\n\tif err := ppsutil.ListPipelineInfo(ctx, a.pipelines, pipeline, request.GetHistory(),\n\t\tfunc(ptr *pps.PipelineInfo) error {\n\t\t\tpipelineVersions[ppsdb.VersionKey(ptr.Pipeline, ptr.Version)] = true\n\t\t\treturn nil\n\t\t}); err != nil {\n\t\treturn err\n\t}\n\n\tjobs := a.jobs.ReadOnly(ctx)\n\tjobInfo := &pps.JobInfo{}\n\t_f := func(string) error {\n\t\tif number == 0 {\n\t\t\treturn errutil.ErrBreak\n\t\t}\n\t\tif request.PaginationMarker != nil {\n\t\t\tcreatedAt := time.Unix(int64(jobInfo.Created.GetSeconds()), int64(jobInfo.Created.GetNanos())).UTC()\n\t\t\tfromTime := time.Unix(int64(request.PaginationMarker.GetSeconds()), int64(request.PaginationMarker.GetNanos())).UTC()\n\t\t\tif createdAt.Equal(fromTime) || !request.Reverse && createdAt.After(fromTime) || request.Reverse && createdAt.Before(fromTime) {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t\tif request.GetDetails() {\n\t\t\tif err := a.getJobDetails(ctx, jobInfo); err != nil {\n\t\t\t\tif auth.IsErrNotAuthorized(err) {\n\t\t\t\t\treturn nil // skip job--see note at top of function\n\t\t\t\t}\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t\tif len(request.GetInputCommit()) > 0 {\n\t\t\t// Only include the job if it's in the set of intersected commitset IDs\n\t\t\tcommitsets, err := a.intersectCommitSets(ctx, request.GetInputCommit())\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tif _, ok := commitsets[jobInfo.Job.Id]; !ok {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t\tif !pipelineVersions[ppsdb.VersionKey(jobInfo.Job.Pipeline, jobInfo.PipelineVersion)] {\n\t\t\treturn nil\n\t\t}\n\n\t\tif ok, err := filterJob(ctx, jobInfo); err != nil {\n\t\t\treturn errors.Wrap(err, \"error filtering job\")\n\t\t} else if !ok {\n\t\t\treturn nil\n\t\t}\n\n\t\t// Erase any AuthToken - this shouldn't be returned to anyone (the workers\n\t\t// won't use this function to get their auth token)\n\t\tjobInfo.AuthToken = \"\"\n\n\t\tif err := resp.Send(jobInfo); err != nil {\n\t\t\treturn errors.Wrap(err, \"error sending job\")\n\t\t}\n\t\tnumber--\n\t\treturn nil\n\t}\n\topts := &col.Options{Target: col.SortByCreateRevision, Order: col.SortDescend}\n\tif request.Reverse {\n\t\topts.Order = col.SortAscend\n\t}\n\tif pipeline != nil {\n\t\terr = jobs.GetByIndex(ppsdb.JobsPipelineIndex, ppsdb.JobsPipelineKey(pipeline), jobInfo, opts, _f)\n\t} else {\n\t\terr = jobs.List(jobInfo, opts, _f)\n\t}\n\tif err != nil && err != errutil.ErrBreak {\n\t\tif errors.Is(err, context.DeadlineExceeded) {\n\t\t\treturn status.Error(codes.DeadlineExceeded, err.Error())\n\t\t}\n\t\treturn errors.EnsureStack(err)\n\t}\n\treturn nil\n}", "func JobListHandler(c *Context, w http.ResponseWriter, r *http.Request) {\n\taccount, err := Authenticate(c, w, r)\n\tif err != nil {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"error\": err,\n\t\t}).Error(\"Authentication failure.\")\n\t\treturn\n\t}\n\n\tif err := r.ParseForm(); err != nil {\n\t\tRhoError{\n\t\t\tCode: CodeUnableToParseQuery,\n\t\t\tMessage: fmt.Sprintf(\"Unable to parse query parameters: %v\", err),\n\t\t\tHint: \"You broke Go's URL parsing somehow! Make URLs that suck less.\",\n\t\t\tRetry: false,\n\t\t}.Log(account).Report(http.StatusBadRequest, w)\n\t\treturn\n\t}\n\n\tq := JobQuery{AccountName: account.Name}\n\tif rawJIDs, ok := r.Form[\"jid\"]; ok {\n\t\tjids := make([]uint64, len(rawJIDs))\n\t\tfor i, rawJID := range rawJIDs {\n\t\t\tif jids[i], err = strconv.ParseUint(rawJID, 10, 64); err != nil {\n\t\t\t\tRhoError{\n\t\t\t\t\tCode: CodeUnableToParseQuery,\n\t\t\t\t\tMessage: fmt.Sprintf(\"Unable to parse JID [%s]: %v\", rawJID, err),\n\t\t\t\t\tHint: \"Please only use valid JIDs.\",\n\t\t\t\t\tRetry: false,\n\t\t\t\t}.Log(account).Report(http.StatusBadRequest, w)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\tq.JIDs = jids\n\t}\n\tif names, ok := r.Form[\"name\"]; ok {\n\t\tq.Names = names\n\t}\n\tif statuses, ok := r.Form[\"status\"]; ok {\n\t\tq.Statuses = statuses\n\t}\n\tif rawLimit := r.FormValue(\"limit\"); rawLimit != \"\" {\n\t\tlimit, err := strconv.ParseInt(rawLimit, 10, 0)\n\t\tif err != nil {\n\t\t\tRhoError{\n\t\t\t\tCode: CodeUnableToParseQuery,\n\t\t\t\tMessage: fmt.Sprintf(\"Unable to parse limit [%s]: %v\", rawLimit, err),\n\t\t\t\tHint: \"Please specify a valid integral limit.\",\n\t\t\t\tRetry: false,\n\t\t\t}.Log(account).Report(http.StatusBadRequest, w)\n\t\t\treturn\n\t\t}\n\n\t\tif limit > 9999 {\n\t\t\tlimit = 9999\n\t\t}\n\t\tif limit < 1 {\n\t\t\tRhoError{\n\t\t\t\tCode: CodeUnableToParseQuery,\n\t\t\t\tMessage: fmt.Sprintf(\"Invalid negative or zero limit [%d]\", limit),\n\t\t\t\tHint: \"Please specify a valid, positive integral limit.\",\n\t\t\t\tRetry: false,\n\t\t\t}.Log(account).Report(http.StatusBadRequest, w)\n\t\t\treturn\n\t\t}\n\t\tq.Limit = int(limit)\n\t} else {\n\t\tq.Limit = 1000\n\t}\n\n\tif rawBefore := r.FormValue(\"before\"); rawBefore != \"\" {\n\t\tbefore, err := strconv.ParseUint(rawBefore, 10, 64)\n\t\tif err != nil {\n\t\t\tRhoError{\n\t\t\t\tCode: CodeUnableToParseQuery,\n\t\t\t\tMessage: fmt.Sprintf(`Unable to parse Before bound [%s]: %v`, rawBefore, err),\n\t\t\t\tHint: \"Please specify a valid integral JID as the lower bound.\",\n\t\t\t\tRetry: false,\n\t\t\t}.Log(account).Report(http.StatusBadRequest, w)\n\t\t\treturn\n\t\t}\n\t\tq.Before = before\n\t}\n\tif rawAfter := r.FormValue(\"after\"); rawAfter != \"\" {\n\t\tafter, err := strconv.ParseUint(rawAfter, 10, 64)\n\t\tif err != nil {\n\t\t\tRhoError{\n\t\t\t\tCode: CodeUnableToParseQuery,\n\t\t\t\tMessage: fmt.Sprintf(`Unable to parse After bound [%s]: %v`, rawAfter, err),\n\t\t\t\tHint: \"Please specify a valid integral JID as the upper bound.\",\n\t\t\t\tRetry: false,\n\t\t\t}.Log(account).Report(http.StatusBadRequest, w)\n\t\t\treturn\n\t\t}\n\t\tq.After = after\n\t}\n\n\tresults, err := c.ListJobs(q)\n\tif err != nil {\n\t\tre := RhoError{\n\t\t\tCode: CodeListFailure,\n\t\t\tMessage: fmt.Sprintf(\"Unable to list jobs: %v\", err),\n\t\t\tHint: \"This is most likely a database problem.\",\n\t\t\tRetry: true,\n\t\t}\n\t\tre.Log(account).Report(http.StatusServiceUnavailable, w)\n\t\treturn\n\t}\n\n\tvar response struct {\n\t\tJobs []SubmittedJob `json:\"jobs\"`\n\t}\n\tresponse.Jobs = results\n\n\tlog.WithFields(log.Fields{\n\t\t\"query\": q,\n\t\t\"result count\": len(results),\n\t\t\"account\": account.Name,\n\t}).Debug(\"Successful job query.\")\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tjson.NewEncoder(w).Encode(response)\n}", "func (s *JobService) List(ctx context.Context, clientTimeOffset int, collectAllChildJobs bool) (*Groups, *http.Response, error) {\n\trequest := Request{\n\t\tAction: JobAction,\n\t\tMethod: \"getGroupInfo\",\n\t\tData: []interface{}{[]interface{}{nil}, clientTimeOffset, collectAllChildJobs},\n\t\tType: \"rpc\",\n\t\tTid: 1,\n\t}\n\n\treq, err := s.client.NewRequest(&request)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tvar groups Groups\n\tr := Response{Data: &groups}\n\tresp, err := s.client.Do(ctx, req, &r)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\treturn &groups, resp, nil\n}", "func (c *FakeTraefikServices) List(ctx context.Context, opts v1.ListOptions) (result *v1alpha1.TraefikServiceList, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewListAction(traefikservicesResource, traefikservicesKind, c.ns, opts), &v1alpha1.TraefikServiceList{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\n\tlabel, _, _ := testing.ExtractFromListOptions(opts)\n\tif label == nil {\n\t\tlabel = labels.Everything()\n\t}\n\tlist := &v1alpha1.TraefikServiceList{ListMeta: obj.(*v1alpha1.TraefikServiceList).ListMeta}\n\tfor _, item := range obj.(*v1alpha1.TraefikServiceList).Items {\n\t\tif label.Matches(labels.Set(item.Labels)) {\n\t\t\tlist.Items = append(list.Items, item)\n\t\t}\n\t}\n\treturn list, err\n}", "func (c *FakeQuobyteServices) List(opts v1.ListOptions) (result *quobyte_com_v1.QuobyteServiceList, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewListAction(quobyteservicesResource, quobyteservicesKind, c.ns, opts), &quobyte_com_v1.QuobyteServiceList{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\n\tlabel, _, _ := testing.ExtractFromListOptions(opts)\n\tif label == nil {\n\t\tlabel = labels.Everything()\n\t}\n\tlist := &quobyte_com_v1.QuobyteServiceList{ListMeta: obj.(*quobyte_com_v1.QuobyteServiceList).ListMeta}\n\tfor _, item := range obj.(*quobyte_com_v1.QuobyteServiceList).Items {\n\t\tif label.Matches(labels.Set(item.Labels)) {\n\t\t\tlist.Items = append(list.Items, item)\n\t\t}\n\t}\n\treturn list, err\n}", "func (table *Timetable) List() []*Task {\n\ttasks := make([]*Task, 0)\n\tfor _, task := range table.schedule {\n\t\ttasks = append(tasks, task)\n\t}\n\treturn tasks\n}", "func listJobs(w io.Writer, projectID string) error {\n\t// projectID := \"my-project-id\"\n\t// jobID := \"my-job-id\"\n\tctx := context.Background()\n\n\tclient, err := bigquery.NewClient(ctx, projectID)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"bigquery.NewClient: %w\", err)\n\t}\n\tdefer client.Close()\n\n\tit := client.Jobs(ctx)\n\t// List up to 10 jobs to demonstrate iteration.\n\tfor i := 0; i < 10; i++ {\n\t\tj, err := it.Next()\n\t\tif err == iterator.Done {\n\t\t\tbreak\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tstate := \"Unknown\"\n\t\tswitch j.LastStatus().State {\n\t\tcase bigquery.Pending:\n\t\t\tstate = \"Pending\"\n\t\tcase bigquery.Running:\n\t\t\tstate = \"Running\"\n\t\tcase bigquery.Done:\n\t\t\tstate = \"Done\"\n\t\t}\n\t\tfmt.Fprintf(w, \"Job %s in state %s\\n\", j.ID(), state)\n\t}\n\treturn nil\n}", "func (c *jxTasks) List(opts v1.ListOptions) (result *v1alpha1.JxTaskList, err error) {\n\tresult = &v1alpha1.JxTaskList{}\n\terr = c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"jxtasks\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tDo().\n\t\tInto(result)\n\treturn\n}", "func (k *Kube) ListJobs(ctx context.Context, in *ListJobsInput) (out *ListJobsOutput, err error) {\n\tif err = k.checkInput(ctx, in); err != nil {\n\t\treturn nil, err\n\t}\n\n\t//List Jobs\n\tjobs := &jobs{}\n\terr = k.visor.ListResources(ctx, kubevisor.ResourceTypeJobs, jobs, nil, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t//Get Events\n\tevents := &events{}\n\terr = k.visor.ListResources(ctx, kubevisor.ResourceTypeEvents, events, nil, []string{\"involvedObject.kind=Job,reason=FailedCreate\"})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t//Get Pods\n\tpods := &pods{}\n\terr = k.visor.ListResources(ctx, kubevisor.ResourceTypePods, pods, nil, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t//get jobs and investigate\n\tout = &ListJobsOutput{}\n\tmapping := map[types.UID]*ListJobItem{}\n\tfor _, job := range jobs.Items {\n\t\tif len(job.Spec.Template.Spec.Containers) != 1 {\n\t\t\tk.logs.Debugf(\"skipping job '%s' in namespace '%s' as it has not just 1 container\", job.Name, job.Namespace)\n\t\t\tcontinue\n\t\t}\n\n\t\tc := job.Spec.Template.Spec.Containers[0]\n\t\titem := &ListJobItem{\n\t\t\tName: job.GetName(),\n\t\t\tImage: c.Image,\n\t\t\tCreatedAt: job.CreationTimestamp.Local(),\n\t\t\tDetails: JobDetails{},\n\t\t}\n\n\t\tif parr := job.Spec.Parallelism; parr != nil {\n\t\t\titem.Details.Parallelism = *parr\n\t\t}\n\n\t\tif dt := job.GetDeletionTimestamp(); dt != nil {\n\t\t\titem.DeletedAt = dt.Local() //mark as deleting\n\t\t}\n\n\t\tif job.Status.StartTime != nil {\n\t\t\titem.ActiveAt = job.Status.StartTime.Local()\n\t\t}\n\n\t\tfor _, dataset := range job.Spec.Template.Spec.Volumes {\n\t\t\tif dataset.FlexVolume != nil {\n\t\t\t\tif dataset.FlexVolume.Options[\"input/dataset\"] != \"\" {\n\t\t\t\t\titem.Input = append(item.Input, dataset.FlexVolume.Options[\"input/dataset\"])\n\t\t\t\t}\n\t\t\t\tif dataset.FlexVolume.Options[\"output/dataset\"] != \"\" {\n\t\t\t\t\titem.Output = append(item.Output, dataset.FlexVolume.Options[\"output/dataset\"])\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\tfor _, cond := range job.Status.Conditions {\n\t\t\tif cond.Status != corev1.ConditionTrue {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tswitch cond.Type {\n\t\t\tcase batchv1.JobComplete:\n\t\t\t\titem.CompletedAt = cond.LastTransitionTime.Local()\n\t\t\tcase batchv1.JobFailed:\n\t\t\t\titem.FailedAt = cond.LastTransitionTime.Local()\n\t\t\t}\n\t\t}\n\t\titem.Memory = job.Spec.Template.Spec.Containers[0].Resources.Requests.Memory().MilliValue()\n\t\titem.VCPU = job.Spec.Template.Spec.Containers[0].Resources.Requests.Cpu().MilliValue()\n\n\t\tmapping[job.UID] = item\n\t\tout.Items = append(out.Items, item)\n\t}\n\n\t//map events to jobs\n\tfor _, ev := range events.Items {\n\t\t_, ok := mapping[ev.InvolvedObject.UID]\n\t\tif ok { //event for one of our jobs\n\t\t\tmapping[ev.InvolvedObject.UID].Details.FailedCreateEvents = append(\n\t\t\t\tmapping[ev.InvolvedObject.UID].Details.FailedCreateEvents,\n\t\t\t\tJobEvent{Message: ev.Message},\n\t\t\t)\n\t\t}\n\t}\n\n\t//map pods to jobs\n\tfor _, pod := range pods.Items {\n\t\tuid, ok := pod.Labels[\"controller-uid\"]\n\t\tif !ok {\n\t\t\tcontinue //not part of a controller\n\t\t}\n\n\t\tjobItem, ok := mapping[types.UID(uid)]\n\t\tif !ok {\n\t\t\tcontinue //not part of any job\n\t\t}\n\n\t\t//technically we can have multiple pods per job (one terminating, unknown etc) so we pick the\n\t\t//one that is created most recently to base our details on\n\t\tif pod.CreationTimestamp.Local().After(jobItem.Details.SeenAt) {\n\t\t\tjobItem.Details.SeenAt = pod.CreationTimestamp.Local() //this pod was created after previous pod\n\t\t} else {\n\t\t\tcontinue //this pod was created before the other one in the item, ignore\n\t\t}\n\n\t\t//the pod phase allows us to distinguish between Pending and Running\n\t\tswitch pod.Status.Phase {\n\t\tcase corev1.PodPending:\n\t\t\tjobItem.Details.Phase = JobDetailsPhasePending\n\t\tcase corev1.PodRunning:\n\t\t\tjobItem.Details.Phase = JobDetailsPhaseRunning\n\t\tcase corev1.PodFailed:\n\t\t\tjobItem.Details.Phase = JobDetailsPhaseFailed\n\t\tcase corev1.PodSucceeded:\n\t\t\tjobItem.Details.Phase = JobDetailsPhaseSucceeded\n\t\tdefault:\n\t\t\tjobItem.Details.Phase = JobDetailsPhaseUnknown\n\t\t}\n\n\t\tfor _, cond := range pod.Status.Conditions {\n\t\t\t//onschedulable is a reason for being pending\n\t\t\tif cond.Type == corev1.PodScheduled {\n\t\t\t\tif cond.Status == corev1.ConditionFalse {\n\t\t\t\t\tif cond.Reason == corev1.PodReasonUnschedulable {\n\t\t\t\t\t\t// From src: \"PodReasonUnschedulable reason in PodScheduled PodCondition means that the scheduler\n\t\t\t\t\t\t// can't schedule the pod right now\"\n\t\t\t\t\t\tjobItem.Details.UnschedulableReason = \"NotYetSchedulable\" //special case\n\t\t\t\t\t\tjobItem.Details.UnschedulableMessage = cond.Message\n\t\t\t\t\t} else {\n\t\t\t\t\t\tjobItem.Details.UnschedulableReason = cond.Reason\n\t\t\t\t\t\tjobItem.Details.UnschedulableMessage = cond.Message\n\t\t\t\t\t}\n\n\t\t\t\t\t//NotScheduled\n\n\t\t\t\t} else if cond.Status == corev1.ConditionTrue {\n\t\t\t\t\tjobItem.Details.Scheduled = true\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\t//container conditions allow us to capture ErrImageNotFound\n\t\tfor _, cstatus := range pod.Status.ContainerStatuses {\n\t\t\tif cstatus.Name != \"main\" { //we only care about the main container\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t//waiting reasons give us ErrImagePull/Backoff\n\t\t\tif cstatus.State.Waiting != nil {\n\t\t\t\tjobItem.Details.WaitingReason = cstatus.State.Waiting.Reason\n\t\t\t\tjobItem.Details.WaitingMessage = cstatus.State.Waiting.Message\n\t\t\t}\n\n\t\t\tif cstatus.State.Terminated != nil {\n\t\t\t\tjobItem.Details.TerminatedReason = cstatus.State.Terminated.Reason\n\t\t\t\tjobItem.Details.TerminatedMessage = cstatus.State.Terminated.Message\n\t\t\t\tjobItem.Details.TerminatedExitCode = cstatus.State.Terminated.ExitCode\n\t\t\t}\n\t\t}\n\t}\n\n\treturn out, nil\n}", "func listPods(client *kubernetes.Clientset, fieldSelector string, labelSelector string) ([]v1.Pod, error) {\n\tlog := log.WithFields(logrus.Fields{\n\t\t\"fieldSelector\": fieldSelector,\n\t\t\"labelSelector\": labelSelector,\n\t})\n\n\tlog.Debug(\"List pods\")\n\tpodList, err := client.CoreV1().Pods(metav1.NamespaceAll).List(metav1.ListOptions{\n\t\tFieldSelector: fieldSelector,\n\t\tLabelSelector: labelSelector,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn podList.Items, nil\n}", "func (r *CompaniesJobsService) List(companyName string) *CompaniesJobsListCall {\n\tc := &CompaniesJobsListCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.companyName = companyName\n\treturn c\n}", "func ListTasks() []*cron.Entry {\n\treturn c.Entries()\n}", "func (nc *NexusConn) TaskList(prefix string, limit int, skip int) (*TaskList, error) {\n\tpar := map[string]interface{}{\n\t\t\"prefix\": prefix,\n\t\t\"limit\": limit,\n\t\t\"skip\": skip,\n\t}\n\tres, err := nc.Exec(\"task.list\", par)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tlist := &TaskList{}\n\tb, err := json.Marshal(res)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\terr = json.Unmarshal(b, list)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn list, nil\n}", "func newCmdJobList(ctx api.Context) *cobra.Command {\n\tvar jsonOutput bool\n\tvar quietOutput bool\n\tcmd := &cobra.Command{\n\t\tUse: \"list\",\n\t\tShort: \"Show all job definitions\",\n\t\tArgs: cobra.NoArgs,\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tclient, err := metronomeClient(ctx)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tjobs, err := client.Jobs(\n\t\t\t\tmetronome.EmbedActiveRun(),\n\t\t\t\tmetronome.EmbedSchedule(),\n\t\t\t\tmetronome.EmbedHistorySummary(),\n\t\t\t)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tif quietOutput {\n\t\t\t\tfor _, job := range jobs {\n\t\t\t\t\tfmt.Fprintln(ctx.Out(), job.ID)\n\t\t\t\t}\n\t\t\t\treturn nil\n\t\t\t}\n\n\t\t\tif jsonOutput {\n\t\t\t\tenc := json.NewEncoder(ctx.Out())\n\t\t\t\tenc.SetIndent(\"\", \" \")\n\t\t\t\treturn enc.Encode(jobs)\n\t\t\t}\n\n\t\t\ttable := cli.NewTable(ctx.Out(), []string{\"ID\", \"STATUS\", \"LAST RUN\"})\n\t\t\tfor _, job := range jobs {\n\t\t\t\ttable.Append([]string{job.ID, job.Status(), job.LastRunStatus()})\n\t\t\t}\n\t\t\ttable.Render()\n\n\t\t\treturn nil\n\t\t},\n\t}\n\tcmd.Flags().BoolVar(&jsonOutput, \"json\", false, \"Print in json format\")\n\tcmd.Flags().BoolVarP(&quietOutput, \"quiet\", \"q\", false, \"Print only IDs of listed jobs\")\n\treturn cmd\n}", "func (s *store) ListJobs() ([]string, error) {\n\tnames := make([]string, 0)\n\tjobsClient := s.client.BatchV1().Jobs(s.namespace)\n\tlist, err := jobsClient.List(metav1.ListOptions{})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfor _, job := range list.Items {\n\t\t// TODO: add label\n\t\t// metadata, err := meta.Accessor(job)\n\t\t// if err != nil {\n\t\t// \treturn nil, err\n\t\t// }\n\t\tnames = append(names, job.GetName())\n\t}\n\treturn names, nil\n}", "func List(ctx *cli.Context) error {\n\tm := task.NewFileManager()\n\ttasks := m.GetAllOpenTasks()\n\n\ttasks = sortTasks(ctx.String(\"sort\"), tasks)\n\n\tfor _, v := range tasks {\n\t\tfmt.Println(v.String())\n\t}\n\treturn nil\n}", "func (z *zpoolctl) List(ctx context.Context, name, options string, properties []string, t string) *execute {\n\targs := []string{\"list\"}\n\tif len(options) > 0 {\n\t\targs = append(args, options)\n\t}\n\tif properties != nil {\n\t\tkv := \"-o \"\n\t\tfor _, v := range properties {\n\t\t\tkv += v + \",\"\n\t\t}\n\t\tkv = strings.TrimSuffix(kv, \",\")\n\t\targs = append(args, kv)\n\t}\n\tif len(t) > 0 {\n\t\targs = append(args, \"-T \"+t)\n\t}\n\targs = append(args, name)\n\treturn &execute{ctx: ctx, name: z.cmd, args: args}\n}", "func (s *runnablesrvc) List(ctx context.Context, p *runnable.ListPayload) (res []*runnable.Runnable, err error) {\n\ts.logger.Print(\"runnable.list\")\n\tidQuery := \"\"\n\tif p.ID != nil {\n\t\tidQuery = *p.ID\n\t}\n\tkindQuery := \"\"\n\tif p.Kind != nil {\n\t\tkindQuery = *p.Kind\n\t}\n\titems, err := s.store.Find(ctx, idQuery, kindQuery, p.Labels)\n\tres = make([]*runnable.Runnable, 0, len(items))\n\tfor _, r := range items {\n\t\tres = append(res, runnableDomainToRest(r))\n\t}\n\treturn res, err\n}", "func Jobs(nomad *NomadServer) ([]Job, int, error) {\n\tjobs := make([]Job, 0)\n\tstatus, err := decodeJSON(url(nomad)+\"/v1/jobs\", &jobs)\n\treturn jobs, status, err\n}", "func ListJobs(w http.ResponseWriter, r *http.Request) {\n\tdefer func() {\n\t\tif err := recover(); err != nil {\n\t\t\t// error handling\n\t\t\tErrorLog(fmt.Sprintf(\"Web list jobs Error: %s\\n\", err))\n\t\t\terrMsg := fmt.Sprintf(\"Error: %s\", err)\n\t\t\thttp.Error(w, errMsg, http.StatusInternalServerError)\n\t\t\tAccessLog(r, http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t}()\n\n\tif r.Method == \"GET\" {\n\t\tif !isLogin(r) {\n\t\t\tLogin(w, r)\n\t\t\treturn\n\t\t}\n\t\tret := JobsOutPut{}\n\t\tApiUrl := fmt.Sprintf(\"%s/jobs\", WEB.ApiUrl)\n\t\tretJson, err := HttpGet(ApiUrl, nil, BASIC_AUTH.User, BASIC_AUTH.Pass)\n\t\terr = json.Unmarshal(retJson, &ret)\n\t\tCheckErr(err)\n\t\t//ErrorLog(ret)\n\t\tif ret.Status != \"OK\" {\n\t\t\tCheckErr(fmt.Errorf(\"%s\", ret.Status))\n\t\t}\n\n\t\tt, err := template.ParseFiles(WEB.TemplatePath+\"/listjobs.html\", WEB.TemplatePath+\"/header.html\", WEB.TemplatePath+\"/footer.html\")\n\t\tCheckErr(err)\n\t\t//err = templates.Execute(w, ret)\n\t\terr = t.Execute(w, ret)\n\t\tCheckErr(err)\n\n\t} else {\n\t\tCheckErr(fmt.Errorf(\"Sorry: Only Accept GET Method\"))\n\t}\n\tAccessLog(r, http.StatusOK)\n\treturn\n}", "func (q *Queue) ListJobs(postJobs bool) ([]*Job, error) {\n\tbucket := jobsBucketName\n\tif postJobs == true {\n\t\tbucket = postJobsBucketName\n\t}\n\tvar jobList []*Job\n\terr := q.db.View(func(tx *nutsdb.Tx) error {\n\t\tentries, err := tx.GetAll(bucket)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfor _, entry := range entries {\n\t\t\tjob, err := DecodeJob(entry.Value)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tjobList = append(jobList, job)\n\t\t}\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn jobList, nil\n}", "func (h *Handler) ListByLabel(labels string) ([]*corev1.Node, error) {\n\tlistOptions := h.Options.ListOptions.DeepCopy()\n\tlistOptions.LabelSelector = labels\n\tnodeList, err := h.clientset.CoreV1().Nodes().List(h.ctx, *listOptions)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn extractList(nodeList), nil\n}", "func (c *iperfTasks) List(opts v1.ListOptions) (result *alpha1.IperfTaskList, err error) {\n\tvar timeout time.Duration\n\tif opts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*opts.TimeoutSeconds) * time.Second\n\t}\n\tresult = &alpha1.IperfTaskList{}\n\terr = c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"iperftasks\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tDo().\n\t\tInto(result)\n\treturn\n}", "func (schematics *SchematicsV1) ListJobsWithContext(ctx context.Context, listJobsOptions *ListJobsOptions) (result *JobList, response *core.DetailedResponse, err error) {\n\terr = core.ValidateStruct(listJobsOptions, \"listJobsOptions\")\n\tif err != nil {\n\t\treturn\n\t}\n\n\tbuilder := core.NewRequestBuilder(core.GET)\n\tbuilder = builder.WithContext(ctx)\n\tbuilder.EnableGzipCompression = schematics.GetEnableGzipCompression()\n\t_, err = builder.ResolveRequestURL(schematics.Service.Options.URL, `/v2/jobs`, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tfor headerName, headerValue := range listJobsOptions.Headers {\n\t\tbuilder.AddHeader(headerName, headerValue)\n\t}\n\n\tsdkHeaders := common.GetSdkHeaders(\"schematics\", \"V1\", \"ListJobs\")\n\tfor headerName, headerValue := range sdkHeaders {\n\t\tbuilder.AddHeader(headerName, headerValue)\n\t}\n\tbuilder.AddHeader(\"Accept\", \"application/json\")\n\n\tif listJobsOptions.Offset != nil {\n\t\tbuilder.AddQuery(\"offset\", fmt.Sprint(*listJobsOptions.Offset))\n\t}\n\tif listJobsOptions.Limit != nil {\n\t\tbuilder.AddQuery(\"limit\", fmt.Sprint(*listJobsOptions.Limit))\n\t}\n\tif listJobsOptions.Sort != nil {\n\t\tbuilder.AddQuery(\"sort\", fmt.Sprint(*listJobsOptions.Sort))\n\t}\n\tif listJobsOptions.Profile != nil {\n\t\tbuilder.AddQuery(\"profile\", fmt.Sprint(*listJobsOptions.Profile))\n\t}\n\tif listJobsOptions.Resource != nil {\n\t\tbuilder.AddQuery(\"resource\", fmt.Sprint(*listJobsOptions.Resource))\n\t}\n\tif listJobsOptions.ActionID != nil {\n\t\tbuilder.AddQuery(\"action_id\", fmt.Sprint(*listJobsOptions.ActionID))\n\t}\n\tif listJobsOptions.List != nil {\n\t\tbuilder.AddQuery(\"list\", fmt.Sprint(*listJobsOptions.List))\n\t}\n\n\trequest, err := builder.Build()\n\tif err != nil {\n\t\treturn\n\t}\n\n\tvar rawResponse map[string]json.RawMessage\n\tresponse, err = schematics.Service.Request(request, &rawResponse)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalModel(rawResponse, \"\", &result, UnmarshalJobList)\n\tif err != nil {\n\t\treturn\n\t}\n\tresponse.Result = result\n\n\treturn\n}", "func (h *Handler) ListByLabel(labels string) ([]*unstructured.Unstructured, error) {\n\tlistOptions := h.Options.ListOptions.DeepCopy()\n\tlistOptions.LabelSelector = labels\n\n\tif err := h.getGVRAndNamespaceScope(); err != nil {\n\t\treturn nil, err\n\t}\n\tif h.isNamespaced {\n\t\treturn extractList(h.dynamicClient.Resource(h.gvr).Namespace(h.namespace).List(h.ctx, *listOptions))\n\t}\n\treturn extractList(h.dynamicClient.Resource(h.gvr).List(h.ctx, *listOptions))\n}", "func (s elasticDLJobNamespaceLister) List(selector labels.Selector) (ret []*v1alpha1.ElasticDLJob, err error) {\n\terr = cache.ListAllByNamespace(s.indexer, s.namespace, selector, func(m interface{}) {\n\t\tret = append(ret, m.(*v1alpha1.ElasticDLJob))\n\t})\n\treturn ret, err\n}", "func (cli *Client) ListTranscodingJobs(pipelineName string) (*api.ListTranscodingJobsResponse, error) {\n\treturn api.ListTranscodingJobs(cli, pipelineName)\n}", "func LabelList(project string) ([]*gitlab.Label, error) {\n\tp, err := FindProject(project)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlist, _, err := lab.Labels.ListLabels(p.ID, &gitlab.ListLabelsOptions{})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn list, nil\n}", "func (c *cronFederatedHPAs) List(ctx context.Context, opts v1.ListOptions) (result *v1alpha1.CronFederatedHPAList, err error) {\n\tvar timeout time.Duration\n\tif opts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*opts.TimeoutSeconds) * time.Second\n\t}\n\tresult = &v1alpha1.CronFederatedHPAList{}\n\terr = c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"cronfederatedhpas\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tDo(ctx).\n\t\tInto(result)\n\treturn\n}", "func (client *Client) ListAsynJobs(request *ListAsynJobsRequest) (response *ListAsynJobsResponse, err error) {\n\tresponse = CreateListAsynJobsResponse()\n\terr = client.DoAction(request, response)\n\treturn\n}", "func ListPods(t testing.TestingT, options *KubectlOptions, filters metav1.ListOptions) []corev1.Pod {\n\tpods, err := ListPodsE(t, options, filters)\n\trequire.NoError(t, err)\n\treturn pods\n}", "func (c *FakeTraincrds) List(opts v1.ListOptions) (result *apisv1.TraincrdList, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewListAction(traincrdsResource, traincrdsKind, c.ns, opts), &apisv1.TraincrdList{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\n\tlabel, _, _ := testing.ExtractFromListOptions(opts)\n\tif label == nil {\n\t\tlabel = labels.Everything()\n\t}\n\tlist := &apisv1.TraincrdList{ListMeta: obj.(*apisv1.TraincrdList).ListMeta}\n\tfor _, item := range obj.(*apisv1.TraincrdList).Items {\n\t\tif label.Matches(labels.Set(item.Labels)) {\n\t\t\tlist.Items = append(list.Items, item)\n\t\t}\n\t}\n\treturn list, err\n}", "func (c *Configuration) JobInfoList(\n\tfiles []string) ([]*job_info.JobInfo, error) {\n\n\tres := make([]*job_info.JobInfo, len(files))\n\tfor i := 0; i < len(files); i++ {\n\t\tres[i] = job_info.NewJobInfo()\n\t\terr := res[i].ReadFromFile(c.JobInfoDir + \"/\" + files[i])\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn res, nil\n}", "func List(input ListInput) *corev1.PodList {\n\tExpect(input.Lister).NotTo(BeNil(), \"input.Lister is required for Pod.List\")\n\tExpect(input.Namespace).NotTo(BeNil(), \"input.Namespace is required for Pod.List\")\n\tExpect(len(input.Labels) == 0).NotTo(BeTrue(), \"input.Labels is required for Pod.List\")\n\n\tBy(fmt.Sprintf(\"Listing pods with labels %v in %s namespace\", input.Labels, input.Namespace))\n\n\tpods := &corev1.PodList{}\n\tExpect(input.Lister.List(context.TODO(), pods, client.InNamespace(input.Namespace), client.MatchingLabels(input.Labels))).Should(Succeed())\n\n\treturn pods\n}", "func List(input ListInput) *corev1.PodList {\n\tExpect(input.Lister).NotTo(BeNil(), \"input.Lister is required for Pod.List\")\n\tExpect(input.Namespace).NotTo(BeNil(), \"input.Namespace is required for Pod.List\")\n\tExpect(len(input.Labels) == 0).NotTo(BeTrue(), \"input.Labels is required for Pod.List\")\n\n\tBy(fmt.Sprintf(\"Listing pods with labels %v in %s namespace\", input.Labels, input.Namespace))\n\n\tpods := &corev1.PodList{}\n\tExpect(input.Lister.List(context.TODO(), pods, client.InNamespace(input.Namespace), client.MatchingLabels(input.Labels))).Should(Succeed())\n\n\treturn pods\n}", "func (p *timeWheel) runJobList(workList *list.List) {\n\tvar jb *job\n\tfor e := workList.Front(); e != nil; e = e.Next() {\n\t\tjb = e.Value.(*job)\n\t\tgo jb.jober.Run()\n\t\tif !jb.oneTime {\n\t\t\tp.add(jb)\n\t\t}\n\t}\n}", "func (c *FakeRobots) List(ctx context.Context, opts v1.ListOptions) (result *v1alpha1.RobotList, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewListAction(robotsResource, robotsKind, c.ns, opts), &v1alpha1.RobotList{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\n\tlabel, _, _ := testing.ExtractFromListOptions(opts)\n\tif label == nil {\n\t\tlabel = labels.Everything()\n\t}\n\tlist := &v1alpha1.RobotList{ListMeta: obj.(*v1alpha1.RobotList).ListMeta}\n\tfor _, item := range obj.(*v1alpha1.RobotList).Items {\n\t\tif label.Matches(labels.Set(item.Labels)) {\n\t\t\tlist.Items = append(list.Items, item)\n\t\t}\n\t}\n\treturn list, err\n}", "func (shim *JobDirectClient) ListJobs(ctx context.Context, in *GraphID, opts ...grpc.CallOption) (Job_ListJobsClient, error) {\n md, _ := metadata.FromOutgoingContext(ctx)\n ictx := metadata.NewIncomingContext(ctx, md)\n\n\tw := &directJobListJobs{ictx, make(chan *QueryJob, 100), in, nil}\n if shim.streamServerInt != nil {\n go func() {\n defer w.close()\n info := grpc.StreamServerInfo{\n FullMethod: \"/gripql.Job/ListJobs\",\n IsServerStream: true,\n }\n w.e = shim.streamServerInt(shim.server, w, &info, _Job_ListJobs_Handler)\n } ()\n return w, nil\n }\n\tgo func() {\n defer w.close()\n\t\tw.e = shim.server.ListJobs(in, w)\n\t}()\n\treturn w, nil\n}", "func (s *TaskService) List() ([]Task, *http.Response, error) {\n\tresObj := new(TaskResponse)\n\tresp, err := s.sling.New().\n\t\tQueryStruct(CreateFunctionParam(\"gettasks\")).\n\t\tReceiveSuccess(resObj)\n\tif resObj != nil && len(resObj.Results) > 0 {\n\t\tif resObj.Results[0].ErrorDesc != nil {\n\t\t\treturn nil, resp, Error{*resObj.Results[0].ErrorDesc}\n\t\t}\n\t\treturn *(&resObj.Results), resp, err\n\t}\n\treturn make([]Task, 0), resp, err\n}", "func (ctrl TerraformJobController) All(c *gin.Context) {\n\tuser := c.MustGet(cUser).(common.User)\n\n\tparser := util.NewQueryParser(c)\n\tmatch := bson.M{}\n\tmatch = parser.Match([]string{\"status\", \"type\", \"failed\"}, match)\n\tmatch = parser.Lookups([]string{\"id\", \"name\", \"labels\"}, match)\n\tquery := db.TerrafromJobs().Find(match)\n\tif order := parser.OrderBy(); order != \"\" {\n\t\tquery.Sort(order)\n\t}\n\n\troles := new(rbac.TerraformJobTemplate)\n\tvar jobs []terraform.Job\n\titer := query.Iter()\n\tvar tmpJob terraform.Job\n\tfor iter.Next(&tmpJob) {\n\t\tif !roles.ReadByID(user, tmpJob.JobTemplateID) {\n\t\t\tcontinue\n\t\t}\n\t\tmetadata.JobMetadata(&tmpJob)\n\t\tjobs = append(jobs, tmpJob)\n\t}\n\tif err := iter.Close(); err != nil {\n\t\tAbortWithError(LogFields{Context: c, Status: http.StatusGatewayTimeout,\n\t\t\tMessage: \"Error while getting job\", Log: logrus.Fields{\n\t\t\t\t\"Error\": err.Error(),\n\t\t\t},\n\t\t})\n\t\treturn\n\t}\n\n\tcount := len(jobs)\n\tpgi := util.NewPagination(c, count)\n\tif pgi.HasPage() {\n\t\tAbortWithError(LogFields{Context: c, Status: http.StatusNotFound,\n\t\t\tMessage: \"#\" + strconv.Itoa(pgi.Page()) + \" page contains no results.\",\n\t\t})\n\t\treturn\n\t}\n\n\tc.JSON(http.StatusOK, common.Response{\n\t\tCount: count,\n\t\tNext: pgi.NextPage(),\n\t\tPrevious: pgi.PreviousPage(),\n\t\tData: jobs[pgi.Skip():pgi.End()],\n\t})\n}", "func OnList(c *grumble.Context) error {\n\tlen := len(config.AppConfig.Plans)\n\tif len == 0 {\n\t\tfmt.Println(\"No plans available. Try \\\"read\\\".\")\n\t\treturn nil\n\t}\n\n\tfor i, plan := range config.AppConfig.Plans {\n\t\tfmt.Println(i+1, plan.Name)\n\t\tfor i, task := range plan.Tasks {\n\t\t\tif task.GetDescription() != \"\" {\n\t\t\t\tfmt.Println(\" \", strconv.Itoa(i+1)+\".\", task.GetDescription())\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func (s *arangoTaskLister) List(selector labels.Selector) (ret []*v2alpha1.ArangoTask, err error) {\n\terr = cache.ListAll(s.indexer, selector, func(m interface{}) {\n\t\tret = append(ret, m.(*v2alpha1.ArangoTask))\n\t})\n\treturn ret, err\n}", "func (re *stubRegistrationService) ListBySelector(ctx context.Context, request common.Selector) (reply common.RegistrationEntries, err error) {\n\treturn reply, err\n}", "func (h *Hub) List(ctx context.Context, request *pb.ListRequest) (*pb.ListReply, error) {\n\tlog.G(h.ctx).Info(\"handling List request\")\n\tvar info = make(map[string]*pb.ListReply_ListValue)\n\th.mu.Lock()\n\tfor k := range h.miners {\n\t\tinfo[k] = new(pb.ListReply_ListValue)\n\t}\n\th.mu.Unlock()\n\n\th.tasksmu.Lock()\n\tfor k, v := range h.tasks {\n\t\tlr, ok := info[v]\n\t\tif ok {\n\t\t\tlr.Values = append(lr.Values, k)\n\t\t\tinfo[v] = lr\n\t\t}\n\t}\n\th.tasksmu.Unlock()\n\treturn &pb.ListReply{Info: info}, nil\n}", "func (t *DRMAATracker) ListJobCategories() ([]string, error) {\n\treturn []string{}, nil\n}", "func (o *RecoverOptions) List(client *ctrlclient.CtrlClient) ([]string, cobra.ShellCompDirective) {\n\tpods, err := o.selectPods(client, []string{})\n\tif err != nil {\n\t\tcommon.PrettyPrint(errors.Wrap(err, \"select pods\").Error(), 0, common.Red)\n\t\treturn nil, cobra.ShellCompDirectiveNoFileComp\n\t}\n\n\tvar names []string\n\tfor _, pod := range pods {\n\t\tnames = append(names, pod.Name)\n\t}\n\n\treturn names, cobra.ShellCompDirectiveNoFileComp\n}", "func (c *Conn) dellJobs(state string) ([]*gofishrf.Task, error) {\n\tendpoint := \"/redfish/v1/Managers/iDRAC.Embedded.1/Oem/Dell/Jobs?$expand=*($levels=1)\"\n\n\tresp, err := c.redfishwrapper.Get(endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif resp.StatusCode != 200 {\n\t\treturn nil, errors.New(\"dell jobs endpoint returned unexpected status code: \" + strconv.Itoa(resp.StatusCode))\n\t}\n\n\tbody, err := io.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdata := dellJobResponseData{}\n\terr = json.Unmarshal(body, &data)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttasks := []*gofishrf.Task{}\n\tfor _, job := range data.Members {\n\t\tif state != \"\" && !strings.EqualFold(job.JobState, state) {\n\t\t\tcontinue\n\t\t}\n\n\t\ttasks = append(tasks, &gofishrf.Task{\n\t\t\tEntity: gofishcommon.Entity{\n\t\t\t\tID: job.ID,\n\t\t\t\tODataID: job.OdataID,\n\t\t\t\tName: job.Name,\n\t\t\t},\n\t\t\tDescription: job.Name,\n\t\t\tPercentComplete: job.PercentComplete,\n\t\t\tStartTime: job.StartTime,\n\t\t\tEndTime: job.CompletionTime,\n\t\t\tTaskState: gofishrf.TaskState(job.JobState),\n\t\t\tTaskStatus: gofishcommon.Health(job.Message), // abuse the TaskStatus to include any status message\n\t\t})\n\t}\n\n\treturn tasks, nil\n}", "func (c *FakeGBPServers) List(opts v1.ListOptions) (result *aciawv1.GBPServerList, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewListAction(gbpserversResource, gbpserversKind, c.ns, opts), &aciawv1.GBPServerList{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\n\tlabel, _, _ := testing.ExtractFromListOptions(opts)\n\tif label == nil {\n\t\tlabel = labels.Everything()\n\t}\n\tlist := &aciawv1.GBPServerList{ListMeta: obj.(*aciawv1.GBPServerList).ListMeta}\n\tfor _, item := range obj.(*aciawv1.GBPServerList).Items {\n\t\tif label.Matches(labels.Set(item.Labels)) {\n\t\t\tlist.Items = append(list.Items, item)\n\t\t}\n\t}\n\treturn list, err\n}", "func (client IotHubResourceClient) ListJobsSender(req *http.Request) (*http.Response, error) {\n\treturn client.Send(req, azure.DoRetryWithRegistration(client.Client))\n}", "func (s *bucketRequestLister) List(selector labels.Selector) (ret []*v1alpha1.BucketRequest, err error) {\n\terr = cache.ListAll(s.indexer, selector, func(m interface{}) {\n\t\tret = append(ret, m.(*v1alpha1.BucketRequest))\n\t})\n\treturn ret, err\n}", "func (client JobClient) ListByAccountSender(req *http.Request) (*http.Response, error) {\n return client.Send(req, azure.DoRetryWithRegistration(client.Client))\n }", "func (c *FakeRedisTriggers) List(ctx context.Context, opts v1.ListOptions) (result *v1beta1.RedisTriggerList, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewListAction(redistriggersResource, redistriggersKind, c.ns, opts), &v1beta1.RedisTriggerList{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\n\tlabel, _, _ := testing.ExtractFromListOptions(opts)\n\tif label == nil {\n\t\tlabel = labels.Everything()\n\t}\n\tlist := &v1beta1.RedisTriggerList{ListMeta: obj.(*v1beta1.RedisTriggerList).ListMeta}\n\tfor _, item := range obj.(*v1beta1.RedisTriggerList).Items {\n\t\tif label.Matches(labels.Set(item.Labels)) {\n\t\t\tlist.Items = append(list.Items, item)\n\t\t}\n\t}\n\treturn list, err\n}", "func (l Leftovers) List(filter string, regex bool) {\n\tl.logger.NoConfirm()\n\tvar deletables []common.Deletable\n\n\tfor _, r := range l.resources {\n\t\tlist, err := r.List(filter, regex)\n\t\tif err != nil {\n\t\t\tl.logger.Println(color.YellowString(err.Error()))\n\t\t}\n\n\t\tdeletables = append(deletables, list...)\n\t}\n\n\tfor _, d := range deletables {\n\t\tl.logger.Println(fmt.Sprintf(\"[%s: %s]\", d.Type(), d.Name()))\n\t}\n}", "func (c *Client) ListFHIRExportJobs(ctx context.Context, params *ListFHIRExportJobsInput, optFns ...func(*Options)) (*ListFHIRExportJobsOutput, error) {\n\tif params == nil {\n\t\tparams = &ListFHIRExportJobsInput{}\n\t}\n\n\tresult, metadata, err := c.invokeOperation(ctx, \"ListFHIRExportJobs\", params, optFns, c.addOperationListFHIRExportJobsMiddlewares)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tout := result.(*ListFHIRExportJobsOutput)\n\tout.ResultMetadata = metadata\n\treturn out, nil\n}", "func Jobs(live bool) Model {\n\t// prepare state options\n\tstateOptions := []Option{\n\t\t{Value: string(axe.Enqueued), Label: \"Enqueued\"},\n\t\t{Value: string(axe.Dequeued), Label: \"Dequeued\"},\n\t\t{Value: string(axe.Completed), Label: \"Completed\"},\n\t\t{Value: string(axe.Failed), Label: \"Failed\"},\n\t\t{Value: string(axe.Cancelled), Label: \"Cancelled\"},\n\t}\n\n\t// prepare model\n\tmodel := Auto(&axe.Model{}, \"job\", \"Job\", \"Jobs\", nil)\n\tmodel.Watchable = live\n\tmodel.Immediate = live\n\tmodel.Creatable = false\n\n\t// filter columns\n\tmodel.Columns = lo.Filter(model.Columns, func(column Column, _ int) bool {\n\t\treturn column.Key != \"events\"\n\t})\n\n\t// patch columns\n\tfor i, column := range model.Columns {\n\t\tif column.Key == \"state\" {\n\t\t\tmodel.Columns[i].Format = FormatMap\n\t\t\tmodel.Columns[i].Options = stateOptions\n\t\t} else if column.Key == \"progress\" {\n\t\t\tmodel.Columns[i].Format = FormatProgress\n\t\t} else if stick.Contains([]string{\"created-at\", \"available-at\", \"started-at\", \"ended-at\", \"finished-at\"}, column.Key) {\n\t\t\tmodel.Columns[i].Format = FormatRelativeDate\n\t\t}\n\t}\n\n\t// patch fields\n\tfor i, field := range model.Fields {\n\t\tif field.Key == \"state\" {\n\t\t\tmodel.Fields[i].Control = ControlSelect\n\t\t\tmodel.Fields[i].Options = stateOptions\n\t\t}\n\t}\n\n\treturn model\n}", "func (l Leftovers) List(filter string, regex bool) {\n\tl.logger.NoConfirm()\n\n\tvar deletables []common.Deletable\n\n\tfor _, r := range l.resources {\n\t\tlist, err := r.List(filter, regex)\n\t\tif err != nil {\n\t\t\tl.logger.Println(color.YellowString(err.Error()))\n\t\t}\n\n\t\tdeletables = append(deletables, list...)\n\t}\n\n\tfor _, d := range deletables {\n\t\tl.logger.Println(fmt.Sprintf(\"[%s: %s]\", d.Type(), d.Name()))\n\t}\n}", "func (schematics *SchematicsV1) ListJobs(listJobsOptions *ListJobsOptions) (result *JobList, response *core.DetailedResponse, err error) {\n\treturn schematics.ListJobsWithContext(context.Background(), listJobsOptions)\n}", "func listPods(ctx context.Context, c client.Client, rokku *rokkuv1alpha1.Rokku) ([]rokkuv1alpha1.PodStatus, error) {\n\tpodList := &corev1.PodList{}\n\tlabelSelector := labels.SelectorFromSet(k8s.LabelsForRokku(rokku.Name))\n\tlistOps := &client.ListOptions{Namespace: rokku.Namespace, LabelSelector: labelSelector}\n\terr := c.List(ctx, podList, listOps)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar pods []rokkuv1alpha1.PodStatus\n\n\tfor _, p := range podList.Items {\n\t\tif p.Status.PodIP == \"\" {\n\t\t\tp.Status.PodIP = \"<pending>\"\n\t\t}\n\n\t\tif p.Status.HostIP == \"\" {\n\t\t\tp.Status.HostIP = \"<pending>\"\n\t\t}\n\n\t\tpods = append(pods, rokkuv1alpha1.PodStatus{\n\t\t\tName: p.Name,\n\t\t\tPodIP: p.Status.PodIP,\n\t\t\tHostIP: p.Status.HostIP,\n\t\t})\n\t}\n\tsort.Slice(pods, func(i, j int) bool {\n\t\treturn pods[i].Name < pods[j].Name\n\t})\n\n\treturn pods, nil\n}", "func (ts *TaskService) List(reqdata *TaskListRequest) (*TaskListResponse, *http.Response, error) {\n\n\tu := fmt.Sprintf(\"tasks/list\")\n\n\tu, err := addOptions(u, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treq, err := ts.client.NewRequest(\"POST\", u, reqdata)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tvar result *TaskListResponse\n\tresp, err := ts.client.Do(req, &result)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\n\treturn result, resp, nil\n}", "func (t *TMDB) Jobs() (*Jobs, error) {\n\tj := new(Jobs)\n\tif err := t.get(j, \"/3/configuration/jobs\", url.Values{}); err != nil {\n\t\treturn nil, err\n\t}\n\treturn j, nil\n}", "func (c *concurrencyControls) List(ctx context.Context, opts v1.ListOptions) (result *v1alpha1.ConcurrencyControlList, err error) {\n\tvar timeout time.Duration\n\tif opts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*opts.TimeoutSeconds) * time.Second\n\t}\n\tresult = &v1alpha1.ConcurrencyControlList{}\n\terr = c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"concurrencycontrols\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tDo(ctx).\n\t\tInto(result)\n\treturn\n}", "func (schedulerClient *Scheduler)GetPodListWithLabels(namespace string, targetLabels map[string]string) []apiv1.Pod {\n\tpodsClient := schedulerClient.clientSet.CoreV1().Pods(namespace)\n\tset:=labels.Set(targetLabels)\n\tpods, err := podsClient.List(metav1.ListOptions{\n\t\tLabelSelector: set.AsSelector().String(),\n\t})\n\n\tfor attempts:=0;err!= nil;{\n\t\tif attempts > 2 {\n\t\t\tlog.Println(err)\n\t\t\treturn nil\t// too many tries\n\t\t}\n\t\tpods, err = podsClient.List(metav1.ListOptions{\n\t\t\tLabelSelector: set.AsSelector().String(),\n\t\t})\n\t\tattempts = attempts+1\n\t}\n\treturn pods.Items\n}", "func (client JobClient) ListByAccountResponder(resp *http.Response) (result JobResourceDescriptionList, err error) {\n err = autorest.Respond(\n resp,\n azure.WithErrorUnlessStatusCode(http.StatusOK),\n autorest.ByUnmarshallingJSON(&result),\n autorest.ByClosing())\n result.Response = autorest.Response{Response: resp}\n return\n }", "func (s *hTTPCheckLister) List(selector labels.Selector) (ret []*v1alpha1.HTTPCheck, err error) {\n\terr = cache.ListAll(s.indexer, selector, func(m interface{}) {\n\t\tret = append(ret, m.(*v1alpha1.HTTPCheck))\n\t})\n\treturn ret, err\n}", "func (s *testRunLister) List(selector labels.Selector) (ret []*v1alpha1.TestRun, err error) {\n\terr = cache.ListAll(s.indexer, selector, func(m interface{}) {\n\t\tret = append(ret, m.(*v1alpha1.TestRun))\n\t})\n\treturn ret, err\n}", "func (client Client) ListPods(labelQuery map[string]string) (api.PodList, error) {\n\tpath := \"pods\"\n\tif labelQuery != nil && len(labelQuery) > 0 {\n\t\tpath += \"?labels=\" + EncodeLabelQuery(labelQuery)\n\t}\n\tvar result api.PodList\n\t_, err := client.rawRequest(\"GET\", path, nil, &result)\n\treturn result, err\n}", "func (uis *UIService) ListJobSpecs(req *v1.ListJobSpecsRequest, srv v1.WerftUI_ListJobSpecsServer) error {\n\tuis.mu.RLock()\n\tdefer uis.mu.RUnlock()\n\n\tfor _, r := range uis.cache {\n\t\terr := srv.Send(r)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (js *JobSession) GetJobs(filter drmaa2interface.JobInfo) ([]drmaa2interface.Job, error) {\n\tvar joblist []drmaa2interface.Job\n\n\tfor _, tracker := range js.tracker {\n\t\tjobs, err := tracker.ListJobs()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tfor _, jobid := range jobs {\n\t\t\tif jinfo, err := tracker.JobInfo(jobid); err != nil {\n\t\t\t\tcontinue\n\t\t\t} else {\n\t\t\t\tif d2hlp.JobInfoMatches(jinfo, filter) {\n\t\t\t\t\t// TODO get template from DB\n\t\t\t\t\tjobtemplate := drmaa2interface.JobTemplate{}\n\n\t\t\t\t\tjob := newJob(jobid, js.name, jobtemplate, tracker)\n\t\t\t\t\tjoblist = append(joblist, drmaa2interface.Job(job))\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn joblist, nil\n}", "func (c *Client) ListFHIRImportJobs(ctx context.Context, params *ListFHIRImportJobsInput, optFns ...func(*Options)) (*ListFHIRImportJobsOutput, error) {\n\tif params == nil {\n\t\tparams = &ListFHIRImportJobsInput{}\n\t}\n\n\tresult, metadata, err := c.invokeOperation(ctx, \"ListFHIRImportJobs\", params, optFns, c.addOperationListFHIRImportJobsMiddlewares)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tout := result.(*ListFHIRImportJobsOutput)\n\tout.ResultMetadata = metadata\n\treturn out, nil\n}", "func (r *ProjectsLocationsMigrationJobsService) List(parent string) *ProjectsLocationsMigrationJobsListCall {\n\tc := &ProjectsLocationsMigrationJobsListCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.parent = parent\n\treturn c\n}", "func (c *restClient) ListPatchJobs(ctx context.Context, req *osconfigpb.ListPatchJobsRequest, opts ...gax.CallOption) *PatchJobIterator {\n\tit := &PatchJobIterator{}\n\treq = proto.Clone(req).(*osconfigpb.ListPatchJobsRequest)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tit.InternalFetch = func(pageSize int, pageToken string) ([]*osconfigpb.PatchJob, string, error) {\n\t\tresp := &osconfigpb.ListPatchJobsResponse{}\n\t\tif pageToken != \"\" {\n\t\t\treq.PageToken = pageToken\n\t\t}\n\t\tif pageSize > math.MaxInt32 {\n\t\t\treq.PageSize = math.MaxInt32\n\t\t} else if pageSize != 0 {\n\t\t\treq.PageSize = int32(pageSize)\n\t\t}\n\t\tbaseUrl, err := url.Parse(c.endpoint)\n\t\tif err != nil {\n\t\t\treturn nil, \"\", err\n\t\t}\n\t\tbaseUrl.Path += fmt.Sprintf(\"/v1beta/%v/patchJobs\", req.GetParent())\n\n\t\tparams := url.Values{}\n\t\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\t\tif req.GetFilter() != \"\" {\n\t\t\tparams.Add(\"filter\", fmt.Sprintf(\"%v\", req.GetFilter()))\n\t\t}\n\t\tif req.GetPageSize() != 0 {\n\t\t\tparams.Add(\"pageSize\", fmt.Sprintf(\"%v\", req.GetPageSize()))\n\t\t}\n\t\tif req.GetPageToken() != \"\" {\n\t\t\tparams.Add(\"pageToken\", fmt.Sprintf(\"%v\", req.GetPageToken()))\n\t\t}\n\n\t\tbaseUrl.RawQuery = params.Encode()\n\n\t\t// Build HTTP headers from client and context metadata.\n\t\thds := append(c.xGoogHeaders, \"Content-Type\", \"application/json\")\n\t\theaders := gax.BuildHeaders(ctx, hds...)\n\t\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\t\tif settings.Path != \"\" {\n\t\t\t\tbaseUrl.Path = settings.Path\n\t\t\t}\n\t\t\thttpReq, err := http.NewRequest(\"GET\", baseUrl.String(), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\thttpReq.Header = headers\n\n\t\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tdefer httpRsp.Body.Close()\n\n\t\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn nil\n\t\t}, opts...)\n\t\tif e != nil {\n\t\t\treturn nil, \"\", e\n\t\t}\n\t\tit.Response = resp\n\t\treturn resp.GetPatchJobs(), resp.GetNextPageToken(), nil\n\t}\n\n\tfetch := func(pageSize int, pageToken string) (string, error) {\n\t\titems, nextPageToken, err := it.InternalFetch(pageSize, pageToken)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tit.items = append(it.items, items...)\n\t\treturn nextPageToken, nil\n\t}\n\n\tit.pageInfo, it.nextFunc = iterator.NewPageInfo(fetch, it.bufLen, it.takeBuf)\n\tit.pageInfo.MaxSize = int(req.GetPageSize())\n\tit.pageInfo.Token = req.GetPageToken()\n\n\treturn it\n}", "func (g genericPlugin) List(gvk schema.GroupVersionKind, namespace string,\n\tclient plugin.KubernetesConnector) ([]helm.KubernetesResource, error) {\n\n\tvar returnData []helm.KubernetesResource\n\treturn returnData, nil\n}" ]
[ "0.7026325", "0.6361638", "0.60010505", "0.5990891", "0.5946092", "0.59099966", "0.5854926", "0.5831763", "0.58041465", "0.576753", "0.5726013", "0.56881773", "0.5670315", "0.5655169", "0.5636831", "0.5607969", "0.56053746", "0.5604373", "0.55717695", "0.55702835", "0.5560738", "0.5554281", "0.5522473", "0.5480123", "0.54763204", "0.5461151", "0.545796", "0.545637", "0.54337776", "0.5423021", "0.540975", "0.5409176", "0.5380621", "0.5375148", "0.53527194", "0.53503627", "0.53384554", "0.5328281", "0.52984667", "0.5296685", "0.52313924", "0.52055115", "0.5194707", "0.517929", "0.5179056", "0.5177619", "0.51598144", "0.5157234", "0.51351905", "0.51223844", "0.51210815", "0.5112276", "0.5110768", "0.50979215", "0.50627935", "0.50507927", "0.5044465", "0.50382227", "0.50303525", "0.5002369", "0.5000827", "0.4993256", "0.4993256", "0.49804652", "0.4962422", "0.49459562", "0.49366048", "0.49301714", "0.49286968", "0.49270067", "0.49144468", "0.49073955", "0.4905108", "0.4886611", "0.48774603", "0.4868701", "0.48656398", "0.48644787", "0.48643804", "0.48552212", "0.48532236", "0.48530334", "0.48501092", "0.48461872", "0.48285493", "0.4827539", "0.48076865", "0.48010543", "0.47996232", "0.47901064", "0.47846866", "0.47844586", "0.47818473", "0.47775698", "0.47745663", "0.4771324", "0.4764884", "0.47565833", "0.47492322", "0.47434136" ]
0.7250088
0
Watch returns a watch.Interface that watches the requested tZCronJobs.
func (c *FakeTZCronJobs) Watch(opts v1.ListOptions) (watch.Interface, error) { return c.Fake. InvokesWatch(testing.NewWatchAction(tzcronjobsResource, c.ns, opts)) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *scheduledJobs) Watch(opts api.ListOptions) (watch.Interface, error) {\n\treturn c.r.Get().\n\t\tPrefix(\"watch\").\n\t\tNamespace(c.ns).\n\t\tResource(\"scheduledjobs\").\n\t\tVersionedParams(&opts, api.ParameterCodec).\n\t\tWatch()\n}", "func (c *aITrainingJobs) Watch(opts metav1.ListOptions) (watch.Interface, error) {\n\tvar timeout time.Duration\n\tif opts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*opts.TimeoutSeconds) * time.Second\n\t}\n\topts.Watch = true\n\treturn c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"aitrainingjobs\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tWatch()\n}", "func (c *NamespacesJobsListCall) Watch(watch bool) *NamespacesJobsListCall {\n\tc.urlParams_.Set(\"watch\", fmt.Sprint(watch))\n\treturn c\n}", "func (s *Scheduler) WatchJobs(ctx context.Context, c chan<- ffs.Job, iid ffs.APIID) error {\n\treturn s.js.Watch(ctx, c, iid)\n}", "func (c *cronFederatedHPAs) Watch(ctx context.Context, opts v1.ListOptions) (watch.Interface, error) {\n\tvar timeout time.Duration\n\tif opts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*opts.TimeoutSeconds) * time.Second\n\t}\n\topts.Watch = true\n\treturn c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"cronfederatedhpas\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tWatch(ctx)\n}", "func WatchJobs(client client.Interface, system v1.SystemID, format printer.Format, w io.Writer) {\n\t// Poll the API for the systems and send it to the channel\n\tjobs := make(chan []v1.Job)\n\n\tgo wait.PollImmediateInfinite(\n\t\t5*time.Second,\n\t\tfunc() (bool, error) {\n\t\t\td, err := client.V1().Systems().Jobs(system).List()\n\t\t\tif err != nil {\n\t\t\t\treturn false, err\n\t\t\t}\n\n\t\t\tjobs <- d\n\t\t\treturn false, nil\n\t\t},\n\t)\n\n\tvar handle func([]v1.Job)\n\tswitch format {\n\tcase printer.FormatTable:\n\t\tt := jobsTable(w)\n\t\thandle = func(jobs []v1.Job) {\n\t\t\tr := jobsTableRows(jobs)\n\t\t\tt.Overwrite(r)\n\t\t}\n\n\tcase printer.FormatJSON:\n\t\tj := printer.NewJSON(w)\n\t\thandle = func(jobs []v1.Job) {\n\t\t\tj.Print(jobs)\n\t\t}\n\n\tdefault:\n\t\tpanic(fmt.Sprintf(\"unexpected format %v\", format))\n\t}\n\n\tfor d := range jobs {\n\t\thandle(d)\n\t}\n}", "func (obs *Observer) Watch(opts metav1.ListOptions) (watch.Interface, error) {\n\treturn obs.client.Namespace(obs.namespace).Watch(opts)\n}", "func Watch(paths ...string) (*Watcher, error) {\n\tevent := make(chan EventItem)\n\terr := watch(paths, event)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Watcher{\n\t\tEvent: event,\n\t}, nil\n}", "func (i *API) WatchJobs(ctx context.Context, c chan<- ffs.Job, jids ...ffs.JobID) error {\n\tvar jobs []ffs.Job\n\tfor _, jid := range jids {\n\t\tj, err := i.sched.GetJob(jid)\n\t\tif err == scheduler.ErrNotFound {\n\t\t\tcontinue\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"getting current job state: %s\", err)\n\t\t}\n\t\tjobs = append(jobs, j)\n\t}\n\n\tch := make(chan ffs.Job, 1)\n\tfor _, j := range jobs {\n\t\tselect {\n\t\tcase ch <- j:\n\t\tdefault:\n\t\t\tlog.Warnf(\"dropped notifying current job state on slow receiver on %s\", i.cfg.ID)\n\t\t}\n\t}\n\tvar err error\n\tgo func() {\n\t\terr = i.sched.WatchJobs(ctx, ch, i.cfg.ID)\n\t\tclose(ch)\n\t}()\n\tfor j := range ch {\n\t\tif len(jids) == 0 {\n\t\t\tc <- j\n\t\t}\n\tJidLoop:\n\t\tfor _, jid := range jids {\n\t\t\tif jid == j.ID {\n\t\t\t\tc <- j\n\t\t\t\tbreak JidLoop\n\t\t\t}\n\t\t}\n\t}\n\tif err != nil {\n\t\treturn fmt.Errorf(\"scheduler listener: %s\", err)\n\t}\n\n\treturn nil\n}", "func (c *ConsulClient) Watch(ctx context.Context, wh *WatchConfig) (IWatcher, error) {\n\tregistryOperationCount.WithLabelValues(env, \"Watch\").Inc()\n\n\tstartTime := time.Now()\n\tdefer func() {\n\t\tregistryOperationTimeTaken.WithLabelValues(env, \"Watch\").Observe(time.Now().Sub(startTime).Seconds())\n\t}()\n\n\tparams := map[string]interface{}{}\n\n\tif wh.WatchType == \"key\" {\n\t\tparams[\"type\"] = wh.WatchType\n\t\tparams[\"key\"] = wh.WatchPath\n\t} else if wh.WatchType == \"keyprefix\" {\n\t\tparams[\"type\"] = wh.WatchType\n\t\tparams[\"prefix\"] = wh.WatchPath\n\t}\n\n\tplan, err := watch.Parse(params)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tcwh := NewConsulWatcher(ctx, wh, plan, c.client)\n\n\treturn cwh, nil\n}", "func (c *klusterlets) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) {\n\tvar timeout time.Duration\n\tif opts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*opts.TimeoutSeconds) * time.Second\n\t}\n\topts.Watch = true\n\treturn c.client.Get().\n\t\tResource(\"klusterlets\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tWatch(ctx)\n}", "func (m *Master) watch(taskName, taskType string) {\n\tctx, cancel := context.WithCancel(context.Background())\n\t// Start the timer to keep track of the task.\n\tgo func(ctx context.Context, timeout chan struct{ taskName, taskType string }, task string) {\n\t\tt := time.NewTimer(taskTimeout)\n\t\tdefer t.Stop()\n\t\tselect {\n\t\tcase <-t.C:\n\t\t\ttimeout <- struct {\n\t\t\t\ttaskName string\n\t\t\t\ttaskType string\n\t\t\t}{taskName, taskType}\n\t\tcase <-ctx.Done():\n\t\t\treturn\n\t\t}\n\t}(ctx, m.timeout, taskName)\n\n\t// Append to the cancelers slice.\n\tm.cancelerMutex.Lock()\n\tdefer m.cancelerMutex.Unlock()\n\tm.cancelers = append(m.cancelers, cancel)\n}", "func (c *volumeSnapshotSchedules) Watch(ctx context.Context, opts v1.ListOptions) (watch.Interface, error) {\n\tvar timeout time.Duration\n\tif opts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*opts.TimeoutSeconds) * time.Second\n\t}\n\topts.Watch = true\n\treturn c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"volumesnapshotschedules\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tWatch(ctx)\n}", "func (f *FFS) WatchJobs(ctx context.Context, ch chan<- JobEvent, jids ...ffs.JobID) error {\n\tjidStrings := make([]string, len(jids))\n\tfor i, jid := range jids {\n\t\tjidStrings[i] = jid.String()\n\t}\n\n\tstream, err := f.client.WatchJobs(ctx, &rpc.WatchJobsRequest{Jids: jidStrings})\n\tif err != nil {\n\t\treturn err\n\t}\n\tgo func() {\n\t\tfor {\n\t\t\treply, err := stream.Recv()\n\t\t\tif err == io.EOF || status.Code(err) == codes.Canceled {\n\t\t\t\tclose(ch)\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tif err != nil {\n\t\t\t\tch <- JobEvent{Err: err}\n\t\t\t\tclose(ch)\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tc, err := util.CidFromString(reply.Job.Cid)\n\t\t\tif err != nil {\n\t\t\t\tch <- JobEvent{Err: err}\n\t\t\t\tclose(ch)\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tdealErrors, err := fromRPCDealErrors(reply.Job.DealErrors)\n\t\t\tif err != nil {\n\t\t\t\tch <- JobEvent{Err: err}\n\t\t\t\tclose(ch)\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tvar status ffs.JobStatus\n\t\t\tswitch reply.Job.Status {\n\t\t\tcase rpc.JobStatus_JOB_STATUS_QUEUED:\n\t\t\t\tstatus = ffs.Queued\n\t\t\tcase rpc.JobStatus_JOB_STATUS_EXECUTING:\n\t\t\t\tstatus = ffs.Executing\n\t\t\tcase rpc.JobStatus_JOB_STATUS_FAILED:\n\t\t\t\tstatus = ffs.Failed\n\t\t\tcase rpc.JobStatus_JOB_STATUS_CANCELED:\n\t\t\t\tstatus = ffs.Canceled\n\t\t\tcase rpc.JobStatus_JOB_STATUS_SUCCESS:\n\t\t\t\tstatus = ffs.Success\n\t\t\tdefault:\n\t\t\t\tstatus = ffs.Unspecified\n\t\t\t}\n\t\t\tjob := ffs.Job{\n\t\t\t\tID: ffs.JobID(reply.Job.Id),\n\t\t\t\tAPIID: ffs.APIID(reply.Job.ApiId),\n\t\t\t\tCid: c,\n\t\t\t\tStatus: status,\n\t\t\t\tErrCause: reply.Job.ErrCause,\n\t\t\t\tDealErrors: dealErrors,\n\t\t\t}\n\t\t\tch <- JobEvent{Job: job}\n\t\t}\n\t}()\n\treturn nil\n}", "func (c *AnalyticsController) runWatches() {\n\tlastResourceVersion := big.NewInt(0)\n\tcurrentResourceVersion := big.NewInt(0)\n\twatchListItems := WatchFuncList(c.kclient, c.client)\n\tfor name := range watchListItems {\n\n\t\t// assign local variable (not in range operator above) so that each\n\t\t// goroutine gets the correct watch function required\n\t\twfnc := watchListItems[name]\n\t\tn := name\n\t\tbackoff := 1 * time.Second\n\n\t\tgo wait.Until(func() {\n\t\t\t// any return from this func only exits that invocation of the func.\n\t\t\t// wait.Until will call it again after its sync period.\n\t\t\twatchLog := log.WithFields(log.Fields{\n\t\t\t\t\"watch\": n,\n\t\t\t})\n\t\t\twatchLog.Infof(\"starting watch\")\n\t\t\tw, err := wfnc.watchFunc(metav1.ListOptions{})\n\t\t\tif err != nil {\n\t\t\t\twatchLog.Errorf(\"error creating watch: %v\", err)\n\t\t\t}\n\n\t\t\twatchLog.Debugf(\"backing off watch for %v seconds\", backoff)\n\t\t\ttime.Sleep(backoff)\n\t\t\tbackoff = backoff * 2\n\t\t\tif backoff > 60*time.Second {\n\t\t\t\tbackoff = 60 * time.Second\n\t\t\t}\n\n\t\t\tif w == nil {\n\t\t\t\twatchLog.Errorln(\"watch function nil, watch not created, returning\")\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tfor {\n\t\t\t\tselect {\n\t\t\t\tcase event, ok := <-w.ResultChan():\n\t\t\t\t\tif !ok {\n\t\t\t\t\t\twatchLog.Warnln(\"watch channel closed unexpectedly, attempting to re-establish\")\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\n\t\t\t\t\tif event.Type == watch.Error {\n\t\t\t\t\t\twatchLog.Errorf(\"watch channel returned error: %s\", spew.Sdump(event))\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\n\t\t\t\t\t// success means the watch is working.\n\t\t\t\t\t// reset the backoff back to 1s for this watch\n\t\t\t\t\tbackoff = 1 * time.Second\n\n\t\t\t\t\tif event.Type == watch.Added || event.Type == watch.Deleted {\n\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\twatchLog.Errorf(\"Unable to create object meta for %v: %v\", event.Object, err)\n\t\t\t\t\t\t\treturn\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\tm, err := meta.Accessor(event.Object)\n\t\t\t\t\t\t// if both resource versions can be converted to numbers\n\t\t\t\t\t\t// and if the current resource version is lower than the\n\t\t\t\t\t\t// last recorded resource version for this resource type\n\t\t\t\t\t\t// then skip the event\n\t\t\t\t\t\tc.mutex.RLock()\n\t\t\t\t\t\tif _, ok := lastResourceVersion.SetString(c.watchResourceVersions[n], 10); ok {\n\t\t\t\t\t\t\tif _, ok = currentResourceVersion.SetString(m.GetResourceVersion(), 10); ok {\n\t\t\t\t\t\t\t\tif lastResourceVersion.Cmp(currentResourceVersion) == 1 {\n\t\t\t\t\t\t\t\t\twatchLog.Debugf(\"ResourceVersion %v is to old (%v)\",\n\t\t\t\t\t\t\t\t\t\tcurrentResourceVersion, c.watchResourceVersions[n])\n\t\t\t\t\t\t\t\t\tc.mutex.RUnlock()\n\t\t\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t\tc.mutex.RUnlock()\n\n\t\t\t\t\t\t// each watch is a separate go routine\n\t\t\t\t\t\tc.mutex.Lock()\n\t\t\t\t\t\tc.watchResourceVersions[n] = m.GetResourceVersion()\n\t\t\t\t\t\tc.mutex.Unlock()\n\n\t\t\t\t\t\tanalytic, err := newEvent(c.typer, event.Object, event.Type)\n\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\twatchLog.Errorf(\"unexpected error creating analytic from watch event %#v\", event.Object)\n\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t// additional info will be set to the analytic and\n\t\t\t\t\t\t\t// an instance queued for all destinations\n\t\t\t\t\t\t\terr := c.AddEvent(analytic)\n\t\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\t\twatchLog.Errorf(\"error adding event: %v - %v\", err, analytic)\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}, 1*time.Millisecond, c.stopChannel)\n\t}\n}", "func (c *jxTasks) Watch(opts v1.ListOptions) (watch.Interface, error) {\n\topts.Watch = true\n\treturn c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"jxtasks\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tWatch()\n}", "func (d *JobManager) watchJobExecution(request *restful.Request, response *restful.Response) {\n\n}", "func (m *Manager) Watch(mObj *models.CrudWatcherCreateArgs, client crude.Watcher) (string, error) {\n\tm.InWObj = mObj\n\treturn m.RetWID, m.RetWErr\n}", "func (c *interacts) Watch(opts meta_v1.ListOptions) (watch.Interface, error) {\n\topts.Watch = true\n\treturn c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"interacts\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tWatch()\n}", "func (c *snapshotRules) Watch(ctx context.Context, opts v1.ListOptions) (watch.Interface, error) {\n\tvar timeout time.Duration\n\tif opts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*opts.TimeoutSeconds) * time.Second\n\t}\n\topts.Watch = true\n\treturn c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"snapshotrules\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tWatch(ctx)\n}", "func (k *Kubernetes) Watch(qname string) error {\n\treturn k.APIConn.Watch(qname)\n}", "func (c *FakeGlueCatalogTables) Watch(opts v1.ListOptions) (watch.Interface, error) {\n\treturn c.Fake.\n\t\tInvokesWatch(testing.NewWatchAction(gluecatalogtablesResource, c.ns, opts))\n\n}", "func (b *Byzcoin) Watch(ctx context.Context) <-chan *ledger.TransactionResult {\n\tc := make(chan *ledger.TransactionResult, 100)\n\tb.bc.Watch(ctx, observer{ch: c})\n\n\treturn c\n}", "func (c *rpcServices) Watch(opts metav1.ListOptions) (watch.Interface, error) {\n\topts.Watch = true\n\treturn c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"rpcservices\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tWatch()\n}", "func (w *TaskWatcher) Watch(ctx context.Context) <-chan *TaskEvent {\n\tc := make(chan *TaskEvent, w.cfg.ChannelSize)\n\tgo w.watch(ctx, c)\n\treturn c\n}", "func (s *stateManager) Watch(watcher *AllocationWatcher) func() {\n\tstopChan := make(chan interface{})\n\ts.stopChan = append(s.stopChan, stopChan)\n\tctx := context.Background()\n\n\tkey := fmt.Sprintf(\"%s/allocations\", etcdPrefix)\n\twatchChan := s.cli.Watch(ctx, key, clientv3.WithPrefix(), clientv3.WithPrevKV())\n\n\tstopFunc := func() {\n\t\tstopChan <- true\n\t}\n\n\t// Start a new thread and watch for changes in etcd\n\tgo s.watchChannel(watchChan, stopChan, watcher)\n\n\treturn stopFunc\n}", "func (fk *FakeRouter) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) {\n\tpanic(\"not implemented\")\n}", "func (cs *checkoutService) Watch(req *healthpb.HealthCheckRequest, server healthpb.Health_WatchServer) error {\n\treturn nil\n}", "func (c *concurrencyControls) Watch(ctx context.Context, opts v1.ListOptions) (watch.Interface, error) {\n\tvar timeout time.Duration\n\tif opts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*opts.TimeoutSeconds) * time.Second\n\t}\n\topts.Watch = true\n\treturn c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"concurrencycontrols\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tWatch(ctx)\n}", "func (r *Registry) Watch(ctx context.Context, serviceName string) (registry.Watcher, error) {\n\treturn newWatcher(ctx, r.opt.Namespace, serviceName, r.consumer)\n}", "func (c *kongs) Watch(opts v1.ListOptions) (watch.Interface, error) {\n\topts.Watch = true\n\treturn c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"kongs\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tWatch()\n}", "func (t *FakeObjectTracker) Watch(gvr schema.GroupVersionResource, name string) (watch.Interface, error) {\n\tif t.fakingOptions.failAll != nil {\n\t\terr := t.fakingOptions.failAll.RunFakeInvocations()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn t.delegatee.Watch(gvr, name)\n}", "func Watch(i Item, conn *redis.Client) error {\n\treply := conn.Cmd(\"WATCH\", i.getWatchKey())\n\tif reply.Err != nil {\n\t\tOnPrimaryFailure()\n\t}\n\treturn reply.Err\n}", "func (a *App) Watch(budget string, bedrooms string, brokers_fee bool, interval time.Duration) {\n\tparams := url.Values{\n\t\t\"max_price\": {budget},\n\t\t\"min_bedrooms\": {bedrooms},\n\t\t\"max_bedrooms\": {bedrooms},\n\t\t\"availabilityMode\": {\"0\"},\n\t\t\"broker_fee\": {Btos(brokers_fee)},\n\t\t\"sale_date\": {\"all+dates\"},\n\t}.Encode()\n\n\tclURL := fmt.Sprintf(\"https://%s.craigslist.org/search/apa?%s\", a.site, params)\n\n\tfor {\n\t\ta.collector.Visit(clURL)\n\t\tlog.Printf(\n\t\t\t\"Found %d new listing(s) on the last scrape.\\n\",\n\t\t\ta.countNewLastScrape,\n\t\t)\n\t\ttime.Sleep(interval)\n\t}\n}", "func Watch(strinput2 string, tc net.Conn, watcherport int, dest *string) {\n\tnetwork.SendDataMessage(&tc, 4, 8, watcherport, strinput2)\n\tgo ListentoWatcherport(watcherport, dest)\n}", "func (api *snapshotrestoreAPI) Watch(handler SnapshotRestoreHandler) error {\n\tapi.ct.startWorkerPool(\"SnapshotRestore\")\n\treturn api.ct.WatchSnapshotRestore(handler)\n}", "func (c *sandboxes) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) {\n\tvar timeout time.Duration\n\tif opts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*opts.TimeoutSeconds) * time.Second\n\t}\n\topts.Watch = true\n\treturn c.client.Get().\n\t\tResource(\"sandboxes\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tWatch(ctx)\n}", "func (k *kubernetes) Watch(opts ...router.WatchOption) (router.Watcher, error) {\n\treturn &watcher{\n\t\tevents: make(chan *router.Event),\n\t}, nil\n}", "func (c *FakeGBPServers) Watch(opts v1.ListOptions) (watch.Interface, error) {\n\treturn c.Fake.\n\t\tInvokesWatch(testing.NewWatchAction(gbpserversResource, c.ns, opts))\n\n}", "func (w *ResourceWatcher) Watch(objType runtime.Object, done <-chan struct{}) {\n\tsource := cache.NewListWatchFromClient(\n\t\tw.client,\n\t\tw.resource.Plural,\n\t\tw.namespace,\n\t\tfields.Everything())\n\t_, controller := cache.NewInformer(\n\t\tsource,\n\n\t\t// The object type.\n\t\tobjType,\n\n\t\t// resyncPeriod\n\t\t// Every resyncPeriod, all resources in the cache will retrigger events.\n\t\t// Set to 0 to disable the resync.\n\t\t0,\n\n\t\t// Your custom resource event handlers.\n\t\tw.resourceEventHandlers)\n\n\tgo controller.Run(done)\n\t<-done\n}", "func (s *HealthServer) Watch(in *healthpb.HealthCheckRequest, srv healthpb.Health_WatchServer) error {\n\treturn status.Error(codes.Unimplemented, \"Watch is not implemented\")\n}", "func (c *FakeImagePullJobs) Watch(ctx context.Context, opts v1.ListOptions) (watch.Interface, error) {\n\treturn c.Fake.\n\t\tInvokesWatch(testing.NewWatchAction(imagepulljobsResource, c.ns, opts))\n\n}", "func (m *MockJobClient) Watch(arg0 string, arg1 v10.ListOptions) (watch.Interface, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"Watch\", arg0, arg1)\n\tret0, _ := ret[0].(watch.Interface)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (c *solrBackups) Watch(opts v1.ListOptions) (watch.Interface, error) {\n\tvar timeout time.Duration\n\tif opts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*opts.TimeoutSeconds) * time.Second\n\t}\n\topts.Watch = true\n\treturn c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"solrbackups\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tWatch()\n}", "func (s *CAServer) Watch(_ *ghc.HealthCheckRequest, _ ghc.Health_WatchServer) error {\n\treturn nil\n}", "func (c *FakeTraefikServices) Watch(ctx context.Context, opts v1.ListOptions) (watch.Interface, error) {\n\treturn c.Fake.\n\t\tInvokesWatch(testing.NewWatchAction(traefikservicesResource, c.ns, opts))\n\n}", "func NewWatch() *Watch {\n\tr := &Watch{\n\t\tActions: make(map[string]WatcherAction, 0),\n\t}\n\n\treturn r\n}", "func (c *Client) Watch(gvr schema.GroupVersionResource, opts ...ListOption) (w watch.Interface, err error) {\n\trestClient, err := c.rest(gvr.GroupVersion())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tctx, cancel := context.WithTimeout(c.ctx, c.timeout)\n\tlistOpts := ListOptions{Raw: &metav1.ListOptions{Watch: true}}\n\tlistOpts.ApplyOptions(opts)\n\tw, err = restClient.Get().\n\t\tTimeout(c.timeout).\n\t\tNamespaceIfScoped(listOpts.Namespace, listOpts.Namespace != \"\").\n\t\tResource(gvr.Resource).\n\t\tVersionedParams(listOpts.AsListOptions(), scheme.ParameterCodec).\n\t\tWatch(ctx)\n\tif err != nil {\n\t\tcancel()\n\t\treturn nil, err\n\t}\n\treturn &watcher{Interface: w, cancel: cancel}, nil\n}", "func (mw *MultiWatcher) Watch(ctx context.Context) {\n\twg := sync.WaitGroup{}\n\twg.Add(len(mw.watchers))\n\tfor _, w := range mw.watchers {\n\t\tgo func(w *Watcher) {\n\t\t\tdefer wg.Done()\n\t\t\tw.Watch(ctx)\n\t\t}(w)\n\t}\n\twg.Wait()\n}", "func (c *kaosRules) Watch(opts meta_v1.ListOptions) (watch.Interface, error) {\n\topts.Watch = true\n\treturn c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"kaosrules\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tWatch()\n}", "func (c *nodes) Watch(opts api.ListOptions) (watch.Interface, error) {\n\treturn c.r.Get().\n\t\tPrefix(\"watch\").\n\t\tNamespace(api.NamespaceAll).\n\t\tResource(c.resourceName()).\n\t\tVersionedParams(&opts, api.ParameterCodec).\n\t\tWatch()\n}", "func (c *FakeRedisTriggers) Watch(ctx context.Context, opts v1.ListOptions) (watch.Interface, error) {\n\treturn c.Fake.\n\t\tInvokesWatch(testing.NewWatchAction(redistriggersResource, c.ns, opts))\n\n}", "func (c *FakeTraincrds) Watch(opts v1.ListOptions) (watch.Interface, error) {\n\treturn c.Fake.\n\t\tInvokesWatch(testing.NewWatchAction(traincrdsResource, c.ns, opts))\n\n}", "func (c *Client) Watch(ctx context.Context) {\n\tt := time.NewTicker(1 * time.Minute)\n\tfor {\n\t\tselect {\n\t\tcase <-t.C:\n\t\t\turls, err := clusterNodes(c.Endpoint)\n\t\t\tif err != nil {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tc.ServerList.SetServers(urls...)\n\t\tcase <-ctx.Done():\n\t\t\tt.Stop()\n\t\t\treturn\n\t\t}\n\t}\n}", "func (w *UniversalCheckWatcher) Watch(allocID, taskName, checkID string, check *structs.ServiceCheck, wr WorkloadRestarter) {\n\tif !check.TriggersRestarts() {\n\t\treturn // check_restart not set; no-op\n\t}\n\n\tc := &restarter{\n\t\tallocID: allocID,\n\t\ttaskName: taskName,\n\t\tcheckID: checkID,\n\t\tcheckName: check.Name,\n\t\ttaskKey: key(allocID + taskName),\n\t\ttask: wr,\n\t\tinterval: check.Interval,\n\t\tgrace: check.CheckRestart.Grace,\n\t\tgraceUntil: time.Now().Add(check.CheckRestart.Grace),\n\t\ttimeLimit: check.Interval * time.Duration(check.CheckRestart.Limit-1),\n\t\tignoreWarnings: check.CheckRestart.IgnoreWarnings,\n\t\tlogger: w.logger.With(\"alloc_id\", allocID, \"task\", taskName, \"check\", check.Name),\n\t}\n\n\tselect {\n\tcase w.checkUpdateCh <- checkWatchUpdate{\n\t\tcheckID: checkID,\n\t\trestart: c,\n\t}: // activate watch\n\tcase <-w.done: // exited; nothing to do\n\t}\n}", "func (c *FakeRobots) Watch(ctx context.Context, opts v1.ListOptions) (watch.Interface, error) {\n\treturn c.Fake.\n\t\tInvokesWatch(testing.NewWatchAction(robotsResource, c.ns, opts))\n\n}", "func (rci *redisClientImplementation) Watch(ctx context.Context, fn func(*redis.Tx) error, keys ...string) error {\n\treturn rci.client.Watch(ctx, fn, keys...)\n}", "func (c *backingservices) Watch(opts kapi.ListOptions) (watch.Interface, error) {\n\treturn c.r.Get().\n\t\tNamespace(c.ns).\n\t\tPrefix(\"watch\").\n\t\tResource(\"backingservices\").\n\t\tVersionedParams(&opts, kapi.ParameterCodec).\n\t\tWatch()\n}", "func (c *ClusterResourceClient) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) {\n\treturn c.clientCache.ClusterOrDie(logicalcluster.Wildcard).Resource(c.resource).Watch(ctx, opts)\n}", "func (s *Session) Watch(watchObject ...map[string]bool) {\n\tobjectString := \"\"\n\tif len(watchObject) == 1 {\n\t\tvar values []string\n\t\tfor k, v := range watchObject[0] {\n\t\t\tvalues = append(values, fmt.Sprintf(`\"%s\":%v`, k, v))\n\t\t}\n\t\tobjectString = fmt.Sprintf(`={%s}`, strings.Join(values, \",\"))\n\t}\n\ts.SendCommand(WatchCommand + objectString)\n}", "func (api *bucketAPI) Watch(handler BucketHandler) error {\n\tapi.ct.startWorkerPool(\"Bucket\")\n\treturn api.ct.WatchBucket(handler)\n}", "func (api *objectAPI) Watch(handler ObjectHandler) error {\n\tapi.ct.startWorkerPool(\"Object\")\n\treturn api.ct.WatchObject(handler)\n}", "func (w *Watcher) Watch() {\n\tch := make(chan struct{})\n\tgo func(stopCh <-chan struct{}) {\n\t\tw.Informer.Informer().AddEventHandler(w.ResourceEventHandlerFuncs)\n\t\tw.Informer.Informer().Run(stopCh)\n\t}(ch)\n\t<-w.StopChannel\n\tclose(ch)\n\tlogrus.Info(\"stoping watcher for \", w.GroupVersionResource)\n}", "func (c *krakenClusters) Watch(opts v1.ListOptions) (watch.Interface, error) {\n\topts.Watch = true\n\treturn c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"krakenclusters\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tWatch()\n}", "func (api *clusterAPI) Watch(handler ClusterHandler) error {\n\tapi.ct.startWorkerPool(\"Cluster\")\n\treturn api.ct.WatchCluster(handler)\n}", "func WatchCommand(args Args, done chan bool) {\n\tfor _, client := range args.ThemeClients {\n\t\tconfig := client.GetConfiguration()\n\t\tclient.Message(\"Spawning %d workers for %s\", config.Concurrency, args.Domain)\n\t\tassetEvents := client.NewFileWatcher(args.Directory, args.NotifyFile)\n\t\tfor i := 0; i < config.Concurrency; i++ {\n\t\t\tgo spawnWorker(assetEvents, client)\n\t\t\tclient.Message(\"%s Worker #%d ready to upload local changes\", config.Domain, i)\n\t\t}\n\t}\n}", "func (c *previewFeatures) Watch(ctx context.Context, opts v1.ListOptions) (watch.Interface, error) {\n\tvar timeout time.Duration\n\tif opts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*opts.TimeoutSeconds) * time.Second\n\t}\n\topts.Watch = true\n\treturn c.client.Get().\n\t\tResource(\"previewfeatures\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tWatch(ctx)\n}", "func (h *HealthImpl) Watch(*v1.HealthCheckRequest, v1.Health_WatchServer) error {\n\treturn nil\n}", "func (api *versionAPI) Watch(handler VersionHandler) error {\n\tapi.ct.startWorkerPool(\"Version\")\n\treturn api.ct.WatchVersion(handler)\n}", "func (c *externalInterfaces) Watch(opts metav1.ListOptions) (watch.Interface, error) {\n\tvar timeout time.Duration\n\tif opts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*opts.TimeoutSeconds) * time.Second\n\t}\n\topts.Watch = true\n\treturn c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"externalinterfaces\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tWatch()\n}", "func (rr *Registry) Watch(ctx context.Context) ([]*WatchEvent, <-chan *WatchEvent, error) {\n\trr.mu.Lock()\n\tdefer rr.mu.Unlock()\n\n\tprefix := rr.prefixPath()\n\n\tgetRes, err := rr.kv.Get(ctx, prefix, etcdv3.WithPrefix())\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tcurrentEvents := make([]*WatchEvent, 0, len(getRes.Kvs))\n\tfor _, kv := range getRes.Kvs {\n\t\treg, err := rr.unmarshalRegistration(kv)\n\t\tif err != nil {\n\t\t\treturn nil, nil, err\n\t\t}\n\t\twev := &WatchEvent{\n\t\t\tKey: string(kv.Key),\n\t\t\tReg: reg,\n\t\t\tType: Create,\n\t\t}\n\t\tcurrentEvents = append(currentEvents, wev)\n\t}\n\n\t// Channel to publish registry changes.\n\twatchEvents := make(chan *WatchEvent)\n\n\t// Write a change or exit the watcher.\n\tput := func(we *WatchEvent) {\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\treturn\n\t\tcase watchEvents <- we:\n\t\t}\n\t}\n\tputTerminalError := func(we *WatchEvent) {\n\t\tgo func() {\n\t\t\tdefer func() {\n\t\t\t\trecover()\n\t\t\t}()\n\t\t\tselect {\n\t\t\tcase <-time.After(10 * time.Minute):\n\t\t\tcase watchEvents <- we:\n\t\t\t}\n\t\t}()\n\t}\n\t// Create a watch-event from an event.\n\tcreateWatchEvent := func(ev *etcdv3.Event) *WatchEvent {\n\t\twev := &WatchEvent{Key: string(ev.Kv.Key)}\n\t\tif ev.IsCreate() {\n\t\t\twev.Type = Create\n\t\t} else if ev.IsModify() {\n\t\t\twev.Type = Modify\n\t\t} else {\n\t\t\twev.Type = Delete\n\t\t\t// Create base registration from just key.\n\t\t\treg := &Registration{}\n\t\t\tgraphType, graphName, err := rr.graphTypeAndNameFromKey(string(ev.Kv.Key))\n\t\t\tif err != nil {\n\t\t\t\tfmt.Println(err)\n\t\t\t}\n\t\t\treg.Type = graphType\n\t\t\treg.Name = graphName\n\t\t\twev.Reg = reg\n\t\t\t// Need to return now because\n\t\t\t// delete events don't contain\n\t\t\t// any data to unmarshal.\n\t\t\treturn wev\n\t\t}\n\t\treg, err := rr.unmarshalRegistration(ev.Kv)\n\t\tif err != nil {\n\t\t\twev.Error = fmt.Errorf(\"%v: failed unmarshaling value: '%s'\", err, ev.Kv.Value)\n\t\t} else {\n\t\t\twev.Reg = reg\n\t\t}\n\t\treturn wev\n\t}\n\n\t// Watch deltas in etcd, with the give prefix, starting\n\t// at the revision of the get call above.\n\tdeltas := rr.client.Watch(ctx, prefix, etcdv3.WithPrefix(), etcdv3.WithRev(getRes.Header.Revision+1))\n\tgo func() {\n\t\tdefer close(watchEvents)\n\t\tfor {\n\t\t\tdelta, open := <-deltas\n\t\t\tif !open {\n\t\t\t\tselect {\n\t\t\t\tcase <-ctx.Done():\n\t\t\t\tdefault:\n\t\t\t\t\tputTerminalError(&WatchEvent{Error: ErrWatchClosedUnexpectedly})\n\t\t\t\t}\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif delta.Err() != nil {\n\t\t\t\tputTerminalError(&WatchEvent{Error: delta.Err()})\n\t\t\t\treturn\n\t\t\t}\n\t\t\tfor _, event := range delta.Events {\n\t\t\t\tput(createWatchEvent(event))\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn currentEvents, watchEvents, nil\n}", "func (c *FakeQuobyteServices) Watch(opts v1.ListOptions) (watch.Interface, error) {\n\treturn c.Fake.\n\t\tInvokesWatch(testing.NewWatchAction(quobyteservicesResource, c.ns, opts))\n\n}", "func (c *gitTracks) Watch(opts v1.ListOptions) (watch.Interface, error) {\n\tvar timeout time.Duration\n\tif opts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*opts.TimeoutSeconds) * time.Second\n\t}\n\topts.Watch = true\n\treturn c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"gittracks\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tWatch()\n}", "func (m *MPD) Watch(ctx context.Context) Watch {\n\treturn goWatch(ctx, m.url)\n}", "func (s *Server) Watch(in *grpc_health_v1.HealthCheckRequest, server grpc_health_v1.Health_WatchServer) error {\n\tresp := &grpc_health_v1.HealthCheckResponse{Status: grpc_health_v1.HealthCheckResponse_SERVING}\n\treturn server.Send(resp)\n}", "func (r workloadEndpoints) Watch(ctx context.Context, opts options.ListOptions) (watch.Interface, error) {\r\n\treturn r.client.resources.Watch(ctx, opts, apiv3.KindWorkloadEndpoint, nil)\r\n}", "func (g *Gulf) Watch(patterns []string, tasks ...string) {}", "func (s *V3Backend) Watch(ctx context.Context, key string) <-chan *args.ChangeEvent {\n\twatchChan := s.Client.Watch(ctx, key, etcd.WithPrefix())\n\ts.changeChan = make(chan *args.ChangeEvent)\n\ts.done = make(chan struct{})\n\n\ts.wg.Add(1)\n\tgo func() {\n\t\tvar resp etcd.WatchResponse\n\t\tvar ok bool\n\t\tdefer s.wg.Done()\n\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase resp, ok = <-watchChan:\n\t\t\t\tif !ok {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tif resp.Canceled {\n\t\t\t\t\ts.changeChan <- NewChangeError(errors.Wrap(resp.Err(),\n\t\t\t\t\t\t\"V3Backend.Watch(): ETCD server cancelled watch\"))\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tfor _, event := range resp.Events {\n\t\t\t\t\ts.changeChan <- NewChangeEvent(event)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\treturn s.changeChan\n}", "func (s *Server) Watch(ctx context.Context, filePath string) (*topo.WatchData, <-chan *topo.WatchData, error) {\n\tlog.Info(\"Starting Kubernetes topo Watch on \", filePath)\n\n\tcurrent := &topo.WatchData{}\n\n\t// get current\n\tinitialCtx, initialCancel := context.WithTimeout(ctx, topo.RemoteOperationTimeout)\n\tdefer initialCancel()\n\n\tcontents, ver, err := s.Get(initialCtx, filePath)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tcurrent.Contents = contents\n\tcurrent.Version = ver\n\n\t// Create the changes channel\n\tchanges := make(chan *topo.WatchData, 10)\n\n\t// Create a signal channel for non-interrupt shutdowns\n\tgracefulShutdown := make(chan struct{})\n\n\tresource, err := s.buildFileResource(filePath, []byte{})\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\t// Create the informer / indexer to watch the single resource\n\trestClient := s.vtKubeClient.TopoV1beta1().RESTClient()\n\tlistwatch := cache.NewListWatchFromClient(restClient, \"vitesstoponodes\", s.namespace, fields.OneTermEqualSelector(\"metadata.name\", resource.Name))\n\n\t// set up index funcs\n\tindexers := cache.Indexers{}\n\tindexers[\"by_parent\"] = indexByParent\n\n\t_, memberInformer := cache.NewIndexerInformer(listwatch, &vtv1beta1.VitessTopoNode{}, 0,\n\t\tcache.ResourceEventHandlerFuncs{\n\t\t\tAddFunc: func(obj any) {\n\t\t\t\tvtn := obj.(*vtv1beta1.VitessTopoNode)\n\t\t\t\tout, err := unpackValue([]byte(vtn.Data.Value))\n\t\t\t\tif err != nil {\n\t\t\t\t\tchanges <- &topo.WatchData{Err: err}\n\t\t\t\t\tclose(gracefulShutdown)\n\t\t\t\t} else {\n\t\t\t\t\tchanges <- &topo.WatchData{\n\t\t\t\t\t\tContents: out,\n\t\t\t\t\t\tVersion: KubernetesVersion(vtn.GetResourceVersion()),\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t},\n\t\t\tUpdateFunc: func(oldObj, newObj any) {\n\t\t\t\tvtn := newObj.(*vtv1beta1.VitessTopoNode)\n\t\t\t\tout, err := unpackValue([]byte(vtn.Data.Value))\n\t\t\t\tif err != nil {\n\t\t\t\t\tchanges <- &topo.WatchData{Err: err}\n\t\t\t\t\tclose(gracefulShutdown)\n\t\t\t\t} else {\n\t\t\t\t\tchanges <- &topo.WatchData{\n\t\t\t\t\t\tContents: out,\n\t\t\t\t\t\tVersion: KubernetesVersion(vtn.GetResourceVersion()),\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t},\n\t\t\tDeleteFunc: func(obj any) {\n\t\t\t\tvtn := obj.(*vtv1beta1.VitessTopoNode)\n\t\t\t\tchanges <- &topo.WatchData{Err: topo.NewError(topo.NoNode, vtn.Name)}\n\t\t\t\tclose(gracefulShutdown)\n\t\t\t},\n\t\t}, indexers)\n\n\t// create control chan for informer and start it\n\tinformerChan := make(chan struct{})\n\tgo memberInformer.Run(informerChan)\n\n\t// Handle interrupts\n\tgo closeOnDone(ctx, filePath, informerChan, gracefulShutdown, changes)\n\n\treturn current, changes, nil\n}", "func (f Forwarder) Watch() {\n\thandler := func(i interface{}) {\n\t\tvar container string\n\t\tvar message string\n\t\tvar context map[string]interface{}\n\t\tdata := i.(map[string]interface{})\n\t\tmetadata := data[\"metadata\"].(map[string]interface{})\n\n\t\ttmp, ok := metadata[\"context\"]\n\t\tif ok {\n\t\t\tcontext = tmp.(map[string]interface{})\n\t\t}\n\n\t\ttmp, ok = context[\"container\"]\n\t\tif ok {\n\t\t\tcontainer = tmp.(string)\n\t\t}\n\n\t\t_, ok = f.Forwards[container]\n\t\tif ok {\n\t\t\ttmp, ok := metadata[\"message\"]\n\t\t\tif ok {\n\t\t\t\tmessage = tmp.(string)\n\t\t\t}\n\t\t\tswitch message {\n\t\t\tcase ContainerStarted:\n\t\t\t\tgo func() {\n\t\t\t\t\t// Wait a few seconds for the newly running container to get an IP address\n\t\t\t\t\ttime.Sleep(2 * time.Second)\n\t\t\t\t\tf.ForwardContainer(container)\n\t\t\t\t}()\n\n\t\t\tcase ContainerStopped:\n\t\t\t\tf.ReverseContainer(container)\n\t\t\t}\n\t\t}\n\n\t}\n\n\tf.Monitor([]string{}, handler)\n}", "func (r *Reloader) Watch(ctx context.Context) error {\n\tif err := r.apply(ctx); err != nil {\n\t\treturn err\n\t}\n\n\ttick := time.NewTicker(r.watchInterval)\n\tdefer tick.Stop()\n\n\tfor {\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\treturn nil\n\t\tcase <-tick.C:\n\t\t}\n\n\t\tif err := r.apply(ctx); err != nil {\n\t\t\tklog.Error(err)\n\t\t}\n\t}\n}", "func (s *svc) Watch(opts ...router.WatchOption) (router.Watcher, error) {\n\trsp, err := s.router.Watch(context.Background(), &pb.WatchRequest{}, s.callOpts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\toptions := router.WatchOptions{\n\t\tService: \"*\",\n\t}\n\tfor _, o := range opts {\n\t\to(&options)\n\t}\n\treturn newWatcher(rsp, options)\n}", "func (kvclient *MockResKVClient) Watch(ctx context.Context, key string, withPrefix bool) chan *kvstore.Event {\n\treturn nil\n}", "func Watch(ctx context.Context, watcher watch.Interface) (chan *Target, chan *Target, chan *Target) {\n\tadded := make(chan *Target)\n\tfinished := make(chan *Target)\n\tdeleted := make(chan *Target)\n\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase e := <-watcher.ResultChan():\n\t\t\t\tif e.Object == nil {\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\tpod := e.Object.(*v1.Pod)\n\n\t\t\t\tswitch e.Type {\n\t\t\t\tcase watch.Added:\n\t\t\t\t\tif pod.Status.Phase != v1.PodRunning {\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\n\t\t\t\t\tfor _, container := range pod.Spec.Containers {\n\t\t\t\t\t\tadded <- NewTarget(pod.Namespace, pod.Name, container.Name)\n\t\t\t\t\t}\n\t\t\t\tcase watch.Modified:\n\t\t\t\t\tswitch pod.Status.Phase {\n\t\t\t\t\tcase v1.PodRunning:\n\t\t\t\t\t\tfor _, container := range pod.Spec.Containers {\n\t\t\t\t\t\t\tadded <- NewTarget(pod.Namespace, pod.Name, container.Name)\n\t\t\t\t\t\t}\n\t\t\t\t\tcase v1.PodSucceeded, v1.PodFailed:\n\t\t\t\t\t\tfor _, container := range pod.Spec.Containers {\n\t\t\t\t\t\t\tfinished <- NewTarget(pod.Namespace, pod.Name, container.Name)\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\tcase watch.Deleted:\n\t\t\t\t\tfor _, container := range pod.Spec.Containers {\n\t\t\t\t\t\tdeleted <- NewTarget(pod.Namespace, pod.Name, container.Name)\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\tcase <-ctx.Done():\n\t\t\t\twatcher.Stop()\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn added, finished, deleted\n}", "func (h *HealthImpl) Watch(in *grpc_health_v1.HealthCheckRequest, stream grpc_health_v1.Health_WatchServer) error {\n\treturn nil\n}", "func (s *Service) Watch(ctx context.Context, chConfig *config.CheckerConfig) {\n\tsourcesFile, err := os.Open(chConfig.Source())\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdefer sourcesFile.Close()\n\n\tfor i := 0; i < cap(s.workerPool); i++ {\n\t\tworker := NewWorker(s.store, s.workerPool)\n\t\tworker.Start(ctx)\n\t}\n\n\tvar parallelRun int\n\tscanner := bufio.NewScanner(sourcesFile)\n\tfor scanner.Scan() && parallelRun < parallelRunMaxQty {\n\t\ts.spawnCheck(ctx, scanner.Text(), chConfig.Interval())\n\t\tparallelRun++\n\t}\n\n\tif err := scanner.Err(); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tselect {\n\tcase <-ctx.Done():\n\t\tfmt.Println(\"Service shutdown.\")\n\t\treturn\n\t}\n}", "func (rbw *roleBindingWatcher) Watch() {\n\tstop := make(chan struct{})\n\tdefer close(stop)\n\tgo rbw.watchController.Run(1, stop)\n\t<-rbw.quit\n}", "func (w *ClusterDynamicClient) Watch(opts metav1.ListOptions) (watch.Interface, error) {\n\treturn w.dClient.Resource(w.resource).Namespace(w.namespace).Watch(w.ctx, opts)\n}", "func (m *manager) Watch(addr string) {\n\tm.watch.Watch(addr)\n}", "func (api *hostAPI) Watch(handler HostHandler) error {\n\tapi.ct.startWorkerPool(\"Host\")\n\treturn api.ct.WatchHost(handler)\n}", "func (c *customReplicationControllers) Watch(opts v1.ListOptions) (watch.Interface, error) {\n\topts.Watch = true\n\treturn c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"customreplicationcontrollers\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tWatch()\n}", "func (srv *HealthServer) Watch(*grpc_health_v1.HealthCheckRequest, grpc_health_v1.Health_WatchServer) error {\n\treturn nil\n}", "func (f *extendedPodFactory) ListWatch(customResourceClient interface{}, ns string, fieldSelector string) cache.ListerWatcher {\n\tclient := customResourceClient.(clientset.Interface)\n\treturn &cache.ListWatch{\n\t\tListFunc: func(opts metav1.ListOptions) (runtime.Object, error) {\n\t\t\treturn client.CoreV1().Pods(ns).List(context.TODO(), opts)\n\t\t},\n\t\tWatchFunc: func(opts metav1.ListOptions) (watch.Interface, error) {\n\t\t\treturn client.CoreV1().Pods(ns).Watch(context.TODO(), opts)\n\t\t},\n\t}\n}", "func (nDB *NetworkDB) Watch(tname, nid string) (*events.Channel, func()) {\n\tvar matcher events.Matcher\n\n\tif tname != \"\" || nid != \"\" {\n\t\tmatcher = events.MatcherFunc(func(ev events.Event) bool {\n\t\t\tvar evt event\n\t\t\tswitch ev := ev.(type) {\n\t\t\tcase CreateEvent:\n\t\t\t\tevt = event(ev)\n\t\t\tcase UpdateEvent:\n\t\t\t\tevt = event(ev)\n\t\t\tcase DeleteEvent:\n\t\t\t\tevt = event(ev)\n\t\t\t}\n\n\t\t\tif tname != \"\" && evt.Table != tname {\n\t\t\t\treturn false\n\t\t\t}\n\n\t\t\tif nid != \"\" && evt.NetworkID != nid {\n\t\t\t\treturn false\n\t\t\t}\n\n\t\t\treturn true\n\t\t})\n\t}\n\n\tch := events.NewChannel(0)\n\tsink := events.Sink(events.NewQueue(ch))\n\n\tif matcher != nil {\n\t\tsink = events.NewFilter(sink, matcher)\n\t}\n\n\tnDB.broadcaster.Add(sink)\n\treturn ch, func() {\n\t\tnDB.broadcaster.Remove(sink)\n\t\tch.Close()\n\t\tsink.Close()\n\t}\n}", "func Watch(ctx context.Context, cliEngine *engine.Engine, task string, t ox.Task) {\n\ttaskCtx, cancel := context.WithCancel(ctx)\n\n\tfiles, err := getWatcherFiles(t.Sources, t.Dir)\n\tif err != nil {\n\t\tutils.PrintError(err)\n\t\treturn\n\t}\n\n\twatcher, err := fsnotify.NewWatcher()\n\tif err != nil {\n\t\tutils.PrintError(err)\n\t\treturn\n\t}\n\tdefer watcher.Close()\n\n\tfor _, file := range files {\n\t\terr = watcher.Add(file)\n\t\tif err != nil {\n\t\t\tutils.PrintError(err)\n\t\t\treturn\n\t\t}\n\t}\n\n\trunOnWatch := func() {\n\t\terr := cliEngine.Run(taskCtx, task)\n\t\tif err != nil {\n\t\t\tutils.PrintError(err)\n\t\t}\n\t}\n\n\tgo runOnWatch()\n\n\tfor {\n\t\tselect {\n\t\tcase event := <-watcher.Events:\n\t\t\tswitch {\n\t\t\tcase event.Op&fsnotify.Write == fsnotify.Write:\n\t\t\t\tfallthrough\n\t\t\tcase event.Op&fsnotify.Create == fsnotify.Create:\n\t\t\t\tfallthrough\n\t\t\tcase event.Op&fsnotify.Remove == fsnotify.Remove:\n\t\t\t\tfallthrough\n\t\t\tcase event.Op&fsnotify.Rename == fsnotify.Rename:\n\t\t\t\tcancel()\n\t\t\t\ttaskCtx, cancel = context.WithCancel(ctx)\n\t\t\t\tgo runOnWatch()\n\t\t\t}\n\t\tcase <-ctx.Done():\n\t\t\tcancel()\n\t\t\treturn\n\t\tcase err := <-watcher.Errors:\n\t\t\tutils.PrintError(err)\n\t\t\tcancel()\n\t\t\treturn\n\t\t}\n\t}\n}", "func (c *configAuditReports) Watch(ctx context.Context, opts v1.ListOptions) (watch.Interface, error) {\n\tvar timeout time.Duration\n\tif opts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*opts.TimeoutSeconds) * time.Second\n\t}\n\topts.Watch = true\n\treturn c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"configauditreports\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tWatch(ctx)\n}", "func (api *tenantAPI) Watch(handler TenantHandler) error {\n\tapi.ct.startWorkerPool(\"Tenant\")\n\treturn api.ct.WatchTenant(handler)\n}", "func (w *watcher) Watch() {\n\tfor {\n\t\tfor watchPath := range w.watchItems {\n\t\t\tfileChanged, err := w.scanChange(watchPath)\n\t\t\tif err != nil {\n\t\t\t\tw.errors <- err\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tif fileChanged != \"\" {\n\t\t\t\tw.events <- fileChanged\n\t\t\t\tstartTime = time.Now()\n\t\t\t}\n\t\t}\n\n\t\ttime.Sleep(time.Duration(w.pollInterval) * time.Millisecond)\n\t}\n}", "func (c *meshPolicies) Watch(opts v1.ListOptions) (watch.Interface, error) {\n\tvar timeout time.Duration\n\tif opts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*opts.TimeoutSeconds) * time.Second\n\t}\n\topts.Watch = true\n\treturn c.client.Get().\n\t\tResource(\"meshpolicies\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tWatch()\n}", "func (api *nodeAPI) Watch(handler NodeHandler) error {\n\tapi.ct.startWorkerPool(\"Node\")\n\treturn api.ct.WatchNode(handler)\n}" ]
[ "0.74233603", "0.73256457", "0.6655384", "0.6595689", "0.65147346", "0.6470929", "0.6460171", "0.6263511", "0.6231087", "0.6198213", "0.6154817", "0.61261344", "0.609842", "0.6087547", "0.6079176", "0.60721064", "0.6070674", "0.6048015", "0.6023506", "0.5992892", "0.59888357", "0.5984638", "0.59837174", "0.5978222", "0.5967192", "0.59362453", "0.5935091", "0.59318316", "0.5918817", "0.5913833", "0.59060955", "0.59042853", "0.5902468", "0.58984804", "0.58912706", "0.58807826", "0.5871967", "0.5864701", "0.5860408", "0.58578014", "0.58546156", "0.5848959", "0.58471096", "0.58297294", "0.58200437", "0.5805387", "0.58040345", "0.58013403", "0.57941645", "0.57937557", "0.5788745", "0.5770938", "0.57551074", "0.57523227", "0.57465774", "0.57320297", "0.57277507", "0.5726726", "0.57246125", "0.57114285", "0.56937915", "0.56866974", "0.56835", "0.56765604", "0.5670506", "0.5663248", "0.5652707", "0.56510425", "0.5649773", "0.56408423", "0.56352985", "0.5629582", "0.56287026", "0.5620038", "0.56185377", "0.5607224", "0.5605088", "0.5589491", "0.5582266", "0.55783486", "0.55759984", "0.55729103", "0.55681646", "0.556439", "0.5564233", "0.5563272", "0.556322", "0.55566156", "0.55529475", "0.5541878", "0.5541754", "0.55272245", "0.55196166", "0.5518036", "0.55009747", "0.5499096", "0.5487914", "0.5485986", "0.5484713", "0.5481674" ]
0.73646355
1
Create takes the representation of a tZCronJob and creates it. Returns the server's representation of the tZCronJob, and an error, if there is any.
func (c *FakeTZCronJobs) Create(tZCronJob *v1alpha1.TZCronJob) (result *v1alpha1.TZCronJob, err error) { obj, err := c.Fake. Invokes(testing.NewCreateAction(tzcronjobsResource, c.ns, tZCronJob), &v1alpha1.TZCronJob{}) if obj == nil { return nil, err } return obj.(*v1alpha1.TZCronJob), err }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *Jobs) CreateJob(in *JobIn) (out *JobOut, err error) {\n\tbody, err := c.call(\"POST\", \"/jobs/\", in)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer body.Close()\n\n\terr = json.NewDecoder(body).Decode(&out)\n\treturn\n}", "func (c *controller) CreateJob(namespace string, job Job) (*apibatchv1.Job, error) {\n\tj := job.Build()\n\treturn c.k8sBatchClient.Jobs(namespace).Create(j)\n}", "func CreateJob(ctx context.Context, p string) (j *Job, err error) {\n\tt := utils.FromTaskContext(ctx)\n\n\tj = &Job{\n\t\tPath: p,\n\t\tMarker: \"\",\n\t}\n\n\tcontent, err := msgpack.Marshal(j)\n\tif err != nil {\n\t\treturn\n\t}\n\n\terr = contexts.DB.Put(constants.FormatJobKey(t, j.Path), content, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func (c *Controller) CreateJob(yaml []byte) (weles.JobID, error) {\n\tj, err := c.jobs.NewJob(yaml)\n\tif err != nil {\n\t\treturn weles.JobID(0), err\n\t}\n\n\tgo c.parser.Parse(j)\n\n\treturn j, nil\n}", "func (c *PlatformGraphQLClient) CreateJob(ctx context.Context, tdoID string, isReprocessJob bool, tasks ...CreateJobTask) (*Job, error) {\n\treq := graphql.NewRequest(`\n\t\tmutation(\n\t\t\t$targetId: ID\n\t\t\t$isReprocessJob: Boolean\n\t\t\t$tasks: [CreateTask!]\n\t\t) {\n\t\t\tcreateJob(input: {\n\t\t\t\ttargetId: $targetId\n\t\t\t\tisReprocessJob: $isReprocessJob\n\t\t\t\ttasks: $tasks\n\t\t\t}) {\n\t\t\t\tid\n\t\t\t\tname\n\t\t\t\ttargetId\n\t\t\t\tstatus\n\t\t\t\ttasks {\n\t\t\t\t\trecords {\n\t\t\t\t\t\tid\n\t\t\t\t\t\tstatus\n\t\t\t\t\t\tengine {\n\t\t\t\t\t\t\tid\n\t\t\t\t\t\t\tname\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t`)\n\n\treq.Var(\"targetId\", tdoID)\n\treq.Var(\"isReprocessJob\", isReprocessJob)\n\treq.Var(\"tasks\", tasks)\n\n\tvar resp struct {\n\t\tResult *Job `json:\"createJob\"`\n\t}\n\treturn resp.Result, c.Run(ctx, req, &resp)\n}", "func (c *cloudSchedulerRESTClient) CreateJob(ctx context.Context, req *schedulerpb.CreateJobRequest, opts ...gax.CallOption) (*schedulerpb.Job, error) {\n\tm := protojson.MarshalOptions{AllowPartial: true, UseEnumNumbers: true}\n\tbody := req.GetJob()\n\tjsonReq, err := m.Marshal(body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v1/%v/jobs\", req.GetParent())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"parent\", url.QueryEscape(req.GetParent()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\topts = append((*c.CallOptions).CreateJob[0:len((*c.CallOptions).CreateJob):len((*c.CallOptions).CreateJob)], opts...)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tresp := &schedulerpb.Job{}\n\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"POST\", baseUrl.String(), bytes.NewReader(jsonReq))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}, opts...)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\treturn resp, nil\n}", "func (c *Client) CreateNewJob(job *Job) error {\n\tbody, err := json.Marshal(job)\n\tif err != nil {\n\t\tlogrus.WithError(err)\n\t\treturn err\n\t}\n\n\turl := fmt.Sprintf(\"http://%s/api/v1/jobs\", c.options.ServerAddr)\n\tresp, err := http.Post(url, \"application/json\", bytes.NewReader(body))\n\tif err != nil {\n\t\tlogrus.WithError(err)\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\n\tbody, err = ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\tlogrus.WithError(err)\n\t}\n\treturn err\n}", "func (s *JobService) Create(ctx context.Context, bij *api_types.InferenceJob) (err error) {\n\n\tbij.CreatedAt = time.Now().UTC()\n\tbij.UpdatedAt = time.Now().UTC()\n\n\tif errs := ValidateJobInput(*bij); len(errs) > 0 {\n\t\treturn odahuErrors.InvalidEntityError{\n\t\t\tEntity: bij.ID,\n\t\t\tValidationErrors: errs,\n\t\t}\n\t}\n\n\tservice, err := s.sRepo.Get(ctx, nil, bij.Spec.InferenceServiceID)\n\tif err != nil {\n\t\tif odahuErrors.IsNotFoundError(err) {\n\t\t\treturn odahuErrors.InvalidEntityError{\n\t\t\t\tEntity: \"job\",\n\t\t\t\tValidationErrors: []error{fmt.Errorf(\"unable to fetch corresponding service: %s\", err)},\n\t\t\t}\n\t\t}\n\t\treturn err\n\t}\n\n\tDefaultJob(bij, service)\n\n\terrs, err := ValidateJob(*bij, s.connGetter, service)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif len(errs) > 0 {\n\t\treturn odahuErrors.InvalidEntityError{\n\t\t\tEntity: bij.ID,\n\t\t\tValidationErrors: errs,\n\t\t}\n\t}\n\n\terr = s.repo.Create(ctx, nil, *bij)\n\n\treturn err\n}", "func (c *scheduledJobs) Create(job *batch.ScheduledJob) (result *batch.ScheduledJob, err error) {\n\tresult = &batch.ScheduledJob{}\n\terr = c.r.Post().Namespace(c.ns).Resource(\"scheduledjobs\").Body(job).Do().Into(result)\n\treturn\n}", "func (p *AuroraSchedulerManagerClient) CreateJob(ctx context.Context, description *JobConfiguration) (r *Response, err error) {\n var _args182 AuroraSchedulerManagerCreateJobArgs\n _args182.Description = description\n var _result183 AuroraSchedulerManagerCreateJobResult\n var meta thrift.ResponseMeta\n meta, err = p.Client_().Call(ctx, \"createJob\", &_args182, &_result183)\n p.SetLastResponseMeta_(meta)\n if err != nil {\n return\n }\n return _result183.GetSuccess(), nil\n}", "func (r *realKubeClient) CreateJob(job *Job) error {\n\turl := \"/apis/extensions/v1beta1/namespaces/\" + job.Metadata[\"namespace\"].(string) + \"/jobs\"\n\tdata, err := json.Marshal(job)\n\tif err != nil {\n\t\treturn err\n\t}\n\tbyteData := bytes.NewReader(data)\n\treturn r.doPost(url, byteData)\n\n}", "func (c *jobsRESTClient) CreateJob(ctx context.Context, req *runpb.CreateJobRequest, opts ...gax.CallOption) (*CreateJobOperation, error) {\n\tm := protojson.MarshalOptions{AllowPartial: true, UseEnumNumbers: true}\n\tbody := req.GetJob()\n\tjsonReq, err := m.Marshal(body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v2/%v/jobs\", req.GetParent())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\tparams.Add(\"jobId\", fmt.Sprintf(\"%v\", req.GetJobId()))\n\tif req.GetValidateOnly() {\n\t\tparams.Add(\"validateOnly\", fmt.Sprintf(\"%v\", req.GetValidateOnly()))\n\t}\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"parent\", url.QueryEscape(req.GetParent()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tresp := &longrunningpb.Operation{}\n\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"POST\", baseUrl.String(), bytes.NewReader(jsonReq))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}, opts...)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\n\toverride := fmt.Sprintf(\"/v2/%s\", resp.GetName())\n\treturn &CreateJobOperation{\n\t\tlro: longrunning.InternalNewOperation(*c.LROClient, resp),\n\t\tpollPath: override,\n\t}, nil\n}", "func CreateJob(ctx *gin.Context) {\n\tlog := logger.RuntimeLog\n\tvar jobModel *model.Job\n\tif err := ctx.BindJSON(&jobModel); err != nil {\n\t\tSendResponse(ctx, err, \"Request Body Invalid\")\n\t}\n\n\tjobNamespace := strings.ToLower(jobModel.JobMeta.AppMeta.Namespace)\n\tjobName := jobModel.JobMeta.AppMeta.Name\n\tzoneName := jobModel.JobMeta.AppMeta.ZoneName\n\n\t// fetch k8s-client handler by zoneName\n\tkclient, err := GetClientByAzCode(zoneName)\n\tif err != nil {\n\t\tlog.WithError(err)\n\t\tSendResponse(ctx, errno.ErrTokenInvalid, nil)\n\t}\n\n\tstartAt := time.Now()\n\t_, err = kclient.BatchV1().Jobs(jobNamespace).Create(makeupJobData(ctx, jobModel))\n\tif err != nil {\n\t\tSendResponse(ctx, err, \"create deployment fail.\")\n\t\treturn\n\t}\n\tlogger.MetricsEmit(\n\t\tJOB_CONST.K8S_LOG_Method_CreateJob,\n\t\tutil.GetReqID(ctx),\n\t\tfloat32(time.Since(startAt)/time.Millisecond),\n\t\terr == err,\n\t)\n\tSendResponse(ctx, errno.OK, fmt.Sprintf(\"Create Job %s success.\", jobName))\n\treturn\n}", "func (c *JobClient) Create() *JobCreate {\n\tmutation := newJobMutation(c.config, OpCreate)\n\treturn &JobCreate{config: c.config, hooks: c.Hooks(), mutation: mutation}\n}", "func (j *JobWorker) CreateJob(request CreateJobRequest) (Job, error) {\n\t// combines stderr and stdout of the job process\n\tvar outputBuffer Buffer\n\n\tcmd := exec.Command(request.Command, request.Args...)\n\tcmd.Stdout = &outputBuffer\n\tcmd.Stderr = &outputBuffer\n\n\t// no need to save the job if it fails to start\n\tif err := cmd.Start(); err != nil {\n\t\terr = fmt.Errorf(\"job failed to start, reason: %v\", err)\n\t\tlog.Errorf(err.Error())\n\t\treturn Job{}, err\n\t}\n\n\t// job started, create job object, assign ID and save the job\n\tjob := Job{\n\t\tID: getNextJobID(),\n\t\tRequest: request,\n\t\tStatus: JobRunning,\n\t\tcmd: cmd,\n\t\toutputBuffer: &outputBuffer,\n\t\twaitForExit: &sync.WaitGroup{},\n\t\tstatusLock: &sync.RWMutex{},\n\t}\n\n\tlog.Infof(\"job %v: started\", job.ID)\n\n\t// save the job in memory for future reference\n\tj.store(job.ID, &job)\n\n\t// wait for job to finish\n\tgo handleFinish(&job)\n\n\treturn job, nil\n}", "func CreateJob(request CreateJobRequest) (CreateJobResponse, error) {\n\tjob, err := entities.NewJob(\n\t\trequest.Name,\n\t\trequest.Command,\n\t\trequest.Tick,\n\t\ttrue,\n\t)\n\tif err != nil {\n\t\treturn CreateJobResponse{}, err\n\t}\n\terr = writer.CreateJob(&job)\n\tif err != nil {\n\t\treturn CreateJobResponse{}, err\n\t}\n\treturn CreateJobResponse{Msg: job.Name + \" created\"}, nil\n}", "func Create(ctx context.Context, options *Options) (_ *JobObject, err error) {\n\tif options == nil {\n\t\toptions = &Options{}\n\t}\n\n\tvar jobName *winapi.UnicodeString\n\tif options.Name != \"\" {\n\t\tjobName, err = winapi.NewUnicodeString(options.Name)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tvar jobHandle windows.Handle\n\tif options.UseNTVariant {\n\t\toa := winapi.ObjectAttributes{\n\t\t\tLength: unsafe.Sizeof(winapi.ObjectAttributes{}),\n\t\t\tObjectName: jobName,\n\t\t\tAttributes: 0,\n\t\t}\n\t\tstatus := winapi.NtCreateJobObject(&jobHandle, winapi.JOB_OBJECT_ALL_ACCESS, &oa)\n\t\tif status != 0 {\n\t\t\treturn nil, winapi.RtlNtStatusToDosError(status)\n\t\t}\n\t} else {\n\t\tvar jobNameBuf *uint16\n\t\tif jobName != nil && jobName.Buffer != nil {\n\t\t\tjobNameBuf = jobName.Buffer\n\t\t}\n\t\tjobHandle, err = windows.CreateJobObject(nil, jobNameBuf)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tdefer func() {\n\t\tif err != nil {\n\t\t\twindows.Close(jobHandle)\n\t\t}\n\t}()\n\n\tjob := &JobObject{\n\t\thandle: jobHandle,\n\t}\n\n\t// If the IOCP we'll be using to receive messages for all jobs hasn't been\n\t// created, create it and start polling.\n\tif options.Notifications {\n\t\tmq, err := setupNotifications(ctx, job)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tjob.mq = mq\n\t}\n\n\tif options.EnableIOTracking {\n\t\tif err := enableIOTracking(jobHandle); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn job, nil\n}", "func (n *Namespace) CreateJob(backend Backend, job *Job) error {\n\n\t// ensure the namespace exists\n\terr := n.checkOrCreateNamespace(backend)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// get JSON for Job\n\tjson, err := job.Serialize()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// create job, overwriting existing\n\terr = backend.WriteKey(getJobPath(n.namespace, job.ID), string(json), false, etcd.PrevNoExist, 0)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"problem creating job %v\", err)\n\t}\n\n\treturn nil\n}", "func (client JobClient) Create(ctx context.Context, resourceGroupName string, automationAccountName string, jobName string, parameters JobCreateParameters, clientRequestID string) (result Job, err error) {\n\tif tracing.IsEnabled() {\n\t\tctx = tracing.StartSpan(ctx, fqdn+\"/JobClient.Create\")\n\t\tdefer func() {\n\t\t\tsc := -1\n\t\t\tif result.Response.Response != nil {\n\t\t\t\tsc = result.Response.Response.StatusCode\n\t\t\t}\n\t\t\ttracing.EndSpan(ctx, sc, err)\n\t\t}()\n\t}\n\tif err := validation.Validate([]validation.Validation{\n\t\t{TargetValue: resourceGroupName,\n\t\t\tConstraints: []validation.Constraint{{Target: \"resourceGroupName\", Name: validation.MaxLength, Rule: 90, Chain: nil},\n\t\t\t\t{Target: \"resourceGroupName\", Name: validation.MinLength, Rule: 1, Chain: nil},\n\t\t\t\t{Target: \"resourceGroupName\", Name: validation.Pattern, Rule: `^[-\\w\\._]+$`, Chain: nil}}},\n\t\t{TargetValue: parameters,\n\t\t\tConstraints: []validation.Constraint{{Target: \"parameters.JobCreateProperties\", Name: validation.Null, Rule: true, Chain: nil}}}}); err != nil {\n\t\treturn result, validation.NewError(\"automation.JobClient\", \"Create\", err.Error())\n\t}\n\n\treq, err := client.CreatePreparer(ctx, resourceGroupName, automationAccountName, jobName, parameters, clientRequestID)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"automation.JobClient\", \"Create\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresp, err := client.CreateSender(req)\n\tif err != nil {\n\t\tresult.Response = autorest.Response{Response: resp}\n\t\terr = autorest.NewErrorWithError(err, \"automation.JobClient\", \"Create\", resp, \"Failure sending request\")\n\t\treturn\n\t}\n\n\tresult, err = client.CreateResponder(resp)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"automation.JobClient\", \"Create\", resp, \"Failure responding to request\")\n\t\treturn\n\t}\n\n\treturn\n}", "func (p *AuroraSchedulerManagerClient) CreateJob(ctx context.Context, description *JobConfiguration) (r *Response, err error) {\n var _args132 AuroraSchedulerManagerCreateJobArgs\n _args132.Description = description\n var _result133 AuroraSchedulerManagerCreateJobResult\n if err = p.Client_().Call(ctx, \"createJob\", &_args132, &_result133); err != nil {\n return\n }\n return _result133.GetSuccess(), nil\n}", "func (s *ContinuousRestoreJobsServiceOp) Create(ctx context.Context, groupID, clusterID string, request *atlas.ContinuousJobRequest) (*atlas.ContinuousJobs, *Response, error) {\n\tif request == nil {\n\t\treturn nil, nil, atlas.NewArgError(\"request\", \"must be set\")\n\t}\n\tif clusterID == \"\" {\n\t\treturn nil, nil, atlas.NewArgError(\"clusterID\", \"must be set\")\n\t}\n\tif groupID == \"\" {\n\t\treturn nil, nil, atlas.NewArgError(\"groupID\", \"must be set\")\n\t}\n\n\tpath := fmt.Sprintf(continuousRestoreJobsPath, groupID, clusterID)\n\n\treq, err := s.Client.NewRequest(ctx, http.MethodPost, path, request)\n\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\troot := new(atlas.ContinuousJobs)\n\tresp, err := s.Client.Do(ctx, req, root)\n\n\treturn root, resp, err\n}", "func CreateJob(name string) *batchv1.Job {\n\tj := &batchv1.Job{}\n\tj.APIVersion = \"batch/v1\"\n\tj.Kind = \"Job\"\n\tj.Name = name\n\tj.Spec.Template.Name = name\n\tj.Spec.Template.Spec.RestartPolicy = \"Never\"\n\treturn j\n}", "func (r *jobs) Create(txID string, job *dbmodels.Job) (*dbmodels.Job, error) {\n\trc := r.db.Database().Collection(viper.GetString(\"db.jobs_collection\"))\n\tctx, cancel := context.WithTimeout(\n\t\tcontext.Background(),\n\t\ttime.Duration(viper.GetInt(\"db.query_timeout_in_sec\"))*time.Second,\n\t)\n\tdefer cancel()\n\n\tinsertResult, err := rc.InsertOne(ctx, job)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to create organization with error %v\", err)\n\t}\n\n\tjob.ID = insertResult.InsertedID.(primitive.ObjectID)\n\treturn job, nil\n}", "func (cl *Client) CreateJob(args CreateJobArgs) error {\n\tif args.CallTime == \"\" {\n\t\treturn fmt.Errorf(\"[Room %d] no call time provided\", args.RoomNumber)\n\t}\n\tparts := strings.Split(args.CallTime, \":\")\n\tif len(parts) != 2 {\n\t\treturn fmt.Errorf(\"[Room %d] wake up time (%s) must be in the following format: hh:mm\", args.RoomNumber, args.CallTime)\n\t}\n\t// Parse wake up time hour and min\n\tcallHour, err := strconv.Atoi(parts[0])\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"[Room %d] wake up hour (%s) in wake up time (%s) is not a number\", args.RoomNumber, parts[0], args.CallTime)\n\t}\n\tcallMin, err := strconv.Atoi(parts[1])\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"[Room %d] wake up min (%s) in wake up time (%s) is not a number\", args.RoomNumber, parts[1], args.CallTime)\n\t}\n\tif args.CallURL == \"\" {\n\t\treturn fmt.Errorf(\"[Room %d] No call URL provided\", args.RoomNumber)\n\t}\n\t// Prepare scheduler job name\n\tjobID := args.SchedulerLocation + \"/jobs/call_at_\" + args.CallTime\n\t// Make sure scheduler job doesn't exist for provided wake up time\n\texistingJob, err := cl.GetJob(jobID)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// Job already exists - do nothing\n\tif existingJob != nil {\n\t\treturn nil\n\t}\n\t// Prepare schedule to call once a day at provided time\n\tschedule := fmt.Sprintf(\"%d %d * * *\", callMin, callHour)\n\treq := &schedulerpb.CreateJobRequest{\n\t\tParent: args.SchedulerLocation,\n\t\tJob: &schedulerpb.Job{\n\t\t\tName: removeColon(jobID),\n\t\t\tTarget: &schedulerpb.Job_HttpTarget{\n\t\t\t\tHttpTarget: &schedulerpb.HttpTarget{\n\t\t\t\t\tUri: args.CallURL,\n\t\t\t\t\tHttpMethod: schedulerpb.HttpMethod_GET,\n\t\t\t\t},\n\t\t\t},\n\t\t\tTimeZone: args.SchedulerTimezone,\n\t\t\tSchedule: schedule,\n\t\t},\n\t}\n\tresp, err := cl.CloudSchedulerClient.CreateJob(cl.ctx, req)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"CreateJob->cl.CloudSchedulerClient.CreateJob\")\n\t}\n\t// Make sure job has been created with correct state\n\tif resp.State != schedulerpb.Job_ENABLED {\n\t\treturn errors.Wrap(fmt.Errorf(\"Found incorrect job state %d in response. Must be %d\", resp.State, schedulerpb.Job_ENABLED), \"CreateJob->cl.CloudSchedulerClient.CreateJob\")\n\t}\n\n\treturn nil\n}", "func (cli *Client) CreateJob(pipelineName, sourceKey, targetKey, presetName string) (*api.CreateJobResponse, error) {\n\treturn api.CreateJob(cli, pipelineName, sourceKey, targetKey, presetName)\n}", "func (c *Client) CreateJob(ctx context.Context, body interface{}) (string, error) {\n\tbodyMap, err := toJSONMap(body)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tresponse, err := c.API.SendRequestWithMaps(ctx, \"POST\", \"/bulk\", bodyMap)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tvar jobID string\n\terr = json.Unmarshal(response[\"jobId\"], &jobID)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn jobID, nil\n}", "func CreateJob(request *http.Request) *Job {\n\tjob := &Job{\n\t\trequest: request,\n\t\tresponse: nil,\n\t\terror: nil,\n\t\tisDone: make(chan bool, 1), // mark as done is non-blocking for worker\n\t\tisLogin: false,\n\t}\n\n\treturn job\n}", "func (db jobDatabase) CreateJob(taskType string, userID uuid.UUID, data string) (mgm.Job, error) {\n\tcon, err := db.mysql.GetConnection()\n\tif err != nil {\n\t\treturn mgm.Job{}, err\n\t}\n\tdefer con.Close()\n\n\tres, err := con.Exec(\"INSERT INTO jobs (type, user, data) VALUES (?,?,?)\", taskType, userID.String(), data)\n\tif err != nil {\n\t\treturn mgm.Job{}, err\n\t}\n\tid, _ := res.LastInsertId()\n\tj := mgm.Job{}\n\terr = con.QueryRow(\"SELECT * FROM jobs WHERE id=?\", id).Scan(&j.ID, &j.Timestamp, &j.Type, &j.User, &j.Data)\n\tif err != nil {\n\t\treturn mgm.Job{}, err\n\t}\n\n\treturn j, nil\n}", "func createJob(t *testing.T, descr string, delete bool) *Job {\n\tcf := func(now *time.Time) (bool, bool) { return now.Seconds()%2 == 0, delete }\n\ttf := func(id string) { t.Logf(\"Performed job %s\\n\", id) }\n\n\treturn NewJob(\"test-server-\"+descr, cf, tf)\n}", "func (schematics *SchematicsV1) CreateJob(createJobOptions *CreateJobOptions) (result *Job, response *core.DetailedResponse, err error) {\n\treturn schematics.CreateJobWithContext(context.Background(), createJobOptions)\n}", "func (db DB) CreateJob(j *Job) error {\n\ts := db.session.Copy()\n\tdefer s.Close()\n\treturn s.DB(app).C(\"jobs\").Insert(j)\n}", "func (o *VRS) CreateJob(child *Job) *bambou.Error {\n\n\treturn bambou.CurrentSession().CreateChild(o, child)\n}", "func Create(client *golangsdk.ServiceClient, opts CreateOptsBuilder) (r JobResult) {\n\treqBody, err := opts.ToServerCreateMap()\n\tif err != nil {\n\t\tr.Err = err\n\t\treturn\n\t}\n\n\t_, r.Err = client.Post(createURL(client), reqBody, &r.Body, &golangsdk.RequestOpts{OkCodes: []int{200}})\n\treturn\n}", "func (c *CloudSchedulerClient) CreateJob(ctx context.Context, req *schedulerpb.CreateJobRequest, opts ...gax.CallOption) (*schedulerpb.Job, error) {\n\treturn c.internalClient.CreateJob(ctx, req, opts...)\n}", "func (r *TransferJobsService) Create(transferjob *TransferJob) *TransferJobsCreateCall {\n\tc := &TransferJobsCreateCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.transferjob = transferjob\n\treturn c\n}", "func createCronjob(client k8sclient.Interface, module *protos.Module) error {\n\tcjConfig := createCronJobConfig(module)\n\tcj, err := client.BatchV1beta1().CronJobs(defaultNS).Create(cjConfig)\n\tif err == nil {\n\t\tlog.Infof(\"Created CronJob %+v\", cj)\n\t} else if errors.IsAlreadyExists(err) {\n\t\tlog.Infof(\"CronJob %+v already exists\", cjConfig)\n\t} else {\n\t\tlog.Errorf(\"Failed to create CronJob %+v with error: %v\", cjConfig, err)\n\t}\n\n\treturn err\n}", "func (r *ProjectsLocationsJobsService) Create(parent string, job *Job) *ProjectsLocationsJobsCreateCall {\n\tc := &ProjectsLocationsJobsCreateCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.parent = parent\n\tc.job = job\n\treturn c\n}", "func (c *aITrainingJobs) Create(aITrainingJob *v1.AITrainingJob) (result *v1.AITrainingJob, err error) {\n\tresult = &v1.AITrainingJob{}\n\terr = c.client.Post().\n\t\tNamespace(c.ns).\n\t\tResource(\"aitrainingjobs\").\n\t\tBody(aITrainingJob).\n\t\tDo().\n\t\tInto(result)\n\treturn\n}", "func (j *ScheduledJob) Create(ctx context.Context, ex sqlutil.InternalExecutor, txn *kv.Txn) error {\n\tif j.rec.ScheduleID != 0 {\n\t\treturn errors.New(\"cannot specify schedule id when creating new cron job\")\n\t}\n\n\tif !j.isDirty() {\n\t\treturn errors.New(\"no settings specified for scheduled job\")\n\t}\n\n\tcols, qargs, err := j.marshalChanges()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\trow, retCols, err := ex.QueryRowExWithCols(ctx, \"sched-create\", txn,\n\t\tsessiondata.InternalExecutorOverride{User: security.RootUserName()},\n\t\tfmt.Sprintf(\"INSERT INTO %s (%s) VALUES(%s) RETURNING schedule_id\",\n\t\t\tj.env.ScheduledJobsTableName(), strings.Join(cols, \",\"), generatePlaceholders(len(qargs))),\n\t\tqargs...,\n\t)\n\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"failed to create new schedule\")\n\t}\n\tif row == nil {\n\t\treturn errors.New(\"failed to create new schedule\")\n\t}\n\n\treturn j.InitFromDatums(row, retCols)\n}", "func (o *orm) CreateJob(ctx context.Context, jobSpec *Job, p pipeline.Pipeline) (Job, error) {\n\tvar jb Job\n\tfor _, task := range p.Tasks {\n\t\tif task.Type() == pipeline.TaskTypeBridge {\n\t\t\t// Bridge must exist\n\t\t\tname := task.(*pipeline.BridgeTask).Name\n\t\t\tbt := bridges.BridgeType{}\n\t\t\tif err := o.db.First(&bt, \"name = ?\", name).Error; err != nil {\n\t\t\t\tif errors.Is(err, gorm.ErrRecordNotFound) {\n\t\t\t\t\treturn jb, errors.Wrap(pipeline.ErrNoSuchBridge, name)\n\t\t\t\t}\n\t\t\t\treturn jb, err\n\t\t\t}\n\t\t}\n\t}\n\n\ttx := postgres.TxFromContext(ctx, o.db)\n\n\t// Autogenerate a job ID if not specified\n\tif jobSpec.ExternalJobID == (uuid.UUID{}) {\n\t\tjobSpec.ExternalJobID = uuid.NewV4()\n\t}\n\n\tswitch jobSpec.Type {\n\tcase DirectRequest:\n\t\terr := tx.Create(&jobSpec.DirectRequestSpec).Error\n\t\tif err != nil {\n\t\t\treturn jb, errors.Wrap(err, \"failed to create DirectRequestSpec for jobSpec\")\n\t\t}\n\t\tjobSpec.DirectRequestSpecID = &jobSpec.DirectRequestSpec.ID\n\tcase FluxMonitor:\n\t\terr := tx.Create(&jobSpec.FluxMonitorSpec).Error\n\t\tif err != nil {\n\t\t\treturn jb, errors.Wrap(err, \"failed to create FluxMonitorSpec for jobSpec\")\n\t\t}\n\t\tjobSpec.FluxMonitorSpecID = &jobSpec.FluxMonitorSpec.ID\n\tcase OffchainReporting:\n\t\tif jobSpec.OffchainreportingOracleSpec.EncryptedOCRKeyBundleID.Valid {\n\t\t\t_, err := o.keyStore.OCR().Get(jobSpec.OffchainreportingOracleSpec.EncryptedOCRKeyBundleID.String)\n\t\t\tif err != nil {\n\t\t\t\treturn jb, errors.Wrapf(ErrNoSuchKeyBundle, \"%v\", jobSpec.OffchainreportingOracleSpec.EncryptedOCRKeyBundleID)\n\t\t\t}\n\t\t}\n\t\tif jobSpec.OffchainreportingOracleSpec.P2PPeerID != nil {\n\t\t\t_, err := o.keyStore.P2P().Get(jobSpec.OffchainreportingOracleSpec.P2PPeerID.Raw())\n\t\t\tif err != nil {\n\t\t\t\treturn jb, errors.Wrapf(ErrNoSuchPeerID, \"%v\", jobSpec.OffchainreportingOracleSpec.P2PPeerID)\n\t\t\t}\n\t\t}\n\t\tif jobSpec.OffchainreportingOracleSpec.TransmitterAddress != nil {\n\t\t\t_, err := o.keyStore.Eth().Get(jobSpec.OffchainreportingOracleSpec.TransmitterAddress.Hex())\n\t\t\tif err != nil {\n\t\t\t\treturn jb, errors.Wrapf(ErrNoSuchTransmitterAddress, \"%v\", jobSpec.OffchainreportingOracleSpec.TransmitterAddress)\n\t\t\t}\n\t\t}\n\n\t\terr := tx.Create(&jobSpec.OffchainreportingOracleSpec).Error\n\t\tif err != nil {\n\t\t\treturn jb, errors.Wrap(err, \"failed to create OffchainreportingOracleSpec for jobSpec\")\n\t\t}\n\t\tjobSpec.OffchainreportingOracleSpecID = &jobSpec.OffchainreportingOracleSpec.ID\n\tcase Keeper:\n\t\terr := tx.Create(&jobSpec.KeeperSpec).Error\n\t\tif err != nil {\n\t\t\treturn jb, errors.Wrap(err, \"failed to create KeeperSpec for jobSpec\")\n\t\t}\n\t\tjobSpec.KeeperSpecID = &jobSpec.KeeperSpec.ID\n\tcase Cron:\n\t\terr := tx.Create(&jobSpec.CronSpec).Error\n\t\tif err != nil {\n\t\t\treturn jb, errors.Wrap(err, \"failed to create CronSpec for jobSpec\")\n\t\t}\n\t\tjobSpec.CronSpecID = &jobSpec.CronSpec.ID\n\tcase VRF:\n\t\terr := tx.Create(&jobSpec.VRFSpec).Error\n\t\tpqErr, ok := err.(*pgconn.PgError)\n\t\tif err != nil && ok && pqErr.Code == \"23503\" {\n\t\t\tif pqErr.ConstraintName == \"vrf_specs_public_key_fkey\" {\n\t\t\t\treturn jb, errors.Wrapf(ErrNoSuchPublicKey, \"%s\", jobSpec.VRFSpec.PublicKey.String())\n\t\t\t}\n\t\t}\n\t\tif err != nil {\n\t\t\treturn jb, errors.Wrap(err, \"failed to create VRFSpec for jobSpec\")\n\t\t}\n\t\tjobSpec.VRFSpecID = &jobSpec.VRFSpec.ID\n\tcase Webhook:\n\t\terr := tx.Create(&jobSpec.WebhookSpec).Error\n\t\tif err != nil {\n\t\t\treturn jb, errors.Wrap(err, \"failed to create WebhookSpec for jobSpec\")\n\t\t}\n\t\tjobSpec.WebhookSpecID = &jobSpec.WebhookSpec.ID\n\t\tfor i, eiWS := range jobSpec.WebhookSpec.ExternalInitiatorWebhookSpecs {\n\t\t\tjobSpec.WebhookSpec.ExternalInitiatorWebhookSpecs[i].WebhookSpecID = jobSpec.WebhookSpec.ID\n\t\t\terr := tx.Create(&jobSpec.WebhookSpec.ExternalInitiatorWebhookSpecs[i]).Error\n\t\t\tif err != nil {\n\t\t\t\treturn jb, errors.Wrapf(err, \"failed to create ExternalInitiatorWebhookSpec for WebhookSpec: %#v\", eiWS)\n\t\t\t}\n\t\t}\n\tdefault:\n\t\tlogger.Fatalf(\"Unsupported jobSpec.Type: %v\", jobSpec.Type)\n\t}\n\n\tpipelineSpecID, err := o.pipelineORM.CreateSpec(ctx, tx, p, jobSpec.MaxTaskDuration)\n\tif err != nil {\n\t\treturn jb, errors.Wrap(err, \"failed to create pipeline spec\")\n\t}\n\tjobSpec.PipelineSpecID = pipelineSpecID\n\terr = tx.Create(jobSpec).Error\n\tif err != nil {\n\t\treturn jb, errors.Wrap(err, \"failed to create job\")\n\t}\n\n\treturn o.FindJob(ctx, jobSpec.ID)\n}", "func NewCreateJobCommand(p *config.KfParams) *cobra.Command {\n\tvar (\n\t\tresourceFlags utils.ResourceFlags\n\t\tschedule string\n\t\tconcurrencyPolicy string\n\t\tasync utils.AsyncFlags\n\t)\n\n\tcmd := &cobra.Command{\n\t\tUse: \"create-job APP_NAME JOB_NAME COMMAND\",\n\t\tShort: \"Create a Job on the App.\",\n\t\tExample: `kf create-job my-app my-job \"sleep 100\"`,\n\t\tArgs: cobra.ExactArgs(3),\n\t\tLong: `The create-job sub-command lets operators create a Job that can be run on a schedule or ad hoc.`,\n\t\tSilenceUsage: true,\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tctx := cmd.Context()\n\t\t\tif err := p.ValidateSpaceTargeted(); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tappName := args[0]\n\t\t\tjobName := args[1]\n\t\t\tcommand := args[2]\n\n\t\t\tclient := client.Get(ctx)\n\n\t\t\tapp, err := client.KfV1alpha1().\n\t\t\t\tApps(p.Space).\n\t\t\t\tGet(ctx, appName, metav1.GetOptions{})\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"failed to get App: %s\", err)\n\t\t\t}\n\n\t\t\tdesiredTaskSchedule := &v1alpha1.TaskSchedule{\n\t\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\t\tName: jobName,\n\t\t\t\t\tNamespace: p.Space,\n\t\t\t\t\tOwnerReferences: []metav1.OwnerReference{\n\t\t\t\t\t\t*kmeta.NewControllerRef(app),\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tSpec: v1alpha1.TaskScheduleSpec{\n\t\t\t\t\tSchedule: placeholderCron,\n\t\t\t\t\tSuspend: true,\n\t\t\t\t\tConcurrencyPolicy: concurrencyPolicy,\n\t\t\t\t\tTaskTemplate: v1alpha1.TaskSpec{\n\t\t\t\t\t\tAppRef: corev1.LocalObjectReference{\n\t\t\t\t\t\t\tName: appName,\n\t\t\t\t\t\t},\n\t\t\t\t\t\tCPU: resourceFlags.CPU(),\n\t\t\t\t\t\tMemory: manifest.CFToSIUnits(resourceFlags.Memory()),\n\t\t\t\t\t\tDisk: manifest.CFToSIUnits(resourceFlags.Disk()),\n\t\t\t\t\t\tCommand: command,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t}\n\n\t\t\tif schedule != \"\" {\n\t\t\t\tdesiredTaskSchedule.Spec.Schedule = schedule\n\t\t\t\tdesiredTaskSchedule.Spec.Suspend = false\n\t\t\t}\n\n\t\t\ttaskSchedule, err := client.KfV1alpha1().\n\t\t\t\tTaskSchedules(p.Space).\n\t\t\t\tCreate(ctx, desiredTaskSchedule, metav1.CreateOptions{})\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"failed to create Job: %s\", err)\n\t\t\t}\n\n\t\t\tlogging.FromContext(ctx).Infof(\"Job %s created.\", taskSchedule.Name)\n\n\t\t\treturn async.WaitFor(\n\t\t\t\tctx,\n\t\t\t\tcmd.OutOrStderr(),\n\t\t\t\t\"Waiting for Job to become ready\",\n\t\t\t\ttime.Second,\n\t\t\t\tfunc() (bool, error) {\n\t\t\t\t\tts, err := client.KfV1alpha1().\n\t\t\t\t\t\tTaskSchedules(p.Space).\n\t\t\t\t\t\tGet(ctx, jobName, metav1.GetOptions{})\n\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn false, err\n\t\t\t\t\t}\n\t\t\t\t\treturn ts.Status.IsReady(), nil\n\t\t\t\t},\n\t\t\t)\n\t\t},\n\t}\n\n\tresourceFlags.Add(cmd)\n\tasync.Add(cmd)\n\n\t// The default is left as \"\" here to determine if the schedule flag was\n\t// provided. When not provided the schedule is defaulted to placeholderCron.\n\tcmd.Flags().StringVarP(\n\t\t&schedule,\n\t\t\"schedule\",\n\t\t\"s\",\n\t\t\"\",\n\t\t\"Cron schedule on which to execute the Job.\",\n\t)\n\n\tcmd.Flags().StringVarP(\n\t\t&concurrencyPolicy,\n\t\t\"concurrency-policy\",\n\t\t\"c\",\n\t\t\"Always\",\n\t\t\"Specifies how to treat concurrent executions of a Job: Always (default), Replace, or Forbid.\",\n\t)\n\n\treturn cmd\n}", "func newCronJobForCR(cr *v1alpha1.Copybird) *v1beta1.CronJob {\n\tlabels := map[string]string{\n\t\t\"app\": cr.Name,\n\t}\n\treturn &v1beta1.CronJob{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tName: cr.Name + \"-cronjob\",\n\t\t\tNamespace: cr.Namespace,\n\t\t\tLabels: labels,\n\t\t},\n\t\tSpec: v1beta1.CronJobSpec{\n\t\t\tSchedule: cr.Spec.Cron,\n\t\t\tJobTemplate: v1beta1.JobTemplateSpec{\n\t\t\t\tSpec: batchv1.JobSpec{\n\t\t\t\t\tTemplate: v1.PodTemplateSpec{\n\t\t\t\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\t\t\t\tName: cr.Name + \"-copybird\",\n\t\t\t\t\t\t\tNamespace: cr.Namespace,\n\t\t\t\t\t\t\tLabels: labels,\n\t\t\t\t\t\t},\n\t\t\t\t\t\tSpec: v1.PodSpec{\n\t\t\t\t\t\t\tContainers: []v1.Container{\n\t\t\t\t\t\t\t\t{\n\t\t\t\t\t\t\t\t\tName: cr.Name,\n\t\t\t\t\t\t\t\t\tImage: \"copybird/copybird\",\n\t\t\t\t\t\t\t\t\tCommand: []string{},\n\t\t\t\t\t\t\t\t\tArgs: []string{},\n\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\tRestartPolicy: \"OnFailure\",\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n}", "func NewJob() *Job {\n\treturn &Job{\n\t\tID: uuid.New().String(),\n\t\tPriority: PriorityNormal,\n\t\tTimestamp: time.Now(),\n\t\tContentType: ContentTypeMsgpack,\n\t}\n}", "func (r *JobRepo) Create(data *model.Job) (*model.Job, error) {\n\tdb, err := db.Instance()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\terr = db.Create(&data).Error\n\treturn data, err\n}", "func CreateJob(board schema.Board, boardID string, categoryID string, ownerID string) {\n\n\tobjID, err := primitive.ObjectIDFromHex(boardID)\n\tcategoryID_int, err := strconv.Atoi(categoryID)\n\tif err != nil {\n\t\t// handle error\n\t\tlog.Fatal(\"CreateJob() str conv ERROR:\", err)\n\t}\n\n\t//get job details from board document inputted through put request\n\tjd := schema.JobDetails{\n\t\tCompany: board.Categories[0].Jobs[0].JobDetails.Company,\n\t\tTitle: board.Categories[0].Jobs[0].JobDetails.Title,\n\t\tLocation: board.Categories[0].Jobs[0].JobDetails.Location,\n\t\tCategory: board.Categories[0].Jobs[0].JobDetails.Category,\n\t\tPostDate: board.Categories[0].Jobs[0].JobDetails.PostDate,\n\t\tDescription: board.Categories[0].Jobs[0].JobDetails.Description,\n\t\tExperience: board.Categories[0].Jobs[0].JobDetails.Experience,\n\t\tURL: board.Categories[0].Jobs[0].JobDetails.URL,\n\t\tDateAdded: board.Categories[0].Jobs[0].JobDetails.DateAdded,\n\t\tSalary: board.Categories[0].Jobs[0].JobDetails.Salary,\n\t\tTasks: board.Categories[0].Jobs[0].JobDetails.Tasks}\n\n\t//stick job details into job var\n\tj := schema.Job{JobDetails: jd, ID: board.Categories[0].Jobs[0].ID}\n\n\tfilter := bson.M{\"_id\": bson.M{\"$eq\": objID}, \"categories.id\": bson.M{\"$eq\": categoryID_int}}\n\tupdate := bson.M{\"$set\": bson.M{\"categories.$.jobs\": j}}\n\n\tupdateResult, err := db.Collection(COLLNAME).UpdateOne(\n\t\tcontext.Background(),\n\t\tfilter,\n\t\tupdate,\n\t)\n\tif err != nil {\n\t\tlog.Fatal(\"CreateJob() ERROR:\", err)\n\t}\n\n\tfmt.Println(\"create job func input: \", j)\n\n\tfmt.Printf(\"Matched %v documents and updated %v documents.\\n\", updateResult.MatchedCount, updateResult.ModifiedCount)\n\tfmt.Println(\"Full Result: \", updateResult)\n\n}", "func (a *SyncApiService) CreateSyncJob(ctx context.Context, syncJob SyncJobCreateParams) (CreateResponse, *http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Post\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t \tsuccessPayload CreateResponse\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/platform/3/sync/jobs\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{ \"application/json\", }\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\n\t\t\"application/json\",\n\t\t}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &syncJob\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn successPayload, nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn successPayload, localVarHttpResponse, err\n\t}\n\tdefer localVarHttpResponse.Body.Close()\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tbodyBytes, _ := ioutil.ReadAll(localVarHttpResponse.Body)\n\t\treturn successPayload, localVarHttpResponse, reportError(\"Status: %v, Body: %s\", localVarHttpResponse.Status, bodyBytes)\n\t}\n\n\tif err = json.NewDecoder(localVarHttpResponse.Body).Decode(&successPayload); err != nil {\n\t\treturn successPayload, localVarHttpResponse, err\n\t}\n\n\n\treturn successPayload, localVarHttpResponse, err\n}", "func NewJob(receipts []uuid.UUID, sendTime time.Time) *Job {\n\treturn &Job{\n\t\tID: uuid.NewUUID(),\n\t\tSendTime: sendTime,\n\t\tSendStatus: READY,\n\t\tReceipts: receipts,\n\t}\n}", "func (cli *Client) CronCreate(ctx context.Context, name string, sv types.Cron) (types.Cron, error) {\n\tvar cron types.Cron\n\tvar r = url.Values{}\n\tr.Set(\"name\", name)\n\tresp, err := cli.post(ctx, \"/crons/create\", r, sv, nil)\n\tif err != nil {\n\t\treturn cron, err\n\t}\n\terr = json.NewDecoder(resp.body).Decode(&cron)\n\tensureReaderClosed(resp)\n\treturn cron, err\n}", "func (r *JobsService) Create(createjobrequest *CreateJobRequest) *JobsCreateCall {\n\tc := &JobsCreateCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.createjobrequest = createjobrequest\n\treturn c\n}", "func (rh *RestHandler) Create(w http.ResponseWriter, r *http.Request) {\n\terr := r.ParseForm()\n\tif err != nil {\n\t\twriteError(w, http.StatusBadRequest, \"unable to parse form values\")\n\t\treturn\n\t}\n\n\tjob := model.Job{}\n\tif r.FormValue(\"partner_id\") == \"\" {\n\t\twriteError(w, http.StatusBadRequest, \"missing partner_id value\")\n\t\treturn\n\t}\n\tjob.PartnerID, err = strconv.ParseInt(r.FormValue(\"partner_id\"), 10, 64)\n\tif err != nil {\n\t\twriteError(w, http.StatusBadRequest, \"invalid partner_id value\")\n\t\treturn\n\t}\n\tif r.FormValue(\"category_id\") == \"\" {\n\t\twriteError(w, http.StatusBadRequest, \"missing category_id value\")\n\t\treturn\n\t}\n\tjob.CategoryID, err = strconv.ParseInt(r.FormValue(\"category_id\"), 10, 64)\n\tif err != nil {\n\t\twriteError(w, http.StatusBadRequest, errors.Wrap(err, \"invalid category_id value\").Error())\n\t\treturn\n\t}\n\tjob.Title = strings.TrimSpace(r.FormValue(\"title\"))\n\tif job.Title == \"\" {\n\t\twriteError(w, http.StatusBadRequest, \"missing or empty title given\")\n\t\treturn\n\t}\n\n\tif r.FormValue(\"expires_at\") == \"\" {\n\t\twriteError(w, http.StatusBadRequest, \"missing expires_at value\")\n\t\treturn\n\t}\n\tnow := time.Now()\n\tjob.ExpiresAt, err = time.ParseInLocation(model.DateFormat, r.FormValue(\"expires_at\"), now.Location())\n\tif err != nil || job.ExpiresAt.IsZero() {\n\t\twriteError(w, http.StatusBadRequest, errors.Wrap(err, \"invalid expiration date\").Error())\n\t\treturn\n\t}\n\t//Times are parsed without hour, so whe have to add the hours until de end of the day\n\tjob.ExpiresAt = job.ExpiresAt.Add(23*time.Hour + 59*time.Minute + 59*time.Second)\n\tif job.ExpiresAt.Before(now) {\n\t\twriteError(w, http.StatusBadRequest, \"job already expired\")\n\t\treturn\n\t}\n\treq := model.RequestCreate{}\n\treq.Job = job\n\n\tencreq, err := crypt.EncryptRequest(rh.cipher, req)\n\tif err != nil {\n\t\twriteError(w, http.StatusInternalServerError, err.Error())\n\t\treturn\n\t}\n\n\t_, err = rh.backend.Create(context.Background(), encreq)\n\tif err != nil {\n\t\twriteError(w, http.StatusInternalServerError, err.Error())\n\t\treturn\n\t}\n\n\twriteResponse(w, http.StatusCreated, nil)\n}", "func (h *CronJobCreateHandler) Handle(ctx context.Context, req types.Request) types.Response {\n\tobj := &batchv1.CronJob{}\n\n\terr := h.Decoder.Decode(req, obj)\n\tif err != nil {\n\t\treturn admission.ErrorResponse(http.StatusBadRequest, err)\n\t}\n\n\tallowed, reason, err := h.validatingCronJobFn(ctx, obj)\n\tif err != nil {\n\t\treturn admission.ErrorResponse(http.StatusInternalServerError, err)\n\t}\n\treturn admission.ValidationResponse(allowed, reason)\n}", "func (c *jobsRESTClient) CreateJobOperation(name string) *CreateJobOperation {\n\toverride := fmt.Sprintf(\"/v2/%s\", name)\n\treturn &CreateJobOperation{\n\t\tlro: longrunning.InternalNewOperation(*c.LROClient, &longrunningpb.Operation{Name: name}),\n\t\tpollPath: override,\n\t}\n}", "func CreateJob(cmd *cobra.Command, args []string) error {\n\n\tclient, err := auth.GetClient()\n\tif err != nil {\n\t\treturn err\n\t}\n\t// Parse all flags\n\n\tvar allowSideEffectsDefault bool\n\tallowSideEffects := &allowSideEffectsDefault\n\terr = flags.ParseFlag(cmd.Flags(), \"allow-side-effects\", &allowSideEffects)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"allow-side-effects\": ` + err.Error())\n\t}\n\tvar collectEventSummaryDefault bool\n\tcollectEventSummary := &collectEventSummaryDefault\n\terr = flags.ParseFlag(cmd.Flags(), \"collect-event-summary\", &collectEventSummary)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"collect-event-summary\": ` + err.Error())\n\t}\n\tvar collectFieldSummaryDefault bool\n\tcollectFieldSummary := &collectFieldSummaryDefault\n\terr = flags.ParseFlag(cmd.Flags(), \"collect-field-summary\", &collectFieldSummary)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"collect-field-summary\": ` + err.Error())\n\t}\n\tvar collectTimeBucketsDefault bool\n\tcollectTimeBuckets := &collectTimeBucketsDefault\n\terr = flags.ParseFlag(cmd.Flags(), \"collect-time-buckets\", &collectTimeBuckets)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"collect-time-buckets\": ` + err.Error())\n\t}\n\tvar earliestDefault string\n\tearliest := &earliestDefault\n\terr = flags.ParseFlag(cmd.Flags(), \"earliest\", &earliest)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"earliest\": ` + err.Error())\n\t}\n\tvar enablePreviewDefault bool\n\tenablePreview := &enablePreviewDefault\n\terr = flags.ParseFlag(cmd.Flags(), \"enable-preview\", &enablePreview)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"enable-preview\": ` + err.Error())\n\t}\n\tvar extractAllFieldsDefault bool\n\textractAllFields := &extractAllFieldsDefault\n\terr = flags.ParseFlag(cmd.Flags(), \"extract-all-fields\", &extractAllFields)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"extract-all-fields\": ` + err.Error())\n\t}\n\tvar extractFieldsDefault string\n\textractFields := &extractFieldsDefault\n\terr = flags.ParseFlag(cmd.Flags(), \"extract-fields\", &extractFields)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"extract-fields\": ` + err.Error())\n\t}\n\tvar latestDefault string\n\tlatest := &latestDefault\n\terr = flags.ParseFlag(cmd.Flags(), \"latest\", &latest)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"latest\": ` + err.Error())\n\t}\n\tvar maxTimeDefault int32\n\tmaxTime := &maxTimeDefault\n\terr = flags.ParseFlag(cmd.Flags(), \"max-time\", &maxTime)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"max-time\": ` + err.Error())\n\t}\n\tvar messages []model.Message\n\terr = flags.ParseFlag(cmd.Flags(), \"messages\", &messages)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"messages\": ` + err.Error())\n\t}\n\tvar moduleDefault string\n\tmodule := &moduleDefault\n\terr = flags.ParseFlag(cmd.Flags(), \"module\", &module)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"module\": ` + err.Error())\n\t}\n\tvar query string\n\terr = flags.ParseFlag(cmd.Flags(), \"query\", &query)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"query\": ` + err.Error())\n\t}\n\tvar relativeTimeAnchorDefault string\n\trelativeTimeAnchor := &relativeTimeAnchorDefault\n\terr = flags.ParseFlag(cmd.Flags(), \"relative-time-anchor\", &relativeTimeAnchor)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"relative-time-anchor\": ` + err.Error())\n\t}\n\tvar requiredFreshnessDefault int32\n\trequiredFreshness := &requiredFreshnessDefault\n\terr = flags.ParseFlag(cmd.Flags(), \"required-freshness\", &requiredFreshness)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"required-freshness\": ` + err.Error())\n\t}\n\tvar statusDefault model.SearchStatus\n\tstatus := &statusDefault\n\terr = flags.ParseFlag(cmd.Flags(), \"status\", &status)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"status\": ` + err.Error())\n\t}\n\tvar timezone interface{}\n\terr = flags.ParseFlag(cmd.Flags(), \"timezone\", &timezone)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"timezone\": ` + err.Error())\n\t}\n\t// Form the request body\n\tgenerated_request_body := model.SearchJob{\n\n\t\tAllowSideEffects: allowSideEffects,\n\t\tCollectEventSummary: collectEventSummary,\n\t\tCollectFieldSummary: collectFieldSummary,\n\t\tCollectTimeBuckets: collectTimeBuckets,\n\t\tEnablePreview: enablePreview,\n\t\tExtractAllFields: extractAllFields,\n\t\tExtractFields: extractFields,\n\t\tMaxTime: maxTime,\n\t\tMessages: messages,\n\t\tModule: module,\n\t\tQuery: query,\n\t\tQueryParameters: &model.QueryParameters{\n\t\t\tEarliest: earliest,\n\t\t\tLatest: latest,\n\t\t\tRelativeTimeAnchor: relativeTimeAnchor,\n\t\t\tTimezone: timezone,\n\t\t},\n\t\tRequiredFreshness: requiredFreshness,\n\t\tStatus: status,\n\t}\n\n\t// Silence Usage\n\tcmd.SilenceUsage = true\n\n\tresp, err := client.SearchService.CreateJob(generated_request_body)\n\tif err != nil {\n\t\treturn err\n\t}\n\tjsonx.Pprint(cmd, resp)\n\treturn nil\n}", "func NewJob(jobType string, data interface{}) Job {\n\tj := Job{\n\t\tuuid: uuid.New().String(),\n\t\tjobType: jobType,\n\t\tdata: data,\n\t}\n\n\t// detect the coordinated request\n\tif req, ok := data.(*request.CoordinatedRequest); ok {\n\t\tj.req = req\n\t}\n\n\treturn j\n}", "func CreateOrUpdate(ctx context.Context, c client.Client, cj *batchv1.CronJob, equal EqualityFunc, mutate MutateFunc) error {\n\tcurrent := &batchv1.CronJob{}\n\tkey := client.ObjectKey{Name: cj.Name, Namespace: cj.Namespace}\n\terr := c.Get(ctx, key, current)\n\tif err != nil {\n\t\tif apierrors.IsNotFound(err) {\n\t\t\terr = c.Create(ctx, cj)\n\n\t\t\tif err == nil {\n\t\t\t\treturn nil\n\t\t\t}\n\n\t\t\treturn kverrors.Wrap(err, \"failed to create cronjob\",\n\t\t\t\t\"name\", cj.Name,\n\t\t\t\t\"namespace\", cj.Namespace,\n\t\t\t)\n\t\t}\n\n\t\treturn kverrors.Wrap(err, \"failed to get cronjob\",\n\t\t\t\"name\", cj.Name,\n\t\t\t\"namespace\", cj.Namespace,\n\t\t)\n\t}\n\n\tif !equal(current, cj) {\n\t\terr := retry.RetryOnConflict(retry.DefaultRetry, func() error {\n\t\t\tif err := c.Get(ctx, key, current); err != nil {\n\t\t\t\treturn kverrors.Wrap(err, \"failed to get cronjob\",\n\t\t\t\t\t\"name\", cj.Name,\n\t\t\t\t\t\"namespace\", cj.Namespace,\n\t\t\t\t)\n\t\t\t}\n\n\t\t\tmutate(current, cj)\n\t\t\tif err := c.Update(ctx, current); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\t\tif err != nil {\n\t\t\treturn kverrors.Wrap(err, \"failed to update cronjob\",\n\t\t\t\t\"name\", cj.Name,\n\t\t\t\t\"namespace\", cj.Namespace,\n\t\t\t)\n\t\t}\n\t\treturn nil\n\t}\n\n\treturn nil\n}", "func (ci *CronImpl) Create(timeZone string) (*cron.Cron, error) {\n\tl, err := time.LoadLocation(timeZone)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn cron.NewWithLocation(l), nil\n}", "func (client *Client) CreateETLJob(request *CreateETLJobRequest) (response *CreateETLJobResponse, err error) {\n\tresponse = CreateCreateETLJobResponse()\n\terr = client.DoAction(request, response)\n\treturn\n}", "func CreateJobFromCronjob(cronJob *batchv1beta1.CronJob) (*batchv1.Job, error) {\n\t// This duplicates the logic used by kubectl to create a Job from a CronJob\n\tannotations := make(map[string]string)\n\tannotations[\"cronjob.kubernetes.io/instantiate\"] = \"manual\"\n\tfor k, v := range cronJob.Spec.JobTemplate.Annotations {\n\t\tannotations[k] = v\n\t}\n\n\tjobDef := &batchv1.Job{\n\t\tTypeMeta: metav1.TypeMeta{APIVersion: batchv1.SchemeGroupVersion.String(), Kind: \"Job\"},\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tName: fmt.Sprintf(\"%s-%d\", cronJob.ObjectMeta.Name, time.Now().Unix()),\n\t\t\tAnnotations: annotations,\n\t\t\tLabels: cronJob.Spec.JobTemplate.Labels,\n\t\t\tOwnerReferences: []metav1.OwnerReference{\n\t\t\t\t*metav1.NewControllerRef(cronJob, appsv1.SchemeGroupVersion.WithKind(\"CronJob\")),\n\t\t\t},\n\t\t},\n\t\tSpec: cronJob.Spec.JobTemplate.Spec,\n\t}\n\n\tif job, err := client.BatchV1().Jobs(cronJob.ObjectMeta.Namespace).Create(jobDef); err != nil {\n\t\treturn nil, err\n\t} else {\n\t\treturn job, nil\n\t}\n}", "func (c APIClient) CreateMonitoringJob(mj *MonitoringJob) error {\n\treturn c.doHTTPBoth(\"PUT\", \"https://api.nsone.net/v1/monitoring/jobs\", mj)\n}", "func (c *JobsClient) CreateJob(ctx context.Context, req *runpb.CreateJobRequest, opts ...gax.CallOption) (*CreateJobOperation, error) {\n\treturn c.internalClient.CreateJob(ctx, req, opts...)\n}", "func CreateBaseJob(message *models.Message) (Processable, error) {\n\treturn BaseJob{\n\t\tmessage: message,\n\t}, nil\n}", "func (s *store) CreateJob(name string, containerDetailsArray []platform.ContainerDetails) error {\n\n\tvar containers []apiv1.Container\n\tfor _, containerDetails := range containerDetailsArray {\n\t\tcontainers = append(containers, apiv1.Container{\n\t\t\tName: containerDetails.Name,\n\t\t\tImage: containerDetails.Image,\n\t\t\tEnv: containerDetails.EnvVars,\n\t\t\tImagePullPolicy: containerDetails.ImagePullPolicy,\n\t\t})\n\t}\n\n\tjobsClient := s.client.BatchV1().Jobs(s.namespace)\n\tjob := &batchv1.Job{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tName: name,\n\t\t},\n\t\tSpec: batchv1.JobSpec{\n\t\t\tBackoffLimit: int32Ptr(5),\n\t\t\tTemplate: apiv1.PodTemplateSpec{\n\t\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\t\tLabels: map[string]string{\n\t\t\t\t\t\t\"app\": \"presidio\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tSpec: apiv1.PodSpec{\n\t\t\t\t\tContainers: containers,\n\t\t\t\t\tRestartPolicy: \"OnFailure\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\n\t_, err := jobsClient.Create(job)\n\treturn err\n}", "func (s *Session) CreateGitLabJob(namePrefix string, spec *protocol.JobSpec, k8sJobParams *K8SJobParameters, cacheBucket string) (*Job, error) {\n\tjob, err := newJobFromGitLab(s, namePrefix, spec, k8sJobParams, cacheBucket)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn job, nil\n}", "func (j *JobWithSchedule) BuildJob() (RawJob, error) {\n\tvar bufferJob bytes.Buffer\n\tencoderJob := gob.NewEncoder(&bufferJob)\n\t// encode the CronJob interface keeping the unit of work\n\t// to execute. We need to use the encodeJob method\n\t// due to how gob interface encoding works.\n\tif err := encodeJob(encoderJob, j.run); err != nil {\n\t\treturn RawJob{}, err\n\t}\n\t// finally, encode the bytes to base64\n\tj.rawJob.SerializedJob = base64.StdEncoding.EncodeToString(bufferJob.Bytes())\n\n\t// now, encode the job input\n\tif err := j.encodeJobInput(); err != nil {\n\t\treturn RawJob{}, err\n\t}\n\treturn j.rawJob, nil\n}", "func (c *JobpositionClient) Create() *JobpositionCreate {\n\tmutation := newJobpositionMutation(c.config, OpCreate)\n\treturn &JobpositionCreate{config: c.config, hooks: c.Hooks(), mutation: mutation}\n}", "func (c *jobsGRPCClient) CreateJobOperation(name string) *CreateJobOperation {\n\treturn &CreateJobOperation{\n\t\tlro: longrunning.InternalNewOperation(*c.LROClient, &longrunningpb.Operation{Name: name}),\n\t}\n}", "func NewJob() *Job {\n\treturn &Job{}\n}", "func NewJob(jobtype string, args ...interface{}) *Job {\n\treturn &Job{\n\t\tType: jobtype,\n\t\tQueue: \"default\",\n\t\tArgs: args,\n\t\tJid: RandomJid(),\n\t\tCreatedAt: time.Now().UTC().Format(time.RFC3339Nano),\n\t\tRetry: 25,\n\t}\n}", "func (r *ReconcileDescheduler) createJob(descheduler *deschedulerv1alpha1.Descheduler) (*batch.Job, error) {\n\tactiveDeadline := int64(100)\n\tlog.Printf(\"Creating descheduler job\")\n\tjob := &batch.Job{\n\t\tTypeMeta: metav1.TypeMeta{\n\t\t\tKind: \"Job\",\n\t\t\tAPIVersion: batch.SchemeGroupVersion.String(),\n\t\t},\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tName: descheduler.Name,\n\t\t\tNamespace: descheduler.Namespace,\n\t\t},\n\t\tSpec: batch.JobSpec{\n\t\t\tActiveDeadlineSeconds: &activeDeadline,\n\t\t\tTemplate: v1.PodTemplateSpec{\n\t\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\t\tName: \"descheduler-job-spec\",\n\t\t\t\t},\n\t\t\t\tSpec: v1.PodSpec{\n\t\t\t\t\tVolumes: []v1.Volume{{\n\t\t\t\t\t\tName: \"policy-volume\",\n\t\t\t\t\t\tVolumeSource: v1.VolumeSource{\n\t\t\t\t\t\t\tConfigMap: &v1.ConfigMapVolumeSource{\n\t\t\t\t\t\t\t\tLocalObjectReference: v1.LocalObjectReference{\n\t\t\t\t\t\t\t\t\tName: descheduler.Name,\n\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t\tRestartPolicy: \"Never\",\n\t\t\t\t\tContainers: []v1.Container{{\n\t\t\t\t\t\tName: \"openshift-descheduler\",\n\t\t\t\t\t\tImage: \"registry.svc.ci.openshift.org/openshift/origin-v4.0:descheduler\", // TODO: Make this configurable too.\n\t\t\t\t\t\tPorts: []v1.ContainerPort{{ContainerPort: 80}},\n\t\t\t\t\t\tResources: v1.ResourceRequirements{\n\t\t\t\t\t\t\tLimits: v1.ResourceList{\n\t\t\t\t\t\t\t\tv1.ResourceCPU: resource.MustParse(\"100m\"),\n\t\t\t\t\t\t\t\tv1.ResourceMemory: resource.MustParse(\"500Mi\"),\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\tRequests: v1.ResourceList{\n\t\t\t\t\t\t\t\tv1.ResourceCPU: resource.MustParse(\"100m\"),\n\t\t\t\t\t\t\t\tv1.ResourceMemory: resource.MustParse(\"500Mi\"),\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t},\n\t\t\t\t\t\tCommand: []string{\"/bin/descheduler\", \"--policy-config-file\", \"/policy-dir/policy.yaml\"},\n\t\t\t\t\t\tVolumeMounts: []v1.VolumeMount{{\n\t\t\t\t\t\t\tMountPath: \"/policy-dir\",\n\t\t\t\t\t\t\tName: \"policy-volume\",\n\t\t\t\t\t\t}},\n\t\t\t\t\t}},\n\t\t\t\t\tServiceAccountName: \"openshift-descheduler\", // TODO: This is hardcoded as of now, find a way to reference it from rbac.yaml.\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\terr := controllerutil.SetControllerReference(descheduler, job, r.scheme)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Error setting owner references %v\", err)\n\t}\n\treturn job, nil\n}", "func NewJob(name string, fun interface{}, args ...interface{}) (*Job, error) {\n\tfunction := reflect.ValueOf(fun)\n\tif function.Kind() != reflect.Func {\n\t\treturn nil, errors.New(\"schedule: jobs can only be created for functions\")\n\t}\n\targuments := make([]reflect.Value, len(args))\n\tfor i, arg := range args {\n\t\targuments[i] = reflect.ValueOf(arg)\n\t}\n\treturn &Job{\n\t\tName: name,\n\t\tfunction: function,\n\t\targs: arguments,\n\t\ttrigger: NewTrigger(),\n\t}, nil\n}", "func (r *ProjectsLocationsMigrationJobsService) Create(parent string, migrationjob *MigrationJob) *ProjectsLocationsMigrationJobsCreateCall {\n\tc := &ProjectsLocationsMigrationJobsCreateCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.parent = parent\n\tc.migrationjob = migrationjob\n\treturn c\n}", "func NewJob(name, ns string, containers ContainerList) *Job {\n\tret := &Job{\n\t\tcore: &batchv1.Job{\n\t\t\tMetadata: objectMeta(name, ns),\n\t\t\tSpec: &batchv1.JobSpec{\n\t\t\t\tTemplate: podTemplateSpec(emptyMap(), containers),\n\t\t\t},\n\t\t},\n\t}\n\tret.core.Spec.Template.Spec.RestartPolicy = k8s.String(\"OnFailure\")\n\treturn ret\n}", "func (client JobClient) CreateResponder(resp *http.Response) (result Job, err error) {\n\terr = autorest.Respond(\n\t\tresp,\n\t\tazure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated),\n\t\tautorest.ByUnmarshallingJSON(&result),\n\t\tautorest.ByClosing())\n\tresult.Response = autorest.Response{Response: resp}\n\treturn\n}", "func NewJob(ctx *pulumi.Context,\n\tname string, args *JobArgs, opts ...pulumi.ResourceOption) (*Job, error) {\n\tif args == nil {\n\t\targs = &JobArgs{}\n\t}\n\n\treplaceOnChanges := pulumi.ReplaceOnChanges([]string{\n\t\t\"location\",\n\t\t\"project\",\n\t})\n\topts = append(opts, replaceOnChanges)\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Job\n\terr := ctx.RegisterResource(\"google-native:dataflow/v1b3:Job\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (o *AggregatedDomain) CreateJob(child *Job) *bambou.Error {\n\n\treturn bambou.CurrentSession().CreateChild(o, child)\n}", "func (cli *Client) CreateJobCustomize(args *api.CreateJobArgs) (*api.CreateJobResponse, error) {\n\treturn api.CreateJobCustomize(cli, args)\n}", "func NewJob(fn func() error, opts ...JobOption) Job {\n\tvar o jobOptions\n\to = defaultJobOptions\n\tfor _, opt := range opts {\n\t\topt(&o)\n\t}\n\n\tif o.externalTrigger {\n\t\treturn &job{\n\t\t\tfn: fn,\n\t\t}\n\t}\n\n\treturn &cronJob{\n\t\tfn: fn,\n\t\tschedule: o.schedule,\n\t\tcron: o.cron,\n\t}\n}", "func createJobWithCustomAttributes(w io.Writer, projectID, companyID, jobTitle string) (*talentpb.Job, error) {\n\tctx := context.Background()\n\n\t// Initialize a job service client.\n\tc, err := talent.NewJobClient(ctx)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"talent.NewJobClient: %w\", err)\n\t}\n\tdefer c.Close()\n\n\t// requisitionID shoud be the unique ID in your system\n\trequisitionID := fmt.Sprintf(\"job-with-custom-attribute-%s\", uuid.Must(uuid.NewV4()).String())\n\tjobToCreate := &talentpb.Job{\n\t\tCompany: fmt.Sprintf(\"projects/%s/companies/%s\", projectID, companyID),\n\t\tRequisitionId: requisitionID,\n\t\tTitle: jobTitle,\n\t\tApplicationInfo: &talentpb.Job_ApplicationInfo{\n\t\t\tUris: []string{\"https://googlesample.com/career\"},\n\t\t},\n\t\tDescription: \"Design, devolop, test, deploy, maintain and improve software.\",\n\t\tLanguageCode: \"en-US\",\n\t\tPromotionValue: 2,\n\t\tEmploymentTypes: []talentpb.EmploymentType{talentpb.EmploymentType_FULL_TIME},\n\t\tAddresses: []string{\"Mountain View, CA\"},\n\t\tCustomAttributes: map[string]*talentpb.CustomAttribute{\n\t\t\t\"someFieldString\": {\n\t\t\t\tFilterable: true,\n\t\t\t\tStringValues: []string{\"someStrVal\"},\n\t\t\t},\n\t\t\t\"someFieldLong\": {\n\t\t\t\tFilterable: true,\n\t\t\t\tLongValues: []int64{900},\n\t\t\t},\n\t\t},\n\t\tCompensationInfo: &talentpb.CompensationInfo{\n\t\t\tEntries: []*talentpb.CompensationInfo_CompensationEntry{\n\t\t\t\t{\n\t\t\t\t\tType: talentpb.CompensationInfo_BASE,\n\t\t\t\t\tUnit: talentpb.CompensationInfo_HOURLY,\n\t\t\t\t\tCompensationAmount: &talentpb.CompensationInfo_CompensationEntry_Amount{\n\t\t\t\t\t\tAmount: &money.Money{\n\t\t\t\t\t\t\tCurrencyCode: \"USD\",\n\t\t\t\t\t\t\tUnits: 1,\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\n\t// Construct a createJob request.\n\treq := &talentpb.CreateJobRequest{\n\t\tParent: fmt.Sprintf(\"projects/%s\", projectID),\n\t\tJob: jobToCreate,\n\t}\n\n\tresp, err := c.CreateJob(ctx, req)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"CreateJob: %w\", err)\n\t}\n\n\tfmt.Fprintf(w, \"Created job with custom attributres: %q\\n\", resp.GetName())\n\tfmt.Fprintf(w, \"Custom long field has value: %v\\n\", resp.GetCustomAttributes()[\"someFieldLong\"].GetLongValues())\n\n\treturn resp, nil\n}", "func CreateAndWaitJob(ns string, job *v1.Job, timeout time.Duration) (*v1.Job, error) {\n\tretJob, err := kubecli.BatchV1().Jobs(ns).Create(job)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tinterval := time.Second\n\terr = retry.Retry(int(timeout/interval), interval, func() (bool, error) {\n\t\tretJob, err = kubecli.BatchV1().Jobs(ns).Get(job.Name, metav1.GetOptions{})\n\t\tif err != nil {\n\t\t\treturn false, err\n\t\t}\n\n\t\tswitch retJob.Status.Active {\n\t\tcase 1:\n\t\t\treturn true, nil\n\t\tdefault:\n\t\t\treturn false, nil\n\t\t}\n\t})\n\tlog.Infof(\"Job '%q' created\", retJob.GetName())\n\treturn retJob, nil\n}", "func NewJob(object dbus.BusObject) *Job {\n\treturn &Job{object}\n}", "func (w *Worker) NewJob(j *templates.Message, reply *bool) (e error) {\n\t// T.Ln(\"received new job\")\n\tif !w.dispatchReady.Load() {\n\t\tD.Ln(\"dispatch not ready\")\n\t\t*reply = true\n\t\treturn\n\t}\n\tif w.templatesMessage != nil {\n\t\tif j.PrevBlock == w.templatesMessage.PrevBlock {\n\t\t\t// T.Ln(\"not a new job\")\n\t\t\t*reply = true\n\t\t\treturn\n\t\t}\n\t}\n\t// D.S(j)\n\t*reply = true\n\tD.Ln(\"halting current work\")\n\tw.stopChan <- struct{}{}\n\t// load the job into the template\n\tif w.templatesMessage == nil {\n\t\tw.templatesMessage = j\n\t} else {\n\t\t*w.templatesMessage = *j\n\t}\n\tD.Ln(\"switching to new job\")\n\tw.startChan <- struct{}{}\n\treturn\n}", "func NewJob(ctx *pulumi.Context,\n\tname string, args *JobArgs, opts ...pulumi.ResourceOption) (*Job, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.RoleArn == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'RoleArn'\")\n\t}\n\tif args.Type == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'Type'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Job\n\terr := ctx.RegisterResource(\"aws-native:databrew:Job\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (c *JobsClient) CreateJobOperation(name string) *CreateJobOperation {\n\treturn c.internalClient.CreateJobOperation(name)\n}", "func (p *SqliteProvider) CreateJob(job *models.Job) error {\n\t_, err := p.db.NamedExec(`INSERT INTO jobs (jid, status, started_at, finished_at, created_by, client_id, multi_job_id, details)\n\t\t\t\t\t\t\t\t\t\t\tVALUES (:jid, :status, :started_at, :finished_at, :created_by, :client_id, :multi_job_id, :details)`,\n\t\tconvertToSqlite(job))\n\tif err != nil {\n\t\t// check if it's \"already exist\" err\n\t\ttypeErr, ok := err.(sqlite3.Error)\n\t\tif ok && typeErr.Code == sqlite3.ErrConstraint {\n\t\t\tp.log.Debugf(\"Job already exist with ID: %s\", job.JID)\n\t\t\treturn nil\n\t\t}\n\t} else {\n\t\tp.log.Debugf(\"Job saved successfully: %v\", *job)\n\t}\n\treturn err\n}", "func (s *grpcServer) CreateJobPost(ctx context.Context, req *pb.CreateJobPostRequest) (*pb.JobPost, error) {\n\t_, rep, err := s.createJobPost.ServeGRPC(ctx, req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn rep.(*pb.JobPost), nil\n}", "func (s *deploymentServer) createCronjob(ctx context.Context, manifest []byte, env []EnvVar, initVariables []EnvVar) error {\n\tdecoder := k8sYaml.NewYAMLOrJSONDecoder(bytes.NewReader(manifest), 1000)\n\n\tj := &apibatch.CronJob{}\n\n\tif err := decoder.Decode(&j); err != nil {\n\t\treturn err\n\t}\n\n\tif len(env) > 0 {\n\t\tcontainers := j.Spec.JobTemplate.Spec.Template.Spec.Containers\n\t\tapplyEnvironment(containers, env)\n\t}\n\n\tinitContainers := j.Spec.JobTemplate.Spec.Template.Spec.InitContainers\n\tif len(initContainers) > 0 {\n\t\tfmt.Println(\"job \" + j.Namespace + \".\" + j.Name + \" has initContainers\")\n\t\tapplyEnvironment(initContainers, initVariables)\n\t} else {\n\t\tfmt.Println(\"job \" + j.Namespace + \".\" + j.Name + \" has not initContainers; bug in config\")\n\t}\n\n\tbatchAPI := s.clientset.BatchV1beta1()\n\tapiJobs := batchAPI.CronJobs(j.Namespace)\n\n\tif _, err := apiJobs.Create(ctx, j, metav1.CreateOptions{}); err != nil {\n\t\treturn fmt.Errorf(\"job create error '%s'\", err.Error())\n\t}\n\treturn nil\n}", "func NewJob(id string, check CheckFunc, task TaskFunc) *Job {\n\treturn &Job{id, check, task}\n}", "func (c *client) startNewJob(ctx context.Context, opts launcher.LaunchOptions, jobInterface v12.JobInterface, ns string, safeName string, safeSha string) ([]runtime.Object, error) {\n\tlog.Logger().Infof(\"about to create a new job for name %s and sha %s\", safeName, safeSha)\n\n\t// lets see if we are using a version stream to store the git operator configuration\n\tfolder := filepath.Join(opts.Dir, \"versionStream\", \"git-operator\")\n\texists, err := files.DirExists(folder)\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"failed to check if folder exists %s\", folder)\n\t}\n\tif !exists {\n\t\t// lets try the original location\n\t\tfolder = filepath.Join(opts.Dir, \".jx\", \"git-operator\")\n\t}\n\n\tjobFileName := \"job.yaml\"\n\n\tfileNamePath := filepath.Join(opts.Dir, \".jx\", \"git-operator\", \"filename.txt\")\n\texists, err = files.FileExists(fileNamePath)\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"failed to check for file %s\", fileNamePath)\n\t}\n\tif exists {\n\t\tdata, err := ioutil.ReadFile(fileNamePath)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrapf(err, \"failed to load file %s\", fileNamePath)\n\t\t}\n\t\tjobFileName = strings.TrimSpace(string(data))\n\t\tif jobFileName == \"\" {\n\t\t\treturn nil, errors.Errorf(\"the job name file %s is empty\", fileNamePath)\n\t\t}\n\t}\n\n\tfileName := filepath.Join(folder, jobFileName)\n\texists, err = files.FileExists(fileName)\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"failed to find file %s in repository %s\", fileName, safeName)\n\t}\n\tif !exists {\n\t\treturn nil, errors.Errorf(\"repository %s does not have a Job file: %s\", safeName, fileName)\n\t}\n\n\tresource := &v1.Job{}\n\terr = yamls.LoadFile(fileName, resource)\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"failed to load Job file %s in repository %s\", fileName, safeName)\n\t}\n\n\tif !opts.NoResourceApply {\n\t\t// now lets check if there is a resources dir\n\t\tresourcesDir := filepath.Join(folder, \"resources\")\n\t\texists, err = files.DirExists(resourcesDir)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrapf(err, \"failed to check if resources directory %s exists in repository %s\", resourcesDir, safeName)\n\t\t}\n\t\tif exists {\n\t\t\tabsDir, err := filepath.Abs(resourcesDir)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, errors.Wrapf(err, \"failed to get absolute resources dir %s\", resourcesDir)\n\t\t\t}\n\n\t\t\tcmd := &cmdrunner.Command{\n\t\t\t\tName: \"kubectl\",\n\t\t\t\tArgs: []string{\"apply\", \"-f\", absDir},\n\t\t\t}\n\t\t\tlog.Logger().Infof(\"running command: %s\", cmd.CLI())\n\t\t\t_, err = c.runner(cmd)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, errors.Wrapf(err, \"failed to apply resources in dir %s\", absDir)\n\t\t\t}\n\t\t}\n\t}\n\n\t// lets try use a maximum of 31 characters and a minimum of 10 for the sha\n\tnamePrefix := trimLength(safeName, 20)\n\n\tid := uuid.New().String()\n\tresourceName := namePrefix + \"-\" + id\n\n\tresource.Name = resourceName\n\n\tif resource.Labels == nil {\n\t\tresource.Labels = map[string]string{}\n\t}\n\tresource.Labels[constants.DefaultSelectorKey] = constants.DefaultSelectorValue\n\tresource.Labels[launcher.RepositoryLabelKey] = safeName\n\tresource.Labels[launcher.CommitShaLabelKey] = safeSha\n\n\tr2, err := jobInterface.Create(ctx, resource, metav1.CreateOptions{})\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"failed to create Job %s in namespace %s\", resourceName, ns)\n\t}\n\tlog.Logger().Infof(\"created Job %s in namespace %s\", resourceName, ns)\n\treturn []runtime.Object{r2}, nil\n}", "func (s *grpcServer) CreateJobFunction(ctx context.Context, req *pb.CreateJobFunctionRequest) (*pb.JobFunction, error) {\n\t_, rep, err := s.createJobFunction.ServeGRPC(ctx, req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn rep.(*pb.JobFunction), nil\n}", "func NewJob(arrTime int) (j *Job) {\n\tj = new(Job)\n\tj.IntAttrs = make(map[string]int)\n\tj.StrAttrs = make(map[string]string)\n\tj.JobId = rand.Int63()\n\tj.ArrTime = arrTime\n\treturn j\n}", "func NewJob(intervel uint64) *Job {\n\treturn &Job{\n\t\tintervel, 0,\n\t\t\"\",\n\t\tJOB_UNIT_TYPE_UNKNOWN,\n\t\t\"\",\n\t\ttime.Unix(0, 0),\n\t\ttime.Unix(0, 0), 0,\n\t\tmake(map[string]interface{}),\n\t\tmake(map[string]([]interface{})),\n\t}\n}", "func new_job(w http.ResponseWriter, req *http.Request) {\n\tfmt.Println(\"Handling connection...\")\n\n\t// Parse the HTTP request.\n\tif err := req.ParseForm(); err != nil {\n\t\tpanic(err)\n\t}\n\n\t// Put the bytes from the request into a file\n\tbuf := new(bytes.Buffer)\n\tbuf.ReadFrom(req.Body)\n\tjobJson := buf.String()\n\n\t// Print out the json.\n\tfmt.Println(jobJson)\n\n\t// Convert string json to job struct\n\tjob := data.JsonToJob([]byte(jobJson))\n\n\t// Run the code and get []byte output\n\toutput := runCode(job.Extension, job.Code, job.FileName)\n\n\t// Send a response back.\n\tw.Write(output)\n}", "func createCronJobConfig(module *protos.Module) *v1beta1.CronJob {\n\tvolumes, volumeMounts := makeVolumes(module)\n\tcontainers := makeContainers(module, volumeMounts)\n\tm := module.GetSpec()\n\n\tsuccessfulJobHistoryLimit := int32(1)\n\tfailedJobsHistoryLimit := int32(1)\n\treturn &v1beta1.CronJob{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tName: module.Name,\n\t\t},\n\t\tSpec: v1beta1.CronJobSpec{\n\t\t\tSchedule: m.Schedule,\n\t\t\tJobTemplate: v1beta1.JobTemplateSpec{\n\t\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\t\tLabels: map[string]string{\n\t\t\t\t\t\t\"name\": module.Name,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tSpec: batchv1.JobSpec{\n\t\t\t\t\tTemplate: v1.PodTemplateSpec{\n\t\t\t\t\t\tSpec: v1.PodSpec{\n\t\t\t\t\t\t\tContainers: containers,\n\t\t\t\t\t\t\tVolumes: volumes,\n\t\t\t\t\t\t\tHostNetwork: true,\n\t\t\t\t\t\t\tRestartPolicy: \"Never\",\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t\tSuccessfulJobsHistoryLimit: &successfulJobHistoryLimit,\n\t\t\tFailedJobsHistoryLimit: &failedJobsHistoryLimit,\n\t\t},\n\t}\n}", "func NewJob() Job {\n\tjobConfig := aurora.NewJobConfiguration()\n\ttaskConfig := aurora.NewTaskConfig()\n\tjobKey := aurora.NewJobKey()\n\n\t// Job Config\n\tjobConfig.Key = jobKey\n\tjobConfig.TaskConfig = taskConfig\n\n\t// Task Config\n\ttaskConfig.Job = jobKey\n\ttaskConfig.Container = aurora.NewContainer()\n\ttaskConfig.Container.Mesos = aurora.NewMesosContainer()\n\n\t// Resources\n\tnumCpus := aurora.NewResource()\n\tramMb := aurora.NewResource()\n\tdiskMb := aurora.NewResource()\n\n\tresources := map[resourceType]*aurora.Resource{CPU: numCpus, RAM: ramMb, DISK: diskMb}\n\ttaskConfig.Resources = []*aurora.Resource{numCpus, ramMb, diskMb}\n\n\tnumCpus.NumCpus = new(float64)\n\tramMb.RamMb = new(int64)\n\tdiskMb.DiskMb = new(int64)\n\n\treturn &AuroraJob{\n\t\tjobConfig: jobConfig,\n\t\tresources: resources,\n\t\tmetadata: make(map[string]*aurora.Metadata),\n\t\tconstraints: make(map[string]*aurora.Constraint),\n\t\tportCount: 0,\n\t}\n}", "func NewJob(name string, f func() (bool, error)) *Job {\n\treturn &Job{Name: name, F: f}\n}", "func CronJobObjectWrapper(create CronJobCreator) ObjectCreator {\n\treturn func(existing runtime.Object) (runtime.Object, error) {\n\t\tif existing != nil {\n\t\t\treturn create(existing.(*batchv1beta1.CronJob))\n\t\t}\n\t\treturn create(&batchv1beta1.CronJob{})\n\t}\n}", "func newJob(job Runnable, priority int) JobEntry {\n\treturn &pt{\n\t\tpriority: priority,\n\t\tjob: job,\n\t\tlock: &sync.Mutex{},\n\t}\n}", "func NewCrontab() *Crontab {\n\tctb := &Crontab{\n\t\tjobs: make(map[string]*Job),\n\t\tcontrol: make(chan *crontabControl),\n\t\tticker: time.NewTicker(1e9),\n\t}\n\n\tgo ctb.backend()\n\n\treturn ctb\n}", "func NewJob(brigadeSVC brigade.Interface, logger log.Logger) subcollector {\n\treturn &job{\n\t\tbrigadeSVC: brigadeSVC,\n\t\tlogger: logger,\n\n\t\tjobInfoDesc: prometheus.NewDesc(\n\t\t\tprometheus.BuildFQName(namespace, jobSubSystem, \"info\"),\n\t\t\t\"Brigade job information.\",\n\t\t\t[]string{\"id\", \"build_id\", \"name\", \"image\"}, nil,\n\t\t),\n\t\tjobStatusDesc: prometheus.NewDesc(\n\t\t\tprometheus.BuildFQName(namespace, jobSubSystem, \"status\"),\n\t\t\t\"Brigade job status.\",\n\t\t\t[]string{\"id\", \"status\"}, nil,\n\t\t),\n\t\tjobDurationDesc: prometheus.NewDesc(\n\t\t\tprometheus.BuildFQName(namespace, jobSubSystem, \"duration_seconds\"),\n\t\t\t\"Brigade job duration in seconds.\",\n\t\t\t[]string{\"id\"}, nil,\n\t\t),\n\t\tjobCreationDesc: prometheus.NewDesc(\n\t\t\tprometheus.BuildFQName(namespace, jobSubSystem, \"create_time_seconds\"),\n\t\t\t\"Brigade job creation time in unix timestamp.\",\n\t\t\t[]string{\"id\"}, nil,\n\t\t),\n\t\tjobStartDesc: prometheus.NewDesc(\n\t\t\tprometheus.BuildFQName(namespace, jobSubSystem, \"start_time_seconds\"),\n\t\t\t\"Brigade job start time in unix timestamp.\",\n\t\t\t[]string{\"id\"}, nil,\n\t\t),\n\t}\n}", "func NewJOB(c *httpclient.Client) *JOB {\n\tclient := httpclient.New()\n\tif c != nil {\n\t\tclient = c\n\t}\n\tsg, err := time.LoadLocation(\"Asia/Singapore\")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn &JOB{\n\t\tClient: client,\n\t\tsg: sg,\n\t}\n}" ]
[ "0.67432356", "0.66305035", "0.66197085", "0.660366", "0.65776867", "0.65504235", "0.6515398", "0.64804167", "0.64739484", "0.64633006", "0.6440657", "0.64136183", "0.6394644", "0.6386326", "0.63756454", "0.6352607", "0.6352166", "0.6351604", "0.63278204", "0.63187474", "0.61663115", "0.6097529", "0.60918176", "0.60858303", "0.6056411", "0.60079664", "0.60018873", "0.59889334", "0.5987655", "0.59824294", "0.59715813", "0.59466535", "0.59411484", "0.5939134", "0.5919835", "0.59179443", "0.5909412", "0.588858", "0.5872428", "0.5867772", "0.58564395", "0.58560216", "0.5836226", "0.5821588", "0.5820258", "0.57934153", "0.57873285", "0.5771265", "0.57706875", "0.574909", "0.57475567", "0.5742802", "0.5736224", "0.57167655", "0.5711978", "0.57080704", "0.57008624", "0.5694179", "0.56782246", "0.5651923", "0.56004727", "0.5588855", "0.5588542", "0.55758214", "0.5561776", "0.5553818", "0.55499995", "0.55419195", "0.55289614", "0.5487731", "0.5477298", "0.5449078", "0.5429621", "0.5355878", "0.53549695", "0.53519845", "0.5345042", "0.5331214", "0.53261244", "0.5317761", "0.53174776", "0.52948606", "0.5278293", "0.5275118", "0.5264793", "0.5264551", "0.5263395", "0.5262155", "0.52611583", "0.52604395", "0.52541876", "0.52397174", "0.5224665", "0.52219546", "0.5209437", "0.52018744", "0.5195033", "0.5194096", "0.5167724", "0.51618254" ]
0.70655316
0
Update takes the representation of a tZCronJob and updates it. Returns the server's representation of the tZCronJob, and an error, if there is any.
func (c *FakeTZCronJobs) Update(tZCronJob *v1alpha1.TZCronJob) (result *v1alpha1.TZCronJob, err error) { obj, err := c.Fake. Invokes(testing.NewUpdateAction(tzcronjobsResource, c.ns, tZCronJob), &v1alpha1.TZCronJob{}) if obj == nil { return nil, err } return obj.(*v1alpha1.TZCronJob), err }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func UpdateJob(w http.ResponseWriter, r *http.Request) {\n\tresponse := services.UpdateJob(r)\n\n\trender.Status(r, response.Code)\n\trender.JSON(w, r, response)\n}", "func (c *cloudSchedulerRESTClient) UpdateJob(ctx context.Context, req *schedulerpb.UpdateJobRequest, opts ...gax.CallOption) (*schedulerpb.Job, error) {\n\tm := protojson.MarshalOptions{AllowPartial: true, UseEnumNumbers: true}\n\tbody := req.GetJob()\n\tjsonReq, err := m.Marshal(body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v1/%v\", req.GetJob().GetName())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\tif req.GetUpdateMask() != nil {\n\t\tupdateMask, err := protojson.Marshal(req.GetUpdateMask())\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tparams.Add(\"updateMask\", string(updateMask[1:len(updateMask)-1]))\n\t}\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"job.name\", url.QueryEscape(req.GetJob().GetName()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\topts = append((*c.CallOptions).UpdateJob[0:len((*c.CallOptions).UpdateJob):len((*c.CallOptions).UpdateJob)], opts...)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tresp := &schedulerpb.Job{}\n\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"PATCH\", baseUrl.String(), bytes.NewReader(jsonReq))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}, opts...)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\treturn resp, nil\n}", "func (c *jobsRESTClient) UpdateJob(ctx context.Context, req *runpb.UpdateJobRequest, opts ...gax.CallOption) (*UpdateJobOperation, error) {\n\tm := protojson.MarshalOptions{AllowPartial: true, UseEnumNumbers: true}\n\tbody := req.GetJob()\n\tjsonReq, err := m.Marshal(body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v2/%v\", req.GetJob().GetName())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\tif req.GetAllowMissing() {\n\t\tparams.Add(\"allowMissing\", fmt.Sprintf(\"%v\", req.GetAllowMissing()))\n\t}\n\tif req.GetValidateOnly() {\n\t\tparams.Add(\"validateOnly\", fmt.Sprintf(\"%v\", req.GetValidateOnly()))\n\t}\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"job.name\", url.QueryEscape(req.GetJob().GetName()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tresp := &longrunningpb.Operation{}\n\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"PATCH\", baseUrl.String(), bytes.NewReader(jsonReq))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}, opts...)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\n\toverride := fmt.Sprintf(\"/v2/%s\", resp.GetName())\n\treturn &UpdateJobOperation{\n\t\tlro: longrunning.InternalNewOperation(*c.LROClient, resp),\n\t\tpollPath: override,\n\t}, nil\n}", "func UpdateJob(cmd *cobra.Command, args []string) error {\n\n\tclient, err := auth.GetClient()\n\tif err != nil {\n\t\treturn err\n\t}\n\t// Parse all flags\n\n\tvar sid string\n\terr = flags.ParseFlag(cmd.Flags(), \"sid\", &sid)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"sid\": ` + err.Error())\n\t}\n\tvar status model.UpdateJobStatus\n\terr = flags.ParseFlag(cmd.Flags(), \"status\", &status)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"status\": ` + err.Error())\n\t}\n\t// Form the request body\n\tgenerated_request_body := model.UpdateJob{\n\n\t\tStatus: status,\n\t}\n\n\t// Silence Usage\n\tcmd.SilenceUsage = true\n\n\tresp, err := client.SearchService.UpdateJob(sid, generated_request_body)\n\tif err != nil {\n\t\treturn err\n\t}\n\tjsonx.Pprint(cmd, resp)\n\treturn nil\n}", "func (c *scheduledJobs) Update(job *batch.ScheduledJob) (result *batch.ScheduledJob, err error) {\n\tresult = &batch.ScheduledJob{}\n\terr = c.r.Put().Namespace(c.ns).Resource(\"scheduledjobs\").Name(job.Name).Body(job).Do().Into(result)\n\treturn\n}", "func (client JobClient) Update(ctx context.Context, resourceGroupName string, accountName string, jobName string, body JobResourcePatchDescription) (result JobUpdateFuture, err error) {\n if tracing.IsEnabled() {\n ctx = tracing.StartSpan(ctx, fqdn + \"/JobClient.Update\")\n defer func() {\n sc := -1\n if result.Response() != nil {\n sc = result.Response().StatusCode\n }\n tracing.EndSpan(ctx, sc, err)\n }()\n }\n req, err := client.UpdatePreparer(ctx, resourceGroupName, accountName, jobName, body)\n if err != nil {\n err = autorest.NewErrorWithError(err, \"microsoftazuremanagementaisupercomputer.JobClient\", \"Update\", nil , \"Failure preparing request\")\n return\n }\n\n result, err = client.UpdateSender(req)\n if err != nil {\n err = autorest.NewErrorWithError(err, \"microsoftazuremanagementaisupercomputer.JobClient\", \"Update\", result.Response(), \"Failure sending request\")\n return\n }\n\n return\n}", "func UpdateJobTask(w http.ResponseWriter, r *http.Request) {\n\tresponse := services.UpdateJobTask(r)\n\n\trender.Status(r, response.Code)\n\trender.JSON(w, r, response)\n}", "func (o *Job) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tjobUpdateCacheMut.RLock()\n\tcache, cached := jobUpdateCache[key]\n\tjobUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tjobColumns,\n\t\t\tjobPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update jobs, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"jobs\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, jobPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(jobType, jobMapping, append(wl, jobPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update jobs row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for jobs\")\n\t}\n\n\tif !cached {\n\t\tjobUpdateCacheMut.Lock()\n\t\tjobUpdateCache[key] = cache\n\t\tjobUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (r *JobRepo) Update(data *model.Job) (*model.Job, error) {\n\tdb, err := db.Instance()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\terr = db.Model(data).Updates(map[string]interface{}{\"status\": data.Status, \"start_time\": data.StartTime, \"end_time\": data.EndTime, \"log\": data.Log}).Error\n\treturn data, err\n}", "func UpdateJob(cmd CmdInterface) {\n\tjob, err := store.FindJob(cmd.Parts[1])\n\tif err != nil {\n\t\tReturnError(cmd, err.Error())\n\t\treturn\n\t}\n\n\tif cmd.Parts[2] == store.JobInProgress {\n\t\tjob.State = store.JobInProgress\n\t\tjob.StartedProcessingAt = time.Now()\n\t} else if cmd.Parts[2] == store.JobDone {\n\t\tjob.State = store.JobDone\n\t\tjob.EndProcessingAt = time.Now()\n\t} else if cmd.Parts[2] == store.JobErrored {\n\t\tjob.State = store.JobErrored\n\t\tjob.Attempts = job.Attempts + 1\n\t} else {\n\t\tReturnError(cmd, \"unknown-state\")\n\t\treturn\n\t}\n\n\tstore.UpdateSession(cmd.User)\n\tjob.Update()\n\n\tdelay := -1\n\tif cmd.Parts[2] == store.JobInProgress {\n\t\t// set the worker as busy\n\t\tcmd.User.Status = store.WorkerBusy\n\t} else {\n\t\t// free this worker\n\t\tcmd.User.Status = store.WorkerAvailable\n\t\t// see if this worker can work again\n\t\tdelay = 2 // in seconds\n\t\tif job.State == store.JobErrored {\n\t\t\tdelay = 1800 // wait 30 minutes before trying again\n\t\t}\n\t\tif store.Options.DelayPolicy == \"IGNORE\" {\n\t\t\tdelay = 0\n\t\t}\n\t}\n\n\t// update the worker status\n\tfor key, worker := range job.Queue.Workers {\n\t\tif worker.Addr == cmd.User.Addr {\n\t\t\tjob.Queue.Workers[key] = cmd.User\n\t\t}\n\t}\n\tjob.Queue.UpdateAndKeep([]string{\"Jobs\"})\n\n\tif delay != -1 {\n\t\tgo DispatchUniversalWithDelay(delay)\n\t}\n\n\tif job.State == store.JobDone {\n\t\tgo job.Expire(store.Options.JobTimeout)\n\t}\n\n\tReturnString(cmd, \"OK\")\n}", "func (c *FakeTZCronJobs) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1alpha1.TZCronJob, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewPatchSubresourceAction(tzcronjobsResource, c.ns, name, pt, data, subresources...), &v1alpha1.TZCronJob{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\treturn obj.(*v1alpha1.TZCronJob), err\n}", "func (db jobDatabase) UpdateJob(job mgm.Job) error {\n\tcon, err := db.mysql.GetConnection()\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer con.Close()\n\n\t//The function states update job, but only the data field ever changes\n\t_, err = con.Exec(\"UPDATE jobs SET data=?\", job.Data)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *Jobs) UpdateJob(id int, in *JobIn) (out *JobOut, err error) {\n\tbody, err := c.call(\"PUT\", fmt.Sprintf(\"/jobs/%d\", id), in)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer body.Close()\n\n\terr = json.NewDecoder(body).Decode(&out)\n\treturn\n}", "func (j *TrainingJob) Update(newJob *spec.MxJob) {\n\tj.send(&jobEvent{\n\t\ttyp: eventModifyJob,\n\t\tcluster: newJob,\n\t})\n}", "func CreateOrUpdate(ctx context.Context, c client.Client, cj *batchv1.CronJob, equal EqualityFunc, mutate MutateFunc) error {\n\tcurrent := &batchv1.CronJob{}\n\tkey := client.ObjectKey{Name: cj.Name, Namespace: cj.Namespace}\n\terr := c.Get(ctx, key, current)\n\tif err != nil {\n\t\tif apierrors.IsNotFound(err) {\n\t\t\terr = c.Create(ctx, cj)\n\n\t\t\tif err == nil {\n\t\t\t\treturn nil\n\t\t\t}\n\n\t\t\treturn kverrors.Wrap(err, \"failed to create cronjob\",\n\t\t\t\t\"name\", cj.Name,\n\t\t\t\t\"namespace\", cj.Namespace,\n\t\t\t)\n\t\t}\n\n\t\treturn kverrors.Wrap(err, \"failed to get cronjob\",\n\t\t\t\"name\", cj.Name,\n\t\t\t\"namespace\", cj.Namespace,\n\t\t)\n\t}\n\n\tif !equal(current, cj) {\n\t\terr := retry.RetryOnConflict(retry.DefaultRetry, func() error {\n\t\t\tif err := c.Get(ctx, key, current); err != nil {\n\t\t\t\treturn kverrors.Wrap(err, \"failed to get cronjob\",\n\t\t\t\t\t\"name\", cj.Name,\n\t\t\t\t\t\"namespace\", cj.Namespace,\n\t\t\t\t)\n\t\t\t}\n\n\t\t\tmutate(current, cj)\n\t\t\tif err := c.Update(ctx, current); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\t\tif err != nil {\n\t\t\treturn kverrors.Wrap(err, \"failed to update cronjob\",\n\t\t\t\t\"name\", cj.Name,\n\t\t\t\t\"namespace\", cj.Namespace,\n\t\t\t)\n\t\t}\n\t\treturn nil\n\t}\n\n\treturn nil\n}", "func (e *engineImpl) updateJob(c context.Context, def catalog.Definition) error {\n\treturn e.txn(c, def.JobID, func(c context.Context, job *CronJob, isNew bool) error {\n\t\tif !isNew && job.Enabled && job.matches(def) {\n\t\t\treturn errSkipPut\n\t\t}\n\t\tif isNew {\n\t\t\t// JobID is <projectID>/<name>, it's ensure by Catalog.\n\t\t\tchunks := strings.Split(def.JobID, \"/\")\n\t\t\tif len(chunks) != 2 {\n\t\t\t\treturn fmt.Errorf(\"unexpected jobID format: %s\", def.JobID)\n\t\t\t}\n\t\t\t*job = CronJob{\n\t\t\t\tJobID: def.JobID,\n\t\t\t\tProjectID: chunks[0],\n\t\t\t\tEnabled: false, // to trigger 'if !oldEnabled' below\n\t\t\t\tSchedule: def.Schedule,\n\t\t\t\tTask: def.Task,\n\t\t\t\tState: JobState{State: JobStateDisabled},\n\t\t\t}\n\t\t}\n\t\toldEnabled := job.Enabled\n\t\toldEffectiveSchedule := job.effectiveSchedule()\n\n\t\t// Update the job in full before running any state changes.\n\t\tjob.Revision = def.Revision\n\t\tjob.RevisionURL = def.RevisionURL\n\t\tjob.Enabled = true\n\t\tjob.Schedule = def.Schedule\n\t\tjob.Task = def.Task\n\n\t\t// Do state machine transitions.\n\t\tif !oldEnabled {\n\t\t\terr := e.rollSM(c, job, func(sm *StateMachine) error { return sm.OnJobEnabled() })\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t\tif job.effectiveSchedule() != oldEffectiveSchedule {\n\t\t\tlogging.Infof(c, \"Job's schedule changed\")\n\t\t\treturn e.rollSM(c, job, func(sm *StateMachine) error { return sm.OnScheduleChange() })\n\t\t}\n\t\treturn nil\n\t})\n}", "func (m *Mongo) UpdateJob(ctx context.Context, id string, job *models.Job) (err error) {\n\treturn m.updateByID(ctx, id, bson.M{\n\t\t\"$set\": job,\n\t\t\"$currentDate\": bson.M{\n\t\t\t\"last_updated\": true,\n\t\t\t\"unique_timestamp\": bson.M{\n\t\t\t\t\"$type\": \"timestamp\",\n\t\t\t},\n\t\t},\n\t})\n}", "func (h *CronJobUpdateHandler) Handle(ctx context.Context, req types.Request) types.Response {\n\tobj := &batchv1.CronJob{}\n\n\terr := h.Decoder.Decode(req, obj)\n\tif err != nil {\n\t\treturn admission.ErrorResponse(http.StatusBadRequest, err)\n\t}\n\n\tallowed, reason, err := h.validatingCronJobFn(ctx, obj)\n\tif err != nil {\n\t\treturn admission.ErrorResponse(http.StatusInternalServerError, err)\n\t}\n\treturn admission.ValidationResponse(allowed, reason)\n}", "func (db DB) UpdateJob(j *Job) error {\n\ts := db.session.Copy()\n\tdefer s.Close()\n\treturn s.DB(app).C(\"jobs\").Update(j.ID, j)\n}", "func (wj *WsubJob) Update(db XODB) error {\n\tvar err error\n\n\t// if doesn't exist, bail\n\tif !wj._exists {\n\t\treturn errors.New(\"update failed: does not exist\")\n\t}\n\n\t// if deleted, bail\n\tif wj._deleted {\n\t\treturn errors.New(\"update failed: marked for deletion\")\n\t}\n\n\t// sql query\n\tconst sqlstr = `UPDATE jlabo.wsub_jobs SET ` +\n\t\t`type = ?, alias = ?, date_start = ?, date_reply = ?, description_fr_FR = ?, description_en_GB = ?, checked_out = ?, checked_out_time = ?, title_fr_FR = ?, title_en_GB = ?, state = ?, created = ?, created_by = ?, modified = ?, modified_by = ?, publish_up = ?, publish_down = ?, attribs = ?, misc_fr_FR = ?, misc_en_GB = ?, logo1 = ?, logo2 = ?, logo3 = ?, logo4 = ?, logo5 = ?, group = ?, keywords = ?` +\n\t\t` WHERE id = ?`\n\n\t// run query\n\tXOLog(sqlstr, wj.Type, wj.Alias, wj.DateStart, wj.DateReply, wj.DescriptionFrFr, wj.DescriptionEnGb, wj.CheckedOut, wj.CheckedOutTime, wj.TitleFrFr, wj.TitleEnGb, wj.State, wj.Created, wj.CreatedBy, wj.Modified, wj.ModifiedBy, wj.PublishUp, wj.PublishDown, wj.Attribs, wj.MiscFrFr, wj.MiscEnGb, wj.Logo1, wj.Logo2, wj.Logo3, wj.Logo4, wj.Logo5, wj.Group, wj.Keywords, wj.ID)\n\t_, err = db.Exec(sqlstr, wj.Type, wj.Alias, wj.DateStart, wj.DateReply, wj.DescriptionFrFr, wj.DescriptionEnGb, wj.CheckedOut, wj.CheckedOutTime, wj.TitleFrFr, wj.TitleEnGb, wj.State, wj.Created, wj.CreatedBy, wj.Modified, wj.ModifiedBy, wj.PublishUp, wj.PublishDown, wj.Attribs, wj.MiscFrFr, wj.MiscEnGb, wj.Logo1, wj.Logo2, wj.Logo3, wj.Logo4, wj.Logo5, wj.Group, wj.Keywords, wj.ID)\n\treturn err\n}", "func (r *DeviceManagementExportJobRequest) Update(ctx context.Context, reqObj *DeviceManagementExportJob) error {\n\treturn r.JSONRequest(ctx, \"PATCH\", \"\", reqObj, nil)\n}", "func (r *ExactMatchLookupJobRequest) Update(ctx context.Context, reqObj *ExactMatchLookupJob) error {\n\treturn r.JSONRequest(ctx, \"PATCH\", \"\", reqObj, nil)\n}", "func (c *restClient) ExecutePatchJob(ctx context.Context, req *osconfigpb.ExecutePatchJobRequest, opts ...gax.CallOption) (*osconfigpb.PatchJob, error) {\n\tm := protojson.MarshalOptions{AllowPartial: true, UseEnumNumbers: true}\n\tjsonReq, err := m.Marshal(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v1beta/%v/patchJobs:execute\", req.GetParent())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"parent\", url.QueryEscape(req.GetParent()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\topts = append((*c.CallOptions).ExecutePatchJob[0:len((*c.CallOptions).ExecutePatchJob):len((*c.CallOptions).ExecutePatchJob)], opts...)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tresp := &osconfigpb.PatchJob{}\n\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"POST\", baseUrl.String(), bytes.NewReader(jsonReq))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}, opts...)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\treturn resp, nil\n}", "func (s *Client) ProcessUpdateJob(job *params.UpdateJob) (client.Response, error) {\n\tjob.Job = \"update\"\n\treturn s.c.Post(job.GetEndpoint(), job)\n}", "func (c *FakeTZCronJobs) UpdateStatus(tZCronJob *v1alpha1.TZCronJob) (*v1alpha1.TZCronJob, error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewUpdateSubresourceAction(tzcronjobsResource, \"status\", c.ns, tZCronJob), &v1alpha1.TZCronJob{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\treturn obj.(*v1alpha1.TZCronJob), err\n}", "func (c *JobClient) Update() *JobUpdate {\n\tmutation := newJobMutation(c.config, OpUpdate)\n\treturn &JobUpdate{config: c.config, hooks: c.Hooks(), mutation: mutation}\n}", "func (c *aITrainingJobs) Update(aITrainingJob *v1.AITrainingJob) (result *v1.AITrainingJob, err error) {\n\tresult = &v1.AITrainingJob{}\n\terr = c.client.Put().\n\t\tNamespace(c.ns).\n\t\tResource(\"aitrainingjobs\").\n\t\tName(aITrainingJob.Name).\n\t\tBody(aITrainingJob).\n\t\tDo().\n\t\tInto(result)\n\treturn\n}", "func (c APIClient) UpdateMonitoringJob(mj *MonitoringJob) error {\n\treturn c.doHTTPBoth(\"POST\", fmt.Sprintf(\"https://api.nsone.net/v1/monitoring/jobs/%s\", mj.Id), mj)\n}", "func PatchJob(ns, name string, updateFunc func(job *v1.Job)) error {\n\toJob, err := kubecli.BatchV1().Jobs(ns).Get(name, metav1.GetOptions{})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tnJob := cloneJob(oJob)\n\tupdateFunc(nJob)\n\tpatchData, err := CreatePatch(oJob, nJob, v1.Job{})\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = kubecli.BatchV1().Jobs(ns).Patch(name, types.StrategicMergePatchType, patchData)\n\treturn err\n}", "func (c *CloudSchedulerClient) UpdateJob(ctx context.Context, req *schedulerpb.UpdateJobRequest, opts ...gax.CallOption) (*schedulerpb.Job, error) {\n\treturn c.internalClient.UpdateJob(ctx, req, opts...)\n}", "func (c *aITrainingJobs) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.AITrainingJob, err error) {\n\tresult = &v1.AITrainingJob{}\n\terr = c.client.Patch(pt).\n\t\tNamespace(c.ns).\n\t\tResource(\"aitrainingjobs\").\n\t\tSubResource(subresources...).\n\t\tName(name).\n\t\tBody(data).\n\t\tDo().\n\t\tInto(result)\n\treturn\n}", "func (d *JobDoc) ToJob() *jobs.Job {\n\treturn &jobs.Job{\n\t\tID: d.ID,\n\t\tName: d.Name,\n\t\tCreatedAt: d.CreatedAt,\n\t\tUpdatedAt: d.UpdatedAt,\n\t}\n}", "func (j *RawJob) toJob() (Job, error) {\n\tjob, jobInput, err := j.decodeSerializedFields()\n\tif err != nil {\n\t\treturn Job{}, err\n\t}\n\tresult := Job{\n\t\tID: j.ID,\n\t\tGroupID: j.GroupID,\n\t\tSuperGroupID: j.SuperGroupID,\n\t\tcronID: j.CronID,\n\t\tCronExpression: j.CronExpression,\n\t\tpaused: j.Paused,\n\t\tcreatedAt: j.CreatedAt,\n\t\tupdatedAt: j.UpdatedAt,\n\t\tJob: job,\n\t\tJobInput: jobInput.OtherInputs,\n\t}\n\treturn result, nil\n}", "func (c *Controller) updateCaffe2Job(old, cur interface{}) {\n\toldCaffe2Job := old.(*api.Caffe2Job)\n\tglog.Infof(\"Updating caffe2job: %s\", oldCaffe2Job.Name)\n\tc.enqueueController(cur)\n}", "func (a *SyncApiService) UpdateSyncJob(ctx context.Context, syncJob SyncJob, syncJobId string) ( *http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Put\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/platform/3/sync/jobs/{SyncJobId}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"SyncJobId\"+\"}\", fmt.Sprintf(\"%v\", syncJobId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{ \"application/json\", }\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\n\t\t\"application/json\",\n\t\t}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &syncJob\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\tdefer localVarHttpResponse.Body.Close()\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tbodyBytes, _ := ioutil.ReadAll(localVarHttpResponse.Body)\n\t\treturn localVarHttpResponse, reportError(\"Status: %v, Body: %s\", localVarHttpResponse.Status, bodyBytes)\n\t}\n\n\treturn localVarHttpResponse, err\n}", "func (r UpdateReplicationJobRequest) Send(ctx context.Context) (*UpdateReplicationJobOutput, error) {\n\tr.Request.SetContext(ctx)\n\terr := r.Request.Send()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn r.Request.Data.(*UpdateReplicationJobOutput), nil\n}", "func UnmarshalJob(m map[string]json.RawMessage, result interface{}) (err error) {\n\tobj := new(Job)\n\terr = core.UnmarshalPrimitive(m, \"command_object\", &obj.CommandObject)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"command_object_id\", &obj.CommandObjectID)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"command_name\", &obj.CommandName)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"command_parameter\", &obj.CommandParameter)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"command_options\", &obj.CommandOptions)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalModel(m, \"inputs\", &obj.Inputs, UnmarshalVariableData)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalModel(m, \"settings\", &obj.Settings, UnmarshalVariableData)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"tags\", &obj.Tags)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"id\", &obj.ID)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"name\", &obj.Name)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"description\", &obj.Description)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"location\", &obj.Location)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"resource_group\", &obj.ResourceGroup)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"submitted_at\", &obj.SubmittedAt)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"submitted_by\", &obj.SubmittedBy)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"start_at\", &obj.StartAt)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"end_at\", &obj.EndAt)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"duration\", &obj.Duration)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalModel(m, \"status\", &obj.Status, UnmarshalJobStatus)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalModel(m, \"data\", &obj.Data, UnmarshalJobData)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"targets_ini\", &obj.TargetsIni)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalModel(m, \"bastion\", &obj.Bastion, UnmarshalTargetResourceset)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalModel(m, \"log_summary\", &obj.LogSummary, UnmarshalJobLogSummary)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"log_store_url\", &obj.LogStoreURL)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"state_store_url\", &obj.StateStoreURL)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"results_url\", &obj.ResultsURL)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"updated_at\", &obj.UpdatedAt)\n\tif err != nil {\n\t\treturn\n\t}\n\treflect.ValueOf(result).Elem().Set(reflect.ValueOf(obj))\n\treturn\n}", "func (d *defaultJobRepository) PatchJobStatus(ctxIn context.Context, jobPatcher JobPatch) error {\n _, span := trace.StartSpan(ctxIn, \"(*defaultJobRepository).PatchJobStatus\")\n defer span.End()\n\n job := &Job{\n Status: jobPatcher.Status,\n ForeignJobID: ForeignJobID{\n BigQueryID: jobPatcher.ForeignJobID.BigQueryID,\n CloudStorageID: jobPatcher.ForeignJobID.CloudStorageID,\n },\n EntityAudit: EntityAudit{\n UpdatedTimestamp: time.Now(),\n },\n }\n\n _, err := d.storageService.DB().Model(job).\n Column(\"status\", \"audit_updated_timestamp\", \"bigquery_extract_job_id\", \"cloudstorage_transfer_job_id\").\n Where(\"audit_deleted_timestamp IS NULL\").\n Where(\"id = ?\", jobPatcher.ID).\n Update()\n\n if err != nil {\n return fmt.Errorf(\"error during executing updating job statement: %s\", err)\n }\n\n return nil\n}", "func (future *JobsUpdateFuture) Result(client JobsClient) (jr JobResource, err error) {\n\tvar done bool\n\tdone, err = future.Done(client)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"databox.JobsUpdateFuture\", \"Result\", future.Response(), \"Polling failure\")\n\t\treturn\n\t}\n\tif !done {\n\t\terr = azure.NewAsyncOpIncompleteError(\"databox.JobsUpdateFuture\")\n\t\treturn\n\t}\n\tsender := autorest.DecorateSender(client, autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n\tif jr.Response.Response, err = future.GetResult(sender); err == nil && jr.Response.Response.StatusCode != http.StatusNoContent {\n\t\tjr, err = client.UpdateResponder(jr.Response.Response)\n\t\tif err != nil {\n\t\t\terr = autorest.NewErrorWithError(err, \"databox.JobsUpdateFuture\", \"Result\", jr.Response.Response, \"Failure responding to request\")\n\t\t}\n\t}\n\treturn\n}", "func (c *Client) ExecutePatchJob(ctx context.Context, req *osconfigpb.ExecutePatchJobRequest, opts ...gax.CallOption) (*osconfigpb.PatchJob, error) {\n\treturn c.internalClient.ExecutePatchJob(ctx, req, opts...)\n}", "func (j *ScheduledJob) Update(ctx context.Context, ex sqlutil.InternalExecutor, txn *kv.Txn) error {\n\tif !j.isDirty() {\n\t\treturn nil\n\t}\n\n\tif j.rec.ScheduleID == 0 {\n\t\treturn errors.New(\"cannot update schedule: missing schedule id\")\n\t}\n\n\tcols, qargs, err := j.marshalChanges()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif len(qargs) == 0 {\n\t\treturn nil // Nothing changed.\n\t}\n\n\tn, err := ex.ExecEx(ctx, \"sched-update\", txn,\n\t\tsessiondata.InternalExecutorOverride{User: security.RootUserName()},\n\t\tfmt.Sprintf(\"UPDATE %s SET (%s) = (%s) WHERE schedule_id = %d\",\n\t\t\tj.env.ScheduledJobsTableName(), strings.Join(cols, \",\"),\n\t\t\tgeneratePlaceholders(len(qargs)), j.ScheduleID()),\n\t\tqargs...,\n\t)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif n != 1 {\n\t\treturn fmt.Errorf(\"expected to update 1 schedule, updated %d instead\", n)\n\t}\n\n\treturn nil\n}", "func (r *ReconcileBackup) updateCronJobStatus(request reconcile.Request) error {\n\tbkp, err := service.FetchBackupCR(request.Name, request.Namespace, r.client)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Check if Cronjob Name or Status was changed, if yes update it\n\tcronJob, err := service.FetchCronJob(bkp.Name, bkp.Namespace, r.client)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Check if CronJob changed, if yes update its status\n\tif err := r.insertUpdateCronJobStatus(cronJob, bkp); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (s *grpcServer) UpdateJobPost(ctx context.Context, req *pb.UpdateJobPostRequest) (*pb.JobPost, error) {\n\t_, rep, err := s.updateJobPost.ServeGRPC(ctx, req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn rep.(*pb.JobPost), nil\n}", "func (c *FakeTZCronJobs) Get(name string, options v1.GetOptions) (result *v1alpha1.TZCronJob, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewGetAction(tzcronjobsResource, c.ns, name), &v1alpha1.TZCronJob{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\treturn obj.(*v1alpha1.TZCronJob), err\n}", "func (m *DataHistoryManager) UpsertJob(job *DataHistoryJob, insertOnly bool) error {\n\tif m == nil {\n\t\treturn ErrNilSubsystem\n\t}\n\tif !m.IsRunning() {\n\t\treturn ErrSubSystemNotStarted\n\t}\n\tif job == nil {\n\t\treturn errNilJob\n\t}\n\tif job.Nickname == \"\" {\n\t\treturn fmt.Errorf(\"upsert job %w\", errNicknameUnset)\n\t}\n\n\tj, err := m.GetByNickname(job.Nickname, false)\n\tif err != nil && !errors.Is(err, errJobNotFound) {\n\t\treturn err\n\t}\n\n\tif insertOnly && j != nil ||\n\t\t(j != nil && j.Status != dataHistoryStatusActive) {\n\t\treturn fmt.Errorf(\"upsert job %w nickname: %s - status: %s \", errNicknameInUse, j.Nickname, j.Status)\n\t}\n\n\tm.m.Lock()\n\tdefer m.m.Unlock()\n\n\terr = m.validateJob(job)\n\tif err != nil {\n\t\treturn err\n\t}\n\ttoUpdate := false\n\tif !insertOnly {\n\t\tfor i := range m.jobs {\n\t\t\tif !strings.EqualFold(m.jobs[i].Nickname, job.Nickname) {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\ttoUpdate = true\n\t\t\tjob.ID = m.jobs[i].ID\n\t\t\tif job.Exchange != \"\" && m.jobs[i].Exchange != job.Exchange {\n\t\t\t\tm.jobs[i].Exchange = job.Exchange\n\t\t\t}\n\t\t\tif job.Asset != \"\" && m.jobs[i].Asset != job.Asset {\n\t\t\t\tm.jobs[i].Asset = job.Asset\n\t\t\t}\n\t\t\tif !job.Pair.IsEmpty() && !m.jobs[i].Pair.Equal(job.Pair) {\n\t\t\t\tm.jobs[i].Pair = job.Pair\n\t\t\t}\n\t\t\tif !job.StartDate.IsZero() && !m.jobs[i].StartDate.Equal(job.StartDate) {\n\t\t\t\tm.jobs[i].StartDate = job.StartDate\n\t\t\t}\n\t\t\tif !job.EndDate.IsZero() && !m.jobs[i].EndDate.Equal(job.EndDate) {\n\t\t\t\tm.jobs[i].EndDate = job.EndDate\n\t\t\t}\n\t\t\tif job.Interval != 0 && m.jobs[i].Interval != job.Interval {\n\t\t\t\tm.jobs[i].Interval = job.Interval\n\t\t\t}\n\t\t\tif job.RunBatchLimit != 0 && m.jobs[i].RunBatchLimit != job.RunBatchLimit {\n\t\t\t\tm.jobs[i].RunBatchLimit = job.RunBatchLimit\n\t\t\t}\n\t\t\tif job.RequestSizeLimit != 0 && m.jobs[i].RequestSizeLimit != job.RequestSizeLimit {\n\t\t\t\tm.jobs[i].RequestSizeLimit = job.RequestSizeLimit\n\t\t\t}\n\t\t\tif job.MaxRetryAttempts != 0 && m.jobs[i].MaxRetryAttempts != job.MaxRetryAttempts {\n\t\t\t\tm.jobs[i].MaxRetryAttempts = job.MaxRetryAttempts\n\t\t\t}\n\t\t\tm.jobs[i].DataType = job.DataType\n\t\t\tm.jobs[i].Status = job.Status\n\t\t\tbreak\n\t\t}\n\t}\n\tif job.ID == uuid.Nil {\n\t\tjob.ID, err = uuid.NewV4()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tjob.rangeHolder, err = kline.CalculateCandleDateRanges(job.StartDate, job.EndDate, job.Interval, uint32(job.RequestSizeLimit))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif !toUpdate {\n\t\tm.jobs = append(m.jobs, job)\n\t}\n\n\tdbJob := m.convertJobToDBModel(job)\n\treturn m.jobDB.Upsert(dbJob)\n}", "func (s *deploymentServer) updateCronjob(ctx context.Context, job *apibatch.CronJob, repo *Repository) (bool, error) {\n\tcontainers := job.Spec.JobTemplate.Spec.Template.Spec.InitContainers\n\n\tif len(containers) > 0 {\n\t\tfmt.Println(\"job \" + job.Namespace + \".\" + job.Name + \" has initContainers\")\n\t} else {\n\t\tfmt.Println(\"job \" + job.Namespace + \".\" + job.Name + \" has not initContainers; can not update\")\n\t}\n\n\tvar grace int64 = 5\n\tpodsAPI := s.clientset.CoreV1().Pods(job.Namespace)\n\tif err := podsAPI.DeleteCollection(\n\t\tctx, metav1.DeleteOptions{GracePeriodSeconds: &grace},\n\t\tmetav1.ListOptions{LabelSelector: \"sia-app=\" + job.Name}); err != nil {\n\t\treturn false, fmt.Errorf(\"could not find and delete pods for restart: %v\", err)\n\t}\n\n\treturn true, nil\n}", "func (j *JobWithSchedule) BuildJob() (RawJob, error) {\n\tvar bufferJob bytes.Buffer\n\tencoderJob := gob.NewEncoder(&bufferJob)\n\t// encode the CronJob interface keeping the unit of work\n\t// to execute. We need to use the encodeJob method\n\t// due to how gob interface encoding works.\n\tif err := encodeJob(encoderJob, j.run); err != nil {\n\t\treturn RawJob{}, err\n\t}\n\t// finally, encode the bytes to base64\n\tj.rawJob.SerializedJob = base64.StdEncoding.EncodeToString(bufferJob.Bytes())\n\n\t// now, encode the job input\n\tif err := j.encodeJobInput(); err != nil {\n\t\treturn RawJob{}, err\n\t}\n\treturn j.rawJob, nil\n}", "func (c *jobsRESTClient) UpdateJobOperation(name string) *UpdateJobOperation {\n\toverride := fmt.Sprintf(\"/v2/%s\", name)\n\treturn &UpdateJobOperation{\n\t\tlro: longrunning.InternalNewOperation(*c.LROClient, &longrunningpb.Operation{Name: name}),\n\t\tpollPath: override,\n\t}\n}", "func (jh *JobHistory) Update() *JobHistoryUpdateOne {\n\treturn (&JobHistoryClient{config: jh.config}).UpdateOne(jh)\n}", "func (a *Scheduler) Run(job *jobs.Job) error {\n\tlogrus.Debugf(\"Processing job: %+v\", job)\n\n\t// If the compliance database says the thing is already running,\n\t// we'll try to insert it into cereal once to make sure it is correct\n\t// Otherwise, compliance has completed and we're waiting for cereal\n\t// to agree\n\tshouldRetry := true\n\tif job.Status == types.StatusRunning {\n\t\tshouldRetry = false\n\t\tlogrus.Warnf(\"job %q (%q) already running\", job.Id, job.Name)\n\t}\n\t// If the job has a recurrence, we update the job schedule\n\tif job.Recurrence != \"\" {\n\t\t// Ensure recurrence rule can be parsed\n\t\t_, err := rrule.StrToRRule(job.Recurrence)\n\t\tif err != nil {\n\t\t\treturn &errorutils.InvalidError{Msg: fmt.Sprintf(\"failed to schedule job %q (%q) invalid job recurrence rule: %v\",\n\t\t\t\tjob.Id, job.Name, err)}\n\t\t}\n\t\terr = a.scanner.UpdateParentJobSchedule(job.Id, job.JobCount, job.Recurrence, job.ScheduledTime)\n\t\tif err != nil {\n\t\t\tlogrus.Errorf(\"error updating status for job %s (%s) : %s\", job.Name, job.Id, err.Error())\n\t\t}\n\t\treturn nil\n\t}\n\n\terr := a.pushWorkflow(job, shouldRetry)\n\tif err != nil {\n\t\tstrErr := fmt.Sprintf(\"Unable to add jobs to inspec agent: %s\", err.Error())\n\t\tlogrus.Error(strErr)\n\t\treturn errors.New(strErr)\n\t}\n\treturn nil\n}", "func (d Dispatcher) Job(id string) (string, error) {\n\tj, err := d.GetBC().FindJob(id)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tjBytes, err := helpers.Serialize(j)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn string(jBytes), nil\n}", "func (client JobClient) UpdateResponder(resp *http.Response) (result JobResourceDescription, err error) {\n err = autorest.Respond(\n resp,\n azure.WithErrorUnlessStatusCode(http.StatusOK,http.StatusCreated,http.StatusAccepted),\n autorest.ByUnmarshallingJSON(&result),\n autorest.ByClosing())\n result.Response = autorest.Response{Response: resp}\n return\n }", "func (c *JobClient) UpdateOne(j *Job) *JobUpdateOne {\n\tmutation := newJobMutation(c.config, OpUpdateOne, withJob(j))\n\treturn &JobUpdateOne{config: c.config, hooks: c.Hooks(), mutation: mutation}\n}", "func (c *jobsGRPCClient) UpdateJobOperation(name string) *UpdateJobOperation {\n\treturn &UpdateJobOperation{\n\t\tlro: longrunning.InternalNewOperation(*c.LROClient, &longrunningpb.Operation{Name: name}),\n\t}\n}", "func (me *PROTECTIONJOBS_IMPL) UpdateProtectionJob (\r\n body *models.ProtectionJobRequest,\r\n id int64) (*models.ProtectionJob, error) {\r\n//validating required parameters\r\n if (body == nil){\r\n return nil,errors.New(\"The parameter 'body' is a required parameter and cannot be nil.\")\r\n} //the endpoint path uri\r\n _pathUrl := \"/public/protectionJobs/{id}\"\r\n\r\n //variable to hold errors\r\n var err error = nil\r\n //process optional template parameters\r\n _pathUrl, err = apihelper.AppendUrlWithTemplateParameters(_pathUrl, map[string]interface{} {\r\n \"id\" : id,\r\n })\r\n if err != nil {\r\n //error in template param handling\r\n return nil, err\r\n }\r\n\r\n //the base uri for api requests\r\n _queryBuilder := configuration.GetBaseURI(configuration.DEFAULT_HOST,me.config);\r\n\r\n //prepare query string for API call\r\n _queryBuilder = _queryBuilder + _pathUrl\r\n\r\n //validate and preprocess url\r\n _queryBuilder, err = apihelper.CleanUrl(_queryBuilder)\r\n if err != nil {\r\n //error in url validation or cleaning\r\n return nil, err\r\n }\r\n if me.config.AccessToken() == nil {\r\n return nil, errors.New(\"Access Token not set. Please authorize the client using client.Authorize()\");\r\n }\r\n //prepare headers for the outgoing request\r\n headers := map[string]interface{} {\r\n \"user-agent\" : \"cohesity-Go-sdk-6.2.0\",\r\n \"accept\" : \"application/json\",\r\n \"content-type\" : \"application/json; charset=utf-8\",\r\n \"Authorization\" : fmt.Sprintf(\"%s %s\",*me.config.AccessToken().TokenType, *me.config.AccessToken().AccessToken),\r\n }\r\n\r\n //prepare API request\r\n _request := unirest.Put(_queryBuilder, headers, body)\r\n //and invoke the API call request to fetch the response\r\n _response, err := unirest.AsString(_request,me.config.SkipSSL());\r\n if err != nil {\r\n //error in API invocation\r\n return nil, err\r\n }\r\n\r\n //error handling using HTTP status codes\r\n if (_response.Code == 0) {\r\n err = apihelper.NewAPIError(\"Error\", _response.Code, _response.RawBody)\r\n } else if (_response.Code < 200) || (_response.Code > 206) { //[200,206] = HTTP OK\r\n err = apihelper.NewAPIError(\"HTTP Response Not OK\", _response.Code, _response.RawBody)\r\n }\r\n if(err != nil) {\r\n //error detected in status code validation\r\n return nil, err\r\n }\r\n\r\n //returning the response\r\n var retVal *models.ProtectionJob = &models.ProtectionJob{}\r\n err = json.Unmarshal(_response.RawBody, &retVal)\r\n\r\n if err != nil {\r\n //error in parsing\r\n return nil, err\r\n }\r\n return retVal, nil\r\n\r\n}", "func (c *cloudSchedulerRESTClient) RunJob(ctx context.Context, req *schedulerpb.RunJobRequest, opts ...gax.CallOption) (*schedulerpb.Job, error) {\n\tm := protojson.MarshalOptions{AllowPartial: true, UseEnumNumbers: true}\n\tjsonReq, err := m.Marshal(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v1/%v:run\", req.GetName())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"name\", url.QueryEscape(req.GetName()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\topts = append((*c.CallOptions).RunJob[0:len((*c.CallOptions).RunJob):len((*c.CallOptions).RunJob)], opts...)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tresp := &schedulerpb.Job{}\n\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"POST\", baseUrl.String(), bytes.NewReader(jsonReq))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}, opts...)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\treturn resp, nil\n}", "func (s *JobStatus) UpdateFromOutput(status string) error {\n\t// output has format given by slurm or other programs:\n\t// elapsed_time status\n\t// 00:02:36 COMPLETED\n\tvals := strings.Fields(status)\n\n\tif len(vals) != 3 {\n\t\treturn errors.New(\"Job not in database \" + status)\n\t}\n\n\ttimestart := vals[0]\n\ttimeend := vals[1]\n\tstatusstr := vals[2]\n\n\tif timestart != \"Unknown\" {\n\t\tftmstring := \"2006-01-02T15:04:05\"\n\t\ttime, err := time.Parse(ftmstring, timestart)\n\t\tif err != nil {\n\t\t\treturn errors.New(\"Wrong JobStatus output: \" + err.Error())\n\t\t}\n\t\ts.StartTime = utils.TimeToString(time)\n\t}\n\n\tif timeend != \"Unknown\" {\n\t\tftmstring := \"2006-01-02T15:04:05\"\n\t\ttime, err := time.Parse(ftmstring, timeend)\n\t\tif err != nil {\n\t\t\treturn errors.New(\"Wrong JobStatus output: \" + err.Error())\n\t\t}\n\t\ts.EndTime = utils.TimeToString(time)\n\t}\n\n\ts.Status, s.Message = ExplainStatus(statusstr)\n\n\treturn nil\n}", "func (t *JobMessage) ToAPI() models.JobMessage {\n\treturn models.JobMessage{\n\t\tCreatedAt: *TimeToDateTime(&t.CreatedAt),\n\t\tMessage: t.Message,\n\t}\n}", "func (c *JobsClient) UpdateJob(ctx context.Context, req *runpb.UpdateJobRequest, opts ...gax.CallOption) (*UpdateJobOperation, error) {\n\treturn c.internalClient.UpdateJob(ctx, req, opts...)\n}", "func UnmarshalJobLite(m map[string]json.RawMessage, result interface{}) (err error) {\n\tobj := new(JobLite)\n\terr = core.UnmarshalPrimitive(m, \"id\", &obj.ID)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"name\", &obj.Name)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"description\", &obj.Description)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"command_object\", &obj.CommandObject)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"command_object_id\", &obj.CommandObjectID)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"command_name\", &obj.CommandName)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"tags\", &obj.Tags)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"location\", &obj.Location)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"resource_group\", &obj.ResourceGroup)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"targets_ini\", &obj.TargetsIni)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"submitted_at\", &obj.SubmittedAt)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"submitted_by\", &obj.SubmittedBy)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"duration\", &obj.Duration)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"start_at\", &obj.StartAt)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"end_at\", &obj.EndAt)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalModel(m, \"status\", &obj.Status, UnmarshalJobStatus)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalModel(m, \"log_summary\", &obj.LogSummary, UnmarshalJobLogSummary)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = core.UnmarshalPrimitive(m, \"updated_at\", &obj.UpdatedAt)\n\tif err != nil {\n\t\treturn\n\t}\n\treflect.ValueOf(result).Elem().Set(reflect.ValueOf(obj))\n\treturn\n}", "func (sched *lqScheduler) updateCostOfJob(jobId uint) (error) {\n\tlog.Debugf(\"Tracking cost of job %d\", jobId)\n\tjob, err := db.Jobs().Get(jobId)\n\tif err != nil {\n\t\treturn lq.NewErrorf(err, \"Failed updating cost of job %d\", jobId)\n\t}\n\n\tresource, err := db.Resources().Get(job.InstanceID)\n\tif err != nil {\n\t\treturn lq.NewErrorf(err, \"Failed updating cost of job %d\", jobId)\n\t}\n\n\tcost, err := sched.engine.GetResourceCostWithAwsApi(resource, time.Unix(0, job.StartTime), time.Unix(0, job.EndTime))\n\tif err != nil {\n\t\treturn lq.NewErrorf(err, \"Failed updating cost of job %d\", jobId)\n\t}\n\n\tlog.Infof(\"Job %d: Total cost $%f\", jobId, cost)\n\terr = db.Jobs().SetTotalCost(job.ID, job.TotalCost + cost)\n\tif err != nil {\n\t\treturn lq.NewErrorf(err, \"Failed updating cost of job %d\", jobId)\n\t}\n\treturn nil\n}", "func (s *Slurm) JobInfo(ctx context.Context, req *api.JobInfoRequest) (*api.JobInfoResponse, error) {\n\tinfo, err := s.client.SJobInfo(req.JobId)\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"could not get job %d info\", req.JobId)\n\t}\n\n\tpInfo, err := mapSInfoToProtoInfo(info)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"could not convert slurm info into proto info\")\n\t}\n\n\tif len(pInfo) == 0 {\n\t\treturn nil, errors.New(\"job info slice is empty, probably invalid scontrol output\")\n\t}\n\n\treturn &api.JobInfoResponse{Info: pInfo}, nil\n}", "func (js *JobSession) RunJob(jt drmaa2interface.JobTemplate) (drmaa2interface.Job, error) {\n\tid, err := js.tracker[0].AddJob(jt)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn newJob(id, js.name, jt, js.tracker[0]), nil\n}", "func (r *Job) ValidateUpdate(old runtime.Object) error {\n\tjoblog.Info(\"validate update\", \"name\", r.Name)\n\toldJob, ok := old.(*Job)\n\tif !ok {\n\t\treturn fmt.Errorf(\"can't validate job update\")\n\t}\n\tclient := jobmgr.GetClient()\n\tjobs := JobList{}\n\tif err := client.List(context.Background(), &jobs); err != nil {\n\t\treturn err\n\t}\n\tfor _, job := range jobs.Items {\n\t\tif job.Spec.Name == oldJob.Spec.Name {\n\t\t\treturn ErrJobExists\n\t\t}\n\t}\n\treturn nil\n}", "func (d DB) Get(id string) (*job.Job, error) {\n\ttemplate := `select to_jsonb(j.job) from (select * from %[1]s where job ->> 'id' = $1) as j;`\n\tquery := fmt.Sprintf(template, TABLE_NAME)\n\tvar r sql.NullString\n\terr := d.conn.QueryRow(query, id).Scan(&r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresult := &job.Job{}\n\tif r.Valid {\n\t\terr = json.Unmarshal([]byte(r.String), &result)\n\t}\n\treturn result, err\n}", "func (d DB) Save(j *job.Job) error {\n\ttemplate := `update %[1]s SET job = $1 where job ->> 'id' = '` + j.Id + `'`\n\n\t_, errFind := d.Get(j.Id)\n\tif errFind != nil {\n\t\ttemplate = `insert into %[1]s (job) values($1);`\n\t}\n\n\tquery := fmt.Sprintf(template, TABLE_NAME)\n\tr, err := json.Marshal(j)\n\tif err != nil {\n\t\treturn err\n\t}\n\ttransaction, err := d.conn.Begin()\n\tif err != nil {\n\t\treturn err\n\t}\n\tstatement, err := transaction.Prepare(query)\n\tif err != nil {\n\t\ttransaction.Rollback() //nolint:errcheck // adding insult to injury\n\t\treturn err\n\t}\n\tdefer statement.Close()\n\t_, err = statement.Exec(string(r))\n\tif err != nil {\n\t\ttransaction.Rollback() //nolint:errcheck // adding insult to injury\n\t\treturn err\n\t}\n\treturn transaction.Commit()\n}", "func ReconcileJob(reqLogger logr.Logger, c client.Client, job *batchv1.Job, wait bool) error {\n\tif err := SetCreationSpecAnnotation(&job.ObjectMeta, job); err != nil {\n\t\treturn err\n\t}\n\n\tfound := &batchv1.Job{}\n\tif err := c.Get(context.TODO(), types.NamespacedName{Name: job.Name, Namespace: job.Namespace}, found); err != nil {\n\t\t// Return API error\n\t\tif client.IgnoreNotFound(err) != nil {\n\t\t\treturn err\n\t\t}\n\t\t// Create the job\n\t\tif err := c.Create(context.TODO(), job); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Check the found job spec\n\tif !CreationSpecsEqual(job.ObjectMeta, found.ObjectMeta) {\n\t\t// We need to delete the job and requeue\n\t\tif err := c.Delete(context.TODO(), found); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif wait {\n\t\t\treturn errors.NewRequeueError(\"Deleting stale job and requeuing\", 5)\n\t\t}\n\t}\n\n\tif wait {\n\t\t// Make sure the job completed\n\t\tif found.Status.Succeeded != 1 {\n\t\t\treturn errors.NewRequeueError(\"Waiting for job to complete\", 3)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (r *CronCrawlerJobReconciler) Reconcile(req ctrl.Request) (ctrl.Result, error) {\n\tctx := context.Background()\n\tlog := r.Log.WithValues(\"croncrawlerjob\", req.NamespacedName)\n\tinstance := &cronjobv1.CronCrawlerJob{}\n\n\t// 1 判断自定义资源是否存在\n\terr := r.Get(ctx, req.NamespacedName, instance)\n\tif err != nil {\n\t\tif errors.IsNotFound(err) {\n\t\t\t// Request object not found, could have been deleted after ctrl request.\n\t\t\t// Owned objects are automatically garbage collected. For additional cleanup logic use finalizers.\n\t\t\t// Return and don't requeue\n\t\t\treturn ctrl.Result{}, nil\n\t\t}\n\t\t// Error reading the object - requeue the request.\n\t\tlog.Error(err, \"Failed to get CronCrawlerJob\")\n\t\treturn ctrl.Result{}, err\n\t}\n\n\tif instance.DeletionTimestamp != nil {\n\t\treturn ctrl.Result{}, err\n\t}\n\n\tjobType := instance.Spec.JobType\n\tschedule := instance.Spec.Schedule\n\n\t// 采集类型为定时任务时,必须指定采集频率\n\tif (jobType == \"cronjob\") && (schedule == \"\") {\n\t\tlog.Error(err, \"`schedule` must be confirmed in a cronjob jobtype!\")\n\t\treturn ctrl.Result{}, err\n\t}\n\n\t// 2 自定义资源已经存在,判断新旧任务类型是否一致,\n\t// 2.1 如果不一致一律删除,注意是普通任务也直接删除,\n\t// 2.2 如果一致,也就是定时任务了,直接更新\n\n\tnewSpec := CrawlerUpdatedData{\n\t\tJobType: instance.Spec.JobType,\n\t\tSchedule: instance.Spec.Schedule,\n\t\tEnv: instance.Spec.Env,\n\t\tDuration: instance.Spec.Duration,\n\t\tPublisherTTL: instance.Spec.Publisher.PublisherTTL,\n\t\tConsumerTTL: instance.Spec.Consumer.ConsumerTTL,\n\t\tParallelism: instance.Spec.Consumer.Parallelism,\n\t}\n\n\tduration, _ := time.ParseDuration(instance.Spec.Duration)\n\tns := instance.Namespace\n\n\tlog.Info(\"instance ns: \" + ns + \", req ns: \" + req.NamespacedName.Namespace)\n\tname := instance.Name\n\tshortDes := ns + \"/\" + name\n\tpublisherName := NameForCronCrawlerJob(name, \"publisher\", jobType)\n\tconsumerName := NameForCronCrawlerJob(name, \"consumer\", jobType)\n\n\t// 判断是不是旧任务入队\n\tif (instance.Annotations != nil) && (instance.Annotations[\"status\"] == \"old\") {\n\t\t// 反序列化\n\t\toldSpec := CrawlerUpdatedData{}\n\t\terr = json.Unmarshal([]byte(instance.Annotations[\"spec\"]), &oldSpec)\n\t\tif err != nil {\n\t\t\treturn ctrl.Result{}, nil\n\t\t}\n\t\t// 判断是否有更新\n\t\tif !reflect.DeepEqual(oldSpec, newSpec) {\n\t\t\t// job type\n\t\t\toldJobType := oldSpec.JobType\n\t\t\tif (jobType != oldJobType) || (oldJobType == \"job\") {\n\t\t\t\t// deletes the older resource\n\t\t\t\tif err := r.DeleteOldJobsByLabels(ns, name, oldJobType); err != nil {\n\t\t\t\t\tlog.Error(err, \"Older \"+oldJobType+\" for \"+shortDes+\" deletes failed\")\n\t\t\t\t\treturn ctrl.Result{}, err\n\t\t\t\t}\n\t\t\t\tlog.Info(\"Older \" + oldJobType + \" for \" + shortDes + \" deletes success\")\n\t\t\t\t// resets the current resource\n\t\t\t\tinstance.Annotations = nil\n\t\t\t\tif err := r.Update(ctx, instance); err != nil {\n\t\t\t\t\treturn ctrl.Result{}, err\n\t\t\t\t}\n\t\t\t\tlog.Info(\"Current \" + jobType + \"'s annotation sets to nil, would be requeued later\")\n\t\t\t\treturn ctrl.Result{Requeue: true, RequeueAfter: duration}, nil\n\t\t\t} else if jobType == \"cronjob\" {\n\t\t\t\t// 更新cronjob,并返回\n\t\t\t\tif err := r.UpdateOldJobs(instance, ns, jobType, publisherName, consumerName); err != nil {\n\t\t\t\t\tlog.Error(err, \"Older \"+jobType+\" for \"+shortDes+\" updates failed\")\n\t\t\t\t\treturn ctrl.Result{}, err\n\t\t\t\t}\n\t\t\t\tlog.Info(\"Older \" + oldJobType + \" for \" + shortDes + \" updates success\")\n\t\t\t\treturn ctrl.Result{}, err\n\n\t\t\t}\n\n\t\t}\n\t}\n\n\t// 3、使用当前instance创建新资源并关联数据返回,只有当获取不到资源时候,才创建资源\n\n\tvar jobPublisher runtime.Object\n\tvar jobConsumers runtime.Object\n\n\tif jobType == \"job\" {\n\t\tjobPublisher = r.NewPublisherJob(instance, publisherName)\n\t\tjobConsumers = r.NewConsumersJob(instance, consumerName)\n\t} else if jobType == \"cronjob\" {\n\t\tjobPublisher = r.NewPublisherCronJob(instance, publisherName)\n\t\tjobConsumers = r.NewConsumersCronJob(instance, consumerName)\n\t} else {\n\t\tlog.Error(err, \"JobType sets wrong\")\n\t\treturn ctrl.Result{}, nil\n\t}\n\n\tif err := r.Get(ctx, types.NamespacedName{Name: publisherName, Namespace: ns}, jobPublisher); err != nil {\n\t\tlog.Info(\"Starting to create \" + jobType + \" for publisher\")\n\t\tif err := r.Create(ctx, jobPublisher); err != nil {\n\t\t\tlog.Error(err, jobType+\" for publisher creates failed\")\n\t\t\treturn ctrl.Result{}, err\n\t\t}\n\t\treturn ctrl.Result{Requeue: true}, nil\n\n\t}\n\tif err := r.Get(ctx, types.NamespacedName{Name: consumerName, Namespace: ns}, jobConsumers); err != nil {\n\t\tlog.Info(\"Starting to create \" + jobType + \" for consumers\")\n\t\tif err := r.Create(ctx, jobConsumers); err != nil {\n\t\t\tlog.Error(err, jobType+\" for Consumers creates failed\")\n\t\t\treturn ctrl.Result{}, err\n\t\t}\n\t\treturn ctrl.Result{Requeue: true}, nil\n\t}\n\n\t// 关联 Annotations\n\tdata, _ := json.Marshal(&newSpec)\n\tif instance.Annotations != nil {\n\t\tinstance.Annotations[\"spec\"] = string(data)\n\t\tinstance.Annotations[\"status\"] = \"old\"\n\t} else {\n\t\tinstance.Annotations = map[string]string{\"spec\": string(data), \"status\": \"new\"}\n\t}\n\n\tif err := r.Update(ctx, instance); err != nil {\n\t\treturn ctrl.Result{}, nil\n\t}\n\n\t// 更新显示状态\n\tpodList := &corev1.PodList{}\n\tlabels := map[string]string{\"crawler-table\": name, \"jobType\": jobType}\n\tlistOption := []client.ListOption{\n\t\tclient.InNamespace(ns),\n\t\tclient.MatchingLabels(labels),\n\t}\n\tif err := r.List(ctx, podList, listOption...); err != nil {\n\t\treturn ctrl.Result{}, nil\n\t}\n\n\tcronCrawlerJobStatus := getPodStatus(podList.Items)\n\tif !reflect.DeepEqual(cronCrawlerJobStatus, instance.Status) {\n\t\tlog.Info(\"Starting to update status\")\n\t\tinstance.Status = *cronCrawlerJobStatus\n\t\terr := r.Status().Update(ctx, instance)\n\t\tif err != nil {\n\t\t\tlog.Error(err, \"Failed to update CronCrawlerJob status\")\n\t\t\treturn ctrl.Result{}, err\n\t\t}\n\t}\n\n\treturn ctrl.Result{}, nil\n}", "func CreateJob(board schema.Board, boardID string, categoryID string, ownerID string) {\n\n\tobjID, err := primitive.ObjectIDFromHex(boardID)\n\tcategoryID_int, err := strconv.Atoi(categoryID)\n\tif err != nil {\n\t\t// handle error\n\t\tlog.Fatal(\"CreateJob() str conv ERROR:\", err)\n\t}\n\n\t//get job details from board document inputted through put request\n\tjd := schema.JobDetails{\n\t\tCompany: board.Categories[0].Jobs[0].JobDetails.Company,\n\t\tTitle: board.Categories[0].Jobs[0].JobDetails.Title,\n\t\tLocation: board.Categories[0].Jobs[0].JobDetails.Location,\n\t\tCategory: board.Categories[0].Jobs[0].JobDetails.Category,\n\t\tPostDate: board.Categories[0].Jobs[0].JobDetails.PostDate,\n\t\tDescription: board.Categories[0].Jobs[0].JobDetails.Description,\n\t\tExperience: board.Categories[0].Jobs[0].JobDetails.Experience,\n\t\tURL: board.Categories[0].Jobs[0].JobDetails.URL,\n\t\tDateAdded: board.Categories[0].Jobs[0].JobDetails.DateAdded,\n\t\tSalary: board.Categories[0].Jobs[0].JobDetails.Salary,\n\t\tTasks: board.Categories[0].Jobs[0].JobDetails.Tasks}\n\n\t//stick job details into job var\n\tj := schema.Job{JobDetails: jd, ID: board.Categories[0].Jobs[0].ID}\n\n\tfilter := bson.M{\"_id\": bson.M{\"$eq\": objID}, \"categories.id\": bson.M{\"$eq\": categoryID_int}}\n\tupdate := bson.M{\"$set\": bson.M{\"categories.$.jobs\": j}}\n\n\tupdateResult, err := db.Collection(COLLNAME).UpdateOne(\n\t\tcontext.Background(),\n\t\tfilter,\n\t\tupdate,\n\t)\n\tif err != nil {\n\t\tlog.Fatal(\"CreateJob() ERROR:\", err)\n\t}\n\n\tfmt.Println(\"create job func input: \", j)\n\n\tfmt.Printf(\"Matched %v documents and updated %v documents.\\n\", updateResult.MatchedCount, updateResult.ModifiedCount)\n\tfmt.Println(\"Full Result: \", updateResult)\n\n}", "func (p *AuroraSchedulerManagerClient) StartJobUpdate(ctx context.Context, request *JobUpdateRequest, message string) (r *Response, err error) {\n var _args198 AuroraSchedulerManagerStartJobUpdateArgs\n _args198.Request = request\n _args198.Message = message\n var _result199 AuroraSchedulerManagerStartJobUpdateResult\n var meta thrift.ResponseMeta\n meta, err = p.Client_().Call(ctx, \"startJobUpdate\", &_args198, &_result199)\n p.SetLastResponseMeta_(meta)\n if err != nil {\n return\n }\n return _result199.GetSuccess(), nil\n}", "func GetJobFromRadixJob(job *v1.RadixJob, jobDeployments []*deploymentModels.DeploymentSummary) *Job {\n\tsteps := GetJobStepsFromRadixJob(job)\n\n\tcreated := radixutils.FormatTime(&job.CreationTimestamp)\n\tif job.Status.Created != nil {\n\t\t// Use this instead, because in a migration this may be more correct\n\t\t// as migrated jobs will have the same creation timestamp in the new cluster\n\t\tcreated = radixutils.FormatTime(job.Status.Created)\n\t}\n\n\tvar jobComponents []*deploymentModels.ComponentSummary\n\tif len(jobDeployments) > 0 {\n\t\tjobComponents = jobDeployments[0].Components\n\t}\n\n\tjobModel := Job{\n\t\tName: job.GetName(),\n\t\tBranch: job.Spec.Build.Branch,\n\t\tCommitID: job.Spec.Build.CommitID,\n\t\tCreated: created,\n\t\tStarted: radixutils.FormatTime(job.Status.Started),\n\t\tEnded: radixutils.FormatTime(job.Status.Ended),\n\t\tStatus: GetStatusFromRadixJobStatus(job.Status, job.Spec.Stop),\n\t\tPipeline: string(job.Spec.PipeLineType),\n\t\tSteps: steps,\n\t\tDeployments: jobDeployments,\n\t\tComponents: jobComponents,\n\t\tTriggeredBy: job.Spec.TriggeredBy,\n\t}\n\tif job.Spec.PipeLineType == v1.Promote {\n\t\tjobModel.PromotedFromEnvironment = job.Spec.Promote.FromEnvironment\n\t\tjobModel.PromotedToEnvironment = job.Spec.Promote.ToEnvironment\n\t\tjobModel.PromotedDeploymentName = job.Spec.Promote.DeploymentName\n\t}\n\n\treturn &jobModel\n}", "func (c *JobpositionClient) Update() *JobpositionUpdate {\n\tmutation := newJobpositionMutation(c.config, OpUpdate)\n\treturn &JobpositionUpdate{config: c.config, hooks: c.Hooks(), mutation: mutation}\n}", "func (m *DataHistoryManager) UpsertJob(job *DataHistoryJob, insertOnly bool) error {\n\tif m == nil {\n\t\treturn ErrNilSubsystem\n\t}\n\tif !m.IsRunning() {\n\t\treturn ErrSubSystemNotStarted\n\t}\n\tif job == nil {\n\t\treturn errNilJob\n\t}\n\tif job.Nickname == \"\" {\n\t\treturn fmt.Errorf(\"upsert job %w\", errNicknameUnset)\n\t}\n\tj, err := m.GetByNickname(job.Nickname, false)\n\tif err != nil && !errors.Is(err, errJobNotFound) {\n\t\treturn err\n\t}\n\tif insertOnly && j != nil ||\n\t\t(j != nil && j.Status != dataHistoryStatusActive) {\n\t\treturn fmt.Errorf(\"upsert job %w nickname: %s - status: %s \", errNicknameInUse, j.Nickname, j.Status)\n\t}\n\tif job.PrerequisiteJobNickname != \"\" {\n\t\tvar p *DataHistoryJob\n\t\tp, err = m.GetByNickname(job.PrerequisiteJobNickname, false)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"upsert job %s could not find prerequisite job nickname %v %w\", job.Nickname, job.PrerequisiteJobNickname, err)\n\t\t}\n\t\tif p.Status != dataHistoryStatusActive && p.Status != dataHistoryStatusPaused {\n\t\t\treturn fmt.Errorf(\"upsert job %s prerequisite job nickname %v already completed %w\", job.Nickname, p.Nickname, errJobInvalid)\n\t\t}\n\t}\n\n\terr = m.validateJob(job)\n\tif err != nil {\n\t\treturn err\n\t}\n\texistingJob, err := m.GetByNickname(job.Nickname, false)\n\tif err != nil && err != errJobNotFound {\n\t\treturn err\n\t}\n\tif existingJob != nil {\n\t\tjob.ID = existingJob.ID\n\t}\n\tif job.ID == uuid.Nil {\n\t\tjob.ID, err = uuid.NewV4()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tinterval := job.Interval\n\tif job.DataType == dataHistoryConvertCandlesDataType {\n\t\tinterval = job.ConversionInterval\n\t}\n\tjob.rangeHolder, err = kline.CalculateCandleDateRanges(job.StartDate, job.EndDate, interval, uint32(job.RequestSizeLimit))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdbJob := m.convertJobToDBModel(job)\n\terr = m.jobDB.Upsert(dbJob)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif job.PrerequisiteJobNickname == \"\" {\n\t\treturn nil\n\t}\n\tjob.Status = dataHistoryStatusPaused\n\treturn m.jobDB.SetRelationshipByNickname(job.PrerequisiteJobNickname, job.Nickname, int64(dataHistoryStatusPaused))\n}", "func (c *scheduledJobs) UpdateStatus(job *batch.ScheduledJob) (result *batch.ScheduledJob, err error) {\n\tresult = &batch.ScheduledJob{}\n\terr = c.r.Put().Namespace(c.ns).Resource(\"scheduledjobs\").Name(job.Name).SubResource(\"status\").Body(job).Do().Into(result)\n\treturn\n}", "func ParseJob(input string) (job *Job, err error) {\n\t// Note: rather than check for errors throughout, all parsing functions\n\t// called by this one simply panic if they encounter a problem; the panics\n\t// are caught here and translated into normal errors.\n\tfail := func(cause interface{}) error {\n\t\treturn fmt.Errorf(\"parse failed: %s\", cause)\n\t}\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\terr = fail(r)\n\t\t}\n\t}()\n\tfields := ws.Split(strings.TrimSpace(input), 6)\n\tif len(fields) != 6 {\n\t\treturn nil, fail(fmt.Errorf(\"got %d fields, want 6\", len(fields)))\n\t}\n\treturn &Job{\n\t\tMinute: expand(clamp(parseField(fields[0], nil), 0, 59)),\n\t\tHour: expand(clamp(parseField(fields[1], nil), 0, 23)),\n\t\tMonthDay: expand(clamp(parseField(fields[2], nil), 1, 31)),\n\t\tMonth: expand(clamp(parseField(fields[3], trMonth), 1, 12)),\n\t\tWeekday: expand(clamp(parseField(fields[4], trWeekday), 0, 6)),\n\t\tCommand: fields[5],\n\t}, nil\n}", "func (r *ProjectsLocationsJobsService) Patch(name string, job *Job) *ProjectsLocationsJobsPatchCall {\n\tc := &ProjectsLocationsJobsPatchCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.name = name\n\tc.job = job\n\treturn c\n}", "func (op *UpdateJobOperation) Poll(ctx context.Context, opts ...gax.CallOption) (*runpb.Job, error) {\n\topts = append([]gax.CallOption{gax.WithPath(op.pollPath)}, opts...)\n\tvar resp runpb.Job\n\tif err := op.lro.Poll(ctx, &resp, opts...); err != nil {\n\t\treturn nil, err\n\t}\n\tif !op.Done() {\n\t\treturn nil, nil\n\t}\n\treturn &resp, nil\n}", "func (c *Client) ControlJob(jobID string, action string) {\n\turl := fmt.Sprintf(\"http://%s/api/v1/jobs/%s\", c.options.ServerAddr, jobID)\n\tresp, err := http.Post(url,\n\t\t\"application/json\",\n\t\tstrings.NewReader(fmt.Sprintf(\"action=%s\", action)),\n\t)\n\tif err != nil {\n\t\tlogrus.WithError(err)\n\t\treturn\n\t}\n\tdefer resp.Body.Close()\n}", "func (j *JobRun) PostUpdate(s gorp.SqlExecutor) error {\n\tjobJSON, err := json.Marshal(j.Job)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tparamsJSON, errP := json.Marshal(j.Parameters)\n\tif errP != nil {\n\t\treturn errP\n\t}\n\n\tquery := \"update workflow_node_run_job set job = $2, variables = $3 where id = $1\"\n\tif n, err := s.Exec(query, j.ID, jobJSON, paramsJSON); err != nil {\n\t\treturn err\n\t} else if n, _ := n.RowsAffected(); n == 0 {\n\t\treturn fmt.Errorf(\"Unable to update workflow_node_run_job id = %d\", j.ID)\n\t}\n\n\treturn nil\n}", "func (in *Job) DeepCopy() *Job {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(Job)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (j *JobWorker) GetJobStatus(id string) (Job, error) {\n\tif job, ok := j.load(id); ok {\n\t\t// make a shallow copy of the job object\n\t\tjobCopy := *job\n\t\tjobCopy.Output = jobCopy.outputBuffer.String()\n\t\treturn jobCopy, nil\n\t} else {\n\t\treturn Job{}, fmt.Errorf(\"job %v: not found\", id)\n\t}\n}", "func (b *RepairCronJob) Reconcile(driver opsdk.Client) (sdk.Object, error) {\n\tvar err error\n\n\tb.configureDesired()\n\n\texisting := &batchv1beta1.CronJob{\n\t\tTypeMeta: GetCronJobTypeMeta(),\n\t\tObjectMeta: b.desired.ObjectMeta,\n\t}\n\terr = driver.Get(existing)\n\tif err != nil {\n\t\treturn nil, errors.New(\"could not get existing\")\n\t}\n\n\tif existing.ResourceVersion != \"\" {\n\t\t// here we have one that is existing and one that is expected\n\t\t// we put our code here to reconcile the two and return\n\t\t// the reconciled object\n\t\tb.desired.ResourceVersion = existing.ResourceVersion\n\t\terr = driver.Update(b.desired)\n\t\treturn b.desired, err\n\t}\n\n\terr = driver.Create(b.desired)\n\treturn b.desired, err\n}", "func (ji *JobInfo) SetJob(job *batch.Job) {\n\tji.Name = job.Name\n\tji.Namespace = job.Namespace\n\tji.Job = job\n}", "func UpdateNodeJobRun(db gorp.SqlExecutor, store cache.Store, p *sdk.Project, j *sdk.WorkflowNodeJobRun) error {\n\tdbj := JobRun(*j)\n\tif _, err := db.Update(&dbj); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *cloudSchedulerRESTClient) CreateJob(ctx context.Context, req *schedulerpb.CreateJobRequest, opts ...gax.CallOption) (*schedulerpb.Job, error) {\n\tm := protojson.MarshalOptions{AllowPartial: true, UseEnumNumbers: true}\n\tbody := req.GetJob()\n\tjsonReq, err := m.Marshal(body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v1/%v/jobs\", req.GetParent())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"parent\", url.QueryEscape(req.GetParent()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\topts = append((*c.CallOptions).CreateJob[0:len((*c.CallOptions).CreateJob):len((*c.CallOptions).CreateJob)], opts...)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tresp := &schedulerpb.Job{}\n\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"POST\", baseUrl.String(), bytes.NewReader(jsonReq))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}, opts...)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\treturn resp, nil\n}", "func GetJob(ctx context.Context, p string) (j *Job, err error) {\n\tt := utils.FromTaskContext(ctx)\n\n\tj = &Job{}\n\n\tcontent, err := contexts.DB.Get(constants.FormatJobKey(t, p), nil)\n\tif err == leveldb.ErrNotFound {\n\t\treturn nil, nil\n\t}\n\tif err != nil {\n\t\treturn\n\t}\n\n\terr = msgpack.Unmarshal(content, j)\n\tif err != nil {\n\t\tlogrus.Panicf(\"Msgpack unmarshal failed for %v.\", err)\n\t}\n\treturn\n}", "func ApplyJob(client batchclientv1.JobsGetter, required *batchv1.Job) (*batchv1.Job, bool, error) {\n\texisting, err := client.Jobs(required.Namespace).Get(required.Name, metav1.GetOptions{})\n\tif apierrors.IsNotFound(err) {\n\t\tactual, err := client.Jobs(required.Namespace).Create(required)\n\t\treturn actual, true, err\n\t}\n\tif err != nil {\n\t\treturn nil, false, err\n\t}\n\t// if we only create this resource, we have no need to continue further\n\tif IsCreateOnly(required) {\n\t\treturn nil, false, nil\n\t}\n\n\tmodified := pointer.BoolPtr(false)\n\tresourcemerge.EnsureJob(modified, existing, *required)\n\tif !*modified {\n\t\treturn existing, false, nil\n\t}\n\n\tactual, err := client.Jobs(required.Namespace).Update(existing)\n\treturn actual, true, err\n}", "func (q *Queue) Update(id int, data []byte) error {\n\tif _, ok := q.data[id]; !ok {\n\t\treturn fmt.Errorf(\"could not find any job with id : %d\", id)\n\t}\n\tq.lock.Lock()\n\tq.data[id] = data\n\tq.lock.Unlock()\n\treturn nil\n}", "func (c *restClient) CancelPatchJob(ctx context.Context, req *osconfigpb.CancelPatchJobRequest, opts ...gax.CallOption) (*osconfigpb.PatchJob, error) {\n\tm := protojson.MarshalOptions{AllowPartial: true, UseEnumNumbers: true}\n\tjsonReq, err := m.Marshal(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v1beta/%v:cancel\", req.GetName())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"name\", url.QueryEscape(req.GetName()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\topts = append((*c.CallOptions).CancelPatchJob[0:len((*c.CallOptions).CancelPatchJob):len((*c.CallOptions).CancelPatchJob)], opts...)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tresp := &osconfigpb.PatchJob{}\n\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"POST\", baseUrl.String(), bytes.NewReader(jsonReq))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}, opts...)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\treturn resp, nil\n}", "func (d *Daemon) JobStatus(ctx context.Context, jobID job.ID) (job.Status, error) {\n\t// Is the job queued, running, or recently finished?\n\tstatus, ok := d.JobStatusCache.Status(jobID)\n\tif ok {\n\t\treturn status, nil\n\t}\n\n\t// Look through the commits for a note referencing this job. This\n\t// means that even if fluxd restarts, we will at least remember\n\t// jobs which have pushed a commit.\n\tnotes, err := d.Repo.NoteRevList(ctx, d.GitConfig.NotesRef)\n\tif err != nil {\n\t\treturn status, errors.Wrap(err, \"enumerating commit notes\")\n\t}\n\tcommits, err := d.Repo.CommitsBefore(ctx, \"HEAD\", false, d.GitConfig.Paths...)\n\tif err != nil {\n\t\treturn status, errors.Wrap(err, \"checking revisions for status\")\n\t}\n\n\tfor _, commit := range commits {\n\t\tif _, ok := notes[commit.Revision]; ok {\n\t\t\tvar n note\n\t\t\tok, err := d.Repo.GetNote(ctx, commit.Revision, d.GitConfig.NotesRef, &n)\n\t\t\tif ok && err == nil && n.JobID == jobID {\n\t\t\t\tstatus = job.Status{\n\t\t\t\t\tStatusString: job.StatusSucceeded,\n\t\t\t\t\tResult: job.Result{\n\t\t\t\t\t\tRevision: commit.Revision,\n\t\t\t\t\t\tSpec: &n.Spec,\n\t\t\t\t\t\tResult: n.Result,\n\t\t\t\t\t},\n\t\t\t\t}\n\t\t\t\treturn status, nil\n\t\t\t}\n\t\t}\n\t}\n\treturn status, unknownJobError(jobID)\n}", "func (cl *Client) GetJob() (*MapReduceJob, error) {\n\tresp, err := cl.client.Get(cl.baseurl)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif resp.StatusCode != http.StatusOK {\n\t\treturn nil, fmt.Errorf(\"Got status: %s\", resp.Status)\n\t}\n\tdefer resp.Body.Close()\n\tdecoder := json.NewDecoder(resp.Body)\n\tjb := &MapReduceJob{}\n\terr = decoder.Decode(jb)\n\treturn jb, err\n}", "func (t *DRMAATracker) JobInfo(jobID string) (drmaa2interface.JobInfo, error) {\n\t// we get the job info when the job is finished - we can also\n\t// use the DRM system specific calls (like on GE)\n\tstate, _, err := t.JobState(jobID)\n\tif err != nil {\n\t\treturn drmaa2interface.JobInfo{}, err\n\t}\n\tif state == drmaa2interface.Failed || state == drmaa2interface.Done {\n\t\t// job is in end state\n\t\tjinfo, err := t.session.Wait(jobID, 60)\n\t\tif err != nil {\n\t\t\treturn drmaa2interface.JobInfo{}, err\n\t\t}\n\t\treturn ConvertDRMAAJobInfoToDRMAA2JobInfo(&jinfo), nil\n\t}\n\treturn drmaa2interface.JobInfo{}, nil\n}", "func Update(ctx *routing.Context) error {\n\tlogger := logger.GetLogInstance(\"\", \"\")\n\tdb := ctx.Get(\"db\").(*gorm.DB)\n\tid := ctx.Param(\"id\")\n\n\tcmp := &dbmodels.Campaign{}\n\n\tif r := jsoniter.Unmarshal(ctx.Request.Body(), &cmp); r != nil {\n\t\tlogger.Error(r)\n\t\treturn r\n\t}\n\n\tif err := db.Where(\"id = ?\", id).First(&cmp).Error; err != nil {\n\t\tlogger.Error(err)\n\t\tctx.Response.SetStatusCode(404)\n\t\tr := models.NewResponse(false, nil, \"cmp not found\")\n\t\treturn ctx.WriteData(r.MustMarshal())\n\t}\n\n\tif err := jsoniter.Unmarshal(ctx.Request.Body(), &cmp); err != nil {\n\t\tctx.Response.SetStatusCode(400)\n\t\tlogger.Error(err)\n\t\tr := models.NewResponse(false, nil, \"unexpected error\")\n\n\t\treturn ctx.WriteData(r.MustMarshal())\n\n\t}\n\n\tdb.Save(&cmp)\n\tr := models.NewResponse(true, cmp, \"OK\")\n\treturn ctx.WriteData(r.MustMarshal())\n}", "func (c *jxTasks) Update(jxTask *v1alpha1.JxTask) (result *v1alpha1.JxTask, err error) {\n\tresult = &v1alpha1.JxTask{}\n\terr = c.client.Put().\n\t\tNamespace(c.ns).\n\t\tResource(\"jxtasks\").\n\t\tName(jxTask.Name).\n\t\tBody(jxTask).\n\t\tDo().\n\t\tInto(result)\n\treturn\n}", "func (c *scheduledJobs) Get(name string) (result *batch.ScheduledJob, err error) {\n\tresult = &batch.ScheduledJob{}\n\terr = c.r.Get().Namespace(c.ns).Resource(\"scheduledjobs\").Name(name).Do().Into(result)\n\treturn\n}", "func (c *Client) PostJob(j models.JobReq) (models.JobResp, error) {\n\tvar out models.JobResp\n\terr := c.doHTTPReq(http.MethodPost,\n\t\tfmt.Sprintf(uriPostJob, j.TaskName), j, nil, &out)\n\treturn out, err\n}", "func (p *AuroraSchedulerManagerClient) ScheduleCronJob(ctx context.Context, description *JobConfiguration) (r *Response, err error) {\n var _args184 AuroraSchedulerManagerScheduleCronJobArgs\n _args184.Description = description\n var _result185 AuroraSchedulerManagerScheduleCronJobResult\n var meta thrift.ResponseMeta\n meta, err = p.Client_().Call(ctx, \"scheduleCronJob\", &_args184, &_result185)\n p.SetLastResponseMeta_(meta)\n if err != nil {\n return\n }\n return _result185.GetSuccess(), nil\n}", "func (j *Job) Save(ctx context.Context) (err error) {\n\tt := utils.FromTaskContext(ctx)\n\n\tcontent, err := msgpack.Marshal(j)\n\tif err != nil {\n\t\treturn\n\t}\n\n\terr = contexts.DB.Put(constants.FormatJobKey(t, j.Path), content, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func (c *FakeTZCronJobs) Watch(opts v1.ListOptions) (watch.Interface, error) {\n\treturn c.Fake.\n\t\tInvokesWatch(testing.NewWatchAction(tzcronjobsResource, c.ns, opts))\n\n}", "func ToJobBase(r repoconfig.JobBase) job.Base {\n\treturn job.Base{\n\t\tName: r.Name,\n\t\tLabels: r.Labels,\n\t\tAnnotations: r.Annotations,\n\t\tMaxConcurrency: r.MaxConcurrency,\n\t\tAgent: r.Agent,\n\t\tCluster: r.Cluster,\n\t\tNamespace: r.Namespace,\n\t\tErrorOnEviction: r.ErrorOnEviction,\n\t\tSourcePath: r.SourcePath,\n\t\tSpec: r.Spec,\n\t\tPipelineRunSpec: r.PipelineRunSpec,\n\t\tUtilityConfig: ToUtilityConfig(r.UtilityConfig),\n\t}\n}" ]
[ "0.63324195", "0.6151295", "0.5906331", "0.5875017", "0.58593154", "0.5774403", "0.57014155", "0.56926346", "0.56737024", "0.56274194", "0.56148297", "0.55691075", "0.55199784", "0.5518581", "0.5503346", "0.54800415", "0.54626197", "0.54567194", "0.5443423", "0.5391119", "0.5387904", "0.53553975", "0.5325324", "0.5317383", "0.52815783", "0.52509904", "0.52488506", "0.5233862", "0.5178498", "0.51769865", "0.5164666", "0.5159494", "0.5153006", "0.51474655", "0.5134767", "0.5125806", "0.50760674", "0.5066904", "0.50606525", "0.50467116", "0.50291497", "0.5015038", "0.4976536", "0.49350363", "0.49273345", "0.49261072", "0.49230397", "0.49165708", "0.4909044", "0.48892802", "0.4879768", "0.4862928", "0.4858789", "0.48513114", "0.48410314", "0.4823932", "0.48155728", "0.48081955", "0.47574648", "0.4746354", "0.47377446", "0.47370726", "0.47314903", "0.47062367", "0.47005504", "0.4693126", "0.46864974", "0.4661405", "0.46519297", "0.46386474", "0.46349353", "0.46332368", "0.46276975", "0.4623595", "0.4605535", "0.4599468", "0.4597147", "0.4595529", "0.45921358", "0.45841867", "0.45766383", "0.4575201", "0.45725963", "0.45678914", "0.45595738", "0.4558914", "0.4552088", "0.45506716", "0.4544059", "0.45402503", "0.45372748", "0.45339382", "0.4532986", "0.45310354", "0.4526138", "0.452545", "0.45235825", "0.45225888", "0.45177943", "0.450787" ]
0.6831923
0
UpdateStatus was generated because the type contains a Status member. Add a +genclient:noStatus comment above the type to avoid generating UpdateStatus().
func (c *FakeTZCronJobs) UpdateStatus(tZCronJob *v1alpha1.TZCronJob) (*v1alpha1.TZCronJob, error) { obj, err := c.Fake. Invokes(testing.NewUpdateSubresourceAction(tzcronjobsResource, "status", c.ns, tZCronJob), &v1alpha1.TZCronJob{}) if obj == nil { return nil, err } return obj.(*v1alpha1.TZCronJob), err }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (i *DeleteOrUpdateInvTask) StatusUpdate(_ *taskrunner.TaskContext, _ object.ObjMetadata) {}", "func (r *Reconciler) UpdateStatus() error {\n\tlog := r.Logger.WithField(\"func\", \"UpdateStatus\")\n\tlog.Infof(\"Updating noobaa status\")\n\tr.NooBaa.Status.ObservedGeneration = r.NooBaa.Generation\n\treturn r.Client.Status().Update(r.Ctx, r.NooBaa)\n}", "func (manager *Manager) updateStatus(jobStatus ingestic.JobStatus) {\n\tmanager.estimatedEndTimeInSec = jobStatus.EstimatedEndTimeInSec\n\tmanager.percentageComplete = jobStatus.PercentageComplete\n\n\tevent := &automate_event.EventMsg{\n\t\tEventID: createEventUUID(),\n\t\tType: &automate_event.EventType{Name: automate_event_type.ProjectRulesUpdateStatus},\n\t\tPublished: ptypes.TimestampNow(),\n\t\tProducer: &automate_event.Producer{\n\t\t\tID: event_ids.ComplianceInspecReportProducerID,\n\t\t},\n\t\tData: &_struct.Struct{\n\t\t\tFields: map[string]*_struct.Value{\n\t\t\t\t\"Completed\": {\n\t\t\t\t\tKind: &_struct.Value_BoolValue{\n\t\t\t\t\t\tBoolValue: jobStatus.Completed,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\t\"PercentageComplete\": {\n\t\t\t\t\tKind: &_struct.Value_NumberValue{\n\t\t\t\t\t\tNumberValue: float64(jobStatus.PercentageComplete),\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\t\"EstimatedTimeCompeleteInSec\": {\n\t\t\t\t\tKind: &_struct.Value_NumberValue{\n\t\t\t\t\t\tNumberValue: float64(jobStatus.EstimatedEndTimeInSec),\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tproject_update_tags.ProjectUpdateIDTag: {\n\t\t\t\t\tKind: &_struct.Value_StringValue{\n\t\t\t\t\t\tStringValue: manager.projectUpdateID,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\n\tpubReq := &automate_event.PublishRequest{Msg: event}\n\t_, err := manager.eventServiceClient.Publish(context.Background(), pubReq)\n\tif err != nil {\n\t\tlogrus.Warnf(\"Publishing Failed event %v\", err)\n\t}\n}", "func Status(status int) {\n\tif r, ok := responseDefinition(); ok {\n\t\tr.Status = status\n\t}\n}", "func (t *Task) updateStatus() {\n\tb, err := json.Marshal(&map[string]interface{}{\n\t\t\"type\": \"update\",\n\t\t\"start\": t.Desc.Start,\n\t\t\"end\": t.Desc.End,\n\t\t\"status\": t.Desc.Status,\n\t})\n\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\torm.UpdateTask(&t.Desc)\n\tsockets.Message(t.Desc.ID, b)\n}", "func (_BaseContent *BaseContentTransactor) UpdateStatus(opts *bind.TransactOpts, status_code *big.Int) (*types.Transaction, error) {\n\treturn _BaseContent.contract.Transact(opts, \"updateStatus\", status_code)\n}", "func (s *SQLProxySpec) UpdateStatus(rsrc interface{}, reconciled []ResourceInfo, err error) {\n\tstatus := rsrc.(*AirflowBaseStatus)\n\tstatus.SQLProxy = ComponentStatus{}\n\tif s != nil {\n\t\tstatus.SQLProxy.update(reconciled, err)\n\t\tif status.SQLProxy.Status != StatusReady {\n\t\t\tstatus.Status = StatusInProgress\n\t\t}\n\t}\n}", "func (s *SimpleBlockFactory) StatusUpdate(block *types.Block) {\n}", "func (r *CustomDomainReconciler) statusUpdate(reqLogger logr.Logger, instance *customdomainv1alpha1.CustomDomain) error {\n\terr := r.Client.Status().Update(context.TODO(), instance)\n\tif err != nil {\n\t\treqLogger.Error(err, fmt.Sprintf(\"Status update for %s failed\", instance.Name))\n\t}\n\t//reqLogger.Info(fmt.Sprintf(\"Status updated for %s\", instance.Name))\n\treturn err\n}", "func (h *Handler) UpdateStatus(w http.ResponseWriter, r *http.Request) {\n\n\tcmd := sigstat.Command{\n\t\tStatus: \"running\",\n\t}\n\n\th.client.CommandService().UpdateStatus(cmd)\n}", "func (b *FollowUpBuilder) Status(value string) *FollowUpBuilder {\n\tb.status = value\n\tb.bitmap_ |= 2048\n\treturn b\n}", "func (r *Reconciler) UpdateStatus() error {\n\terr := r.Client.Status().Update(r.Ctx, r.NooBaaAccount)\n\tif err != nil {\n\t\tr.Logger.Errorf(\"UpdateStatus: %s\", err)\n\t\treturn err\n\t}\n\tr.Logger.Infof(\"UpdateStatus: Done\")\n\treturn nil\n}", "func (w *ClusterDynamicClient) UpdateStatus(obj *unstructured.Unstructured, options metav1.UpdateOptions) (*unstructured.Unstructured, error) {\n\treturn w.dClient.Resource(w.resource).Namespace(w.namespace).UpdateStatus(w.ctx, obj, options)\n}", "func (m *SchemaExtension) SetStatus(value *string)() {\n err := m.GetBackingStore().Set(\"status\", value)\n if err != nil {\n panic(err)\n }\n}", "func (v Validator) UpdateStatus(newStatus sdk.BondStatus) Validator {\n\tv.Status = newStatus\n\treturn v\n}", "func (r *ManagedServicePollRequest) Status(value int) *ManagedServicePollRequest {\n\tr.statuses = append(r.statuses, value)\n\treturn r\n}", "func (m *ThreatAssessmentRequest) SetStatus(value *ThreatAssessmentStatus)() {\n m.status = value\n}", "func (r *ReconcilerBase) UpdateStatus(obj runtime.Object) error {\n\treturn r.GetClient().Status().Update(context.Background(), obj)\n}", "func (m *Manager) updateStatus(ctx context.Context, status *MessageStatus) {\n\terr := m.r.UpdateStatus(ctx, status)\n\tif err != nil {\n\t\tlog.Log(ctx, errors.Wrap(err, \"update message status\"))\n\t}\n}", "func updateStatus(args []string) {\n\tdata, klocworkURL := formBaseRequest(\"update_status\")\n\tdata.Set(\"project\", args[0])\n\tdata.Set(\"ids\", args[1])\n\tdata.Set(\"status\", args[2])\n\n\tsendRequest(klocworkURL, data)\n\n}", "func (f *Friend) SetStatus(status int) {\n\tf.status = status\n}", "func (t *StatusTracker) UpdateStatus() {\n\tif t.failures < t.halfOpenThreshold {\n\t\tt.status = Closed\n\t} else if t.failures < t.openThreshold {\n\t\tt.status = HalfOpen\n\t} else {\n\t\tt.status = Open\n\t}\n}", "func StatusUpdate(pkt event.Packet) client.RegistryFunc {\n\treturn func(clients client.Registry) error {\n\t\tfrom := pkt.UIDs()[0]\n\n\t\tif _, ok := clients[from]; !ok {\n\t\t\treturn fmt.Errorf(\"for packet numbered %v client %v is not connected\", pkt.Sequence(), from)\n\t\t}\n\n\t\ttargetClient := clients[from]\n\n\t\tfor uid := range targetClient.Followers {\n\t\t\tfollower, ok := clients[uid]\n\t\t\tif !ok {\n\t\t\t\t// Client is no longer present, delete from followers\n\t\t\t\tdelete(targetClient.Followers, uid)\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tif !follower.IsActive() {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tif err := follower.Send(pkt); err != nil {\n\t\t\t\tlog.Debug(fmt.Sprintf(\"notify.StatusUpdate: for client %v, got error %#q\", uid, err))\n\t\t\t\tdelete(targetClient.Followers, uid)\n\n\t\t\t\tclient.UnregisterFunc(uid)(clients)\n\t\t\t}\n\t\t}\n\n\t\treturn nil\n\t}\n}", "func (c *rpcServices) UpdateStatus(rpcService *v1.RpcService) (result *v1.RpcService, err error) {\n\tresult = &v1.RpcService{}\n\terr = c.client.Put().\n\t\tNamespace(c.ns).\n\t\tResource(\"rpcservices\").\n\t\tName(rpcService.Name).\n\t\tSubResource(\"status\").\n\t\tBody(rpcService).\n\t\tDo().\n\t\tInto(result)\n\treturn\n}", "func (m *RetentionEventStatus) SetStatus(value *EventStatusType)() {\n err := m.GetBackingStore().Set(\"status\", value)\n if err != nil {\n panic(err)\n }\n}", "func (w *WaitTask) StatusUpdate(taskContext *TaskContext, id object.ObjMetadata) {\n\tw.mu.Lock()\n\tdefer w.mu.Unlock()\n\n\tif klog.V(5).Enabled() {\n\t\tstatus := taskContext.ResourceCache().Get(id).Status\n\t\tklog.Infof(\"status update (object: %q, status: %q)\", id, status)\n\t}\n\n\tswitch {\n\tcase w.pending.Contains(id):\n\t\tswitch {\n\t\tcase w.changedUID(taskContext, id):\n\t\t\t// replaced\n\t\t\tw.handleChangedUID(taskContext, id)\n\t\t\tw.pending = w.pending.Remove(id)\n\t\tcase w.reconciledByID(taskContext, id):\n\t\t\t// reconciled - remove from pending & send event\n\t\t\terr := taskContext.InventoryManager().SetSuccessfulReconcile(id)\n\t\t\tif err != nil {\n\t\t\t\t// Object never applied or deleted!\n\t\t\t\tklog.Errorf(\"Failed to mark object as successful reconcile: %v\", err)\n\t\t\t}\n\t\t\tw.pending = w.pending.Remove(id)\n\t\t\tw.sendEvent(taskContext, id, event.ReconcileSuccessful)\n\t\tcase w.failedByID(taskContext, id):\n\t\t\t// failed - remove from pending & send event\n\t\t\terr := taskContext.InventoryManager().SetFailedReconcile(id)\n\t\t\tif err != nil {\n\t\t\t\t// Object never applied or deleted!\n\t\t\t\tklog.Errorf(\"Failed to mark object as failed reconcile: %v\", err)\n\t\t\t}\n\t\t\tw.pending = w.pending.Remove(id)\n\t\t\tw.failed = append(w.failed, id)\n\t\t\tw.sendEvent(taskContext, id, event.ReconcileFailed)\n\t\t\t// default - still pending\n\t\t}\n\tcase !w.Ids.Contains(id):\n\t\t// not in wait group - ignore\n\t\treturn\n\tcase w.skipped(taskContext, id):\n\t\t// skipped - ignore\n\t\treturn\n\tcase w.failed.Contains(id):\n\t\t// If a failed resource becomes current before other\n\t\t// resources have completed/timed out, we consider it\n\t\t// current.\n\t\tif w.reconciledByID(taskContext, id) {\n\t\t\t// reconciled - remove from pending & send event\n\t\t\terr := taskContext.InventoryManager().SetSuccessfulReconcile(id)\n\t\t\tif err != nil {\n\t\t\t\t// Object never applied or deleted!\n\t\t\t\tklog.Errorf(\"Failed to mark object as successful reconcile: %v\", err)\n\t\t\t}\n\t\t\tw.failed = w.failed.Remove(id)\n\t\t\tw.sendEvent(taskContext, id, event.ReconcileSuccessful)\n\t\t} else if !w.failedByID(taskContext, id) {\n\t\t\t// If a resource is no longer reported as Failed and is not Reconciled,\n\t\t\t// they should just go back to InProgress.\n\t\t\terr := taskContext.InventoryManager().SetPendingReconcile(id)\n\t\t\tif err != nil {\n\t\t\t\t// Object never applied or deleted!\n\t\t\t\tklog.Errorf(\"Failed to mark object as pending reconcile: %v\", err)\n\t\t\t}\n\t\t\tw.failed = w.failed.Remove(id)\n\t\t\tw.pending = append(w.pending, id)\n\t\t\tw.sendEvent(taskContext, id, event.ReconcilePending)\n\t\t}\n\t\t// else - still failed\n\tdefault:\n\t\t// reconciled - check if unreconciled\n\t\tif !w.reconciledByID(taskContext, id) {\n\t\t\t// unreconciled - add to pending & send event\n\t\t\terr := taskContext.InventoryManager().SetPendingReconcile(id)\n\t\t\tif err != nil {\n\t\t\t\t// Object never applied or deleted!\n\t\t\t\tklog.Errorf(\"Failed to mark object as pending reconcile: %v\", err)\n\t\t\t}\n\t\t\tw.pending = append(w.pending, id)\n\t\t\tw.sendEvent(taskContext, id, event.ReconcilePending)\n\t\t}\n\t\t// else - still reconciled\n\t}\n\n\tklog.V(3).Infof(\"wait task progress: %d/%d\", len(w.Ids)-len(w.pending), len(w.Ids))\n\n\t// If we no longer have any pending resources, the WaitTask\n\t// can be completed.\n\tif len(w.pending) == 0 {\n\t\t// all reconciled, so exit\n\t\tklog.V(3).Infof(\"all objects reconciled or skipped (name: %q)\", w.TaskName)\n\t\tw.cancelFunc()\n\t}\n}", "func (r *Reconciler) setStatus(\n\tctx context.Context,\n\tinstance *v1alpha1.ServiceBindingRequest,\n\tstatus string,\n) error {\n\tinstance.Status.BindingStatus = status\n\treturn r.client.Status().Update(ctx, instance)\n}", "func (r *Reconciler) updateStatus(ctx reconcileRequestContext, processingJobStatus string) error {\n\treturn r.updateStatusWithAdditional(ctx, processingJobStatus, \"\")\n}", "func (m *TeamsAsyncOperation) SetStatus(value *TeamsAsyncOperationStatus)() {\n m.status = value\n}", "func (dtm *DfgetTaskManager) UpdateStatus(ctx context.Context, clientID, taskID, status string) error {\n\tdfgetTask, err := dtm.getDfgetTask(clientID, taskID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif dfgetTask.Status != types.DfGetTaskStatusSUCCESS {\n\t\tdfgetTask.Status = status\n\t}\n\n\treturn nil\n}", "func (r *ManagedServiceUpdateResponse) Status() int {\n\tif r == nil {\n\t\treturn 0\n\t}\n\treturn r.status\n}", "func (nr *namedReceiver) UpdateStatus(ctx context.Context, status *MessageStatus) error {\n\treturn nr.Receiver.UpdateStatus(ctx, status.wrap(ctx, nr.ns))\n}", "func (souo *SubjectsOfferedUpdateOne) SetSTATUS(b bool) *SubjectsOfferedUpdateOne {\n\tsouo.mutation.SetSTATUS(b)\n\treturn souo\n}", "func (k *KubernetesScheduler) StatusUpdate(driver mesos.SchedulerDriver, taskStatus *mesos.TaskStatus) {\n\tlog.Infof(\"Received status update %v\\n\", taskStatus)\n\n\tk.Lock()\n\tdefer k.Unlock()\n\n\tswitch taskStatus.GetState() {\n\tcase mesos.TaskState_TASK_STAGING:\n\t\tk.handleTaskStaging(taskStatus)\n\tcase mesos.TaskState_TASK_STARTING:\n\t\tk.handleTaskStarting(taskStatus)\n\tcase mesos.TaskState_TASK_RUNNING:\n\t\tk.handleTaskRunning(taskStatus)\n\tcase mesos.TaskState_TASK_FINISHED:\n\t\tk.handleTaskFinished(taskStatus)\n\tcase mesos.TaskState_TASK_FAILED:\n\t\tk.handleTaskFailed(taskStatus)\n\tcase mesos.TaskState_TASK_KILLED:\n\t\tk.handleTaskKilled(taskStatus)\n\tcase mesos.TaskState_TASK_LOST:\n\t\tk.handleTaskLost(taskStatus)\n\t}\n}", "func ChangeStatus(c *server.Context) error {\n\tvar (\n\t\terr error\n\t\tchangeReq struct {\n\t\t\tIDs []uint32 `json:\"ids\" validate:\"required,min=1\"`\n\t\t\tStatus int8 `json:\"status\" validate:\"required,eq=-1|eq=1|eq=2|eq=3\"`\n\t\t}\n\t)\n\n\tisAdmin := c.Request().Context().Value(\"user\").(jwtgo.MapClaims)[util.IsAdmin].(bool)\n\tif !isAdmin {\n\t\tlogger.Error(\"You don't have access\")\n\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrToken, nil)\n\t}\n\n\terr = c.JSONBody(&changeReq)\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrInvalidParam, nil)\n\t}\n\n\terr = c.Validate(changeReq)\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrInvalidParam, nil)\n\t}\n\n\tconn, err := mysql.Pool.Get()\n\tdefer mysql.Pool.Release(conn)\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrMysql, nil)\n\t}\n\n\t// status: -1 -> delete or sold out\n\t// 1 -> common\n\t// 2 -> promotion\n\t// 3 -> today new wares\n\terr = ware.Service.ChangeStatus(conn, changeReq.IDs, changeReq.Status)\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrMysql, nil)\n\t}\n\n\tlogger.Info(\"change ware status success\")\n\treturn core.WriteStatusAndDataJSON(c, constants.ErrSucceed, nil)\n}", "func updateAgentStatus(status, message string) error {\n\t// IMP - To be removed once the model is in production\n\tif agent.cfg == nil || agent.cfg.GetAgentName() == \"\" {\n\t\treturn nil\n\t}\n\n\tif agent.agentResource != nil {\n\t\tagentResourceType := getAgentResourceType()\n\t\tresource := createAgentStatusSubResource(agentResourceType, status, message)\n\t\tbuffer, err := json.Marshal(resource)\n\t\tif err != nil {\n\t\t\treturn nil\n\t\t}\n\n\t\tsubResURL := agent.cfg.GetEnvironmentURL() + \"/\" + agentResourceType + \"/\" + agent.cfg.GetAgentName() + \"/status\"\n\t\t_, err = agent.apicClient.ExecuteAPI(coreapi.PUT, subResURL, nil, buffer)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (m *ScheduleItem) SetStatus(value *FreeBusyStatus)() {\n err := m.GetBackingStore().Set(\"status\", value)\n if err != nil {\n panic(err)\n }\n}", "func (s *MySQLSpec) UpdateStatus(rsrc interface{}, reconciled []ResourceInfo, err error) {\n\tstatus := rsrc.(*AirflowBaseStatus)\n\tstatus.MySQL = ComponentStatus{}\n\tif s != nil {\n\t\tstatus.MySQL.update(reconciled, err)\n\t\tif status.MySQL.Status != StatusReady {\n\t\t\tstatus.Status = StatusInProgress\n\t\t}\n\t}\n}", "func (m *LongRunningOperation) SetStatus(value *LongRunningOperationStatus)() {\n err := m.GetBackingStore().Set(\"status\", value)\n if err != nil {\n panic(err)\n }\n}", "func (client *ClientRPCMethods) Status(in *string, response *ServerStatus) (err error) {\n\t*response = *client.client.callback.Status()\n\treturn nil\n}", "func (clt *client) setStatus(newStatus Status) {\n\tclt.statusLock.Lock()\n\tclt.status = newStatus\n\tclt.statusLock.Unlock()\n}", "func (ktuo *KqiTargetUpdateOne) SetStatus(b bool) *KqiTargetUpdateOne {\n\tktuo.mutation.SetStatus(b)\n\treturn ktuo\n}", "func (UserService) UpdateStatus(dto dto.UserEditStatusDto) int64 {\n\tu := userDao.Get(dto.Id, false)\n\t//u.Status = dto.Status\n\tc := userDao.Update(&u, map[string]interface{}{\n\t\t\"status\": dto.Status,\n\t})\n\treturn c.RowsAffected\n}", "func (self *client) GetStatus() {\n\n}", "func UpdateStatus(status, description string) {\n\tupdateAgentStatus(status, description)\n}", "func updateStatus(e *event) {\n\tfMap := followers[e.from]\n\tfor _, f := range fMap {\n\t\tif h, ok := clients.Get(f); ok {\n\t\t\th.Write(e)\n\t\t}\n\t}\n}", "func (s *NFSStoreSpec) UpdateStatus(rsrc interface{}, reconciled []ResourceInfo, err error) {\n\tstatus := rsrc.(*AirflowBaseStatus)\n\tstatus.Storage = ComponentStatus{}\n\tif s != nil {\n\t\tstatus.Storage.update(reconciled, err)\n\t\tif status.Storage.Status != StatusReady {\n\t\t\tstatus.Status = StatusInProgress\n\t\t}\n\t}\n}", "func (c *CSV) UpdateStatus(a *config.AppContext) error {\n\t/*\n\t * We will update the status\n\t */\n\treturn a.Db.Model(&c.Resource).Updates(map[string]interface{}{\n\t\t\"status\": c.Resource.Status,\n\t}).Error\n}", "func (a *DefaultApiService) Status(ctx _context.Context) ApiStatusRequest {\n\treturn ApiStatusRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t}\n}", "func (s *AirflowUISpec) UpdateStatus(rsrc interface{}, reconciled []ResourceInfo, err error) {\n\tstatus := rsrc.(*AirflowClusterStatus)\n\tstatus.UI = ComponentStatus{}\n\tif s != nil {\n\t\tstatus.UI.update(reconciled, err)\n\t\tif status.UI.Status != StatusReady {\n\t\t\tstatus.Status = StatusInProgress\n\t\t}\n\t}\n}", "func (puo *ProfileUpdateOne) SetStatus(b bool) *ProfileUpdateOne {\n\tpuo.mutation.SetStatus(b)\n\treturn puo\n}", "func (c *client) Status(u *model.User, r *model.Repo, b *model.Build, link string) error {\n\tclient := c.newClientToken(u.Token)\n\n\tstatus := getStatus(b.Status)\n\tdesc := getDesc(b.Status)\n\n\t_, err := client.CreateStatus(\n\t\tr.Owner,\n\t\tr.Name,\n\t\tb.Commit,\n\t\tgitea.CreateStatusOption{\n\t\t\tState: status,\n\t\t\tTargetURL: link,\n\t\t\tDescription: desc,\n\t\t\tContext: c.Context,\n\t\t},\n\t)\n\n\treturn err\n}", "func (b *ServiceClusterBuilder) Status(value string) *ServiceClusterBuilder {\n\tb.status = value\n\tb.bitmap_ |= 256\n\treturn b\n}", "func (m *AccessPackageAssignment) SetStatus(value *string)() {\n m.status = value\n}", "func (o *Status) Update() {\n o.Time = time.Now()\n}", "func (c *cachestub) UpdateRecordStatus(token, version string, status cache.RecordStatusT, timestamp int64, metadata []cache.MetadataStream) error {\n\treturn nil\n}", "func (oiu *OrderInfoUpdate) SetStatus(i int8) *OrderInfoUpdate {\n\toiu.mutation.ResetStatus()\n\toiu.mutation.SetStatus(i)\n\treturn oiu\n}", "func (m *ThreatAssessmentRequest) SetStatus(value *ThreatAssessmentStatus)() {\n err := m.GetBackingStore().Set(\"status\", value)\n if err != nil {\n panic(err)\n }\n}", "func (oiuo *OrderInfoUpdateOne) SetStatus(i int8) *OrderInfoUpdateOne {\n\toiuo.mutation.ResetStatus()\n\toiuo.mutation.SetStatus(i)\n\treturn oiuo\n}", "func (b *ClusterBuilder) Status(value *ClusterStatusBuilder) *ClusterBuilder {\n\tb.status = value\n\tif value != nil {\n\t\tb.bitmap_ |= 549755813888\n\t} else {\n\t\tb.bitmap_ &^= 549755813888\n\t}\n\treturn b\n}", "func UpdateStatus(APIstub shim.ChaincodeStubInterface, args []string, txnID string) sc.Response {\n\tfmt.Println(\"UpdateStatus Initial\")\n\tfmt.Println(args)\n\texistingClaimAsBytes, _ := APIstub.GetState(args[0])\n\n\tclaim := Claim{}\n\tjson.Unmarshal(existingClaimAsBytes, &claim)\n\n\tclaim.Status = args[1]\n\n\tclaimAsBytes, _ := json.Marshal(claim)\n\tAPIstub.PutState(args[0], claimAsBytes)\n\n\ttimestamp, _ := APIstub.GetTxTimestamp()\n\ttimestampAsInt := timestamp.GetSeconds()\n\tisotimestamp := time.Unix(timestampAsInt, 0).Format(time.RFC3339)\n\ttxnDetails := []string{txnID, \"CSU - Claim Status Update\", isotimestamp, args[1], claim.ID}\n\tfmt.Println(txnDetails)\n\tfmt.Println(txn.Add(APIstub, txnDetails))\n\ttxn.Add(APIstub, txnDetails)\n\treturn shim.Success(claimAsBytes)\n\n}", "func (sry *Sryun) Status(user *model.User, repo *model.Repo, build *model.Build, link string) error {\n\treturn nil\n}", "func (m *IndustryDataRunActivity) SetStatus(value *IndustryDataActivityStatus)() {\n err := m.GetBackingStore().Set(\"status\", value)\n if err != nil {\n panic(err)\n }\n}", "func (r *ReferenceAdapter) StatusUpdate() error {\n\terr := r.kubeClient.Status().Update(context.TODO(), r.ProjectReference)\n\tif err != nil {\n\t\treturn operrors.Wrap(err, fmt.Sprintf(\"failed to update ProjectReference status of %s\", r.ProjectReference.Name))\n\t}\n\n\treturn nil\n}", "func (ktu *KqiTargetUpdate) SetStatus(b bool) *KqiTargetUpdate {\n\tktu.mutation.SetStatus(b)\n\treturn ktu\n}", "func (m *SynchronizationJob) SetStatus(value SynchronizationStatusable)() {\n err := m.GetBackingStore().Set(\"status\", value)\n if err != nil {\n panic(err)\n }\n}", "func (d *Device) SetStatus(status status.FieldDeviceStatus) { d.status = status }", "func (sou *SubjectsOfferedUpdate) SetSTATUS(b bool) *SubjectsOfferedUpdate {\n\tsou.mutation.SetSTATUS(b)\n\treturn sou\n}", "func (s *SchedulerSpec) UpdateStatus(rsrc interface{}, reconciled []ResourceInfo, err error) {\n\tstatus := rsrc.(*AirflowClusterStatus)\n\tstatus.Scheduler = SchedulerStatus{}\n\tif s != nil {\n\t\tstatus.Scheduler.Resources.update(reconciled, err)\n\t\tif status.Scheduler.Resources.Status != StatusReady {\n\t\t\tstatus.Status = StatusInProgress\n\t\t}\n\t}\n}", "func (client MockStatusClient) Update(context ctx.Context, object ctrlClient.Object, options ...ctrlClient.UpdateOption) error {\n\tkindKey, err := buildKindKey(object)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tclient.rawClient.fillInMaps(kindKey)\n\n\tjsonData, err := json.Marshal(object)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tobjectKey, err := buildJSONObjectKey(jsonData)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = client.rawClient.checkPresence(kindKey, objectKey)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\texistingObject := client.rawClient.data[kindKey][objectKey]\n\n\tnewObject := make(map[string]interface{})\n\terr = json.Unmarshal(jsonData, &newObject)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\texistingObject[\"status\"] = newObject[\"status\"]\n\tclient.rawClient.data[kindKey][objectKey] = existingObject\n\n\treturn nil\n}", "func (m *WorkbookOperation) SetStatus(value *WorkbookOperationStatus)() {\n m.status = value\n}", "func (driver *MesosExecutorDriver) SendStatusUpdate(taskStatus *mesosproto.TaskStatus) (mesosproto.Status, error) {\n\tlog.Infoln(\"Sending status update\")\n\n\tdriver.mutex.Lock()\n\tdefer driver.mutex.Unlock()\n\n\tif taskStatus.GetState() == mesosproto.TaskState_TASK_STAGING {\n\t\tlog.Errorf(\"Executor is not allowed to send TASK_STAGING status update. Aborting!\\n\")\n\t\tdriver.Abort()\n\t\terr := fmt.Errorf(\"Attempted to send TASK_STAGING status update\")\n\t\tdriver.Executor.Error(driver, err.Error())\n\t\treturn driver.status, err\n\t}\n\n\t// Set up status update.\n\tupdate := driver.makeStatusUpdate(taskStatus)\n\tlog.Infof(\"Executor sending status update %v\\n\", update.String())\n\n\t// Capture the status update.\n\tdriver.updates[uuid.UUID(update.GetUuid()).String()] = update\n\n\t// Put the status update in the message.\n\tmessage := &mesosproto.StatusUpdateMessage{\n\t\tUpdate: update,\n\t\tPid: proto.String(driver.self.String()),\n\t}\n\t// Send the message.\n\tif err := driver.messenger.Send(driver.slaveUPID, message); err != nil {\n\t\tlog.Errorf(\"Failed to send %v: %v\\n\")\n\t\treturn driver.status, err\n\t}\n\treturn driver.status, nil\n}", "func (client *Client) ServiceStatus(request *ServiceStatusRequest) (response *ServiceStatusResponse, err error) {\nresponse = CreateServiceStatusResponse()\nerr = client.DoAction(request, response)\nreturn\n}", "func (svc *Service) updateContractStatus(ctx context.Context, req *apistructs.UpdateContractReq, client *apistructs.ClientModel, access *apistructs.APIAccessesModel,\n\tcontract *apistructs.ContractModel) error {\n\torg, err := svc.getOrg(context.Background(), req.OrgID)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to GetOrg\")\n\t}\n\n\tif req.Body.Status == nil {\n\t\treturn nil\n\t}\n\n\t// do something depends on contract status\n\tstatus := req.Body.Status.ToLower()\n\tswitch status {\n\tcase apistructs.ContractApproved:\n\t\t// the manager change the contract status to \"approved\", call the api-gateway to grant to the client\n\t\tif err := bdl.Bdl.GrantEndpointToClient(strconv.FormatUint(req.OrgID, 10), req.Identity.UserID,\n\t\t\tclient.ClientID, access.EndpointID); err != nil {\n\t\t\treturn err\n\t\t}\n\tcase apistructs.ContractDisapproved:\n\t\t// do nothing with api-gateway\n\tcase apistructs.ContractUnapproved:\n\t\t// call the api-gateway to revoke the grant\n\t\tif err := bdl.Bdl.RevokeEndpointFromClient(strconv.FormatUint(req.OrgID, 10), req.Identity.UserID,\n\t\t\tclient.ClientID, access.EndpointID); err != nil {\n\t\t\treturn err\n\t\t}\n\tdefault:\n\t\treturn errors.New(\"invalid contract status\")\n\t}\n\n\ttimeNow := time.Now()\n\ttx := dbclient.Tx()\n\tdefer tx.RollbackUnlessCommitted()\n\n\tif err := tx.Model(contract).\n\t\tWhere(map[string]interface{}{\"org_id\": req.OrgID, \"id\": req.URIParams.ContractID}).\n\t\tUpdates(map[string]interface{}{\"status\": status, \"updated_at\": timeNow}).\n\t\tError; err != nil {\n\t\treturn errors.Wrap(err, \"failed to Updates contractModel\")\n\t}\n\n\taction := status2Action(status)\n\tif err := tx.Create(&apistructs.ContractRecordModel{\n\t\tID: 0,\n\t\tOrgID: req.OrgID,\n\t\tContractID: contract.ID,\n\t\tAction: fmt.Sprintf(\"%s对该调用申请的授权\", action),\n\t\tCreatorID: req.Identity.UserID,\n\t\tCreatedAt: timeNow,\n\t}).Error; err != nil {\n\t\treturn errors.Wrap(err, \"failed to Create contractRecordModel\")\n\t}\n\n\ttx.Commit()\n\n\t// notification by mail and in-site letter\n\tgo svc.contractMsgToUser(req.OrgID, contract.CreatorID, access.AssetName, client,\n\t\tsvc.ApprovalResultFromStatus(ctx, status, org.Locale))\n\n\treturn nil\n}", "func (s *FlowerSpec) UpdateStatus(rsrc interface{}, reconciled []ResourceInfo, err error) {\n\tstatus := rsrc.(*AirflowClusterStatus)\n\tstatus.Flower = ComponentStatus{}\n\tif s != nil {\n\t\tstatus.Flower.update(reconciled, err)\n\t\tif status.Flower.Status != StatusReady {\n\t\t\tstatus.Status = StatusInProgress\n\t\t}\n\t}\n}", "func (c *Controller) updateStatus(name types.NamespacedName, update func(status *CmdStatus), stillHasSameProcNum func() bool) {\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\n\tif !stillHasSameProcNum() {\n\t\treturn\n\t}\n\n\tcmd, ok := c.updateCmds[name]\n\tif !ok {\n\t\treturn\n\t}\n\n\tupdate(&cmd.Status)\n\n\terr := c.client.Status().Update(c.globalCtx, cmd)\n\tif err != nil && !apierrors.IsNotFound(err) {\n\t\tc.st.Dispatch(store.NewErrorAction(fmt.Errorf(\"syncing to apiserver: %v\", err)))\n\t\treturn\n\t}\n\n\tc.st.Dispatch(local.NewCmdUpdateStatusAction(cmd))\n}", "func (m *ConnectorStatusDetails) SetStatus(value *ConnectorHealthState)() {\n err := m.GetBackingStore().Set(\"status\", value)\n if err != nil {\n panic(err)\n }\n}", "func (sh *StatusHandler) Status(ctx context.Context, in *empty.Empty, out *proto.StatusResponse) error {\n\tout.OK = true\n\tout.Address = sh.address\n\n\treturn nil\n}", "func (r *NodeReconciler) updateStatus(ctx context.Context, node *filecoinv1alpha1.Node) error {\n\t// TODO: update after multi-client support\n\tnode.Status.Client = \"lotus\"\n\n\tif err := r.Status().Update(ctx, node); err != nil {\n\t\tr.Log.Error(err, \"unable to update filecoin node status\")\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (r *Client) UpdateStatus(ctx context.Context, obj client.Object, state alertmanagerv1alpha1.State, requeueTime ...float64) (ctrl.Result, error) {\n\tlog := log.Logger(ctx, \"controllers.common\", \"common\", \"UpdateStatus\")\n\n\tif err := r.Status().Update(ctx, obj); err != nil {\n\t\tlog.Error(err, \"Unable to update status\", \"status\", state)\n\t\tr.Recorder.Event(obj, v1.EventTypeWarning, string(alertmanagerv1alpha1.Error), \"Unable to create/update status due to error \"+err.Error())\n\t\treturn ctrl.Result{RequeueAfter: 30 * time.Second}, nil\n\t}\n\n\tif state != alertmanagerv1alpha1.Error {\n\t\treturn ctrl.Result{}, nil\n\t}\n\n\t//if wait time is specified, requeue it after provided time\n\tif len(requeueTime) == 0 {\n\t\trequeueTime[0] = 0\n\t}\n\n\tlog.Info(\"Requeue time\", \"time\", requeueTime[0])\n\treturn ctrl.Result{RequeueAfter: time.Duration(requeueTime[0]) * time.Millisecond}, nil\n}", "func (w *ServerInterfaceWrapper) UpdateUserStatus(ctx echo.Context) error {\n\tvar err error\n\t// ------------- Path parameter \"id\" -------------\n\tvar id int\n\n\terr = runtime.BindStyledParameter(\"simple\", false, \"id\", ctx.Param(\"id\"), &id)\n\tif err != nil {\n\t\treturn echo.NewHTTPError(http.StatusBadRequest, fmt.Sprintf(\"Invalid format for parameter id: %s\", err))\n\t}\n\n\tctx.Set(\"OAuth.Scopes\", []string{\"\"})\n\n\t// Invoke the callback with all the unmarshalled arguments\n\terr = w.Handler.UpdateUserStatus(ctx, id)\n\treturn err\n}", "func (v1 *EndpointsV1) GenStatusEndpoint(r *http.Request) interface{} {\n\treturn Status{\n\t\tStatus: \"ok\",\n\t\tName: \"Goscout\",\n\t\tVersion: gitVersion(),\n\t\tServerTime: time.Now().String(),\n\t\tServerTimeEpoch: time.Now().UnixNano() / int64(time.Millisecond),\n\t\tAPIEnabled: true,\n\t\tCareportalEnabled: true,\n\t\tBoluscalcEnabled: true,\n\t\tSettings: &StatusSettings{},\n\t\tExtendedSettings: &StatusExtendedSettings{\n\t\t\tDeviceStatus: struct {\n\t\t\t\tAdvanced bool `json:\"advanced\"`\n\t\t\t}{\n\t\t\t\tAdvanced: true,\n\t\t\t},\n\t\t},\n\t\tAuthorized: nil,\n\t}\n}", "func (r *reflectedStatusAccessor) SetStatus(status interface{}) {\n\tif r != nil && r.status.IsValid() && r.status.CanSet() {\n\t\tr.status.Set(reflect.ValueOf(status))\n\t}\n}", "func (o *Cause) UnsetStatus() {\n\to.Status.Unset()\n}", "func (r *ClusterPollRequest) Status(value int) *ClusterPollRequest {\n\tr.statuses = append(r.statuses, value)\n\treturn r\n}", "func (service *PrivateLinkService) SetStatus(status genruntime.ConvertibleStatus) error {\n\t// If we have exactly the right type of status, assign it\n\tif st, ok := status.(*PrivateLinkService_STATUS_PrivateLinkService_SubResourceEmbedded); ok {\n\t\tservice.Status = *st\n\t\treturn nil\n\t}\n\n\t// Convert status to required version\n\tvar st PrivateLinkService_STATUS_PrivateLinkService_SubResourceEmbedded\n\terr := status.ConvertStatusTo(&st)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to convert status\")\n\t}\n\n\tservice.Status = st\n\treturn nil\n}", "func InstanceViewStatus_STATUSGenerator() gopter.Gen {\n\tif instanceViewStatus_STATUSGenerator != nil {\n\t\treturn instanceViewStatus_STATUSGenerator\n\t}\n\n\tgenerators := make(map[string]gopter.Gen)\n\tAddIndependentPropertyGeneratorsForInstanceViewStatus_STATUS(generators)\n\tinstanceViewStatus_STATUSGenerator = gen.Struct(reflect.TypeOf(InstanceViewStatus_STATUS{}), generators)\n\n\treturn instanceViewStatus_STATUSGenerator\n}", "func UpdateStatus(db gorp.SqlExecutor, id int64, status string) error {\n\t_, err := db.Exec(\"UPDATE cds_migration SET status = $1 WHERE id = $2\", status, id)\n\treturn err\n}", "func (s *PostgresSpec) UpdateStatus(rsrc interface{}, reconciled []ResourceInfo, err error) {\n\tstatus := rsrc.(*AirflowBaseStatus)\n\tstatus.Postgres = ComponentStatus{}\n\tif s != nil {\n\t\tstatus.Postgres.update(reconciled, err)\n\t\tif status.Postgres.Status != StatusReady {\n\t\t\tstatus.Status = StatusInProgress\n\t\t}\n\t}\n}", "func (c *scheduledJobs) UpdateStatus(job *batch.ScheduledJob) (result *batch.ScheduledJob, err error) {\n\tresult = &batch.ScheduledJob{}\n\terr = c.r.Put().Namespace(c.ns).Resource(\"scheduledjobs\").Name(job.Name).SubResource(\"status\").Body(job).Do().Into(result)\n\treturn\n}", "func (c *Client) Status() error {\n\tclient, err := c.client(false)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\trel, err := url.Parse(\"status\")\n\tif err != nil {\n\t\t// This indicates a programming error.\n\t\tpanic(err)\n\t}\n\n\tresp, err := client.Get(c.baseURL.ResolveReference(rel).String())\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\n\tif resp.StatusCode != http.StatusNoContent {\n\t\treturn ErrInvalidServiceBehavior\n\t}\n\treturn nil\n}", "func (pu *ProfileUpdate) SetStatus(b bool) *ProfileUpdate {\n\tpu.mutation.SetStatus(b)\n\treturn pu\n}", "func Status(status string) error {\n\treturn SdNotify(\"STATUS=\" + status)\n}", "func (v *Status) Update() error {\n\treturn nil\n}", "func (r *ChartReconciler) UpdateStatus(c *stablev1.Chart) error {\n\tif err := r.Status().Update(ctx, c); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func VirtualMachinePatchStatus_STATUSGenerator() gopter.Gen {\n\tif virtualMachinePatchStatus_STATUSGenerator != nil {\n\t\treturn virtualMachinePatchStatus_STATUSGenerator\n\t}\n\n\tgenerators := make(map[string]gopter.Gen)\n\tAddRelatedPropertyGeneratorsForVirtualMachinePatchStatus_STATUS(generators)\n\tvirtualMachinePatchStatus_STATUSGenerator = gen.Struct(reflect.TypeOf(VirtualMachinePatchStatus_STATUS{}), generators)\n\n\treturn virtualMachinePatchStatus_STATUSGenerator\n}", "func (puo *PostUpdateOne) SetStatus(i int8) *PostUpdateOne {\n\tpuo.mutation.ResetStatus()\n\tpuo.mutation.SetStatus(i)\n\treturn puo\n}", "func (suite *TestManagerSuite) TestManagerUpdateStatus() {\n\tl, err := suite.m.GetBy(\"d1000\", \"ruuid\", []string{v1.MimeTypeNativeReport})\n\trequire.NoError(suite.T(), err)\n\trequire.Equal(suite.T(), 1, len(l))\n\n\toldSt := l[0].Status\n\n\terr = suite.m.UpdateStatus(\"tid001\", job.SuccessStatus.String(), 10000)\n\trequire.NoError(suite.T(), err)\n\n\tl, err = suite.m.GetBy(\"d1000\", \"ruuid\", []string{v1.MimeTypeNativeReport})\n\trequire.NoError(suite.T(), err)\n\trequire.Equal(suite.T(), 1, len(l))\n\n\tassert.NotEqual(suite.T(), oldSt, l[0].Status)\n\tassert.Equal(suite.T(), job.SuccessStatus.String(), l[0].Status)\n}", "func (pu *PostUpdate) SetStatus(i int8) *PostUpdate {\n\tpu.mutation.ResetStatus()\n\tpu.mutation.SetStatus(i)\n\treturn pu\n}", "func (s *eremeticScheduler) StatusUpdate(driver sched.SchedulerDriver, status *mesos.TaskStatus) {\n\tid := status.TaskId.GetValue()\n\n\tlog.Debugf(\"Received task status [%s] for task [%s]\", status.State.String(), id)\n\n\ttask, err := database.ReadTask(id)\n\tif err != nil {\n\t\tlog.Debugf(\"Error reading task from database: %s\", err)\n\t}\n\n\tif task.ID == \"\" {\n\t\ttask = types.EremeticTask{\n\t\t\tID: id,\n\t\t\tSlaveId: status.SlaveId.GetValue(),\n\t\t}\n\t}\n\n\tif !task.IsRunning() && *status.State == mesos.TaskState_TASK_RUNNING {\n\t\tTasksRunning.Inc()\n\t}\n\n\tif types.IsTerminal(status.State) {\n\t\tTasksTerminated.With(prometheus.Labels{\"status\": status.State.String()}).Inc()\n\t\tif task.WasRunning() {\n\t\t\tTasksRunning.Dec()\n\t\t}\n\t}\n\n\ttask.UpdateStatus(types.Status{\n\t\tStatus: status.State.String(),\n\t\tTime: time.Now().Unix(),\n\t})\n\n\tif *status.State == mesos.TaskState_TASK_FAILED && !task.WasRunning() {\n\t\tif task.Retry >= maxRetries {\n\t\t\tlog.Warnf(\"giving up on %s after %d retry attempts\", id, task.Retry)\n\t\t} else {\n\t\t\tlog.Infof(\"task %s was never running. re-scheduling\", id)\n\t\t\ttask.UpdateStatus(types.Status{\n\t\t\t\tStatus: mesos.TaskState_TASK_STAGING.String(),\n\t\t\t\tTime: time.Now().Unix(),\n\t\t\t})\n\t\t\ttask.Retry++\n\t\t\tgo func() {\n\t\t\t\tQueueSize.Inc()\n\t\t\t\ts.tasks <- id\n\t\t\t}()\n\t\t}\n\t}\n\n\tif types.IsTerminal(status.State) {\n\t\thandler.NotifyCallback(&task)\n\t}\n\n\tdatabase.PutTask(&task)\n}", "func (comp *Competition) UpdateStatus(statusID int) error {\n\tif comp.statusID >= statusID && comp.statusID != 0 {\n\t\treturn errors.New(\"cannot revert competition status\")\n\t}\n\tif comp.statusID == CompetitionStatusClosed || comp.statusID == CompetitionStatusCancelled {\n\t\treturn errors.New(\"competition status is locked\")\n\t}\n\tcomp.statusID = statusID\n\treturn nil\n}" ]
[ "0.6663045", "0.6435942", "0.6429158", "0.635048", "0.6284993", "0.62775517", "0.6203051", "0.6202162", "0.61929494", "0.618797", "0.61788225", "0.61285967", "0.6122142", "0.60842466", "0.6038285", "0.60271007", "0.6014957", "0.59953797", "0.59856516", "0.597921", "0.59776646", "0.59532565", "0.593746", "0.5922131", "0.59207433", "0.59121835", "0.58814716", "0.58724874", "0.5869423", "0.58671516", "0.5864305", "0.5863733", "0.58631915", "0.5852475", "0.5847568", "0.5837685", "0.5835927", "0.58247745", "0.58088094", "0.58031774", "0.5786231", "0.57851446", "0.57780606", "0.5776926", "0.57693136", "0.5765461", "0.57574177", "0.5746787", "0.574417", "0.57411295", "0.573268", "0.57267946", "0.57240874", "0.57195604", "0.57145727", "0.5712587", "0.5711367", "0.5709272", "0.57057804", "0.57030886", "0.57029295", "0.5695423", "0.5687883", "0.5679442", "0.5677195", "0.56765896", "0.56744564", "0.5672378", "0.56699944", "0.56695646", "0.5667887", "0.5664797", "0.5661475", "0.5659772", "0.5654756", "0.56543714", "0.5647724", "0.56453705", "0.56410736", "0.56381613", "0.5636736", "0.56303465", "0.56278896", "0.56261224", "0.56252176", "0.5623226", "0.56204957", "0.56199557", "0.56199163", "0.5607087", "0.5601604", "0.5596981", "0.5589353", "0.5582313", "0.5563451", "0.55618626", "0.5561496", "0.55580205", "0.55481654", "0.5548146", "0.5546109" ]
0.0
-1
Delete takes name of the tZCronJob and deletes it. Returns an error if one occurs.
func (c *FakeTZCronJobs) Delete(name string, options *v1.DeleteOptions) error { _, err := c.Fake. Invokes(testing.NewDeleteAction(tzcronjobsResource, c.ns, name), &v1alpha1.TZCronJob{}) return err }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *scheduledJobs) Delete(name string, options *api.DeleteOptions) (err error) {\n\treturn c.r.Delete().Namespace(c.ns).Resource(\"scheduledjobs\").Name(name).Body(options).Do().Error()\n}", "func (c *aITrainingJobs) Delete(name string, options *metav1.DeleteOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"aitrainingjobs\").\n\t\tName(name).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func DeleteJob(ctx *gin.Context) {\n\tlog := logger.RuntimeLog\n\tzoneName := ctx.Param(\"zone\")\n\tnamespace := ctx.Param(\"ns\")\n\tname := ctx.Param(\"name\")\n\n\tkclient, err := GetClientByAzCode(zoneName)\n\tif err != nil {\n\t\tSendResponse(ctx, errno.ErrBind, nil)\n\t\treturn\n\t}\n\n\tlog.Info(\"Deleting deployment...\")\n\n\tdeletePolicy := metav1.DeletePropagationForeground\n\n\tstartAt := time.Now()\n\terr = kclient.BatchV1().Jobs(namespace).Delete(name, &metav1.DeleteOptions{\n\t\tPropagationPolicy: &deletePolicy,\n\t})\n\tlogger.MetricsEmit(\n\t\tJOB_CONST.K8S_LOG_Method_DeleteJob,\n\t\tutil.GetReqID(ctx),\n\t\tfloat32(time.Since(startAt)/time.Millisecond),\n\t\terr == nil || errors.IsNotFound(err),\n\t)\n\n\tif err != nil {\n\t\tif errors.IsNotFound(err) {\n\t\t\tlog.Infof(\"Job %s not found in k8s\", name)\n\t\t\tSendResponse(ctx, err, nil)\n\t\t\treturn\n\t\t}\n\t\tSendResponse(ctx, err, nil)\n\t\treturn\n\t}\n\n\tSendResponse(ctx, errno.OK, nil)\n\treturn\n}", "func (j *Job) Delete(ctx context.Context) (err error) {\n\tctx = trace.StartSpan(ctx, \"cloud.google.com/go/bigquery.Job.Delete\")\n\tdefer func() { trace.EndSpan(ctx, err) }()\n\n\tcall := j.c.bqs.Jobs.Delete(j.projectID, j.jobID).Context(ctx)\n\tif j.location != \"\" {\n\t\tcall = call.Location(j.location)\n\t}\n\tsetClientHeader(call.Header())\n\n\treturn runWithRetry(ctx, func() (err error) {\n\t\tsCtx := trace.StartSpan(ctx, \"bigquery.jobs.delete\")\n\t\terr = call.Do()\n\t\ttrace.EndSpan(sCtx, err)\n\t\treturn err\n\t})\n}", "func (j *Job) Delete() error {\n\treturn j.conn.Delete(j.ID)\n}", "func Delete(ctx context.Context, c client.Client, key client.ObjectKey) error {\n\tcj := New(key.Name, key.Namespace, nil).Build()\n\n\tif err := c.Delete(ctx, cj, &client.DeleteOptions{}); err != nil {\n\t\treturn kverrors.Wrap(err, \"failed to delete cronjob\",\n\t\t\t\"name\", cj.Name,\n\t\t\t\"namespace\", cj.Namespace,\n\t\t)\n\t}\n\n\treturn nil\n}", "func (r *JobsService) Delete(name string) *JobsDeleteCall {\n\tc := &JobsDeleteCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.name = name\n\treturn c\n}", "func (client JobClient) Delete(ctx context.Context, resourceGroupName string, accountName string, jobName string) (result JobDeleteFuture, err error) {\n if tracing.IsEnabled() {\n ctx = tracing.StartSpan(ctx, fqdn + \"/JobClient.Delete\")\n defer func() {\n sc := -1\n if result.Response() != nil {\n sc = result.Response().StatusCode\n }\n tracing.EndSpan(ctx, sc, err)\n }()\n }\n req, err := client.DeletePreparer(ctx, resourceGroupName, accountName, jobName)\n if err != nil {\n err = autorest.NewErrorWithError(err, \"microsoftazuremanagementaisupercomputer.JobClient\", \"Delete\", nil , \"Failure preparing request\")\n return\n }\n\n result, err = client.DeleteSender(req)\n if err != nil {\n err = autorest.NewErrorWithError(err, \"microsoftazuremanagementaisupercomputer.JobClient\", \"Delete\", result.Response(), \"Failure sending request\")\n return\n }\n\n return\n}", "func DeleteJob(ctx context.Context, p string) (err error) {\n\tt := utils.FromTaskContext(ctx)\n\n\treturn contexts.DB.Delete(constants.FormatJobKey(t, p), nil)\n}", "func (k *k8sService) deleteCronJob(name string) error {\n\terr := k.client.BatchV1beta1().CronJobs(defaultNS).Delete(name, nil)\n\tif err == nil {\n\t\tlog.Infof(\"Deleted CronJob %v\", name)\n\t} else if errors.IsNotFound(err) {\n\t\tlog.Infof(\"CronJob %v not found\", name)\n\t\terr = nil\n\t} else {\n\t\tlog.Errorf(\"Failed to delete CronJob %v with error: %v\", name, err)\n\t}\n\n\treturn err\n}", "func (c *Clientset) DeleteJob(namespace, name string) error {\n\treturn c.kubernetes.BatchV1().Jobs(namespace).Delete(name, &defaultDeleteOptions)\n}", "func (wj *WsubJob) Delete(db XODB) error {\n\tvar err error\n\n\t// if doesn't exist, bail\n\tif !wj._exists {\n\t\treturn nil\n\t}\n\n\t// if deleted, bail\n\tif wj._deleted {\n\t\treturn nil\n\t}\n\n\t// sql query\n\tconst sqlstr = `DELETE FROM jlabo.wsub_jobs WHERE id = ?`\n\n\t// run query\n\tXOLog(sqlstr, wj.ID)\n\t_, err = db.Exec(sqlstr, wj.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set deleted\n\twj._deleted = true\n\n\treturn nil\n}", "func TestDeleteJob(t *testing.T) {\n\tc := &kc{}\n\tif err := deleteJob(c, \"job-name\", github.PullRequest{}); err != nil {\n\t\tt.Fatalf(\"Didn't expect error deleting job: %v\", err)\n\t}\n\t// The default kube.Job has nil parallelism and 0 succeeded pods. Ensure\n\t// that we explicitly zeroed parallelism.\n\tif c.job.Spec.Parallelism == nil || *c.job.Spec.Parallelism != 0 {\n\t\tt.Error(\"Didn't set parallelism to 0.\")\n\t}\n}", "func DeleteJob(w http.ResponseWriter, r *http.Request) {\n\tresponse := services.DeleteJob(r)\n\n\trender.Status(r, response.Code)\n\trender.JSON(w, r, response)\n}", "func DeleteJob(ns, name string) error {\n\tdeletePolicy := metav1.DeletePropagationForeground\n\terr := kubecli.BatchV1().Jobs(ns).Delete(name, &metav1.DeleteOptions{\n\t\tPropagationPolicy: &deletePolicy,\n\t})\n\tif err != nil && !apierrors.IsNotFound(err) {\n\t\treturn err\n\t}\n\n\terr = DeletePodsByLabel(ns, map[string]string{\"job-name\": name})\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *cloudSchedulerRESTClient) DeleteJob(ctx context.Context, req *schedulerpb.DeleteJobRequest, opts ...gax.CallOption) error {\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v1/%v\", req.GetName())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"name\", url.QueryEscape(req.GetName()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\treturn gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"DELETE\", baseUrl.String(), nil)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\t// Returns nil if there is no error, otherwise wraps\n\t\t// the response code and body into a non-nil error\n\t\treturn googleapi.CheckResponse(httpRsp)\n\t}, opts...)\n}", "func (ctb *Crontab) DeleteJob(id string) {\n\tctb.control <- &crontabControl{opJobDel, id}\n}", "func DeleteJob(cmd CmdInterface) {\n\tjob, err := store.FindJob(cmd.Parts[1])\n\tif err != nil {\n\t\tReturnError(cmd, err.Error())\n\t\treturn\n\t}\n\terr = job.Delete()\n\tif err != nil {\n\t\tReturnError(cmd, err.Error())\n\t\treturn\n\t}\n\n\tReturnString(cmd, \"OK\")\n}", "func (d *Disque) Delete(jobID string) (err error) {\n\t_, err = d.call(\"DELJOB\", redis.Args{}.Add(jobID))\n\treturn\n}", "func (r *ProjectsLocationsJobsService) Delete(name string) *ProjectsLocationsJobsDeleteCall {\n\tc := &ProjectsLocationsJobsDeleteCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.name = name\n\treturn c\n}", "func (s *store) DeleteJob(name string) error {\n\tjobsClient := s.client.BatchV1().Jobs(s.namespace)\n\terr := jobsClient.Delete(name, &metav1.DeleteOptions{})\n\treturn err\n}", "func TestCrontabDelete(t *testing.T) {\n\tctb := NewCrontab()\n\tjob := createJob(t, \"delete\", true)\n\n\tctb.AddJob(job)\n\ttime.Sleep(10 * 1e9)\n\tctb.Stop()\n}", "func (c *JobClient) Delete() *JobDelete {\n\tmutation := newJobMutation(c.config, OpDelete)\n\treturn &JobDelete{config: c.config, hooks: c.Hooks(), mutation: mutation}\n}", "func (r *ProjectsLocationsMigrationJobsService) Delete(name string) *ProjectsLocationsMigrationJobsDeleteCall {\n\tc := &ProjectsLocationsMigrationJobsDeleteCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.name = name\n\treturn c\n}", "func (o *Job) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no Job provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), jobPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"jobs\\\" WHERE \\\"id\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete from jobs\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by delete for jobs\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn rowsAff, nil\n}", "func (s *service) DeleteJob(id string) error {\n\tfoundJob, err := s.repo.GetJob(id)\n\tif err != nil {\n\t\treturn httperror.NewNotFoundError(\"job not found\")\n\t}\n\ts.mu.Lock()\n\tjobCh, ok := s.jobWorkers[foundJob.ID]\n\ts.mu.Unlock()\n\tif ok {\n\t\tjobCh <- struct{}{}\n\t}\n\treturn s.repo.DeleteJob(foundJob)\n}", "func (c *Client) DeleteJob(jobID string) error {\n\tclient := &http.Client{}\n\turl := fmt.Sprintf(\"http://%s/api/v1/jobs/%s\", c.options.ServerAddr, jobID)\n\treq, err := http.NewRequest(\"DELETE\", url, nil)\n\tif err != nil {\n\t\tlogrus.WithError(err)\n\t\treturn err\n\t}\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\tlogrus.WithError(err)\n\t\treturn err\n\t}\n\n\tdefer resp.Body.Close()\n\n\t_, err = ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\tlogrus.WithError(err)\n\t}\n\treturn err\n}", "func (m *DataHistoryManager) DeleteJob(nickname, id string) error {\n\tif m == nil {\n\t\treturn ErrNilSubsystem\n\t}\n\tif atomic.LoadInt32(&m.started) == 0 {\n\t\treturn ErrSubSystemNotStarted\n\t}\n\tif nickname == \"\" && id == \"\" {\n\t\treturn errNicknameIDUnset\n\t}\n\tif nickname != \"\" && id != \"\" {\n\t\treturn errOnlyNicknameOrID\n\t}\n\tvar dbJob *datahistoryjob.DataHistoryJob\n\tvar err error\n\tm.m.Lock()\n\tdefer m.m.Unlock()\n\tfor i := range m.jobs {\n\t\tif strings.EqualFold(m.jobs[i].Nickname, nickname) ||\n\t\t\tm.jobs[i].ID.String() == id {\n\t\t\tdbJob = m.convertJobToDBModel(m.jobs[i])\n\t\t\tm.jobs = append(m.jobs[:i], m.jobs[i+1:]...)\n\t\t\tbreak\n\t\t}\n\t}\n\tif dbJob == nil {\n\t\tif nickname != \"\" {\n\t\t\tdbJob, err = m.jobDB.GetByNickName(nickname)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t} else {\n\t\t\tdbJob, err = m.jobDB.GetByID(id)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\tif dbJob.Status != int64(dataHistoryStatusActive) {\n\t\tstatus := dataHistoryStatus(dbJob.Status)\n\t\treturn fmt.Errorf(\"job: %v status: %s error: %w\", dbJob.Nickname, status, errCanOnlyDeleteActiveJobs)\n\t}\n\tdbJob.Status = int64(dataHistoryStatusRemoved)\n\terr = m.jobDB.Upsert(dbJob)\n\tif err != nil {\n\t\treturn err\n\t}\n\tlog.Infof(log.DataHistory, \"deleted job %v\", dbJob.Nickname)\n\treturn nil\n}", "func deleteTestJob(t *testing.T, db *mongo.Database, id string) {\n\toid, err := primitive.ObjectIDFromHex(id)\n\tif err != nil {\n\t\tt.Fatalf(\"failed to parse object ID: %v\", err)\n\t}\n\tm := bson.M{\"_id\": oid}\n\tif err := db.Collection(jobCollectionName).FindOneAndDelete(context.Background(), m).Err(); err != nil {\n\t\tt.Fatalf(\"failed to delete: %s\", err)\n\t}\n}", "func (sched *StdScheduler) DeleteJob(key int) error {\n\tsched.Lock()\n\tdefer sched.Unlock()\n\n\tfor i, item := range *sched.Queue {\n\t\tif item.Job.Key() == key {\n\t\t\tsched.Queue.Remove(i)\n\t\t\treturn nil\n\t\t}\n\t}\n\n\treturn errors.New(\"No Job with the given Key found\")\n}", "func (a *apiServer) DeleteJob(ctx context.Context, request *pps.DeleteJobRequest) (response *emptypb.Empty, retErr error) {\n\tif request.GetJob() == nil {\n\t\treturn nil, errors.New(\"job cannot be nil\")\n\t}\n\tensurePipelineProject(request.Job.Pipeline)\n\tif err := a.txnEnv.WithWriteContext(ctx, func(txnCtx *txncontext.TransactionContext) error {\n\t\treturn a.deleteJobInTransaction(txnCtx, request)\n\t}); err != nil {\n\t\treturn nil, err\n\t}\n\tclearJobCache(a.env.GetPachClient(ctx), ppsdb.JobKey(request.Job))\n\treturn &emptypb.Empty{}, nil\n}", "func (c *jobsRESTClient) DeleteJob(ctx context.Context, req *runpb.DeleteJobRequest, opts ...gax.CallOption) (*DeleteJobOperation, error) {\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v2/%v\", req.GetName())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\tif req.GetEtag() != \"\" {\n\t\tparams.Add(\"etag\", fmt.Sprintf(\"%v\", req.GetEtag()))\n\t}\n\tif req.GetValidateOnly() {\n\t\tparams.Add(\"validateOnly\", fmt.Sprintf(\"%v\", req.GetValidateOnly()))\n\t}\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"name\", url.QueryEscape(req.GetName()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tresp := &longrunningpb.Operation{}\n\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"DELETE\", baseUrl.String(), nil)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}, opts...)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\n\toverride := fmt.Sprintf(\"/v2/%s\", resp.GetName())\n\treturn &DeleteJobOperation{\n\t\tlro: longrunning.InternalNewOperation(*c.LROClient, resp),\n\t\tpollPath: override,\n\t}, nil\n}", "func Delete(name string) error {\n\treturn <-delete(name)\n}", "func (db jobDatabase) DeleteJob(job mgm.Job) error {\n\tcon, err := db.mysql.GetConnection()\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer con.Close()\n\n\t_, err = con.Exec(\"DELETE FROM jobs WHERE id=?\", job.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (t *DRMAATracker) DeleteJob(jobID string) error {\n\tt.Lock()\n\tdefer t.Unlock()\n\t// job needs to be in an end state\n\tps, err := t.session.JobPs(jobID)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif ps != drmaa.PsDone && ps != drmaa.PsFailed {\n\t\treturn fmt.Errorf(\"job is not in an end state (%v)\", ps)\n\t}\n\tt.store.RemoveJob(jobID)\n\treturn nil\n}", "func (dt *DockerTracker) DeleteJob(jobid string) error {\n\tif err := dt.check(); err != nil {\n\t\treturn err\n\t}\n\tif state := dt.JobState(jobid); state != drmaa2interface.Done && state != drmaa2interface.Failed {\n\t\treturn errors.New(\"job is not in an end-state\")\n\t}\n\treturn dt.cli.ContainerRemove(context.Background(),\n\t\tjobid,\n\t\ttypes.ContainerRemoveOptions{\n\t\t\tForce: true,\n\t\t\tRemoveLinks: false,\n\t\t\tRemoveVolumes: true,\n\t\t},\n\t)\n}", "func (op *DeleteJobOperation) Name() string {\n\treturn op.lro.Name()\n}", "func (r *OptimJob) ValidateDelete() error {\n\toptimjoblog.Info(\"validate delete\", \"name\", r.Name)\n\n\t// TODO(user): fill in your validation logic upon object deletion.\n\treturn nil\n}", "func (d DB) Delete(id string) error {\n\tquery := fmt.Sprintf(`delete from %v where job ->> 'id' = $1;`, TABLE_NAME)\n\t_, err := d.conn.Exec(query, id)\n\treturn err\n}", "func (c *cronFederatedHPAs) Delete(ctx context.Context, name string, opts v1.DeleteOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"cronfederatedhpas\").\n\t\tName(name).\n\t\tBody(&opts).\n\t\tDo(ctx).\n\t\tError()\n}", "func (c *controller) DeleteJob(namespace, jobName string) error {\n\treturn c.k8sBatchClient.Jobs(namespace).Delete(jobName, &metav1.DeleteOptions{})\n}", "func Delete(name string) {\n\tkt.Remove(name)\n}", "func (regionEnv *RegionEnv) deleteJob(job *unstructured.Unstructured, wait bool) error {\n\tgvk := schema.GroupVersionKind{Group: \"batch\", Version: \"v1\", Kind: \"Job\"}\n\tregionEnv.Logger.Infof(\"Deleting Job %s\", job.GetName())\n\tmapping, _ := regionEnv.Mapper.RESTMapping(gvk.GroupKind(), gvk.Version)\n\tdynamicInterface := regionEnv.DynamicClient.Resource(mapping.Resource).Namespace(regionEnv.Cfg.Namespace)\n\n\tpropagation := metav1.DeletePropagationForeground\n\tdeleteOptions := metav1.DeleteOptions{\n\t\tGracePeriodSeconds: int64Ref(0),\n\t\tPropagationPolicy: &propagation,\n\t}\n\n\tif wait {\n\t\tlistOptions := metav1.ListOptions{\n\t\t\tFieldSelector: fmt.Sprintf(\"metadata.name=%s\", job.GetName()),\n\t\t\tTimeoutSeconds: int64Ref(30),\n\t\t}\n\t\twatcher, err := dynamicInterface.Watch(regionEnv.Context, listOptions)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer watcher.Stop()\n\n\t\twatchCh := watcher.ResultChan()\n\t\terr = dynamicInterface.Delete(regionEnv.Context, job.GetName(), deleteOptions)\n\t\tif err != nil {\n\t\t\tif errors.IsNotFound(err) {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\treturn err\n\t\t}\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-regionEnv.Context.Done():\n\t\t\t\treturn regionEnv.Context.Err()\n\n\t\t\tcase event, ok := <-watchCh:\n\t\t\t\tif !ok {\n\t\t\t\t\treturn fmt.Errorf(\"timeout deleting job\")\n\t\t\t\t}\n\n\t\t\t\tif event.Type == watch.Deleted {\n\t\t\t\t\treturn nil\n\t\t\t\t}\n\t\t\t\tif event.Type == watch.Error {\n\t\t\t\t\tregionEnv.Logger.Warnf(\"error reported during job deletion: %v\", event.Object)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t} else {\n\t\terr := dynamicInterface.Delete(regionEnv.Context, job.GetName(), deleteOptions)\n\t\tif err != nil {\n\t\t\tif errors.IsNotFound(err) {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (js *JobService) DeleteJob(id uint) (*entity.Job, []error) {\n\tusr, errs := js.jobRepo.DeleteJob(id)\n\n\tif len(errs) > 0 {\n\t\treturn nil, errs\n\t}\n\n\treturn usr, nil\n}", "func (b *Bucket) Delete(ctx context.Context, name string) error {\n\treturn b.bkt.Object(name).Delete(ctx)\n}", "func Delete(client *golangsdk.ServiceClient, opts DeleteOpts) (r JobResult) {\n\treqBody, err := opts.ToServerDeleteMap()\n\tif err != nil {\n\t\tr.Err = err\n\t\treturn\n\t}\n\t_, r.Err = client.Post(deleteURL(client), reqBody, &r.Body, &golangsdk.RequestOpts{OkCodes: []int{200}})\n\treturn\n}", "func (r DeleteJobRequest) Send(ctx context.Context) (*DeleteJobOutput, error) {\n\tr.Request.SetContext(ctx)\n\terr := r.Request.Send()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn r.Request.Data.(*DeleteJobOutput), nil\n}", "func deleteGCPJob(context *cli.Context) {\n\tconfig := getConfig(context)\n\tgcp := getGCP(config)\n\n\terr := gcp.DeleteJob(context.String(\"job-id\"))\n\tif err != nil {\n\t\tfmt.Printf(\"Failed to delete GCP job %s: %s\\n\", context.String(\"job-id\"), err)\n\t} else {\n\t\tfmt.Printf(\"Deleted GCP job %s\\n\", context.String(\"job-id\"))\n\t}\n}", "func (c *jxTasks) Delete(name string, options *v1.DeleteOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"jxtasks\").\n\t\tName(name).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *jobsRESTClient) DeleteJobOperation(name string) *DeleteJobOperation {\n\toverride := fmt.Sprintf(\"/v2/%s\", name)\n\treturn &DeleteJobOperation{\n\t\tlro: longrunning.InternalNewOperation(*c.LROClient, &longrunningpb.Operation{Name: name}),\n\t\tpollPath: override,\n\t}\n}", "func removeJob(id string, port string) {\n\tresponse := makePutRequest(fmt.Sprintf(\"http://localhost%s/jobs/delete\", port), bytes.NewBuffer([]byte(id+\" \"+fmt.Sprintf(\"%d\", os.Getuid()))))\n\tfmt.Println(response)\n}", "func (r *Job) ValidateDelete() error {\n\treturn nil\n}", "func (c *JobClient) DeleteOne(j *Job) *JobDeleteOne {\n\treturn c.DeleteOneID(j.ID)\n}", "func DeleteJobTask(w http.ResponseWriter, r *http.Request) {\n\tresponse := services.DeleteJobTask(r)\n\n\trender.Status(r, response.Code)\n\trender.JSON(w, r, response)\n}", "func (cli *Client) CronDelete(ctx context.Context, id string) error {\n\tv := url.Values{}\n\tresp, err := cli.delete(ctx, \"/crons/\"+id, v, nil)\n\tensureReaderClosed(resp)\n\treturn err\n}", "func (c *jobsGRPCClient) DeleteJobOperation(name string) *DeleteJobOperation {\n\treturn &DeleteJobOperation{\n\t\tlro: longrunning.InternalNewOperation(*c.LROClient, &longrunningpb.Operation{Name: name}),\n\t}\n}", "func (c *Cron) JobRemove(expr string, cmd func()) error {\n\tc.statusLock.RLock()\n\tdefer c.statusLock.RUnlock()\n\n\tcmdName, err := obtainFuncName(cmd)\n\tif err != nil {\n\t\treturn errors.New(\"invalid cmd: \" + err.Error())\n\t}\n\n\tswitch c.status {\n\tcase statusInitial:\n\t\tc.jobListLock.Lock()\n\t\tfor item := c.jobList.Front(); item != c.jobList.Back(); item = item.Next() {\n\t\t\tj := item.Value.(*job)\n\t\t\tif j.expr == expr && j.cmdName == cmdName {\n\t\t\t\t_ = c.jobList.Remove(item)\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tc.jobListLock.Unlock()\n\t\treturn nil\n\tcase statusRunning:\n\t\tj, err := obtainJob(expr, cmdName, cmd)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tc.jobRecordLock.Lock()\n\t\tdefer c.jobRecordLock.Unlock()\n\t\tj.updateNextTime(time.Now())\n\t\tvar jobList *list.List\n\t\tjobNodeList := c.jobRecord.Find(j.nextTime)\n\t\tswitch len(jobNodeList) {\n\t\tcase 0:\n\t\t\treturn nil\n\t\tcase 1:\n\t\t\tjobList = jobNodeList[0].V.(*list.List)\n\t\t\tvar deleteJob *list.Element\n\t\t\tfor item := jobList.Front(); item != nil; item = item.Next() {\n\t\t\t\tif item.Value.(*job).cmdName == cmdName {\n\t\t\t\t\tdeleteJob = item\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\tif deleteJob != nil {\n\t\t\t\tjobList.Remove(deleteJob)\n\t\t\t\tif jobList.Len() <= 0 {\n\t\t\t\t\tc.jobRecord.DeleteByKey(j.nextTime)\n\t\t\t\t}\n\t\t\t}\n\t\tdefault:\n\t\t\treturn errors.New(\"jobRecord error\")\n\t\t}\n\t\treturn nil\n\tdefault:\n\t\treturn errors.New(\"invalid status\")\n\t}\n}", "func (y *Yeelight) CronDel(t string) string {\n\tcmd := `{\"id\":14,\"method\":\"cron_del\",\"params\":[` + t + `]}`\n\treturn y.request(cmd)\n}", "func (c *FakeImagePullJobs) Delete(ctx context.Context, name string, opts v1.DeleteOptions) error {\n\t_, err := c.Fake.\n\t\tInvokes(testing.NewDeleteAction(imagepulljobsResource, c.ns, name), &v1alpha1.ImagePullJob{})\n\n\treturn err\n}", "func (s *JobDB) DeleteJob(jobToDelete BusInfoJob) {\n\tchatID := jobToDelete.ChatID\n\n\tuserKey := []byte(strconv.FormatInt(chatID, 10))\n\n\tdb, err := bolt.Open(s.dbFile, 0600, nil)\n\tif err != nil {\n\t\tlog.Fatalln(err)\n\t}\n\tdefer db.Close()\n\n\terr = db.Update(func(tx *bolt.Tx) error {\n\n\t\tb := tx.Bucket([]byte(s.userBucket))\n\t\tif b == nil {\n\t\t\treturn nil\n\t\t}\n\n\t\tv := b.Get(userKey)\n\t\tstoredJobs := []BusInfoJob{}\n\t\tjson.Unmarshal(v, &storedJobs)\n\n\t\t// Remove job and store the remaining back to the key\n\t\tremainingJobs := storedJobs[:0]\n\t\tfor _, job := range storedJobs {\n\t\t\tif job != jobToDelete {\n\t\t\t\tremainingJobs = append(remainingJobs, job)\n\t\t\t}\n\t\t}\n\t\tencRemainingJobs, err := json.Marshal(remainingJobs)\n\t\tif err != nil {\n\t\t\tlog.Fatalln(err)\n\t\t}\n\t\tb.Put(userKey, encRemainingJobs)\n\n\t\t// Check and remove from the other Job bucket if ChatID has no jobs for that day anymore\n\t\tremovedJobDay := jobToDelete.Weekday\n\t\tremainingJobsForDay := s.getJobsByChatIDandDay(chatID, removedJobDay, tx)\n\n\t\tif len(remainingJobsForDay) == 0 {\n\t\t\ts.deleteChatIDFromDayLookup(chatID, removedJobDay, tx)\n\t\t}\n\n\t\treturn nil\n\t})\n}", "func (c *peer) Delete(name string) error {\n\tkey := path.Join(c.prefix, name)\n\treturn c.store.Del(key)\n}", "func (o *orm) DeleteJob(ctx context.Context, id int32) error {\n\terr := o.db.Exec(`\n\t\tWITH deleted_jobs AS (\n\t\t\tDELETE FROM jobs WHERE id = ? RETURNING\n\t\t\t\tpipeline_spec_id,\n\t\t\t\toffchainreporting_oracle_spec_id,\n\t\t\t\tkeeper_spec_id,\n\t\t\t\tcron_spec_id,\n\t\t\t\tflux_monitor_spec_id,\n\t\t\t\tvrf_spec_id,\n\t\t\t\twebhook_spec_id,\n\t\t\t\tdirect_request_spec_id\n\t\t),\n\t\tdeleted_oracle_specs AS (\n\t\t\tDELETE FROM offchainreporting_oracle_specs WHERE id IN (SELECT offchainreporting_oracle_spec_id FROM deleted_jobs)\n\t\t),\n\t\tdeleted_keeper_specs AS (\n\t\t\tDELETE FROM keeper_specs WHERE id IN (SELECT keeper_spec_id FROM deleted_jobs)\n\t\t),\n\t\tdeleted_cron_specs AS (\n\t\t\tDELETE FROM cron_specs WHERE id IN (SELECT cron_spec_id FROM deleted_jobs)\n\t\t),\n\t\tdeleted_fm_specs AS (\n\t\t\tDELETE FROM flux_monitor_specs WHERE id IN (SELECT flux_monitor_spec_id FROM deleted_jobs)\n\t\t),\n\t\tdeleted_vrf_specs AS (\n\t\t\tDELETE FROM vrf_specs WHERE id IN (SELECT vrf_spec_id FROM deleted_jobs)\n\t\t),\n\t\tdeleted_webhook_specs AS (\n\t\t\tDELETE FROM webhook_specs WHERE id IN (SELECT webhook_spec_id FROM deleted_jobs)\n\t\t),\n\t\tdeleted_dr_specs AS (\n\t\t\tDELETE FROM direct_request_specs WHERE id IN (SELECT direct_request_spec_id FROM deleted_jobs)\n\t\t)\n\t\tDELETE FROM pipeline_specs WHERE id IN (SELECT pipeline_spec_id FROM deleted_jobs)\n\t`, id).Error\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"DeleteJob failed to delete job\")\n\t}\n\n\treturn nil\n}", "func (d *Delegate) BeforeJobDeleted(spec job.Job) {}", "func (j *TrainingJob) delete() {\n\tj.gc.CollectJob(j.job.Metadata.Name, garbagecollection.NullUID)\n}", "func (c *nodes) Delete(name string) error {\n\treturn c.r.Delete().Resource(c.resourceName()).Name(name).Do().Error()\n}", "func Delete(name string) error {\n\tprofile := &performancev2.PerformanceProfile{}\n\tif err := testclient.Client.Get(context.TODO(), types.NamespacedName{Name: name}, profile); err != nil {\n\t\tif errors.IsNotFound(err) {\n\t\t\treturn nil\n\t\t}\n\t\treturn err\n\t}\n\n\tif err := testclient.Client.Delete(context.TODO(), profile); err != nil {\n\t\treturn err\n\t}\n\tkey := client.ObjectKey{\n\t\tName: name,\n\t}\n\treturn WaitForDeletion(key, 2*time.Minute)\n}", "func (a *Agent) Kill(checkID string) error {\n\tjob, err := a.storage.Get(checkID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tcmd := exec.Command(\"kubectl\", \"delete\", \"pod\", \"--grace-period\", \"0\", job.CheckID)\n\n\treturn cmd.Run()\n}", "func (dt *Tracker) DeleteJob(jobid string) error {\n\treturn dt.processTracker.DeleteJob(jobid)\n}", "func (schematics *SchematicsV1) DeleteJob(deleteJobOptions *DeleteJobOptions) (response *core.DetailedResponse, err error) {\n\treturn schematics.DeleteJobWithContext(context.Background(), deleteJobOptions)\n}", "func (b *Bucket) Delete(_ context.Context, name string) error {\n\treturn b.client.DeleteObject(b.name, name)\n}", "func (s *grpcServer) DeleteJobPost(ctx context.Context, req *pb.DeleteJobPostRequest) (*pb.DeleteJobPostResponse, error) {\n\t_, rep, err := s.deleteJobPost.ServeGRPC(ctx, req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn rep.(*pb.DeleteJobPostResponse), nil\n}", "func (c APIClient) DeleteMonitoringJob(id string) error {\n\treturn c.doHTTPDelete(fmt.Sprintf(\"https://api.nsone.net/v1/monitoring/jobs/%s\", id))\n}", "func (c *CloudSchedulerClient) DeleteJob(ctx context.Context, req *schedulerpb.DeleteJobRequest, opts ...gax.CallOption) error {\n\treturn c.internalClient.DeleteJob(ctx, req, opts...)\n}", "func (si *SyncIndexJob) Delete(enable bool) {\n\tsi.SyncJob.Delete(enable)\n\tsi.IndexJob.Delete(enable)\n}", "func DeleteJob(cmd *cobra.Command, args []string) error {\n\n\tclient, err := auth.GetClient()\n\tif err != nil {\n\t\treturn err\n\t}\n\t// Parse all flags\n\n\tvar earliestDefault string\n\tearliest := &earliestDefault\n\terr = flags.ParseFlag(cmd.Flags(), \"earliest\", &earliest)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"earliest\": ` + err.Error())\n\t}\n\tvar extractFieldsDefault string\n\textractFields := &extractFieldsDefault\n\terr = flags.ParseFlag(cmd.Flags(), \"extract-fields\", &extractFields)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"extract-fields\": ` + err.Error())\n\t}\n\tvar index string\n\terr = flags.ParseFlag(cmd.Flags(), \"index\", &index)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"index\": ` + err.Error())\n\t}\n\tvar latestDefault string\n\tlatest := &latestDefault\n\terr = flags.ParseFlag(cmd.Flags(), \"latest\", &latest)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"latest\": ` + err.Error())\n\t}\n\tvar maxTimeDefault int32\n\tmaxTime := &maxTimeDefault\n\terr = flags.ParseFlag(cmd.Flags(), \"max-time\", &maxTime)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"max-time\": ` + err.Error())\n\t}\n\tvar messages []model.Message\n\terr = flags.ParseFlag(cmd.Flags(), \"messages\", &messages)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"messages\": ` + err.Error())\n\t}\n\tvar module string\n\terr = flags.ParseFlag(cmd.Flags(), \"module\", &module)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"module\": ` + err.Error())\n\t}\n\tvar predicate string\n\terr = flags.ParseFlag(cmd.Flags(), \"predicate\", &predicate)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"predicate\": ` + err.Error())\n\t}\n\tvar relativeTimeAnchorDefault string\n\trelativeTimeAnchor := &relativeTimeAnchorDefault\n\terr = flags.ParseFlag(cmd.Flags(), \"relative-time-anchor\", &relativeTimeAnchor)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"relative-time-anchor\": ` + err.Error())\n\t}\n\tvar statusDefault model.SearchStatus\n\tstatus := &statusDefault\n\terr = flags.ParseFlag(cmd.Flags(), \"status\", &status)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"status\": ` + err.Error())\n\t}\n\tvar timezone interface{}\n\terr = flags.ParseFlag(cmd.Flags(), \"timezone\", &timezone)\n\tif err != nil {\n\t\treturn fmt.Errorf(`error parsing \"timezone\": ` + err.Error())\n\t}\n\t// Form the request body\n\tgenerated_request_body := model.DeleteSearchJob{\n\n\t\tExtractFields: extractFields,\n\t\tIndex: index,\n\t\tMaxTime: maxTime,\n\t\tMessages: messages,\n\t\tModule: module,\n\t\tPredicate: predicate,\n\t\tQueryParameters: &model.QueryParameters{\n\t\t\tEarliest: earliest,\n\t\t\tLatest: latest,\n\t\t\tRelativeTimeAnchor: relativeTimeAnchor,\n\t\t\tTimezone: timezone,\n\t\t},\n\t\tStatus: status,\n\t}\n\n\t// Silence Usage\n\tcmd.SilenceUsage = true\n\n\tresp, err := client.SearchService.DeleteJob(generated_request_body)\n\tif err != nil {\n\t\treturn err\n\t}\n\tjsonx.Pprint(cmd, resp)\n\treturn nil\n}", "func (m *PrintersItemJobsPrintJobItemRequestBuilder) Delete(ctx context.Context, requestConfiguration *PrintersItemJobsPrintJobItemRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": ia572726a95efa92ddd544552cd950653dc691023836923576b2f4bf716cf204a.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": ia572726a95efa92ddd544552cd950653dc691023836923576b2f4bf716cf204a.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func (b *Bucket) Delete(_ context.Context, name string) error {\n\tdelete(b.objects, name)\n\treturn nil\n}", "func (d Delegate) BeforeJobDeleted(spec job.Job) {\n}", "func (s *server) Delete(ctx context.Context, body *pb.NameHolder) (*pb.DeletionResponse, error) {\n\tappName := body.GetName()\n\tfilter := types.M{\n\t\tmongo.NameKey: appName,\n\t\tmongo.InstanceTypeKey: mongo.AppInstance,\n\t}\n\n\tnode, _ := redis.FetchAppNode(appName)\n\tgo redis.DecrementServiceLoad(ServiceName, node)\n\tgo redis.RemoveApp(appName)\n\tgo diskCleanup(appName)\n\n\tif configs.CloudflareConfig.PlugIn {\n\t\tgo cloudflare.DeleteRecord(appName, mongo.AppInstance)\n\t}\n\n\t_, err := mongo.DeleteInstance(filter)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &pb.DeletionResponse{Success: true}, nil\n}", "func (task SchTask) Delete(taskname string, own, force bool) string {\n\tcmd := &exec.Cmd{}\n\n\tif Debug {\n\t\treturn dbgMessage\n\t}\n\n\tif own {\n\t\ttaskname = task.prefix + taskname\n\t}\n\n\tif !force {\n\t\tcmd = exec.Command(task.bin, _Delete.Command, _Delete.taskname, taskname)\n\t} else {\n\t\tcmd = exec.Command(task.bin, _Delete.Command, _Delete.taskname, taskname, _Delete.force)\n\t}\n\n\toutput, err := cmd.CombinedOutput()\n\tcatch(output, err)\n\n\treturn string(output)\n}", "func (c *kaosRules) Delete(name string, options *meta_v1.DeleteOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"kaosrules\").\n\t\tName(name).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (k *kubernetes) Delete(s *runtime.Service) error {\n\tk.Lock()\n\tdefer k.Unlock()\n\n\t// create new kubernetes micro service\n\tservice := newService(s, runtime.CreateOptions{\n\t\tType: k.options.Type,\n\t})\n\n\tlog.Debugf(\"Runtime queueing service %s for delete action\", service.Name)\n\n\t// queue service for removal\n\tk.queue <- &task{\n\t\taction: stop,\n\t\tservice: service,\n\t}\n\n\treturn nil\n}", "func (r *ResourceGroup) Delete(t *testing.T, name string) {\n\tr.logger.Printf(\"ResourceGroup.Delete: deleting %q\", name)\n\n\tr.client.Delete(name, r.ctx.Done())\n\n\tresgroup, err := r.client.Get(name)\n\tif err != nil {\n\t\tr.logger.Printf(\"ResourceGroup.Delete finished\")\n\t\treturn\n\t}\n\tr.logger.Printf(\"ResourceGroup.Delete: still exists, checking if deprovisioning\")\n\tif resgroup.Properties == nil {\n\t\tt.Fatal(\"ResourceGroup.Delete: resgroup does not have properties\")\n\t}\n\tif resgroup.Properties.ProvisioningState == nil {\n\t\tt.Fatal(\"ResourceGroup.Delete: resgroup does not have ProvisioningState\")\n\t}\n\tprovisioningState := *resgroup.Properties.ProvisioningState\n\texpectedState := \"Deleting\"\n\tif *resgroup.Properties.ProvisioningState != expectedState {\n\t\tt.Fatalf(\"ResourceGroup.Delete: expected state[%s] got [%s]\", expectedState, provisioningState)\n\t}\n\tr.logger.Printf(\"ResourceGroup.Delete: resgroup is deprovisioning, should be ok\")\n}", "func (c *DeleteCommand) CommandName() string {\n\treturn \"delete\"\n}", "func (c *Client) DeleteJob(jobID string, purge bool) error {\n\tparams := url.Values{}\n\tparams.Set(\"purge\", strconv.FormatBool(purge))\n\treturn c.doHTTPReq(http.MethodDelete,\n\t\tfmt.Sprintf(uriDeleteJob, jobID), params, nil, nil)\n}", "func (f *DMAClient) Delete(name string, options *meta_v1.DeleteOptions) error {\n\treturn f.client.Delete().\n\t\tNamespace(f.namespace).Resource(f.plural).\n\t\tName(name).Body(options).Do(context.Background()).\n\t\tError()\n}", "func (r DeleteReplicationJobRequest) Send(ctx context.Context) (*DeleteReplicationJobOutput, error) {\n\tr.Request.SetContext(ctx)\n\terr := r.Request.Send()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn r.Request.Data.(*DeleteReplicationJobOutput), nil\n}", "func (c *JobpositionClient) Delete() *JobpositionDelete {\n\tmutation := newJobpositionMutation(c.config, OpDelete)\n\treturn &JobpositionDelete{config: c.config, hooks: c.Hooks(), mutation: mutation}\n}", "func (r *ProjectsInstancesTablesService) Delete(name string) *ProjectsInstancesTablesDeleteCall {\n\tc := &ProjectsInstancesTablesDeleteCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.name = name\n\treturn c\n}", "func (client JobClient) DeleteSender(req *http.Request) (future JobDeleteFuture, err error) {\n var resp *http.Response\n resp, err = client.Send(req, azure.DoRetryWithRegistration(client.Client))\n if err != nil {\n return\n }\n future.Future, err = azure.NewFutureFromResponse(resp)\n return\n }", "func Delete(db gorp.SqlExecutor, id string) error {\n\taccessTokenID, err := db.SelectNullStr(\"SELECT auth_consumer_id FROM worker WHERE id = $1\", id)\n\tif err != nil {\n\t\treturn sdk.WithStack(err)\n\t}\n\tquery := `DELETE FROM worker WHERE id = $1`\n\tif _, err := db.Exec(query, id); err != nil {\n\t\treturn sdk.WithStack(err)\n\t}\n\n\tif accessTokenID.Valid {\n\t\tif err := authentication.DeleteConsumerByID(db, accessTokenID.String); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (s *gcBlobTaskStore) Delete(ctx context.Context, b *models.GCBlobTask) error {\n\tdefer metrics.InstrumentQuery(\"gc_blob_task_delete\")()\n\n\tq := \"DELETE FROM gc_blob_review_queue WHERE digest = decode($1, 'hex')\"\n\tdgst, err := NewDigest(b.Digest)\n\tif err != nil {\n\t\treturn err\n\t}\n\tres, err := s.db.ExecContext(ctx, q, dgst)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"deleting GC blob task: %w\", err)\n\t}\n\tcount, err := res.RowsAffected()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"deleting GC blob task: %w\", err)\n\t}\n\tif count == 0 {\n\t\treturn fmt.Errorf(\"GC blob task not found\")\n\t}\n\n\treturn nil\n}", "func (c *SubresourceClient) Delete(namespace, name string) (e error) {\n\tif c.Error != \"\" {\n\t\te = fmt.Errorf(c.Error)\n\t}\n\treturn\n}", "func (c *client) Delete(ctx context.Context, name string) error {\n\tlcn, err := c.Get(ctx, name)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif len(*lcn) == 0 {\n\t\treturn fmt.Errorf(\"Location [%s] not found\", name)\n\t}\n\n\trequest, err := c.getLocationRequest(wssdcloudcommon.Operation_DELETE, name, &(*lcn)[0])\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = c.LocationAgentClient.Invoke(ctx, request)\n\n\treturn err\n}", "func KubeDelete(object string, name string) string {\n\tvar outputstring string\n\tif object == \"\" || name == \"\" {\n\t\toutputstring = \"\"\n\t} else {\n\t\toutputstring = fmt.Sprintf(\"delete %s %s\", object, name)\n\t}\n\treturn KubeCommand(outputstring)\n}", "func (c *MockResourceGroupsClient) Delete(ctx context.Context, name string) error {\n\tif _, ok := c.RGs[name]; !ok {\n\t\treturn fmt.Errorf(\"%s does not exist\", name)\n\t}\n\tdelete(c.RGs, name)\n\treturn nil\n}", "func (jz *Jzon) Delete(k string) (err error) {\n\tif jz.Type != JzTypeObj {\n\t\treturn expectTypeOf(JzTypeObj, jz.Type)\n\t}\n\n\tdelete(jz.data.(map[string]*Jzon), k)\n\treturn nil\n}", "func (c *Controller) Delete(ctx context.Context, name string) error {\n\tregistry, err := c.Get(ctx, name)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tcID := registry.Status.ContainerID\n\tif cID == \"\" {\n\t\treturn fmt.Errorf(\"container not running registry: %s\", name)\n\t}\n\n\treturn c.dockerClient.ContainerRemove(ctx, registry.Status.ContainerID, types.ContainerRemoveOptions{\n\t\tForce: true,\n\t})\n}", "func (c *CleanernameClient) Delete() *CleanernameDelete {\n\tmutation := newCleanernameMutation(c.config, OpDelete)\n\treturn &CleanernameDelete{config: c.config, hooks: c.Hooks(), mutation: mutation}\n}", "func Delete(c *gophercloud.ServiceClient, archivePolicyName string) (r DeleteResult) {\n\trequestOpts := &gophercloud.RequestOpts{\n\t\tMoreHeaders: map[string]string{\n\t\t\t\"Accept\": \"application/json, */*\",\n\t\t},\n\t}\n\t_, r.Err = c.Delete(deleteURL(c, archivePolicyName), requestOpts)\n\treturn\n}" ]
[ "0.6969676", "0.6850797", "0.677984", "0.65925467", "0.6563074", "0.65428567", "0.638948", "0.63471174", "0.6333265", "0.6281954", "0.62698877", "0.62488395", "0.62485254", "0.62442154", "0.6225446", "0.61785704", "0.61037046", "0.6042961", "0.60361505", "0.60282016", "0.59873086", "0.5949002", "0.5934434", "0.5932314", "0.5894964", "0.58783036", "0.58406067", "0.5779367", "0.5755479", "0.5736274", "0.5719671", "0.5705899", "0.56882274", "0.5679226", "0.56687623", "0.5624857", "0.5621487", "0.5617393", "0.55879134", "0.557537", "0.55618465", "0.55091494", "0.550154", "0.5498929", "0.5495197", "0.5485994", "0.54811394", "0.5474711", "0.54561615", "0.54434633", "0.54430425", "0.5439657", "0.5434705", "0.5432411", "0.53987414", "0.5394799", "0.53817904", "0.53451407", "0.5344321", "0.5330046", "0.53005487", "0.5294891", "0.5261934", "0.5251391", "0.5245323", "0.52448267", "0.5231152", "0.5226026", "0.5223468", "0.5202288", "0.5186228", "0.5141925", "0.51389885", "0.51345736", "0.51341784", "0.51315355", "0.5130306", "0.51234955", "0.51224023", "0.5110456", "0.510204", "0.50928766", "0.5088419", "0.5076905", "0.5070402", "0.5069371", "0.50679624", "0.5066939", "0.50543016", "0.5048702", "0.50379133", "0.50320035", "0.5017373", "0.5015411", "0.5015244", "0.50123364", "0.50107414", "0.50018865", "0.49977577", "0.49930578" ]
0.6458323
6
DeleteCollection deletes a collection of objects.
func (c *FakeTZCronJobs) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error { action := testing.NewDeleteCollectionAction(tzcronjobsResource, c.ns, listOptions) _, err := c.Fake.Invokes(action, &v1alpha1.TZCronJobList{}) return err }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *kongs) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"kongs\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *stewards) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOptions.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"stewards\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *sandboxes) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOpts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tResource(\"sandboxes\").\n\t\tVersionedParams(&listOpts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(&opts).\n\t\tDo(ctx).\n\t\tError()\n}", "func (c *inMemoryProviders) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOptions.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"inmemoryproviders\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *awsMediaStoreContainers) DeleteCollection(options *meta_v1.DeleteOptions, listOptions meta_v1.ListOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"awsmediastorecontainers\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *demos) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOptions.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"demos\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *rpcServices) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"rpcservices\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (w *ClusterDynamicClient) DeleteCollection(options metav1.DeleteOptions, listOptions metav1.ListOptions) error {\n\treturn w.dClient.Resource(w.resource).Namespace(w.namespace).DeleteCollection(w.ctx, options, listOptions)\n}", "func (c *acceptableUsePolicies) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOptions.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"acceptableusepolicies\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *FakeRabbitmqs) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(rabbitmqsResource, c.ns, listOptions)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.RabbitmqList{})\n\treturn err\n}", "func (c *interacts) DeleteCollection(options *meta_v1.DeleteOptions, listOptions meta_v1.ListOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"interacts\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *awsAutoscalingPolicies) DeleteCollection(options *meta_v1.DeleteOptions, listOptions meta_v1.ListOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"awsautoscalingpolicies\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *cronFederatedHPAs) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOpts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"cronfederatedhpas\").\n\t\tVersionedParams(&listOpts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(&opts).\n\t\tDo(ctx).\n\t\tError()\n}", "func (c *cloudFormationTemplates) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"cloudformationtemplates\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *staticFabricNetworkAttachments) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOpts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"staticfabricnetworkattachments\").\n\t\tVersionedParams(&listOpts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(&opts).\n\t\tDo(ctx).\n\t\tError()\n}", "func (c *meshPolicies) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOptions.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tResource(\"meshpolicies\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *kaosRules) DeleteCollection(options *meta_v1.DeleteOptions, listOptions meta_v1.ListOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"kaosrules\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *quarksStatefulSets) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOpts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"quarksstatefulsets\").\n\t\tVersionedParams(&listOpts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(&opts).\n\t\tDo(ctx).\n\t\tError()\n}", "func (c *snapshotRules) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOpts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"snapshotrules\").\n\t\tVersionedParams(&listOpts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(&opts).\n\t\tDo(ctx).\n\t\tError()\n}", "func (c *aITrainingJobs) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOptions.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"aitrainingjobs\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *distccClientClaims) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOptions.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"distccclientclaims\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *awsServiceDiscoveryServices) DeleteCollection(options *meta_v1.DeleteOptions, listOptions meta_v1.ListOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"awsservicediscoveryservices\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func deleteCollection(c client.Client, collection *kabanerov1alpha1.Collection, reqLogger logr.Logger) error {\n\t// Clean up existing collection assets.\n\terr := cleanupAssets(context.TODO(), collection, c, reqLogger)\n\tif err != nil {\n\t\treqLogger.Error(err, \"Error during cleanup processing.\")\n\t\treturn err\n\t}\n\n\t// Remove the finalizer entry from the instance.\n\terr = removeCollectionFinalizer(context.TODO(), collection, c, reqLogger)\n\tif err != nil {\n\t\treqLogger.Error(err, \"Error while attempting to remove the finalizer.\")\n\t\treturn err\n\t}\n\n\t// Delete the collection.\n\terr = c.Delete(context.TODO(), collection)\n\tif err != nil {\n\t\treqLogger.Error(err, \"Error while attempting to remove the finalizer.\")\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *FakeRobots) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(robotsResource, c.ns, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.RobotList{})\n\treturn err\n}", "func (c *sonicwallNetworkPolicies) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOpts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"sonicwallnetworkpolicies\").\n\t\tVersionedParams(&listOpts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(&opts).\n\t\tDo(ctx).\n\t\tError()\n}", "func (c *FakeIotDpses) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(iotdpsesResource, c.ns, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.IotDpsList{})\n\treturn err\n}", "func (c *concurrencyControls) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOpts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"concurrencycontrols\").\n\t\tVersionedParams(&listOpts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(&opts).\n\t\tDo(ctx).\n\t\tError()\n}", "func (c *gitTracks) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOptions.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"gittracks\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *FakeIngressBackends) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(ingressbackendsResource, c.ns, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.IngressBackendList{})\n\treturn err\n}", "func (c *advancedDeployments) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOpts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"advanceddeployments\").\n\t\tVersionedParams(&listOpts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(&opts).\n\t\tDo(ctx).\n\t\tError()\n}", "func (c *awsOrganizationsPolicyAttachments) DeleteCollection(options *meta_v1.DeleteOptions, listOptions meta_v1.ListOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"awsorganizationspolicyattachments\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *FakeQuobyteServices) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(quobyteservicesResource, c.ns, listOptions)\n\n\t_, err := c.Fake.Invokes(action, &quobyte_com_v1.QuobyteServiceList{})\n\treturn err\n}", "func (c *Collection) Delete() error {\n\tif c.Name == \"ALL_MEDIA_AUTO_COLLECTION\" {\n\t\treturn ErrAllSaved\n\t}\n\tinsta := c.insta\n\n\tdata, err := json.Marshal(\n\t\tmap[string]string{\n\t\t\t\"_uid\": toString(insta.Account.ID),\n\t\t\t\"_uuid\": insta.uuid,\n\t\t},\n\t)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, _, err = insta.sendRequest(\n\t\t&reqOptions{\n\t\t\tEndpoint: fmt.Sprintf(urlCollectionDelete, c.ID),\n\t\t\tIsPost: true,\n\t\t\tQuery: generateSignature(data),\n\t\t},\n\t)\n\treturn err\n}", "func (c *volumeSnapshotSchedules) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOpts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"volumesnapshotschedules\").\n\t\tVersionedParams(&listOpts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(&opts).\n\t\tDo(ctx).\n\t\tError()\n}", "func (c *horizontalPodAutoscalers) DeleteCollection(ctx context.Context, options *metav1.DeleteOptions, listOptions metav1.ListOptions) error {\n\tns := c.getNs(c.ns)\n\terr := c.hpa.DeleteCollection(ctx, options, listOptions)\n\tif err != nil {\n\t\tcode := getStatusCode(err)\n\t\tc.metrics.ClientMetricErrors.WithLabelValues(\"hpa\", \"deletecollection\", \"\", ns, code).Inc()\n\t\treturn err\n\t}\n\tc.metrics.ClientMetricSuccess.WithLabelValues(\"hpa\", \"deletecollection\", \"\", ns).Inc()\n\treturn nil\n}", "func (c *FakeRedisTriggers) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(redistriggersResource, c.ns, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &v1beta1.RedisTriggerList{})\n\treturn err\n}", "func (c *klusterlets) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOpts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tResource(\"klusterlets\").\n\t\tVersionedParams(&listOpts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(&opts).\n\t\tDo(ctx).\n\t\tError()\n}", "func (c *solrBackups) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOptions.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"solrbackups\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *userProjectBindings) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOptions.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tResource(\"userprojectbindings\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *jxTasks) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"jxtasks\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *FakeManagedCertificates) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(managedcertificatesResource, c.ns, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &networkinggkeiov1.ManagedCertificateList{})\n\treturn err\n}", "func (c *awsSesReceiptRuleSets) DeleteCollection(options *meta_v1.DeleteOptions, listOptions meta_v1.ListOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"awssesreceiptrulesets\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *FakeTraefikServices) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(traefikservicesResource, c.ns, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.TraefikServiceList{})\n\treturn err\n}", "func (c *destinationPolicies) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"destinationpolicies\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *FakeSystemBackups) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(systembackupsResource, c.ns, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &v1beta2.SystemBackupList{})\n\treturn err\n}", "func (c *FakeSnatPolicies) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewRootDeleteCollectionAction(snatpoliciesResource, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &acisnatv1.SnatPolicyList{})\n\treturn err\n}", "func (c *iperfTasks) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOptions.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"iperftasks\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *routeGroups) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOptions.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"routegroups\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *kuberhealthyChecks) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOptions.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"khchecks\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(options).\n\t\tDo(context.TODO()).\n\t\tError()\n}", "func (c *previewFeatures) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOpts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tResource(\"previewfeatures\").\n\t\tVersionedParams(&listOpts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(&opts).\n\t\tDo(ctx).\n\t\tError()\n}", "func (c *tiKVGroups) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOptions.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"tikvgroups\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *FakeDaemonstools) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(daemonstoolsResource, c.ns, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &v1beta1.DaemonstoolList{})\n\treturn err\n}", "func (c *FakeHelms) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(helmsResource, c.ns, listOptions)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.HelmList{})\n\treturn err\n}", "func (c *googleCloudStorageSources) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOpts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"googlecloudstoragesources\").\n\t\tVersionedParams(&listOpts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(&opts).\n\t\tDo(ctx).\n\t\tError()\n}", "func (c *FakeKuberaBackups) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(kuberabackupsResource, c.ns, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &backuprestorev1.KuberaBackupList{})\n\treturn err\n}", "func (c *FakeGBPServers) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(gbpserversResource, c.ns, listOptions)\n\n\t_, err := c.Fake.Invokes(action, &aciawv1.GBPServerList{})\n\treturn err\n}", "func (c *networkServicesEndpointPolicies) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOpts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"networkservicesendpointpolicies\").\n\t\tVersionedParams(&listOpts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(&opts).\n\t\tDo(ctx).\n\t\tError()\n}", "func (c *FakeExportedServiceSets) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(exportedservicesetsResource, c.ns, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &federationv1.ExportedServiceSetList{})\n\treturn err\n}", "func (c *secrets) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOpts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second\n\t}\n\n\treturn c.client.Delete().\n\t\tResource(\"secrets\").\n\t\tVersionedParams(listOpts).\n\t\tTimeout(timeout).\n\t\tBody(&opts).\n\t\tDo(ctx).\n\t\tError()\n}", "func (c *krakenClusters) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"krakenclusters\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *FakeRuleEndpoints) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(ruleendpointsResource, c.ns, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &rulesv1.RuleEndpointList{})\n\treturn err\n}", "func (c *FakeSpidermans) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(spidermansResource, c.ns, listOptions)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.SpidermanList{})\n\treturn err\n}", "func (c *FakeAwsApiGatewayVpcLinks) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(awsapigatewayvpclinksResource, c.ns, listOptions)\n\n\t_, err := c.Fake.Invokes(action, &aws_v1.AwsApiGatewayVpcLinkList{})\n\treturn err\n}", "func (c *FakeGlueCatalogTables) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(gluecatalogtablesResource, c.ns, listOptions)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.GlueCatalogTableList{})\n\treturn err\n}", "func (c *FakeEndpointsServices) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(endpointsservicesResource, c.ns, listOptions)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.EndpointsServiceList{})\n\treturn err\n}", "func (c *globalThreatFeeds) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOpts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tResource(\"globalthreatfeeds\").\n\t\tVersionedParams(&listOpts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(&opts).\n\t\tDo(ctx).\n\t\tError()\n}", "func (c *FakeAwsIamGroupPolicyAttachments) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(awsiamgrouppolicyattachmentsResource, c.ns, listOptions)\n\n\t_, err := c.Fake.Invokes(action, &aws_v1.AwsIamGroupPolicyAttachmentList{})\n\treturn err\n}", "func (c *customReplicationControllers) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"customreplicationcontrollers\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *FakeVaultSecretClaims) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(vaultsecretclaimsResource, c.ns, listOptions)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.VaultSecretClaimList{})\n\treturn err\n}", "func (c *FakeAzureEventHubsSources) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(azureeventhubssourcesResource, c.ns, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.AzureEventHubsSourceList{})\n\treturn err\n}", "func (c *FakeClusterImageSets) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewRootDeleteCollectionAction(clusterimagesetsResource, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &hivev1.ClusterImageSetList{})\n\treturn err\n}", "func (c *FakePrunes) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(prunesResource, c.ns, listOptions)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.PruneList{})\n\treturn err\n}", "func (c *FakeTraincrds) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(traincrdsResource, c.ns, listOptions)\n\n\t_, err := c.Fake.Invokes(action, &apisv1.TraincrdList{})\n\treturn err\n}", "func (c *awsIamRolePolicyAttachments) DeleteCollection(options *meta_v1.DeleteOptions, listOptions meta_v1.ListOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"awsiamrolepolicyattachments\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *FakeAcceptableUsePolicies) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(acceptableusepoliciesResource, c.ns, listOptions)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha.AcceptableUsePolicyList{})\n\treturn err\n}", "func (c *awsDmsCertificates) DeleteCollection(options *meta_v1.DeleteOptions, listOptions meta_v1.ListOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"awsdmscertificates\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *FakeNetworkConnectivityHubs) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(networkconnectivityhubsResource, c.ns, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &v1beta1.NetworkConnectivityHubList{})\n\treturn err\n}", "func (impl *ServerServerGroup) DeleteCollection() ([]base.ModelInterface, *base.ErrorResponse) {\n\tvar (\n\t\tname = impl.ResourceName()\n\t\trecordCollection = impl.NewEntityCollection()\n\t\tc = impl.GetConnection()\n\t)\n\n\t// We need transaction to ensure the total and the query count is consistent.\n\ttx := c.Begin()\n\tif err := tx.Error; err != nil {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"resource\": name,\n\t\t\t\"error\": err,\n\t\t}).Warn(\"Delete collection in DB failed, start transaction failed.\")\n\t\treturn nil, base.NewErrorResponseTransactionError()\n\t}\n\n\tif err := tx.Where(\"\\\"ServerGroupID\\\" <> ?\", DefaultServerGroupID).Find(recordCollection).Error; err != nil {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"resource\": name,\n\t\t\t\"error\": err,\n\t\t}).Warn(\"Delete collection in DB failed, find resource failed.\")\n\t\treturn nil, base.NewErrorResponseTransactionError()\n\t}\n\n\tif err := tx.Where(\"\\\"ServerGroupID\\\" <> ?\", DefaultServerGroupID).Delete(entity.ServerServerGroup{}).Error; err != nil {\n\t\ttx.Rollback()\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"resource\": name,\n\t\t\t\"error\": err,\n\t\t}).Warn(\"Delete collection in DB failed, delete resources failed, transaction rollback.\")\n\t\treturn nil, base.NewErrorResponseTransactionError()\n\t}\n\tret, errorResp := impl.TemplateImpl.ConvertFindResultToModel(recordCollection)\n\tif errorResp != nil {\n\t\ttx.Rollback()\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"resource\": name,\n\t\t\t\"errorResp\": errorResp.ID,\n\t\t}).Warn(\"Delete collection in DB failed, convert find result failed, transaction rollback.\")\n\t\treturn nil, errorResp\n\t}\n\tif err := tx.Commit().Error; err != nil {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"resource\": name,\n\t\t\t\"error\": err,\n\t\t}).Warn(\"Delete collection in DB failed, commit failed.\")\n\t\treturn nil, base.NewErrorResponseTransactionError()\n\t}\n\treturn ret, nil\n}", "func (c *FakeClusterManagers) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewRootDeleteCollectionAction(clustermanagersResource, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &operatorv1.ClusterManagerList{})\n\treturn err\n}", "func (c *FakeAWSSNSTargets) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(awssnstargetsResource, c.ns, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.AWSSNSTargetList{})\n\treturn err\n}", "func (c *FakeL4Rules) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(l4rulesResource, c.ns, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha2.L4RuleList{})\n\treturn err\n}", "func (c *podSecurityPolicySubjectReviews) DeleteCollection(options *pkg_api.DeleteOptions, listOptions pkg_api.ListOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"podsecuritypolicysubjectreviews\").\n\t\tVersionedParams(&listOptions, pkg_api.ParameterCodec).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (a *ManagementApiService) DeleteCollection(ctx _context.Context, applicationId int32, campaignId int32, collectionId int32) apiDeleteCollectionRequest {\n\treturn apiDeleteCollectionRequest{\n\t\tapiService: a,\n\t\tctx: ctx,\n\t\tapplicationId: applicationId,\n\t\tcampaignId: campaignId,\n\t\tcollectionId: collectionId,\n\t}\n}", "func (c *FakeCloudwatchEventTargets) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(cloudwatcheventtargetsResource, c.ns, listOptions)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.CloudwatchEventTargetList{})\n\treturn err\n}", "func (c *FakePodNetworkings) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewRootDeleteCollectionAction(podnetworkingsResource, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &v1beta1.PodNetworkingList{})\n\treturn err\n}", "func (c *FakeWorkflowTemplates) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(workflowtemplatesResource, c.ns, listOptions)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.WorkflowTemplateList{})\n\treturn err\n}", "func (c *FakeApiGatewayMethodSettingses) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(apigatewaymethodsettingsesResource, c.ns, listOptions)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.ApiGatewayMethodSettingsList{})\n\treturn err\n}", "func (c *FakeProxyRoutes) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(proxyroutesResource, c.ns, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.ProxyRouteList{})\n\treturn err\n}", "func (c *clusterVulnerabilityReports) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOpts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tResource(\"clustervulnerabilityreports\").\n\t\tVersionedParams(&listOpts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(&opts).\n\t\tDo(ctx).\n\t\tError()\n}", "func (c *FakeTridentSnapshotInfos) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(tridentsnapshotinfosResource, c.ns, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &netappv1.TridentSnapshotInfoList{})\n\treturn err\n}", "func (c *tiDBGroups) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\tvar timeout time.Duration\n\tif listOptions.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second\n\t}\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"tidbgroups\").\n\t\tVersionedParams(&listOptions, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tBody(options).\n\t\tDo().\n\t\tError()\n}", "func (c *FakeGoogleServiceAccounts) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(googleserviceaccountsResource, c.ns, listOptions)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.GoogleServiceAccountList{})\n\treturn err\n}", "func (c *FakeDaskClusters) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(daskclustersResource, c.ns, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &kubernetesdaskorgv1.DaskClusterList{})\n\treturn err\n}", "func (c *FakeHealthCheckPolicies) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(healthcheckpoliciesResource, c.ns, listOptions)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.HealthCheckPolicyList{})\n\treturn err\n}", "func (c *FakeGameServerAllocations) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(gameserverallocationsResource, c.ns, listOptions)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.GameServerAllocationList{})\n\treturn err\n}", "func (c *FakeGoogleCloudPubSubSources) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(googlecloudpubsubsourcesResource, c.ns, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.GoogleCloudPubSubSourceList{})\n\treturn err\n}", "func (c *FakeExampleOperators) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(exampleoperatorsResource, c.ns, listOptions)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.ExampleOperatorList{})\n\treturn err\n}", "func (c *FakeFederatedNotificationManagers) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewRootDeleteCollectionAction(federatednotificationmanagersResource, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &v1beta2.FederatedNotificationManagerList{})\n\treturn err\n}", "func (c *FakeIamUserPolicyAttachments) DeleteCollection(options *v1.DeleteOptions, listOptions v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(iamuserpolicyattachmentsResource, c.ns, listOptions)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.IamUserPolicyAttachmentList{})\n\treturn err\n}", "func (p PostgresPersister) DeleteCollection(id string) error {\n\tvar dbid int32\n\tfmt.Sscanf(id, p.pathPrefix+model.CollectionIDFormat, &dbid)\n\t_, err := p.db.Exec(\"DELETE FROM collection WHERE id = $1\", dbid)\n\treturn translateError(err)\n}", "func (c *FakeAppEngineFlexibleAppVersions) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {\n\taction := testing.NewDeleteCollectionAction(appengineflexibleappversionsResource, c.ns, listOpts)\n\n\t_, err := c.Fake.Invokes(action, &v1alpha1.AppEngineFlexibleAppVersionList{})\n\treturn err\n}" ]
[ "0.7699723", "0.7570869", "0.75703704", "0.75454164", "0.7514351", "0.7488343", "0.7451056", "0.74491966", "0.74294275", "0.7426684", "0.74233156", "0.7389707", "0.7370067", "0.7361063", "0.7334545", "0.7327154", "0.7316015", "0.7310656", "0.729753", "0.7297466", "0.72921824", "0.7283851", "0.7271686", "0.7263111", "0.7241649", "0.7207388", "0.72064424", "0.72014874", "0.7199653", "0.71760017", "0.7173545", "0.71662444", "0.7160453", "0.71478593", "0.7147578", "0.7136736", "0.7133718", "0.7133091", "0.7126527", "0.71187246", "0.7116073", "0.711103", "0.7105441", "0.70986146", "0.7094344", "0.70904547", "0.7070048", "0.706444", "0.70625335", "0.70598567", "0.705932", "0.70576197", "0.7049842", "0.70479715", "0.7044902", "0.70419455", "0.70357305", "0.7031671", "0.70080435", "0.70072424", "0.7006902", "0.6990322", "0.6986337", "0.69836634", "0.69733405", "0.696775", "0.6961", "0.69535977", "0.6953", "0.6952913", "0.694083", "0.6937553", "0.69319034", "0.69314396", "0.69269407", "0.69251764", "0.6922217", "0.6912924", "0.6908678", "0.69035363", "0.68884", "0.68861294", "0.6878824", "0.6864794", "0.6864632", "0.6861005", "0.6834033", "0.6830367", "0.6822361", "0.68028736", "0.6801885", "0.67925274", "0.67880195", "0.6786985", "0.6781987", "0.677914", "0.6764247", "0.6763957", "0.6762543", "0.6757195", "0.6756878" ]
0.0
-1
Patch applies the patch and returns the patched tZCronJob.
func (c *FakeTZCronJobs) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1alpha1.TZCronJob, err error) { obj, err := c.Fake. Invokes(testing.NewPatchSubresourceAction(tzcronjobsResource, c.ns, name, pt, data, subresources...), &v1alpha1.TZCronJob{}) if obj == nil { return nil, err } return obj.(*v1alpha1.TZCronJob), err }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *aITrainingJobs) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.AITrainingJob, err error) {\n\tresult = &v1.AITrainingJob{}\n\terr = c.client.Patch(pt).\n\t\tNamespace(c.ns).\n\t\tResource(\"aitrainingjobs\").\n\t\tSubResource(subresources...).\n\t\tName(name).\n\t\tBody(data).\n\t\tDo().\n\t\tInto(result)\n\treturn\n}", "func PatchJob(ns, name string, updateFunc func(job *v1.Job)) error {\n\toJob, err := kubecli.BatchV1().Jobs(ns).Get(name, metav1.GetOptions{})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tnJob := cloneJob(oJob)\n\tupdateFunc(nJob)\n\tpatchData, err := CreatePatch(oJob, nJob, v1.Job{})\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = kubecli.BatchV1().Jobs(ns).Patch(name, types.StrategicMergePatchType, patchData)\n\treturn err\n}", "func (r *ProjectsLocationsJobsService) Patch(name string, job *Job) *ProjectsLocationsJobsPatchCall {\n\tc := &ProjectsLocationsJobsPatchCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.name = name\n\tc.job = job\n\treturn c\n}", "func (c *restClient) ExecutePatchJob(ctx context.Context, req *osconfigpb.ExecutePatchJobRequest, opts ...gax.CallOption) (*osconfigpb.PatchJob, error) {\n\tm := protojson.MarshalOptions{AllowPartial: true, UseEnumNumbers: true}\n\tjsonReq, err := m.Marshal(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v1beta/%v/patchJobs:execute\", req.GetParent())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"parent\", url.QueryEscape(req.GetParent()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\topts = append((*c.CallOptions).ExecutePatchJob[0:len((*c.CallOptions).ExecutePatchJob):len((*c.CallOptions).ExecutePatchJob)], opts...)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tresp := &osconfigpb.PatchJob{}\n\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"POST\", baseUrl.String(), bytes.NewReader(jsonReq))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}, opts...)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\treturn resp, nil\n}", "func (c *restClient) GetPatchJob(ctx context.Context, req *osconfigpb.GetPatchJobRequest, opts ...gax.CallOption) (*osconfigpb.PatchJob, error) {\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v1beta/%v\", req.GetName())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"name\", url.QueryEscape(req.GetName()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\topts = append((*c.CallOptions).GetPatchJob[0:len((*c.CallOptions).GetPatchJob):len((*c.CallOptions).GetPatchJob)], opts...)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tresp := &osconfigpb.PatchJob{}\n\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"GET\", baseUrl.String(), nil)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}, opts...)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\treturn resp, nil\n}", "func (r *TransferJobsService) Patch(jobName string, updatetransferjobrequest *UpdateTransferJobRequest) *TransferJobsPatchCall {\n\tc := &TransferJobsPatchCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.jobName = jobName\n\tc.updatetransferjobrequest = updatetransferjobrequest\n\treturn c\n}", "func (c *Client) GetPatchJob(ctx context.Context, req *osconfigpb.GetPatchJobRequest, opts ...gax.CallOption) (*osconfigpb.PatchJob, error) {\n\treturn c.internalClient.GetPatchJob(ctx, req, opts...)\n}", "func (m *MockJobClient) Patch(arg0, arg1 string, arg2 types.PatchType, arg3 []byte, arg4 ...string) (*v1.Job, error) {\n\tm.ctrl.T.Helper()\n\tvarargs := []interface{}{arg0, arg1, arg2, arg3}\n\tfor _, a := range arg4 {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"Patch\", varargs...)\n\tret0, _ := ret[0].(*v1.Job)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (c *jxTasks) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1alpha1.JxTask, err error) {\n\tresult = &v1alpha1.JxTask{}\n\terr = c.client.Patch(pt).\n\t\tNamespace(c.ns).\n\t\tResource(\"jxtasks\").\n\t\tSubResource(subresources...).\n\t\tName(name).\n\t\tBody(data).\n\t\tDo().\n\t\tInto(result)\n\treturn\n}", "func SchedulePatch(ctx context.Context, env evergreen.Environment, patchId string, version *model.Version, patchUpdateReq model.PatchUpdate) (int, error) {\n\tvar err error\n\tp, err := patch.FindOneId(patchId)\n\tif err != nil {\n\t\treturn http.StatusInternalServerError, errors.Wrapf(err, \"loading patch '%s'\", patchId)\n\t}\n\tif p == nil {\n\t\treturn http.StatusBadRequest, errors.Errorf(\"patch '%s' not found\", patchId)\n\t}\n\n\tif p.IsCommitQueuePatch() {\n\t\treturn http.StatusBadRequest, errors.New(\"can't schedule commit queue patch\")\n\t}\n\tprojectRef, err := model.FindMergedProjectRef(p.Project, p.Version, true)\n\tif err != nil {\n\t\treturn http.StatusInternalServerError, errors.Wrapf(err, \"finding project ref '%s' for version '%s'\", p.Project, p.Version)\n\t}\n\tif projectRef == nil {\n\t\treturn http.StatusInternalServerError, errors.Errorf(\"project '%s' for version '%s' not found\", p.Project, p.Version)\n\t}\n\n\tstatusCode, err := model.ConfigurePatch(ctx, env.Settings(), p, version, projectRef, patchUpdateReq)\n\tif err != nil {\n\t\treturn statusCode, err\n\t}\n\tif p.Version != \"\" { // if the version already exists, no more to do\n\t\treturn http.StatusOK, nil\n\t}\n\n\t// create a separate context from the one the caller has so that the caller\n\t// can't interrupt the db operations here\n\tnewCxt := context.Background()\n\t// Process additional patch trigger aliases added via UI.\n\t// Child patches created with the CLI --trigger-alias flag go through a separate flow, so ensure that new child patches are also created before the parent is finalized.\n\tif err := ProcessTriggerAliases(ctx, p, projectRef, env, patchUpdateReq.PatchTriggerAliases); err != nil {\n\t\treturn http.StatusInternalServerError, errors.Wrap(err, \"processing patch trigger aliases\")\n\t}\n\tif len(patchUpdateReq.PatchTriggerAliases) > 0 {\n\t\tp.Triggers.Aliases = patchUpdateReq.PatchTriggerAliases\n\t\tif err = p.SetTriggerAliases(); err != nil {\n\t\t\treturn http.StatusInternalServerError, errors.Wrapf(err, \"attaching trigger aliases '%s'\", p.Id.Hex())\n\t\t}\n\t}\n\t_, err = model.FinalizePatch(newCxt, p, p.GetRequester(), \"\")\n\tif err != nil {\n\t\treturn http.StatusInternalServerError, errors.Wrap(err, \"finalizing patch\")\n\t}\n\n\tif p.IsGithubPRPatch() {\n\t\tjob := NewGithubStatusUpdateJobForNewPatch(p.Id.Hex())\n\t\tif err := evergreen.GetEnvironment().LocalQueue().Put(newCxt, job); err != nil {\n\t\t\treturn http.StatusInternalServerError, errors.Wrap(err, \"adding GitHub status update job to queue\")\n\t\t}\n\t}\n\treturn http.StatusOK, nil\n}", "func (importedCertificateMetadataPatch *ImportedCertificateMetadataPatch) AsPatch() (_patch map[string]interface{}, err error) {\n\tvar jsonData []byte\n\tjsonData, err = json.Marshal(importedCertificateMetadataPatch)\n\tif err == nil {\n\t\terr = json.Unmarshal(jsonData, &_patch)\n\t}\n\treturn\n}", "func MockCronJob() batchv1beta1.CronJob {\n\tp := MockPod()\n\treturn batchv1beta1.CronJob{\n\t\tSpec: batchv1beta1.CronJobSpec{\n\t\t\tJobTemplate: batchv1beta1.JobTemplateSpec{\n\t\t\t\tSpec: batchv1.JobSpec{\n\t\t\t\t\tTemplate: corev1.PodTemplateSpec{Spec: p.Spec},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n}", "func (m *PrintersItemJobsPrintJobItemRequestBuilder) Patch(ctx context.Context, body iadcd81124412c61e647227ecfc4449d8bba17de0380ddda76f641a29edf2b242.PrintJobable, requestConfiguration *PrintersItemJobsPrintJobItemRequestBuilderPatchRequestConfiguration)(iadcd81124412c61e647227ecfc4449d8bba17de0380ddda76f641a29edf2b242.PrintJobable, error) {\n requestInfo, err := m.ToPatchRequestInformation(ctx, body, requestConfiguration);\n if err != nil {\n return nil, err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": ia572726a95efa92ddd544552cd950653dc691023836923576b2f4bf716cf204a.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": ia572726a95efa92ddd544552cd950653dc691023836923576b2f4bf716cf204a.CreateODataErrorFromDiscriminatorValue,\n }\n res, err := m.BaseRequestBuilder.RequestAdapter.Send(ctx, requestInfo, iadcd81124412c61e647227ecfc4449d8bba17de0380ddda76f641a29edf2b242.CreatePrintJobFromDiscriminatorValue, errorMapping)\n if err != nil {\n return nil, err\n }\n if res == nil {\n return nil, nil\n }\n return res.(iadcd81124412c61e647227ecfc4449d8bba17de0380ddda76f641a29edf2b242.PrintJobable), nil\n}", "func (r *JobsService) Patch(name string, updatejobrequest *UpdateJobRequest) *JobsPatchCall {\n\tc := &JobsPatchCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.name = name\n\tc.updatejobrequest = updatejobrequest\n\treturn c\n}", "func SchedulePatch(ctx context.Context, patchId string, version *model.Version, patchUpdateReq PatchUpdate) (error, int, string, string) {\n\tvar err error\n\tp, err := patch.FindOneId(patchId)\n\tif err != nil {\n\t\treturn errors.Errorf(\"error loading patch: %s\", err), http.StatusInternalServerError, \"\", \"\"\n\t}\n\n\t// only modify parameters if the patch hasn't been finalized\n\tif patchUpdateReq.ParametersModel != nil && p.Version == \"\" {\n\t\tvar parameters []patch.Parameter\n\t\tfor _, param := range patchUpdateReq.ParametersModel {\n\t\t\tparameters = append(parameters, param.ToService())\n\t\t}\n\t\tif err = p.SetParameters(parameters); err != nil {\n\t\t\treturn errors.Errorf(\"error setting patch parameters: %s\", err), http.StatusInternalServerError, \"\", \"\"\n\t\t}\n\t}\n\n\tif p.IsCommitQueuePatch() {\n\t\treturn errors.New(\"can't schedule commit queue patch\"), http.StatusBadRequest, \"\", \"\"\n\t}\n\n\t// Unmarshal the project config and set it in the project context\n\tproject := &model.Project{}\n\tif _, err = model.LoadProjectInto(ctx, []byte(p.PatchedConfig), nil, p.Project, project); err != nil {\n\t\treturn errors.Errorf(\"Error unmarshaling project config: %v\", err), http.StatusInternalServerError, \"\", \"\"\n\t}\n\n\taddDisplayTasksToPatchReq(&patchUpdateReq, *project)\n\ttasks := model.VariantTasksToTVPairs(patchUpdateReq.VariantsTasks)\n\n\ttasks.ExecTasks, err = model.IncludeDependencies(project, tasks.ExecTasks, p.GetRequester())\n\tgrip.Warning(message.WrapError(err, message.Fields{\n\t\t\"message\": \"error including dependencies for patch\",\n\t\t\"patch\": patchId,\n\t}))\n\n\tif err = model.ValidateTVPairs(project, tasks.ExecTasks); err != nil {\n\t\treturn err, http.StatusBadRequest, \"\", \"\"\n\t}\n\n\t// update the description for both reconfigured and new patches\n\tif err = p.SetDescription(patchUpdateReq.Description); err != nil {\n\t\treturn errors.Wrap(err, \"Error setting description\"), http.StatusInternalServerError, \"\", \"\"\n\t}\n\n\t// update the description for both reconfigured and new patches\n\tif err = p.SetVariantsTasks(tasks.TVPairsToVariantTasks()); err != nil {\n\t\treturn errors.Wrap(err, \"Error setting description\"), http.StatusInternalServerError, \"\", \"\"\n\t}\n\n\t// create a separate context from the one the callar has so that the caller\n\t// can't interrupt the db operations here\n\tnewCxt := context.Background()\n\n\tprojectRef, err := model.FindMergedProjectRef(project.Identifier)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"unable to find project ref\"), http.StatusInternalServerError, \"\", \"\"\n\t}\n\tif projectRef == nil {\n\t\treturn errors.Errorf(\"project '%s' not found\", project.Identifier), http.StatusInternalServerError, \"\", \"\"\n\t}\n\n\tif p.Version != \"\" {\n\t\tp.Activated = true\n\t\t// This patch has already been finalized, just add the new builds and tasks\n\t\tif version == nil {\n\t\t\treturn errors.Errorf(\"Couldn't find patch for id %v\", p.Version), http.StatusInternalServerError, \"\", \"\"\n\t\t}\n\t\t// First add new tasks to existing builds, if necessary\n\t\terr = model.AddNewTasksForPatch(context.Background(), p, version, project, tasks, projectRef.Identifier)\n\t\tif err != nil {\n\t\t\treturn errors.Wrapf(err, \"Error creating new tasks for version `%s`\", version.Id), http.StatusInternalServerError, \"\", \"\"\n\t\t}\n\n\t\terr = model.AddNewBuildsForPatch(newCxt, p, version, project, tasks, projectRef)\n\t\tif err != nil {\n\t\t\treturn errors.Wrapf(err, \"Error creating new builds for version `%s`\", version.Id), http.StatusInternalServerError, \"\", \"\"\n\t\t}\n\n\t\treturn nil, http.StatusOK, \"Builds and tasks successfully added to patch.\", version.Id\n\n\t} else {\n\t\tsettings, err := evergreen.GetConfig()\n\t\tif err != nil {\n\t\t\treturn err, http.StatusInternalServerError, \"\", \"\"\n\t\t}\n\t\tgithubOauthToken, err := settings.GetGithubOauthToken()\n\t\tif err != nil {\n\t\t\treturn err, http.StatusInternalServerError, \"\", \"\"\n\t\t}\n\t\tp.Activated = true\n\t\terr = p.SetVariantsTasks(tasks.TVPairsToVariantTasks())\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"Error setting patch variants and tasks\"), http.StatusInternalServerError, \"\", \"\"\n\t\t}\n\n\t\t// Process additional patch trigger aliases added via UI.\n\t\t// Child patches created with the CLI --trigger-alias flag go through a separate flow, so ensure that new child patches are also created before the parent is finalized.\n\t\tchildPatchIds, err := units.ProcessTriggerAliases(ctx, p, projectRef, evergreen.GetEnvironment(), patchUpdateReq.PatchTriggerAliases)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"Error processing patch trigger aliases\"), http.StatusInternalServerError, \"\", \"\"\n\t\t}\n\t\tif len(childPatchIds) > 0 {\n\t\t\tif err = p.SetChildPatches(); err != nil {\n\t\t\t\treturn errors.Wrapf(err, \"error attaching child patches '%s'\", p.Id.Hex()), http.StatusInternalServerError, \"\", \"\"\n\t\t\t}\n\t\t\tp.Triggers.Aliases = patchUpdateReq.PatchTriggerAliases\n\t\t\tif err = p.SetTriggerAliases(); err != nil {\n\t\t\t\treturn errors.Wrapf(err, \"error attaching trigger aliases '%s'\", p.Id.Hex()), http.StatusInternalServerError, \"\", \"\"\n\t\t\t}\n\t\t}\n\n\t\trequester := p.GetRequester()\n\t\tver, err := model.FinalizePatch(newCxt, p, requester, githubOauthToken)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"Error finalizing patch\"), http.StatusInternalServerError, \"\", \"\"\n\t\t}\n\t\tif requester == evergreen.PatchVersionRequester {\n\t\t\tgrip.Info(message.Fields{\n\t\t\t\t\"operation\": \"patch creation\",\n\t\t\t\t\"message\": \"finalized patch\",\n\t\t\t\t\"from\": \"UI\",\n\t\t\t\t\"patch_id\": p.Id,\n\t\t\t\t\"variants\": p.BuildVariants,\n\t\t\t\t\"tasks\": p.Tasks,\n\t\t\t\t\"variant_tasks\": p.VariantsTasks,\n\t\t\t\t\"alias\": p.Alias,\n\t\t\t})\n\t\t}\n\n\t\tif p.IsGithubPRPatch() {\n\t\t\tjob := units.NewGithubStatusUpdateJobForNewPatch(p.Id.Hex())\n\t\t\tif err := evergreen.GetEnvironment().LocalQueue().Put(newCxt, job); err != nil {\n\t\t\t\treturn errors.Wrap(err, \"Error adding github status update job to queue\"), http.StatusInternalServerError, \"\", \"\"\n\t\t\t}\n\t\t}\n\n\t\treturn nil, http.StatusOK, \"Patch builds are scheduled.\", ver.Id\n\t}\n}", "func (privateCertificateConfigurationIntermediateCAPatch *PrivateCertificateConfigurationIntermediateCAPatch) AsPatch() (_patch map[string]interface{}, err error) {\n\tvar jsonData []byte\n\tjsonData, err = json.Marshal(privateCertificateConfigurationIntermediateCAPatch)\n\tif err == nil {\n\t\terr = json.Unmarshal(jsonData, &_patch)\n\t}\n\treturn\n}", "func (c *FakeTZCronJobs) Update(tZCronJob *v1alpha1.TZCronJob) (result *v1alpha1.TZCronJob, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewUpdateAction(tzcronjobsResource, c.ns, tZCronJob), &v1alpha1.TZCronJob{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\treturn obj.(*v1alpha1.TZCronJob), err\n}", "func CronJobReconciler(reportName string, mrc *kubermaticv1.MeteringReportConfiguration, caBundleName string, getRegistry registry.ImageRewriter, seed *kubermaticv1.Seed) reconciling.NamedCronJobReconcilerFactory {\n\treturn func() (string, reconciling.CronJobReconciler) {\n\t\treturn reportName, func(job *batchv1.CronJob) (*batchv1.CronJob, error) {\n\t\t\tvar args []string\n\t\t\targs = append(args, fmt.Sprintf(\"--ca-bundle=%s\", \"/opt/ca-bundle/ca-bundle.pem\"))\n\t\t\targs = append(args, fmt.Sprintf(\"--prometheus-api=http://%s.%s.svc\", prometheus.Name, seed.Namespace))\n\t\t\targs = append(args, fmt.Sprintf(\"--output-dir=%s\", reportName))\n\t\t\targs = append(args, fmt.Sprintf(\"--output-prefix=%s\", seed.Name))\n\n\t\t\tif mrc.Monthly {\n\t\t\t\targs = append(args, \"--last-month\")\n\t\t\t} else {\n\t\t\t\targs = append(args, fmt.Sprintf(\"--last-number-of-days=%d\", mrc.Interval))\n\t\t\t}\n\n\t\t\t// needs to be last\n\t\t\targs = append(args, mrc.Types...)\n\n\t\t\tif job.Labels == nil {\n\t\t\t\tjob.Labels = make(map[string]string)\n\t\t\t}\n\t\t\tjob.Labels[common.NameLabel] = reportName\n\t\t\tjob.Labels[common.ComponentLabel] = meteringName\n\n\t\t\tjob.Spec.Schedule = mrc.Schedule\n\t\t\tjob.Spec.JobTemplate.Spec.Parallelism = pointer.Int32(1)\n\t\t\tjob.Spec.JobTemplate.Spec.Template.Spec.ServiceAccountName = \"\"\n\t\t\tjob.Spec.JobTemplate.Spec.Template.Spec.DeprecatedServiceAccount = \"\"\n\t\t\tjob.Spec.JobTemplate.Spec.Template.Spec.RestartPolicy = corev1.RestartPolicyOnFailure\n\t\t\tjob.Spec.JobTemplate.Spec.Template.Spec.ImagePullSecrets = []corev1.LocalObjectReference{{Name: resources.ImagePullSecretName}}\n\n\t\t\tjob.Spec.JobTemplate.Spec.Template.Spec.Containers = []corev1.Container{\n\t\t\t\t{\n\t\t\t\t\tName: reportName,\n\t\t\t\t\tImage: getMeteringImage(getRegistry),\n\t\t\t\t\tImagePullPolicy: corev1.PullIfNotPresent,\n\t\t\t\t\tCommand: []string{\"/metering\"},\n\t\t\t\t\tArgs: args,\n\t\t\t\t\tEnv: []corev1.EnvVar{\n\t\t\t\t\t\t{\n\t\t\t\t\t\t\tName: \"S3_ENDPOINT\",\n\t\t\t\t\t\t\tValueFrom: &corev1.EnvVarSource{\n\t\t\t\t\t\t\t\tSecretKeyRef: &corev1.SecretKeySelector{\n\t\t\t\t\t\t\t\t\tLocalObjectReference: corev1.LocalObjectReference{\n\t\t\t\t\t\t\t\t\t\tName: SecretName,\n\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t\tKey: Endpoint,\n\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t},\n\t\t\t\t\t\t{\n\t\t\t\t\t\t\tName: \"S3_BUCKET\",\n\t\t\t\t\t\t\tValueFrom: &corev1.EnvVarSource{\n\t\t\t\t\t\t\t\tSecretKeyRef: &corev1.SecretKeySelector{\n\t\t\t\t\t\t\t\t\tLocalObjectReference: corev1.LocalObjectReference{\n\t\t\t\t\t\t\t\t\t\tName: SecretName,\n\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t\tKey: Bucket,\n\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t},\n\t\t\t\t\t\t{\n\t\t\t\t\t\t\tName: \"ACCESS_KEY_ID\",\n\t\t\t\t\t\t\tValueFrom: &corev1.EnvVarSource{\n\t\t\t\t\t\t\t\tSecretKeyRef: &corev1.SecretKeySelector{\n\t\t\t\t\t\t\t\t\tLocalObjectReference: corev1.LocalObjectReference{\n\t\t\t\t\t\t\t\t\t\tName: SecretName,\n\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t\tKey: AccessKey,\n\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t},\n\t\t\t\t\t\t{\n\t\t\t\t\t\t\tName: \"SECRET_ACCESS_KEY\",\n\t\t\t\t\t\t\tValueFrom: &corev1.EnvVarSource{\n\t\t\t\t\t\t\t\tSecretKeyRef: &corev1.SecretKeySelector{\n\t\t\t\t\t\t\t\t\tLocalObjectReference: corev1.LocalObjectReference{\n\t\t\t\t\t\t\t\t\t\tName: SecretName,\n\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t\tKey: SecretKey,\n\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t\tVolumeMounts: []corev1.VolumeMount{\n\t\t\t\t\t\t{\n\t\t\t\t\t\t\tName: \"ca-bundle\",\n\t\t\t\t\t\t\tMountPath: \"/opt/ca-bundle/\",\n\t\t\t\t\t\t\tReadOnly: true,\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t}\n\n\t\t\tjob.Spec.JobTemplate.Spec.Template.Spec.Volumes = []corev1.Volume{\n\t\t\t\t{\n\t\t\t\t\tName: \"ca-bundle\",\n\t\t\t\t\tVolumeSource: corev1.VolumeSource{\n\t\t\t\t\t\tConfigMap: &corev1.ConfigMapVolumeSource{\n\t\t\t\t\t\t\tLocalObjectReference: corev1.LocalObjectReference{\n\t\t\t\t\t\t\t\tName: caBundleName,\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t}\n\t\t\treturn job, nil\n\t\t}\n\t}\n}", "func (c *FakeImagePullJobs) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts v1.PatchOptions, subresources ...string) (result *v1alpha1.ImagePullJob, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewPatchSubresourceAction(imagepulljobsResource, c.ns, name, pt, data, subresources...), &v1alpha1.ImagePullJob{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\treturn obj.(*v1alpha1.ImagePullJob), err\n}", "func (c *Client) ExecutePatchJob(ctx context.Context, req *osconfigpb.ExecutePatchJobRequest, opts ...gax.CallOption) (*osconfigpb.PatchJob, error) {\n\treturn c.internalClient.ExecutePatchJob(ctx, req, opts...)\n}", "func (k *kubectlContext) Patch(args ...string) error {\n\tout, err := k.do(append([]string{\"patch\"}, args...)...)\n\tk.t.Log(string(out))\n\treturn err\n}", "func (d *defaultJobRepository) PatchJobStatus(ctxIn context.Context, jobPatcher JobPatch) error {\n _, span := trace.StartSpan(ctxIn, \"(*defaultJobRepository).PatchJobStatus\")\n defer span.End()\n\n job := &Job{\n Status: jobPatcher.Status,\n ForeignJobID: ForeignJobID{\n BigQueryID: jobPatcher.ForeignJobID.BigQueryID,\n CloudStorageID: jobPatcher.ForeignJobID.CloudStorageID,\n },\n EntityAudit: EntityAudit{\n UpdatedTimestamp: time.Now(),\n },\n }\n\n _, err := d.storageService.DB().Model(job).\n Column(\"status\", \"audit_updated_timestamp\", \"bigquery_extract_job_id\", \"cloudstorage_transfer_job_id\").\n Where(\"audit_deleted_timestamp IS NULL\").\n Where(\"id = ?\", jobPatcher.ID).\n Update()\n\n if err != nil {\n return fmt.Errorf(\"error during executing updating job statement: %s\", err)\n }\n\n return nil\n}", "func (c *restClient) CancelPatchJob(ctx context.Context, req *osconfigpb.CancelPatchJobRequest, opts ...gax.CallOption) (*osconfigpb.PatchJob, error) {\n\tm := protojson.MarshalOptions{AllowPartial: true, UseEnumNumbers: true}\n\tjsonReq, err := m.Marshal(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v1beta/%v:cancel\", req.GetName())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"name\", url.QueryEscape(req.GetName()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\topts = append((*c.CallOptions).CancelPatchJob[0:len((*c.CallOptions).CancelPatchJob):len((*c.CallOptions).CancelPatchJob)], opts...)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tresp := &osconfigpb.PatchJob{}\n\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"POST\", baseUrl.String(), bytes.NewReader(jsonReq))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}, opts...)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\treturn resp, nil\n}", "func (privateCertificateConfigurationRootCAPatch *PrivateCertificateConfigurationRootCAPatch) AsPatch() (_patch map[string]interface{}, err error) {\n\tvar jsonData []byte\n\tjsonData, err = json.Marshal(privateCertificateConfigurationRootCAPatch)\n\tif err == nil {\n\t\terr = json.Unmarshal(jsonData, &_patch)\n\t}\n\treturn\n}", "func (o HttpRuleOutput) Patch() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v HttpRule) *string { return v.Patch }).(pulumi.StringPtrOutput)\n}", "func Patch(expectations []*RunExpectation) func(t *testing.T) {\n\t_mocker = &runMocker{\n\t\texpectations: expectations,\n\t\tptr: 0,\n\t\tlength: len(expectations),\n\t}\n\n\treturn func(t *testing.T) {\n\t\tif expectation := _mocker.expectation(); expectation != nil {\n\t\t\tt.Errorf(\"execkit-mock: missing call: %v\", expectation)\n\t\t\tt.FailNow()\n\t\t}\n\t\t_mocker = nil\n\t}\n}", "func (privateCertificateConfigurationTemplatePatch *PrivateCertificateConfigurationTemplatePatch) AsPatch() (_patch map[string]interface{}, err error) {\n\tvar jsonData []byte\n\tjsonData, err = json.Marshal(privateCertificateConfigurationTemplatePatch)\n\tif err == nil {\n\t\terr = json.Unmarshal(jsonData, &_patch)\n\t}\n\treturn\n}", "func (publicCertificateMetadataPatch *PublicCertificateMetadataPatch) AsPatch() (_patch map[string]interface{}, err error) {\n\tvar jsonData []byte\n\tjsonData, err = json.Marshal(publicCertificateMetadataPatch)\n\tif err == nil {\n\t\terr = json.Unmarshal(jsonData, &_patch)\n\t}\n\treturn\n}", "func (privateCertificateMetadataPatch *PrivateCertificateMetadataPatch) AsPatch() (_patch map[string]interface{}, err error) {\n\tvar jsonData []byte\n\tjsonData, err = json.Marshal(privateCertificateMetadataPatch)\n\tif err == nil {\n\t\terr = json.Unmarshal(jsonData, &_patch)\n\t}\n\treturn\n}", "func (c *iperfTasks) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *alpha1.IperfTask, err error) {\n\tresult = &alpha1.IperfTask{}\n\terr = c.client.Patch(pt).\n\t\tNamespace(c.ns).\n\t\tResource(\"iperftasks\").\n\t\tSubResource(subresources...).\n\t\tName(name).\n\t\tBody(data).\n\t\tDo().\n\t\tInto(result)\n\treturn\n}", "func (r *Reconciler) patch(\n\tctx context.Context, object client.Object,\n\tpatch client.Patch, options ...client.PatchOption,\n) error {\n\toptions = append([]client.PatchOption{r.Owner}, options...)\n\treturn r.Client.Patch(ctx, object, patch, options...)\n}", "func (c *cronFederatedHPAs) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts v1.PatchOptions, subresources ...string) (result *v1alpha1.CronFederatedHPA, err error) {\n\tresult = &v1alpha1.CronFederatedHPA{}\n\terr = c.client.Patch(pt).\n\t\tNamespace(c.ns).\n\t\tResource(\"cronfederatedhpas\").\n\t\tName(name).\n\t\tSubResource(subresources...).\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tBody(data).\n\t\tDo(ctx).\n\t\tInto(result)\n\treturn\n}", "func (publicCertificateConfigurationCALetsEncryptPatch *PublicCertificateConfigurationCALetsEncryptPatch) AsPatch() (_patch map[string]interface{}, err error) {\n\tvar jsonData []byte\n\tjsonData, err = json.Marshal(publicCertificateConfigurationCALetsEncryptPatch)\n\tif err == nil {\n\t\terr = json.Unmarshal(jsonData, &_patch)\n\t}\n\treturn\n}", "func (c *krakenClusters) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1alpha1.KrakenCluster, err error) {\n\tresult = &v1alpha1.KrakenCluster{}\n\terr = c.client.Patch(pt).\n\t\tNamespace(c.ns).\n\t\tResource(\"krakenclusters\").\n\t\tSubResource(subresources...).\n\t\tName(name).\n\t\tBody(data).\n\t\tDo().\n\t\tInto(result)\n\treturn\n}", "func MockJob() batchv1.Job {\n\tp := MockPod()\n\treturn batchv1.Job{\n\t\tSpec: batchv1.JobSpec{\n\t\t\tTemplate: corev1.PodTemplateSpec{Spec: p.Spec},\n\t\t},\n\t}\n}", "func Patch() int {\n\treturn patch\n}", "func CreateOrUpdate(ctx context.Context, c client.Client, cj *batchv1.CronJob, equal EqualityFunc, mutate MutateFunc) error {\n\tcurrent := &batchv1.CronJob{}\n\tkey := client.ObjectKey{Name: cj.Name, Namespace: cj.Namespace}\n\terr := c.Get(ctx, key, current)\n\tif err != nil {\n\t\tif apierrors.IsNotFound(err) {\n\t\t\terr = c.Create(ctx, cj)\n\n\t\t\tif err == nil {\n\t\t\t\treturn nil\n\t\t\t}\n\n\t\t\treturn kverrors.Wrap(err, \"failed to create cronjob\",\n\t\t\t\t\"name\", cj.Name,\n\t\t\t\t\"namespace\", cj.Namespace,\n\t\t\t)\n\t\t}\n\n\t\treturn kverrors.Wrap(err, \"failed to get cronjob\",\n\t\t\t\"name\", cj.Name,\n\t\t\t\"namespace\", cj.Namespace,\n\t\t)\n\t}\n\n\tif !equal(current, cj) {\n\t\terr := retry.RetryOnConflict(retry.DefaultRetry, func() error {\n\t\t\tif err := c.Get(ctx, key, current); err != nil {\n\t\t\t\treturn kverrors.Wrap(err, \"failed to get cronjob\",\n\t\t\t\t\t\"name\", cj.Name,\n\t\t\t\t\t\"namespace\", cj.Namespace,\n\t\t\t\t)\n\t\t\t}\n\n\t\t\tmutate(current, cj)\n\t\t\tif err := c.Update(ctx, current); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\t\tif err != nil {\n\t\t\treturn kverrors.Wrap(err, \"failed to update cronjob\",\n\t\t\t\t\"name\", cj.Name,\n\t\t\t\t\"namespace\", cj.Namespace,\n\t\t\t)\n\t\t}\n\t\treturn nil\n\t}\n\n\treturn nil\n}", "func (p *Patcher) Patch(c []byte) error {\n\tb, err := read(p.r)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar a interface{}\n\n\tif err := jsoniter.Unmarshal(c, &a); err != nil {\n\t\treturn err\n\t}\n\n\treturn p.write(a, b)\n}", "func NewPatch() *Patch {\n\treturn &Patch{\n\t\tpatchOps: []*patchOp{},\n\t}\n}", "func (r *BrokerResource) GetPatch(owner, slug string, id int) (string, error) {\n\tdata := []byte{}\n\t// uri, err := url.Parse(\"https://api.bitbucket.org/1.0\" + path)\n\t// https://bitbucket.org/!api/2.0/repositories/tdburke/test_mymysql/pullrequests/1/patch\n\n\tpath := fmt.Sprintf(\"/repositories/tdburke/test_mymysql/pullrequests/1/patch\")\n\n\tfmt.Println(path)\n\n\tif err := r.client.do(\"GET\", path, nil, nil, &data); err != nil {\n\t\tfmt.Println(\"Get error:\", err)\n\t\treturn \"\", err\n\t}\n\n\tfmt.Println(data)\n\n\tif len(data) == 0 {\n\t\treturn \"\", ErrNotFound\n\t}\n\n\treturn \"\", nil\n}", "func (p *Patch) Patch() error {\n\tif p == nil {\n\t\treturn errors.New(\"patch is nil\")\n\t}\n\tif err := isPatchable(p.target, p.redirection); err != nil {\n\t\treturn err\n\t}\n\tif err := applyPatch(p); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (fkw *FakeClientWrapper) Patch(ctx context.Context, obj runtime.Object,\n\tpatch k8sCl.Patch, opts ...k8sCl.PatchOption) error {\n\treturn fkw.client.Patch(ctx, obj, patch, opts...)\n}", "func (c *kaosRules) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.KaosRule, err error) {\n\tresult = &v1.KaosRule{}\n\terr = c.client.Patch(pt).\n\t\tNamespace(c.ns).\n\t\tResource(\"kaosrules\").\n\t\tSubResource(subresources...).\n\t\tName(name).\n\t\tBody(data).\n\t\tDo().\n\t\tInto(result)\n\treturn\n}", "func (c *customReplicationControllers) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *customrc.CustomReplicationController, err error) {\n\tresult = &customrc.CustomReplicationController{}\n\terr = c.client.Patch(pt).\n\t\tNamespace(c.ns).\n\t\tResource(\"customreplicationcontrollers\").\n\t\tSubResource(subresources...).\n\t\tName(name).\n\t\tBody(data).\n\t\tDo().\n\t\tInto(result)\n\treturn\n}", "func (configurationPatch *ConfigurationPatch) AsPatch() (_patch map[string]interface{}, err error) {\n\tvar jsonData []byte\n\tjsonData, err = json.Marshal(configurationPatch)\n\tif err == nil {\n\t\terr = json.Unmarshal(jsonData, &_patch)\n\t}\n\treturn\n}", "func (s *Nap) Patch(pathURL string) *Nap {\n\ts.method = MethodPatch\n\treturn s.Path(pathURL)\n}", "func (c *Client) CancelPatchJob(ctx context.Context, req *osconfigpb.CancelPatchJobRequest, opts ...gax.CallOption) (*osconfigpb.PatchJob, error) {\n\treturn c.internalClient.CancelPatchJob(ctx, req, opts...)\n}", "func CronJobObjectWrapper(create CronJobCreator) ObjectCreator {\n\treturn func(existing runtime.Object) (runtime.Object, error) {\n\t\tif existing != nil {\n\t\t\treturn create(existing.(*batchv1beta1.CronJob))\n\t\t}\n\t\treturn create(&batchv1beta1.CronJob{})\n\t}\n}", "func (c *FakeTZCronJobs) Get(name string, options v1.GetOptions) (result *v1alpha1.TZCronJob, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewGetAction(tzcronjobsResource, c.ns, name), &v1alpha1.TZCronJob{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\treturn obj.(*v1alpha1.TZCronJob), err\n}", "func (b *RepairCronJob) Reconcile(driver opsdk.Client) (sdk.Object, error) {\n\tvar err error\n\n\tb.configureDesired()\n\n\texisting := &batchv1beta1.CronJob{\n\t\tTypeMeta: GetCronJobTypeMeta(),\n\t\tObjectMeta: b.desired.ObjectMeta,\n\t}\n\terr = driver.Get(existing)\n\tif err != nil {\n\t\treturn nil, errors.New(\"could not get existing\")\n\t}\n\n\tif existing.ResourceVersion != \"\" {\n\t\t// here we have one that is existing and one that is expected\n\t\t// we put our code here to reconcile the two and return\n\t\t// the reconciled object\n\t\tb.desired.ResourceVersion = existing.ResourceVersion\n\t\terr = driver.Update(b.desired)\n\t\treturn b.desired, err\n\t}\n\n\terr = driver.Create(b.desired)\n\treturn b.desired, err\n}", "func (k *kubeclient) Patch(name string, pt types.PatchType,\n\tpatchObj []byte) (*apis.UpgradeResult, error) {\n\tif strings.TrimSpace(name) == \"\" {\n\t\treturn nil, errors.New(\"failed to patch upgrade result: missing upgradeResult name\")\n\t}\n\tcs, err := k.getClientOrCached()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn k.patch(cs, name, pt, patchObj, k.namespace)\n}", "func (r *versionResolver) Patch(ctx context.Context, obj *restModel.APIVersion) (*restModel.APIPatch, error) {\n\tif !evergreen.IsPatchRequester(*obj.Requester) {\n\t\treturn nil, nil\n\t}\n\tapiPatch, err := data.FindPatchById(*obj.Id)\n\tif err != nil {\n\t\treturn nil, InternalServerError.Send(ctx, fmt.Sprintf(\"Couldn't find a patch with id '%s': %s\", *obj.Id, err.Error()))\n\t}\n\treturn apiPatch, nil\n}", "func (iface *Iface) patch(dev Patchable) {\n\tiface.patched = dev\n}", "func (r *ProjectsPatchPoliciesService) Patch(name string, patchpolicy *PatchPolicy) *ProjectsPatchPoliciesPatchCall {\n\tc := &ProjectsPatchPoliciesPatchCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.name = name\n\tc.patchpolicy = patchpolicy\n\treturn c\n}", "func (w *ClusterDynamicClient) Patch(name string, pt types.PatchType, data []byte, options metav1.PatchOptions, subresources ...string) (*unstructured.Unstructured, error) {\n\treturn w.dClient.Resource(w.resource).Namespace(w.namespace).Patch(w.ctx, name, pt, data, options, subresources...)\n}", "func (a *APITest) Patch(url string) *Request {\n\ta.request.method = http.MethodPatch\n\ta.request.url = url\n\treturn a.request\n}", "func (m *PrintersItemJobsPrintJobItemRequestBuilder) ToPatchRequestInformation(ctx context.Context, body iadcd81124412c61e647227ecfc4449d8bba17de0380ddda76f641a29edf2b242.PrintJobable, requestConfiguration *PrintersItemJobsPrintJobItemRequestBuilderPatchRequestConfiguration)(*i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.RequestInformation, error) {\n requestInfo := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.NewRequestInformation()\n requestInfo.UrlTemplate = m.BaseRequestBuilder.UrlTemplate\n requestInfo.PathParameters = m.BaseRequestBuilder.PathParameters\n requestInfo.Method = i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.PATCH\n requestInfo.Headers.Add(\"Accept\", \"application/json\")\n err := requestInfo.SetContentFromParsable(ctx, m.BaseRequestBuilder.RequestAdapter, \"application/json\", body)\n if err != nil {\n return nil, err\n }\n if requestConfiguration != nil {\n requestInfo.Headers.AddAll(requestConfiguration.Headers)\n requestInfo.AddRequestOptions(requestConfiguration.Options)\n }\n return requestInfo, nil\n}", "func (p *Patcher) Patch(ctx context.Context, name string, patch duck.JSONPatch) (*v1alpha1.AWSSNSSource, error) {\n\tjsonPatch, err := json.Marshal(patch)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"applying JSON patch: %w\", err)\n\t}\n\n\topts := metav1.PatchOptions{\n\t\tFieldManager: p.component,\n\t}\n\n\treturn p.cli.Patch(ctx, name, types.JSONPatchType, jsonPatch, opts, \"status\")\n}", "func (c *Client) ListPatchJobs(ctx context.Context, req *osconfigpb.ListPatchJobsRequest, opts ...gax.CallOption) *PatchJobIterator {\n\treturn c.internalClient.ListPatchJobs(ctx, req, opts...)\n}", "func (c *FakeRedisTriggers) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts v1.PatchOptions, subresources ...string) (result *v1beta1.RedisTrigger, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewPatchSubresourceAction(redistriggersResource, c.ns, name, pt, data, subresources...), &v1beta1.RedisTrigger{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\treturn obj.(*v1beta1.RedisTrigger), err\n}", "func (c *Client) patch(rawURL string, authenticate bool, expectedStatus int, in interface{}, out interface{}) error {\n\terr := c.do(rawURL, \"PATCH\", authenticate, expectedStatus, in, out)\n\treturn errio.Error(err)\n}", "func (secretGroupPatch *SecretGroupPatch) AsPatch() (_patch map[string]interface{}, err error) {\n\tvar jsonData []byte\n\tjsonData, err = json.Marshal(secretGroupPatch)\n\tif err == nil {\n\t\terr = json.Unmarshal(jsonData, &_patch)\n\t}\n\treturn\n}", "func patch(c client.Client, opCond *operators.OperatorCondition, newCond meta.Condition) error {\n\tnewCond.LastTransitionTime = meta.Now()\n\tpatchData, err := json.Marshal([]*patcher.JSONPatch{\n\t\tpatcher.NewJSONPatch(\"add\", \"/spec/conditions\", []meta.Condition{newCond})})\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to generate patch request body for Condition %v: %w\", newCond, err)\n\t}\n\tif err = c.Patch(context.TODO(), opCond, client.RawPatch(types.JSONPatchType, patchData)); err != nil {\n\t\treturn fmt.Errorf(\"unable to apply patch %s to OperatorCondition %s: %w\", patchData, opCond.GetName(), err)\n\t}\n\treturn nil\n}", "func (m *ZebraFotaConnectorRequestBuilder) Patch(ctx context.Context, body ie233ee762e29b4ba6970aa2a2efce4b7fde11697ca9ea81099d0f8269309c1be.ZebraFotaConnectorable, requestConfiguration *ZebraFotaConnectorRequestBuilderPatchRequestConfiguration)(ie233ee762e29b4ba6970aa2a2efce4b7fde11697ca9ea81099d0f8269309c1be.ZebraFotaConnectorable, error) {\n requestInfo, err := m.ToPatchRequestInformation(ctx, body, requestConfiguration);\n if err != nil {\n return nil, err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n res, err := m.BaseRequestBuilder.RequestAdapter.Send(ctx, requestInfo, ie233ee762e29b4ba6970aa2a2efce4b7fde11697ca9ea81099d0f8269309c1be.CreateZebraFotaConnectorFromDiscriminatorValue, errorMapping)\n if err != nil {\n return nil, err\n }\n if res == nil {\n return nil, nil\n }\n return res.(ie233ee762e29b4ba6970aa2a2efce4b7fde11697ca9ea81099d0f8269309c1be.ZebraFotaConnectorable), nil\n}", "func NewJobReconciler(ctx context.Context, config *config.Config, mgr manager.Manager, podLogGetter PodLogGetter) (reconcile.Reconciler, error) {\n\tversionedSecretStore := versionedsecretstore.NewVersionedSecretStore(mgr.GetClient())\n\n\treturn &ReconcileJob{\n\t\tctx: ctx,\n\t\tconfig: config,\n\t\tclient: mgr.GetClient(),\n\t\tpodLogGetter: podLogGetter,\n\t\tscheme: mgr.GetScheme(),\n\t\tversionedSecretStore: versionedSecretStore,\n\t}, nil\n}", "func (c *concurrencyControls) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts v1.PatchOptions, subresources ...string) (result *v1alpha1.ConcurrencyControl, err error) {\n\tresult = &v1alpha1.ConcurrencyControl{}\n\terr = c.client.Patch(pt).\n\t\tNamespace(c.ns).\n\t\tResource(\"concurrencycontrols\").\n\t\tName(name).\n\t\tSubResource(subresources...).\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tBody(data).\n\t\tDo(ctx).\n\t\tInto(result)\n\treturn\n}", "func (akp *AKPatches) ApprovePatch(ctx context.Context, params interface{}) ([]byte, error) {\n\tpostData, err := json.Marshal(params)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\trequest, err := http.NewRequest(\"POST\", akp.client.Server+\"/api/v1.0/AKPatches.ApprovePatch\",\n\t\tbytes.NewBuffer(postData))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\traw, err := akp.client.Do(ctx, request, nil)\n\treturn raw, err\n}", "func (j *Job) Job() JobFunc {\n\treturn j.job\n}", "func (o HttpRuleResponseOutput) Patch() pulumi.StringOutput {\n\treturn o.ApplyT(func(v HttpRuleResponse) string { return v.Patch }).(pulumi.StringOutput)\n}", "func (c *Client) Patch(rawurl string, in, out interface{}) error {\n\treturn c.Do(rawurl, \"PATCH\", in, out)\n}", "func NewRepairCronJob(cc *v1alpha1.CassandraCluster) *RepairCronJob {\n\treturn &RepairCronJob{\n\t\tcluster: cc,\n\t}\n}", "func (c *FakeTZCronJobs) Watch(opts v1.ListOptions) (watch.Interface, error) {\n\treturn c.Fake.\n\t\tInvokesWatch(testing.NewWatchAction(tzcronjobsResource, c.ns, opts))\n\n}", "func newCronJobForCR(cr *v1alpha1.Copybird) *v1beta1.CronJob {\n\tlabels := map[string]string{\n\t\t\"app\": cr.Name,\n\t}\n\treturn &v1beta1.CronJob{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tName: cr.Name + \"-cronjob\",\n\t\t\tNamespace: cr.Namespace,\n\t\t\tLabels: labels,\n\t\t},\n\t\tSpec: v1beta1.CronJobSpec{\n\t\t\tSchedule: cr.Spec.Cron,\n\t\t\tJobTemplate: v1beta1.JobTemplateSpec{\n\t\t\t\tSpec: batchv1.JobSpec{\n\t\t\t\t\tTemplate: v1.PodTemplateSpec{\n\t\t\t\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\t\t\t\tName: cr.Name + \"-copybird\",\n\t\t\t\t\t\t\tNamespace: cr.Namespace,\n\t\t\t\t\t\t\tLabels: labels,\n\t\t\t\t\t\t},\n\t\t\t\t\t\tSpec: v1.PodSpec{\n\t\t\t\t\t\t\tContainers: []v1.Container{\n\t\t\t\t\t\t\t\t{\n\t\t\t\t\t\t\t\t\tName: cr.Name,\n\t\t\t\t\t\t\t\t\tImage: \"copybird/copybird\",\n\t\t\t\t\t\t\t\t\tCommand: []string{},\n\t\t\t\t\t\t\t\t\tArgs: []string{},\n\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\tRestartPolicy: \"OnFailure\",\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n}", "func (c *solrBackups) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1beta1.SolrBackup, err error) {\n\tresult = &v1beta1.SolrBackup{}\n\terr = c.client.Patch(pt).\n\t\tNamespace(c.ns).\n\t\tResource(\"solrbackups\").\n\t\tSubResource(subresources...).\n\t\tName(name).\n\t\tBody(data).\n\t\tDo().\n\t\tInto(result)\n\treturn\n}", "func (s *gatewayClient) Patch(o *v1alpha3.Gateway, patchType types.PatchType, data []byte, subresources ...string) (*v1alpha3.Gateway, error) {\n\tobj, err := s.objectClient.Patch(o.Name, o, patchType, data, subresources...)\n\treturn obj.(*v1alpha3.Gateway), err\n}", "func (f5 *f5LTM) patch(url string, payload interface{}, result interface{}) error {\n\treturn f5.restRequestPayload(\"PATCH\", url, payload, result)\n}", "func (c *restClient) ListPatchJobs(ctx context.Context, req *osconfigpb.ListPatchJobsRequest, opts ...gax.CallOption) *PatchJobIterator {\n\tit := &PatchJobIterator{}\n\treq = proto.Clone(req).(*osconfigpb.ListPatchJobsRequest)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tit.InternalFetch = func(pageSize int, pageToken string) ([]*osconfigpb.PatchJob, string, error) {\n\t\tresp := &osconfigpb.ListPatchJobsResponse{}\n\t\tif pageToken != \"\" {\n\t\t\treq.PageToken = pageToken\n\t\t}\n\t\tif pageSize > math.MaxInt32 {\n\t\t\treq.PageSize = math.MaxInt32\n\t\t} else if pageSize != 0 {\n\t\t\treq.PageSize = int32(pageSize)\n\t\t}\n\t\tbaseUrl, err := url.Parse(c.endpoint)\n\t\tif err != nil {\n\t\t\treturn nil, \"\", err\n\t\t}\n\t\tbaseUrl.Path += fmt.Sprintf(\"/v1beta/%v/patchJobs\", req.GetParent())\n\n\t\tparams := url.Values{}\n\t\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\t\tif req.GetFilter() != \"\" {\n\t\t\tparams.Add(\"filter\", fmt.Sprintf(\"%v\", req.GetFilter()))\n\t\t}\n\t\tif req.GetPageSize() != 0 {\n\t\t\tparams.Add(\"pageSize\", fmt.Sprintf(\"%v\", req.GetPageSize()))\n\t\t}\n\t\tif req.GetPageToken() != \"\" {\n\t\t\tparams.Add(\"pageToken\", fmt.Sprintf(\"%v\", req.GetPageToken()))\n\t\t}\n\n\t\tbaseUrl.RawQuery = params.Encode()\n\n\t\t// Build HTTP headers from client and context metadata.\n\t\thds := append(c.xGoogHeaders, \"Content-Type\", \"application/json\")\n\t\theaders := gax.BuildHeaders(ctx, hds...)\n\t\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\t\tif settings.Path != \"\" {\n\t\t\t\tbaseUrl.Path = settings.Path\n\t\t\t}\n\t\t\thttpReq, err := http.NewRequest(\"GET\", baseUrl.String(), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\thttpReq.Header = headers\n\n\t\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tdefer httpRsp.Body.Close()\n\n\t\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn nil\n\t\t}, opts...)\n\t\tif e != nil {\n\t\t\treturn nil, \"\", e\n\t\t}\n\t\tit.Response = resp\n\t\treturn resp.GetPatchJobs(), resp.GetNextPageToken(), nil\n\t}\n\n\tfetch := func(pageSize int, pageToken string) (string, error) {\n\t\titems, nextPageToken, err := it.InternalFetch(pageSize, pageToken)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tit.items = append(it.items, items...)\n\t\treturn nextPageToken, nil\n\t}\n\n\tit.pageInfo, it.nextFunc = iterator.NewPageInfo(fetch, it.bufLen, it.takeBuf)\n\tit.pageInfo.MaxSize = int(req.GetPageSize())\n\tit.pageInfo.Token = req.GetPageToken()\n\n\treturn it\n}", "func (c *kongs) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1alpha1.Kong, err error) {\n\tresult = &v1alpha1.Kong{}\n\terr = c.client.Patch(pt).\n\t\tNamespace(c.ns).\n\t\tResource(\"kongs\").\n\t\tSubResource(subresources...).\n\t\tName(name).\n\t\tBody(data).\n\t\tDo().\n\t\tInto(result)\n\treturn\n}", "func (client MockStatusClient) Patch(context ctx.Context, object ctrlClient.Object, patch ctrlClient.Patch, options ...ctrlClient.PatchOption) error {\n\treturn fmt.Errorf(\"not implemented\")\n}", "func (conn Connection) Patch(cmd string, content, result interface{}) (resp *http.Response, err error) {\n\treturn conn.Send(http.MethodPatch, cmd, content, result)\n}", "func Patch(t testing.TB, dest, value interface{}) {\n\tNew(t).Patch(dest, value)\n}", "func (res *Resource) Patch(storage store.Update) {\n\tres.HandleFuncC(\n\t\tpat.Patch(patID),\n\t\tfunc(ctx context.Context, w http.ResponseWriter, r *http.Request) {\n\t\t\tres.patchHandler(ctx, w, r, storage)\n\t\t},\n\t)\n\n\tres.addRoute(patch, patID)\n}", "func (secretMetadataPatch *SecretMetadataPatch) AsPatch() (_patch map[string]interface{}, err error) {\n\tvar jsonData []byte\n\tjsonData, err = json.Marshal(secretMetadataPatch)\n\tif err == nil {\n\t\terr = json.Unmarshal(jsonData, &_patch)\n\t}\n\treturn\n}", "func (c *interacts) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.Interact, err error) {\n\tresult = &v1.Interact{}\n\terr = c.client.Patch(pt).\n\t\tNamespace(c.ns).\n\t\tResource(\"interacts\").\n\t\tSubResource(subresources...).\n\t\tName(name).\n\t\tBody(data).\n\t\tDo().\n\t\tInto(result)\n\treturn\n}", "func (mr *MockJobClientMockRecorder) Patch(arg0, arg1, arg2, arg3 interface{}, arg4 ...interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\tvarargs := append([]interface{}{arg0, arg1, arg2, arg3}, arg4...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"Patch\", reflect.TypeOf((*MockJobClient)(nil).Patch), varargs...)\n}", "func RawPatch(patchType types.PatchType, data []byte) Patch {\n\treturn &patch{patchType, data}\n}", "func (secretVersionMetadataPatch *SecretVersionMetadataPatch) AsPatch() (_patch map[string]interface{}, err error) {\n\tvar jsonData []byte\n\tjsonData, err = json.Marshal(secretVersionMetadataPatch)\n\tif err == nil {\n\t\terr = json.Unmarshal(jsonData, &_patch)\n\t}\n\treturn\n}", "func (m *ThreatSubmissionEmailThreatsEmailThreatSubmissionItemRequestBuilder) Patch(ctx context.Context, body i084fa7ab3bba802bf5cc3b408e230cc64c167a57976e0d42c37e17154afd5b78.EmailThreatSubmissionable, requestConfiguration *ThreatSubmissionEmailThreatsEmailThreatSubmissionItemRequestBuilderPatchRequestConfiguration)(i084fa7ab3bba802bf5cc3b408e230cc64c167a57976e0d42c37e17154afd5b78.EmailThreatSubmissionable, error) {\n requestInfo, err := m.ToPatchRequestInformation(ctx, body, requestConfiguration);\n if err != nil {\n return nil, err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n res, err := m.BaseRequestBuilder.RequestAdapter.Send(ctx, requestInfo, i084fa7ab3bba802bf5cc3b408e230cc64c167a57976e0d42c37e17154afd5b78.CreateEmailThreatSubmissionFromDiscriminatorValue, errorMapping)\n if err != nil {\n return nil, err\n }\n if res == nil {\n return nil, nil\n }\n return res.(i084fa7ab3bba802bf5cc3b408e230cc64c167a57976e0d42c37e17154afd5b78.EmailThreatSubmissionable), nil\n}", "func (conn Connection) Patch(cmd string, content, result interface{}) (effect *SideEffect, resp *http.Response, err error) {\n\treturn conn.Send(http.MethodPatch, cmd, content, result)\n}", "func (m *FeatureRolloutPolicyItemRequestBuilder) Patch(ctx context.Context, body iadcd81124412c61e647227ecfc4449d8bba17de0380ddda76f641a29edf2b242.FeatureRolloutPolicyable, requestConfiguration *FeatureRolloutPolicyItemRequestBuilderPatchRequestConfiguration)(iadcd81124412c61e647227ecfc4449d8bba17de0380ddda76f641a29edf2b242.FeatureRolloutPolicyable, error) {\n requestInfo, err := m.CreatePatchRequestInformation(ctx, body, requestConfiguration);\n if err != nil {\n return nil, err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": ia572726a95efa92ddd544552cd950653dc691023836923576b2f4bf716cf204a.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": ia572726a95efa92ddd544552cd950653dc691023836923576b2f4bf716cf204a.CreateODataErrorFromDiscriminatorValue,\n }\n res, err := m.requestAdapter.SendAsync(ctx, requestInfo, iadcd81124412c61e647227ecfc4449d8bba17de0380ddda76f641a29edf2b242.CreateFeatureRolloutPolicyFromDiscriminatorValue, errorMapping)\n if err != nil {\n return nil, err\n }\n if res == nil {\n return nil, nil\n }\n return res.(iadcd81124412c61e647227ecfc4449d8bba17de0380ddda76f641a29edf2b242.FeatureRolloutPolicyable), nil\n}", "func (c *FakeKuberaBackups) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts v1.PatchOptions, subresources ...string) (result *backuprestorev1.KuberaBackup, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewPatchSubresourceAction(kuberabackupsResource, c.ns, name, pt, data, subresources...), &backuprestorev1.KuberaBackup{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\treturn obj.(*backuprestorev1.KuberaBackup), err\n}", "func (c *meshPolicies) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1alpha1.MeshPolicy, err error) {\n\tresult = &v1alpha1.MeshPolicy{}\n\terr = c.client.Patch(pt).\n\t\tResource(\"meshpolicies\").\n\t\tSubResource(subresources...).\n\t\tName(name).\n\t\tBody(data).\n\t\tDo().\n\t\tInto(result)\n\treturn\n}", "func (c *FakeProjects) Patch(name string, pt pkg_api.PatchType, data []byte, subresources ...string) (result *api.Project, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(core.NewPatchSubresourceAction(projectsResource, c.ns, name, data, subresources...), &api.Project{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\treturn obj.(*api.Project), err\n}", "func (c *klusterlets) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.Klusterlet, err error) {\n\tresult = &v1.Klusterlet{}\n\terr = c.client.Patch(pt).\n\t\tResource(\"klusterlets\").\n\t\tName(name).\n\t\tSubResource(subresources...).\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tBody(data).\n\t\tDo(ctx).\n\t\tInto(result)\n\treturn\n}", "func (r *ProjectsPatchJobsService) Get(name string) *ProjectsPatchJobsGetCall {\n\tc := &ProjectsPatchJobsGetCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.name = name\n\treturn c\n}", "func (client *MockClient) Patch(context ctx.Context, object ctrlClient.Object, patch ctrlClient.Patch, options ...ctrlClient.PatchOption) error {\n\treturn fmt.Errorf(\"Not implemented\")\n}", "func (r *CronCrawlerJobReconciler) Reconcile(req ctrl.Request) (ctrl.Result, error) {\n\tctx := context.Background()\n\tlog := r.Log.WithValues(\"croncrawlerjob\", req.NamespacedName)\n\tinstance := &cronjobv1.CronCrawlerJob{}\n\n\t// 1 判断自定义资源是否存在\n\terr := r.Get(ctx, req.NamespacedName, instance)\n\tif err != nil {\n\t\tif errors.IsNotFound(err) {\n\t\t\t// Request object not found, could have been deleted after ctrl request.\n\t\t\t// Owned objects are automatically garbage collected. For additional cleanup logic use finalizers.\n\t\t\t// Return and don't requeue\n\t\t\treturn ctrl.Result{}, nil\n\t\t}\n\t\t// Error reading the object - requeue the request.\n\t\tlog.Error(err, \"Failed to get CronCrawlerJob\")\n\t\treturn ctrl.Result{}, err\n\t}\n\n\tif instance.DeletionTimestamp != nil {\n\t\treturn ctrl.Result{}, err\n\t}\n\n\tjobType := instance.Spec.JobType\n\tschedule := instance.Spec.Schedule\n\n\t// 采集类型为定时任务时,必须指定采集频率\n\tif (jobType == \"cronjob\") && (schedule == \"\") {\n\t\tlog.Error(err, \"`schedule` must be confirmed in a cronjob jobtype!\")\n\t\treturn ctrl.Result{}, err\n\t}\n\n\t// 2 自定义资源已经存在,判断新旧任务类型是否一致,\n\t// 2.1 如果不一致一律删除,注意是普通任务也直接删除,\n\t// 2.2 如果一致,也就是定时任务了,直接更新\n\n\tnewSpec := CrawlerUpdatedData{\n\t\tJobType: instance.Spec.JobType,\n\t\tSchedule: instance.Spec.Schedule,\n\t\tEnv: instance.Spec.Env,\n\t\tDuration: instance.Spec.Duration,\n\t\tPublisherTTL: instance.Spec.Publisher.PublisherTTL,\n\t\tConsumerTTL: instance.Spec.Consumer.ConsumerTTL,\n\t\tParallelism: instance.Spec.Consumer.Parallelism,\n\t}\n\n\tduration, _ := time.ParseDuration(instance.Spec.Duration)\n\tns := instance.Namespace\n\n\tlog.Info(\"instance ns: \" + ns + \", req ns: \" + req.NamespacedName.Namespace)\n\tname := instance.Name\n\tshortDes := ns + \"/\" + name\n\tpublisherName := NameForCronCrawlerJob(name, \"publisher\", jobType)\n\tconsumerName := NameForCronCrawlerJob(name, \"consumer\", jobType)\n\n\t// 判断是不是旧任务入队\n\tif (instance.Annotations != nil) && (instance.Annotations[\"status\"] == \"old\") {\n\t\t// 反序列化\n\t\toldSpec := CrawlerUpdatedData{}\n\t\terr = json.Unmarshal([]byte(instance.Annotations[\"spec\"]), &oldSpec)\n\t\tif err != nil {\n\t\t\treturn ctrl.Result{}, nil\n\t\t}\n\t\t// 判断是否有更新\n\t\tif !reflect.DeepEqual(oldSpec, newSpec) {\n\t\t\t// job type\n\t\t\toldJobType := oldSpec.JobType\n\t\t\tif (jobType != oldJobType) || (oldJobType == \"job\") {\n\t\t\t\t// deletes the older resource\n\t\t\t\tif err := r.DeleteOldJobsByLabels(ns, name, oldJobType); err != nil {\n\t\t\t\t\tlog.Error(err, \"Older \"+oldJobType+\" for \"+shortDes+\" deletes failed\")\n\t\t\t\t\treturn ctrl.Result{}, err\n\t\t\t\t}\n\t\t\t\tlog.Info(\"Older \" + oldJobType + \" for \" + shortDes + \" deletes success\")\n\t\t\t\t// resets the current resource\n\t\t\t\tinstance.Annotations = nil\n\t\t\t\tif err := r.Update(ctx, instance); err != nil {\n\t\t\t\t\treturn ctrl.Result{}, err\n\t\t\t\t}\n\t\t\t\tlog.Info(\"Current \" + jobType + \"'s annotation sets to nil, would be requeued later\")\n\t\t\t\treturn ctrl.Result{Requeue: true, RequeueAfter: duration}, nil\n\t\t\t} else if jobType == \"cronjob\" {\n\t\t\t\t// 更新cronjob,并返回\n\t\t\t\tif err := r.UpdateOldJobs(instance, ns, jobType, publisherName, consumerName); err != nil {\n\t\t\t\t\tlog.Error(err, \"Older \"+jobType+\" for \"+shortDes+\" updates failed\")\n\t\t\t\t\treturn ctrl.Result{}, err\n\t\t\t\t}\n\t\t\t\tlog.Info(\"Older \" + oldJobType + \" for \" + shortDes + \" updates success\")\n\t\t\t\treturn ctrl.Result{}, err\n\n\t\t\t}\n\n\t\t}\n\t}\n\n\t// 3、使用当前instance创建新资源并关联数据返回,只有当获取不到资源时候,才创建资源\n\n\tvar jobPublisher runtime.Object\n\tvar jobConsumers runtime.Object\n\n\tif jobType == \"job\" {\n\t\tjobPublisher = r.NewPublisherJob(instance, publisherName)\n\t\tjobConsumers = r.NewConsumersJob(instance, consumerName)\n\t} else if jobType == \"cronjob\" {\n\t\tjobPublisher = r.NewPublisherCronJob(instance, publisherName)\n\t\tjobConsumers = r.NewConsumersCronJob(instance, consumerName)\n\t} else {\n\t\tlog.Error(err, \"JobType sets wrong\")\n\t\treturn ctrl.Result{}, nil\n\t}\n\n\tif err := r.Get(ctx, types.NamespacedName{Name: publisherName, Namespace: ns}, jobPublisher); err != nil {\n\t\tlog.Info(\"Starting to create \" + jobType + \" for publisher\")\n\t\tif err := r.Create(ctx, jobPublisher); err != nil {\n\t\t\tlog.Error(err, jobType+\" for publisher creates failed\")\n\t\t\treturn ctrl.Result{}, err\n\t\t}\n\t\treturn ctrl.Result{Requeue: true}, nil\n\n\t}\n\tif err := r.Get(ctx, types.NamespacedName{Name: consumerName, Namespace: ns}, jobConsumers); err != nil {\n\t\tlog.Info(\"Starting to create \" + jobType + \" for consumers\")\n\t\tif err := r.Create(ctx, jobConsumers); err != nil {\n\t\t\tlog.Error(err, jobType+\" for Consumers creates failed\")\n\t\t\treturn ctrl.Result{}, err\n\t\t}\n\t\treturn ctrl.Result{Requeue: true}, nil\n\t}\n\n\t// 关联 Annotations\n\tdata, _ := json.Marshal(&newSpec)\n\tif instance.Annotations != nil {\n\t\tinstance.Annotations[\"spec\"] = string(data)\n\t\tinstance.Annotations[\"status\"] = \"old\"\n\t} else {\n\t\tinstance.Annotations = map[string]string{\"spec\": string(data), \"status\": \"new\"}\n\t}\n\n\tif err := r.Update(ctx, instance); err != nil {\n\t\treturn ctrl.Result{}, nil\n\t}\n\n\t// 更新显示状态\n\tpodList := &corev1.PodList{}\n\tlabels := map[string]string{\"crawler-table\": name, \"jobType\": jobType}\n\tlistOption := []client.ListOption{\n\t\tclient.InNamespace(ns),\n\t\tclient.MatchingLabels(labels),\n\t}\n\tif err := r.List(ctx, podList, listOption...); err != nil {\n\t\treturn ctrl.Result{}, nil\n\t}\n\n\tcronCrawlerJobStatus := getPodStatus(podList.Items)\n\tif !reflect.DeepEqual(cronCrawlerJobStatus, instance.Status) {\n\t\tlog.Info(\"Starting to update status\")\n\t\tinstance.Status = *cronCrawlerJobStatus\n\t\terr := r.Status().Update(ctx, instance)\n\t\tif err != nil {\n\t\t\tlog.Error(err, \"Failed to update CronCrawlerJob status\")\n\t\t\treturn ctrl.Result{}, err\n\t\t}\n\t}\n\n\treturn ctrl.Result{}, nil\n}", "func (p *Permission) GetPatch() (map[string]interface{}, error) {\n\tpatch := make(map[string]interface{})\n\n\tif p.Owner != \"\" {\n\t\tpatch[\"owner\"] = p.Owner\n\t}\n\tif p.Username != \"\" {\n\t\treturn nil, errors.NewUnsupportedPatchError(\"permission\", \"username\")\n\t}\n\tif p.Password != \"\" {\n\t\treturn nil, errors.NewUnsupportedPatchError(\"permission\", \"password\")\n\t}\n\tif p.Creator != \"\" {\n\t\treturn nil, errors.NewUnsupportedPatchError(\"permission\", \"creator\")\n\t}\n\tif p.Categories != nil {\n\t\tpatch[\"categories\"] = p.Categories\n\t\tif p.ACLs != nil {\n\t\t\tif err := p.ValidateACLs(p.ACLs...); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tpatch[\"acls\"] = p.ACLs\n\t\t} else {\n\t\t\tpatch[\"acls\"] = category.ACLsFor(p.Categories...)\n\t\t}\n\t}\n\tif p.Ops != nil {\n\t\tpatch[\"ops\"] = p.Ops\n\t}\n\tif p.Indices != nil {\n\t\tpatch[\"indices\"] = p.Indices\n\t}\n\tif p.Sources != nil {\n\t\tif err := validateSources(p.Sources); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tpatch[\"sources\"] = p.Sources\n\t}\n\tif p.Referers != nil {\n\t\tif err := validateReferers(p.Referers); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tpatch[\"referers\"] = p.Referers\n\t}\n\tif p.CreatedAt != \"\" {\n\t\treturn nil, errors.NewUnsupportedPatchError(\"permission\", \"created_at\")\n\t}\n\tif p.TTL.String() != \"0s\" {\n\t\tpatch[\"ttl\"] = p.TTL\n\t}\n\t// Cannot patch individual limits to 0\n\tif p.Limits != nil {\n\t\tlimits := make(map[string]interface{})\n\t\tif p.Limits.IPLimit != 0 {\n\t\t\tlimits[\"ip_limit\"] = p.Limits.IPLimit\n\t\t}\n\t\tif p.Limits.DocsLimit != 0 {\n\t\t\tlimits[\"docs_limit\"] = p.Limits.DocsLimit\n\t\t}\n\t\tif p.Limits.SearchLimit != 0 {\n\t\t\tlimits[\"search_limit\"] = p.Limits.SearchLimit\n\t\t}\n\t\tif p.Limits.IndicesLimit != 0 {\n\t\t\tlimits[\"indices_limit\"] = p.Limits.IndicesLimit\n\t\t}\n\t\tif p.Limits.CatLimit != 0 {\n\t\t\tlimits[\"cat_limit\"] = p.Limits.CatLimit\n\t\t}\n\t\tif p.Limits.ClustersLimit != 0 {\n\t\t\tlimits[\"clusters_limit\"] = p.Limits.ClustersLimit\n\t\t}\n\t\tif p.Limits.MiscLimit != 0 {\n\t\t\tlimits[\"misc_limit\"] = p.Limits.MiscLimit\n\t\t}\n\t\tpatch[\"limits\"] = limits\n\t}\n\tif p.Description != \"\" {\n\t\tpatch[\"description\"] = p.Description\n\t}\n\n\treturn patch, nil\n}", "func (patchwork *Patchwork) Patch(p func(repo github.Repository, directory string)) {\n\tpatchwork.patch = p\n}", "func (c *Client) GetPullRequestPatch(owner, repo string, index int64) ([]byte, *Response, error) {\n\treturn c.getPullRequestDiffOrPatch(owner, repo, \"patch\", index)\n}" ]
[ "0.65886015", "0.6050717", "0.5993012", "0.596517", "0.5955357", "0.5934546", "0.590108", "0.5899239", "0.58721954", "0.58264655", "0.58226603", "0.5789603", "0.5764795", "0.56989354", "0.56664795", "0.5655391", "0.5622445", "0.55950445", "0.5582619", "0.5572436", "0.55122024", "0.5509819", "0.5493515", "0.5467452", "0.54611623", "0.5452704", "0.54519206", "0.54488164", "0.54329014", "0.542342", "0.53558403", "0.53411984", "0.53188825", "0.53164566", "0.52973384", "0.5296197", "0.5291173", "0.5284976", "0.5275422", "0.52712405", "0.5254057", "0.52218664", "0.52156854", "0.5203509", "0.5200281", "0.5175431", "0.5167096", "0.5166818", "0.51624763", "0.51604414", "0.5135272", "0.5122456", "0.5110396", "0.5108944", "0.51071966", "0.5104698", "0.5100967", "0.50969064", "0.50923085", "0.5087412", "0.5081836", "0.507987", "0.5071326", "0.506018", "0.50506705", "0.5044912", "0.5044874", "0.50429446", "0.503574", "0.50283605", "0.50270146", "0.5019732", "0.50152475", "0.5011132", "0.4999827", "0.49974167", "0.49882483", "0.49831957", "0.49818143", "0.49811736", "0.49797258", "0.49686924", "0.49679682", "0.4967205", "0.49651557", "0.49649954", "0.49648088", "0.49632284", "0.4963003", "0.4949359", "0.49386314", "0.49369028", "0.49336937", "0.4928209", "0.49189508", "0.49185193", "0.4917954", "0.49167362", "0.49013034", "0.4897014" ]
0.7325353
0
NewHappyDevFormatter returns a new instance of HappyDevFormatter.
func NewHappyDevFormatter(name string) *HappyDevFormatter { jf := NewJSONFormatter(name) return &HappyDevFormatter{ name: name, jsonFormatter: jf, } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewFormatter() *formatter {\n\treturn &formatter{}\n}", "func NewFormatter() (*Formatter, error) {\n\tlexer, err := NewLexer()\n\n\tif err != nil {\n\t\treturn &Formatter{}, err\n\t}\n\n\treturn &Formatter{&lexer}, nil\n}", "func New() *Formatter {\n\treturn &Formatter{\n\t\tformats: map[string]Format{},\n\t}\n}", "func NewFormatter() *Formatter {\n\treturn &Formatter{\n\t\tPassColor: color.New(color.FgGreen),\n\t\tFailColor: color.New(color.FgRed),\n\t\tStringMaxLength: 0,\n\t\tDisabledColor: false,\n\t\tIndent: 2,\n\t}\n}", "func newZapDevelopmentConfig() zap.Config {\n\tc := zap.NewDevelopmentConfig()\n\t// This displays log messages in a format compatible with the zap-pretty print library\n\tc.EncoderConfig = zapcore.EncoderConfig{\n\t\tTimeKey: \"ts\",\n\t\tLevelKey: \"level\",\n\t\tNameKey: \"logger\",\n\t\tCallerKey: \"caller\",\n\t\tMessageKey: \"msg\",\n\t\tStacktraceKey: \"stacktrace\",\n\t\tLineEnding: zapcore.DefaultLineEnding,\n\t\tEncodeLevel: zapcore.CapitalColorLevelEncoder,\n\t\tEncodeTime: zapcore.EpochTimeEncoder,\n\t\tEncodeDuration: zapcore.SecondsDurationEncoder,\n\t\tEncodeCaller: zapcore.ShortCallerEncoder,\n\t}\n\treturn c\n}", "func NewFormatter(withoutUnit bool, duration time.Duration) *Formatter {\n\treturn &Formatter{withoutUnit, duration}\n}", "func NewFormatter() *Formatter {\n\treturn &Formatter{\n\t\tKeyColor: ansi.NewColor(ansi.FgBlue, ansi.Bold),\n\t\tStringColor: nil,\n\t\tBoolColor: ansi.NewColor(ansi.FgYellow, ansi.Bold),\n\t\tNumberColor: ansi.NewColor(ansi.FgMagenta),\n\t\tNullColor: ansi.NewColor(ansi.Bold),\n\t\tDisableColor: false,\n\t\tIndent: 2,\n\t}\n}", "func NewHumanizeFormatter() *Formatter {\n\treturn &Formatter{\n\t\tCompact: false,\n\t\tDateTimeFormat: \"2006-01-02T15:04:05\",\n\t\tFallback: &logrus.JSONFormatter{},\n\t}\n}", "func New(options ...Option) *Formatter {\n\tf := &Formatter{\n\t\ttabStr: []byte(\" \"),\n\t\tnewline: []byte(\"\\n\"),\n\t\ttextFormatters: func(tag Tag) TextFormatter {\n\t\t\treturn nil\n\t\t},\n\t}\n\tfor _, option := range options {\n\t\toption(f)\n\t}\n\treturn f\n}", "func NewFormatter(w io.Writer, fmt string) (*Formatter, error) {\n\tif fmt == \"\" {\n\t\tfmt = DefaultTemplate\n\t}\n\ttmpl, err := template.New(\"out\").Parse(fmt)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Formatter{\n\t\toutput: w,\n\t\ttemplate: tmpl,\n\t\tColorize: false,\n\t\tShowFields: true,\n\t\tMaxFieldLength: 30,\n\t\tShowPrefix: true,\n\t\tShowSuffix: true,\n\t\tExcludeFields: defaultExcludes,\n\t}, nil\n}", "func Formatter() *formatter {\n\treturn &formatter{}\n}", "func NewFormatter(providers map[string]Provider) Formatter {\n\treturn Formatter{\n\t\tproviders: providers,\n\t}\n}", "func (hd *HappyDevFormatter) Format(writer io.Writer, level int, msg string, args []interface{}) {\n\tbuf := pool.Get()\n\tdefer pool.Put(buf)\n\n\tif len(args) == 1 {\n\t\targs = append(args, 0)\n\t\tcopy(args[1:], args[0:])\n\t\targs[0] = singleArgKey\n\t}\n\n\t// warn about reserved, bad and complex keys\n\tfor i := 0; i < len(args); i += 2 {\n\t\tisReserved, err := isReservedKey(args[i])\n\t\tif err != nil {\n\t\t\tInternalLog.Error(\"Key is not a string.\", \"err\", fmt.Errorf(\"args[%d]=%v\", i, args[i]))\n\t\t} else if isReserved {\n\t\t\tInternalLog.Fatal(\"Key conflicts with reserved key. Avoiding using single rune keys.\", \"key\", args[i].(string))\n\t\t} else {\n\t\t\t// Ensure keys are simple strings. The JSONFormatter doesn't escape\n\t\t\t// keys as a performance tradeoff. This panics if the JSON key\n\t\t\t// value has a different value than a simple quoted string.\n\t\t\tkey, isOK := args[i].(string)\n\t\t\tif !isOK {\n\t\t\t\tpanic(\"Key is invalid or was omitted\")\n\t\t\t}\n\t\t\tfor _, aRune := range key {\n\t\t\t\tif aRune == '\"' || aRune == '\\\\' {\n\t\t\t\t\tpanic(\"Key is invalid. Use simpler key for: \" + fmt.Sprintf(\"%q\", key))\n\t\t\t\t}\n\t\t\t\tif int(aRune) < 32 {\n\t\t\t\t\tpanic(\"Key contains control character. Use simpler key for: \" + fmt.Sprintf(\"%q\", key))\n\t\t\t\t}\n\t\t\t}\n\t\t\tb, err := json.Marshal(key)\n\t\t\tif err != nil {\n\t\t\t\tpanic(\"Key is invalid. \" + err.Error())\n\t\t\t}\n\t\t\tif string(b) != `\"`+key+`\"` {\n\t\t\t\tpanic(\"Key is complex. Use simpler key for: \" + fmt.Sprintf(\"%q\", key))\n\t\t\t}\n\t\t}\n\t}\n\n\t// use the production JSON formatter to format the log first. This\n\t// ensures JSON will marshal/unmarshal correctly in production.\n\tentry := hd.jsonFormatter.LogEntry(level, msg, args)\n\n\t// reset the column tracker used for fancy formatting\n\tcol := 0\n\n\t// timestamp\n\tbuf.WriteString(theme.Misc)\n\thd.writeString(buf, entry[KeyMap.Time].(string), &col)\n\tif !disableColors {\n\t\tbuf.WriteString(ansi.Reset)\n\t}\n\n\t// emphasize warnings and errors\n\tmessage, context, color := hd.getLevelContext(level, entry)\n\tif message == \"\" {\n\t\tmessage = entry[KeyMap.Message].(string)\n\t}\n\n\t// DBG, INF ...\n\thd.set(buf, \"\", entry[KeyMap.Level].(string), color, &col)\n\t// logger name\n\thd.set(buf, \"\", entry[KeyMap.Name], theme.Misc, &col)\n\t// message from user\n\thd.set(buf, \"\", message, theme.Message, &col)\n\n\t// Preserve key order in the sequencethey were added by developer.This\n\t// makes it easier for developers to follow the log.\n\torder := []string{}\n\tlenArgs := len(args)\n\tfor i := 0; i < len(args); i += 2 {\n\t\tif i+1 >= lenArgs {\n\t\t\tcontinue\n\t\t}\n\t\tif key, ok := args[i].(string); ok {\n\t\t\torder = append(order, key)\n\t\t} else {\n\t\t\torder = append(order, badKeyAtIndex(i))\n\t\t}\n\t}\n\n\tfor _, key := range order {\n\t\t// skip reserved keys which were already added to buffer above\n\t\tisReserved, err := isReservedKey(key)\n\t\tif err != nil {\n\t\t\tpanic(\"key is invalid. Should never get here. \" + err.Error())\n\t\t} else if isReserved {\n\t\t\tcontinue\n\t\t}\n\t\thd.set(buf, key, entry[key], theme.Value, &col)\n\t}\n\n\taddLF := true\n\thasCallStack := entry[KeyMap.CallStack] != nil\n\t// WRN,ERR file, line number context\n\n\tif context != \"\" && context != \"\\n\" {\n\t\t// warnings and traces are single line, space can be optimized\n\t\tif level == LevelTrace || (level == LevelWarn && !hasCallStack) {\n\t\t\t// gets rid of \"in \"\n\t\t\tidx := strings.IndexRune(context, 'n')\n\t\t\thd.set(buf, \"in\", context[idx+2:], theme.Value, &col)\n\t\t} else {\n\t\t\tbuf.WriteRune('\\n')\n\t\t\tif !disableColors {\n\t\t\t\tbuf.WriteString(color)\n\t\t\t}\n\t\t\taddLF = context[len(context)-1:] != \"\\n\"\n\t\t\tbuf.WriteString(context)\n\t\t\tif !disableColors {\n\t\t\t\tbuf.WriteString(ansi.Reset)\n\t\t\t}\n\t\t}\n\t} else if hasCallStack {\n\t\thd.set(buf, \"\", entry[KeyMap.CallStack], color, &col)\n\t}\n\tif addLF {\n\t\tbuf.WriteRune('\\n')\n\t}\n\tbuf.WriteTo(writer)\n}", "func NewFormatter(t string) Formatter {\n\tswitch t {\n\tcase \"stdout\":\n\t\treturn newSTDOUTFormatter()\n\tcase \"plain\":\n\t\treturn newPlainFormatter()\n\tcase \"json\":\n\t\treturn newJSONFormatter()\n\tcase \"yaml\":\n\t\treturn newYamlFormatter()\n\tdefault:\n\t\treturn newSTDOUTFormatter()\n\t}\n}", "func NewHumanizeFormatterFromEnv() (logrus.Formatter, error) {\n\tdefaultFormatter := NewHumanizeFormatter()\n\n\tprovidedFormat, err := parseFormatFromEnv()\n\n\tif err != nil {\n\t\treturn defaultFormatter, err\n\t}\n\n\tswitch providedFormat {\n\tcase FormatCompact:\n\t\tdefaultFormatter.Compact = true\n\t\treturn defaultFormatter, nil\n\tcase FormatFull:\n\t\tdefaultFormatter.Compact = false\n\t\treturn defaultFormatter, nil\n\tcase FormatJSON:\n\t\treturn defaultFormatter.Fallback, nil\n\t}\n\n\treturn defaultFormatter, nil\n}", "func NewFormat(ctx context.Context, client *github.Client, debug bool) *Format {\n\treturn &Format{ctx: ctx, client: client, debug: debug}\n}", "func NewFormatter(format string, colored bool) (*Formatter, error) {\n\tfm := new(Formatter)\n\tfm.colored = colored\n\tif err := fm.SetFormat(format); err != nil {\n\t\treturn nil, err\n\t}\n\treturn fm, nil\n}", "func NewFormatter(printID bool) *Formatter {\n\tf := &Formatter{\n\t\tprintID: printID,\n\t\tw: uilive.New(),\n\t\ts: NewSpinner(),\n\t}\n\treturn f\n}", "func WithTab(tab string) Option { return func(f *Formatter) { f.tabStr = []byte(tab) } }", "func NewMockFormatter(ctrl *gomock.Controller) *MockFormatter {\n\tmock := &MockFormatter{ctrl: ctrl}\n\tmock.recorder = &MockFormatterMockRecorder{mock}\n\treturn mock\n}", "func New() *Formatter {\n\tf := &Formatter{\n\t\tLabels: map[string]string{},\n\t}\n\treturn f\n}", "func NewConsoleFormatter(color bool, opts ...OptionFormatter) *ConsoleFormatter {\n\tf := &ConsoleFormatter{\n\t\tConsoleTimestampFormat: \"2018-01-02/15:04:05\",\n\t\tConsoleLevelFormat: \"%.1s\",\n\t\tConsoleLogFormat: \"%localtime% %LEVEL% %msg% %fields%\",\n\t\tConsoleCallerFormat: \"%file%:%line% %fun%()\",\n\t\tConsoleFieldSep: \", \",\n\t\tConsoleFieldKVSep: \":\",\n\t\tConsoleFieldsWrap: \" [%s]\", // \"「%s」\"\n\t\tConsoleLogColor: color,\n\t}\n\t// call option functions on instance to set options on it\n\tfor _, opt := range opts {\n\t\topt(f)\n\t}\n\treturn f\n}", "func newPrettyStdout(w io.Writer) *prettyStdout {\n\treturn &prettyStdout{\n\t\twriter: w,\n\t\tbuffer: bytes.NewBuffer([]byte{}),\n\t}\n}", "func newInfluxFormatter() influxFormatter {\n\treturn influxFormatter{}\n}", "func WithFormatter(formatter Formatter) Option {\n\treturn option{\n\t\ttable: func(enc *TableEncoder) error {\n\t\t\tenc.formatter = formatter\n\t\t\treturn nil\n\t\t},\n\t\texpanded: func(enc *ExpandedEncoder) error {\n\t\t\tenc.formatter = formatter\n\t\t\treturn nil\n\t\t},\n\t\tjson: func(enc *JSONEncoder) error {\n\t\t\tenc.formatter = formatter\n\t\t\treturn nil\n\t\t},\n\t\tunaligned: func(enc *UnalignedEncoder) error {\n\t\t\tenc.formatter = formatter\n\t\t\treturn nil\n\t\t},\n\t\ttemplate: func(enc *TemplateEncoder) error {\n\t\t\tenc.formatter = formatter\n\t\t\treturn nil\n\t\t},\n\t\tcrosstab: func(view *CrosstabView) error {\n\t\t\tview.formatter = formatter\n\t\t\treturn nil\n\t\t},\n\t}\n}", "func NewFormatter(format string) Formatter {\n\tif fun, ok := formatMap[format]; ok {\n\t\treturn fun()\n\t}\n\treturn nil\n}", "func New() *zap.SugaredLogger {\n\tvar config zap.Config\n\tif os.Getenv(\"LOG_LEVEL\") == \"prod\" {\n\t\tconfig = zap.NewProductionConfig()\n\t} else {\n\t\tconfig = zap.NewDevelopmentConfig()\n\t\tconfig.EncoderConfig.EncodeLevel = zapcore.CapitalColorLevelEncoder\n\t}\n\tconfig.EncoderConfig.TimeKey = \"timestamp\"\n\tconfig.EncoderConfig.EncodeTime = zapcore.ISO8601TimeEncoder\n\tlogger, _ := config.Build()\n\treturn logger.Sugar()\n}", "func NewDevLogger() (*zap.Logger, error) {\n\tlogCfg := zap.NewProductionConfig()\n\tlogCfg.Level = zap.NewAtomicLevelAt(zap.DebugLevel)\n\t// logCfg.DisableStacktrace = true\n\tlogCfg.EncoderConfig = encoderCfg\n\treturn logCfg.Build()\n}", "func New() gournal.Appender {\n\treturn &appender{zap.New(zap.NewJSONEncoder())}\n}", "func NewHumanizer(tag language.Tag, options ... interface{}) Humanizer {\n return &humanizer{\n Tag: tag,\n NF: lxstrconv.NewDecimalFormat(tag),\n Printer: message.NewPrinter(tag),\n }\n}", "func WithFormatter(v logrus.Formatter) Option {\n\treturn formatterOption{f: v}\n}", "func newLogFormatter(globals server.Globals) handlers.LogFormatter {\n\treturn logFormatter{\n\t\tlogger: globals.Logger,\n\t}.format\n}", "func NewFormatter(t Type, fileName string, logger *log.Logger) Formatter {\n\tswitch t {\n\tcase StdoutType:\n\t\treturn newStdoutFormatter(logger)\n\tcase FileType:\n\t\treturn newFileFormatter(fileName, logger)\n\tcase JSONType:\n\t\treturn newJSONFormatter(fileName, logger)\n\tcase YamlType:\n\t\treturn newYamlFormatter(fileName, logger)\n\tcase SkopeoType:\n\t\treturn newSkopeoFormatter(fileName, logger)\n\tdefault:\n\t\treturn newStdoutFormatter(logger)\n\t}\n}", "func NewParser(f string) (*LogHandler, error) {\n\tif !ValidFormat(f) {\n\t\treturn nil, fmt.Errorf(\"%s is not a valid parser format\", f)\n\t}\n\n\tvar p = &LogHandler{}\n\n\tif f == RFC5424Name || f == RFC5424Standard {\n\t\tp.Parser = &parser.RFC5424{}\n\t\tp.Fmt = RFC5424Standard\n\t} else if f == WatchguardName || f == WatchguardFirebox {\n\t\tp.Parser = &parser.Watchguard{}\n\t\tp.Fmt = WatchguardFirebox\n\t}\n\n\tlog.Printf(\"input format parser created for %s\", f)\n\tp.Stats = stats.Add\n\tp.Parser.Init()\n\treturn p, nil\n}", "func NewJSON() Formatter {\n\treturn &jsonFormat{\n\t\tOutput: os.Stdout,\n\t}\n}", "func (b *Basic) Formatter() Formatter {\n\treturn b.formatter\n}", "func NewDumper(ctx context.Context, project, instance, database string, out io.Writer, timestamp *time.Time, bulkSize uint, tables []string) (*Dumper, error) {\n\tdbPath := fmt.Sprintf(\"projects/%s/instances/%s/databases/%s\", project, instance, database)\n\tclient, err := spanner.NewClientWithConfig(ctx, dbPath, spanner.ClientConfig{\n\t\tSessionPoolConfig: spanner.SessionPoolConfig{\n\t\t\tMinOpened: 1,\n\t\t\tMaxOpened: 1,\n\t\t},\n\t})\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to create spanner client: %v\", err)\n\t}\n\n\tvar opts []option.ClientOption\n\tif emulatorAddr := os.Getenv(\"SPANNER_EMULATOR_HOST\"); emulatorAddr != \"\" {\n\t\temulatorOpts := []option.ClientOption{\n\t\t\toption.WithEndpoint(emulatorAddr),\n\t\t\toption.WithGRPCDialOption(grpc.WithInsecure()),\n\t\t\toption.WithoutAuthentication(),\n\t\t}\n\t\topts = append(opts, emulatorOpts...)\n\t}\n\tadminClient, err := adminapi.NewDatabaseAdminClient(ctx, opts...)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to create spanner admin client: %v\", err)\n\t}\n\n\tif bulkSize == 0 {\n\t\tbulkSize = defaultBulkSize\n\t}\n\n\td := &Dumper{\n\t\tproject: project,\n\t\tinstance: instance,\n\t\tdatabase: database,\n\t\ttables: map[string]bool{},\n\t\tout: out,\n\t\ttimestamp: timestamp,\n\t\tbulkSize: bulkSize,\n\t\tclient: client,\n\t\tadminClient: adminClient,\n\t}\n\n\tfor _, table := range tables {\n\t\td.tables[strings.Trim(table, \"`\")] = true\n\t}\n\treturn d, nil\n}", "func New() (*zap.SugaredLogger, error) {\n\tzapConfig := zap.NewProductionConfig()\n\tzapConfig.DisableCaller = true\n\tzapConfig.DisableStacktrace = true\n\tzapConfig.EncoderConfig.EncodeTime = zapcore.ISO8601TimeEncoder\n\n\tzapLogger, err := zapConfig.Build()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn zapLogger.Sugar(), nil\n}", "func New(sink io.Writer) Debug {\n\tflags := log.Lshortfile\n\treturn log.New(sink, \"\", flags)\n}", "func NewParser() parser.IngressAnnotation {\n\treturn waf{}\n}", "func NewHdwalletCmd(client ioctl.Client) *cobra.Command {\n\thdwalletShorts, _ := client.SelectTranslation(_hdwalletCmdShorts)\n\tcmd := &cobra.Command{\n\t\tUse: \"hdwallet\",\n\t\tShort: hdwalletShorts,\n\t}\n\tcmd.AddCommand(NewHdwalletCreateCmd(client))\n\tcmd.AddCommand(NewHdwalletDeleteCmd(client))\n\tcmd.AddCommand(NewHdwalletDeriveCmd(client))\n\tcmd.AddCommand(NewHdwalletExportCmd(client))\n\tcmd.AddCommand(NewHdwalletImportCmd(client))\n\treturn cmd\n}", "func NewDumper(c Consumer, f Filter, s Storage, p time.Duration) *Dumper {\n\treturn &Dumper{consumer: c, filter: f, storage: s, logPeriod: p}\n}", "func newDiagnosticTextWriter(out io.Writer, files map[string]*hcl.File) hcl.DiagnosticWriter {\n\tconst outputWidth = 0\n\tconst enableColor = true\n\treturn hcl.NewDiagnosticTextWriter(out, files, outputWidth, enableColor)\n}", "func NewFlattener(configuration *DataTransformerConf) *Flattener {\n\n\tconfiguration.isSHAKE = isShakeAlgorithm(configuration.HashingAlgorithm)\n\n\tf := &Flattener{\n\t\tdataProcessorCore: dataProcessorCore{\n\t\t\tconfiguration: configuration,\n\t\t\tpointMap: sync.Map{},\n\t\t\tterminateChan: make(chan struct{}, 1),\n\t\t\tloggers: logh.CreateContextualLogger(\"pkg\", \"timeline/flattener\"),\n\t\t},\n\t}\n\n\tf.parent = f\n\n\treturn f\n}", "func New() *Haikunator {\n\treturn &Haikunator{\n\t\tAdjectives: adjectives,\n\t\tNouns: nouns,\n\t\tDelimiter: \"-\",\n\t\tTokenLength: 4,\n\t\tTokenHex: false,\n\t\tTokenChars: numbers,\n\t\tRandom: rand.New(rand.NewSource(time.Now().UnixNano())),\n\t}\n}", "func NewFLATE() *FLATEFormat { return &FLATEFormat{} }", "func newInfoDumper(w io.Writer, indent string, b boxLike, version int, flags uint32) *infoDumper {\n\tbd := infoDumper{w, indent, b, nil}\n\tif version == -1 {\n\t\tbd.write(\"[%s] size=%d\", b.Type(), b.Size())\n\t} else if version >= 0 {\n\t\tbd.write(\"[%s] size=%d version=%d flags=%06x\", b.Type(), b.Size(), version, flags)\n\t} else { // version = -2\n\t\tbd.write(\"GroupingType %q size=%d\", b.Type(), b.Size())\n\t}\n\treturn &bd\n}", "func NewJSONFormatter(keys map[string]string, full bool) (Formatter, error) {\n\tif len(keys) > 0 {\n\t\tstructure := true\n\t\tmapping := map[string]string{\n\t\t\t\"name\": \"name\", \"time\": \"time\", \"level\": \"level\", \"message\": \"message\",\n\t\t\t\"fields\": \"fields\", \"caller\": \"caller\", \"stack\": \"stack\",\n\t\t}\n\t\tfor key, value := range keys {\n\t\t\tif mapping[key] == \"\" {\n\t\t\t\t// We require that the key-name map must be pure.\n\t\t\t\treturn nil, fmt.Errorf(\"invalid json formatter key %q\", key)\n\t\t\t}\n\t\t\t// We ignore the case where all fields are mapped as empty, which is more practical.\n\t\t\tif value != \"\" && mapping[key] != value {\n\t\t\t\tstructure = false\n\t\t\t\tmapping[key] = value\n\t\t\t}\n\t\t}\n\t\t// when the json field cannot be predicted in advance, we use map to package the log data.\n\t\t// is there a better solution to improve the efficiency of json serialization?\n\t\tif !structure {\n\t\t\treturn NewJSONFormatterFromPool(newJSONFormatterMapPool(full, mapping)), nil\n\t\t}\n\t}\n\t// In most cases, the performance of json serialization of structure is higher than\n\t// that of json serialization of map. When the json field name has not changed, we\n\t// try to use structure for json serialization.\n\treturn NewJSONFormatterFromPool(newJSONFormatterObjectPool(full)), nil\n}", "func NewDevController(streamStat StreamStatRead) *devController {\n\treturn &devController{\n\t\tGrid: ui.NewGrid(),\n\t\tbodyChart: widgets.NewPieChart(),\n\t\tbodyTable: widgets.NewTable(),\n\t\tfootText: widgets.NewParagraph(),\n\t\tstreamStat: streamStat,\n\t}\n}", "func NewStandardFormatter() (standardFormatter *StandardFormatter) {\n\treturn &StandardFormatter{\n\t\tappendNewLine: true,\n\t\tdateTimeLayout: \"2006-01-02 15:04:05\",\n\t\tlayout: \"%(dateTime) [%(logLevel)] (%(pid)) %(program) %(loggerName) %(fileName) %(lineNum) %(message)\",\n\t\tmutex: new(sync.RWMutex),\n\t}\n}", "func NewParser(builder builder.TestSuitesBuilder, stream bool) parser.TestOutputParser {\n\treturn &testOutputParser{\n\t\tbuilder: builder,\n\t\tstream: stream,\n\t}\n}", "func newCMLogger(name string, chainId string, logger *zap.SugaredLogger, logLevel log.LOG_LEVEL) *CMLogger {\n\treturn &CMLogger{name: name, chainId: chainId, SugaredLogger: logger, logLevel: logLevel}\n}", "func NewConsoleFormatter() Formatter {\n\treturn new(consoleFormatter)\n}", "func NewDiffFormat(source string) formatter.Format {\n\tswitch source {\n\tcase formatter.TableFormatKey:\n\t\treturn defaultDiffTableFormat\n\t}\n\treturn formatter.Format(source)\n}", "func NewFlattener() plugins.SFHandler {\n\treturn new(Flattener)\n}", "func NewMinimalLineFormatter() Formatter {\n\treturn &LineFormatter{LineFormat: LINE_FORMAT_MINIMAL}\n}", "func NewFormatter() *PlainTextFormatter {\n\treturn &PlainTextFormatter{}\n}", "func newParser(br string) (*parser, []string, error) {\n\tloc, err := bugreportutils.TimeZone(br)\n\tif err != nil {\n\t\treturn nil, []string{}, err\n\t}\n\tpm, warnings := bugreportutils.ExtractPIDMappings(br)\n\t// Extract the year and month from the bugreport dumpstate line.\n\td, err := bugreportutils.DumpState(br)\n\tif err != nil {\n\t\treturn nil, warnings, fmt.Errorf(\"could not find dumpstate information in the bugreport: %v\", err)\n\t}\n\tbuf := new(bytes.Buffer)\n\treturn &parser{\n\t\treferenceYear: d.Year(),\n\t\treferenceMonth: d.Month(),\n\t\tloc: loc,\n\t\tbuf: buf,\n\t\tcsvState: csv.NewState(buf, true),\n\t\tpidMappings: pm,\n\t}, warnings, nil\n}", "func NewJsonFormatter() Formatter {\n\treturn JsonFormatter{}\n}", "func NewCmdDev(out io.Writer) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"dev\",\n\t\tShort: \"Runs a pipeline file in development mode\",\n\t\tArgs: cobra.NoArgs,\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\treturn dev(out, filename)\n\t\t},\n\t}\n\tAddRunDevFlags(cmd)\n\tAddDevFlags(cmd)\n\treturn cmd\n}", "func NewGoshawk(ctx context.Context, cfg *configpb.GoshawkConfig, hc *http.Client, scanOpts scanner.ScannerOptions) (*Goshawk, error) {\n\tif cfg.DestLog == nil {\n\t\treturn nil, errors.New(\"no source log config found\")\n\t}\n\tif cfg.SourceLog == nil || len(cfg.SourceLog) == 0 {\n\t\treturn nil, errors.New(\"no source log config found\")\n\t}\n\n\tdest, err := logConfigFromProto(cfg.DestLog, hc)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to parse dest log config: %v\", err)\n\t}\n\tseenNames := make(map[string]bool)\n\torigins := make(map[string]*originLog)\n\tfor _, lc := range cfg.SourceLog {\n\t\tbase, err := logConfigFromProto(lc, hc)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"failed to parse source log config: %v\", err)\n\t\t}\n\t\tif _, ok := seenNames[base.Name]; ok {\n\t\t\treturn nil, fmt.Errorf(\"duplicate source logs for name %s\", base.Name)\n\t\t}\n\t\tseenNames[base.Name] = true\n\n\t\tif _, ok := origins[base.URL]; ok {\n\t\t\treturn nil, fmt.Errorf(\"duplicate source logs for url %s\", base.URL)\n\t\t}\n\t\torigins[base.URL] = &originLog{\n\t\t\tlogConfig: *base,\n\t\t\tsths: make(chan *x509ext.LogSTHInfo, cfg.BufferSize),\n\t\t}\n\t}\n\n\thawk := Goshawk{dest: dest, origins: origins}\n\tscanOpts.Matcher = &hawk\n\tif scanOpts.Matcher.(scanner.Matcher) == nil {\n\t\treturn nil, fmt.Errorf(\"hawk does not satisfy scanner.Matcher interface\")\n\t}\n\thawk.scanOpts = scanOpts\n\treturn &hawk, nil\n}", "func WithDevelopmentLogger(opts ...zap.Option) Option {\n\treturn func(s *SVC) error {\n\t\ts.zapOpts = append(s.zapOpts, opts...)\n\t\tlogger, atom := s.newLogger(\n\t\t\tzapcore.DebugLevel,\n\t\t\tzapcore.NewJSONEncoder(zap.NewProductionEncoderConfig()),\n\t\t)\n\t\tlogger = logger.With(zap.String(\"app\", s.Name), zap.String(\"version\", s.Version))\n\t\treturn assignLogger(s, logger, atom)\n\t}\n}", "func NewParser(waf *engine.Waf) (*Parser, error) {\n\tif waf == nil {\n\t\treturn nil, errors.New(\"must use a valid waf instance\")\n\t}\n\tp := &Parser{\n\t\tWaf: waf,\n\t\tdefaultActions: []string{},\n\t\tDisabledDirectives: []string{},\n\t}\n\treturn p, nil\n}", "func New() (*MinConf, error) {\n\tjson := `{\n\t\t\"$\": {\n\t\t\t\"envs\": {\"development\": \"dev ENV ARGV\"},\n\t\t},\n\t\t\"dev\": {}\n\t}`\n\treturn NewFromString(json)\n}", "func WithOptionalFormatter(enabled bool, formatter ByteFormatter) OutputFormatter {\n\treturn func(i io.Writer, input []byte) []byte {\n\t\tif enabled {\n\t\t\treturn formatter(input)\n\t\t}\n\t\treturn input\n\t}\n}", "func NewCliOpFormatter(workingDirPath, dataDirPath string) CliOpFormatter {\n\treturn CliOpFormatter{workingDirPath, dataDirPath}\n}", "func CreateCryptoFormatter(format string) *AcraCryptoFormatter {\n\tvar formatter *AcraCryptoFormatter\n\tswitch strings.ToLower(format) {\n\tcase JSONFormatString:\n\t\tformatter = NewCryptoFormatter(JSONFormatter())\n\tcase CefFormatString:\n\t\tformatter = NewCryptoFormatter(CEFFormatter())\n\tdefault:\n\t\tformatter = NewCryptoFormatter(TextFormatter())\n\t}\n\treturn formatter\n}", "func (f *Formattable) Formatter() Formatter {\n\tif f.formatter == nil {\n\t\tf.formatter = NewTextFormatter()\n\t}\n\treturn f.formatter\n}", "func New(output *os.File, NoColors ...bool) *Tiny {\n\tnocolors := false\n\tif len(NoColors) > 0 {\n\t\tnocolors = NoColors[0]\n\t}\n\tl := logrus.New()\n\tl.SetLevel(logrus.DebugLevel)\n\tlog.SetOutput(output)\n\tl.SetFormatter(&f.Formatter{\n\t\tNoColors: nocolors,\n\t\tHideKeys: true,\n\t\tFieldsOrder: []string{\"component\", \"category\"},\n\t})\n\treturn &Tiny{l}\n}", "func NewZKHappyPathTest(setup ZKTestSetup) *ZKTest {\n\thappyPathTest := zkHappyPathTest{\n\t\ttestSetup: setup,\n\t\tinfo: log.New(os.Stderr, \"INFO: \", log.Ldate|log.Ltime|log.Lshortfile),\n\t\terror: log.New(os.Stderr, \"ERROR: \", log.Ldate|log.Ltime|log.Lshortfile),\n\t}\n\tzkHappyTest := ZKTest{\n\t\tDesc: \"Happy Path Test\",\n\t\tCallbacks: &happyPathTest,\n\t\tErr: nil,\n\t}\n\n\treturn &zkHappyTest\n}", "func NewParser(logger logr.Logger, schema *Schema) *Parser {\n\treturn &Parser{\n\t\tlogger: logger.WithName(\"json-schema-parser\"),\n\t\tschema: schema,\n\t}\n}", "func newParser(filename string, b []byte, opts ...Option) *parser {\n\tp := &parser{\n\t\tfilename: filename,\n\t\terrs: new(errList),\n\t\tdata: b,\n\t\tpt: savepoint{position: position{line: 1}},\n\t\trecover: true,\n\t}\n\tp.setOptions(opts)\n\treturn p\n}", "func makeStartLine(formatter logFormatter, format string, args ...interface{}) *buffer {\n\tentry := makeUnstructuredEntry(\n\t\tcontext.Background(),\n\t\tseverity.UNKNOWN, /* header - ignored */\n\t\t0, /* header - ignored */\n\t\t2, /* depth */\n\t\ttrue, /* redactable */\n\t\tformat,\n\t\targs...)\n\tentry.header = true\n\tentry.tags = configTagsBuffer\n\treturn formatter.formatEntry(entry)\n}", "func NewLogger(dev bool) *zap.Logger {\n\tvar cfg zap.Config\n\tif dev {\n\t\tcfg = zap.NewDevelopmentConfig()\n\t} else {\n\t\tcfg = zap.NewProductionConfig()\n\t}\n\n\tlogger, err := cfg.Build()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn logger\n}", "func DefaultDevelopmentConfig(minLevel zapcore.Level) zap.Config {\n\tloggerConf := zap.NewDevelopmentConfig()\n\tloggerConf.Level.SetLevel(minLevel)\n\tloggerConf.EncoderConfig.EncodeLevel = zapcore.CapitalColorLevelEncoder\n\n\treturn loggerConf\n}", "func newParser(filename string, b []byte, opts ...Option) *parser {\n\tstats := Stats{\n\t\tChoiceAltCnt: make(map[string]map[string]int),\n\t}\n\n\tp := &parser{\n\t\tfilename: filename,\n\t\terrs: new(errList),\n\t\tdata: b,\n\t\tpt: savepoint{position: position{line: 1}},\n\t\trecover: true,\n\t\tcur: current{\n\t\t\tglobalStore: make(storeDict),\n\t\t},\n\t\tmaxFailPos: position{col: 1, line: 1},\n\t\tmaxFailExpected: make([]string, 0, 20),\n\t\tStats: &stats,\n\t\t// start rule is rule [0] unless an alternate entrypoint is specified\n\t\tentrypoint: g.rules[0].name,\n\t}\n\tp.setOptions(opts)\n\n\tif p.maxExprCnt == 0 {\n\t\tp.maxExprCnt = math.MaxUint64\n\t}\n\n\treturn p\n}", "func New() *Logger {\n\tconf := zap.NewDevelopmentConfig()\n\tl, err := conf.Build(\n\t\tzap.AddStacktrace(zap.FatalLevel),\n\t)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tz := l.Sugar()\n\treturn &Logger{z}\n}", "func newParser(invocation string, verbose bool) *parser {\n\tp := &parser{\n\t\tScanner: &scanner.Scanner{},\n\t\toriginal: invocation,\n\t\tverbose: verbose,\n\t\terrors: make([]error, 0),\n\t}\n\tp.Init(strings.NewReader(invocation))\n\tp.Error = func(s *scanner.Scanner, msg string) {\n\t\tp.errors = append(p.errors, p.error(s.Position, msg))\n\t}\n\treturn p\n}", "func newParser(filename string, b []byte, opts ...Option) *parser {\n\tstats := Stats{\n\t\tChoiceAltCnt: make(map[string]map[string]int),\n\t}\n\n\tp := &parser{\n\t\tfilename: filename,\n\t\terrs: new(errList),\n\t\tdata: b,\n\t\tpt: savepoint{position: position{line: 1}},\n\t\trecover: true,\n\t\tcur: current{\n\t\t\tstate: make(storeDict),\n\t\t\tglobalStore: make(storeDict),\n\t\t},\n\t\tmaxFailPos: position{col: 1, line: 1},\n\t\tmaxFailExpected: make([]string, 0, 20),\n\t\tStats: &stats,\n\t\t// start rule is rule [0] unless an alternate entrypoint is specified\n\t\tentrypoint: g.rules[0].name,\n\t}\n\tp.setOptions(opts)\n\n\tif p.maxExprCnt == 0 {\n\t\tp.maxExprCnt = math.MaxUint64\n\t}\n\n\treturn p\n}", "func newParser(filename string, b []byte, opts ...Option) *parser {\n\tstats := Stats{\n\t\tChoiceAltCnt: make(map[string]map[string]int),\n\t}\n\n\tp := &parser{\n\t\tfilename: filename,\n\t\terrs: new(errList),\n\t\tdata: b,\n\t\tpt: savepoint{position: position{line: 1}},\n\t\trecover: true,\n\t\tcur: current{\n\t\t\tstate: make(storeDict),\n\t\t\tglobalStore: make(storeDict),\n\t\t},\n\t\tmaxFailPos: position{col: 1, line: 1},\n\t\tmaxFailExpected: make([]string, 0, 20),\n\t\tStats: &stats,\n\t\t// start rule is rule [0] unless an alternate entrypoint is specified\n\t\tentrypoint: g.rules[0].name,\n\t}\n\tp.setOptions(opts)\n\n\tif p.maxExprCnt == 0 {\n\t\tp.maxExprCnt = math.MaxUint64\n\t}\n\n\treturn p\n}", "func newParser(filename string, b []byte, opts ...Option) *parser {\n\tstats := Stats{\n\t\tChoiceAltCnt: make(map[string]map[string]int),\n\t}\n\n\tp := &parser{\n\t\tfilename: filename,\n\t\terrs: new(errList),\n\t\tdata: b,\n\t\tpt: savepoint{position: position{line: 1}},\n\t\trecover: true,\n\t\tcur: current{\n\t\t\tstate: make(storeDict),\n\t\t\tglobalStore: make(storeDict),\n\t\t},\n\t\tmaxFailPos: position{col: 1, line: 1},\n\t\tmaxFailExpected: make([]string, 0, 20),\n\t\tStats: &stats,\n\t\t// start rule is rule [0] unless an alternate entrypoint is specified\n\t\tentrypoint: g.rules[0].name,\n\t}\n\tp.setOptions(opts)\n\n\tif p.maxExprCnt == 0 {\n\t\tp.maxExprCnt = math.MaxUint64\n\t}\n\n\treturn p\n}", "func NewSimpleLineFormatter() Formatter {\n\treturn &LineFormatter{LineFormat: LINE_FORMAT_SIMPLE}\n}", "func NewDumper(specs ...string) (Dumper, error) {\n\tlog.Info(\"creating request dumper...\")\n\n\td := &dumper{\n\t\tmethods: make(map[string]*spec),\n\t\tregexps: []*spec{},\n\t\thandlers: make(map[string]Handler),\n\t\tresolve: false,\n\t\tenabled: true,\n\t}\n\n\tRegisterDefaultHandlers(d)\n\n\tfor _, spec := range specs {\n\t\tif err := d.Parse(spec); err != nil {\n\t\t\treturn &dumper{}, err\n\t\t}\n\t}\n\n\treturn d, nil\n}", "func New(clk gpio.PinOut, data gpio.PinIn) (*Dev, error) {\n\tif err := data.In(gpio.PullDown, gpio.FallingEdge); err != nil {\n\t\treturn nil, err\n\t}\n\tif err := clk.Out(gpio.Low); err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Dev{\n\t\tname: \"hx711{\" + clk.Name() + \", \" + data.Name() + \"}\",\n\t\tinputMode: CHANNEL_A_GAIN_128,\n\t\tclk: clk,\n\t\tdata: data,\n\t\tdone: nil,\n\t}, nil\n}", "func NewFromZap(logger *zap.Logger) Logger {\n\treturn zapLogger{logger: logger.Sugar()}\n}", "func Parser() *HJSON {\n\treturn &HJSON{}\n}", "func New() *Devenv {\n\treturn &Devenv{\n\t\tConfig: config.New(),\n\t}\n}", "func RunDevWith(plugin plugin.YomoObjectPlugin, endpoint string, formatter OutputFormatter) {\n\tgo func() {\n\t\tlogger.Infof(\"plugin service [%s] start... [%s]\", plugin.Name(), endpoint)\n\n\t\t// activation service\n\t\tframework.NewServer(endpoint, plugin)\n\t}()\n\n\tyomoEchoClient, err := util.QuicClient(configs.DefaultEchoConf.EchoServerAddr)\n\t//yomoEchoClient, err := util.QuicClient(\"localhost:11520\")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tyomoPluginClient, err := util.QuicClient(endpoint)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tgo io.Copy(yomoPluginClient, yomoEchoClient) // nolint\n\n\t// select formatter\n\tvar w io.Writer\n\tswitch formatter {\n\tcase OutputHexString:\n\t\tw = &hexStringFormatter{}\n\tcase OutputPacketPrinter:\n\t\tw = &packetPrinterFormatter{}\n\tcase OutputEchoData:\n\t\tw = &echoDataFormatter{}\n\tcase OutputThermometerData:\n\t\tw = &thermometerDataFormatter{}\n\tdefault:\n\t\tw = &packetPrinterFormatter{}\n\t}\n\tgo util.CopyTo(w, yomoPluginClient) // nolint\n\n\tfor {\n\t\ttime.Sleep(time.Second)\n\t\t_, err = yomoEchoClient.Write([]byte(\"ping\"))\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t}\n}", "func NewDevNullLogger() *DevNullLogger {\n\treturn &DevNullLogger{}\n}", "func NewDev(ifName string, frameFilter FrameFilter) (dev Dev, err error) {\n\td := new(bpfDev)\n\td.name = ifName\n\td.filter = frameFilter\n\td.fd, err = getBpfFd()\n\tif err != nil {\n\t\treturn\n\t}\n\terr = ifReq(d.fd, ifName)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tvar bufLen int\n\tbufLen, err = ioCtl(d.fd)\n\tif err != nil {\n\t\treturn\n\t}\n\t_, err = d.getMTU()\n\tif err != nil {\n\t\treturn\n\t}\n\t_, err = d.getHardwareAddr()\n\tif err != nil {\n\t\treturn\n\t}\n\n\td.readBuf = make([]byte, bufLen)\n\n\tdev = d\n\n\treturn\n}", "func newTransactorSetup(t *testing.T, prng *rand.Rand, hideSignHash bool, signer types.Signer, chainID int64) test.TransactorSetup {\n\twalletSeed := make([]byte, 20)\n\tprng.Read(walletSeed)\n\tmnemonic, err := hdwallet.NewMnemonicFromEntropy(walletSeed)\n\trequire.NoError(t, err)\n\n\trawHDWallet, err := hdwallet.NewFromMnemonic(mnemonic)\n\trequire.NoError(t, err)\n\trequire.NotNil(t, rawHDWallet)\n\n\tvar wrappedWallet accounts.Wallet = rawHDWallet\n\tif hideSignHash {\n\t\twrappedWallet = &noSignHash{rawHDWallet}\n\t}\n\thdWallet, err := hd.NewWallet(wrappedWallet, hd.DefaultRootDerivationPath.String(), 0)\n\trequire.NoError(t, err)\n\trequire.NotNil(t, hdWallet)\n\n\tvalidAcc, err := hdWallet.NewAccount()\n\trequire.NoError(t, err)\n\trequire.NotNil(t, validAcc)\n\n\treturn test.TransactorSetup{\n\t\tSigner: signer,\n\t\tChainID: chainID,\n\t\tTr: hd.NewTransactor(hdWallet.Wallet(), signer),\n\t\tValidAcc: accounts.Account{Address: wallet.AsEthAddr(validAcc.Address())},\n\t\tMissingAcc: accounts.Account{Address: common.HexToAddress(missingAddr)},\n\t}\n}", "func NewOptab(cfg printer.Config) Interface {\n\treturn GoFmt(&optab{cfg: cfg})\n}", "func newPrinter(format OutputFormat, writerType WriterType) *printer {\n\treturn &printer{Formatter: NewFormatter(format, writerType), Encoder: NewEncoder()}\n}", "func NewDebugger() *Debugger {\n\td := &Debugger{\n\t\tvm: otto.New(),\n\t\tcallbacks: make(map[string][]otto.Value),\n\t\tStep: make(chan bool, 1),\n\t\tCont: make(chan bool, 1),\n\t}\n\n\t// Make the \"on\" function available to the js vm\n\td.vm.Set(\"on\", func(call otto.FunctionCall) otto.Value {\n\t\tname, _ := call.Argument(0).ToString()\n\t\td.callbacks[name] = append(d.callbacks[name], call.Argument(1))\n\t\treturn otto.Value{}\n\t})\n\n\t// add the pretty print functions\n\td.vm.Run(prettPrintSrc)\n\n\treturn d\n}", "func NewParser(handler *MetricHandler) *Parser {\n\treturn &Parser{\n\t\tmachine: NewMachine(handler),\n\t\thandler: handler,\n\t}\n}", "func NewTimeFormatter(timestamps bool) TimeFormatter {\n\ttimeFormatter := timeFormatter(timestamps)\n\treturn &timeFormatter\n}", "func New(level string, writer string, prettyprint string) Logger {\n\tvar lg Logger\n\tlg.level = stringToLevel()[level]\n\tlg.logger = json.NewEncoder(stringToWriter(writer))\n\tif prettyprint == \"true\" {\n\t\tlg.logger.SetIndent(\"\", \" \")\n\t}\n\n\tvar process = strings.Split(os.Args[0], \"/\")\n\tlg.json.Process = process[len(process)-1]\n\n\treturn lg\n}", "func NewFzShade() *FzShade {\n\treturn (*FzShade)(allocFzShadeMemory(1))\n}", "func (app *builder) WithLexer(lexer lexers.Lexer) Builder {\n\tapp.lexer = lexer\n\treturn app\n}", "func NewDevAttr() DevAttributes {\n\ta := DevAttributes{\n\t\tErrlogHistSize: 60, // default max number of lines in errlog history\n\t}\n\n\treturn a\n}" ]
[ "0.58646077", "0.57912886", "0.5618458", "0.5608604", "0.5532328", "0.5477053", "0.5396334", "0.5378421", "0.53731", "0.5371514", "0.5264039", "0.522683", "0.51359427", "0.510094", "0.50950074", "0.50814706", "0.50802374", "0.50709385", "0.50024474", "0.49971724", "0.49211594", "0.48947728", "0.4888736", "0.48830217", "0.4814764", "0.4742475", "0.4734575", "0.47208446", "0.46312985", "0.4606049", "0.45994863", "0.4591475", "0.45698628", "0.45687413", "0.4551745", "0.4551159", "0.4513645", "0.45111623", "0.45060566", "0.44826496", "0.4480269", "0.4478963", "0.44587994", "0.44542027", "0.44450176", "0.43984708", "0.43840626", "0.43836546", "0.4381013", "0.43742892", "0.43717396", "0.43708387", "0.43495077", "0.43423748", "0.43417743", "0.43381855", "0.4324555", "0.4319212", "0.4318146", "0.43074825", "0.4299204", "0.4293432", "0.4265537", "0.42619202", "0.4259851", "0.42494068", "0.42408615", "0.42374232", "0.42329702", "0.42260924", "0.421728", "0.42093146", "0.42084625", "0.42084256", "0.41973355", "0.41945374", "0.419212", "0.4191624", "0.41846508", "0.41846508", "0.41846508", "0.4176978", "0.41751426", "0.41697592", "0.41695875", "0.4168626", "0.41663098", "0.41661644", "0.41658297", "0.41602135", "0.41602117", "0.41601077", "0.41532314", "0.41528502", "0.41502038", "0.41463628", "0.41435772", "0.41421032", "0.4142075", "0.41281828" ]
0.85738707
0
Write a string and tracks the position of the string so we can break lines cleanly. Do not send ANSI escape sequences, just raw strings
func (hd *HappyDevFormatter) writeString(buf bufferWriter, s string, col *int) { buf.WriteString(s) *col += len(s) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (w *win) WriteString(str string) {\n\tw.Addr(\"#%d\", w.pAddr)\n\tdata := []byte(str + \"\\n\")\n\tw.writeData(data)\n\n\tnr := utf8.RuneCount(data)\n\tw.pAddr += nr\n\tw.eAddr += nr\n}", "func (t *tScreen) writeString(s string) {\r\n\tif t.buffering {\r\n\t\tio.WriteString(&t.buf, s)\r\n\t} else {\r\n\t\tio.WriteString(t.c, s)\r\n\t}\r\n}", "func (tb *Textbox) WriteString(s string) (int, error) {\n\treturn tb.WriteRunes([]rune(s))\n}", "func (res Responder) WriteString(s string) int {\n\tn := res.writeInline(binDOLLAR, strconv.Itoa(len(s)))\n\tm, _ := res.b.WriteString(s)\n\tres.b.Write(binCRLF)\n\treturn n + m + 2\n}", "func WriteString(data []byte, str string, stype string_t, pos *int, l int) {\n switch stype {\n case NULLSTR:\n checkSize(len(data[*pos:]), len(str))\n // Write the string and then terminate with 0x00 byte.\n copy(data[*pos:], str)\n checkSize(len(data[*pos:]), len(str) + 1)\n *pos += len(str)\n data[*pos] = 0x00\n *pos++\n\n case LENENCSTR:\n // Write the encoded length.\n WriteLenEncInt(data, uint64(len(str)), pos)\n // Then write the string as a FIXEDSTR.\n WriteString(data, str, FIXEDSTR, pos, l)\n\n case FIXEDSTR:\n\n checkSize(len(data[*pos:]), l)\n // Pads the string with 0's to fill the specified length l.\n copy(data[*pos:*pos+l], str)\n *pos += l\n\n case EOFSTR:\n\n checkSize(len(data[*pos:]), len(str))\n // Copies the string into the data.\n *pos += copy(data[*pos:], str)\n }\n}", "func (n *ninjaWriterWithWrap) WriteString(s string) (written int, noError error) {\n\t// Always return the full length of the string and a nil error.\n\t// ninjaWriterWithWrap doesn't return errors to the caller, it saves them until Flush()\n\twritten = len(s)\n\n\tif n.err != nil {\n\t\treturn\n\t}\n\n\tconst spaceLen = 1\n\tif !n.space {\n\t\t// No space is pending, so a line wrap can't be inserted before this, so just write\n\t\t// the string.\n\t\tn.lineLen += len(s)\n\t\t_, n.err = n.writer.WriteString(s)\n\t} else if n.lineLen+len(s)+spaceLen > n.maxLineLen {\n\t\t// A space is pending, and the pending strings plus the current string would exceed the\n\t\t// maximum line length. Wrap and indent before the pending space and strings, then write\n\t\t// the pending and current strings.\n\t\t_, n.err = n.writer.WriteString(\" $\\n\")\n\t\tif n.err != nil {\n\t\t\treturn\n\t\t}\n\t\t_, n.err = n.writer.WriteString(indentString[:indentWidth*2])\n\t\tif n.err != nil {\n\t\t\treturn\n\t\t}\n\t\tn.lineLen = indentWidth*2 + n.pendingLen\n\t\ts = strings.TrimLeftFunc(s, unicode.IsSpace)\n\t\tn.pending = append(n.pending, s)\n\t\tn.writePending()\n\n\t\tn.space = false\n\t} else {\n\t\t// A space is pending but the current string would not reach the maximum line length,\n\t\t// add it to the pending list.\n\t\tn.pending = append(n.pending, s)\n\t\tn.pendingLen += len(s)\n\t\tn.lineLen += len(s)\n\t}\n\n\treturn\n}", "func (w *VT100Writer) WriteStr(data string) {\n\tw.Write([]byte(data))\n\t//fmt.Fprintf(os.Stderr, \"\\x1b[2m>>\\x1b[m '\\x1b[33m%s\\x1b[m'\\n\", data)\n}", "func (hw *HighlightedWriter) WriteStr(data string) {\n\tif hw.writingInput && data != \"\" {\n\t\t// If prompt is trying to write input text, intercept and replace the write with syntax highlighted text\n\t\tpartial := shellContext.cmd.String() + data\n\t\titerator, err := shellContext.lexer.Tokenise(nil, partial)\n\t\tif err != nil {\n\t\t\thw.delegate.WriteStr(data)\n\t\t\treturn\n\t\t}\n\t\tmodified := strings.Builder{}\n\t\tshellContext.formatter.Format(&modified, shellContext.style, iterator)\n\t\tlines := strings.Split(modified.String(), \"\\n\")\n\t\tif strings.Contains(data, \"\\n\") {\n\t\t\t// data was a complete line, so we should output the next-to-last\n\t\t\thw.delegate.WriteRawStr(lines[len(lines)-2] + \"\\n\")\n\t\t} else {\n\t\t\thw.delegate.WriteRawStr(lines[len(lines)-1])\n\t\t}\n\t\treturn\n\t}\n\thw.delegate.WriteStr(data)\n}", "func (p *printer) writeString(pos token.Position, s string, isLit bool) {\n\tif p.out.Column == 1 {\n\t\tif p.Config.Mode&SourcePos != 0 {\n\t\t\tp.writeLineDirective(pos)\n\t\t}\n\t\tp.writeIndent()\n\t}\n\n\tif pos.IsValid() {\n\t\t// update p.pos (if pos is invalid, continue with existing p.pos)\n\t\t// Note: Must do this after handling line beginnings because\n\t\t// writeIndent updates p.pos if there's indentation, but p.pos\n\t\t// is the position of s.\n\t\tp.pos = pos\n\t}\n\n\tif isLit {\n\t\t// Protect s such that is passes through the tabwriter\n\t\t// unchanged. Note that valid Go programs cannot contain\n\t\t// tabwriter.Escape bytes since they do not appear in legal\n\t\t// UTF-8 sequences.\n\t\tp.output = append(p.output, tabwriter.Escape)\n\t\tif p.lastTok == token.CSTRING {\n\t\t\tp.output = append(p.output, 'C')\n\t\t}\n\t}\n\n\tif debug {\n\t\tp.output = append(p.output, fmt.Sprintf(\"/*%s*/\", pos)...) // do not update p.pos!\n\t}\n\tp.output = append(p.output, s...)\n\n\t// update positions\n\tnlines := 0\n\tvar li int // index of last newline; valid if nlines > 0\n\tfor i := 0; i < len(s); i++ {\n\t\t// Raw string literals may contain any character except back quote (`).\n\t\tif ch := s[i]; ch == '\\n' || ch == '\\f' {\n\t\t\t// account for line break\n\t\t\tnlines++\n\t\t\tli = i\n\t\t\t// A line break inside a literal will break whatever column\n\t\t\t// formatting is in place; ignore any further alignment through\n\t\t\t// the end of the line.\n\t\t\tp.endAlignment = true\n\t\t}\n\t}\n\tp.pos.Offset += len(s)\n\tif nlines > 0 {\n\t\tp.pos.Line += nlines\n\t\tp.out.Line += nlines\n\t\tc := len(s) - li\n\t\tp.pos.Column = c\n\t\tp.out.Column = c\n\t} else {\n\t\tp.pos.Column += len(s)\n\t\tp.out.Column += len(s)\n\t}\n\n\tif isLit {\n\t\tp.output = append(p.output, tabwriter.Escape)\n\t}\n\n\tp.last = p.pos\n}", "func WriteString() {\n\tfmt.Println(\"----------------> WriteString\")\n\ts := \" world\"\n\tbuf := bytes.NewBufferString(\"hello\")\n\tfmt.Println(buf.String())\n\n\t//write string at then end of buffer\n\tbuf.WriteString(s)\n\n\tfmt.Println(buf.String())\n}", "func (hw *HighlightedWriter) WriteRawStr(data string) {\n\thw.delegate.WriteRawStr(data)\n}", "func (w *VT100Writer) WriteRawStr(data string) {\n\tw.WriteRaw([]byte(data))\n}", "func writeString(w http.ResponseWriter, body string, status int) {\n\twriteBody(w, []byte(body), status, \"text/plain\")\n}", "func WriteString(buffer []byte, offset int, value string) {\n WriteBytes(buffer, offset, []byte(value))\n}", "func (g *Generator) WriteString(s string) (int, error) {\n\treturn g.o.Write([]byte(s))\n}", "func (w *ByteWriter) WriteString(val string, offset int) (int, error) {\n\t_, err := w.Write([]byte(val), offset)\n\treturn offset + len(val), err\n}", "func WriteString(w http.ResponseWriter, str string) {\n\tw.Header().Set(\"Content-Type\", \"text/plain\")\n\tw.WriteHeader(http.StatusOK)\n\t_, err := w.Write([]byte(str))\n\tif err != nil {\n\t\tLogWarning(errors.WrapPrefix(err, \"failed to write response\", 0))\n\t}\n}", "func (w Writer) writeString(s string) error {\n\t_, err := w.w.WriteString(s)\n\treturn err\n}", "func (res Responder) WriteInlineString(s string) int {\n\treturn res.writeInline(binPLUS, s)\n}", "func (stream *Stream) WriteRawString(s string) {\n\tstream.write([]byte(s))\n}", "func (b *defaultByteBuffer) WriteString(s string) (n int, err error) {\n\tif b.status&BitWritable == 0 {\n\t\treturn -1, errors.New(\"unwritable buffer, cannot support WriteString\")\n\t}\n\tn = len(s)\n\tb.ensureWritable(n)\n\tcopy(b.buff[b.writeIdx:b.writeIdx+n], s)\n\tb.writeIdx += n\n\treturn\n}", "func (w *Writer) WriteString(msg string) {\n\tif w.err != nil {\n\t\treturn\n\t}\n\tw.b = AppendString(w.b, msg)\n}", "func (e *Escpos) Write(data string) (int, error) {\n\treturn e.WriteRaw([]byte(data))\n}", "func (buffer *Buffer) WriteString(s string) {\n\tif buffer == nil || buffer.B == nil {\n\t\treturn\n\t}\n\n\tif _, err := buffer.B.WriteString(s); err != nil {\n\t\t_, _ = fmt.Fprintf(os.Stderr, \"buffer write error: %v\\n\", err)\n\t\tbuffer.Error = err\n\t}\n}", "func (w *Writer) WriteString(str string) {\n\tlength := util.UTF16Length(str)\n\tswitch {\n\tcase length == 0:\n\t\tw.writeByte(TagEmpty)\n\tcase length < 0:\n\t\tw.WriteBytes(*(*[]byte)(unsafe.Pointer(&str)))\n\tcase length == 1:\n\t\tw.writeByte(TagUTF8Char)\n\t\tw.writeString(str)\n\tdefault:\n\t\tsetWriterRef(w, nil, nil)\n\t\twriteString(w, str, length)\n\t}\n}", "func (g *ginGzipWriter) WriteString(s string) (int, error) {\n\treturn g.wrapper.Write([]byte(s))\n}", "func (ts *System) WriteLine(str string) {\n\tts.pages[ts.page].lines[ts.pages[ts.page].line] = &line{}\n\n\tline := \"\"\n\tfor _, char := range strings.Split(str, \"\") {\n\t\tif char == \"\\t\" {\n\t\t\tchar = \" \"\n\t\t}\n\n\t\tline += char\n\n\t\tif ts.pages[ts.page].editable {\n\t\t\tts.needsDraw = append(ts.needsDraw, char)\n\t\t}\n\t}\n\n\tif !ts.pages[ts.page].editable {\n\t\tts.delegateKeyPress(engo.Key(-1), &input.Modifiers{Output: true, Line: &line})\n\t\tts.delegateKeyPress(engo.KeyEnter, &input.Modifiers{Ignore: true, Output: true})\n\t} else {\n\t\tts.needsDraw = append(ts.needsDraw, \"\\n\")\n\t}\n}", "func WriteString(conn io.Writer, str string) error {\n\tvalue := fmt.Sprintf(\"%s\\n\", str)\n\tbytes := []byte(value)\n\t_, err := conn.Write(bytes)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *Ctx) WriteString(s string) (int, error) {\n\tc.Response.AppendBodyString(s)\n\treturn len(s), nil\n}", "func (out *OutBuffer) WriteString(s string) {\n\tout.Append([]byte(s)...)\n}", "func (s *StreamBuffer) WriteStringRun(str string) (n int, err error) {\n\treturn s.WriteRuneRun([]rune(str))\n}", "func writeString(w io.Writer, s string) (n int, err error) {\n\ttype stringWriter interface {\n\t\tWriteString(string) (n int, err error)\n\t}\n\tif sw, ok := w.(stringWriter); ok {\n\t\treturn sw.WriteString(s) // Avoid copy string\n\t}\n\treturn w.Write([]byte(s)) // Using temporary copy\n}", "func (dm *dataManager) writeString(address uint, str string) (err ProcessException) {\n\tdata := []byte(str)\n\n\terr = dm.process.WriteBytes(address, data)\n\n\treturn\n}", "func (r *RingBuffer) WriteString(s string) (n int, err error) {\n\tbs := String2Bytes(s)\n\n\tr.mu.Lock()\n\tdefer r.mu.Unlock()\n\tn, err = r.write(bs)\n\treturn n, err\n}", "func (fs *Fs) WriteString(file *os.File, string string) (int, error) {\n\treturn file.WriteString(string) // #nosec G304\n}", "func (hc *HTTPContext) WriteString(data string) {\n\thc.W.Write([]byte(data))\n}", "func WriteString(w Writer, s string) (int, error) {\n\tif w == nil {\n\t\treturn 0, ErrMissingWriter\n\t}\n\treturn io.WriteString(w, s)\n}", "func (r *SizeRotator) WriteString(str string) (n int, err error) {\r\n\treturn r.Write([]byte(str))\r\n}", "func (rb *recordBuilder) WriteString(s string) (n int, err error) {\n\treturn rb.content.WriteString(s)\n}", "func (bw *BufWriter) RawString(val string) {\n\tif bw.Error != nil {\n\t\treturn\n\t}\n\t_, bw.Error = bw.writer.WriteString(val)\n}", "func WriteString(w io.Writer, data string) error {\n\treturn WriteBytes(w, []byte(data))\n}", "func (this *FileAppender) WriteString(level base.LogLevel, location string, dtime time.Time, message string, args ...interface{}) {\n\tif !this.isDispose {\n\t\t//fmt.Printf(\"FileAppender->WriteString->writeStringChan %s \\r\\n\", message)\n\t\tthis.bufferChan <- this.Formatter().Format(level, location, dtime, message, args...)\n\t}\n}", "func (req *Request) WriteString(s string) (int, error) {\n\treturn req.res.Write([]byte(s))\n}", "func (dw *DataWriter) WriteString(value string) error {\n\tbytes := []rune(value)\n\tbytesNumber := uint16(len(bytes))\n\terr := binary.Write(dw.w, binary.BigEndian, bytesNumber)\n\tif err != nil {\n\t\treturn err\n\t}\n\trunes := []rune(value)\n\tfor _, r := range runes {\n\t\t_, err = dw.w.WriteRune(r)\n\t}\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (w *multiWriter) WriteString(b string) (int, error) {\n\tw.ctx.Set(\"response\", b)\n\treturn w.ResponseWriter.Write([]byte(b))\n}", "func (b *Buf) WriteString(str string) {\n\tn := b.grow(len(str))\n\tcopy(b.b[n:], str)\n}", "func (c *fakeRedisConn) WriteString(str string) { c.rsp = append(c.rsp, str) }", "func (d *Default) WriteString(s string) {\n\td.Writer.Write([]byte(s))\n}", "func (w *Writer) RawString(s string) {\n\tw.buf = append(w.buf, s...)\n}", "func writeString(buf *bytes.Buffer, a []byte) {\n\t// test to see what sort of encoding is best to use\n\tencoding := tokenEnc\n\tacc := make([]byte, len(a), len(a))\n\tfor i, c := range a {\n\t\tacc[i] = c\n\t\tswitch {\n\t\tcase bytes.IndexByte(tokenChar, c) > -1:\n\t\t\tcontinue\n\t\tcase (encoding == tokenEnc) && bytes.IndexByte(stringEncChar, c) > -1:\n\t\t\tencoding = quotedEnc\n\t\t\tstrAcc := make([]byte, i, len(a))\n\t\t\tcopy(strAcc, acc)\n\t\t\tfor j := i; j < len(a); j++ {\n\t\t\t\tc := a[j]\n\t\t\t\tif bytes.IndexByte(stringEncChar, c) < 0 {\n\t\t\t\t\tencoding = base64Enc\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t\tswitch c {\n\t\t\t\tcase '\\b':\n\t\t\t\t\tacc = append(strAcc, []byte(\"\\\\b\")...)\n\t\t\t\tcase '\\t':\n\t\t\t\t\tstrAcc = append(strAcc, []byte(\"\\\\t\")...)\n\t\t\t\tcase '\\v':\n\t\t\t\t\tstrAcc = append(strAcc, []byte(\"\\\\v\")...)\n\t\t\t\tcase '\\n':\n\t\t\t\t\tstrAcc = append(strAcc, []byte(\"\\\\n\")...)\n\t\t\t\tcase '\\f':\n\t\t\t\t\tstrAcc = append(strAcc, []byte(\"\\\\f\")...)\n\t\t\t\tcase '\"':\n\t\t\t\t\tstrAcc = append(strAcc, []byte(\"\\\\\\\"\")...)\n\t\t\t\tcase '\\'':\n\t\t\t\t\tstrAcc = append(strAcc, []byte(\"'\")...)\n\t\t\t\tcase '\\\\':\n\t\t\t\t\tstrAcc = append(strAcc, []byte(\"\\\\\\\\\")...)\n\t\t\t\tcase '\\r':\n\t\t\t\t\tstrAcc = append(strAcc, []byte(\"\\\\r\")...)\n\t\t\t\tdefault:\n\t\t\t\t\tstrAcc = append(strAcc, c)\n\t\t\t\t}\n\t\t\t}\n\t\t\tif encoding == quotedEnc {\n\t\t\t\tbuf.WriteString(\"\\\"\")\n\t\t\t\tbuf.Write(strAcc)\n\t\t\t\tbuf.WriteString(\"\\\"\")\n\t\t\t\treturn\n\t\t\t}\n\t\tdefault:\n\t\t\tencoding = base64Enc\n\t\t\tbreak\n\t\t}\n\t}\n\tswitch encoding {\n\tcase base64Enc:\n\t\tbuf.WriteString(\"|\" + base64Encoding.EncodeToString(acc) + \"|\")\n\tcase tokenEnc:\n\t\tbuf.Write(acc)\n\tdefault:\n\t\tpanic(\"Encoding is neither base64 nor token\")\n\t}\n\n}", "func (c *Context) String(s string) {\n\tconst eom = \"\\n\"\n\tif !strings.HasSuffix(s, eom) {\n\t\t// sends it now, ending the message.\n\t\ts += eom\n\t}\n\t_, err := c.writer.WriteString(s)\n\tif err != nil {\n\t\tc.Error(err)\n\t}\n}", "func (e *encoder) encodeString(s string, color []byte) {\n\tif color != nil {\n\t\tsetColor(e.w, color)\n\t}\n\te.w.WriteByte('\"')\n\tstart := 0\n\tfor i := 0; i < len(s); {\n\t\tif b := s[i]; b < utf8.RuneSelf {\n\t\t\tif ' ' <= b && b <= '~' && b != '\"' && b != '\\\\' {\n\t\t\t\ti++\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif start < i {\n\t\t\t\te.w.WriteString(s[start:i])\n\t\t\t}\n\t\t\tswitch b {\n\t\t\tcase '\"':\n\t\t\t\te.w.WriteString(`\\\"`)\n\t\t\tcase '\\\\':\n\t\t\t\te.w.WriteString(`\\\\`)\n\t\t\tcase '\\b':\n\t\t\t\te.w.WriteString(`\\b`)\n\t\t\tcase '\\f':\n\t\t\t\te.w.WriteString(`\\f`)\n\t\t\tcase '\\n':\n\t\t\t\te.w.WriteString(`\\n`)\n\t\t\tcase '\\r':\n\t\t\t\te.w.WriteString(`\\r`)\n\t\t\tcase '\\t':\n\t\t\t\te.w.WriteString(`\\t`)\n\t\t\tdefault:\n\t\t\t\tconst hex = \"0123456789abcdef\"\n\t\t\t\te.w.WriteString(`\\u00`)\n\t\t\t\te.w.WriteByte(hex[b>>4])\n\t\t\t\te.w.WriteByte(hex[b&0xF])\n\t\t\t}\n\t\t\ti++\n\t\t\tstart = i\n\t\t\tcontinue\n\t\t}\n\t\tc, size := utf8.DecodeRuneInString(s[i:])\n\t\tif c == utf8.RuneError && size == 1 {\n\t\t\tif start < i {\n\t\t\t\te.w.WriteString(s[start:i])\n\t\t\t}\n\t\t\te.w.WriteString(`\\ufffd`)\n\t\t\ti += size\n\t\t\tstart = i\n\t\t\tcontinue\n\t\t}\n\t\ti += size\n\t}\n\tif start < len(s) {\n\t\te.w.WriteString(s[start:])\n\t}\n\te.w.WriteByte('\"')\n\tif color != nil {\n\t\tsetColor(e.w, resetColor)\n\t}\n}", "func (ctx *Context) WriteString(content string) {\n\tctx.ResponseWriter.Write([]byte(content))\n}", "func (l *LogCache) WriteString(p string) (n int, err error) {\n\tnow := time.Now().Local().Format(\"2006-01-02 15:04:05: \")\n\tw := now + p + \"\\n\"\n\tl.Write([]byte(w))\n\t// l.buffer = append(l.buffer, []byte(w)...)\n\t// n, err = l.buf.WriteString(w)\n\treturn\n}", "func (t *Token) writeString(s string) {\n\tt.strBuilder.WriteString(s)\n}", "func (dw *DataWriter) WriteUTF(value string) error {\n\treturn dw.WriteString(value)\n}", "func WriteString(line string, filename string) error {\n\treturn WriteStrings([]string{line}, filename, \"\")\n}", "func WriteString(p thrift.TProtocol, value, name string, field int16) error {\n\treturn WriteStringWithContext(context.Background(), p, value, name, field)\n}", "func (b *Bytes) WriteString(s string) (int, error) {\n\tn := b.Len()\n\tb.grow(n + len(s))\n\tcopy((*b.p)[n:], s)\n\treturn len(s), nil\n}", "func (xs *Sheet) WriteStr(row int, col int, value string, format *Format) int {\n\tfo := uintptr(0)\n\tif nil != format {\n\t\tfo = format.self\n\t}\n\ttmp, _, _ := xs.xb.lib.NewProc(\"xlSheetWriteStrW\").\n\t\tCall(xs.self, I(row), I(col), S(value), fo)\n\treturn int(tmp)\n}", "func drawASCII(text string) {\n\tio.Copy(os.Stdout, strings.NewReader(text))\n}", "func (ctx *Context) WriteString(content string) *HTTPError {\n\tctx.setDefaultHeaders()\n\tctx.SetHeader(\"Content-Length\", strconv.Itoa(len(content)), true)\n\t// set the default content-type\n\tctx.WriteHeader(http.StatusOK)\n\n\tif _, err := ctx.ResponseWriter.Write([]byte(content)); err != nil {\n\t\treturn serverError(err)\n\t}\n\treturn nil\n}", "func (c *Connection) WriteLine(raw string) {\n\tc.conn.Writer.PrintfLine(\"%s\", raw)\n}", "func (r *RotatingFile) WriteString(s string) (int, error) {\n\tdefer r.lock.Unlock()\n\tr.lock.Lock()\n\treturn r.file.WriteString(s)\n}", "func (ed *Editor) putString(s string) {\n\ted.buffer.ClearSel(ed.dot)\n\taddr := ed.buffer.InsertString(ed.dot.From, s)\n\ted.dot.To = addr\n}", "func (w *Writer) WriteString(v string) error {\n\t_, err := w.out.Write([]byte(v))\n\treturn err\n}", "func (tw *terminalWriter) Write(p []byte) (n int, err error) {\n\tif len(p) == 0 {\n\t\treturn 0, nil\n\t}\n\tif tw.emulator == nil {\n\t\treturn tw.wrappedWriter.Write(p)\n\t}\n\t// Emulate terminal by extracting commands and executing them\n\ttotalWritten := 0\n\tstart := 0 // indicates start of the next chunk\n\tend := len(p)\n\tfor current := 0; current < end; current++ {\n\t\tif tw.inSequence {\n\t\t\t// inside escape sequence\n\t\t\ttw.command = append(tw.command, p[current])\n\t\t\tif isAnsiCommandChar(p[current]) {\n\t\t\t\tif !isXtermOscSequence(tw.command, p[current]) {\n\t\t\t\t\t// found the last command character.\n\t\t\t\t\t// Now we have a complete command.\n\t\t\t\t\tnchar, err := tw.emulator.HandleOutputCommand(tw.fd, tw.command)\n\t\t\t\t\ttotalWritten += nchar\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn totalWritten, err\n\t\t\t\t\t}\n\n\t\t\t\t\t// clear the command\n\t\t\t\t\t// don't include current character again\n\t\t\t\t\ttw.command = tw.command[:0]\n\t\t\t\t\tstart = current + 1\n\t\t\t\t\ttw.inSequence = false\n\t\t\t\t}\n\t\t\t}\n\t\t} else {\n\t\t\tif p[current] == ANSI_ESCAPE_PRIMARY {\n\t\t\t\t// entering escape sequnce\n\t\t\t\ttw.inSequence = true\n\t\t\t\t// indicates end of \"normal sequence\", write whatever you have so far\n\t\t\t\tif len(p[start:current]) > 0 {\n\t\t\t\t\tnw, err := tw.emulator.WriteChars(tw.fd, tw.wrappedWriter, p[start:current])\n\t\t\t\t\ttotalWritten += nw\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn totalWritten, err\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\t// include the current character as part of the next sequence\n\t\t\t\ttw.command = append(tw.command, p[current])\n\t\t\t}\n\t\t}\n\t}\n\t// note that so far, start of the escape sequence triggers writing out of bytes to console.\n\t// For the part _after_ the end of last escape sequence, it is not written out yet. So write it out\n\tif !tw.inSequence {\n\t\t// assumption is that we can't be inside sequence and therefore command should be empty\n\t\tif len(p[start:]) > 0 {\n\t\t\tnw, err := tw.emulator.WriteChars(tw.fd, tw.wrappedWriter, p[start:])\n\t\t\ttotalWritten += nw\n\t\t\tif err != nil {\n\t\t\t\treturn totalWritten, err\n\t\t\t}\n\t\t}\n\t}\n\treturn totalWritten, nil\n\n}", "func Write(w http.ResponseWriter, status int, text string) {\n\tWriteBytes(w, status, []byte(text))\n}", "func responseString(rw http.ResponseWriter, statusCode int, s string) {\n\trw.Header().Set(\"Content-Type\", \"text/plain; charset=utf-8\")\n\trw.WriteHeader(statusCode)\n\trw.Write([]byte(s))\n}", "func puts(s tcell.Screen, style tcell.Style, x, y int, str string) {\n\ti := 0\n\tvar deferred []rune\n\tdwidth := 0\n\tzwj := false\n\tfor _, r := range str {\n\t\tif r == '\\u200d' {\n\t\t\tif len(deferred) == 0 {\n\t\t\t\tdeferred = append(deferred, ' ')\n\t\t\t\tdwidth = 1\n\t\t\t}\n\t\t\tdeferred = append(deferred, r)\n\t\t\tzwj = true\n\t\t\tcontinue\n\t\t}\n\t\tif zwj {\n\t\t\tdeferred = append(deferred, r)\n\t\t\tzwj = false\n\t\t\tcontinue\n\t\t}\n\t\tswitch runewidth.RuneWidth(r) {\n\t\tcase 0:\n\t\t\tif len(deferred) == 0 {\n\t\t\t\tdeferred = append(deferred, ' ')\n\t\t\t\tdwidth = 1\n\t\t\t}\n\t\tcase 1:\n\t\t\tif len(deferred) != 0 {\n\t\t\t\ts.SetContent(x+i, y, deferred[0], deferred[1:], style)\n\t\t\t\ti += dwidth\n\t\t\t}\n\t\t\tdeferred = nil\n\t\t\tdwidth = 1\n\t\tcase 2:\n\t\t\tif len(deferred) != 0 {\n\t\t\t\ts.SetContent(x+i, y, deferred[0], deferred[1:], style)\n\t\t\t\ti += dwidth\n\t\t\t}\n\t\t\tdeferred = nil\n\t\t\tdwidth = 2\n\t\t}\n\t\tdeferred = append(deferred, r)\n\t}\n\tif len(deferred) != 0 {\n\t\ts.SetContent(x+i, y, deferred[0], deferred[1:], style)\n\t\ti += dwidth\n\t}\n}", "func (c *genericCatch) appendString(s string, pos Position) {\n\tc.line += s\n\n\t// if it ends in a newline\n\tif s[len(s)-1] == '\\n' {\n\t\tif !c.firstNewline && len(c.line) > 2 {\n\t\t\tc.firstNewline = true // start a new one if the previous one ended in newline\n\n\t\t\tafterTrim := strings.TrimLeft(c.line, \"\\t \")\n\t\t\tdifference := len(c.line) - len(afterTrim)\n\t\t\tif difference != 0 {\n\t\t\t\tc.removeIndent = c.line[:difference]\n\t\t\t}\n\t\t}\n\t\tc.finishLine()\n\t}\n\n\t// the location is empty, so this is the first item\n\tif len(c.positioned) == 0 {\n\t\tc.pushContent(s, pos)\n\t\treturn\n\t}\n\n\t// append an existing string\n\tswitch v := c.lastContent().(type) {\n\tcase string:\n\t\tif v != \"\" && v[len(v)-1] == '\\n' {\n\t\t\t// start a new one if the previous one ended in newline\n\t\t\tc.pushContent(s, pos)\n\t\t} else {\n\t\t\t// other append the current string\n\t\t\tc.positioned[len(c.positioned)-1].content = v + s\n\t\t}\n\tdefault:\n\t\tc.pushContent(s, pos)\n\t}\n}", "func WriteString(wtr io.Writer, s string) error {\n\treturn WriteBytes(wtr, []byte(s))\n}", "func (s *String) Write(p []byte) (n int, err error) {\n\ts.value += string(p)\n\treturn len(p), nil\n}", "func (a ReverseHttpFile) WriteString(s string) (int, error) {\n\treturn 0, syscall.EPERM\n}", "func (tf *Temp) WriteString(contents string) (int, error) {\n\ttf.Lock()\n\tdefer tf.Unlock()\n\n\twritten, err := tf.file.WriteString(contents)\n\treturn written, ex.New(err)\n}", "func (e *Escpos) WriteRaw(data []byte) (n int, err error) {\n\tif len(data) > 0 {\n\t\tif e.Verbose {\n\t\t\tlog.Println(\"Writing %d bytes: %s\\n\", len(data), data)\n\t\t}\n\t\te.dst.Write(data)\n\t} else {\n\t\tif e.Verbose {\n\t\t\tlog.Println(\"Wrote NO bytes\\n\")\n\t\t}\n\t}\n\n\treturn 0, nil\n}", "func (bw *BufWriter) String(val string) {\n\tif bw.Error != nil {\n\t\treturn\n\t}\n\tbw.stringBuf = String(val, bw.stringBuf[:0])\n\t_, bw.Error = bw.writer.Write(bw.stringBuf)\n}", "func String(w http.ResponseWriter, status int, format string,\n\targs ...interface{}) error {\n\tSetContentType(w, TextPlainCharsetUTF8)\n\tw.WriteHeader(status)\n\t_, err := fmt.Fprintf(w, format, args...)\n\treturn err\n}", "func (ctx *Context) WriteString(body string) {\n\tctx.Response.Body = ioutil.NopCloser(strings.NewReader(body))\n}", "func (d *Display) Write(txt string) *Display {\n\tfor _, char := range txt {\n\t\tswitch {\n\t\tcase char >= 32 && char <= 126:\n\t\t\td.sendData(uint8(char)) // code 32 to 126 from font table matchs exactly ASCII\n\t\tcase char == 216:\n\t\t\td.sendData(174) // Ø\n\t\tcase char == 224:\n\t\t\td.sendData(133) // à\n\t\tcase char == 226:\n\t\t\td.sendData(131) // â\n\t\tcase char == 228:\n\t\t\td.sendData(132) // ä\n\t\tcase char == 232:\n\t\t\td.sendData(138) // è\n\t\tcase char == 233:\n\t\t\td.sendData(130) // é\n\t\tcase char == 234:\n\t\t\td.sendData(136) // ê\n\t\tcase char == 235:\n\t\t\td.sendData(137) // ë\n\t\tcase char == 238:\n\t\t\td.sendData(140) // î\n\t\tcase char == 239:\n\t\t\td.sendData(139) // ï\n\t\tcase char == 241:\n\t\t\td.sendData(155) // ñ\n\t\tcase char == 244:\n\t\t\td.sendData(148) // ô\n\t\tcase char == 246:\n\t\t\td.sendData(149) // ö\n\t\tcase char == 248:\n\t\t\td.sendData(175) // ø\n\t\tcase char == 249:\n\t\t\td.sendData(151) // ù\n\t\tcase char == 251:\n\t\t\td.sendData(150) // û\n\t\tcase char == 252:\n\t\t\td.sendData(129) // ü\n\t\tcase char == 255:\n\t\t\td.sendData(152) // ÿ\n\t\tcase char == 231:\n\t\t\td.sendData(135) // ç\n\t\tdefault:\n\t\t\td.sendData(159) // ¿\n\t\t}\n\t}\n\treturn d\n}", "func (pb *PhilosopherBase) WriteString(s string) {\n\tforkState := \"\"\n\tswitch {\n\tcase pb.HoldsFork(pb.LeftFork()) && pb.HoldsFork(pb.RightFork()):\n\t\tforkState = fmt.Sprintf(\", holds forks %d and %d\", pb.leftForkID(), pb.rightForkID())\n\tcase pb.HoldsFork(pb.LeftFork()):\n\t\tforkState = fmt.Sprintf(\", holds fork %d\", pb.leftForkID())\n\tcase pb.HoldsFork(pb.RightFork()):\n\t\tforkState = fmt.Sprintf(\", holds fork %d\", pb.rightForkID())\n\t}\n\n\tscreen.WriteScreenLine(ScreenPos+pb.ID, 1, fmt.Sprintf(\"%s (%d,%s) %s%s\", pb.Name, pb.ID, pb.State, s, forkState))\n}", "func (w *RESPWriter) writeStr(s string) {\n\tw.buf.WriteRune(respSimpleString)\n\tw.buf.WriteString(s)\n\tw.buf.Write(DELIMS)\n}", "func putString(log log.T, byteArray []byte, offsetStart int, offsetEnd int, inputString string) (err error) {\n\tbyteArrayLength := len(byteArray)\n\tif offsetStart > byteArrayLength-1 || offsetEnd > byteArrayLength-1 || offsetStart > offsetEnd || offsetStart < 0 {\n\t\tlog.Error(\"putString failed: Offset is invalid.\")\n\t\treturn errors.New(\"Offset is outside the byte array.\")\n\t}\n\n\tif offsetEnd-offsetStart+1 < len(inputString) {\n\t\tlog.Error(\"putString failed: Not enough space to save the string.\")\n\t\treturn errors.New(\"Not enough space to save the string.\")\n\t}\n\n\t// wipe out the array location first and then insert the new value.\n\tfor i := offsetStart; i <= offsetEnd; i++ {\n\t\tbyteArray[i] = ' '\n\t}\n\n\tcopy(byteArray[offsetStart:offsetEnd+1], inputString)\n\treturn nil\n}", "func (l *RedLogger) WriteString(s string) (n int, err error) {\n\treturn l.Write([]byte(s))\n}", "func WriteChar(char rune) {}", "func (b *printer) UnsafeString(s string) {\n\tw := escapeWriter{w: &b.buf, enclose: true, strip: true}\n\t_, _ = w.Write([]byte(s))\n}", "func (s stdout) WriteAt(p []byte, off int64) (n int, err error) {\n\treturn os.Stdout.Write(p)\n}", "func (e *Encoder) WriteString(s string) error {\n\te.prepareNext(String)\n\tvar err error\n\tif e.out, err = appendString(e.out, s); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *SodaClient) Write(sendMsg string) {\n\tmsg := strings.TrimSpace(sendMsg)\n\n\tbuf := []byte(msg)\n\n\t_, err := c.conn.Write(buf) // returns string length of write and potential write errors\n\n\tif err != nil {\n\t\tfmt.Println(msg, err)\n\t}\n}", "func (w *Worker) WriteString(s string) *Worker {\n\tw.WriteData <- s\n\treturn w\n}", "func putln(s tcell.Screen, style tcell.Style, str string, row int) {\n\tputs(s, style, 0, row, str)\n\trow++\n}", "func (r *renderer) write(s string, unescaped bool) {\n\tif r.indentNext {\n\t\tr.indentNext = false\n\t\tr.w.WriteString(r.indent)\n\t}\n\tif !unescaped {\n\t\ts = html.EscapeString(s)\n\t}\n\tr.w.WriteString(s)\n}", "func (e *Engine) WritePointsString(ptstr ...string) error {\n\tpoints, err := models.ParsePointsString(strings.Join(ptstr, \"\\n\"))\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn e.writePoints(points...)\n}", "func (t *TextGame) drawStr(x int, y int, str string, textColor tcell.Style) {\n\tsizeX, sizeY := t.screen.Size()\n\tif (x < 0) || (y < 0) || (y > sizeY) {\n\t\treturn\n\t}\n\tfor row := 0; row < len(str); row++ {\n\t\tscreenX := x + row\n\t\tif screenX > sizeX {\n\t\t\tbreak\n\t\t}\n\t\tt.screen.SetContent(screenX, y, rune(str[row]), nil, textColor)\n\t}\n}", "func (s *Status) WriteText(w http.ResponseWriter) error {\n\tw.Header().Set(\"Content-Type\", \"text/plain\")\n\tw.WriteHeader(s.Code)\n\t_, err := io.WriteString(w, s.String())\n\treturn err\n}", "func (i *IRC) writeLine(command string, args ...string) (int, error) {\n\tl := util.MakeSimpleIRCLine(command, args...)\n\tlBytes, err := l.LineBytes()\n\n\tif err != nil {\n\t\treturn -1, err\n\t}\n\n\treturn i.write(lBytes, !(command == \"PING\" && i.SuppressPing))\n}", "func (b *printer) SafeString(s SafeString) {\n\tw := escapeWriter{w: &b.buf, enclose: false}\n\t_, _ = w.Write([]byte(s))\n}", "func (c *Ctx) RenderString(code int, s string) {\n\tc.W.WriteHeader(code)\n\tc.W.Write([]byte(s))\n}", "func (w PrintWriter) Write(p []byte) (int, error) {\n\tw.P(string(bytes.TrimSuffix(p, []byte(\"\\n\"))))\n\treturn len(p), nil\n}", "func quoteString(w stringWriter, s string) {\n\tmustWriteByte(w, '\"')\n\tconst hex = \"0123456789abcdef\"\n\tstart := 0\n\tfor i := 0; i < len(s); {\n\t\t// Single-byte code points.\n\t\tif b := s[i]; b < utf8.RuneSelf {\n\t\t\tif b >= ' ' && b != '\"' && b != '\\\\' {\n\t\t\t\ti++\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif start < i {\n\t\t\t\tmustWriteString(w, s[start:i])\n\t\t\t}\n\t\t\tswitch b {\n\t\t\tcase '\\\\', '\"':\n\t\t\t\tmustWriteByte(w, '\\\\')\n\t\t\t\tmustWriteByte(w, b)\n\t\t\tcase '\\n':\n\t\t\t\tmustWriteByte(w, '\\\\')\n\t\t\t\tmustWriteByte(w, 'n')\n\t\t\tcase '\\r':\n\t\t\t\tmustWriteByte(w, '\\\\')\n\t\t\t\tmustWriteByte(w, 'r')\n\t\t\tcase '\\t':\n\t\t\t\tmustWriteByte(w, '\\\\')\n\t\t\t\tmustWriteByte(w, 't')\n\t\t\tdefault:\n\t\t\t\t// This encodes bytes < 0x20 except for \\t, \\n and \\r.\n\t\t\t\tmustWriteString(w, `\\u00`)\n\t\t\t\tmustWriteByte(w, hex[b>>4])\n\t\t\t\tmustWriteByte(w, hex[b&0xF])\n\t\t\t}\n\t\t\ti++\n\t\t\tstart = i\n\t\t\tcontinue\n\t\t}\n\t\t// Multi-byte code points.\n\t\tc, size := utf8.DecodeRuneInString(s[i:])\n\t\tif c == utf8.RuneError && size == 1 {\n\t\t\t// Transform invalid code points into unicode\n\t\t\t// \"replacement character\".\n\t\t\tif start < i {\n\t\t\t\tmustWriteString(w, s[start:i])\n\t\t\t}\n\t\t\tmustWriteString(w, `\\ufffd`)\n\t\t\ti += size\n\t\t\tstart = i\n\t\t\tcontinue\n\t\t}\n\t\t// U+2028 is LINE SEPARATOR.\n\t\t// U+2029 is PARAGRAPH SEPARATOR.\n\t\t// They are both technically valid characters in JSON strings,\n\t\t// but don't work in JSONP, which has to be evaluated as JavaScript,\n\t\t// and can lead to security holes there. It is valid JSON to\n\t\t// escape them, so we do so unconditionally.\n\t\t// See http://timelessrepo.com/json-isnt-a-javascript-subset for discussion.\n\t\tif c == '\\u2028' || c == '\\u2029' {\n\t\t\tif start < i {\n\t\t\t\tmustWriteString(w, s[start:i])\n\t\t\t}\n\t\t\tmustWriteString(w, `\\u202`)\n\t\t\tmustWriteByte(w, hex[c&0xF])\n\t\t\ti += size\n\t\t\tstart = i\n\t\t\tcontinue\n\t\t}\n\t\ti += size\n\t}\n\tif start < len(s) {\n\t\tmustWriteString(w, s[start:])\n\t}\n\tmustWriteByte(w, '\"')\n}" ]
[ "0.7148465", "0.68846196", "0.6405198", "0.63903034", "0.63707054", "0.6252102", "0.62228155", "0.6213513", "0.61829394", "0.6159723", "0.61487716", "0.6142807", "0.6104324", "0.6083521", "0.6053321", "0.60494804", "0.59812474", "0.59790057", "0.59725416", "0.5962015", "0.5931091", "0.5903109", "0.58810264", "0.58547485", "0.5853804", "0.584679", "0.5827082", "0.5810823", "0.57887936", "0.5764917", "0.5756591", "0.57534593", "0.5734897", "0.572531", "0.5708331", "0.5695969", "0.5695828", "0.5654389", "0.563187", "0.5629236", "0.5613827", "0.5580804", "0.55747813", "0.5571958", "0.5568782", "0.5562181", "0.55538", "0.5536249", "0.5530105", "0.5528905", "0.55139196", "0.549532", "0.5493346", "0.5489989", "0.54852307", "0.5473126", "0.5468874", "0.546153", "0.545594", "0.5454994", "0.5449373", "0.5442471", "0.54291", "0.5427089", "0.5424086", "0.54098916", "0.5398294", "0.5396626", "0.5396228", "0.53841513", "0.53721255", "0.536501", "0.53641725", "0.53494865", "0.5343484", "0.53305525", "0.53253496", "0.53170663", "0.53099245", "0.53032494", "0.53029644", "0.5299773", "0.52662563", "0.5236616", "0.5223877", "0.5223821", "0.5220725", "0.52189314", "0.52066517", "0.5204058", "0.5201115", "0.5189886", "0.5188123", "0.5186921", "0.5173045", "0.5169155", "0.51652044", "0.5161109", "0.51569945", "0.5124608" ]
0.63136524
5
Format a log entry.
func (hd *HappyDevFormatter) Format(writer io.Writer, level int, msg string, args []interface{}) { buf := pool.Get() defer pool.Put(buf) if len(args) == 1 { args = append(args, 0) copy(args[1:], args[0:]) args[0] = singleArgKey } // warn about reserved, bad and complex keys for i := 0; i < len(args); i += 2 { isReserved, err := isReservedKey(args[i]) if err != nil { InternalLog.Error("Key is not a string.", "err", fmt.Errorf("args[%d]=%v", i, args[i])) } else if isReserved { InternalLog.Fatal("Key conflicts with reserved key. Avoiding using single rune keys.", "key", args[i].(string)) } else { // Ensure keys are simple strings. The JSONFormatter doesn't escape // keys as a performance tradeoff. This panics if the JSON key // value has a different value than a simple quoted string. key, isOK := args[i].(string) if !isOK { panic("Key is invalid or was omitted") } for _, aRune := range key { if aRune == '"' || aRune == '\\' { panic("Key is invalid. Use simpler key for: " + fmt.Sprintf("%q", key)) } if int(aRune) < 32 { panic("Key contains control character. Use simpler key for: " + fmt.Sprintf("%q", key)) } } b, err := json.Marshal(key) if err != nil { panic("Key is invalid. " + err.Error()) } if string(b) != `"`+key+`"` { panic("Key is complex. Use simpler key for: " + fmt.Sprintf("%q", key)) } } } // use the production JSON formatter to format the log first. This // ensures JSON will marshal/unmarshal correctly in production. entry := hd.jsonFormatter.LogEntry(level, msg, args) // reset the column tracker used for fancy formatting col := 0 // timestamp buf.WriteString(theme.Misc) hd.writeString(buf, entry[KeyMap.Time].(string), &col) if !disableColors { buf.WriteString(ansi.Reset) } // emphasize warnings and errors message, context, color := hd.getLevelContext(level, entry) if message == "" { message = entry[KeyMap.Message].(string) } // DBG, INF ... hd.set(buf, "", entry[KeyMap.Level].(string), color, &col) // logger name hd.set(buf, "", entry[KeyMap.Name], theme.Misc, &col) // message from user hd.set(buf, "", message, theme.Message, &col) // Preserve key order in the sequencethey were added by developer.This // makes it easier for developers to follow the log. order := []string{} lenArgs := len(args) for i := 0; i < len(args); i += 2 { if i+1 >= lenArgs { continue } if key, ok := args[i].(string); ok { order = append(order, key) } else { order = append(order, badKeyAtIndex(i)) } } for _, key := range order { // skip reserved keys which were already added to buffer above isReserved, err := isReservedKey(key) if err != nil { panic("key is invalid. Should never get here. " + err.Error()) } else if isReserved { continue } hd.set(buf, key, entry[key], theme.Value, &col) } addLF := true hasCallStack := entry[KeyMap.CallStack] != nil // WRN,ERR file, line number context if context != "" && context != "\n" { // warnings and traces are single line, space can be optimized if level == LevelTrace || (level == LevelWarn && !hasCallStack) { // gets rid of "in " idx := strings.IndexRune(context, 'n') hd.set(buf, "in", context[idx+2:], theme.Value, &col) } else { buf.WriteRune('\n') if !disableColors { buf.WriteString(color) } addLF = context[len(context)-1:] != "\n" buf.WriteString(context) if !disableColors { buf.WriteString(ansi.Reset) } } } else if hasCallStack { hd.set(buf, "", entry[KeyMap.CallStack], color, &col) } if addLF { buf.WriteRune('\n') } buf.WriteTo(writer) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (h *AuditLogHandler) Format(e *log.Entry) ([]byte, error) {\n\th.entry = e\n\treturn h.formatter.Format(e)\n}", "func (f *formatter) Format(entry *logrus.Entry) ([]byte, error) {\n\toutput := f.LogFormat\n\tif output == \"\" {\n\t\toutput = defaultLogFormat\n\t}\n\n\ttimestampFormat := f.TimestampFormat\n\tif timestampFormat == \"\" {\n\t\ttimestampFormat = defaultTimestampFormat\n\t}\n\n\toutput = strings.Replace(output, \"%time%\", entry.Time.Format(timestampFormat), 1)\n\n\tlevel := strings.ToUpper(entry.Level.String())[0:1]\n\toutput = strings.Replace(output, \"%lvl%\", level, 1)\n\n\tif strings.Contains(output, locationTag) {\n\t\toutput = strings.Replace(output, locationTag, f.when(locationSkip), 1)\n\t}\n\n\toutput = strings.Replace(output, \"%msg%\", entry.Message, 1)\n\n\tfor k, v := range entry.Data {\n\t\tif s, ok := v.(string); ok {\n\t\t\toutput = strings.Replace(output, \"%\"+k+\"%\", s, 1)\n\t\t}\n\t}\n\n\treturn []byte(output), nil\n}", "func (l *logStashFormatter) Format(entry *log.Entry) ([]byte, error) {\n\treturn []byte(\n\t\tfmt.Sprintf(\"%s - fees_service - %s - %s\",\n\t\t\tentry.Time.Format(l.TimestampFormat),\n\t\t\tstrings.ToUpper(entry.Level.String()),\n\t\t\tentry.Message),\n\t), nil\n}", "func (gf *gohilLogFormatter) Format(entry *logrus.Entry) ([]byte, error) {\n\tif entry == nil {\n\t\treturn nil, errors.New(\"logger entry is nil\")\n\t}\n\n\tif entry.Buffer == nil {\n\t\treturn nil, errors.New(\"logger empty entry buffer\")\n\t}\n\n\ttime := entry.Time.UTC().Format(rfc3339Milli)\n\tlevel := strings.ToUpper(entry.Level.String())\n\n\tlogLineData := entry.Buffer\n\tlogLineData.WriteString(\n\t\tfmt.Sprintf(\"%s %s %s\\n\", time, level, entry.Message))\n\n\treturn logLineData.Bytes(), nil\n}", "func (f *LogFormatter) Format(entry *logrus.Entry) ([]byte, error) {\n\tif f.frm != nil {\n\t\treturn f.frm(entry)\n\t}\n\n\tgenericFormatter := logrus.TextFormatter{PadLevelText: true, TimestampFormat: \"1-2|15:04:05.000\", FullTimestamp: true}\n\n\treturn genericFormatter.Format(entry)\n}", "func (l *Line) Format(entry logger.Entry) string {\n\treturn fmt.Sprintf(l.format, entry.Message, entry.Level, entry.Context)\n}", "func (fmttr *Formatter) Format(entry *logrus.Entry) ([]byte, error) {\n\n\tfmttr.terminalInitOnce.Do(func() {\n\t\terr := fmttr.initTermInfo(entry)\n\t\tif err != nil {\n\t\t\tfmt.Printf(\"Failed to initialize terminal with err %s\", err)\n\t\t}\n\t})\n\n\ttimeFormat := fmttr.DateTimeFormat\n\n\ttpl := \"\\n%s [%s]: %s\"\n\n\t// Get the three fields that _aren't_ part of entry.Data\n\t// Every time you format a date in Golang, a kitten weeps\n\tts := entry.Time.Format(timeFormat)\n\tmsg := entry.Message\n\tlevel := entry.Level\n\n\tline := fmt.Sprintf(tpl, ts, level, msg)\n\tfields, err := fmttr.renderFields(entry)\n\tif err != nil {\n\t\treturn []byte{}, err\n\t}\n\n\terrMsg, err := fmttr.renderError(entry)\n\tif err != nil {\n\t\treturn []byte{}, err\n\t}\n\n\treturn []byte(line + fields + errMsg), nil\n}", "func (f *ServiceFormatter) Format(e *log.Entry) ([]byte, error) {\n\te.Data[\"epochTimeMillis\"] = e.Time.UnixNano() / int64(time.Millisecond)\n\te.Data[\"service\"] = f.svcName\n\treturn f.Formatter.Format(e)\n}", "func (f *Formatter) Format(entry *logrus.Entry) ([]byte, error) {\n\tvar b *bytes.Buffer\n\tif entry.Buffer != nil {\n\t\tb = entry.Buffer\n\t} else {\n\t\tb = &bytes.Buffer{}\n\t}\n\tb.WriteString(entry.Message + \"\\n\")\n\treturn b.Bytes(), nil\n}", "func (f *textFormatter) Format(entry *logrus.Entry) ([]byte, error) {\n\tb := &bytes.Buffer{}\n\n\tlevelText := strings.ToUpper(entry.Level.String())[0:4]\n\ttimeStamp := entry.Time.Format(\"2006/01/02 15:04:05.000000\")\n\tif len(entry.Data) > 0 {\n\t\tfmt.Fprintf(b, \"%s: %s %-44s \", levelText, timeStamp, entry.Message)\n\t\tfor k, v := range entry.Data {\n\t\t\tfmt.Fprintf(b, \" %s=%v\", k, v)\n\t\t}\n\t} else {\n\t\t// No padding when there's no fields\n\t\tfmt.Fprintf(b, \"%s: %s %s\", levelText, timeStamp, entry.Message)\n\t}\n\n\tb.WriteByte('\\n')\n\treturn b.Bytes(), nil\n}", "func (f *CLIFormatter) Format(entry *logrus.Entry) ([]byte, error) {\n\treturn []byte(entry.Message + \"\\n\"), nil\n}", "func (f *MyFormatter) Format(entry *logrus.Entry) ([]byte, error) {\n\tvar b *bytes.Buffer\n\tkeys := make([]string, 0, len(entry.Data))\n\tfor k := range entry.Data {\n\t\tkeys = append(keys, k)\n\t}\n\tsort.Strings(keys)\n\n\t// if entry.Buffer != nil {\n\t// \tb = entry.Buffer\n\t// } else {\n\tb = &bytes.Buffer{}\n\t// }\n\n\tlevelColor := getPrintColored(entry)\n\tlevelText := strings.ToUpper(entry.Level.String())\n\n\tdepth := getDepth()\n\tif depth > 0 {\n\t\tdepth = depth - 1\n\t}\n\tpc, file, line, _ := runtime.Caller(depth)\n\n\tpcName := runtime.FuncForPC(pc).Name()\n\tindex := strings.LastIndex(pcName, \".\")\n\tpkg := \"\"\n\tmodel := \"\"\n\tif index != -1 {\n\t\tpkg = pcName[:index]\n\t\tmodel = pcName[index+1:]\n\t} else {\n\t\tpkg = pcName\n\t\tmodel = pcName\n\t}\n\n\tpaths := strings.Split(file, \"/\")\n\tfile = paths[len(paths)-1]\n\n\tfmtStr := \"\"\n\tif logrus.IsTerminal() {\n\t\tfmtStr = \"\\x1b[%dm[%s] %s%d %s [%s] %-4s\\x1b[0m\"\n\t\tfmt.Fprintf(b, fmtStr, levelColor, entry.Time.Format(time.RFC3339), file+\":\", line, pkg, model, levelText)\n\t} else {\n\t\tfmtStr = \"[%s] %s%d %s [%s] %-4s\"\n\t\tfmt.Fprintf(b, fmtStr, entry.Time.Format(time.RFC3339), file+\":\", line, pkg, model, levelText)\n\t}\n\n\tfor _, k := range keys {\n\t\tv := entry.Data[k]\n\t\tif logrus.IsTerminal() {\n\t\t\tfmt.Fprintf(b, \" \\x1b[%dm%s\\x1b[0m=\", levelColor, k)\n\t\t} else {\n\t\t\tfmt.Fprintf(b, \" %s\", k)\n\t\t}\n\n\t\tf.appendValue(b, v)\n\t}\n\n\tfmt.Fprintf(b, \" %-44s \\n\", entry.Message)\n\n\treturn b.Bytes(), nil\n}", "func (locale *locale) Format(log *logrus.Entry) ([]byte, error) {\n\tlog.Time = log.Time.In(locale.Location)\n\n\treturn locale.Formatter.Format(log)\n}", "func (f *StandardFormatter) Format(loggerName string, log LogEvent) (formattedLog string, err error) {\n\tf.mutex.RLock()\n\tdefer f.mutex.RUnlock()\n\tlogMessage := log.Message()\n\tif f.appendNewLine {\n\t\tif len(logMessage) == 0 || logMessage[len(logMessage) - 1:] != \"\\n\" {\n\t\t\tlogMessage = logMessage + \"\\n\"\n\t\t}\n\t}\n\treplacer := strings.NewReplacer(\n\t\t\"%(dateTime)\", log.Time().Format(f.dateTimeLayout),\n\t\t\"%(logLevel)\", log.LogLevel(),\n\t\t\"%(logLevelNum)\", strconv.Itoa(int(log.LogLevelNum())),\n\t\t\"%(program)\", log.Program(),\n\t\t\"%(pid)\", strconv.Itoa(log.Pid()),\n\t\t\"%(hostname)\", log.Hostname(),\n\t\t\"%(loggerName)\", loggerName,\n\t\t\"%(programCounter)\", strconv.FormatUint(uint64(log.Pc()), 16),\n\t\t\"%(fileName)\", log.FileName(),\n\t\t\"%(shortFileName)\", filepath.Base(log.FileName()),\n\t\t\"%(lineNum)\", strconv.Itoa(log.LineNum()),\n\t\t\"%(message)\", logMessage)\n\treturn replacer.Replace(f.layout), nil\n}", "func (f *ServerFormatter) Format(entry *lgrus.Entry) ([]byte, error) {\n\n\tkeys := make([]string, 0, len(entry.Data))\n\tfor k := range entry.Data {\n\t\tkeys = append(keys, k)\n\t}\n\n\tvar b *bytes.Buffer\n\tif entry.Buffer != nil {\n\t\tb = entry.Buffer\n\t} else {\n\t\tb = &bytes.Buffer{}\n\t}\n\n\tf.printLog(b, entry, keys)\n\n\tb.WriteByte('\\n')\n\treturn b.Bytes(), nil\n}", "func (f *klogFormatter) Format(entry *logrus.Entry) ([]byte, error) {\n\tb := &bytes.Buffer{}\n\tfilePath := entry.Caller.File\n\tfilePathArray := strings.Split(filePath, \"/\")\n\tfileName := filePathArray[len(filePathArray)-1]\n\tpidString := strconv.Itoa(os.Getpid())\n\n\t// logrus has seven logging levels: Trace, Debug, Info, Warning, Error, Fatal and Panic.\n\tb.WriteString(strings.ToUpper(entry.Level.String()[:1]))\n\tb.WriteString(entry.Time.Format(\"0102 15:04:05.000000\"))\n\tb.WriteString(\" \")\n\tfor i := 0; i < (7 - len(pidString)); i++ {\n\t\tb.WriteString(\" \")\n\t}\n\tb.WriteString(pidString)\n\tb.WriteString(\" \")\n\tb.WriteString(fileName)\n\tb.WriteString(\":\")\n\tfmt.Fprint(b, entry.Caller.Line)\n\tb.WriteString(\"] \")\n\tif entry.Message != \"\" {\n\t\tb.WriteString(entry.Message)\n\t}\n\tb.WriteByte('\\n')\n\n\treturn b.Bytes(), nil\n}", "func (formatter *LogEntriesDropperFormatter) Format(entry *logrus.Entry) ([]byte, error) {\n\tif entry.Logger.Level >= entry.Level {\n\t\treturn formatter.OriginalFormatter.Format(entry)\n\t}\n\treturn []byte(\"\"), nil\n}", "func (f *TextFormatter) Format(entry *logrus.Entry) ([]byte, error) {\n\tvar b *bytes.Buffer\n\tkeys := make([]string, 0, len(entry.Data))\n\tfor k := range entry.Data {\n\t\tkeys = append(keys, k)\n\t}\n\n\tif !f.DisableSorting {\n\t\tsort.Strings(keys)\n\t}\n\tif entry.Buffer != nil {\n\t\tb = entry.Buffer\n\t} else {\n\t\tb = &bytes.Buffer{}\n\t}\n\n\ttimestampFormat := f.TimestampFormat\n\tif timestampFormat == \"\" {\n\t\ttimestampFormat = defaultTimestampFormat\n\t}\n\n\tif !f.DisableTimestamp {\n\t\tf.appendKeyValue(b, \"time\", entry.Time.Format(timestampFormat))\n\t}\n\tf.appendKeyValue(b, \"level\", entry.Level.String())\n\tif entry.Message != \"\" {\n\t\tf.appendKeyValue(b, \"msg\", entry.Message)\n\t}\n\tfor _, key := range keys {\n\t\tf.appendKeyValue(b, key, entry.Data[key])\n\t}\n\n\tb.WriteString(\"\\r\\n\")\n\treturn b.Bytes(), nil\n}", "func (f *TextFormatter) Format(entry *logrus.Entry) ([]byte, error) {\n\tkeys := make([]string, 0, len(entry.Data))\n\tlevelValue := FieldValueLevel\n\tfor k := range entry.Data {\n\t\tif k == FieldKeyLevel {\n\t\t\tlevelValue = entry.Data[k].(string)\n\t\t}\n\t\tkeys = append(keys, k)\n\t}\n\n\tfixedKeys := make([]string, 0, 3+len(entry.Data))\n\tif !f.DisableTimestamp {\n\t\tfixedKeys = append(fixedKeys, FieldKeyTime)\n\t}\n\n\tif levelValue != FieldValueLevel {\n\t\tfixedKeys = append(fixedKeys, FieldKeyLevel)\n\t}\n\n\tif _, ok := entry.Data[FieldKeyMsg]; entry.Message != \"\" && !ok {\n\t\tfixedKeys = append(fixedKeys, FieldKeyMsg)\n\t}\n\n\tfixedKeys = append(fixedKeys, keys...)\n\n\tvar b *bytes.Buffer\n\tif entry.Buffer != nil {\n\t\tb = entry.Buffer\n\t} else {\n\t\tb = &bytes.Buffer{}\n\t}\n\n\tf.terminalInitOnce.Do(func() { f.init(entry) })\n\n\tif f.TimestampFormat == \"\" {\n\t\tf.TimestampFormat = defaultTimestampFormat\n\t}\n\tif f.isColored() {\n\t\tf.printColored(b, entry, keys, f.TimestampFormat, levelValue)\n\t} else {\n\t\tfor _, key := range fixedKeys {\n\t\t\tvar value interface{}\n\t\t\tswitch key {\n\t\t\tcase FieldKeyTime:\n\t\t\t\tvalue = entry.Time.Format(f.TimestampFormat)\n\t\t\tcase FieldKeyLevel:\n\t\t\t\tvalue = levelValue\n\t\t\tcase FieldKeyMsg:\n\t\t\t\tif entry.Message != \"\" {\n\t\t\t\t\tvalue = entry.Message\n\t\t\t\t} else {\n\t\t\t\t\tvalue = entry.Data[key]\n\t\t\t\t}\n\t\t\tdefault:\n\t\t\t\tvalue = entry.Data[key]\n\t\t\t}\n\t\t\tf.appendKeyValue(b, key, value)\n\t\t}\n\t}\n\n\tb.WriteByte('\\n')\n\treturn b.Bytes(), nil\n}", "func (f *RuniacFormatter) Format(entry *logrus.Entry) ([]byte, error) {\n\n\tvar b *bytes.Buffer\n\tif entry.Buffer != nil {\n\t\tb = entry.Buffer\n\t} else {\n\t\tb = &bytes.Buffer{}\n\t}\n\n\tif f.isColored() {\n\t\tf.prependColored(b, entry.Level)\n\t} else {\n\t\t_, _ = fmt.Fprintf(b, \"[%s] \", strings.ToUpper(entry.Level.String()))\n\t}\n\n\tif _, ok := entry.Data[\"action\"]; ok {\n\n\t\tstepId := []string{}\n\n\t\tstepId = appendIfSet(stepId, entry, \"track\")\n\t\tstepId = appendIfSet(stepId, entry, \"step\")\n\t\tstepId = appendIfSet(stepId, entry, \"regionDeployType\")\n\t\tstepId = appendIfSet(stepId, entry, \"region\")\n\n\t\t_, _ = fmt.Fprintf(b, \"(%s %s) \", entry.Data[\"action\"], strings.Join(stepId, \"/\"))\n\t}\n\n\t_, _ = fmt.Fprintf(b, \"%s\", entry.Message)\n\n\tif err, ok := entry.Data[\"error\"]; ok {\n\t\tb.WriteString(fmt.Sprintf(\" (%s)\", err))\n\t}\n\n\tif f.isColored() {\n\t\tf.postpendColored(b)\n\t}\n\n\tb.WriteByte('\\n')\n\n\treturn b.Bytes(), nil\n}", "func (cf *CustomFormatter) Format(entry *logrus.Entry) ([]byte, error) {\n\t// Define log Level colors and Level text\n\tvar levelColor int\n\tvar levelText string\n\n\tswitch entry.Level {\n\tcase logrus.TraceLevel:\n\t\tlevelText = \"TRACE\"\n\t\tlevelColor = lightGray\n\tcase logrus.DebugLevel:\n\t\tlevelText = \"DEBUG\"\n\t\tlevelColor = lightGray\n\tcase logrus.InfoLevel:\n\t\tlevelText = \"INFO\"\n\t\tlevelColor = cyan\n\tcase logrus.WarnLevel:\n\t\tlevelText = \"WARN\"\n\t\tlevelColor = yellow\n\tcase logrus.ErrorLevel:\n\t\tlevelText = \"ERROR\"\n\t\tlevelColor = red\n\tcase logrus.FatalLevel:\n\t\tlevelText = \"FATAL\"\n\t\tlevelColor = red\n\tcase logrus.PanicLevel:\n\t\tlevelText = \"PANIC\"\n\t\tlevelColor = red\n\tdefault:\n\t\tlevelText = \"NONE\"\n\t\tlevelColor = lightGray\n\t}\n\n\t// Removes all the break line at the end\n\tfor strings.HasSuffix(entry.Message, \"\\n\") {\n\t\tentry.Message = strings.TrimSuffix(entry.Message, \"\\n\")\n\t}\n\n\t// Format the message\n\tmsg := fmt.Sprintf(\"%s \\x1b[%dm%5s\\x1b[0m %s %s\\n\",\n\t\tentry.Time.Format(timeFormat),\n\t\tlevelColor,\n\t\tlevelText,\n\t\tcf.Command,\n\t\tentry.Message,\n\t)\n\n\treturn []byte(msg), nil\n}", "func (f *LineFormatter) Format(r *Record) {\n\treplacer := strings.NewReplacer(\n\t\t\"%datetime%\", r.Time,\n\t\t\"%channel%\", r.Channel,\n\t\t\"%level_name%\", r.LevelName,\n\t\t\"%message%\", r.Message,\n\t\t\"%extra%\", r.Extra.String())\n\n\tr.Formatted = replacer.Replace(f.LineFormat)\n}", "func (tf *moduleFormatter) Format(entry *logrus.Entry) ([]byte, error) {\n\tentry.Data[moduleLogKey] = fmt.Sprintf(\"%s->pid:%d\", tf.module, os.Getegid())\n\tentry.Data[sysIDLogKey] = viper.GetString(\"sys.id\")\n\tentry.Data[sysNameLogKey] = viper.GetString(\"sys.name\")\n\tentry.Data[svcIDLogKey] = serviceName\n\n\terrHappened := entry.Level == logrus.ErrorLevel || entry.Level == logrus.FatalLevel\n\tif errHappened {\n\t\tentry.Data[isCRLogKey] = entry.Level == logrus.FatalLevel\n\t}\n\n\tdata := make(logrus.Fields)\n\tfor k, v := range entry.Data {\n\t\tdata[k] = v\n\t}\n\n\tprefixFieldClashes(data, tf.FieldMap, entry.HasCaller())\n\tkeys := make([]string, 0, len(data))\n\tfor k := range data {\n\t\tkeys = append(keys, k)\n\t}\n\n\tfixedKeys := make([]string, 0, 3+len(data))\n\tif !tf.DisableTimestamp {\n\t\tfixedKeys = append(fixedKeys, resolve(tf.FieldMap, logrus.FieldKeyTime))\n\t}\n\tfixedKeys = append(fixedKeys, resolve(tf.FieldMap, logrus.FieldKeyLevel))\n\tfixedKeys = append(fixedKeys, resolve(tf.FieldMap, logrus.FieldKeyMsg))\n\n\tvar funcVal, fileVal string\n\tif entry.HasCaller() {\n\t\tif tf.CallerPrettyfier != nil {\n\t\t\tfuncVal, fileVal = tf.CallerPrettyfier(entry.Caller)\n\t\t} else {\n\t\t\tfuncVal = entry.Caller.Function\n\t\t\tfileVal = fmt.Sprintf(\"%s:%d\", entry.Caller.File, entry.Caller.Line)\n\t\t}\n\t}\n\n\tif errHappened {\n\t\tfixedKeys = append(fixedKeys, errMsgLogKey)\n\t}\n\n\tif entry.Level == logrus.WarnLevel {\n\t\tfixedKeys = append(fixedKeys, warnMsgLogKey)\n\t}\n\n\tfixedKeys = append(fixedKeys, keys...)\n\n\tvar b *bytes.Buffer\n\tif entry.Buffer != nil {\n\t\tb = entry.Buffer\n\t} else {\n\t\tb = &bytes.Buffer{}\n\t}\n\n\ttimestampFormat := tf.TimestampFormat\n\tif timestampFormat == \"\" {\n\t\ttimestampFormat = defaultTimeFormat\n\t}\n\tfor _, key := range fixedKeys {\n\t\tvar value interface{}\n\t\tswitch {\n\t\tcase key == resolve(tf.FieldMap, logrus.FieldKeyTime):\n\t\t\tvalue = entry.Time.Format(timestampFormat)\n\t\tcase key == resolve(tf.FieldMap, logrus.FieldKeyLevel):\n\t\t\tvalue = entry.Level.String()\n\t\tcase key == resolve(tf.FieldMap, logrus.FieldKeyMsg):\n\t\t\tif entry.Level >= logrus.InfoLevel {\n\t\t\t\tvalue = fmt.Sprintf(\"%s@%s:%s\", fileVal, funcVal, entry.Message)\n\t\t\t} else {\n\t\t\t\tvalue = fmt.Sprintf(\"%s@%s\", fileVal, funcVal)\n\t\t\t}\n\t\tcase key == errMsgLogKey:\n\t\t\tvar ok bool\n\t\t\tvalue, ok = data[logrusErrorKey]\n\t\t\tif ok {\n\t\t\t\tvalue = data[logrusErrorKey]\n\t\t\t}\n\n\t\t\tif entry.Message != \"\" {\n\t\t\t\tvalue = fmt.Sprintf(\"%v:%s\", value, entry.Message)\n\t\t\t}\n\t\tcase key == warnMsgLogKey:\n\t\t\tvalue = entry.Message\n\t\tcase key == logrusErrorKey:\n\t\t\tcontinue\n\t\tdefault:\n\t\t\tvalue = data[key]\n\t\t}\n\t\ttf.appendKeyValue(b, key, value)\n\t}\n\n\tb.WriteByte('\\n')\n\treturn b.Bytes(), nil\n}", "func (f Colored) Format(entry *logrus.Entry) ([]byte, error) {\n\tvar b *bytes.Buffer\n\tif entry.Buffer != nil {\n\t\tb = entry.Buffer\n\t} else {\n\t\tb = &bytes.Buffer{}\n\t}\n\n\tf.printColored(b, entry)\n\tb.WriteByte('\\n')\n\treturn b.Bytes(), nil\n}", "func (f *TextFormatter) Format(entry *logrus.Entry) ([]byte, error) {\n\tvar b *bytes.Buffer\n\tif entry.Buffer != nil {\n\t\tb = entry.Buffer\n\t} else {\n\t\tb = new(bytes.Buffer)\n\t}\n\n\t// Remove a single newline if it already exists in the message to keep\n\t// the behavior of logrus text_formatter the same as the stdlib log package\n\tentry.Message = strings.TrimSuffix(entry.Message, \"\\n\")\n\n\tcaller := \"\"\n\n\tif entry.HasCaller() {\n\t\tfuncVal := fmt.Sprintf(\"%s()\", entry.Caller.Function)\n\t\tfileVal := fmt.Sprintf(\"%s:%d\", entry.Caller.File, entry.Caller.Line)\n\t\tcaller = fileVal + \" \" + funcVal + \" \"\n\t}\n\n\tif entry.Level >= logrus.DebugLevel {\n\t\tnow := time.Now().UTC().Format(\"2006-01-02 15:04:05.999999999Z\")\n\t\tfmt.Fprintf(b, \"%s %s: \", strings.ToUpper(entry.Level.String()), now)\n\t}\n\tfmt.Fprintf(b, \"%s%-44s\\n\", caller, entry.Message)\n\n\treturn b.Bytes(), nil\n}", "func (f DefaultFormatter) Format(event *Event) string {\n\tyear, month, day := event.Time.Date()\n\thour, minute, second := event.Time.Clock()\n\tlevelString := event.Level.String()\n\trightAlignedLevel := strings.Repeat(\" \", 8-len(levelString)) + levelString\n\tmsg := event.Msg\n\tif len(event.Args) > 0 {\n\t\tmsg = fmt.Sprintf(event.Msg, event.Args...)\n\t}\n\tlines := strings.Split(msg, \"\\n\")\n\tfor i, line := range lines {\n\t\tlines[i] = \"\\t\" + line\n\t}\n\tmsg = strings.Join(lines, \"\\n\")\n\treturn fmt.Sprintf(\n\t\t\"%d-%02d-%02d %02d:%02d:%02d: %s: %s: at %s in %s, line %d:\\n%s\\n\\n\",\n\t\tyear, month, day, hour, minute, second,\n\t\trightAlignedLevel, event.Name, event.FuncName,\n\t\tfilepath.Base(event.File), event.Line,\n\t\tstrings.TrimRightFunc(msg, unicode.IsSpace))\n}", "func (t *Target) Format(item *LogItem) string {\n\t// call user-defined formatter if exists\n\tif t.formatter != nil {\n\t\treturn t.formatter.Format(item)\n\t}\n\n\t// default log format: [time][logId][name][level][trace]: message\\n\n\treturn fmt.Sprintf(\"[%s][%s][%s][%s]%s: %s\\n\",\n\t\titem.When.Format(\"2006/01/02 15:04:05.000\"),\n\t\titem.LogId,\n\t\titem.Name,\n\t\tLevelToString(item.Level),\n\t\titem.Trace,\n\t\titem.Message,\n\t)\n}", "func (f *ConsoleFormatter) Format(e *logrus.Entry) ([]byte, error) {\n\tbuf := new(bytes.Buffer)\n\t// Remove a single newline if it already exists in the message to keep\n\t// the behavior of logrus text_formatter the same as the stdlib log package\n\tstrings.TrimSuffix(e.Message, \"\\n\")\n\tmsg := f.colorize(e.Message, DefaultColor, e)\n\toutput := strings.Replace(f.ConsoleLogFormat, \"%msg%\", msg, 1)\n\toutput = f.formatTime(output, e)\n\toutput = f.formatLevel(output, e)\n\toutput = f.formatCaller(output, e)\n\toutput = f.formatFields(output, e)\n\tfmt.Fprint(buf, output, \"\\n\")\n\treturn buf.Bytes(), nil\n}", "func (ev *Event) Format(format string) string {\n\treturn fmt.Sprintf(\"Start um %s - Name %s\\n\", ev.Start.Format(\"15:04\"), ev.Name)\n\t//return fmt.Sprintf(\"Start um %s - Name %s\\n%s\\n\", ev.Start.Format(\"15:04\"), ev.Name, ev.Info)\n}", "func (f *LogstashFormatter) Format(\n\tb *bytes.Buffer,\n\tlvl Level,\n\tctx Context,\n\tmsg string,\n\ttrace []byte,\n) *bytes.Buffer {\n\tb.WriteString(`{\"@timestamp\":\"`)\n\tb.WriteString(time.Now().Format(time.RFC3339Nano))\n\tb.WriteString(`\",\"@version\":1,\"level\":\"`)\n\tb.WriteString(LevelToString(lvl))\n\tb.WriteString(`\",\"`)\n\n\tfor k, v := range ctx {\n\t\tb.WriteString(k)\n\t\tb.WriteString(`\":`)\n\t\tf.appendValue(b, v)\n\t\tb.WriteString(`,\"`)\n\t}\n\n\tb.WriteString(`message\":`)\n\tb.WriteString(strconv.Quote(string(msg)))\n\tif len(trace) > 0 {\n\t\tb.WriteString(`,\"trace\":`)\n\t\tb.WriteString(strconv.Quote(string(trace)))\n\t}\n\tb.WriteString(`}`)\n\tb.WriteByte('\\n')\n\treturn b\n}", "func (f *CustomJSONFormatter) Format(entry *logrus.Entry) ([]byte, error) {\n\tdata := make(logrus.Fields, len(entry.Data)+3)\n\tfor k, v := range entry.Data {\n\t\tswitch v := v.(type) {\n\t\tcase error:\n\t\t\t// Otherwise errors are ignored by `encoding/json`\n\t\t\t// https://github.com/sirupsen/logrus/issues/137\n\t\t\tdata[k] = v.Error()\n\t\tdefault:\n\t\t\tdata[k] = v\n\t\t}\n\t}\n\n\tprefixFieldClashes(data, f.FieldMap)\n\n\ttimestampFormat := f.TimestampFormat\n\tif timestampFormat == \"\" {\n\t\ttimestampFormat = defaultTimestampFormat\n\t}\n\n\tif !f.DisableTimestamp {\n\t\tdata[f.FieldMap.resolve(FieldKeyTime)] = entry.Time.Format(timestampFormat)\n\t}\n\tdata[f.FieldMap.resolve(FieldKeyMsg)] = entry.Message\n\tdata[f.FieldMap.resolve(FieldKeyLevel)] = entry.Level.String()\n\n\tvar serialized []byte\n\tvar err error\n\n\tif f.PrettyPrint {\n\t\tserialized, err = json.MarshalIndent(data, \"\", \" \")\n\t} else {\n\t\tserialized, err = json.Marshal(data)\n\t}\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to marshal fields to JSON, %v\", err)\n\t}\n\treturn append(serialized, '\\n'), nil\n}", "func (f *ZipkinFormatter) Format(entry *logrus.Entry) ([]byte, error) {\n\tdata := fmt.Sprintf(\"%s\\n\", entry.Message)\n\n\treturn []byte(data), nil\n}", "func (s *logrusFormatter) Format(pkg string, l capnslog.LogLevel, _ int, entries ...interface{}) {\n\tfor _, entry := range entries {\n\t\tstr := fmt.Sprint(entry)\n\t\tswitch l {\n\t\tcase capnslog.CRITICAL:\n\t\t\ts.logWithPkg(pkg).Fatal(str)\n\t\tcase capnslog.ERROR:\n\t\t\ts.logWithPkg(pkg).Error(str)\n\t\tcase capnslog.WARNING:\n\t\t\ts.logWithPkg(pkg).Warning(str)\n\t\tcase capnslog.NOTICE:\n\t\t\ts.logWithPkg(pkg).Warning(str)\n\t\tcase capnslog.INFO:\n\t\t\ts.logWithPkg(pkg).Info(str)\n\t\tcase capnslog.DEBUG:\n\t\t\ts.logWithPkg(pkg).Debug(str)\n\t\tcase capnslog.TRACE:\n\t\t\ts.logWithPkg(pkg).Debug(str)\n\t\tdefault:\n\t\t\tpanic(\"Unhandled loglevel\")\n\t\t}\n\t}\n}", "func (c *TextHandler) Log(e log.Entry) {\n\tb := new(bytes.Buffer)\n\n\tb.Reset()\n\tb.WriteString(e.Timestamp.In(loc).Format(defaultFormat))\n\tfmt.Fprint(b, \" \")\n\tfmt.Fprintf(b, \"%-6s\", e.Level.String())\n\tfmt.Fprint(b, \" \")\n\n\tfor _, f := range e.Fields {\n\t\tfmt.Fprint(b, f.Key)\n\t\tfmt.Fprint(b, \"=\")\n\t\tfmt.Fprintf(b, \"%-10s\", f.Value)\n\t\tfmt.Fprint(b, \" \")\n\t}\n\tb.WriteString(e.Message)\n\tfmt.Println(b.String())\n}", "func formatLog() {\n\tFormatter := new(log.TextFormatter)\n\tFormatter.TimestampFormat = \"02-01-2006 15:04:05\"\n\tFormatter.FullTimestamp = true\n\tlog.SetFormatter(Formatter)\n}", "func (f *GlogFormatter) Format(context log.LogContext) []byte {\n\tres := &bytes.Buffer{}\n\n\tfile := context.File\n\tslash := len(file) - 1\n\tfor ; slash >= 0; slash-- {\n\t\tif file[slash] == filepath.Separator {\n\t\t\tbreak\n\t\t}\n\t}\n\tif slash >= 0 {\n\t\tfile = file[slash+1:]\n\t}\n\n\t_, month, day := context.Time.Date()\n\thour, minute, second := context.Time.Clock()\n\tf.tmp[0] = log.UcShortestSeverityStrings[log.SeverityToIndex(context.Severity)][0]\n\tlog.TwoDigits(&f.tmp, 1, int(month))\n\tlog.TwoDigits(&f.tmp, 3, day)\n\tf.tmp[5] = ' '\n\tlog.TwoDigits(&f.tmp, 6, hour)\n\tf.tmp[8] = ':'\n\tlog.TwoDigits(&f.tmp, 9, minute)\n\tf.tmp[11] = ':'\n\tlog.TwoDigits(&f.tmp, 12, second)\n\tf.tmp[14] = '.'\n\tlog.NDigits(&f.tmp, 6, 15, context.Time.Nanosecond()/1000)\n\tf.tmp[21] = ' '\n\tlog.NDigits(&f.tmp, 5, 22, context.Pid)\n\tf.tmp[27] = ' '\n\tres.Write(f.tmp[:28])\n\tres.WriteString(file)\n\tf.tmp[0] = ':'\n\tn := log.Itoa(&f.tmp, 1, context.Line)\n\tf.tmp[n+1] = ']'\n\tf.tmp[n+2] = ' '\n\tres.Write(f.tmp[:n+3])\n\tmessage := \"\"\n\tif context.Format != nil {\n\t\tmessage = fmt.Sprintf(*context.Format, context.Args...)\n\t} else {\n\t\tmessage = fmt.Sprint(context.Args...)\n\t}\n\n\tres.WriteString(message)\n\n\tl := len(message)\n\tif l > 0 && message[l-1] != '\\n' {\n\t\tres.WriteRune('\\n')\n\t}\n\n\treturn res.Bytes()\n}", "func (g *Gelf) Format(rec *logr.LogRec, level logr.Level, buf *bytes.Buffer) (*bytes.Buffer, error) {\n\tif buf == nil {\n\t\tbuf = &bytes.Buffer{}\n\t}\n\tenc := gojay.BorrowEncoder(buf)\n\tdefer func() {\n\t\tenc.Release()\n\t}()\n\n\tgr := gelfRecord{\n\t\tLogRec: rec,\n\t\tGelf: g,\n\t\tlevel: level,\n\t\tsorter: g.FieldSorter,\n\t}\n\n\terr := enc.EncodeObject(gr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbuf.WriteByte(0)\n\treturn buf, nil\n}", "func (f *Formatter) Format(entry *Entry, raw json.RawMessage, prefix, suffix []byte) error {\n\tcolor.NoColor = !f.Colorize\n\tf.enhance(entry)\n\n\terr := f.outputSimple(prefix, f.ShowPrefix)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = f.template.Execute(f.output, entry)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tf.outputFields(entry, raw)\n\n\terr = f.outputSimple(suffix, f.ShowSuffix)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = stacktrace(f.output, raw)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = f.output.Write(NewLine)\n\treturn err\n}", "func (f *formatter) Format(pool *bufferPool, level logLevel, msg string) *bytes.Buffer {\n\tvar buf = pool.get()\n\tbuf.WriteString(fmt.Sprintf(\"%c %06v\", level.abbr(), atomic.LoadUint32(&currentTime)))\n\t_, file, line, ok := runtime.Caller(3)\n\tif ok {\n\t\tvar i int\n\t\tfor i = len(file) - 1; i >= 0; i-- {\n\t\t\tif file[i] == '/' {\n\t\t\t\tif i == len(file)-1 {\n\t\t\t\t\tbreak\n\t\t\t\t} else {\n\t\t\t\t\ti++\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tbuf.WriteString(fmt.Sprintf(\" [%v:%v] %v\\n\", file[i:], line, msg))\n\t}\n\treturn buf\n}", "func (f logFormatter) format(_ io.Writer, params handlers.LogFormatterParams) {\n\tf.logger.With(\n\t\tzap.Duration(\"duration\", time.Since(params.TimeStamp)),\n\t\tzap.String(\"url\", params.URL.String()),\n\t\tzap.String(\"method\", params.Request.Method),\n\t\tzap.String(\"remote_addr\", params.Request.RemoteAddr),\n\t\tzap.String(\"user_agent\", params.Request.UserAgent()),\n\t\tzap.String(\"mode\", \"access_log\"),\n\t).Info(\"Handled HTTP request\")\n}", "func FormatEntry(entry *Entry) string {\n\tbox := packr.New(\"data\", \"./data\")\n\n\tf, err := box.FindString(\"entries/entry.md\")\n\tif err != nil {\n\t\tlogger.Fatalf(\"Cannot get embedded 'entries/README.md' asset: %v\", err)\n\t}\n\n\ttmpl, err := template.New(\"entry.md\").Parse(f)\n\tif err != nil {\n\t\tlogger.Fatalf(\"Cannot get parse template for embedded 'entries/README.md' asset: %v\", err)\n\t}\n\n\tvar out bytes.Buffer\n\ttmpl.Execute(&out, entry)\n\n\treturn out.String()\n}", "func formatLog(level string, format string, v ...interface{}) string {\n\tvar msg string\n\tif len(v) > 0 {\n\t\tmsg = fmt.Sprintf(format, v...)\n\t} else {\n\t\t// Don't run the message through fmt.Sprintf if no args were\n\t\t// supplied. This avoids the %!(MISSING) spam that fmt adds if\n\t\t// it sees an unmatched formatting string (e.g. a %s without a\n\t\t// matching arg). This makes it easy to log.Info(someVariable)\n\t\t// without worrying about junk in the output should it contain\n\t\t// a % sign.\n\t\tmsg = format\n\t}\n\ttimestr := time.Now().UTC().Format(\"2006-01-02 15:04:05.000000\")\n\treturn fmt.Sprintf(\"%s [%s]: %s\\n\", timestr, level, msg)\n}", "func (f *customTextFormatter) Format(entry *logrus.Entry) ([]byte, error) {\n\toriginalText, err := f.TextFormatter.Format(entry)\n\treturn append([]byte(\"▶ \"), originalText...), err\n}", "func FormatLog(flat map[string]interface{}) string {\n\tvar str string\n\tsliceOfKeys := []string{\"timestamp\", \"level\", \"deploymentId\", WorkFlowID.String(), ExecutionID.String(), NodeID.String(), InstanceID.String(), InterfaceName.String(), OperationName.String(), TypeID.String(), \"content\"}\n\tfor _, k := range sliceOfKeys {\n\t\tif val, ok := flat[k].(string); ok {\n\t\t\tif k != \"content\" {\n\t\t\t\tstr += \"[\" + val + \"]\"\n\t\t\t} else {\n\t\t\t\tstr += val\n\t\t\t}\n\t\t} else {\n\t\t\tstr += \"[]\"\n\t\t}\n\n\t}\n\treturn str\n}", "func (c *CollorHandler) Log(e log.Entry) {\n\n\tcolor := defaultColors[e.Level]\n\n\tb := new(bytes.Buffer)\n\tb.Reset()\n\tb.WriteString(e.Timestamp.In(loc).Format(defaultFormat))\n\tb.WriteString(\" \")\n\tfmt.Fprintf(b, \"%s%-6s%s\", ansi.Bold+color, e.Level.String(), ansi.BoldOff+ansi.Reset)\n\n\tfor _, f := range e.Fields {\n\t\tfmt.Fprint(b, ansi.Bold)\n\t\tfmt.Fprint(b, f.Key)\n\t\tfmt.Fprint(b, ansi.BoldOff)\n\t\tfmt.Fprint(b, \"=\")\n\t\tfmt.Fprint(b, ansi.Italics)\n\t\tfmt.Fprintf(b, \"%-10s\", f.Value)\n\t\tfmt.Fprint(b, ansi.ItalicsOff)\n\t\tfmt.Fprint(b, \" \")\n\t}\n\tb.WriteString(e.Message)\n\n\tfmt.Println(b.String())\n}", "func (f *FakeFormatter) Format(b *bytes.Buffer, lvl golog.Level, ctx golog.Context, msg string, trace []byte) *bytes.Buffer {\n\tb.WriteString(msg)\n\treturn b\n}", "func TestLogFormat(t *testing.T) {\n\n\t// Log test message\n\tvar message string = \"this is a log entry\"\n\n\t// Create log entry\n\tInfo(message)\n\n\t// The log format should look like this:\n\t// INFO 2020/09/22 17:34:16 INFO this is a log entry\n\tvar pattern string = fmt.Sprintf(\"[A-Z]{1,4} [0-9]{1,4}/[0-9]{1,2}/[0-9]{1,2} [0-9]{1,2}:[0-9]{1,2}:[0-9]{1,2} %s\", message)\n\tr, err := regexp.Compile(pattern)\n\n\tif err != nil {\n\t\tt.Log(err)\n\t\tt.Fail()\n\t}\n\n\tmatched := r.MatchString(buf.String())\n\tt.Log(matched)\n\n\tif matched != true {\n\t\tt.Log(\"The log entry did not match the regex pattern\")\n\n\t\t// Print the log output\n\t\tt.Log(buf.String())\n\t\tt.Fail()\n\t}\n}", "func (u utcFormatter) Format(e *logrus.Entry) ([]byte, error) {\n\te.Time = e.Time.UTC()\n\treturn u.TextFormatter.Format(e)\n}", "func (f FormatterFunc) Format(e *Event) string { return f(e) }", "func (nullFormatter) Format(e *logrus.Entry) ([]byte, error) {\n\treturn []byte{}, nil\n}", "func (j *JSON) Format(rec *logr.LogRec, stacktrace bool, buf *bytes.Buffer) (*bytes.Buffer, error) {\n\tj.once.Do(j.applyDefaultKeyNames)\n\n\tif buf == nil {\n\t\tbuf = &bytes.Buffer{}\n\t}\n\tenc := gojay.BorrowEncoder(buf)\n\tdefer func() {\n\t\tenc.Release()\n\t}()\n\n\tsorter := j.ContextSorter\n\tif sorter == nil {\n\t\tsorter = j.defaultContextSorter\n\t}\n\n\tjlr := JSONLogRec{\n\t\tLogRec: rec,\n\t\tJSON: j,\n\t\tstacktrace: stacktrace,\n\t\tsorter: sorter,\n\t}\n\n\terr := enc.EncodeObject(jlr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbuf.WriteByte('\\n')\n\treturn buf, nil\n}", "func (fa *FileAppender) writeLogItem(log *LogItem) {\n\n\tts := toolkit.GetTimeStamp(log.t0)\n\n\tif fa.logFile == nil {\n\t\t// output to console\n\t\tfmt.Println(ts, log.severity, log.file, log.line, log.info)\n\t\treturn\n\t}\n\n\tif fa.logFormat == \"\" {\n\t\t// no logformat specified\n\t\tfa.logFile.WriteString(log.file)\n\t\tfa.logFile.WriteString(\" \")\n\t\tfa.logFile.WriteString(log.line)\n\t\tfa.logFile.WriteString(\" \")\n\t\tfa.logFile.WriteString(log.info)\n\t\tfa.logFile.WriteString(\"\\n\")\n\t\treturn\n\t}\n\n\ts := fa.getLogFormat(log.severity)\n\n\tfor true {\n\t\tidx := strings.Index(s, \"%\")\n\t\tif idx > 0 {\n\t\t\tfa.logFile.WriteString(s[0:idx])\n\t\t\ts = s[idx:]\n\t\t} else if idx == 0 {\n\t\t\tif strings.Index(s, \"%V\") == 0 {\n\t\t\t\tfa.logFile.WriteString(getLogSeverityString(log.severity))\n\t\t\t\ts = s[2:]\n\t\t\t} else if strings.Index(s, \"%T\") == 0 {\n\t\t\t\tfa.logFile.WriteString(ts)\n\t\t\t\ts = s[2:]\n\t\t\t} else if strings.Index(s, \"%I\") == 0 {\n\t\t\t\tfa.logFile.WriteString(log.info)\n\t\t\t\ts = s[2:]\n\t\t\t} else if strings.Index(s, \"%F\") == 0 {\n\t\t\t\tfa.logFile.WriteString(path.Base(log.file))\n\t\t\t\ts = s[2:]\n\t\t\t} else if strings.Index(s, \"%L\") == 0 {\n\t\t\t\tfa.logFile.WriteString(log.line)\n\t\t\t\ts = s[2:]\n\t\t\t} else {\n\t\t\t\t// skip the %\n\t\t\t\ts = s[1:]\n\t\t\t}\n\t\t} else {\n\t\t\t// idx < 0\n\t\t\tfa.logFile.WriteString(s)\n\t\t\tbreak\n\t\t}\n\t}\n\n\tfa.logFile.WriteString(\"\\n\")\n}", "func (l *Logger) FormatHeader(t time.Time, funcName string, file string, line int, levelName string) {\n\tl.buf = l.buf[:0]\n\tbuf := &l.buf\n\t*buf = append(*buf, l.prefix...)\n\tif len(l.prefix) > 0 {\n\t\t*buf = append(*buf, \" | \"...)\n\t}\n\tif l.flag&(Ldate|Ltime|Lmicroseconds) != 0 {\n\t\tif l.flag&LUTC != 0 {\n\t\t\tt = t.UTC()\n\t\t}\n\t\tif l.flag&Ldate != 0 {\n\t\t\tyear, month, day := t.Date()\n\t\t\titoa(buf, year, 4)\n\t\t\t*buf = append(*buf, '/')\n\t\t\titoa(buf, int(month), 2)\n\t\t\t*buf = append(*buf, '/')\n\t\t\titoa(buf, day, 2)\n\t\t\t*buf = append(*buf, ' ')\n\t\t}\n\t\tif l.flag&(Ltime|Lmicroseconds) != 0 {\n\t\t\thour, min, sec := t.Clock()\n\t\t\titoa(buf, hour, 2)\n\t\t\t*buf = append(*buf, ':')\n\t\t\titoa(buf, min, 2)\n\t\t\t*buf = append(*buf, ':')\n\t\t\titoa(buf, sec, 2)\n\t\t\tif l.flag&Lmicroseconds != 0 {\n\t\t\t\t*buf = append(*buf, '.')\n\t\t\t\titoa(buf, t.Nanosecond()/1e3, 6)\n\t\t\t}\n\t\t\t//*buf = append(*buf, ' ')\n\t\t}\n\t}\n\t*buf = append(*buf, \" | \"...)\n\tif l.flag&(Lshortfile|Llongfile) != 0 {\n\t\tif l.flag&Lshortfile != 0 {\n\t\t\tshort := file\n\t\t\tfor i := len(file) - 1; i > 0; i-- {\n\t\t\t\tif file[i] == '/' {\n\t\t\t\t\tshort = file[i+1:]\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\tfile = short\n\t\t}\n\n\t\tl.cache = l.cache[:0]\n\t\tl.cache = append(l.cache, funcName...)\n\t\tl.cache = append(l.cache, ':')\n\t\tl.cache = append(l.cache, file...)\n\t\tl.cache = append(l.cache, ':')\n\t\titoa(&l.cache, line, -1)\n\n\t\t*buf = append(*buf, fmt.Sprintf(\"%-60s\", string(l.cache))...)\n\t}\n\n\tid := l.GetGID()\n\tif len(id) > 0 {\n\t\t*buf = append(*buf, \" | \"...)\n\t\t*buf = append(*buf, fmt.Sprintf(\"%-3s\", id)...)\n\t}\n\n\tif len(levelName) > 0 {\n\t\t*buf = append(*buf, \" | \"...)\n\t\t*buf = append(*buf, fmt.Sprintf(\"%-5s\", levelName)...)\n\t}\n\n}", "func adjustLog(level string, msg string) string {\r\n\treturn fmt.Sprintf(log_format, level, getTime(), msg)\r\n}", "func (f *consoleFormatter) Format(e Entity, b *bytes.Buffer) (err error) {\n\tif name := e.Name(); name != \"\" {\n\t\tb.WriteString(name + \" \")\n\t}\n\tif tm := e.TimeString(); tm != \"\" {\n\t\tb.WriteString(\"[\" + tm + \"]\")\n\t}\n\tb.WriteString(\"[\" + e.Level().ColorfulShortCapitalString() + \"] \")\n\tb.WriteString(e.Message())\n\tif caller := e.Caller(); caller != \"\" {\n\t\tb.WriteString(\" \" + caller)\n\t}\n\tif fields := e.Fields(); len(fields) > 0 {\n\t\tb.WriteString(\" \" + internal.FormatFieldsToText(e.Fields()))\n\t}\n\tif stack := e.Stack(); len(stack) > 0 {\n\t\t// In the console, in order to be able to display the stack information better,\n\t\t// we will use a separate line to display the stack information.\n\t\tfor i, j := 0, len(stack); i < j; i++ {\n\t\t\tb.WriteString(\"\\n\\t\" + stack[i])\n\t\t}\n\t}\n\tb.WriteByte('\\n')\n\treturn\n}", "func (formatter JsonFormatter) Format(fields entries.Fields, level string, message string) string {\n\tformattedMetadata := make([]string, len(fields)+2)\n\tindex := 2\n\n\tformattedMetadata[0] = formatStringMessageAsJson(\"level\", level)\n\tformattedMetadata[1] = formatStringMessageAsJson(\"message\", message)\n\tfor key, value := range fields {\n\t\tif _, typeOk := value.(string); typeOk {\n\t\t\tformattedMetadata[index] = formatStringMessageAsJson(key, value)\n\t\t} else {\n\t\t\tformattedMetadata[index] = fmt.Sprintf(\"\\\"%v\\\": %v\", key, value)\n\t\t}\n\t\tindex++\n\t}\n\n\tjoinedMetadata := strings.Join(formattedMetadata, \", \")\n\treturn fmt.Sprintf(\"{ %v }\", joinedMetadata)\n}", "func formatLogMessage(class int, format string, args ...interface{}) string {\n\tif class < 0 || class >= len(loggers) {\n\t\tWriteLog(InternalLogger, \"ERROR: Invalid LogMessage() class %d\", class)\n\n\t\treturn \"\"\n\t}\n\n\tclassName := loggers[class].name\n\ts := fmt.Sprintf(format, args...)\n\n\tsequenceMux.Lock()\n\tdefer sequenceMux.Unlock()\n\n\tsequence = sequence + 1\n\tsequenceString := fmt.Sprintf(\"%d\", sequence)\n\n\tif LogTimeStampFormat == \"\" {\n\t\tLogTimeStampFormat = \"2006-01-02 15:04:05\"\n\t}\n\n\ts = fmt.Sprintf(\"[%s] %-5s %-7s: %s\", time.Now().Format(LogTimeStampFormat), sequenceString, className, s)\n\n\treturn s\n}", "func (l *Logger) formatLevel(lv string) string {\n\n\tspaces := \"\"\n\tif len(lv) == 4 {\n\t\tspaces = \" \"\n\t}\n\treturn lv + spaces\n}", "func (f Formatter) Format(txt string) (string, error) {\n\ttokens, err := f.l.Scan([]byte(txt))\n\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"Failed to format: %q\", err)\n\t}\n\n\tvar s strings.Builder\n\n\tinSession := true\n\tinPerformance := false\n\n\tfor _, tok := range tokens {\n\t\tswitch tok.Name() {\n\t\tcase \"DATE\":\n\t\t\ts.WriteString(\"@ \")\n\t\t\ts.WriteString(tok.Value())\n\t\t\ts.WriteString(\"\\r\\n\")\n\t\tcase \"FAILS\":\n\t\t\ts.WriteString(\" \")\n\t\t\ts.WriteString(tok.Value())\n\t\t\ts.WriteString(\"f\")\n\t\tcase \"LOAD\":\n\t\t\tinPerformance = true\n\t\t\ts.WriteString(\"\\r\\n\")\n\t\t\ts.WriteString(\" \")\n\t\t\ts.WriteString(tok.Value())\n\t\tcase \"METADATA\":\n\t\t\ts.WriteString(\"\\r\\n\")\n\t\t\ts.WriteString(spacer(inSession, inPerformance))\n\t\t\ts.WriteString(\"* \")\n\t\t\ts.WriteString(tok.Value())\n\t\tcase \"MOVEMENT\", \"MOVEMENT_SS\":\n\t\t\tinSession = false\n\t\t\tinPerformance = false\n\t\t\ts.WriteString(\"\\r\\n\\r\\n\")\n\t\t\tif tok.Value() == \"MOVEMENT_SS\" {\n\t\t\t\ts.WriteString(\"+ \")\n\t\t\t}\n\t\t\ts.WriteString(tok.Value())\n\t\t\ts.WriteString(\":\")\n\t\tcase \"NOTE\":\n\t\t\ts.WriteString(\"\\r\\n\")\n\t\t\ts.WriteString(spacer(inSession, inPerformance))\n\t\t\ts.WriteString(\"* \")\n\t\t\ts.WriteString(tok.Value())\n\t\tcase \"REPS\":\n\t\t\ts.WriteString(\" \")\n\t\t\ts.WriteString(tok.Value())\n\t\t\ts.WriteString(\"r\")\n\t\tcase \"SETS\":\n\t\t\ts.WriteString(\" \")\n\t\t\ts.WriteString(tok.Value())\n\t\t\ts.WriteString(\"s\")\n\t\t}\n\t}\n\n\treturn s.String(), nil\n}", "func logf(format string, a ...interface{}) {\n\tif strings.HasSuffix(format, \"\\n\") {\n\t\tfmt.Fprintf(writer, format, a...)\n\t} else {\n\t\tfmt.Fprintf(writer, format+\"\\n\", a...)\n\t}\n}", "func (handler *ConsoleLogHandler) Format() LogFormat {\r\n return JSONFormat\r\n}", "func Format(level types.LogLevel, delimiter, tag, prefix, msg string, kvs ...string) string {\n\tout := types.LogLevels[level]\n\tif tag != \"\" {\n\t\tout += delimiter + tag\n\t}\n\tif prefix != \"\" {\n\t\tout += delimiter + prefix\n\t}\n\tout += delimiter + msg\n\tif len(kvs) > 0 {\n\t\tout += delimiter\n\t\tkey := kvs[0]\n\t\tmp := map[string]string{key: \"\"}\n\n\t\tfor i := 1; i < len(kvs); i++ {\n\t\t\tif i%2 != 0 {\n\t\t\t\tmp[key] = kvs[i]\n\t\t\t} else {\n\t\t\t\tkey = kvs[i]\n\t\t\t}\n\t\t}\n\t\tba, err := json.Marshal(mp)\n\t\tif err != nil {\n\t\t\t// TODO process err\n\t\t} else {\n\t\t\tout += string(ba)\n\t\t}\n\t}\n\treturn out\n}", "func formatLine(cfg *scrapeconfig.WindowsEventsTargetConfig, event win_eventlog.Event) (string, error) {\n\tstructuredEvent := Event{\n\t\tSource: event.Source.Name,\n\t\tChannel: event.Channel,\n\t\tComputer: event.Computer,\n\t\tEventID: event.EventID,\n\t\tVersion: event.Version,\n\t\tLevel: event.Level,\n\t\tTask: event.Task,\n\t\tOpcode: event.Opcode,\n\t\tLevelText: event.LevelText,\n\t\tTaskText: event.TaskText,\n\t\tOpcodeText: event.OpcodeText,\n\t\tKeywords: event.Keywords,\n\t\tTimeCreated: event.TimeCreated.SystemTime,\n\t\tEventRecordID: event.EventRecordID,\n\t}\n\n\tif !cfg.ExcludeEventData {\n\t\tstructuredEvent.EventData = string(event.EventData.InnerXML)\n\t}\n\tif !cfg.ExcludeUserData {\n\t\tstructuredEvent.UserData = string(event.UserData.InnerXML)\n\t}\n\tif !cfg.ExcludeEventMessage {\n\t\tstructuredEvent.Message = event.Message\n\t}\n\tif event.Correlation.ActivityID != \"\" || event.Correlation.RelatedActivityID != \"\" {\n\t\tstructuredEvent.Correlation = &Correlation{\n\t\t\tActivityID: event.Correlation.ActivityID,\n\t\t\tRelatedActivityID: event.Correlation.RelatedActivityID,\n\t\t}\n\t}\n\t// best effort to get the username of the event.\n\tif event.Security.UserID != \"\" {\n\t\tvar userName string\n\t\tusid, err := syscall.StringToSid(event.Security.UserID)\n\t\tif err == nil {\n\t\t\tusername, domain, _, err := usid.LookupAccount(\"\")\n\t\t\tif err == nil {\n\t\t\t\tuserName = fmt.Sprint(domain, \"\\\\\", username)\n\t\t\t}\n\t\t}\n\t\tstructuredEvent.Security = &Security{\n\t\t\tUserID: event.Security.UserID,\n\t\t\tUserName: userName,\n\t\t}\n\t}\n\tif event.Execution.ProcessID != 0 {\n\t\tstructuredEvent.Execution = &Execution{\n\t\t\tProcessID: event.Execution.ProcessID,\n\t\t\tThreadID: event.Execution.ThreadID,\n\t\t}\n\t\t_, _, processName, err := win_eventlog.GetFromSnapProcess(event.Execution.ProcessID)\n\t\tif err == nil {\n\t\t\tstructuredEvent.Execution.ProcessName = processName\n\t\t}\n\t}\n\treturn jsoniter.MarshalToString(structuredEvent)\n}", "func (rs HookRecords) Format(w fmt.State, c rune) {\n\tif !w.Flag('+') {\n\t\t// Without %+v, fall back to String().\n\t\tio.WriteString(w, rs.String())\n\t\treturn\n\t}\n\n\tfor _, r := range rs {\n\t\tfmt.Fprintf(w, \"\\n%s took %v from:\\n\\t%+v\",\n\t\t\tfxreflect.FuncName(r.Func),\n\t\t\tr.Runtime,\n\t\t\tr.CallerFrame)\n\t}\n\tfmt.Fprintf(w, \"\\n\")\n}", "func (e EnumEntry) Format(w fmt.State, verb rune) {\n\tif len(e.Attributes) > 0 {\n\t\tfmt.Fprintf(w, \"%v %v\", e.Attributes, e.Name)\n\t} else {\n\t\tfmt.Fprint(w, e.Name)\n\t}\n}", "func (p *Plain) Format(rec *logr.LogRec, stacktrace bool, buf *bytes.Buffer) (*bytes.Buffer, error) {\n\tdelim := p.Delim\n\tif delim == \"\" {\n\t\tdelim = \" \"\n\t}\n\tif buf == nil {\n\t\tbuf = &bytes.Buffer{}\n\t}\n\n\ttimestampFmt := p.TimestampFormat\n\tif timestampFmt == \"\" {\n\t\ttimestampFmt = logr.DefTimestampFormat\n\t}\n\n\tif !p.DisableTimestamp {\n\t\tvar arr [128]byte\n\t\ttbuf := rec.Time().AppendFormat(arr[:0], timestampFmt)\n\t\tbuf.Write(tbuf)\n\t\tbuf.WriteString(delim)\n\t}\n\tif !p.DisableLevel {\n\t\tfmt.Fprintf(buf, \"%v%s\", rec.Level().Name, delim)\n\t}\n\tif !p.DisableMsg {\n\t\tfmt.Fprint(buf, rec.Msg(), delim)\n\t}\n\tif !p.DisableContext {\n\t\tctx := rec.Fields()\n\t\tif len(ctx) > 0 {\n\t\t\tlogr.WriteFields(buf, ctx, \" \")\n\t\t}\n\t}\n\tif stacktrace && !p.DisableStacktrace {\n\t\tframes := rec.StackFrames()\n\t\tif len(frames) > 0 {\n\t\t\tbuf.WriteString(\"\\n\")\n\t\t\tlogr.WriteStacktrace(buf, rec.StackFrames())\n\t\t}\n\t}\n\tbuf.WriteString(\"\\n\")\n\treturn buf, nil\n}", "func makeUnstructuredEntry(\n\tctx context.Context,\n\ts Severity,\n\tc Channel,\n\tdepth int,\n\tredactable bool,\n\tformat string,\n\targs ...interface{},\n) (res logEntry) {\n\tres = makeEntry(ctx, s, c, depth+1)\n\n\tres.structured = false\n\n\tif redactable {\n\t\tvar buf redact.StringBuilder\n\t\tif len(args) == 0 {\n\t\t\t// TODO(knz): Remove this legacy case.\n\t\t\tbuf.Print(redact.Safe(format))\n\t\t} else if len(format) == 0 {\n\t\t\tbuf.Print(args...)\n\t\t} else {\n\t\t\tbuf.Printf(format, args...)\n\t\t}\n\t\tres.payload = makeRedactablePayload(buf.RedactableString())\n\t} else {\n\t\tvar buf strings.Builder\n\t\tformatArgs(&buf, format, args...)\n\t\tres.payload = makeUnsafePayload(buf.String())\n\t}\n\n\treturn res\n}", "func (l *loggingT) formatHeader(s severity.Severity, file string, line int) *buffer.Buffer {\n\tbuf := buffer.GetBuffer()\n\tif l.skipHeaders {\n\t\treturn buf\n\t}\n\tnow := timeNow()\n\tbuf.FormatHeader(s, file, line, now)\n\treturn buf\n}", "func formattingEvent(event string, data struct {\n\tValue string\n\tDisabledFormatting bool\n}, id string) string {\n\tvar eventMsg bytes.Buffer\n\tif event != \"\" {\n\t\teventMsg.WriteString(fmt.Sprintf(\"event:%s\\n\", strings.Replace(event, \"\\n\", \"\", -1)))\n\t}\n\tif data.Value != \"\" {\n\t\tif !data.DisabledFormatting {\n\t\t\tlines := strings.Split(data.Value, \"\\n\")\n\t\t\tfor _, line := range lines {\n\t\t\t\teventMsg.WriteString(fmt.Sprintf(\"data:%s\\n\", line))\n\t\t\t}\n\t\t} else {\n\t\t\teventMsg.WriteString(fmt.Sprintf(\"data:%s\\n\", data.Value))\n\t\t}\n\t}\n\tif id != \"\" {\n\t\teventMsg.WriteString(fmt.Sprintf(\"id:%s\\n\", strings.Replace(id, \"\\n\", \"\", -1)))\n\t}\n\n\teventMsg.WriteString(\"\\n\\n\")\n\treturn eventMsg.String()\n}", "func (f *recordingSource) formatRecord(record *record) string {\n\tf.scratch.Reset()\n\tf.scratch.WriteString(record.Typ.String())\n\tfor _, arg := range record.Args {\n\t\tf.scratch.WriteByte('\\t')\n\t\tf.scratch.WriteString(formatValueWithType(arg))\n\t}\n\treturn f.scratch.String()\n}", "func (d *DefaultFieldsFormatter) Format(entry *logrus.Entry) ([]byte, error) {\n\tif entry != nil {\n\t\tfor k, v := range d.DefaultFields {\n\t\t\tif _, exists := entry.Data[k]; !exists {\n\t\t\t\tentry.Data[k] = v\n\t\t\t}\n\t\t}\n\t}\n\treturn d.WrappedFormatter.Format(entry)\n}", "func (ts TimeStamp) Format(f string) string {\n\treturn ts.FormatInLocation(f, setting.DefaultUILocation)\n}", "func (f *JSONFormatter) Format(level, file, callstack string, msg interface{}) string {\n\tvar jsonstr string\n\tf.Date = time.Now().Format(\"2006-01-02 15:04:05\")\n\tf.Lev = level\n\tf.Msg = msg\n\tf.File = file\n\tif len(callstack) > 0 {\n\t\tf.Callstack = strings.Split(callstack, \"\\n\")\n\t\tfor i, line := range f.Callstack {\n\t\t\tf.Callstack[i] = strings.TrimLeft(line, \"\\t\")\n\t\t}\n\t}\n\n\tdata, err := json.Marshal(&f)\n\tif err != nil {\n\t\tjsonstr = fmt.Sprintf(`{\"date\":\"%s\",\"lev\":\"%s\",\"error\":\"msg to json failed, %s\"}`, f.Date, f.Lev, err.Error())\n\t} else {\n\t\tjsonstr = string(data)\n\t}\n\n\tif !f.indent {\n\t\treturn jsonstr\n\t}\n\n\tvar out bytes.Buffer\n\terr = json.Indent(&out, data, \"\", \"\\t\")\n\tif err != nil {\n\t\treturn jsonstr\n\t}\n\n\treturn string(out.Bytes())\n}", "func (c *T) Logf(format string, args ...interface{})", "func (e EnumEntry) Format(w fmt.State, verb rune) {\n\tif e.IsInternal {\n\t\tfmt.Fprint(w, \"[[internal]] \")\n\t}\n\tfmt.Fprint(w, e.Name)\n}", "func (l *Logger) formatHeader(buf *[]byte, t time.Time, file string, line int) {\n\t*buf = append(*buf, l.prefix...)\n\tif l.flag&(Ldate|Ltime|Lmicroseconds) != 0 {\n\t\tif l.flag&LUTC != 0 {\n\t\t\tt = t.UTC()\n\t\t}\n\t\tif l.flag&Ldate != 0 {\n\t\t\tyear, month, day := t.Date()\n\t\t\titoa(buf, year, 4)\n\t\t\t*buf = append(*buf, '/')\n\t\t\titoa(buf, int(month), 2)\n\t\t\t*buf = append(*buf, '/')\n\t\t\titoa(buf, day, 2)\n\t\t\t*buf = append(*buf, ' ')\n\t\t}\n\t\tif l.flag&(Ltime|Lmicroseconds) != 0 {\n\t\t\thour, min, sec := t.Clock()\n\t\t\titoa(buf, hour, 2)\n\t\t\t*buf = append(*buf, ':')\n\t\t\titoa(buf, min, 2)\n\t\t\t*buf = append(*buf, ':')\n\t\t\titoa(buf, sec, 2)\n\t\t\tif l.flag&Lmicroseconds != 0 {\n\t\t\t\t*buf = append(*buf, '.')\n\t\t\t\titoa(buf, t.Nanosecond()/1e3, 6)\n\t\t\t}\n\t\t\t*buf = append(*buf, ' ')\n\t\t}\n\t}\n\tif l.flag&(Lshortfile|Llongfile) != 0 {\n\t\tif l.flag&Lshortfile != 0 {\n\t\t\tshort := file\n\t\t\tfor i := len(file) - 1; i > 0; i-- {\n\t\t\t\tif file[i] == '/' {\n\t\t\t\t\tshort = file[i+1:]\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\tfile = short\n\t\t}\n\t\t*buf = append(*buf, file...)\n\t\t*buf = append(*buf, ':')\n\t\titoa(buf, line, -1)\n\t\t*buf = append(*buf, \": \"...)\n\t}\n}", "func handleLogEntry(l *LogEntry, key, value string) (err error) {\n\tif l == nil {\n\t\treturn errors.New(\"invalid LogEntry\")\n\t}\n\n\tif err = checkKeyValueValid(key, value); err != nil {\n\t\treturn fmt.Errorf(\"%v (entry: %+v)\", err, l)\n\t}\n\n\tswitch key {\n\tcase \"container\":\n\t\tl.Container = value\n\n\tcase \"level\":\n\t\tl.Level = value\n\n\tcase \"msg\":\n\t\tl.Msg = value\n\n\tcase \"name\":\n\t\tl.Name = value\n\n\tcase \"pid\":\n\t\tpid := 0\n\t\tif value != \"\" {\n\t\t\tpid, err = strconv.Atoi(value)\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"failed to parse pid from value %v (entry: %+v, key: %v): %v\", value, l, key, err)\n\t\t\t}\n\t\t}\n\n\t\tl.Pid = pid\n\n\tcase \"sandbox\":\n\t\tl.Sandbox = value\n\n\tcase \"source\":\n\t\tl.Source = value\n\n\tcase \"time\":\n\t\tt, err := parseTime(value)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"failed to parse time for value %v (entry: %+v, key: %v): %v\", value, l, key, err)\n\t\t}\n\n\t\tl.Time = t\n\n\tdefault:\n\t\tif v, exists := l.Data[key]; exists {\n\t\t\treturn fmt.Errorf(\"key %q already exists in map with value %q (entry: %+v)\", key, v, l)\n\t\t}\n\n\t\t// non-standard fields are stored here\n\t\tl.Data[key] = value\n\t}\n\n\treturn nil\n}", "func BuildFormatterFromConfig(ca *config.Accessor) (logging.StringFormatter, error) {\n\n\tvar mode string\n\tvar err error\n\n\tentryPath := \"LogWriting.Format.Entry\"\n\n\tif mode, err = ca.StringVal(entryPath); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif mode == textEntryMode {\n\n\t\tlmf := new(logging.LogMessageFormatter)\n\n\t\tif err := ca.Populate(\"LogWriting.Format\", lmf); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tif lmf.PrefixFormat == \"\" && lmf.PrefixPreset == \"\" {\n\t\t\tlmf.PrefixPreset = logging.FrameworkPresetPrefix\n\t\t}\n\n\t\treturn lmf, lmf.Init()\n\t} else if mode == jsonEntryMode {\n\n\t\tjmf := new(logging.JSONLogFormatter)\n\n\t\tcfg := new(logging.JSONConfig)\n\n\t\tca.Populate(\"LogWriting.Format.JSON\", cfg)\n\t\tjmf.Config = cfg\n\n\t\tcfg.UTC, _ = ca.BoolVal(\"LogWriting.Format.UtcTimes\")\n\n\t\tcfg.ParsedFields = logging.ConvertFields(cfg.Fields)\n\n\t\tif err := logging.ValidateJSONFields(cfg.ParsedFields); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tif mb, err := logging.CreateMapBuilder(cfg); err == nil {\n\t\t\tjmf.MapBuilder = mb\n\t\t} else {\n\t\t\treturn nil, err\n\t\t}\n\n\t\treturn jmf, nil\n\t}\n\n\treturn nil, fmt.Errorf(\"%s is a not a supported value for %s. Should be %s or %s\", mode, entryPath, textEntryMode, jsonEntryMode)\n\n}", "func (node OnDup) Format(buf *TrackedBuffer) {\n\tif node == nil {\n\t\treturn\n\t}\n\tbuf.astPrintf(node, \" on duplicate key update %v\", UpdateExprs(node))\n}", "func (c *B) Logf(format string, args ...interface{})", "func (h *handler) HandleLog(e *log.Entry) error {\n\th.Lock()\n\tdefer h.Unlock()\n\n\t// if we have a blank line we don't want padding or prefixes\n\tif e.Message == \"\" {\n\t\t_, _ = fmt.Fprintln(h.writer)\n\t\treturn nil\n\t}\n\n\tif h.wantTime {\n\t\tts := h.timer().Format(timeFmt)\n\t\t_, _ = fmt.Fprint(h.writer, appendMissingLineFeed(fmt.Sprintf(\"%s %s%s\", ts, formatLevel(e.Level), e.Message)))\n\t\treturn nil\n\t}\n\n\t_, _ = fmt.Fprint(h.writer, appendMissingLineFeed(fmt.Sprintf(\"%s%s\", formatLevel(e.Level), e.Message)))\n\n\treturn nil\n}", "func (c *Console) handleLog(entries <-chan *log.Entry) {\n\n\tvar e *log.Entry\n\tvar color log.ANSIEscSeq\n\tvar l int\n\n\tfor e = range entries {\n\n\t\tl = len(e.Fields)\n\t\tcolor = c.colors[e.Level]\n\n\t\tif c.miniTimestamp {\n\t\t\tif l == 0 {\n\t\t\t\tfmt.Fprintf(c.writer, \"%s%6s%s[%04d] %s\", color, e.Level, c.ansiReset, c.parseMiniTimestamp(), e.Message)\n\t\t\t} else {\n\t\t\t\tfmt.Fprintf(c.writer, \"%s%6s%s[%04d] %-25s\", color, e.Level, c.ansiReset, c.parseMiniTimestamp(), e.Message)\n\t\t\t}\n\t\t} else {\n\t\t\tif l == 0 {\n\t\t\t\tfmt.Fprintf(c.writer, \"%s%6s%s[%s] %s\", color, e.Level, c.ansiReset, e.Timestamp.Format(c.timestampFormat), e.Message)\n\t\t\t} else {\n\t\t\t\tfmt.Fprintf(c.writer, \"%s%6s%s[%s] %-25s\", color, e.Level, c.ansiReset, e.Timestamp.Format(c.timestampFormat), e.Message)\n\t\t\t}\n\t\t}\n\n\t\tfor _, f := range e.Fields {\n\t\t\tfmt.Fprintf(c.writer, \" %s%s%s=%v\", color, f.Key, c.ansiReset, f.Value)\n\t\t}\n\n\t\tfmt.Fprintln(c.writer)\n\n\t\te.WG.Done()\n\t}\n}", "func Log(format string, a ...interface{}) {\n\ts := fmt.Sprintf(format, a...)\n\ts = fmt.Sprintf(\"%v: %s\", time.Now().Format(\"2006-01-02T15:04:05.000\"), s)\n\tfmt.Println(s)\n}", "func (e *Entry) Infof(format string, v ...interface{}) {\n\tif LogLevelInfo >= e.logger.Level {\n\t\te.logger.printf(LogLevelInfo, e.Context, format, v...)\n\t}\n}", "func (entry *Entry) Logf(level Level, format string, args ...interface{}) {\n\t(*lrs.Entry)(entry).Logf((lrs.Level)(level), format, args)\n}", "func (node *Insert) Format(buf *TrackedBuffer) {\n\tswitch node.Action {\n\tcase InsertAct:\n\t\tbuf.astPrintf(node, \"%s %v%sinto %v%v%v %v%v\",\n\t\t\tInsertStr,\n\t\t\tnode.Comments, node.Ignore.ToString(),\n\t\t\tnode.Table, node.Partitions, node.Columns, node.Rows, node.OnDup)\n\tcase ReplaceAct:\n\t\tbuf.astPrintf(node, \"%s %v%sinto %v%v%v %v%v\",\n\t\t\tReplaceStr,\n\t\t\tnode.Comments, node.Ignore.ToString(),\n\t\t\tnode.Table, node.Partitions, node.Columns, node.Rows, node.OnDup)\n\tdefault:\n\t\tbuf.astPrintf(node, \"%s %v%sinto %v%v%v %v%v\",\n\t\t\t\"Unkown Insert Action\",\n\t\t\tnode.Comments, node.Ignore.ToString(),\n\t\t\tnode.Table, node.Partitions, node.Columns, node.Rows, node.OnDup)\n\t}\n\n}", "func (a *Adapter) formatMessage(ctx context.Context, logLevel string, message string, options ...interface{}) string {\n\n\tvar now = time.Now().Format(\"2006/01/02 15:04:05.000000\")\n\tvar level = a.setTag(logLevel)\n\tvar uuid = \"NONE\"\n\tvar trace = \"NONE\"\n\n\tid, ok := ctx.Value(internal.ID).(string)\n\tif ok {\n\t\tuuid = id\n\t}\n\n\tpoints, ok := ctx.Value(internal.TraceKey).(string)\n\tif ok {\n\t\ttrace = points\n\t}\n\n\tif len(options) == 0 {\n\t\treturn fmt.Sprintf(\"%s %s [%s] [%s] [%s]\", now, level, uuid, trace, message)\n\t}\n\n\treturn fmt.Sprintf(\"%s %s [%s] [%s] [%s] : %v\", now, level, uuid, trace, message, options)\n}", "func ParseLogEntry(args []string) (LogEntry, error) {\n\tvar entry LogEntry\n\tdurs := strings.Split(args[0], \"-\")\n\tif len(durs) != 2 {\n\t\treturn LogEntry{}, fmt.Errorf(\"(%s) is not a valid time format\", args[0])\n\t}\n\n\t// UGLY AF Time parsing... Let's redo that.\n\t// Probably we should start processing elements in reverse and strip message and project, leaving only the time\n\t// definition. Then parse time formats.\n\tyear, month, day := time.Now().Date()\n\tstarttime, err := time.Parse(ShortDate+\" 15:04\", strconv.Itoa(year)+\"/\"+fmt.Sprintf(\"%02d\", month)+\"/\"+fmt.Sprintf(\"%02d\", day)+\" \"+durs[0])\n\tif err != nil {\n\t\tstarttime, err = time.Parse(ShortDate+\" 15:04\", durs[0])\n\t\tif err != nil {\n\t\t\treturn LogEntry{}, err\n\t\t}\n\t}\n\n\tendtime, err := time.Parse(ShortDate+\" 15:04\", strconv.Itoa(year)+\"/\"+fmt.Sprintf(\"%02d\", month)+\"/\"+fmt.Sprintf(\"%02d\", day)+\" \"+durs[1])\n\tif err != nil {\n\t\tendtime, err = time.Parse(ShortDate+\" 15:04\", durs[1])\n\t\tif err != nil {\n\t\t\treturn LogEntry{}, err\n\t\t}\n\t}\n\n\tswitch len(args) {\n\tcase 2:\n\t\tmessage := args[1]\n\n\t\tentry = LogEntry{StartTime: starttime, EndTime: endtime, Project: \"default\", Message: message}\n\n\tcase 3:\n\t\tmessage := args[2]\n\t\tproject := args[1]\n\n\t\tentry = LogEntry{StartTime: starttime, EndTime: endtime, Project: project, Message: message}\n\t}\n\n\treturn entry, nil\n}", "func (ColourFormatter) Format(span Span) string {\n\tduration, _ := time.ParseDuration(fmt.Sprintf(\"%dns\", span.Duration))\n\n\ttags := \"\"\n\n\tfor k, v := range span.Meta {\n\t\ttags = fmt.Sprintf(\"%s %s:%s\", tags, color.CyanString(k), strconv.Quote(v))\n\t}\n\n\treturn fmt.Sprintf(\n\t\t\"[Trace] %s %s %s %s %s %s%s\\n\",\n\t\tcolor.HiCyanString(span.Service),\n\t\tcolor.GreenString(span.Operation),\n\t\tcolor.MagentaString(span.Resource),\n\t\tcolor.WhiteString(span.Type),\n\t\tcolor.YellowString(\"%s\", duration),\n\t\tcolor.WhiteString(\"%d / %d\", span.ParentID, span.SpanID),\n\t\ttags,\n\t)\n}", "func (entry *Entry) Logf(level Level, format string, args ...interface{}) {\n\tif entry.Logger.IsLevelEnabled(level) {\n\t\tentry.Log(level, fmt.Sprintf(format, args...))\n\t}\n}", "func parseLogFmtData(reader io.Reader, file string, ignoreMissingFields bool) (LogEntries, error) {\n\tentries := LogEntries{}\n\n\td := logfmt.NewDecoder(reader)\n\n\tline := uint64(0)\n\n\t// A record is a single line\n\tfor d.ScanRecord() {\n\t\tline++\n\t\tvar keyvals kvPairs\n\n\t\t// split the line into key/value pairs\n\t\tfor d.ScanKeyval() {\n\t\t\tkey := string(d.Key())\n\t\t\tvalue := string(d.Value())\n\n\t\t\t// If agent debug is enabled, every gRPC request (\"req\")\n\t\t\t// is logged. Since most such requests contain the\n\t\t\t// container ID as a `container_id` field, extract and\n\t\t\t// save it when present.\n\t\t\t//\n\t\t\t// See: https://github.com/kata-containers/agent/blob/master/protocols/grpc/agent.proto\n\t\t\t//\n\t\t\t// Note that we save the container ID in addition to\n\t\t\t// the original value.\n\t\t\tif key == \"req\" {\n\t\t\t\tmatches := agentContainerIDRE.FindSubmatch([]byte(value))\n\t\t\t\tif matches != nil {\n\t\t\t\t\tcontainerID := string(matches[1])\n\n\t\t\t\t\tpair := kvPair{\n\t\t\t\t\t\tkey: \"container\",\n\t\t\t\t\t\tvalue: containerID,\n\t\t\t\t\t}\n\n\t\t\t\t\t// save key/value pair\n\t\t\t\t\tkeyvals = append(keyvals, pair)\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tpair := kvPair{\n\t\t\t\tkey: key,\n\t\t\t\tvalue: value,\n\t\t\t}\n\n\t\t\t// save key/value pair\n\t\t\tkeyvals = append(keyvals, pair)\n\t\t}\n\n\t\tif err := d.Err(); err != nil {\n\t\t\treturn LogEntries{},\n\t\t\t\tfmt.Errorf(\"failed to parse file %q, line %d: %v (keyvals: %+v)\",\n\t\t\t\t\tfile, line, err, keyvals)\n\n\t\t}\n\n\t\tentry, err := createLogEntry(file, line, keyvals)\n\t\tif err != nil {\n\t\t\treturn LogEntries{}, err\n\t\t}\n\n\t\terr = entry.Check(ignoreMissingFields)\n\t\tif err != nil {\n\t\t\treturn LogEntries{}, err\n\t\t}\n\n\t\tentries.Entries = append(entries.Entries, entry)\n\t}\n\n\tif d.Err() != nil {\n\t\treturn LogEntries{},\n\t\t\tfmt.Errorf(\"failed to parse file %q line %d: %v\", file, line, d.Err())\n\t}\n\n\treturn entries, nil\n}", "func FormatStyle(f formatStyle) {\n\tcurrentFormat = f\n\tsetupLogger(logFile, logFile, logFile, logFile, logFile)\n}", "func (node *Commit) Format(buf *TrackedBuffer) {\n\tbuf.WriteString(\"commit\")\n}", "func formatIssue(issue *maintner.GitHubIssue) []byte {\n\tconst timeFormat = \"15:04 on 2 Jan 2006\"\n\tvar buf bytes.Buffer\n\tfmt.Fprintf(&buf, \"%s\\ncreated %s at %s\\n\\n%s\\n\",\n\t\tissue.Title,\n\t\tformatUser(issue.User),\n\t\tissue.Created.Format(timeFormat),\n\t\twrap(\"\\t\", issue.Body))\n\n\ttype update struct {\n\t\ttime time.Time\n\t\tprinted []byte\n\t}\n\tvar updates []update\n\tissue.ForeachComment(func(comment *maintner.GitHubComment) error {\n\t\tvar buf bytes.Buffer\n\t\tfmt.Fprintf(&buf, \"comment %s at %s\\n\\n%s\\n\",\n\t\t\tformatUser(comment.User),\n\t\t\tcomment.Created.Format(timeFormat),\n\t\t\twrap(\"\\t\", comment.Body))\n\t\tupdates = append(updates, update{comment.Created, buf.Bytes()})\n\t\treturn nil\n\t})\n\tissue.ForeachEvent(func(event *maintner.GitHubIssueEvent) error {\n\t\tvar buf bytes.Buffer\n\t\tswitch event.Type {\n\t\tcase \"closed\", \"reopened\":\n\t\t\tfmt.Fprintf(&buf, \"%s %s at %s\\n\\n\",\n\t\t\t\tevent.Type,\n\t\t\t\tformatUser(event.Actor),\n\t\t\t\tevent.Created.Format(timeFormat))\n\t\tdefault:\n\t\t\t// TODO(adg): other types\n\t\t}\n\t\tupdates = append(updates, update{event.Created, buf.Bytes()})\n\t\treturn nil\n\t})\n\tsort.Slice(updates, func(i, j int) bool {\n\t\treturn updates[i].time.Before(updates[j].time)\n\t})\n\tfor _, u := range updates {\n\t\tbuf.Write(u.printed)\n\t}\n\treturn buf.Bytes()\n}", "func setLogFormat(format string) {\n\n\tswitch format {\n\tcase \"TEXT\":\n\t\tlogrus.SetFormatter(&logrus.TextFormatter{})\n\tcase \"JSON\":\n\t\tlogrus.SetFormatter(&logrus.JSONFormatter{})\n\tdefault:\n\t\t//We are going to begin using Kibana for logging, Kibana requires logs outputted as JSON to build dashboards for\n\t\tlogrus.SetFormatter(&logrus.JSONFormatter{})\n\t}\n}", "func (c CommitMessage) Format() string {\n\t// HEADER\n\tvar msg = c.Type\n\tif c.Scope != \"\" {\n\t\tmsg += \"(\" + c.Scope + \")\"\n\t}\n\tmsg += \": \" + c.Subject + \"\\n\\n\"\n\n\t// BODY\n\tif c.Body != \"\" {\n\t\tmsg += wordwrap.WrapString(c.Body, 79) + \"\\n\\n\"\n\t}\n\t// FOOTER\n\tif c.Close != \"\" {\n\t\tmsg += c.issuesList()\n\t}\n\treturn strings.TrimSpace(msg)\n}", "func FormatRequestLog(req *restful.Request) string {\n\treqURI := \"\"\n\tif req.Request.URL != nil {\n\t\treqURI = req.Request.URL.RequestURI()\n\t}\n\n\treturn fmt.Sprintf(RequestLogString, req.Request.Proto, req.Request.Method,\n\t\treqURI, req.Request.RemoteAddr)\n}", "func DbLoggerFormatter(r *ViewSetRunTime, v ...interface{}) string {\n\n\tdblogLevel, _ := v[0].(string)\n\tl := &LogFormat{\n\t\tTimestamp: time.Now().Format(time.RFC3339),\n\t\tVersion: GlobalTrinity.setting.GetProjectVersion(),\n\t\tMessage: r.Gcontext.Errors.String(),\n\t\tLoggerName: \"\",\n\t\tThreadName: \"\",\n\t\tLevel: \"\",\n\t\tHostname: \"hostname\",\n\t\tModuleName: GlobalTrinity.setting.GetProjectName(),\n\t\tTraceID: r.TraceID,\n\t\t// Latency: params.Latency,\n\t\tClientIP: r.Gcontext.ClientIP(),\n\t\tHTTPMethod: r.Gcontext.Request.Method,\n\t\tHTTPPath: r.Gcontext.Request.URL.RequestURI(),\n\t\tHTTPStatusCode: r.Gcontext.Writer.Status(),\n\t\tBodySize: r.Gcontext.Writer.Size(),\n\t\tUserID: r.Gcontext.GetInt64(\"UserID\"),\n\t\tUsername: r.Gcontext.GetString(\"Username\"),\n\t\tErrorDetail: r.Gcontext.GetString(\"ErrorDetail\"),\n\t\tDBLogOrigin: fmt.Sprint(gorm.LogFormatter(v...)),\n\t}\n\tif dblogLevel == \"sql\" {\n\t\tl.DBRunningFile = fmt.Sprint(v[1])\n\t\tl.DBRunningTime, _ = v[2].(time.Duration)\n\t\tl.DBSQL = fmt.Sprint(v[3])\n\t\tl.DBParams = fmt.Sprint(v[4])\n\t\tl.DBEffectedRows = fmt.Sprint(v[5])\n\t}\n\treturn l.GetString()\n}", "func (node Argument) Format(buf *TrackedBuffer) {\n\tbuf.WriteArg(string(node))\n}", "func SetFormat(v string) {\n\tlog.SetFormat(v)\n}" ]
[ "0.75642234", "0.7525652", "0.7511716", "0.7501286", "0.74737483", "0.747359", "0.72316986", "0.70732135", "0.7003363", "0.69529295", "0.6900151", "0.6899977", "0.68932986", "0.6867317", "0.68633306", "0.6830301", "0.6806357", "0.6727993", "0.6656632", "0.6656256", "0.6641873", "0.6631874", "0.6600519", "0.65839857", "0.65586895", "0.6485496", "0.6455386", "0.63429034", "0.63375425", "0.6334442", "0.6328312", "0.6257293", "0.62567997", "0.62397045", "0.62273425", "0.6223774", "0.61984026", "0.61618865", "0.6141984", "0.6086046", "0.6077044", "0.60562193", "0.60488427", "0.6038531", "0.60362136", "0.603199", "0.6001361", "0.59854126", "0.5943373", "0.5907314", "0.58913", "0.57692873", "0.5768981", "0.5756823", "0.5755852", "0.57316947", "0.5717292", "0.5691044", "0.5678024", "0.56766707", "0.56758815", "0.5663223", "0.5646094", "0.5641495", "0.5634903", "0.5621264", "0.5610773", "0.56081337", "0.56058663", "0.5603703", "0.5583089", "0.55624664", "0.55617416", "0.55459833", "0.554566", "0.55452985", "0.55270517", "0.55243075", "0.5518944", "0.550195", "0.5500231", "0.54844505", "0.54820734", "0.54694355", "0.5454859", "0.54539686", "0.5443611", "0.5440716", "0.5429201", "0.5422071", "0.5398095", "0.5389252", "0.53839856", "0.53805554", "0.5375383", "0.5371057", "0.53671217", "0.53565145", "0.5354626", "0.5351468" ]
0.5486447
81
NewSearchableSnapshots provides a builder for the SearchableSnapshots struct.
func NewSearchableSnapshotsBuilder() *SearchableSnapshotsBuilder { r := SearchableSnapshotsBuilder{ &SearchableSnapshots{}, } return &r }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (rb *SearchableSnapshotsBuilder) Build() SearchableSnapshots {\n\treturn *rb.v\n}", "func newSearch(inst *Instagram) *Search {\n\tsearch := &Search{\n\t\tinst: inst,\n\t}\n\treturn search\n}", "func (*publisherSearchIdxR) NewStruct() *publisherSearchIdxR {\n\treturn &publisherSearchIdxR{}\n}", "func NewListSnapshotsRequest(server string) (*http.Request, error) {\n\tvar err error\n\n\tqueryUrl, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbasePath := fmt.Sprintf(\"/snapshot\")\n\tif basePath[0] == '/' {\n\t\tbasePath = basePath[1:]\n\t}\n\n\tqueryUrl, err = queryUrl.Parse(basePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq, err := http.NewRequest(\"GET\", queryUrl.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn req, nil\n}", "func NewSnapshot(version string,\n\tendpoints []types.Resource,\n\tclusters []types.Resource,\n\troutes []types.Resource,\n\tlisteners []types.Resource,\n\truntimes []types.Resource) Snapshot {\n\tout := Snapshot{}\n\tout.Resources[types.Endpoint] = NewResources(version, endpoints)\n\tout.Resources[types.Cluster] = NewResources(version, clusters)\n\tout.Resources[types.Route] = NewResources(version, routes)\n\tout.Resources[types.Listener] = NewResources(version, listeners)\n\tout.Resources[types.Runtime] = NewResources(version, runtimes)\n\treturn out\n}", "func New(slice *[]int) (*Search, error) {\n\tvar s = Search{slice: slice}\n\treturn &s, nil\n}", "func (m *Nitro) NewSnapshot() (*Snapshot, error) {\n\tbuf := m.snapshots.MakeBuf()\n\tdefer m.snapshots.FreeBuf(buf)\n\n\t// Stitch all local gclists from all writers to create snapshot gclist\n\tvar head, tail *skiplist.Node\n\n\tfor w := m.wlist; w != nil; w = w.next {\n\t\tif tail == nil {\n\t\t\thead = w.gchead\n\t\t\ttail = w.gctail\n\t\t} else if w.gchead != nil {\n\t\t\ttail.SetLink(w.gchead)\n\t\t\ttail = w.gctail\n\t\t}\n\n\t\tw.gchead = nil\n\t\tw.gctail = nil\n\n\t\t// Update global stats\n\t\tm.store.Stats.Merge(&w.slSts1)\n\t\tatomic.AddInt64(&m.itemsCount, w.count)\n\t\tw.count = 0\n\t}\n\n\tsnap := &Snapshot{db: m, sn: m.GetCurrSn(), refCount: 1, count: m.ItemsCount()}\n\tm.snapshots.Insert(unsafe.Pointer(snap), CompareSnapshot, buf, &m.snapshots.Stats)\n\tsnap.gclist = head\n\tnewSn := atomic.AddUint32(&m.currSn, 1)\n\tif newSn == math.MaxUint32 {\n\t\treturn nil, ErrMaxSnapshotsLimitReached\n\t}\n\n\treturn snap, nil\n}", "func New(notifier MountNotifier, statePath string) *Snapshot {\n\t// Initialize default library driver\n\ts := &Snapshot{\n\t\tnotifier: newLocalStateNotifier(statePath, notifier),\n\t}\n\ts.driver = csicommon.NewCSIDriver(\"container-snapshot.openshift.io\", \"0.0.1\", uuid.New())\n\tif s.driver == nil {\n\t\tpanic(\"unable to initialize driver\")\n\t}\n\ts.driver.AddControllerServiceCapabilities([]csi.ControllerServiceCapability_RPC_Type{\n\t\tcsi.ControllerServiceCapability_RPC_CREATE_DELETE_VOLUME,\n\t})\n\ts.driver.AddVolumeCapabilityAccessModes([]csi.VolumeCapability_AccessMode_Mode{\n\t\tcsi.VolumeCapability_AccessMode_SINGLE_NODE_WRITER,\n\t\tcsi.VolumeCapability_AccessMode_MULTI_NODE_MULTI_WRITER,\n\t})\n\n\t// Create GRPC servers\n\ts.identity = &identityServer{\n\t\tDefaultIdentityServer: csicommon.NewDefaultIdentityServer(s.driver),\n\t}\n\ts.node = &nodeServer{\n\t\tDefaultNodeServer: csicommon.NewDefaultNodeServer(s.driver),\n\t\tnotifier: s.notifier,\n\t}\n\ts.controller = &controllerServer{\n\t\tDefaultControllerServer: csicommon.NewDefaultControllerServer(s.driver),\n\t}\n\treturn s\n}", "func (r *SearchREST) New() runtime.Object {\n\treturn &searchapis.Search{}\n}", "func NewSearch() *Search {\n\treturn &Search{}\n}", "func NewSearch() *Search {\n\ts := &Search{}\n\tdefaults.SetDefaults(s)\n\treturn s\n}", "func (*repositoryR) NewStruct() *repositoryR {\n\treturn &repositoryR{}\n}", "func (c *Client) NewSearch(searchType string, params *SearchParams) *Search {\n\treturn &Search{\n\t\tclient: c,\n\t\tType: searchType,\n\t\tParams: params,\n\t\tDeleted: false,\n\t}\n}", "func New(meta Metadata, data []byte) (*Snapshot, error) {\n\n\tvar b bytes.Buffer\n\tgw, err := gzip.NewWriterLevel(&b, gzip.BestSpeed)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error building gzip writer: %w\", err)\n\t}\n\tgw.Write(data)\n\tgw.Close()\n\n\treturn &Snapshot{meta: meta, data: b.Bytes()}, nil\n}", "func (txn TxnProbe) NewScanner(start, end []byte, batchSize int, reverse bool) (*txnsnapshot.Scanner, error) {\n\tsnapshot := txnsnapshot.SnapshotProbe{KVSnapshot: txn.GetSnapshot()}\n\treturn snapshot.NewScanner(start, end, batchSize, reverse)\n}", "func (ht *historyTable) newSearch() {\n\tfor i := range ht {\n\t\tht[i].stat /= 8\n\t}\n}", "func NewSnapshot(kind string, startRevision, lastRevision int64, compressionSuffix string, isFinal bool) *brtypes.Snapshot {\n\tsnap := &brtypes.Snapshot{\n\t\tKind: kind,\n\t\tStartRevision: startRevision,\n\t\tLastRevision: lastRevision,\n\t\tCreatedOn: time.Now().UTC(),\n\t\tCompressionSuffix: compressionSuffix,\n\t\tIsFinal: isFinal,\n\t}\n\tsnap.GenerateSnapshotName()\n\treturn snap\n}", "func NewSearchIdleBuilder() *SearchIdleBuilder {\n\tr := SearchIdleBuilder{\n\t\t&SearchIdle{},\n\t}\n\n\treturn &r\n}", "func NewSearchHandler(s registry.Searchable) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tp := &registry.SearchParams{}\n\t\tswitch r.Header.Get(\"Content-Type\") {\n\t\tcase \"application/json\":\n\t\t\tif err := json.NewDecoder(r.Body).Decode(p); err != nil {\n\t\t\t\tapiutil.WriteErrResponse(w, http.StatusBadRequest, err)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif p.Limit == 0 {\n\t\t\t\tp.Limit = defaultLimit\n\t\t\t}\n\t\tdefault:\n\t\t\t// read form values\n\t\t\tvar err error\n\t\t\tif p.Limit, err = apiutil.ReqParamInt(\"limit\", r); err != nil {\n\t\t\t\tp.Limit = defaultLimit\n\t\t\t\terr = nil\n\t\t\t}\n\t\t\tif p.Offset, err = apiutil.ReqParamInt(\"offset\", r); err != nil {\n\t\t\t\tp.Offset = defaultOffset\n\t\t\t\terr = nil\n\t\t\t}\n\t\t\tp.Q = r.FormValue(\"q\")\n\t\t}\n\t\tswitch r.Method {\n\t\tcase \"GET\":\n\t\t\tresults, err := s.Search(*p)\n\t\t\tif err != nil {\n\t\t\t\tapiutil.WriteErrResponse(w, http.StatusBadRequest, err)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tapiutil.WriteResponse(w, results)\n\t\t\treturn\n\t\t}\n\t}\n}", "func NewSearchBucket()(*SearchBucket) {\n m := &SearchBucket{\n }\n m.SetAdditionalData(make(map[string]interface{}));\n return m\n}", "func NewBinarySearch() *BinarySearch {\n\treturn &BinarySearch{}\n}", "func NewSnapshotter(root string) (snapshots.Snapshotter, error) {\n\tfsType, err := getFileSystemType(root)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif strings.ToLower(fsType) != \"ntfs\" {\n\t\treturn nil, errors.Wrapf(errdefs.ErrInvalidArgument, \"%s is not on an NTFS volume - only NTFS volumes are supported\", root)\n\t}\n\n\tif err := os.MkdirAll(root, 0700); err != nil {\n\t\treturn nil, err\n\t}\n\tms, err := storage.NewMetaStore(filepath.Join(root, \"metadata.db\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err := os.Mkdir(filepath.Join(root, \"snapshots\"), 0700); err != nil && !os.IsExist(err) {\n\t\treturn nil, err\n\t}\n\n\treturn &snapshotter{\n\t\troot: root,\n\t\tms: ms,\n\t}, nil\n}", "func newSearchService(sling *sling.Sling) *SearchService {\n\treturn &SearchService{\n\t\tsling: sling.Path(\"search/\"),\n\t}\n}", "func (mdb *memdbSlice) NewSnapshot(ts *common.TsVbuuid, commit bool) (SnapshotInfo, error) {\n\n\tmdb.waitPersist()\n\n\tqc := atomic.LoadInt64(&mdb.qCount)\n\tif qc > 0 {\n\t\tcommon.CrashOnError(errors.New(\"Slice Invariant Violation - commit with pending mutations\"))\n\t}\n\n\tmdb.isDirty = false\n\n\t// Coming here means that cmdCh is empty and flush has finished for this index\n\tatomic.StoreUint32(&mdb.flushActive, 0)\n\n\tsnap, err := mdb.mainstore.NewSnapshot()\n\tif err == memdb.ErrMaxSnapshotsLimitReached {\n\t\tlogging.Warnf(\"Maximum snapshots limit reached for indexer. Restarting indexer...\")\n\t\tos.Exit(0)\n\t}\n\n\tnewSnapshotInfo := &memdbSnapshotInfo{\n\t\tTs: ts,\n\t\tMainSnap: snap,\n\t\tCommitted: commit,\n\t}\n\tmdb.setCommittedCount()\n\n\treturn newSnapshotInfo, err\n}", "func NewSnapshot(options etcd.Options) (store.Snapshot, error) {\n\tcli, err := etcd.NewClient(options)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &snapshot{\n\t\tclient: cli,\n\t\tkey: DefaultKey,\n\t}, nil\n}", "func NewCreateImageFromSnapshotsRequestWithoutParam() *CreateImageFromSnapshotsRequest {\n\n return &CreateImageFromSnapshotsRequest{\n JDCloudRequest: core.JDCloudRequest{\n URL: \"/regions/{regionId}/images:createImageFromSnapshots\",\n Method: \"POST\",\n Header: nil,\n Version: \"v1\",\n },\n }\n}", "func SearchBarNew() (*SearchBar, error) {\n\tc := C.gtk_search_bar_new()\n\tif c == nil {\n\t\treturn nil, nilPtrErr\n\t}\n\tobj := glib.Take(unsafe.Pointer(c))\n\treturn wrapSearchBar(obj), nil\n}", "func NewStreamBackupSearch(storage storage.ExternalStorage, comparator Comparator, searchKey []byte) *StreamBackupSearch {\n\tbs := &StreamBackupSearch{\n\t\tstorage: storage,\n\t\tcomparator: comparator,\n\t}\n\n\tbs.searchKey = codec.EncodeBytes([]byte{}, searchKey)\n\treturn bs\n}", "func newNativeSnapshot(c *C.rocksdb_snapshot_t) *Snapshot {\n\treturn &Snapshot{c: c}\n}", "func NewSearchDir() *SearchDir {\n newObj := SearchDir {\n DoneChan: make(chan bool),\n ErrChan: make(chan string),\n FileChan: make(chan string),\n }\n\n return &newObj\n}", "func New(tp elastictransport.Interface) *Search {\n\tr := &Search{\n\t\ttransport: tp,\n\t\tvalues: make(url.Values),\n\t\theaders: make(http.Header),\n\t\tbuf: gobytes.NewBuffer(nil),\n\n\t\treq: NewRequest(),\n\t}\n\n\treturn r\n}", "func New(prototype Aggregate, opts ...Option) *Repository {\n\tt := reflect.TypeOf(prototype)\n\tif t.Kind() == reflect.Ptr {\n\t\tt = t.Elem()\n\t}\n\n\tr := &Repository{\n\t\tprototype: t,\n\t\tstore: newMemoryStore(),\n\t\tserializer: NewJSONSerializer(),\n\t}\n\n\tfor _, opt := range opts {\n\t\topt(r)\n\t}\n\n\treturn r\n}", "func NewSearcher(lat_tiles, lng_tiles int) Searcher {\n\tsearcher := Searcher{}\n\tsearcher.lat_tiles = lat_tiles\n\tsearcher.lng_tiles = lng_tiles\n\tsearcher.locatable_map = newLocatableMap(lat_tiles, lng_tiles)\n\treturn searcher\n}", "func New(t opentracing.Tracer, geoconn, rateconn *grpc.ClientConn) *Search {\n\treturn &Search{\n\t\tgeoClient: geo.NewGeoClient(geoconn),\n\t\trateClient: rate.NewRateClient(rateconn),\n\t\ttracer: t,\n\t}\n}", "func NewSettingsSearchBuilder() *SettingsSearchBuilder {\n\tr := SettingsSearchBuilder{\n\t\t&SettingsSearch{},\n\t}\n\n\treturn &r\n}", "func (c *restClient) ListSnapshots(ctx context.Context, req *netapppb.ListSnapshotsRequest, opts ...gax.CallOption) *SnapshotIterator {\n\tit := &SnapshotIterator{}\n\treq = proto.Clone(req).(*netapppb.ListSnapshotsRequest)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tit.InternalFetch = func(pageSize int, pageToken string) ([]*netapppb.Snapshot, string, error) {\n\t\tresp := &netapppb.ListSnapshotsResponse{}\n\t\tif pageToken != \"\" {\n\t\t\treq.PageToken = pageToken\n\t\t}\n\t\tif pageSize > math.MaxInt32 {\n\t\t\treq.PageSize = math.MaxInt32\n\t\t} else if pageSize != 0 {\n\t\t\treq.PageSize = int32(pageSize)\n\t\t}\n\t\tbaseUrl, err := url.Parse(c.endpoint)\n\t\tif err != nil {\n\t\t\treturn nil, \"\", err\n\t\t}\n\t\tbaseUrl.Path += fmt.Sprintf(\"/v1/%v/snapshots\", req.GetParent())\n\n\t\tparams := url.Values{}\n\t\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\t\tif req.GetFilter() != \"\" {\n\t\t\tparams.Add(\"filter\", fmt.Sprintf(\"%v\", req.GetFilter()))\n\t\t}\n\t\tif req.GetOrderBy() != \"\" {\n\t\t\tparams.Add(\"orderBy\", fmt.Sprintf(\"%v\", req.GetOrderBy()))\n\t\t}\n\t\tif req.GetPageSize() != 0 {\n\t\t\tparams.Add(\"pageSize\", fmt.Sprintf(\"%v\", req.GetPageSize()))\n\t\t}\n\t\tif req.GetPageToken() != \"\" {\n\t\t\tparams.Add(\"pageToken\", fmt.Sprintf(\"%v\", req.GetPageToken()))\n\t\t}\n\n\t\tbaseUrl.RawQuery = params.Encode()\n\n\t\t// Build HTTP headers from client and context metadata.\n\t\thds := append(c.xGoogHeaders, \"Content-Type\", \"application/json\")\n\t\theaders := gax.BuildHeaders(ctx, hds...)\n\t\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\t\tif settings.Path != \"\" {\n\t\t\t\tbaseUrl.Path = settings.Path\n\t\t\t}\n\t\t\thttpReq, err := http.NewRequest(\"GET\", baseUrl.String(), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\thttpReq.Header = headers\n\n\t\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tdefer httpRsp.Body.Close()\n\n\t\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn nil\n\t\t}, opts...)\n\t\tif e != nil {\n\t\t\treturn nil, \"\", e\n\t\t}\n\t\tit.Response = resp\n\t\treturn resp.GetSnapshots(), resp.GetNextPageToken(), nil\n\t}\n\n\tfetch := func(pageSize int, pageToken string) (string, error) {\n\t\titems, nextPageToken, err := it.InternalFetch(pageSize, pageToken)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tit.items = append(it.items, items...)\n\t\treturn nextPageToken, nil\n\t}\n\n\tit.pageInfo, it.nextFunc = iterator.NewPageInfo(fetch, it.bufLen, it.takeBuf)\n\tit.pageInfo.MaxSize = int(req.GetPageSize())\n\tit.pageInfo.Token = req.GetPageToken()\n\n\treturn it\n}", "func newSnapshotQueue(n int, w int, l logger.Logger) chan *fragment {\n\tch := make(chan *fragment, n)\n\tfor i := 0; i < w; i++ {\n\t\tgo snapshotQueueWorker(ch, l)\n\t}\n\treturn ch\n}", "func NewSearcher(cfg *Config, repos KeyValueStorer) searcher {\n\treturn searcher{\n\t\tcfg: cfg,\n\t\trepos: repos,\n\t}\n}", "func newSnapshotCache() cache.SnapshotCache {\n\treturn cache.NewSnapshotCache(false, tbnProxyNodeHash{}, consoleLogger{})\n}", "func NewSnapshot(ctx *pulumi.Context,\n\tname string, args *SnapshotArgs, opts ...pulumi.ResourceOption) (*Snapshot, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.VolumeId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'VolumeId'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Snapshot\n\terr := ctx.RegisterResource(\"aws-native:fsx:Snapshot\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (s *API) ListSnapshots(req *ListSnapshotsRequest, opts ...scw.RequestOption) (*ListSnapshotsResponse, error) {\n\tvar err error\n\n\tif req.Region == \"\" {\n\t\tdefaultRegion, _ := s.client.GetDefaultRegion()\n\t\treq.Region = defaultRegion\n\t}\n\n\tdefaultPageSize, exist := s.client.GetDefaultPageSize()\n\tif (req.PageSize == nil || *req.PageSize == 0) && exist {\n\t\treq.PageSize = &defaultPageSize\n\t}\n\n\tquery := url.Values{}\n\tparameter.AddToQuery(query, \"name\", req.Name)\n\tparameter.AddToQuery(query, \"order_by\", req.OrderBy)\n\tparameter.AddToQuery(query, \"instance_id\", req.InstanceID)\n\tparameter.AddToQuery(query, \"organization_id\", req.OrganizationID)\n\tparameter.AddToQuery(query, \"project_id\", req.ProjectID)\n\tparameter.AddToQuery(query, \"page\", req.Page)\n\tparameter.AddToQuery(query, \"page_size\", req.PageSize)\n\n\tif fmt.Sprint(req.Region) == \"\" {\n\t\treturn nil, errors.New(\"field Region cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"GET\",\n\t\tPath: \"/rdb/v1/regions/\" + fmt.Sprint(req.Region) + \"/snapshots\",\n\t\tQuery: query,\n\t\tHeaders: http.Header{},\n\t}\n\n\tvar resp ListSnapshotsResponse\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func (r *ProjectsLocationsVolumesSnapshotsService) List(parent string) *ProjectsLocationsVolumesSnapshotsListCall {\n\tc := &ProjectsLocationsVolumesSnapshotsListCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.parent = parent\n\treturn c\n}", "func (s *Snapshot) NewIter(o *IterOptions) (*Iterator, error) {\n\treturn s.NewIterWithContext(context.Background(), o)\n}", "func newSearcher() *defaultSearcher {\n\treturn &defaultSearcher{\n\t\tpathStringer: new(defaultPathStringer),\n\t}\n}", "func (*kvstoreR) NewStruct() *kvstoreR {\n\treturn &kvstoreR{}\n}", "func NewSnapshot(ctx *pulumi.Context,\n\tname string, args *SnapshotArgs, opts ...pulumi.ResourceOption) (*Snapshot, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.InstanceId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'InstanceId'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Snapshot\n\terr := ctx.RegisterResource(\"alicloud:databasefilesystem/snapshot:Snapshot\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (client *WebAppsClient) listSnapshotsCreateRequest(ctx context.Context, resourceGroupName string, name string, options *WebAppsListSnapshotsOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Web/sites/{name}/snapshots\"\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif name == \"\" {\n\t\treturn nil, errors.New(\"parameter name cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{name}\", url.PathEscape(name))\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-02-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func (client *HearthstoneAPI) newCardCollectionSearch() *cardCollectionSearch {\n\treturn &cardCollectionSearch{\n\t\turl: client.apiURL,\n\t\tlocale: client.locale,\n\t\toptionalString: make(map[string]string),\n\t\toptionalInt: make(map[string]int),\n\t}\n}", "func NewRepository(templateInfos map[string]Info) Repository {\n\tif templateInfos == nil {\n\t\treturn repo{\n\t\t\tinfo: make(map[string]Info),\n\t\t\tallSupportedTmpls: make([]string, 0),\n\t\t\ttmplToBuilderNames: make(map[string]string),\n\t\t}\n\t}\n\n\tallSupportedTmpls := make([]string, len(templateInfos))\n\ttmplToBuilderNames := make(map[string]string)\n\n\tfor t, v := range templateInfos {\n\t\tallSupportedTmpls = append(allSupportedTmpls, t)\n\t\ttmplToBuilderNames[t] = v.BldrInterfaceName\n\t}\n\treturn repo{info: templateInfos, tmplToBuilderNames: tmplToBuilderNames, allSupportedTmpls: allSupportedTmpls}\n}", "func NewSnapshot(ctx *pulumi.Context,\n\tname string, args *SnapshotArgs, opts ...pulumi.ResourceOption) (*Snapshot, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.NamespaceName == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'NamespaceName'\")\n\t}\n\tif args.SnapshotName == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'SnapshotName'\")\n\t}\n\tvar resource Snapshot\n\terr := ctx.RegisterResource(\"aws:redshiftserverless/snapshot:Snapshot\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func NewBuilder(nc sous.Inserter, drh string, sourceShell, scratchShell shell.Shell, ls logging.LogSink) (*Builder, error) {\n\tb := &Builder{\n\t\tImageMapper: nc,\n\t\tDockerRegistryHost: drh,\n\t\tSourceShell: sourceShell,\n\t\tScratchShell: scratchShell,\n\t\tlog: ls,\n\t}\n\n\tfiles, err := scratchShell.List()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif len(files) != 0 {\n\t\treturn nil, fmt.Errorf(\"scratch dir %s was not empty\", scratchShell.Dir())\n\t}\n\n\treturn b, nil\n}", "func (s *ImagesByRepositoryRegistryStorage) New() interface{} {\n\treturn &imageapi.ImageRepositoryMapping{}\n}", "func newVolumeSnapshotSchedules(c *StorkV1alpha1Client, namespace string) *volumeSnapshotSchedules {\n\treturn &volumeSnapshotSchedules{\n\t\tclient: c.RESTClient(),\n\t\tns: namespace,\n\t}\n}", "func (es *EventuallyFileOnlySnapshot) NewIter(o *IterOptions) (*Iterator, error) {\n\treturn es.NewIterWithContext(context.Background(), o)\n}", "func New(tableName string) *Builder {\n\tb := &Builder{\n\t\tstrToBucket: map[string]stringInfo{},\n\t\tbuckets: [][]byte{nil}, // initialize with first bucket.\n\t}\n\tb.rootMeta = &metaData{\n\t\tb: b,\n\t\ttypeInfo: &typeInfo{},\n\t}\n\treturn b\n}", "func (client *Client) ListSnapshotsWithOptions(request *ListSnapshotsRequest, runtime *util.RuntimeOptions) (_result *ListSnapshotsResponse, _err error) {\n\t_err = util.ValidateModel(request)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\tquery := map[string]interface{}{}\n\tif !tea.BoolValue(util.IsUnset(request.AuthTimeout)) {\n\t\tquery[\"AuthTimeout\"] = request.AuthTimeout\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.PageNo)) {\n\t\tquery[\"PageNo\"] = request.PageNo\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.PageSize)) {\n\t\tquery[\"PageSize\"] = request.PageSize\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.SnapshotType)) {\n\t\tquery[\"SnapshotType\"] = request.SnapshotType\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.VideoId)) {\n\t\tquery[\"VideoId\"] = request.VideoId\n\t}\n\n\treq := &openapi.OpenApiRequest{\n\t\tQuery: openapiutil.Query(query),\n\t}\n\tparams := &openapi.Params{\n\t\tAction: tea.String(\"ListSnapshots\"),\n\t\tVersion: tea.String(\"2017-03-21\"),\n\t\tProtocol: tea.String(\"HTTPS\"),\n\t\tPathname: tea.String(\"/\"),\n\t\tMethod: tea.String(\"POST\"),\n\t\tAuthType: tea.String(\"AK\"),\n\t\tStyle: tea.String(\"RPC\"),\n\t\tReqBodyType: tea.String(\"formData\"),\n\t\tBodyType: tea.String(\"json\"),\n\t}\n\t_result = &ListSnapshotsResponse{}\n\t_body, _err := client.CallApi(params, req, runtime)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\t_err = tea.Convert(_body, &_result)\n\treturn _result, _err\n}", "func NewSnapshotter(baseDirPath string) (*KopiaSnapshotter, error) {\n\tks := &KopiaSnapshotter{}\n\n\tif err := ks.initializeConnector(baseDirPath); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn ks, nil\n}", "func NewSearchLooksOK() *SearchLooksOK {\n\treturn &SearchLooksOK{}\n}", "func NewRunningStateSearchIntervalBuilder() *RunningStateSearchIntervalBuilder {\n\tr := RunningStateSearchIntervalBuilder{\n\t\t&RunningStateSearchInterval{},\n\t}\n\n\treturn &r\n}", "func (s *SnapshotsServiceOp) List(ctx context.Context, opt *ListOptions) ([]Snapshot, *Response, error) {\n\treturn s.list(ctx, opt, nil)\n}", "func New() Builder {\n\treturn newBuilder()\n}", "func (ks *KopiaSnapshotter) ListSnapshots() ([]string, error) {\n\treturn ks.snap.ListSnapshots()\n}", "func New() *Builder {\n\treturn &Builder{\n\t\tHeaders: make(map[string]string),\n\t\tQuerys: make(map[string]string),\n\t\tlogger: log.New(os.Stdout, \"\", log.LstdFlags),\n\t\ttimeout: time.Duration(20 * time.Second),\n\t}\n}", "func (s *SnapshotsServiceOp) List(opt *ListOptions) ([]Snapshot, *Response, error) {\n\treturn s.list(opt, nil)\n}", "func New() *Versions {\n\tvar obj Versions\n\treturn &obj\n}", "func NewSnapshotsFromProto(s []*auctioneerrpc.BatchSnapshotResponse) []*Snapshot {\n\tresult := make([]*Snapshot, len(s))\n\tfor idx, snapshot := range s {\n\t\tresult[idx] = &Snapshot{\n\t\t\tVersion: snapshot.Version,\n\t\t\tBatchID: hex.EncodeToString(\n\t\t\t\tsnapshot.BatchId,\n\t\t\t),\n\t\t\tPrevBatchID: hex.EncodeToString(\n\t\t\t\tsnapshot.PrevBatchId,\n\t\t\t),\n\t\t\tBatchTxID: snapshot.BatchTxId,\n\t\t\tBatchTx: hex.EncodeToString(\n\t\t\t\tsnapshot.BatchTx,\n\t\t\t),\n\t\t\tBatchTxFeeRateSatPerKw: snapshot.BatchTxFeeRateSatPerKw,\n\t\t\tCreationTimestampNs: snapshot.CreationTimestampNs,\n\t\t\tMatchedMarkets: snapshot.MatchedMarkets,\n\t\t}\n\t}\n\n\treturn result\n}", "func NewSnapshotHandler(c cache.SnapshotCache, resources []ResourceCache, logger logrus.FieldLogger) *SnapshotHandler {\n\n\tsh := &SnapshotHandler{\n\t\tsnapshotCache: c,\n\t\tresources: parseResources(resources),\n\t\tFieldLogger: logger,\n\t}\n\n\treturn sh\n}", "func NewWalker(searchDomain *Domain, searchParameters *Parameters, searchObjectives *MultiObjective) Walker {\n\n\t// generate uuid\n\tuuid := uuid.NewV4()\n\n\t// create, and return the walker\n\twalker := Walker{\n\t\tId: uuid,\n\t\tSearchDomain: searchDomain,\n\t\tSearchParameters: searchParameters,\n\t\tSearchObjectives: searchObjectives,\n\t}\n\n\treturn walker\n}", "func NewSnapshotter(snapshotCache cache.SnapshotCache,\n\tclusters *cluster.ClusterAggregator,\n\tingressListeners *listener.IngressListenerAggregator,\n\tegressListeners *listener.EgressListenerAggregator) *Snapshotter {\n\treturn &Snapshotter{\n\t\tsnapshotCache: snapshotCache,\n\t\tclusters: clusters,\n\t\tingressListeners: ingressListeners,\n\t\tegressListeners: egressListeners,\n\t\tevents: make(chan interface{}),\n\t}\n}", "func NewSearchRequest(src string, budget uint64, keywords []string) *SearchRequest {\n\treturn &SearchRequest{\n\t\tOrigin: src,\n\t\tBudget: budget,\n\t\tKeywords: keywords,\n\t}\n}", "func (*buildingR) NewStruct() *buildingR {\n\treturn &buildingR{}\n}", "func New() Tree {\n\treturn &binarySearchTree{}\n}", "func NewSearchAlterationOptions()(*SearchAlterationOptions) {\n m := &SearchAlterationOptions{\n }\n m.SetAdditionalData(make(map[string]interface{}));\n return m\n}", "func newWatchCommand(c *cli.Context) (Command, error) {\n\trepo, err := newRepositoryRequestInfo(c)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &watchCommand{repo: repo, jsonPaths: c.StringSlice(\"jsonpath\"), streaming: c.Bool(\"streaming\")}, nil\n}", "func NewSnapshotter(clientFactory *dockerclient.ClientFactory) *Snapshotter {\n\treturn &Snapshotter{\n\t\tclientFactory: clientFactory,\n\t}\n}", "func newScanner(limit *uint64) *scanner {\n\treturn &scanner{limit: limit}\n}", "func (m *GraphBaseServiceClient) Search()(*i286f3babd79fe9ec3b0f52b6ed5910842c0adaeff02be1843d0e01c56d9ba6d9.SearchRequestBuilder) {\n return i286f3babd79fe9ec3b0f52b6ed5910842c0adaeff02be1843d0e01c56d9ba6d9.NewSearchRequestBuilderInternal(m.pathParameters, m.requestAdapter);\n}", "func (m *GraphBaseServiceClient) Search()(*i286f3babd79fe9ec3b0f52b6ed5910842c0adaeff02be1843d0e01c56d9ba6d9.SearchRequestBuilder) {\n return i286f3babd79fe9ec3b0f52b6ed5910842c0adaeff02be1843d0e01c56d9ba6d9.NewSearchRequestBuilderInternal(m.pathParameters, m.requestAdapter);\n}", "func Snapshot(ctx context.Context, logger log.Logger, db database.DB, query string, monitorID int64, settings *schema.Settings) error {\n\tsearchClient := client.NewSearchClient(logger, db, search.Indexed(), search.SearcherURLs())\n\tinputs, err := searchClient.Plan(ctx, \"V3\", nil, query, search.Streaming, settings, envvar.SourcegraphDotComMode())\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tclients := searchClient.JobClients()\n\tplanJob, err := jobutil.NewPlanJob(inputs, inputs.Plan)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\thook := func(ctx context.Context, db database.DB, gs commit.GitserverClient, args *gitprotocol.SearchRequest, repoID api.RepoID, _ commit.DoSearchFunc) error {\n\t\treturn snapshotHook(ctx, db, gs, args, monitorID, repoID)\n\t}\n\n\tplanJob, err = addCodeMonitorHook(planJob, hook)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// HACK(camdencheek): limit the concurrency of the commit search job\n\t// because the db passed into this function might actually be a transaction\n\t// and transactions cannot be used concurrently.\n\tplanJob = limitConcurrency(planJob)\n\n\t_, err = planJob.Run(ctx, clients, streaming.NewNullStream())\n\treturn err\n}", "func NewSearchResult() *SearchResult {\n\treturn &SearchResult{\n\t\tMatches: make(map[string]map[string]map[string]string),\n\t\tErrors: make(map[string]*errs.StructError),\n\t\tRepos: make(map[string]*Repo),\n\t\tDurations: SearchDurations{},\n\t}\n}", "func newSTIBuilder(client DockerClient, dockerSocket string, build *api.Build,\n\tbuilderFactory stiBuilderFactory, configValidator stiConfigValidator) *STIBuilder {\n\t// just create instance\n\treturn &STIBuilder{\n\t\tdockerClient: client,\n\t\tdockerSocket: dockerSocket,\n\t\tbuild: build,\n\t\tbuilderFactory: builderFactory,\n\t\tconfigValidator: configValidator,\n\t}\n}", "func newStorage() *storage {\n\tr := make(map[string][]byte)\n\treturn &storage{\n\t\trepository: r,\n\t}\n}", "func NewSnapshot(ctx *pulumi.Context,\n\tname string, args *SnapshotArgs, opts ...pulumi.ResourceOption) (*Snapshot, error) {\n\tif args == nil || args.SourceDisk == nil {\n\t\treturn nil, errors.New(\"missing required argument 'SourceDisk'\")\n\t}\n\tif args == nil {\n\t\targs = &SnapshotArgs{}\n\t}\n\tvar resource Snapshot\n\terr := ctx.RegisterResource(\"gcp:compute/snapshot:Snapshot\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func New(name, key string) (*GetMatchHistory, *SummonerName, error) {\n\n\tif len(key) == 0 || len(key) == 0 {\n\t\treturn nil, nil, fmt.Errorf(\"One of your variables is empty, \\nname:%v \\nkey:%v \", name, key)\n\t}\n\trespBody, err := GetInfo(nameURL, name, key)\n\tif err != nil {\n\t\treturn nil, nil, fmt.Errorf(\"Somethings gone wrong: %v\", err)\n\t}\n\tnewStruct, err := MaKeSummonerName(respBody)\n\n\tif err != nil {\n\t\treturn nil, nil, fmt.Errorf(\"Couldn't build struct after polling api %v\", err)\n\t}\n\treturn &GetMatchHistory{name: name, key: key, encryptedID: newStruct.ID}, &newStruct, nil\n}", "func NewVersionsRequestBuilder(rawUrl string, requestAdapter i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.RequestAdapter)(*VersionsRequestBuilder) {\n urlParams := make(map[string]string)\n urlParams[\"request-raw-url\"] = rawUrl\n return NewVersionsRequestBuilderInternal(urlParams, requestAdapter)\n}", "func NewVersionsRequestBuilderInternal(pathParameters map[string]string, requestAdapter i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.RequestAdapter)(*VersionsRequestBuilder) {\n m := &VersionsRequestBuilder{\n }\n m.urlTemplate = \"{+baseurl}/users/{user%2Did}/drives/{drive%2Did}/list/items/{listItem%2Did}/versions{?%24top,%24skip,%24search,%24filter,%24count,%24orderby,%24select,%24expand}\";\n urlTplParams := make(map[string]string)\n for idx, item := range pathParameters {\n urlTplParams[idx] = item\n }\n m.pathParameters = urlTplParams;\n m.requestAdapter = requestAdapter;\n return m\n}", "func (m *AccessReviewHistoryDefinitionItemRequestBuilder) Instances()(*i6b98e970eebf9cd01ff666e44d13e3bb10488d6747910f84a6d2c2578cefa739.InstancesRequestBuilder) {\n return i6b98e970eebf9cd01ff666e44d13e3bb10488d6747910f84a6d2c2578cefa739.NewInstancesRequestBuilderInternal(m.pathParameters, m.requestAdapter);\n}", "func (c *kubernetesBindingsController) SnapshotsFrom(bindingNames ...string) map[string][]ObjectAndFilterResult {\n\tres := map[string][]ObjectAndFilterResult{}\n\n\tfor _, bindingName := range bindingNames {\n\t\t// Initialize all keys with empty arrays.\n\t\tres[bindingName] = make([]ObjectAndFilterResult, 0)\n\n\t\tsnapshot := c.SnapshotsFor(bindingName)\n\t\tif snapshot != nil {\n\t\t\tres[bindingName] = snapshot\n\t\t}\n\t}\n\n\treturn res\n}", "func NewViewRequestBuilder() viewRequestBuilder {\n\treturn viewRequestBuilder{\n\t\toperations: make(map[model.Fingerprint]ops),\n\t}\n}", "func NewRegistry(s Storage) WatchingRegistry {\n\treturn &storage{s}\n}", "func NewRegistry(s Storage) WatchingRegistry {\n\treturn &storage{s}\n}", "func NewGetSnapshotRequest(server string, id string) (*http.Request, error) {\n\tvar err error\n\n\tvar pathParam0 string\n\n\tpathParam0, err = runtime.StyleParam(\"simple\", false, \"id\", id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tqueryUrl, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbasePath := fmt.Sprintf(\"/snapshot/%s\", pathParam0)\n\tif basePath[0] == '/' {\n\t\tbasePath = basePath[1:]\n\t}\n\n\tqueryUrl, err = queryUrl.Parse(basePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq, err := http.NewRequest(\"GET\", queryUrl.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn req, nil\n}", "func (*filesStorageR) NewStruct() *filesStorageR {\n\treturn &filesStorageR{}\n}", "func New(b []byte) (*Finding, error) {\n\tvar f Finding\n\tif err := json.Unmarshal(b, &f.Containerscanner); err != nil {\n\t\treturn nil, err\n\t}\n\treturn &f, nil\n}", "func New(lruSize int) *DriverStorage {\n\ts := new(DriverStorage)\n\ts.drivers = make(map[int]*Driver)\n\ts.locations = rtreego.NewTree(2, 25, 50)\n\ts.mu = new(sync.RWMutex)\n\ts.lruSize = lruSize\n\treturn s\n}", "func New() *BinarySearchTree {\n\treturn &BinarySearchTree{root: nil}\n}", "func New(options ...Option) Gocache {\n\tg := newDefaultGocache()\n\n\tfor _, opt := range options {\n\t\topt(g)\n\t}\n\n\tfor i := 0; i < int(g.ShardsCount); i++ {\n\t\tg.shards[i] = newDefaultShard()\n\t}\n\n\treturn g\n}", "func NewBinarySearch[T constraints.Ordered]() *BinarySearch[T] {\n\treturn &BinarySearch[T]{\n\t\tRoot: nil,\n\t\t_NIL: nil,\n\t}\n}", "func NewSnapshot(ctx *pulumi.Context,\n\tname string, args *SnapshotArgs, opts ...pulumi.ResourceOption) (*Snapshot, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.DiskId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'DiskId'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Snapshot\n\terr := ctx.RegisterResource(\"alicloud:ecs/snapshot:Snapshot\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func NewSearchApi() *SearchApi {\n\tfmt.Sprintf(strings.Title(\"\"), \"\")\n\tconfig := GetDefaultConfiguration()\n\treturn &SearchApi{\n\t\tConfiguration: config,\n\t}\n}" ]
[ "0.6317776", "0.5406593", "0.533186", "0.51976943", "0.5071362", "0.49973714", "0.49941468", "0.4989026", "0.49606875", "0.4947597", "0.48980603", "0.48810616", "0.4875463", "0.48387536", "0.48254812", "0.48051634", "0.4798278", "0.47925445", "0.47883177", "0.47691983", "0.47319025", "0.47147787", "0.46966177", "0.46798125", "0.46669382", "0.4664848", "0.4641311", "0.46270984", "0.46102977", "0.46054685", "0.46000788", "0.45828968", "0.4580479", "0.45739263", "0.45670322", "0.45498693", "0.4544284", "0.4540242", "0.45190844", "0.45097157", "0.4469394", "0.44464743", "0.44314152", "0.441955", "0.4417618", "0.44143406", "0.43995634", "0.43976468", "0.4394686", "0.43944177", "0.43883127", "0.438463", "0.43682167", "0.4364961", "0.43593502", "0.43439195", "0.43380335", "0.43293175", "0.4326291", "0.43249992", "0.43187037", "0.43169948", "0.43131515", "0.43118674", "0.43053955", "0.43021122", "0.42951873", "0.42939505", "0.42900535", "0.42811093", "0.4275028", "0.42740184", "0.4264327", "0.4243506", "0.42409095", "0.42400306", "0.42345008", "0.42345008", "0.42344815", "0.42166534", "0.42162764", "0.42093334", "0.42055166", "0.4202526", "0.41992557", "0.4197917", "0.41971973", "0.41940588", "0.4191341", "0.4191054", "0.4191054", "0.41902357", "0.4184074", "0.41777006", "0.41725388", "0.41708237", "0.41701108", "0.41693664", "0.41680443", "0.41666928" ]
0.7982566
0
Build finalize the chain and returns the SearchableSnapshots struct
func (rb *SearchableSnapshotsBuilder) Build() SearchableSnapshots { return *rb.v }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewSearchableSnapshotsBuilder() *SearchableSnapshotsBuilder {\n\tr := SearchableSnapshotsBuilder{\n\t\t&SearchableSnapshots{},\n\t}\n\n\treturn &r\n}", "func (rb *SnapshotShardsStatusBuilder) Build() SnapshotShardsStatus {\n\treturn *rb.v\n}", "func (s *SnapshotsServiceOp) list(opt *ListOptions, listOpt *listSnapshotOptions) ([]Snapshot, *Response, error) {\n\tpath := snapshotBasePath\n\tpath, err := addOptions(path, opt)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tpath, err = addOptions(path, listOpt)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treq, err := s.client.NewRequest(\"GET\", path, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\troot := new(snapshotsRoot)\n\tresp, err := s.client.Do(req, root)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\tif l := root.Links; l != nil {\n\t\tresp.Links = l\n\t}\n\n\treturn root.Snapshots, resp, err\n}", "func (t liveUpdateStateTree) createResultSet() store.BuildResultSet {\n\tiTargetID := t.iTarget.ID()\n\tstate := t.iTargetState\n\tres := state.LastSuccessfulResult\n\n\tliveUpdatedContainerIDs := []container.ID{}\n\tfor _, c := range state.RunningContainers {\n\t\tliveUpdatedContainerIDs = append(liveUpdatedContainerIDs, c.ContainerID)\n\t}\n\n\tresultSet := store.BuildResultSet{}\n\tresultSet[iTargetID] = store.NewLiveUpdateBuildResult(res.TargetID(), liveUpdatedContainerIDs)\n\n\t// Invalidate all the image builds for images we depend on.\n\t// Otherwise, the image builder will think the existing image ID\n\t// is valid and won't try to rebuild it.\n\tfor _, id := range t.hasFileChangesIDs {\n\t\tif id != iTargetID {\n\t\t\tresultSet[id] = nil\n\t\t}\n\t}\n\n\treturn resultSet\n}", "func BuildSnapshot(tokenAddress string, provider string, block int64) {\n\n\t// Create an IPC based RPC connection to a remote node\n\tconn, err := ethclient.Dial(provider)\n\tif err != nil {\n\t\tlog.Fatalf(\"Failed to connect to the Ethereum client: %v\", err)\n\t}\n\n\t// Instantiate the contract to query contract balance\n\ttoken, err := NewERC20Token(common.HexToAddress(tokenAddress), conn)\n\tif err != nil {\n\t\tlog.Fatalf(\"Failed to instantiate a Token contract: %v\", err)\n\t}\n\n\t// Returns all wallets located in the token according to Etherscan\n\tarrayOfWallets := GetTokenWallets(tokenAddress, block)\n\n\t// Set global value to amount of addresses\n\tnumAddress = len(arrayOfWallets)\n\n\t// Compare the Geth values to our Etherscan results\n\tCheckGethValues(&arrayOfWallets, block, token)\n\n\t// Pulls the token name from the contract\n\ttokenName, err := token.Name(nil)\n\tif err != nil {\n\t\tlog.Printf(\"Failed to retrieve token name: %v\", err)\n\t}\n\n\t// Write the information to local csv file\n\tWriteToCsv(arrayOfWallets, tokenName, strconv.FormatInt(block, 10))\n}", "func (s *SnapshotsServiceOp) list(ctx context.Context, opt *ListOptions, listOpt *listSnapshotOptions) ([]Snapshot, *Response, error) {\n\tpath := snapshotBasePath\n\tpath, err := addOptions(path, opt)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tpath, err = addOptions(path, listOpt)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treq, err := s.client.NewRequest(ctx, http.MethodGet, path, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\troot := new(snapshotsRoot)\n\tresp, err := s.client.Do(ctx, req, root)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\tif l := root.Links; l != nil {\n\t\tresp.Links = l\n\t}\n\n\treturn root.Snapshots, resp, err\n}", "func getSyncVisitChainIter(db *mgo.Database, sqlDb *sql.DB) (*mgo.Iter, error) {\n\tsourceMatch := bson.M{\n\t\t\"event\": \"visit\",\n\t}\n\n\t// optionally add date-range filtering on `date`\n\tdateRange, err := getBeforeAfterFilter()\n\tif err != nil {\n\t\treturn nil, err\n\t} else if len(dateRange) > 0 {\n\t\tsourceMatch[\"date\"] = dateRange\n\t}\n\n\t// Build a projection map for just the fields we need for deserialization of our record types\n\tsourceProject, err := BuildProjection(reflect.TypeOf(syncVisitChainInputRecord{}))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Build a big honking aggregation pipeline to include blob lookups for DOM/screenshot\n\tbigHonkingQuery := []bson.M{\n\t\t{\"$match\": sourceMatch},\n\t\t{\"$group\": bson.M{\n\t\t\t\"_id\": \"$page\",\n\t\t\t\"last_when\": bson.M{\"$max\": \"$date\"},\n\t\t\t\"links\": bson.M{\"$sum\": 1},\n\t\t\t\"urls\": bson.M{\"$push\": \"$url\"},\n\t\t}},\n\t\t{\"$match\": bson.M{\"links\": bson.M{\"$gt\": 1}}},\n\t\t{\"$project\": sourceProject},\n\t}\n\tif rawSkip, ok := os.LookupEnv(\"SKIP\"); ok {\n\t\tif skip, err := strconv.Atoi(rawSkip); err == nil {\n\t\t\tbigHonkingQuery = append(bigHonkingQuery, bson.M{\"$skip\": skip})\n\t\t} else {\n\t\t\tlog.Printf(\"getSyncVisitChainIter: WARNING, malformed 'SKIP' ENV var '%s' (%v)\\n\", rawSkip, err)\n\t\t}\n\t}\n\tif rawLimit, ok := os.LookupEnv(\"LIMIT\"); ok {\n\t\tif limit, err := strconv.Atoi(rawLimit); err == nil {\n\t\t\tbigHonkingQuery = append(bigHonkingQuery, bson.M{\"$limit\": limit})\n\t\t} else {\n\t\t\tlog.Printf(\"getSyncVisitChainIter: WARNING, malformed 'LIMIT' ENV var '%s' (%v)\\n\", rawLimit, err)\n\t\t}\n\t}\n\treturn db.C(\"events\").Pipe(bigHonkingQuery).AllowDiskUse().Iter(), nil\n}", "func (c *testChain) buildChain() {\n\tpending := make(map[string]*testchainNode, len(c.Nodes))\n\tfor k, v := range c.Nodes {\n\t\tpending[k] = v\n\t}\n\n\t// AUMs with a parent need to know their hash, so we\n\t// only compute AUMs who's parents have been computed\n\t// each iteration. Since at least the genesis AUM\n\t// had no parent, theres always a path to completion\n\t// in O(n+1) where n is the number of AUMs.\n\tc.AUMs = make(map[string]AUM, len(c.Nodes))\n\tc.AUMHashes = make(map[string]AUMHash, len(c.Nodes))\n\tfor i := 0; i < len(c.Nodes)+1; i++ {\n\t\tif len(pending) == 0 {\n\t\t\treturn\n\t\t}\n\n\t\tnext := make([]*testchainNode, 0, 10)\n\t\tfor _, v := range pending {\n\t\t\tif _, parentPending := pending[v.Parent]; !parentPending {\n\t\t\t\tnext = append(next, v)\n\t\t\t}\n\t\t}\n\n\t\tfor _, v := range next {\n\t\t\taum := c.makeAUM(v)\n\t\t\th := aum.Hash()\n\n\t\t\tc.AUMHashes[v.Name] = h\n\t\t\tc.AUMs[v.Name] = aum\n\t\t\tdelete(pending, v.Name)\n\t\t}\n\t}\n\tpanic(\"unexpected: incomplete despite len(Nodes)+1 iterations\")\n}", "func Snapshot(ctx context.Context, logger log.Logger, db database.DB, query string, monitorID int64, settings *schema.Settings) error {\n\tsearchClient := client.NewSearchClient(logger, db, search.Indexed(), search.SearcherURLs())\n\tinputs, err := searchClient.Plan(ctx, \"V3\", nil, query, search.Streaming, settings, envvar.SourcegraphDotComMode())\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tclients := searchClient.JobClients()\n\tplanJob, err := jobutil.NewPlanJob(inputs, inputs.Plan)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\thook := func(ctx context.Context, db database.DB, gs commit.GitserverClient, args *gitprotocol.SearchRequest, repoID api.RepoID, _ commit.DoSearchFunc) error {\n\t\treturn snapshotHook(ctx, db, gs, args, monitorID, repoID)\n\t}\n\n\tplanJob, err = addCodeMonitorHook(planJob, hook)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// HACK(camdencheek): limit the concurrency of the commit search job\n\t// because the db passed into this function might actually be a transaction\n\t// and transactions cannot be used concurrently.\n\tplanJob = limitConcurrency(planJob)\n\n\t_, err = planJob.Run(ctx, clients, streaming.NewNullStream())\n\treturn err\n}", "func (c *CacheManager) BuildAvailables() {\n\t/* Iterate over videoDir (1st level only) and extract filenames */\n\tdir, err := os.Open(c.videoDir)\n\tif err != nil { return }\n\tdefer dir.Close()\n\tfileInfos, err := dir.Readdir(-1)\n\tif err != nil { return }\n\tfor _, fi := range fileInfos {\n\t\tc.availables = append(c.availables, Available{\n\t\t\tProto : \"file\",\n\t\t\tName : utils.RemoveExtension(fi.Name()),\n\t\t\tPath : filepath.Join(c.videoDir, fi.Name()),\n\t\t})\n\t}\n}", "func (ob *OrderBookGroup) collectSnapshots() {\n\tgo func() {\n\t\tfor {\n\t\t\ttime.Sleep(snapshotInterval)\n\n\t\t\tdata, err := ob.Get()\n\t\t\tif err != nil {\n\t\t\t\tlog.Println(\"[POLONIEX] Error loading orderbook snapshot: \", err)\n\t\t\t}\n\t\t\tfor _, book := range data {\n\t\t\t\tif len(book.Buy) > 0 || len(book.Sell) > 0 {\n\t\t\t\t\tob.publish(book, \"s\", nil)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n}", "func QuerySnapshotsUtil(ctx context.Context, m cnsvolume.Manager, snapshotQueryFilter cnstypes.CnsSnapshotQueryFilter,\n\tmaxEntries int64) ([]cnstypes.CnsSnapshotQueryResultEntry, string, error) {\n\tlog := logger.GetLogger(ctx)\n\tvar allQuerySnapshotResults []cnstypes.CnsSnapshotQueryResultEntry\n\tvar snapshotQuerySpec cnstypes.CnsSnapshotQuerySpec\n\tvar batchSize int64\n\tmaxIteration := int64(1)\n\tisMaxIterationSet := false\n\tif snapshotQueryFilter.SnapshotQuerySpecs == nil {\n\t\tlog.Infof(\"Attempting to retrieve all the Snapshots available in the vCenter inventory.\")\n\t} else {\n\t\tsnapshotQuerySpec = snapshotQueryFilter.SnapshotQuerySpecs[0]\n\t\tlog.Infof(\"Invoking QuerySnapshots with spec: %+v\", snapshotQuerySpec)\n\t}\n\t// Check if cursor is specified, if not set a default cursor.\n\tif snapshotQueryFilter.Cursor == nil {\n\t\t// Setting the default limit(128) explicitly.\n\t\tsnapshotQueryFilter = cnstypes.CnsSnapshotQueryFilter{\n\t\t\tCursor: &cnstypes.CnsCursor{\n\t\t\t\tOffset: 0,\n\t\t\t\tLimit: DefaultQuerySnapshotLimit,\n\t\t\t},\n\t\t}\n\t\tbatchSize = DefaultQuerySnapshotLimit\n\t} else {\n\t\tbatchSize = snapshotQueryFilter.Cursor.Limit\n\t}\n\titeration := int64(1)\n\tfor {\n\t\tif iteration > maxIteration {\n\t\t\t// Exceeds the max number of results that can be handled by callers.\n\t\t\tnextToken := strconv.FormatInt(snapshotQueryFilter.Cursor.Offset, 10)\n\t\t\tlog.Infof(\"the number of results: %d approached max-entries: %d for \"+\n\t\t\t\t\"limit: %d in iteration: %d, returning with next-token: %s\",\n\t\t\t\tlen(allQuerySnapshotResults), maxEntries, batchSize, iteration, nextToken)\n\t\t\treturn allQuerySnapshotResults, nextToken, nil\n\t\t}\n\t\tlog.Infof(\"invoking QuerySnapshots in iteration: %d with offset: %d and limit: %d, current total \"+\n\t\t\t\"results: %d\", iteration, snapshotQueryFilter.Cursor.Offset, snapshotQueryFilter.Cursor.Limit,\n\t\t\tlen(allQuerySnapshotResults))\n\t\tsnapshotQueryResult, err := m.QuerySnapshots(ctx, snapshotQueryFilter)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"querySnapshots failed for snapshotQueryFilter: %v. Err=%+v\", snapshotQueryFilter, err)\n\t\t\treturn nil, \"\", err\n\t\t}\n\t\tif snapshotQueryResult == nil {\n\t\t\tlog.Infof(\"Observed empty SnapshotQueryResult\")\n\t\t\tbreak\n\t\t}\n\t\tif len(snapshotQueryResult.Entries) == 0 {\n\t\t\tlog.Infof(\"QuerySnapshots retrieved no results for the spec: %+v\", snapshotQuerySpec)\n\t\t}\n\t\t// Update the max iteration.\n\t\t// isMaxIterationSet ensures that the max iterations are set only once, this is to ensure that the number of\n\t\t// results are lower than the max entries supported by caller in a busy system which has increasing number\n\t\t// total records.\n\t\tif !isMaxIterationSet {\n\t\t\tif snapshotQueryResult.Cursor.TotalRecords < maxEntries {\n\t\t\t\t// If the total number of records is less than max entries supported by caller then\n\t\t\t\t// all results can be retrieved in a loop, when the results are returned no next-token is expected to be set.\n\t\t\t\t// Example:\n\t\t\t\t// maxEntries=200, totalRecords=150, batchSize=128\n\t\t\t\t// maxIteration=2\n\t\t\t\t// iteration-1: 128 results, iteration-2: 22 results\n\t\t\t\t// total results returned: 150\n\t\t\t\t// offset=0\n\t\t\t\tmaxRecords := snapshotQueryResult.Cursor.TotalRecords\n\t\t\t\tnumOfIterationsForAllResults := float64(maxRecords) / float64(batchSize)\n\t\t\t\tmaxIteration = int64(math.Ceil(numOfIterationsForAllResults))\n\t\t\t\tlog.Infof(\"setting max iteration to %d for total records count: %d\", maxIteration, maxRecords)\n\t\t\t} else {\n\t\t\t\t// All results cannot be returned to caller, in this case the expectation is return as many results with a\n\t\t\t\t// nextToken.\n\t\t\t\t// Example:\n\t\t\t\t// maxEntries=150, totalRecords=200, batchSize=128\n\t\t\t\t// maxIteration=1\n\t\t\t\t// iteration-1: 128 results\n\t\t\t\t// total results returned: 128\n\t\t\t\t// offset= 1, callers are expected to call with new offset as next token.\n\t\t\t\tmaxRecords := maxEntries\n\t\t\t\tnumOfIterationsForAllResults := float64(maxRecords) / float64(batchSize)\n\t\t\t\tmaxIteration = int64(math.Floor(numOfIterationsForAllResults))\n\t\t\t\tlog.Infof(\"setting max iteration to %d for total records count: %d and max limit: %d\",\n\t\t\t\t\tmaxIteration, snapshotQueryResult.Cursor.TotalRecords, maxRecords)\n\t\t\t}\n\t\t\tisMaxIterationSet = true\n\t\t}\n\n\t\tallQuerySnapshotResults = append(allQuerySnapshotResults, snapshotQueryResult.Entries...)\n\t\tlog.Infof(\"%d more snapshots to be queried\",\n\t\t\tsnapshotQueryResult.Cursor.TotalRecords-snapshotQueryResult.Cursor.Offset)\n\t\tif snapshotQueryResult.Cursor.Offset == snapshotQueryResult.Cursor.TotalRecords {\n\t\t\tlog.Infof(\"QuerySnapshots retrieved all records (%d) for the SnapshotQuerySpec: %+v in %d iterations\",\n\t\t\t\tsnapshotQueryResult.Cursor.TotalRecords, snapshotQuerySpec, iteration)\n\t\t\tbreak\n\t\t}\n\t\titeration++\n\t\tsnapshotQueryFilter.Cursor = &snapshotQueryResult.Cursor\n\t}\n\treturn allQuerySnapshotResults, \"\", nil\n}", "func (rb *SearchIdleBuilder) Build() SearchIdle {\n\treturn *rb.v\n}", "func (o *GetMoveHistoryURL) Build() (*url.URL, error) {\n\tvar _result url.URL\n\n\tvar _path = \"/move/{locator}/history\"\n\n\tlocator := o.Locator\n\tif locator != \"\" {\n\t\t_path = strings.Replace(_path, \"{locator}\", locator, -1)\n\t} else {\n\t\treturn nil, errors.New(\"locator is required on GetMoveHistoryURL\")\n\t}\n\n\t_basePath := o._basePath\n\tif _basePath == \"\" {\n\t\t_basePath = \"/ghc/v1\"\n\t}\n\t_result.Path = golangswaggerpaths.Join(_basePath, _path)\n\n\tqs := make(url.Values)\n\n\tvar pageQ string\n\tif o.Page != nil {\n\t\tpageQ = swag.FormatInt64(*o.Page)\n\t}\n\tif pageQ != \"\" {\n\t\tqs.Set(\"page\", pageQ)\n\t}\n\n\tvar perPageQ string\n\tif o.PerPage != nil {\n\t\tperPageQ = swag.FormatInt64(*o.PerPage)\n\t}\n\tif perPageQ != \"\" {\n\t\tqs.Set(\"perPage\", perPageQ)\n\t}\n\n\t_result.RawQuery = qs.Encode()\n\n\treturn &_result, nil\n}", "func NewSnapshotsFromProto(s []*auctioneerrpc.BatchSnapshotResponse) []*Snapshot {\n\tresult := make([]*Snapshot, len(s))\n\tfor idx, snapshot := range s {\n\t\tresult[idx] = &Snapshot{\n\t\t\tVersion: snapshot.Version,\n\t\t\tBatchID: hex.EncodeToString(\n\t\t\t\tsnapshot.BatchId,\n\t\t\t),\n\t\t\tPrevBatchID: hex.EncodeToString(\n\t\t\t\tsnapshot.PrevBatchId,\n\t\t\t),\n\t\t\tBatchTxID: snapshot.BatchTxId,\n\t\t\tBatchTx: hex.EncodeToString(\n\t\t\t\tsnapshot.BatchTx,\n\t\t\t),\n\t\t\tBatchTxFeeRateSatPerKw: snapshot.BatchTxFeeRateSatPerKw,\n\t\t\tCreationTimestampNs: snapshot.CreationTimestampNs,\n\t\t\tMatchedMarkets: snapshot.MatchedMarkets,\n\t\t}\n\t}\n\n\treturn result\n}", "func (tg *TradesGroup) collectSnapshots() {\n\tgo func() {\n\t\tfor {\n\t\t\ttime.Sleep(snapshotInterval)\n\n\t\t\tdata, err := tg.Get()\n\t\t\tif err != nil {\n\t\t\t\ttg.publish(nil, \"s\", err)\n\t\t\t}\n\n\t\t\tfor _, smb := range tg.symbols {\n\t\t\t\tvar symbolTrades []schemas.Trade\n\t\t\t\tfor _, trd := range data {\n\t\t\t\t\tif trd.Symbol == smb.Name {\n\t\t\t\t\t\tsymbolTrades = append(symbolTrades, trd)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tif len(symbolTrades) > 0 {\n\t\t\t\t\ttg.publish(symbolTrades, \"s\", nil)\n\t\t\t\t}\n\t\t\t}\n\n\t\t}\n\t}()\n}", "func (c *cachestub) Build(records []cache.Record) error {\n\treturn nil\n}", "func getRelatedBuilds(c context.Context, now *timestamppb.Timestamp, client buildbucketpb.BuildsClient, b *buildbucketpb.Build) ([]*ui.Build, error) {\n\tvar bs []string\n\tfor _, buildset := range protoutil.BuildSets(b) {\n\t\t// HACK(hinoka): Remove the commit/git/ buildsets because we know they're redundant\n\t\t// with the commit/gitiles/ buildsets, and we don't need to ask Buildbucket twice.\n\t\tif strings.HasPrefix(buildset, \"commit/git/\") {\n\t\t\tcontinue\n\t\t}\n\t\tbs = append(bs, buildset)\n\t}\n\tif len(bs) == 0 {\n\t\t// No buildset? No builds.\n\t\treturn nil, nil\n\t}\n\n\t// Do the search request.\n\t// Use multiple requests instead of a single batch request.\n\t// A single large request is CPU bound to a single GAE instance on the buildbucket side.\n\t// Multiple requests allows the use of multiple GAE instances, therefore more parallelism.\n\tresps := make([]*buildbucketpb.SearchBuildsResponse, len(bs))\n\tif err := parallel.WorkPool(8, func(ch chan<- func() error) {\n\t\tfor i, buildset := range bs {\n\t\t\ti := i\n\t\t\tbuildset := buildset\n\t\t\tch <- func() (err error) {\n\t\t\t\tlogging.Debugf(c, \"Searching for %s (%d)\", buildset, i)\n\t\t\t\tresps[i], err = client.SearchBuilds(c, searchBuildset(buildset, summaryBuildsMask))\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Dedupe builds.\n\t// It's possible since we've made multiple requests that we got back the same builds\n\t// multiple times.\n\tseen := map[int64]bool{} // set of build IDs.\n\tresult := []*ui.Build{}\n\tfor _, resp := range resps {\n\t\tfor _, rb := range resp.GetBuilds() {\n\t\t\tif seen[rb.Id] {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tseen[rb.Id] = true\n\t\t\tresult = append(result, &ui.Build{\n\t\t\t\tBuild: rb,\n\t\t\t\tNow: now,\n\t\t\t})\n\t\t}\n\t}\n\n\t// Sort builds by ID.\n\tsort.Slice(result, func(i, j int) bool { return result[i].Id < result[j].Id })\n\n\treturn result, nil\n}", "func findNewArtifacts(ctx context.Context, invID invocations.ID, arts []*artifactCreationRequest) ([]*artifactCreationRequest, error) {\n\t// artifacts are not expected to exist in most cases, and this map would likely\n\t// be empty.\n\ttype state struct {\n\t\thash string\n\t\tsize int64\n\t}\n\tvar states map[string]state\n\tks := spanner.KeySets()\n\tfor _, a := range arts {\n\t\tks = spanner.KeySets(invID.Key(a.parentID(), a.artifactID), ks)\n\t}\n\tvar b spanutil.Buffer\n\terr := span.Read(ctx, \"Artifacts\", ks, []string{\"ParentId\", \"ArtifactId\", \"RBECASHash\", \"Size\"}).Do(\n\t\tfunc(row *spanner.Row) (err error) {\n\t\t\tvar pid, aid string\n\t\t\tvar hash string\n\t\t\tvar size int64\n\t\t\tif err = b.FromSpanner(row, &pid, &aid, &hash, &size); err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif states == nil {\n\t\t\t\tstates = make(map[string]state)\n\t\t\t}\n\t\t\t// The artifact exists.\n\t\t\tstates[invID.Key(pid, aid).String()] = state{hash, size}\n\t\t\treturn\n\t\t},\n\t)\n\tif err != nil {\n\t\treturn nil, appstatus.Errorf(codes.Internal, \"%s\", err)\n\t}\n\n\tnewArts := make([]*artifactCreationRequest, 0, len(arts)-len(states))\n\tfor _, a := range arts {\n\t\tst, ok := states[invID.Key(a.parentID(), a.artifactID).String()]\n\t\tif ok && a.size != st.size {\n\t\t\treturn nil, appstatus.Errorf(codes.AlreadyExists, `%q: exists w/ different size: %d != %d`, a.name(invID), a.size, st.size)\n\t\t}\n\n\t\t// Save the hash, so that it can be reused in the post-verification\n\t\t// after rbecase.UpdateBlob().\n\t\tif a.hash == \"\" {\n\t\t\th := sha256.Sum256(a.data)\n\t\t\ta.hash = artifacts.AddHashPrefix(hex.EncodeToString(h[:]))\n\t\t}\n\n\t\tswitch {\n\t\tcase !ok:\n\t\t\tnewArts = append(newArts, a)\n\t\tcase a.hash != st.hash:\n\t\t\treturn nil, appstatus.Errorf(codes.AlreadyExists, `%q: exists w/ different hash`, a.name(invID))\n\t\tdefault:\n\t\t\t// artifact exists\n\t\t}\n\t}\n\treturn newArts, nil\n}", "func (m *Nitro) NewSnapshot() (*Snapshot, error) {\n\tbuf := m.snapshots.MakeBuf()\n\tdefer m.snapshots.FreeBuf(buf)\n\n\t// Stitch all local gclists from all writers to create snapshot gclist\n\tvar head, tail *skiplist.Node\n\n\tfor w := m.wlist; w != nil; w = w.next {\n\t\tif tail == nil {\n\t\t\thead = w.gchead\n\t\t\ttail = w.gctail\n\t\t} else if w.gchead != nil {\n\t\t\ttail.SetLink(w.gchead)\n\t\t\ttail = w.gctail\n\t\t}\n\n\t\tw.gchead = nil\n\t\tw.gctail = nil\n\n\t\t// Update global stats\n\t\tm.store.Stats.Merge(&w.slSts1)\n\t\tatomic.AddInt64(&m.itemsCount, w.count)\n\t\tw.count = 0\n\t}\n\n\tsnap := &Snapshot{db: m, sn: m.GetCurrSn(), refCount: 1, count: m.ItemsCount()}\n\tm.snapshots.Insert(unsafe.Pointer(snap), CompareSnapshot, buf, &m.snapshots.Stats)\n\tsnap.gclist = head\n\tnewSn := atomic.AddUint32(&m.currSn, 1)\n\tif newSn == math.MaxUint32 {\n\t\treturn nil, ErrMaxSnapshotsLimitReached\n\t}\n\n\treturn snap, nil\n}", "func (b *BuilderMux) Build(ctx context.Context, out io.Writer, tags tag.ImageTags, resolver platform.Resolver, artifacts []*latestV1.Artifact) ([]graph.Artifact, error) {\n\tm := make(map[PipelineBuilder]bool)\n\tfor _, a := range artifacts {\n\t\tm[b.byImageName[a.ImageName]] = true\n\t}\n\n\tfor builder := range m {\n\t\tif err := builder.PreBuild(ctx, out); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tbuilder := func(ctx context.Context, out io.Writer, artifact *latestV1.Artifact, tag string, platforms platform.Matcher) (string, error) {\n\t\tp := b.byImageName[artifact.ImageName]\n\t\tpl, err := filterBuildEnvSupportedPlatforms(p.SupportedPlatforms(), platforms)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tplatforms = pl\n\n\t\tartifactBuilder := p.Build(ctx, out, artifact)\n\t\thooksOpts, err := hooks.NewBuildEnvOpts(artifact, tag, p.PushImages())\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tr := hooks.BuildRunner(artifact.LifecycleHooks, hooksOpts)\n\t\tvar built string\n\t\tif err = r.RunPreHooks(ctx, out); err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tif built, err = artifactBuilder(ctx, out, artifact, tag, platforms); err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tif err = r.RunPostHooks(ctx, out); err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\treturn built, nil\n\t}\n\tar, err := InOrder(ctx, out, tags, resolver, artifacts, builder, b.concurrency, b.store)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfor builder := range m {\n\t\tif err := builder.PostBuild(ctx, out); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn ar, nil\n}", "func (rb *SettingsSearchBuilder) Build() SettingsSearch {\n\treturn *rb.v\n}", "func (s *API) ListSnapshots(req *ListSnapshotsRequest, opts ...scw.RequestOption) (*ListSnapshotsResponse, error) {\n\tvar err error\n\n\tif req.Region == \"\" {\n\t\tdefaultRegion, _ := s.client.GetDefaultRegion()\n\t\treq.Region = defaultRegion\n\t}\n\n\tdefaultPageSize, exist := s.client.GetDefaultPageSize()\n\tif (req.PageSize == nil || *req.PageSize == 0) && exist {\n\t\treq.PageSize = &defaultPageSize\n\t}\n\n\tquery := url.Values{}\n\tparameter.AddToQuery(query, \"name\", req.Name)\n\tparameter.AddToQuery(query, \"order_by\", req.OrderBy)\n\tparameter.AddToQuery(query, \"instance_id\", req.InstanceID)\n\tparameter.AddToQuery(query, \"organization_id\", req.OrganizationID)\n\tparameter.AddToQuery(query, \"project_id\", req.ProjectID)\n\tparameter.AddToQuery(query, \"page\", req.Page)\n\tparameter.AddToQuery(query, \"page_size\", req.PageSize)\n\n\tif fmt.Sprint(req.Region) == \"\" {\n\t\treturn nil, errors.New(\"field Region cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"GET\",\n\t\tPath: \"/rdb/v1/regions/\" + fmt.Sprint(req.Region) + \"/snapshots\",\n\t\tQuery: query,\n\t\tHeaders: http.Header{},\n\t}\n\n\tvar resp ListSnapshotsResponse\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func (c *restClient) ListSnapshots(ctx context.Context, req *netapppb.ListSnapshotsRequest, opts ...gax.CallOption) *SnapshotIterator {\n\tit := &SnapshotIterator{}\n\treq = proto.Clone(req).(*netapppb.ListSnapshotsRequest)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tit.InternalFetch = func(pageSize int, pageToken string) ([]*netapppb.Snapshot, string, error) {\n\t\tresp := &netapppb.ListSnapshotsResponse{}\n\t\tif pageToken != \"\" {\n\t\t\treq.PageToken = pageToken\n\t\t}\n\t\tif pageSize > math.MaxInt32 {\n\t\t\treq.PageSize = math.MaxInt32\n\t\t} else if pageSize != 0 {\n\t\t\treq.PageSize = int32(pageSize)\n\t\t}\n\t\tbaseUrl, err := url.Parse(c.endpoint)\n\t\tif err != nil {\n\t\t\treturn nil, \"\", err\n\t\t}\n\t\tbaseUrl.Path += fmt.Sprintf(\"/v1/%v/snapshots\", req.GetParent())\n\n\t\tparams := url.Values{}\n\t\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\t\tif req.GetFilter() != \"\" {\n\t\t\tparams.Add(\"filter\", fmt.Sprintf(\"%v\", req.GetFilter()))\n\t\t}\n\t\tif req.GetOrderBy() != \"\" {\n\t\t\tparams.Add(\"orderBy\", fmt.Sprintf(\"%v\", req.GetOrderBy()))\n\t\t}\n\t\tif req.GetPageSize() != 0 {\n\t\t\tparams.Add(\"pageSize\", fmt.Sprintf(\"%v\", req.GetPageSize()))\n\t\t}\n\t\tif req.GetPageToken() != \"\" {\n\t\t\tparams.Add(\"pageToken\", fmt.Sprintf(\"%v\", req.GetPageToken()))\n\t\t}\n\n\t\tbaseUrl.RawQuery = params.Encode()\n\n\t\t// Build HTTP headers from client and context metadata.\n\t\thds := append(c.xGoogHeaders, \"Content-Type\", \"application/json\")\n\t\theaders := gax.BuildHeaders(ctx, hds...)\n\t\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\t\tif settings.Path != \"\" {\n\t\t\t\tbaseUrl.Path = settings.Path\n\t\t\t}\n\t\t\thttpReq, err := http.NewRequest(\"GET\", baseUrl.String(), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\thttpReq.Header = headers\n\n\t\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tdefer httpRsp.Body.Close()\n\n\t\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn nil\n\t\t}, opts...)\n\t\tif e != nil {\n\t\t\treturn nil, \"\", e\n\t\t}\n\t\tit.Response = resp\n\t\treturn resp.GetSnapshots(), resp.GetNextPageToken(), nil\n\t}\n\n\tfetch := func(pageSize int, pageToken string) (string, error) {\n\t\titems, nextPageToken, err := it.InternalFetch(pageSize, pageToken)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tit.items = append(it.items, items...)\n\t\treturn nextPageToken, nil\n\t}\n\n\tit.pageInfo, it.nextFunc = iterator.NewPageInfo(fetch, it.bufLen, it.takeBuf)\n\tit.pageInfo.MaxSize = int(req.GetPageSize())\n\tit.pageInfo.Token = req.GetPageToken()\n\n\treturn it\n}", "func Build(queryFunc queryFunc) PubsubHandlerFunc {\n\t// This is the pubsub handler\n\treturn func(ctx context.Context, m *pubsub.Message) error {\n\t\tdefer LogTiming(time.Now(), \"Collect\")\n\n\t\tbucketName := string(m.Data)\n\t\tif !strings.HasPrefix(bucketName, \"source-\") {\n\t\t\treturn fmt.Errorf(\"message must be a bucket name starting with 'source-': %v\", bucketName)\n\t\t}\n\n\t\t// Create GCP client and get a handle on the bucket\n\t\tclient, err := storage.NewClient(context.Background())\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tbucket := client.Bucket(bucketName)\n\n\t\t// Load source config from the bucket\n\t\tvar config map[string]string\n\t\terr = unmarshalFromBucket(bucket, \"config.json\", &config)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Cannot read config.json: %v\", err)\n\t\t}\n\n\t\t// Load state from previous runs\n\t\tvar checksums map[string]string\n\t\terr = unmarshalFromBucket(bucket, \"state.json\", &checksums)\n\t\tif err != nil {\n\t\t\tif err != storage.ErrObjectNotExist {\n\t\t\t\treturn fmt.Errorf(\"Cannot read state.json: %v\", err)\n\t\t\t}\n\t\t\t// Use a default empty value instead\n\t\t\tchecksums = map[string]string{}\n\t\t}\n\n\t\t// Collect API data\n\t\titems, err := queryFunc(config)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// Write items to the bucket\n\t\tfor id, item := range items {\n\t\t\t// Only write this instance if it has changed\n\t\t\tstart := time.Now()\n\t\t\tpreviousChecksum, present := checksums[id]\n\t\t\tsha := sha1.New()\n\t\t\t_, err = sha.Write(item.Data)\n\t\t\tif err != nil {\n\t\t\t\tlog.Printf(\"failed to compute SHA: %v\", err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tchecksum := hex.EncodeToString(sha.Sum(nil))\n\t\t\tif !present || checksum != previousChecksum {\n\t\t\t\terr = writeToBucket(bucket, id, item.Typ, item.Version, item.Data)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tchecksums[id] = checksum\n\t\t\t\tlog.Printf(\"Change: %s\", id)\n\t\t\t\tlog.Printf(\"Timing: Write: %dms\", uint(time.Since(start).Seconds()*1000)) // Milliseconds not supported in Go 1.11\n\t\t\t} else {\n\t\t\t\tlog.Printf(\"No change found: %s\", id)\n\t\t\t}\n\t\t}\n\t\tif err := pruneBucket(bucket, items); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tchecksums = pruneChecksums(checksums, items)\n\t\t// Write state back to the bucket\n\t\tdata, err := json.Marshal(checksums)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Cannot marshal the value: %v\", err)\n\t\t}\n\t\terr = writeToBucket(bucket, \"state.json\", \"\", \"\", data)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}\n}", "func (s *SnapshotsServiceOp) List(ctx context.Context, opt *ListOptions) ([]Snapshot, *Response, error) {\n\treturn s.list(ctx, opt, nil)\n}", "func makeBuckets(current, prev summaryMap, l *logrus.Entry) []*agentpb.MetricsBucket {\n\tres := make([]*agentpb.MetricsBucket, 0, len(current))\n\n\tfor digest, currentESS := range current {\n\t\tprevESS := prev[digest]\n\t\tif prevESS == nil {\n\t\t\tprevESS = &eventsStatementsSummaryByDigest{}\n\t\t}\n\n\t\tswitch {\n\t\tcase currentESS.CountStar == prevESS.CountStar:\n\t\t\t// Another way how this is possible is if events_statements_summary_by_digest was truncated,\n\t\t\t// and then the same number of queries were made.\n\t\t\t// Currently, we can't differentiate between those situations.\n\t\t\t// TODO We probably could by using first_seen/last_seen columns.\n\t\t\tl.Tracef(\"Skipped due to the same number of queries: %s.\", currentESS)\n\t\t\tcontinue\n\t\tcase currentESS.CountStar < prevESS.CountStar:\n\t\t\tl.Debugf(\"Truncate detected. Treating as a new query: %s.\", currentESS)\n\t\t\tprevESS = &eventsStatementsSummaryByDigest{}\n\t\tcase prevESS.CountStar == 0:\n\t\t\tl.Debugf(\"New query: %s.\", currentESS)\n\t\tdefault:\n\t\t\tl.Debugf(\"Normal query: %s.\", currentESS)\n\t\t}\n\n\t\tcount := inc(currentESS.CountStar, prevESS.CountStar)\n\t\tfingerprint, isTruncated := truncate.Query(*currentESS.DigestText)\n\t\tmb := &agentpb.MetricsBucket{\n\t\t\tCommon: &agentpb.MetricsBucket_Common{\n\t\t\t\tSchema: pointer.GetString(currentESS.SchemaName), // TODO can it be NULL?\n\t\t\t\tQueryid: *currentESS.Digest,\n\t\t\t\tFingerprint: fingerprint,\n\t\t\t\tIsTruncated: isTruncated,\n\t\t\t\tNumQueries: count,\n\t\t\t\tNumQueriesWithErrors: inc(currentESS.SumErrors, prevESS.SumErrors),\n\t\t\t\tNumQueriesWithWarnings: inc(currentESS.SumWarnings, prevESS.SumWarnings),\n\t\t\t\tAgentType: inventorypb.AgentType_QAN_MYSQL_PERFSCHEMA_AGENT,\n\t\t\t},\n\t\t\tMysql: &agentpb.MetricsBucket_MySQL{},\n\t\t}\n\n\t\tfor _, p := range []struct {\n\t\t\tvalue float32 // result value: currentESS.SumXXX-prevESS.SumXXX\n\t\t\tsum *float32 // MetricsBucket.XXXSum field to write value\n\t\t\tcnt *float32 // MetricsBucket.XXXCnt field to write count\n\t\t}{\n\t\t\t// in order of events_statements_summary_by_digest columns\n\n\t\t\t// convert picoseconds to seconds\n\t\t\t{inc(currentESS.SumTimerWait, prevESS.SumTimerWait) / 1000000000000, &mb.Common.MQueryTimeSum, &mb.Common.MQueryTimeCnt},\n\t\t\t{inc(currentESS.SumLockTime, prevESS.SumLockTime) / 1000000000000, &mb.Mysql.MLockTimeSum, &mb.Mysql.MLockTimeCnt},\n\n\t\t\t{inc(currentESS.SumRowsAffected, prevESS.SumRowsAffected), &mb.Mysql.MRowsAffectedSum, &mb.Mysql.MRowsAffectedCnt},\n\t\t\t{inc(currentESS.SumRowsSent, prevESS.SumRowsSent), &mb.Mysql.MRowsSentSum, &mb.Mysql.MRowsSentCnt},\n\t\t\t{inc(currentESS.SumRowsExamined, prevESS.SumRowsExamined), &mb.Mysql.MRowsExaminedSum, &mb.Mysql.MRowsExaminedCnt},\n\n\t\t\t{inc(currentESS.SumCreatedTmpDiskTables, prevESS.SumCreatedTmpDiskTables), &mb.Mysql.MTmpDiskTablesSum, &mb.Mysql.MTmpDiskTablesCnt},\n\t\t\t{inc(currentESS.SumCreatedTmpTables, prevESS.SumCreatedTmpTables), &mb.Mysql.MTmpTablesSum, &mb.Mysql.MTmpTablesCnt},\n\t\t\t{inc(currentESS.SumSelectFullJoin, prevESS.SumSelectFullJoin), &mb.Mysql.MFullJoinSum, &mb.Mysql.MFullJoinCnt},\n\t\t\t{inc(currentESS.SumSelectFullRangeJoin, prevESS.SumSelectFullRangeJoin), &mb.Mysql.MSelectFullRangeJoinSum, &mb.Mysql.MSelectFullRangeJoinCnt},\n\t\t\t{inc(currentESS.SumSelectRange, prevESS.SumSelectRange), &mb.Mysql.MSelectRangeSum, &mb.Mysql.MSelectRangeCnt},\n\t\t\t{inc(currentESS.SumSelectRangeCheck, prevESS.SumSelectRangeCheck), &mb.Mysql.MSelectRangeCheckSum, &mb.Mysql.MSelectRangeCheckCnt},\n\t\t\t{inc(currentESS.SumSelectScan, prevESS.SumSelectScan), &mb.Mysql.MFullScanSum, &mb.Mysql.MFullScanCnt},\n\n\t\t\t{inc(currentESS.SumSortMergePasses, prevESS.SumSortMergePasses), &mb.Mysql.MMergePassesSum, &mb.Mysql.MMergePassesCnt},\n\t\t\t{inc(currentESS.SumSortRange, prevESS.SumSortRange), &mb.Mysql.MSortRangeSum, &mb.Mysql.MSortRangeCnt},\n\t\t\t{inc(currentESS.SumSortRows, prevESS.SumSortRows), &mb.Mysql.MSortRowsSum, &mb.Mysql.MSortRowsCnt},\n\t\t\t{inc(currentESS.SumSortScan, prevESS.SumSortScan), &mb.Mysql.MSortScanSum, &mb.Mysql.MSortScanCnt},\n\n\t\t\t{inc(currentESS.SumNoIndexUsed, prevESS.SumNoIndexUsed), &mb.Mysql.MNoIndexUsedSum, &mb.Mysql.MNoIndexUsedCnt},\n\t\t\t{inc(currentESS.SumNoGoodIndexUsed, prevESS.SumNoGoodIndexUsed), &mb.Mysql.MNoGoodIndexUsedSum, &mb.Mysql.MNoGoodIndexUsedCnt},\n\t\t} {\n\t\t\tif p.value != 0 {\n\t\t\t\t*p.sum = p.value\n\t\t\t\t*p.cnt = count\n\t\t\t}\n\t\t}\n\n\t\tres = append(res, mb)\n\t}\n\n\treturn res\n}", "func CollectReleaseSpecsAndProviderLinks(mStruct *manifest.Manifest, baseDir string, namespace string) (map[string]map[string]manifest.JobSpec, map[string]map[string]manifest.JobLink, error) {\n\t// Contains YAML.load('.../release_name/job_name/job.MF')\n\tjobReleaseSpecs := map[string]map[string]manifest.JobSpec{}\n\n\t// Lists every link provided by the job\n\tjobProviderLinks := map[string]map[string]manifest.JobLink{}\n\n\tfor _, instanceGroup := range mStruct.InstanceGroups {\n\t\tfor jobIdx, job := range instanceGroup.Jobs {\n\t\t\t// make sure a map entry exists for the current job release\n\t\t\tif _, ok := jobReleaseSpecs[job.Release]; !ok {\n\t\t\t\tjobReleaseSpecs[job.Release] = map[string]manifest.JobSpec{}\n\t\t\t}\n\n\t\t\t// load job.MF into jobReleaseSpecs[job.Release][job.Name]\n\t\t\tif _, ok := jobReleaseSpecs[job.Release][job.Name]; !ok {\n\t\t\t\tjobMFFilePath := filepath.Join(baseDir, \"jobs-src\", job.Release, job.Name, \"job.MF\")\n\t\t\t\tjobMfBytes, err := ioutil.ReadFile(jobMFFilePath)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn nil, nil, err\n\t\t\t\t}\n\n\t\t\t\tjobSpec := manifest.JobSpec{}\n\t\t\t\tif err := yaml.Unmarshal([]byte(jobMfBytes), &jobSpec); err != nil {\n\t\t\t\t\treturn nil, nil, err\n\t\t\t\t}\n\t\t\t\tjobReleaseSpecs[job.Release][job.Name] = jobSpec\n\t\t\t}\n\n\t\t\t// spec of the current jobs release/name\n\t\t\tspec := jobReleaseSpecs[job.Release][job.Name]\n\n\t\t\t// Generate instance spec for each ig instance\n\t\t\t// This will be stored inside the current job under\n\t\t\t// job.properties.bosh_containerization\n\t\t\tvar jobsInstances []manifest.JobInstance\n\t\t\tfor i := 0; i < instanceGroup.Instances; i++ {\n\n\t\t\t\t// TODO: Understand whether there are negative side-effects to using this\n\t\t\t\t// default\n\t\t\t\tazs := []string{\"\"}\n\t\t\t\tif len(instanceGroup.Azs) > 0 {\n\t\t\t\t\tazs = instanceGroup.Azs\n\t\t\t\t}\n\n\t\t\t\tfor _, az := range azs {\n\t\t\t\t\tindex := len(jobsInstances)\n\t\t\t\t\tname := fmt.Sprintf(\"%s-%s\", instanceGroup.Name, job.Name)\n\t\t\t\t\tid := fmt.Sprintf(\"%v-%v-%v\", instanceGroup.Name, index, job.Name)\n\t\t\t\t\t// TODO: not allowed to hardcode svc.cluster.local\n\t\t\t\t\taddress := fmt.Sprintf(\"%s.%s.svc.cluster.local\", id, namespace)\n\n\t\t\t\t\tjobsInstances = append(jobsInstances, manifest.JobInstance{\n\t\t\t\t\t\tAddress: address,\n\t\t\t\t\t\tAZ: az,\n\t\t\t\t\t\tID: id,\n\t\t\t\t\t\tIndex: index,\n\t\t\t\t\t\tInstance: i,\n\t\t\t\t\t\tName: name,\n\t\t\t\t\t})\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// set jobs.properties.bosh_containerization.instances with the ig instances\n\t\t\tinstanceGroup.Jobs[jobIdx].Properties.BOSHContainerization.Instances = jobsInstances\n\n\t\t\t// Create a list of fully evaluated links provided by the current job\n\t\t\t// These is specified in the job release job.MF file\n\t\t\tif spec.Provides != nil {\n\t\t\t\tvar properties map[string]interface{}\n\n\t\t\t\tfor _, provider := range spec.Provides {\n\t\t\t\t\tproperties = map[string]interface{}{}\n\t\t\t\t\tfor _, property := range provider.Properties {\n\t\t\t\t\t\t// generate a nested struct of map[string]interface{} when\n\t\t\t\t\t\t// a property is of the form foo.bar\n\t\t\t\t\t\tif strings.Contains(property, \".\") {\n\t\t\t\t\t\t\tpropertyStruct := RetrieveNestedProperty(spec, property)\n\t\t\t\t\t\t\tproperties = propertyStruct\n\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\tproperties[property] = RetrievePropertyDefault(spec, property)\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t\t// Override default spec values with explicit settings from the\n\t\t\t\t\t// current bosh deployment manifest, this should be done under each\n\t\t\t\t\t// job, inside a `properties` key.\n\t\t\t\t\tfor propertyName := range properties {\n\t\t\t\t\t\tif explicitSetting, ok := LookUpProperty(job, propertyName); ok {\n\t\t\t\t\t\t\tproperties[propertyName] = explicitSetting\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t\tproviderName := provider.Name\n\t\t\t\t\tproviderType := provider.Type\n\n\t\t\t\t\t// instance_group.job can override the link name through the\n\t\t\t\t\t// instance_group.job.provides, via the \"as\" key\n\t\t\t\t\tif instanceGroup.Jobs[jobIdx].Provides != nil {\n\t\t\t\t\t\tif value, ok := instanceGroup.Jobs[jobIdx].Provides[providerName]; ok {\n\t\t\t\t\t\t\tswitch value.(type) {\n\t\t\t\t\t\t\tcase map[interface{}]interface{}:\n\t\t\t\t\t\t\t\tif overrideLinkName, ok := value.(map[interface{}]interface{})[\"as\"]; ok {\n\t\t\t\t\t\t\t\t\tproviderName = fmt.Sprintf(\"%v\", overrideLinkName)\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\tdefault:\n\t\t\t\t\t\t\t\treturn nil, nil, fmt.Errorf(\"unexpected type detected: %T, should have been a map\", value)\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\n\t\t\t\t\tif providers, ok := jobProviderLinks[providerType]; ok {\n\t\t\t\t\t\tif _, ok := providers[providerName]; ok {\n\t\t\t\t\t\t\treturn nil, nil, fmt.Errorf(\"multiple providers for link: name=%s type=%s\", providerName, providerType)\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\n\t\t\t\t\tif _, ok := jobProviderLinks[providerType]; !ok {\n\t\t\t\t\t\tjobProviderLinks[providerType] = map[string]manifest.JobLink{}\n\t\t\t\t\t}\n\n\t\t\t\t\t// construct the jobProviderLinks of the current job that provides\n\t\t\t\t\t// a link\n\t\t\t\t\tjobProviderLinks[providerType][providerName] = manifest.JobLink{\n\t\t\t\t\t\tInstances: jobsInstances,\n\t\t\t\t\t\tProperties: properties,\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn jobReleaseSpecs, jobProviderLinks, nil\n}", "func (o LookupInstanceResultOutput) Build() BuildResponseOutput {\n\treturn o.ApplyT(func(v LookupInstanceResult) BuildResponse { return v.Build }).(BuildResponseOutput)\n}", "func (s *SnapshotStatusService) buildURL() (string, url.Values, error) {\n\tvar err error\n\tvar path string\n\n\tif s.repository != \"\" {\n\t\tif len(s.snapshot) > 0 {\n\t\t\tpath, err = uritemplates.Expand(\"/_snapshot/{repository}/{snapshot}/_status\", map[string]string{\n\t\t\t\t\"repository\": s.repository,\n\t\t\t\t\"snapshot\": strings.Join(s.snapshot, \",\"),\n\t\t\t})\n\t\t} else {\n\t\t\tpath, err = uritemplates.Expand(\"/_snapshot/{repository}/_status\", map[string]string{\n\t\t\t\t\"repository\": s.repository,\n\t\t\t})\n\t\t}\n\t} else {\n\t\tpath, err = uritemplates.Expand(\"/_snapshot/_status\", nil)\n\t}\n\tif err != nil {\n\t\treturn \"\", url.Values{}, err\n\t}\n\n\t// Add query string parameters\n\tparams := url.Values{}\n\tif v := s.pretty; v != nil {\n\t\tparams.Set(\"pretty\", fmt.Sprint(*v))\n\t}\n\tif v := s.human; v != nil {\n\t\tparams.Set(\"human\", fmt.Sprint(*v))\n\t}\n\tif v := s.errorTrace; v != nil {\n\t\tparams.Set(\"error_trace\", fmt.Sprint(*v))\n\t}\n\tif len(s.filterPath) > 0 {\n\t\tparams.Set(\"filter_path\", strings.Join(s.filterPath, \",\"))\n\t}\n\tif s.masterTimeout != \"\" {\n\t\tparams.Set(\"master_timeout\", s.masterTimeout)\n\t}\n\tif v := s.ignoreUnavailable; v != nil {\n\t\tparams.Set(\"ignore_unavailable\", fmt.Sprint(*v))\n\t}\n\treturn path, params, nil\n}", "func (s *SnapshotsServiceOp) List(opt *ListOptions) ([]Snapshot, *Response, error) {\n\treturn s.list(opt, nil)\n}", "func (s Store) Build(c *Config) {\n\tlogger.Critical(\"Building lease tables\")\n\tleaseList := NetList(c.BaseIP, c.Subnet)\n\tll := LeaseList{}\n\tfor count, i := range leaseList {\n\t\tl := &Lease{}\n\t\tl.ID = int64(count)\n\t\tl.Created = time.Now()\n\t\tl.IP = i.String()\n\t\tll.Leases = append(ll.Leases, l)\n\t}\n\ts.leases = ll\n\t// Reserve the following\n\t// - network address\n\ts.Reserve(leaseList[0])\n\t// - self\n\ts.Reserve(c.BaseIP)\n\t// - broadcast\n\ts.Reserve(leaseList[len(leaseList)-1])\n\ts.leases.Save(s.DBname)\n}", "func NewSnapshot(version string,\n\tendpoints []types.Resource,\n\tclusters []types.Resource,\n\troutes []types.Resource,\n\tlisteners []types.Resource,\n\truntimes []types.Resource) Snapshot {\n\tout := Snapshot{}\n\tout.Resources[types.Endpoint] = NewResources(version, endpoints)\n\tout.Resources[types.Cluster] = NewResources(version, clusters)\n\tout.Resources[types.Route] = NewResources(version, routes)\n\tout.Resources[types.Listener] = NewResources(version, listeners)\n\tout.Resources[types.Runtime] = NewResources(version, runtimes)\n\treturn out\n}", "func TestCommandQueueGetSnapshotWithChildren(t *testing.T) {\n\tdefer leaktest.AfterTest(t)()\n\n\tcq := NewCommandQueue(true /* covering optimization */)\n\tcmd1 := add(cq, roachpb.Key(\"a\"), nil, false, nil)\n\tcmd2 := add(cq, roachpb.Key(\"a\"), nil, true, []*cmd{cmd1})\n\t// the following creates a node with two children because it has two spans\n\t// only the children show up in the snapshot.\n\tcq.add(true, zeroTS, []*cmd{cmd2}, []roachpb.Span{\n\t\t{Key: roachpb.Key(\"a\"), EndKey: roachpb.Key(\"b\")},\n\t\t{Key: roachpb.Key(\"d\"), EndKey: roachpb.Key(\"f\")},\n\t})\n\n\tsnapshot := cq.GetSnapshot()\n\n\tassertExpectedPrereqs(t, snapshot, map[int64][]int64{\n\t\t1: {},\n\t\t2: {1},\n\t\t4: {2},\n\t\t5: {2},\n\t})\n}", "func (b *MonitoringStackResourcesBuilder) Build() (object *MonitoringStackResources, err error) {\n\tobject = new(MonitoringStackResources)\n\tobject.bitmap_ = b.bitmap_\n\tif b.limits != nil {\n\t\tobject.limits, err = b.limits.Build()\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\tif b.requests != nil {\n\t\tobject.requests, err = b.requests.Build()\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\treturn\n}", "func Build(schema *schema.Schema, peers peers.Peers, logger logrus.FieldLogger) (*graphql.Field, error) {\n\tgetKinds := graphql.Fields{}\n\n\tif len(schema.Actions.Classes) == 0 && len(schema.Things.Classes) == 0 {\n\t\treturn nil, fmt.Errorf(\"there are no Actions or Things classes defined yet\")\n\t}\n\n\tcb := newClassBuilder(schema, peers, logger)\n\n\tif len(schema.Actions.Classes) > 0 {\n\t\tactions, err := cb.actions()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tgetKinds[\"Actions\"] = &graphql.Field{\n\t\t\tName: \"GetActions\",\n\t\t\tDescription: descriptions.GetActions,\n\t\t\tType: actions,\n\t\t\tResolve: func(p graphql.ResolveParams) (interface{}, error) {\n\t\t\t\t// Does nothing; pass through the filters\n\t\t\t\treturn p.Source, nil\n\t\t\t},\n\t\t}\n\t}\n\n\tif len(schema.Things.Classes) > 0 {\n\t\tthings, err := cb.things()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tgetKinds[\"Things\"] = &graphql.Field{\n\t\t\tName: \"GetThings\",\n\t\t\tDescription: descriptions.GetThings,\n\t\t\tType: things,\n\t\t\tResolve: func(p graphql.ResolveParams) (interface{}, error) {\n\t\t\t\t// Does nothing; pass through the filters\n\t\t\t\treturn p.Source, nil\n\t\t\t},\n\t\t}\n\t}\n\n\treturn &graphql.Field{\n\t\tName: \"Get\",\n\t\tDescription: descriptions.Get,\n\t\tType: graphql.NewObject(graphql.ObjectConfig{\n\t\t\tName: \"GetObj\",\n\t\t\tFields: getKinds,\n\t\t\tDescription: descriptions.GetObj,\n\t\t}),\n\t\tResolve: func(p graphql.ResolveParams) (interface{}, error) {\n\t\t\treturn p.Source, nil\n\t\t},\n\t}, nil\n}", "func builderImpl(\n\tc context.Context, masterName, builderName string, limit int, cursor string) (\n\t*resp.Builder, error) {\n\n\tvar thisCursor *datastore.Cursor\n\tif cursor != \"\" {\n\t\ttmpCur, err := datastore.DecodeCursor(c, cursor)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"bad cursor: %s\", err)\n\t\t}\n\t\tthisCursor = &tmpCur\n\t}\n\n\tresult := &resp.Builder{\n\t\tName: builderName,\n\t}\n\tmaster, internal, t, err := getMasterJSON(c, masterName)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif clock.Now(c).Sub(t) > 2*time.Minute {\n\t\twarning := fmt.Sprintf(\n\t\t\t\"WARNING: Master data is stale (last updated %s)\", t)\n\t\tlogging.Warningf(c, warning)\n\t\tresult.Warning = warning\n\t}\n\n\tp, ok := master.Builders[builderName]\n\tif !ok {\n\t\t// This long block is just to return a good error message when an invalid\n\t\t// buildbot builder is specified.\n\t\tkeys := make([]string, 0, len(master.Builders))\n\t\tfor k := range master.Builders {\n\t\t\tkeys = append(keys, k)\n\t\t}\n\t\tsort.Strings(keys)\n\t\treturn nil, errBuilderNotFound{masterName, builderName, keys}\n\t}\n\t// Extract pending builds out of the master json.\n\tresult.PendingBuilds = make([]*resp.BuildSummary, len(p.PendingBuildStates))\n\tlogging.Debugf(c, \"Number of pending builds: %d\", len(p.PendingBuildStates))\n\tfor i, pb := range p.PendingBuildStates {\n\t\tstart := time.Unix(int64(pb.SubmittedAt), 0).UTC()\n\t\tresult.PendingBuilds[i] = &resp.BuildSummary{\n\t\t\tPendingTime: resp.Interval{\n\t\t\t\tStarted: start,\n\t\t\t\tDuration: clock.Now(c).UTC().Sub(start),\n\t\t\t},\n\t\t}\n\t\tresult.PendingBuilds[i].Blame = make([]*resp.Commit, len(pb.Source.Changes))\n\t\tfor j, cm := range pb.Source.Changes {\n\t\t\tresult.PendingBuilds[i].Blame[j] = &resp.Commit{\n\t\t\t\tAuthorEmail: cm.Who,\n\t\t\t\tCommitURL: cm.Revlink,\n\t\t\t}\n\t\t}\n\t}\n\n\tbaseURL := \"https://build.chromium.org/p/\"\n\tif internal {\n\t\tbaseURL = \"https://uberchromegw.corp.google.com/i/\"\n\t}\n\tresult.MachinePool = summarizeSlavePool(baseURL+master.Name, p.Slaves, master.Slaves)\n\n\t// This is CPU bound anyways, so there's no need to do this in parallel.\n\tfinishedBuilds, nextCursor, err := getBuilds(c, masterName, builderName, true, limit, thisCursor)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif prevCursor, ok := maybeSetGetCursor(c, thisCursor, nextCursor, limit); ok {\n\t\tif prevCursor == nil {\n\t\t\t// Magic string to signal display prev without cursor\n\t\t\tresult.PrevCursor = \"EMPTY\"\n\t\t} else {\n\t\t\tresult.PrevCursor = (*prevCursor).String()\n\t\t}\n\t}\n\tif nextCursor != nil {\n\t\tresult.NextCursor = (*nextCursor).String()\n\t}\n\t// Cursor is not needed for current builds.\n\tcurrentBuilds, _, err := getBuilds(c, masterName, builderName, false, 0, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// currentBuilds is presented in reversed order, so flip it\n\tfor i, j := 0, len(currentBuilds)-1; i < j; i, j = i+1, j-1 {\n\t\tcurrentBuilds[i], currentBuilds[j] = currentBuilds[j], currentBuilds[i]\n\t}\n\tresult.CurrentBuilds = currentBuilds\n\n\tfor _, fb := range finishedBuilds {\n\t\tif fb != nil {\n\t\t\tresult.FinishedBuilds = append(result.FinishedBuilds, fb)\n\t\t}\n\t}\n\treturn result, nil\n}", "func (b *Builder) Build() []*metricsstore.MetricsStore {\n\tif b.whiteBlackList == nil {\n\t\tpanic(\"whiteBlackList should not be nil\")\n\t}\n\n\tcollectors := []*metricsstore.MetricsStore{}\n\tactiveCollectorNames := []string{}\n\n\tfor _, c := range b.enabledCollectors {\n\t\tconstructor, ok := availableCollectors[c]\n\t\tif !ok {\n\t\t\tklog.Fatalf(\"collector %s is not correct\", c)\n\t\t}\n\n\t\tcollector := constructor(b)\n\t\tactiveCollectorNames = append(activeCollectorNames, c)\n\t\tcollectors = append(collectors, collector)\n\n\t}\n\n\tklog.Infof(\"Active collectors: %s\", strings.Join(activeCollectorNames, \",\"))\n\n\treturn collectors\n}", "func ListSnapshots(sg *snapshotgroup.SnapshotGroup) ([]GeminiSnapshot, error) {\n\tclient := kube.GetClient()\n\tsnapshots, err := client.SnapshotClient.Namespace(sg.ObjectMeta.Namespace).List(metav1.ListOptions{})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tGeminiSnapshots := []GeminiSnapshot{}\n\tfor _, snapshot := range snapshots.Items {\n\t\tsnapshotMeta, err := meta.Accessor(&snapshot)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tannotations := snapshotMeta.GetAnnotations()\n\t\tif managedBy, ok := annotations[managedByAnnotation]; !ok || managedBy != managerName {\n\t\t\tcontinue\n\t\t}\n\t\tif annotations[GroupNameAnnotation] != sg.ObjectMeta.Name {\n\t\t\tcontinue\n\t\t}\n\t\ttimestampStr := annotations[TimestampAnnotation]\n\t\ttimestamp, err := strconv.Atoi(timestampStr)\n\t\tif err != nil {\n\t\t\tklog.Errorf(\"%s/%s: failed to parse unix timestamp %s for %s\", sg.ObjectMeta.Namespace, sg.ObjectMeta.Name, timestampStr, snapshotMeta.GetName())\n\t\t\tcontinue\n\t\t}\n\t\tintervals := []string{}\n\t\tintervalsStr := annotations[IntervalsAnnotation]\n\t\tif intervalsStr != \"\" {\n\t\t\tintervals = strings.Split(intervalsStr, intervalsSeparator)\n\t\t}\n\t\tGeminiSnapshots = append(GeminiSnapshots, GeminiSnapshot{\n\t\t\tNamespace: snapshotMeta.GetNamespace(),\n\t\t\tName: snapshotMeta.GetName(),\n\t\t\tTimestamp: time.Unix(int64(timestamp), 0),\n\t\t\tIntervals: intervals,\n\t\t\tRestore: annotations[RestoreAnnotation],\n\t\t})\n\t}\n\tsort.Slice(GeminiSnapshots, func(i, j int) bool {\n\t\treturn GeminiSnapshots[j].Timestamp.Before(GeminiSnapshots[i].Timestamp)\n\t})\n\treturn GeminiSnapshots, nil\n}", "func (r *Repo) GetSnapshots() ([]*Snapshot, error) {\n\tout, err := exec.Command(resticCmd, \"-r\", r.Path, \"-p\",\n\t\tr.Passwordfile, \"snapshots\", \"--json\").Output()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tfmt.Printf(\"%d bytes of output\\n\", len(out))\n\n\tvar snaps []*Snapshot\n\terr = json.Unmarshal(out, &snaps)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn snaps, nil\n}", "func (rb *SummaryBuilder) Build() Summary {\n\treturn *rb.v\n}", "func (rb *QueryWatchBuilder) Build() QueryWatch {\n\treturn *rb.v\n}", "func getBuilds(c context.Context, masterName, builderName string) ([]*resp.BuildRef, error) {\n\tresult := []*resp.BuildRef{}\n\tbs, err := model.GetBuilds(c, []model.BuildRoot{\n\t\tmodel.GetBuildRoot(c, masterName, builderName)}, 25)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tfor _, b := range bs[0] {\n\t\tmb := &resp.MiloBuild{\n\t\t\tSummary: resp.BuildComponent{\n\t\t\t\tStarted: b.ExecutionTime.Format(time.RFC3339),\n\t\t\t\t// TODO(hinoka/martiniss): Also get the real finished time and duration.\n\t\t\t\tFinished: b.ExecutionTime.Format(time.RFC3339),\n\t\t\t\tStatus: func(status string) resp.Status {\n\t\t\t\t\tswitch status {\n\t\t\t\t\tcase \"SUCCESS\":\n\t\t\t\t\t\treturn resp.Success\n\t\t\t\t\tcase \"FAILURE\":\n\t\t\t\t\t\treturn resp.Failure\n\t\t\t\t\tdefault:\n\t\t\t\t\t\t// TODO(hinoka): Also implement the other status types.\n\t\t\t\t\t\treturn resp.InfraFailure\n\t\t\t\t\t}\n\t\t\t\t}(b.UserStatus),\n\t\t\t\t// TODO(martiniss): Implement summary text.\n\t\t\t\tText: []string{\"Coming Soon....\"},\n\t\t\t},\n\t\t\t// TODO(hinoka/martiniss): Also get the repo so it's a real sourcestamp so\n\t\t\t// that the commit can be linkable.\n\t\t\tSourceStamp: &resp.SourceStamp{\n\t\t\t\tCommit: resp.Commit{\n\t\t\t\t\tRevision: b.Revisions[0].Digest,\n\t\t\t\t},\n\t\t\t},\n\t\t}\n\t\tresult = append(result, &resp.BuildRef{\n\t\t\tURL: func(url string) string {\n\t\t\t\tr, err := regexp.Compile(\".*/builds/(\\\\d+)/.*\")\n\t\t\t\tif err != nil {\n\t\t\t\t\tpanic(err)\n\t\t\t\t}\n\t\t\t\treturn r.FindStringSubmatch(url)[1]\n\t\t\t}(b.BuildLogKey),\n\t\t\tBuild: mb,\n\t\t})\n\t}\n\treturn result, nil\n}", "func (rb *ShardsRecordBuilder) Build() ShardsRecord {\n\treturn *rb.v\n}", "func AggregateSnapshots(snapshots []Snapshot) Snapshot {\n\tresult := createSnapshot()\n\n\tfor _, s := range snapshots {\n\t\tresult.mergeFrom(s)\n\t}\n\n\treturn result\n}", "func (c *kubernetesBindingsController) SnapshotsFrom(bindingNames ...string) map[string][]ObjectAndFilterResult {\n\tres := map[string][]ObjectAndFilterResult{}\n\n\tfor _, bindingName := range bindingNames {\n\t\t// Initialize all keys with empty arrays.\n\t\tres[bindingName] = make([]ObjectAndFilterResult, 0)\n\n\t\tsnapshot := c.SnapshotsFor(bindingName)\n\t\tif snapshot != nil {\n\t\t\tres[bindingName] = snapshot\n\t\t}\n\t}\n\n\treturn res\n}", "func (b *Builder) Build() car.Cars {\n\tvar carPark car.Cars\n\n\tfor _, currentModel := range b.orderedCars {\n\t\tcurrentCar := factory.MakeCar(currentModel)\n\n\t\tif currentCar == nil {\n\t\t\treturn nil\n\t\t}\n\n\t\tcarPark = append(carPark, currentCar)\n\t}\n\tb.orderedCars = car.Cars{}\n\treturn carPark\n}", "func (kvm *Clone) Snapshot(wr io.Writer) error {\n\tkvm.mu.RLock()\n\tdefer kvm.mu.RUnlock()\n\tdata, err := json.Marshal(kvm.keys)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif _, err := wr.Write(data); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (ks *KopiaSnapshotter) ListSnapshots() ([]string, error) {\n\treturn ks.snap.ListSnapshots()\n}", "func (syn *kubeSyncer) performSnapshot() ([]model.KVPair, map[string]bool, resourceVersions) {\n\topts := k8sapi.ListOptions{}\n\tversions := resourceVersions{}\n\tvar snap []model.KVPair\n\tvar keys map[string]bool\n\n\t// Loop until we successfully are able to accesss the API.\n\tfor {\n\t\t// Initialize the values to return.\n\t\tsnap = []model.KVPair{}\n\t\tkeys = map[string]bool{}\n\n\t\t// Get Namespaces (Profiles)\n\t\tlog.Info(\"Syncing Namespaces\")\n\t\tnsList, err := syn.kc.clientSet.Namespaces().List(opts)\n\t\tif err != nil {\n\t\t\tlog.Warnf(\"Error syncing Namespaces, retrying: %s\", err)\n\t\t\ttime.Sleep(1 * time.Second)\n\t\t\tcontinue\n\t\t}\n\t\tlog.Info(\"Received Namespace List() response\")\n\n\t\tversions.namespaceVersion = nsList.ListMeta.ResourceVersion\n\t\tfor _, ns := range nsList.Items {\n\t\t\t// The Syncer API expects a profile to be broken into its underlying\n\t\t\t// components - rules, tags, labels.\n\t\t\tprofile, err := syn.kc.converter.namespaceToProfile(&ns)\n\t\t\tif err != nil {\n\t\t\t\tlog.Panicf(\"%s\", err)\n\t\t\t}\n\t\t\trules, tags, labels := compat.ToTagsLabelsRules(profile)\n\t\t\trules.Revision = profile.Revision\n\t\t\ttags.Revision = profile.Revision\n\t\t\tlabels.Revision = profile.Revision\n\n\t\t\t// Also create a Policy for this Namespace.\n\t\t\tpolicy, err := syn.kc.converter.namespaceToPolicy(&ns)\n\t\t\tif err != nil {\n\t\t\t\tlog.Panicf(\"%s\", err)\n\t\t\t}\n\n\t\t\tsnap = append(snap, *rules, *tags, *labels, *policy)\n\t\t\tkeys[rules.Key.String()] = true\n\t\t\tkeys[tags.Key.String()] = true\n\t\t\tkeys[labels.Key.String()] = true\n\t\t\tkeys[policy.Key.String()] = true\n\t\t}\n\n\t\t// Get NetworkPolicies (Policies)\n\t\tlog.Info(\"Syncing NetworkPolicy\")\n\t\tnpList := extensions.NetworkPolicyList{}\n\t\terr = syn.kc.clientSet.Extensions().RESTClient().\n\t\t\tGet().\n\t\t\tResource(\"networkpolicies\").\n\t\t\tTimeout(10 * time.Second).\n\t\t\tDo().Into(&npList)\n\t\tif err != nil {\n\t\t\tlog.Warnf(\"Error querying NetworkPolicies during snapshot, retrying: %s\", err)\n\t\t\ttime.Sleep(1 * time.Second)\n\t\t\tcontinue\n\t\t}\n\t\tlog.Info(\"Received NetworkPolicy List() response\")\n\n\t\tversions.networkPolicyVersion = npList.ListMeta.ResourceVersion\n\t\tfor _, np := range npList.Items {\n\t\t\tpol, _ := syn.kc.converter.networkPolicyToPolicy(&np)\n\t\t\tsnap = append(snap, *pol)\n\t\t\tkeys[pol.Key.String()] = true\n\t\t}\n\n\t\t// Get Pods (WorkloadEndpoints)\n\t\tlog.Info(\"Syncing Pods\")\n\t\tpoList, err := syn.kc.clientSet.Pods(\"\").List(opts)\n\t\tif err != nil {\n\t\t\tlog.Warnf(\"Error querying Pods during snapshot, retrying: %s\", err)\n\t\t\ttime.Sleep(1 * time.Second)\n\t\t\tcontinue\n\t\t}\n\t\tlog.Info(\"Received Pod List() response\")\n\n\t\tversions.podVersion = poList.ListMeta.ResourceVersion\n\t\tfor _, po := range poList.Items {\n\t\t\t// Ignore any updates for pods which are not ready / valid.\n\t\t\tif !syn.kc.converter.isCalicoPod(&po) {\n\t\t\t\tlog.Debugf(\"Skipping pod %s/%s\", po.ObjectMeta.Namespace, po.ObjectMeta.Name)\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// Convert to a workload endpoint.\n\t\t\twep, err := syn.kc.converter.podToWorkloadEndpoint(&po)\n\t\t\tif err != nil {\n\t\t\t\tlog.WithError(err).Error(\"Failed to convert pod to workload endpoint\")\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tsnap = append(snap, *wep)\n\t\t\tkeys[wep.Key.String()] = true\n\t\t}\n\n\t\t// Sync GlobalConfig.\n\t\tlog.Info(\"Syncing GlobalConfig\")\n\t\tconfList, err := syn.kc.listGlobalConfig(model.GlobalConfigListOptions{})\n\t\tif err != nil {\n\t\t\tlog.Warnf(\"Error querying GlobalConfig during snapshot, retrying: %s\", err)\n\t\t\ttime.Sleep(1 * time.Second)\n\t\t\tcontinue\n\t\t}\n\t\tlog.Info(\"Received GlobalConfig List() response\")\n\n\t\tfor _, c := range confList {\n\t\t\tsnap = append(snap, *c)\n\t\t\tkeys[c.Key.String()] = true\n\t\t}\n\n\t\t// Sync IP Pools.\n\t\tlog.Info(\"Syncing IP Pools\")\n\t\tpoolList, err := syn.kc.List(model.IPPoolListOptions{})\n\t\tif err != nil {\n\t\t\tlog.Warnf(\"Error querying IP Pools during snapshot, retrying: %s\", err)\n\t\t\ttime.Sleep(1 * time.Second)\n\t\t\tcontinue\n\t\t}\n\t\tlog.Info(\"Received IP Pools List() response\")\n\n\t\tfor _, p := range poolList {\n\t\t\tsnap = append(snap, *p)\n\t\t\tkeys[p.Key.String()] = true\n\t\t}\n\n\t\tlog.Info(\"Syncing Nodes\")\n\t\tnoList, err := syn.kc.clientSet.Nodes().List(opts)\n\t\tif err != nil {\n\t\t\tlog.Warnf(\"Error syncing Nodes, retrying: %s\", err)\n\t\t\ttime.Sleep(1 * time.Second)\n\t\t\tcontinue\n\t\t}\n\t\tlog.Info(\"Received Node List() response\")\n\n\t\tversions.nodeVersion = noList.ListMeta.ResourceVersion\n\t\tfor _, no := range noList.Items {\n\t\t\tnode, err := resources.K8sNodeToCalico(&no)\n\t\t\tif err != nil {\n\t\t\t\tlog.Panicf(\"%s\", err)\n\t\t\t}\n\t\t\tif node != nil {\n\t\t\t\tsnap = append(snap, *node)\n\t\t\t\tkeys[node.Key.String()] = true\n\t\t\t}\n\t\t}\n\n\t\t// Include ready state.\n\t\tready, err := syn.kc.getReadyStatus(model.ReadyFlagKey{})\n\t\tif err != nil {\n\t\t\tlog.Warnf(\"Error querying ready status during snapshot, retrying: %s\", err)\n\t\t\ttime.Sleep(1 * time.Second)\n\t\t\tcontinue\n\t\t}\n\t\tsnap = append(snap, *ready)\n\t\tkeys[ready.Key.String()] = true\n\n\t\tlog.Infof(\"Snapshot resourceVersions: %+v\", versions)\n\t\tlog.Debugf(\"Created snapshot: %+v\", snap)\n\t\treturn snap, keys, versions\n\t}\n}", "func (m *VirtualEndpoint) GetSnapshots()([]CloudPcSnapshotable) {\n val, err := m.GetBackingStore().Get(\"snapshots\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.([]CloudPcSnapshotable)\n }\n return nil\n}", "func (f *FileSnapshotStore) getSnapshots() ([]*fileSnapshotMeta, error) {\n\t// Get the eligible snapshots\n\tsnapshots, err := os.ReadDir(f.path)\n\tif err != nil {\n\t\tf.logger.Printf(\"[ERR] snapshot: Failed to scan snapshot dir: %v\", err)\n\t\treturn nil, err\n\t}\n\n\t// Populate the metadata\n\tvar snapMeta []*fileSnapshotMeta\n\tfor _, snap := range snapshots {\n\t\t// Ignore any files\n\t\tif !snap.IsDir() {\n\t\t\tcontinue\n\t\t}\n\n\t\t// Ignore any temporary snapshots\n\t\tdirName := snap.Name()\n\t\tif strings.HasSuffix(dirName, tmpSuffix) {\n\t\t\tf.logger.Printf(\"[WARN] snapshot: Found temporary snapshot: %v\", dirName)\n\t\t\tcontinue\n\t\t}\n\n\t\t// Try to read the meta data\n\t\tmeta, err := f.readMeta(dirName)\n\t\tif err != nil {\n\t\t\tf.logger.Printf(\"[WARN] snapshot: Failed to read metadata for %v: %v\", dirName, err)\n\t\t\tcontinue\n\t\t}\n\n\t\t// Append, but only return up to the retain count\n\t\tsnapMeta = append(snapMeta, meta)\n\t}\n\n\t// Sort the snapshot, reverse so we get new -> old\n\tsort.Sort(sort.Reverse(snapMetaSlice(snapMeta)))\n\n\treturn snapMeta, nil\n}", "func (m *Nitro) GetSnapshots() []*Snapshot {\n\tvar snaps []*Snapshot\n\tbuf := m.snapshots.MakeBuf()\n\tdefer m.snapshots.FreeBuf(buf)\n\titer := m.snapshots.NewIterator(CompareSnapshot, buf)\n\titer.SeekFirst()\n\tfor ; iter.Valid(); iter.Next() {\n\t\tsnaps = append(snaps, (*Snapshot)(iter.Get()))\n\t}\n\n\treturn snaps\n}", "func FindFilteredSnapshots(ctx context.Context, be restic.Lister, loader restic.LoaderUnpacked, hosts []string, tags []restic.TagList, paths []string, snapshotIDs []string) <-chan *restic.Snapshot {\n\tout := make(chan *restic.Snapshot)\n\tgo func() {\n\t\tdefer close(out)\n\t\tbe, err := backend.MemorizeList(ctx, be, restic.SnapshotFile)\n\t\tif err != nil {\n\t\t\tWarnf(\"could not load snapshots: %v\\n\", err)\n\t\t\treturn\n\t\t}\n\n\t\terr = restic.FindFilteredSnapshots(ctx, be, loader, hosts, tags, paths, snapshotIDs, func(id string, sn *restic.Snapshot, err error) error {\n\t\t\tif err != nil {\n\t\t\t\tWarnf(\"Ignoring %q: %v\\n\", id, err)\n\t\t\t} else {\n\t\t\t\tselect {\n\t\t\t\tcase <-ctx.Done():\n\t\t\t\t\treturn ctx.Err()\n\t\t\t\tcase out <- sn:\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\t\tif err != nil {\n\t\t\tWarnf(\"could not load snapshots: %v\\n\", err)\n\t\t}\n\t}()\n\treturn out\n}", "func (r *Repository) Collect(ctx context.Context, live liveset.Liveset) error {\n\tvar w walker\n\tw.Init(r)\n\tvar (\n\t\tn int\n\t\tsize int64\n\t)\n\tfor w.Scan() {\n\t\tif live != nil && live.Contains(w.Digest()) {\n\t\t\tcontinue\n\t\t}\n\t\tsize += w.Info().Size()\n\t\tif err := os.Remove(w.Path()); err != nil {\n\t\t\tr.Log.Errorf(\"remove %q: %v\", w.Path(), err)\n\t\t}\n\t\t// Clean up object subdirectories. (Ignores failure when nonempty.)\n\t\tos.Remove(filepath.Dir(w.Path()))\n\t\tn++\n\t}\n\tif live != nil {\n\t\tr.Log.Printf(\"collected %v objects (%s)\", n, data.Size(size))\n\t}\n\treturn w.Err()\n}", "func Build(o parser.Object) Query {\n\troot := Query{IsTop: true}\n\tq := &root\n\tt := o.Type\n\tindex := 0\n\n\tfor {\n\t\tq.Index = index\n\t\tq.Type = t.RawTypeName()\n\n\t\tswitch kind := t.(type) {\n\t\tcase ast.BuiltIn:\n\t\t\tq.IsBuiltin = true\n\t\t\tt = nil\n\t\tcase ast.Array:\n\t\t\tq.IsArray = true\n\t\t\tt = kind.Element\n\t\tcase ast.Map:\n\t\t\tq.IsMap = true\n\t\t\tq.KeyType = kind.Key.RawTypeName()\n\t\t\tt = kind.Value\n\t\tcase ast.Struct:\n\t\t\tq.IsStruct = true\n\t\t\tfor _, f := range kind.Fields {\n\t\t\t\tq.Fields = append(q.Fields, Query{\n\t\t\t\t\tName: f.Name,\n\t\t\t\t\tAlias: f.Alias,\n\t\t\t\t\tIsBuiltin: true,\n\t\t\t\t\tIndex: index + 1,\n\t\t\t\t\tType: f.Type.RawTypeName(),\n\t\t\t\t})\n\t\t\t}\n\t\t\tt = nil\n\t\t}\n\n\t\tif t == nil {\n\t\t\tbreak\n\t\t}\n\n\t\tnext := &Query{}\n\n\t\tq.Next = next\n\n\t\tq = next\n\t\tindex++\n\t}\n\n\treturn root\n}", "func (vdb *VersionedDB) buildCommitters(updates *statedb.UpdateBatch) ([]batch, error) {\n\tnamespaces := updates.GetUpdatedNamespaces()\n\n\tvar nsCommitterBuilder []batch\n\tfor _, ns := range namespaces {\n\t\tnsUpdates := updates.GetUpdates(ns)\n\t\tdb, err := vdb.getNamespaceDBHandle(ns)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tnsRevs := vdb.committedDataCache.revs[ns]\n\t\tif nsRevs == nil {\n\t\t\tnsRevs = make(nsRevisions)\n\t\t}\n\t\t// for each namespace, construct one builder with the corresponding mongodb handle and mongo revisions\n\t\t// that are already loaded into cache (during validation phase)\n\t\tnsCommitterBuilder = append(nsCommitterBuilder, &nsCommittersBuilder{updates: nsUpdates, db: db, revisions: nsRevs})\n\t}\n\tif err := executeBatches(nsCommitterBuilder); err != nil {\n\t\treturn nil, err\n\t}\n\t// accumulate results across namespaces (one or more batches of `subNsCommitter` for a namespace from each builder)\n\tvar combinedSubNsCommitters []batch\n\tfor _, b := range nsCommitterBuilder {\n\t\tcombinedSubNsCommitters = append(combinedSubNsCommitters, b.(*nsCommittersBuilder).subNsCommitters...)\n\t}\n\treturn combinedSubNsCommitters, nil\n}", "func NewListSnapshotsRequest(server string) (*http.Request, error) {\n\tvar err error\n\n\tqueryUrl, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbasePath := fmt.Sprintf(\"/snapshot\")\n\tif basePath[0] == '/' {\n\t\tbasePath = basePath[1:]\n\t}\n\n\tqueryUrl, err = queryUrl.Parse(basePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq, err := http.NewRequest(\"GET\", queryUrl.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn req, nil\n}", "func (self *GoCodeGenerator) Build(m meta.HasDataDefinitions) error {\n\tmain := &objectRef{\n\t\tMeta: m,\n\t}\n\tself.all = []*objectRef{main}\n\treturn self.build(main)\n}", "func (rb *ArchiveBuilder) Build() Archive {\n\treturn *rb.v\n}", "func (b *Executor) Build(ctx context.Context, stages imagebuilder.Stages) (imageID string, ref reference.Canonical, err error) {\n\tif len(stages) == 0 {\n\t\treturn \"\", nil, errors.New(\"building: no stages to build\")\n\t}\n\tvar cleanupImages []string\n\tcleanupStages := make(map[int]*StageExecutor)\n\n\tstdout := b.out\n\tif b.quiet {\n\t\tb.out = io.Discard\n\t}\n\n\tcleanup := func() error {\n\t\tvar lastErr error\n\t\t// Clean up any containers associated with the final container\n\t\t// built by a stage, for stages that succeeded, since we no\n\t\t// longer need their filesystem contents.\n\n\t\tb.stagesLock.Lock()\n\t\tfor _, stage := range cleanupStages {\n\t\t\tif err := stage.Delete(); err != nil {\n\t\t\t\tlogrus.Debugf(\"Failed to cleanup stage containers: %v\", err)\n\t\t\t\tlastErr = err\n\t\t\t}\n\t\t}\n\t\tcleanupStages = nil\n\t\tb.stagesLock.Unlock()\n\n\t\t// Clean up any builders that we used to get data from images.\n\t\tfor _, builder := range b.containerMap {\n\t\t\tif err := builder.Delete(); err != nil {\n\t\t\t\tlogrus.Debugf(\"Failed to cleanup image containers: %v\", err)\n\t\t\t\tlastErr = err\n\t\t\t}\n\t\t}\n\t\tb.containerMap = nil\n\t\t// Clean up any intermediate containers associated with stages,\n\t\t// since we're not keeping them for debugging.\n\t\tif b.removeIntermediateCtrs {\n\t\t\tif err := b.deleteSuccessfulIntermediateCtrs(); err != nil {\n\t\t\t\tlogrus.Debugf(\"Failed to cleanup intermediate containers: %v\", err)\n\t\t\t\tlastErr = err\n\t\t\t}\n\t\t}\n\t\t// Remove images from stages except the last one, since we're\n\t\t// not going to use them as a starting point for any new\n\t\t// stages.\n\t\tfor i := range cleanupImages {\n\t\t\tremoveID := cleanupImages[len(cleanupImages)-i-1]\n\t\t\tif removeID == imageID {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif _, err := b.store.DeleteImage(removeID, true); err != nil {\n\t\t\t\tlogrus.Debugf(\"failed to remove intermediate image %q: %v\", removeID, err)\n\t\t\t\tif b.forceRmIntermediateCtrs || !errors.Is(err, storage.ErrImageUsedByContainer) {\n\t\t\t\t\tlastErr = err\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tcleanupImages = nil\n\n\t\tif b.rusageLogFile != nil && b.rusageLogFile != b.out {\n\t\t\t// we deliberately ignore the error here, as this\n\t\t\t// function can be called multiple times\n\t\t\tif closer, ok := b.rusageLogFile.(interface{ Close() error }); ok {\n\t\t\t\tcloser.Close()\n\t\t\t}\n\t\t}\n\t\treturn lastErr\n\t}\n\n\tdefer func() {\n\t\tif cleanupErr := cleanup(); cleanupErr != nil {\n\t\t\tif err == nil {\n\t\t\t\terr = cleanupErr\n\t\t\t} else {\n\t\t\t\terr = fmt.Errorf(\"%v: %w\", cleanupErr.Error(), err)\n\t\t\t}\n\t\t}\n\t}()\n\n\t// dependencyMap contains dependencyInfo for each stage,\n\t// dependencyInfo is used later to mark if a particular\n\t// stage is needed by target or not.\n\tdependencyMap := make(map[string]*stageDependencyInfo)\n\t// Build maps of every named base image and every referenced stage root\n\t// filesystem. Individual stages can use them to determine whether or\n\t// not they can skip certain steps near the end of their stages.\n\tfor stageIndex, stage := range stages {\n\t\tdependencyMap[stage.Name] = &stageDependencyInfo{Name: stage.Name, Position: stage.Position}\n\t\tnode := stage.Node // first line\n\t\tfor node != nil { // each line\n\t\t\tfor _, child := range node.Children { // tokens on this line, though we only care about the first\n\t\t\t\tswitch strings.ToUpper(child.Value) { // first token - instruction\n\t\t\t\tcase \"FROM\":\n\t\t\t\t\tif child.Next != nil { // second token on this line\n\t\t\t\t\t\t// If we have a fromOverride, replace the value of\n\t\t\t\t\t\t// image name for the first FROM in the Containerfile.\n\t\t\t\t\t\tif b.fromOverride != \"\" {\n\t\t\t\t\t\t\tchild.Next.Value = b.fromOverride\n\t\t\t\t\t\t\tb.fromOverride = \"\"\n\t\t\t\t\t\t}\n\t\t\t\t\t\tbase := child.Next.Value\n\t\t\t\t\t\tif base != \"scratch\" {\n\t\t\t\t\t\t\tif replaceBuildContext, ok := b.additionalBuildContexts[child.Next.Value]; ok {\n\t\t\t\t\t\t\t\tif replaceBuildContext.IsImage {\n\t\t\t\t\t\t\t\t\tchild.Next.Value = replaceBuildContext.Value\n\t\t\t\t\t\t\t\t\tbase = child.Next.Value\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\theadingArgs := argsMapToSlice(stage.Builder.HeadingArgs)\n\t\t\t\t\t\t\tuserArgs := argsMapToSlice(stage.Builder.Args)\n\t\t\t\t\t\t\t// append heading args so if --build-arg key=value is not\n\t\t\t\t\t\t\t// specified but default value is set in Containerfile\n\t\t\t\t\t\t\t// via `ARG key=value` so default value can be used.\n\t\t\t\t\t\t\tuserArgs = append(headingArgs, userArgs...)\n\t\t\t\t\t\t\tbaseWithArg, err := imagebuilder.ProcessWord(base, userArgs)\n\t\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\t\treturn \"\", nil, fmt.Errorf(\"while replacing arg variables with values for format %q: %w\", base, err)\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\tb.baseMap[baseWithArg] = true\n\t\t\t\t\t\t\tlogrus.Debugf(\"base for stage %d: %q\", stageIndex, base)\n\t\t\t\t\t\t\t// Check if selected base is not an additional\n\t\t\t\t\t\t\t// build context and if base is a valid stage\n\t\t\t\t\t\t\t// add it to current stage's dependency tree.\n\t\t\t\t\t\t\tif _, ok := b.additionalBuildContexts[baseWithArg]; !ok {\n\t\t\t\t\t\t\t\tif _, ok := dependencyMap[baseWithArg]; ok {\n\t\t\t\t\t\t\t\t\t// update current stage's dependency info\n\t\t\t\t\t\t\t\t\tcurrentStageInfo := dependencyMap[stage.Name]\n\t\t\t\t\t\t\t\t\tcurrentStageInfo.Needs = append(currentStageInfo.Needs, baseWithArg)\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\tcase \"ADD\", \"COPY\":\n\t\t\t\t\tfor _, flag := range child.Flags { // flags for this instruction\n\t\t\t\t\t\tif strings.HasPrefix(flag, \"--from=\") {\n\t\t\t\t\t\t\t// TODO: this didn't undergo variable and\n\t\t\t\t\t\t\t// arg expansion, so if the previous stage\n\t\t\t\t\t\t\t// was named using argument values, we might\n\t\t\t\t\t\t\t// not record the right value here.\n\t\t\t\t\t\t\trootfs := strings.TrimPrefix(flag, \"--from=\")\n\t\t\t\t\t\t\tb.rootfsMap[rootfs] = true\n\t\t\t\t\t\t\tlogrus.Debugf(\"rootfs needed for COPY in stage %d: %q\", stageIndex, rootfs)\n\t\t\t\t\t\t\t// Populate dependency tree and check\n\t\t\t\t\t\t\t// if following ADD or COPY needs any other\n\t\t\t\t\t\t\t// stage.\n\t\t\t\t\t\t\tstageName := rootfs\n\t\t\t\t\t\t\theadingArgs := argsMapToSlice(stage.Builder.HeadingArgs)\n\t\t\t\t\t\t\tuserArgs := argsMapToSlice(stage.Builder.Args)\n\t\t\t\t\t\t\t// append heading args so if --build-arg key=value is not\n\t\t\t\t\t\t\t// specified but default value is set in Containerfile\n\t\t\t\t\t\t\t// via `ARG key=value` so default value can be used.\n\t\t\t\t\t\t\tuserArgs = append(headingArgs, userArgs...)\n\t\t\t\t\t\t\tbaseWithArg, err := imagebuilder.ProcessWord(stageName, userArgs)\n\t\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\t\treturn \"\", nil, fmt.Errorf(\"while replacing arg variables with values for format %q: %w\", stageName, err)\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\tstageName = baseWithArg\n\t\t\t\t\t\t\t// If --from=<index> convert index to name\n\t\t\t\t\t\t\tif index, err := strconv.Atoi(stageName); err == nil {\n\t\t\t\t\t\t\t\tstageName = stages[index].Name\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t// Check if selected base is not an additional\n\t\t\t\t\t\t\t// build context and if base is a valid stage\n\t\t\t\t\t\t\t// add it to current stage's dependency tree.\n\t\t\t\t\t\t\tif _, ok := b.additionalBuildContexts[stageName]; !ok {\n\t\t\t\t\t\t\t\tif _, ok := dependencyMap[stageName]; ok {\n\t\t\t\t\t\t\t\t\t// update current stage's dependency info\n\t\t\t\t\t\t\t\t\tcurrentStageInfo := dependencyMap[stage.Name]\n\t\t\t\t\t\t\t\t\tcurrentStageInfo.Needs = append(currentStageInfo.Needs, stageName)\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\tcase \"RUN\":\n\t\t\t\t\tfor _, flag := range child.Flags { // flags for this instruction\n\t\t\t\t\t\t// We need to populate dependency tree of stages\n\t\t\t\t\t\t// if it is using `--mount` and `from=` field is set\n\t\t\t\t\t\t// and `from=` points to a stage consider it in\n\t\t\t\t\t\t// dependency calculation.\n\t\t\t\t\t\tif strings.HasPrefix(flag, \"--mount=\") && strings.Contains(flag, \"from\") {\n\t\t\t\t\t\t\tmountFlags := strings.TrimPrefix(flag, \"--mount=\")\n\t\t\t\t\t\t\tfields := strings.Split(mountFlags, \",\")\n\t\t\t\t\t\t\tfor _, field := range fields {\n\t\t\t\t\t\t\t\tif strings.HasPrefix(field, \"from=\") {\n\t\t\t\t\t\t\t\t\tfromField := strings.SplitN(field, \"=\", 2)\n\t\t\t\t\t\t\t\t\tif len(fromField) > 1 {\n\t\t\t\t\t\t\t\t\t\tmountFrom := fromField[1]\n\t\t\t\t\t\t\t\t\t\t// Check if this base is a stage if yes\n\t\t\t\t\t\t\t\t\t\t// add base to current stage's dependency tree\n\t\t\t\t\t\t\t\t\t\t// but also confirm if this is not in additional context.\n\t\t\t\t\t\t\t\t\t\tif _, ok := b.additionalBuildContexts[mountFrom]; !ok {\n\t\t\t\t\t\t\t\t\t\t\t// Treat from as a rootfs we need to preserve\n\t\t\t\t\t\t\t\t\t\t\tb.rootfsMap[mountFrom] = true\n\t\t\t\t\t\t\t\t\t\t\tif _, ok := dependencyMap[mountFrom]; ok {\n\t\t\t\t\t\t\t\t\t\t\t\t// update current stage's dependency info\n\t\t\t\t\t\t\t\t\t\t\t\tcurrentStageInfo := dependencyMap[stage.Name]\n\t\t\t\t\t\t\t\t\t\t\t\tcurrentStageInfo.Needs = append(currentStageInfo.Needs, mountFrom)\n\t\t\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\t\t\treturn \"\", nil, fmt.Errorf(\"invalid value for field `from=`: %q\", fromField[1])\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t\tnode = node.Next // next line\n\t\t}\n\t\t// Last stage is always target stage.\n\t\t// Since last/target stage is processed\n\t\t// let's calculate dependency map of stages\n\t\t// so we can mark stages which can be skipped.\n\t\tif stage.Position == (len(stages) - 1) {\n\t\t\tmarkDependencyStagesForTarget(dependencyMap, stage.Name)\n\t\t}\n\t}\n\tb.warnOnUnsetBuildArgs(stages, dependencyMap, b.args)\n\n\ttype Result struct {\n\t\tIndex int\n\t\tImageID string\n\t\tRef reference.Canonical\n\t\tError error\n\t}\n\n\tch := make(chan Result, len(stages))\n\n\tif b.stagesSemaphore == nil {\n\t\tb.stagesSemaphore = semaphore.NewWeighted(int64(len(stages)))\n\t}\n\n\tvar wg sync.WaitGroup\n\twg.Add(len(stages))\n\n\tgo func() {\n\t\tcancel := false\n\t\tfor stageIndex := range stages {\n\t\t\tindex := stageIndex\n\t\t\t// Acquire the semaphore before creating the goroutine so we are sure they\n\t\t\t// run in the specified order.\n\t\t\tif err := b.stagesSemaphore.Acquire(ctx, 1); err != nil {\n\t\t\t\tcancel = true\n\t\t\t\tb.lastError = err\n\t\t\t\tch <- Result{\n\t\t\t\t\tIndex: index,\n\t\t\t\t\tError: err,\n\t\t\t\t}\n\t\t\t\twg.Done()\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tb.stagesLock.Lock()\n\t\t\tcleanupStages := cleanupStages\n\t\t\tb.stagesLock.Unlock()\n\t\t\tgo func() {\n\t\t\t\tdefer b.stagesSemaphore.Release(1)\n\t\t\t\tdefer wg.Done()\n\t\t\t\tif cancel || cleanupStages == nil {\n\t\t\t\t\tvar err error\n\t\t\t\t\tif stages[index].Name != strconv.Itoa(index) {\n\t\t\t\t\t\terr = fmt.Errorf(\"not building stage %d: build canceled\", index)\n\t\t\t\t\t} else {\n\t\t\t\t\t\terr = fmt.Errorf(\"not building stage %d (%s): build canceled\", index, stages[index].Name)\n\t\t\t\t\t}\n\t\t\t\t\tch <- Result{\n\t\t\t\t\t\tIndex: index,\n\t\t\t\t\t\tError: err,\n\t\t\t\t\t}\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\t// Skip stage if it is not needed by TargetStage\n\t\t\t\t// or any of its dependency stages and `SkipUnusedStages`\n\t\t\t\t// is not set to `false`.\n\t\t\t\tif stageDependencyInfo, ok := dependencyMap[stages[index].Name]; ok {\n\t\t\t\t\tif !stageDependencyInfo.NeededByTarget && b.skipUnusedStages != types.OptionalBoolFalse {\n\t\t\t\t\t\tlogrus.Debugf(\"Skipping stage with Name %q and index %d since its not needed by the target stage\", stages[index].Name, index)\n\t\t\t\t\t\tch <- Result{\n\t\t\t\t\t\t\tIndex: index,\n\t\t\t\t\t\t\tError: nil,\n\t\t\t\t\t\t}\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tstageID, stageRef, stageErr := b.buildStage(ctx, cleanupStages, stages, index)\n\t\t\t\tif stageErr != nil {\n\t\t\t\t\tcancel = true\n\t\t\t\t\tch <- Result{\n\t\t\t\t\t\tIndex: index,\n\t\t\t\t\t\tError: stageErr,\n\t\t\t\t\t}\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\tch <- Result{\n\t\t\t\t\tIndex: index,\n\t\t\t\t\tImageID: stageID,\n\t\t\t\t\tRef: stageRef,\n\t\t\t\t\tError: nil,\n\t\t\t\t}\n\t\t\t}()\n\t\t}\n\t}()\n\tgo func() {\n\t\twg.Wait()\n\t\tclose(ch)\n\t}()\n\n\tfor r := range ch {\n\t\tstage := stages[r.Index]\n\n\t\tb.stagesLock.Lock()\n\t\tb.terminatedStage[stage.Name] = r.Error\n\t\tb.terminatedStage[strconv.Itoa(stage.Position)] = r.Error\n\n\t\tif r.Error != nil {\n\t\t\tb.stagesLock.Unlock()\n\t\t\tb.lastError = r.Error\n\t\t\treturn \"\", nil, r.Error\n\t\t}\n\n\t\t// If this is an intermediate stage, make a note of the ID, so\n\t\t// that we can look it up later.\n\t\tif r.Index < len(stages)-1 && r.ImageID != \"\" {\n\t\t\tb.imageMap[stage.Name] = r.ImageID\n\t\t\t// We're not populating the cache with intermediate\n\t\t\t// images, so add this one to the list of images that\n\t\t\t// we'll remove later.\n\t\t\tif !b.layers {\n\t\t\t\tcleanupImages = append(cleanupImages, r.ImageID)\n\t\t\t}\n\t\t}\n\t\tif r.Index == len(stages)-1 {\n\t\t\timageID = r.ImageID\n\t\t\tref = r.Ref\n\t\t}\n\t\tb.stagesLock.Unlock()\n\t}\n\n\tif len(b.unusedArgs) > 0 {\n\t\tunusedList := make([]string, 0, len(b.unusedArgs))\n\t\tfor k := range b.unusedArgs {\n\t\t\tunusedList = append(unusedList, k)\n\t\t}\n\t\tsort.Strings(unusedList)\n\t\tfmt.Fprintf(b.out, \"[Warning] one or more build args were not consumed: %v\\n\", unusedList)\n\t}\n\n\t// Add additional tags and print image names recorded in storage\n\tif dest, err := b.resolveNameToImageRef(b.output); err == nil {\n\t\tswitch dest.Transport().Name() {\n\t\tcase storageTransport.Transport.Name():\n\t\t\timg, err := storageTransport.Transport.GetStoreImage(b.store, dest)\n\t\t\tif err != nil {\n\t\t\t\treturn imageID, ref, fmt.Errorf(\"locating just-written image %q: %w\", transports.ImageName(dest), err)\n\t\t\t}\n\t\t\tif len(b.additionalTags) > 0 {\n\t\t\t\tif err = util.AddImageNames(b.store, \"\", b.systemContext, img, b.additionalTags); err != nil {\n\t\t\t\t\treturn imageID, ref, fmt.Errorf(\"setting image names to %v: %w\", append(img.Names, b.additionalTags...), err)\n\t\t\t\t}\n\t\t\t\tlogrus.Debugf(\"assigned names %v to image %q\", img.Names, img.ID)\n\t\t\t}\n\t\t\t// Report back the caller the tags applied, if any.\n\t\t\timg, err = storageTransport.Transport.GetStoreImage(b.store, dest)\n\t\t\tif err != nil {\n\t\t\t\treturn imageID, ref, fmt.Errorf(\"locating just-written image %q: %w\", transports.ImageName(dest), err)\n\t\t\t}\n\t\t\tfor _, name := range img.Names {\n\t\t\t\tfmt.Fprintf(b.out, \"Successfully tagged %s\\n\", name)\n\t\t\t}\n\n\t\tdefault:\n\t\t\tif len(b.additionalTags) > 0 {\n\t\t\t\tb.logger.Warnf(\"don't know how to add tags to images stored in %q transport\", dest.Transport().Name())\n\t\t\t}\n\t\t}\n\t}\n\n\tif err := cleanup(); err != nil {\n\t\treturn \"\", nil, err\n\t}\n\tlogrus.Debugf(\"printing final image id %q\", imageID)\n\tif b.iidfile != \"\" {\n\t\tif err = os.WriteFile(b.iidfile, []byte(\"sha256:\"+imageID), 0644); err != nil {\n\t\t\treturn imageID, ref, fmt.Errorf(\"failed to write image ID to file %q: %w\", b.iidfile, err)\n\t\t}\n\t} else {\n\t\tif _, err := stdout.Write([]byte(imageID + \"\\n\")); err != nil {\n\t\t\treturn imageID, ref, fmt.Errorf(\"failed to write image ID to stdout: %w\", err)\n\t\t}\n\t}\n\treturn imageID, ref, nil\n}", "func newSnapshotQueue(n int, w int, l logger.Logger) chan *fragment {\n\tch := make(chan *fragment, n)\n\tfor i := 0; i < w; i++ {\n\t\tgo snapshotQueueWorker(ch, l)\n\t}\n\treturn ch\n}", "func (rb *WatchRecordStatsBuilder) Build() WatchRecordStats {\n\treturn *rb.v\n}", "func getFullBuilds(c context.Context, masterName, builderName string, finished bool) ([]*buildbotBuild, error) {\n\t// TODO(hinoka): Builder specific structs.\n\tq := ds.NewQuery(\"buildbotBuild\")\n\tq = q.Eq(\"finished\", finished)\n\tq = q.Eq(\"master\", masterName)\n\tq = q.Eq(\"builder\", builderName)\n\tq = q.Order(\"-number\")\n\tq.Finalize()\n\t// Ignore the cursor, we don't need it.\n\tbuildbots, _, err := runBuildsQuery(c, q, 25)\n\treturn buildbots, err\n}", "func List(params Params) (*models.RepositoryConfigs, error) {\n\tif err := params.Validate(); err != nil {\n\t\treturn nil, err\n\t}\n\n\trepo, err := params.V1API.PlatformConfigurationSnapshots.GetSnapshotRepositories(\n\t\tplatform_configuration_snapshots.NewGetSnapshotRepositoriesParams(),\n\t\tparams.AuthWriter,\n\t)\n\tif err != nil {\n\t\treturn nil, api.UnwrapError(err)\n\t}\n\n\treturn repo.Payload, nil\n}", "func (s *SnapshotStatusService) Snapshot(snapshots ...string) *SnapshotStatusService {\n\ts.snapshot = append(s.snapshot, snapshots...)\n\treturn s\n}", "func (rb *IndexingStatsBuilder) Build() IndexingStats {\n\treturn *rb.v\n}", "func (vm *VM) ListMatchingSnapshots(regexes []string) ([]Snapshot, error) {\n\t// argument validity checking\n\texprs := make([]*regexp.Regexp, 0, len(regexes))\n\tfor _, arg := range regexes {\n\t\tregex, err := regexp.Compile(arg)\n\t\tif err != nil {\n\t\t\terr = fmt.Errorf(\"unable to compile regular expression %s: %s\", arg,\n\t\t\t\terr)\n\t\t\treturn nil, err\n\t\t}\n\t\texprs = append(exprs, regex)\n\t}\n\n\tif len(exprs) == 0 {\n\t\treturn nil, fmt.Errorf(\"no regular expression was specified\")\n\t}\n\n\t// retrieve all snapshots from libvirt\n\tinstances, err := vm.Instance.ListAllSnapshots(0)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"unable to retrieve snapshots for VM %s: %s\",\n\t\t\tvm.Descriptor.Name, err)\n\t\treturn nil, err\n\t}\n\n\tmatchedSnapshots := make([]Snapshot, 0, len(instances))\n\n\t// loop over snapshots and check for a match with the given\n\t// regular expressions\n\tfor _, instance := range instances {\n\n\t\t// retrieve and unmarshal the descriptor of snapshot\n\t\txml, err := instance.GetXMLDesc(0)\n\t\tif err != nil {\n\t\t\terr = fmt.Errorf(\"unable to get XML descriptor of snapshot: %s\", err)\n\t\t\tvm.Logger.Warnf(\"Skipping snapshot: %s\", err)\n\t\t\tcontinue\n\t\t}\n\n\t\tdescriptor := libvirtxml.DomainSnapshot{}\n\t\terr = descriptor.Unmarshal(xml)\n\t\tif err != nil {\n\t\t\terr = fmt.Errorf(\"unable to unmarshal the XML descriptor of snapshot: %s\", err)\n\t\t\tvm.Logger.Warn(\"Skipping snapshot: %s\", err)\n\t\t\tcontinue\n\t\t}\n\n\t\t// checking for a matching regular expression\n\t\tfound := false\n\t\tfor _, regex := range exprs {\n\t\t\tif regex.Find([]byte(descriptor.Name)) != nil {\n\t\t\t\tfound = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\n\t\tif found {\n\t\t\t// the caller is responsible for calling domain.Free() on the returned\n\t\t\t// domains\n\t\t\tmatchedSnapshot := Snapshot{\n\t\t\t\tInstance: instance,\n\t\t\t\tDescriptor: descriptor,\n\t\t\t}\n\t\t\tmatchedSnapshots = append(matchedSnapshots, matchedSnapshot)\n\t\t} else {\n\t\t\t// we do not need the instance here anymore\n\t\t\terr = instance.Free()\n\t\t\tif err != nil {\n\t\t\t\tvm.Logger.Warnf(\"unable to free snapshot %s: %s\",\n\t\t\t\t\tdescriptor.Name,\n\t\t\t\t\terr,\n\t\t\t\t)\n\t\t\t}\n\t\t}\n\t}\n\n\t// sort snapshots according to their creation date increasingly\n\tsorter := SnapshotSorter{\n\t\tSnapshots: &matchedSnapshots,\n\t}\n\tsort.Sort(&sorter)\n\n\treturn matchedSnapshots, nil\n}", "func (a StoriesAllStoriesNotModified) construct() StoriesAllStoriesClass { return &a }", "func (client *Client) ListSnapshotsWithOptions(request *ListSnapshotsRequest, runtime *util.RuntimeOptions) (_result *ListSnapshotsResponse, _err error) {\n\t_err = util.ValidateModel(request)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\tquery := map[string]interface{}{}\n\tif !tea.BoolValue(util.IsUnset(request.AuthTimeout)) {\n\t\tquery[\"AuthTimeout\"] = request.AuthTimeout\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.PageNo)) {\n\t\tquery[\"PageNo\"] = request.PageNo\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.PageSize)) {\n\t\tquery[\"PageSize\"] = request.PageSize\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.SnapshotType)) {\n\t\tquery[\"SnapshotType\"] = request.SnapshotType\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.VideoId)) {\n\t\tquery[\"VideoId\"] = request.VideoId\n\t}\n\n\treq := &openapi.OpenApiRequest{\n\t\tQuery: openapiutil.Query(query),\n\t}\n\tparams := &openapi.Params{\n\t\tAction: tea.String(\"ListSnapshots\"),\n\t\tVersion: tea.String(\"2017-03-21\"),\n\t\tProtocol: tea.String(\"HTTPS\"),\n\t\tPathname: tea.String(\"/\"),\n\t\tMethod: tea.String(\"POST\"),\n\t\tAuthType: tea.String(\"AK\"),\n\t\tStyle: tea.String(\"RPC\"),\n\t\tReqBodyType: tea.String(\"formData\"),\n\t\tBodyType: tea.String(\"json\"),\n\t}\n\t_result = &ListSnapshotsResponse{}\n\t_body, _err := client.CallApi(params, req, runtime)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\t_err = tea.Convert(_body, &_result)\n\treturn _result, _err\n}", "func (b *VolumeSnapshotClassBuilder) Result() *snapshotv1api.VolumeSnapshotClass {\n\treturn b.object\n}", "func (b *STSBuilder) Build() (object *STS, err error) {\n\tobject = new(STS)\n\tobject.bitmap_ = b.bitmap_\n\tobject.oidcEndpointURL = b.oidcEndpointURL\n\tobject.externalID = b.externalID\n\tif b.operatorIAMRoles != nil {\n\t\tobject.operatorIAMRoles = make([]*OperatorIAMRole, len(b.operatorIAMRoles))\n\t\tfor i, v := range b.operatorIAMRoles {\n\t\t\tobject.operatorIAMRoles[i], err = v.Build()\n\t\t\tif err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}\n\tobject.roleARN = b.roleARN\n\treturn\n}", "func (b *SubscriptionMetricsBuilder) Build() (object *SubscriptionMetrics, err error) {\n\tobject = new(SubscriptionMetrics)\n\tobject.bitmap_ = b.bitmap_\n\tobject.cloudProvider = b.cloudProvider\n\tif b.computeNodesCpu != nil {\n\t\tobject.computeNodesCpu, err = b.computeNodesCpu.Build()\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\tif b.computeNodesMemory != nil {\n\t\tobject.computeNodesMemory, err = b.computeNodesMemory.Build()\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\tif b.computeNodesSockets != nil {\n\t\tobject.computeNodesSockets, err = b.computeNodesSockets.Build()\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\tobject.consoleUrl = b.consoleUrl\n\tif b.cpu != nil {\n\t\tobject.cpu, err = b.cpu.Build()\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\tobject.criticalAlertsFiring = b.criticalAlertsFiring\n\tobject.healthState = b.healthState\n\tif b.memory != nil {\n\t\tobject.memory, err = b.memory.Build()\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\tif b.nodes != nil {\n\t\tobject.nodes, err = b.nodes.Build()\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\tobject.openshiftVersion = b.openshiftVersion\n\tobject.operatingSystem = b.operatingSystem\n\tobject.operatorsConditionFailing = b.operatorsConditionFailing\n\tobject.region = b.region\n\tif b.sockets != nil {\n\t\tobject.sockets, err = b.sockets.Build()\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\tobject.state = b.state\n\tobject.stateDescription = b.stateDescription\n\tif b.storage != nil {\n\t\tobject.storage, err = b.storage.Build()\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\tobject.subscriptionCpuTotal = b.subscriptionCpuTotal\n\tobject.subscriptionObligationExists = b.subscriptionObligationExists\n\tobject.subscriptionSocketTotal = b.subscriptionSocketTotal\n\tif b.upgrade != nil {\n\t\tobject.upgrade, err = b.upgrade.Build()\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\treturn\n}", "func GetEntriesAndNextToken(req *csi.ListSnapshotsRequest, snapshots []compute.Snapshot) (*csi.ListSnapshotsResponse, error) {\n\tif req == nil {\n\t\treturn nil, status.Errorf(codes.Aborted, \"request is nil\")\n\t}\n\n\tvar err error\n\tstart := 0\n\tif req.StartingToken != \"\" {\n\t\tstart, err = strconv.Atoi(req.StartingToken)\n\t\tif err != nil {\n\t\t\treturn nil, status.Errorf(codes.Aborted, \"ListSnapshots starting token(%s) parsing with error: %v\", req.StartingToken, err)\n\n\t\t}\n\t\tif start >= len(snapshots) {\n\t\t\treturn nil, status.Errorf(codes.Aborted, \"ListSnapshots starting token(%d) is greater than total number of snapshots\", start)\n\t\t}\n\t\tif start < 0 {\n\t\t\treturn nil, status.Errorf(codes.Aborted, \"ListSnapshots starting token(%d) can not be negative\", start)\n\t\t}\n\t}\n\n\tmaxEntries := len(snapshots) - start\n\tif req.MaxEntries > 0 && int(req.MaxEntries) < maxEntries {\n\t\tmaxEntries = int(req.MaxEntries)\n\t}\n\tentries := []*csi.ListSnapshotsResponse_Entry{}\n\tfor count := 0; start < len(snapshots) && count < maxEntries; start++ {\n\t\tif (req.SourceVolumeId != \"\" && req.SourceVolumeId == GetSourceVolumeID(&snapshots[start])) || req.SourceVolumeId == \"\" {\n\t\t\tcsiSnapshot, err := GenerateCSISnapshot(req.SourceVolumeId, &snapshots[start])\n\t\t\tif err != nil {\n\t\t\t\treturn nil, fmt.Errorf(\"failed to generate snapshot entry: %v\", err)\n\t\t\t}\n\t\t\tentries = append(entries, &csi.ListSnapshotsResponse_Entry{Snapshot: csiSnapshot})\n\t\t\tcount++\n\t\t}\n\t}\n\n\tnextToken := len(snapshots)\n\tif start < len(snapshots) {\n\t\tnextToken = start\n\t}\n\n\tlistSnapshotResp := &csi.ListSnapshotsResponse{\n\t\tEntries: entries,\n\t\tNextToken: strconv.Itoa(nextToken),\n\t}\n\n\treturn listSnapshotResp, nil\n}", "func (r *ReleasesImageBuilder) Build(releases model.Releases) error {\n\n\tif r.WorkerCount < 1 {\n\t\treturn fmt.Errorf(\"Invalid worker count %d\", r.WorkerCount)\n\t}\n\n\tif r.OutputDirectory != \"\" {\n\t\tr.DockerRegistry = \"\"\n\t\tr.DockerOrganization = \"\"\n\t}\n\n\tdockerManager, err := newDockerImageBuilder()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Error connecting to docker: %s\", err.Error())\n\t}\n\n\tif r.OutputDirectory != \"\" {\n\t\tif err = os.MkdirAll(r.OutputDirectory, 0755); err != nil {\n\t\t\treturn fmt.Errorf(\"Error creating output directory: %s\", err)\n\t\t}\n\t}\n\n\tworkerLib.MaxJobs = 1\n\tworker := workerLib.NewWorker()\n\n\tresultsCh := make(chan error)\n\tabort := make(chan struct{})\n\tfor _, release := range releases {\n\t\tworker.Add(releaseBuildJob{\n\t\t\trelease: release,\n\t\t\tbuilder: r,\n\t\t\tdockerManager: dockerManager,\n\t\t\tresultsCh: resultsCh,\n\t\t\tabort: abort,\n\t\t})\n\t}\n\n\tgo worker.RunUntilDone()\n\n\taborted := false\n\tfor i := 0; i < len(releases); i++ {\n\t\tresult := <-resultsCh\n\t\tif result != nil {\n\t\t\tif !aborted {\n\t\t\t\tclose(abort)\n\t\t\t\taborted = true\n\t\t\t}\n\t\t\terr = result\n\t\t}\n\t}\n\n\treturn err\n}", "func (rb *SimpleQueryStringQueryBuilder) Build() SimpleQueryStringQuery {\n\treturn *rb.v\n}", "func New(notifier MountNotifier, statePath string) *Snapshot {\n\t// Initialize default library driver\n\ts := &Snapshot{\n\t\tnotifier: newLocalStateNotifier(statePath, notifier),\n\t}\n\ts.driver = csicommon.NewCSIDriver(\"container-snapshot.openshift.io\", \"0.0.1\", uuid.New())\n\tif s.driver == nil {\n\t\tpanic(\"unable to initialize driver\")\n\t}\n\ts.driver.AddControllerServiceCapabilities([]csi.ControllerServiceCapability_RPC_Type{\n\t\tcsi.ControllerServiceCapability_RPC_CREATE_DELETE_VOLUME,\n\t})\n\ts.driver.AddVolumeCapabilityAccessModes([]csi.VolumeCapability_AccessMode_Mode{\n\t\tcsi.VolumeCapability_AccessMode_SINGLE_NODE_WRITER,\n\t\tcsi.VolumeCapability_AccessMode_MULTI_NODE_MULTI_WRITER,\n\t})\n\n\t// Create GRPC servers\n\ts.identity = &identityServer{\n\t\tDefaultIdentityServer: csicommon.NewDefaultIdentityServer(s.driver),\n\t}\n\ts.node = &nodeServer{\n\t\tDefaultNodeServer: csicommon.NewDefaultNodeServer(s.driver),\n\t\tnotifier: s.notifier,\n\t}\n\ts.controller = &controllerServer{\n\t\tDefaultControllerServer: csicommon.NewDefaultControllerServer(s.driver),\n\t}\n\treturn s\n}", "func (os *OpenStack) FindSnapshot(tags map[string]string) ([]string, []string, error) {\n\tvar snapshotIDs, statuses []string\n\tss, err := os.snapshotService()\n\tif err != nil || ss == nil {\n\t\tglog.Errorf(\"Unable to initialize cinder client for region: %s\", os.region)\n\t\treturn snapshotIDs, statuses, fmt.Errorf(\"Failed to find snapshot by tags %v: %v\", tags, err)\n\t}\n\n\topts := SnapshotListOpts{}\n\tsnapshots, err := ss.listSnapshots(opts)\n\n\tif err != nil {\n\t\tglog.Errorf(\"Failed to list snapshots. Error: %v\", err)\n\t\treturn snapshotIDs, statuses, err\n\t}\n\tglog.Infof(\"Listed [%v] snapshots.\", len(snapshots))\n\n\tglog.Infof(\"Looking for matching tags [%#v] in snapshots.\", tags)\n\t// Loop around to find the snapshot with the matching input metadata\n\t// NOTE(xyang): Metadata based filtering for snapshots is supported by Cinder volume API\n\t// microversion 3.21 and above. Currently the OpenStack Cloud Provider only supports V2.0.\n\t// Revisit this later when V3.0 is supported.\n\tfor _, snapshot := range snapshots {\n\t\tglog.Infof(\"Looking for matching tags in snapshot [%#v].\", snapshot)\n\t\tnamespaceVal, ok := snapshot.Metadata[ctrlsnap.CloudSnapshotCreatedForVolumeSnapshotNamespaceTag]\n\t\tif ok {\n\t\t\tif tags[ctrlsnap.CloudSnapshotCreatedForVolumeSnapshotNamespaceTag] == namespaceVal {\n\t\t\t\tnameVal, ok := snapshot.Metadata[ctrlsnap.CloudSnapshotCreatedForVolumeSnapshotNameTag]\n\t\t\t\tif ok {\n\t\t\t\t\tif tags[ctrlsnap.CloudSnapshotCreatedForVolumeSnapshotNameTag] == nameVal {\n\t\t\t\t\t\tuidVal, ok := snapshot.Metadata[ctrlsnap.CloudSnapshotCreatedForVolumeSnapshotUIDTag]\n\t\t\t\t\t\tif ok {\n\t\t\t\t\t\t\tif tags[ctrlsnap.CloudSnapshotCreatedForVolumeSnapshotUIDTag] == uidVal {\n\t\t\t\t\t\t\t\ttimeVal, ok := snapshot.Metadata[ctrlsnap.CloudSnapshotCreatedForVolumeSnapshotTimestampTag]\n\t\t\t\t\t\t\t\tif ok {\n\t\t\t\t\t\t\t\t\tif tags[ctrlsnap.CloudSnapshotCreatedForVolumeSnapshotTimestampTag] == timeVal {\n\t\t\t\t\t\t\t\t\t\tsnapshotIDs = append(snapshotIDs, snapshot.ID)\n\t\t\t\t\t\t\t\t\t\tstatuses = append(statuses, snapshot.Status)\n\t\t\t\t\t\t\t\t\t\tglog.Infof(\"Add snapshot [%#v].\", snapshot)\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn snapshotIDs, statuses, nil\n}", "func (rb *HoltMovingAverageAggregationBuilder) Build() HoltMovingAverageAggregation {\n\treturn *rb.v\n}", "func (c *context) buildItems(scope string, ptrType r.Type, out *mutant.BlockMutations) (ret block.Args, err error) {\n\tvar args block.Args\n\tstructType := ptrType.Elem()\n\t// a field (ex. enum) followed by a mutation will vanish;\n\t// collapsing into the invisible dummy input used for tracking mutations.\n\t// we need to flush those fields into a separate visible dummy input.\n\t// ( or stop the mutation from hiding, but that needs more data in Mutation/s )\n\tvar standaloneFields int\n\tfor i, cnt := 0, structType.NumField(); i < cnt; i++ {\n\t\tif field := structType.Field(i); len(field.PkgPath) == 0 {\n\t\t\tif field.Name != block.NextStatement {\n\t\t\t\tname := pascal.ToCaps(field.Name)\n\t\t\t\tif len(scope) > 0 {\n\t\t\t\t\t// ex. \"a, FIELD\"\n\t\t\t\t\tname = block.Scope(scope, name)\n\t\t\t\t}\n\t\t\t\tif desc, e := c.itemDesc(name, &field, out); e != nil {\n\t\t\t\t\terr = errutil.Append(err, e)\n\t\t\t\t} else if len(desc) > 0 {\n\t\t\t\t\tswitch desc[option.Type] {\n\t\t\t\t\tcase block.StatementInput, block.ValueInput:\n\t\t\t\t\t\tstandaloneFields = 0\n\t\t\t\t\t\tbreak\n\t\t\t\t\tcase block.DummyInput:\n\t\t\t\t\t\tif standaloneFields > 0 {\n\t\t\t\t\t\t\t// inject a dummy input before the new pending input to separate fields\n\t\t\t\t\t\t\tvisibleDummy := block.Dict{option.Type: block.DummyInput}\n\t\t\t\t\t\t\targs.AddArg(visibleDummy)\n\t\t\t\t\t\t}\n\t\t\t\t\t\tbreak\n\t\t\t\t\tdefault:\n\t\t\t\t\t\tstandaloneFields++\n\t\t\t\t\t}\n\t\t\t\t\targs.AddArg(desc)\n\t\t\t\t\t// atoms that are fields need to be encapsulated ( pushed down into ) a dummy input\n\t\t\t\t\t// otherwise they will get their own blank input, and RemoveAtoms will fail.\n\t\t\t\t\t// ( resulting in expansion of each mui input block into 4x workspace fields )\n\t\t\t\t\tif standaloneFields > 0 && len(scope) > 0 {\n\t\t\t\t\t\tvisibleDummy := block.Dict{option.Name: name, option.Type: block.DummyInput}\n\t\t\t\t\t\targs.AddArg(visibleDummy)\n\t\t\t\t\t\tstandaloneFields = 0\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tif err == nil {\n\t\tret = args\n\t}\n\treturn\n}", "func (b *ObjectBuilder) Build() JSON {\n\tif b.pairs == nil {\n\t\tpanic(msgModifyAfterBuild)\n\t}\n\torders := make([]int, len(b.pairs))\n\tfor i := range orders {\n\t\torders[i] = i\n\t}\n\tsorter := pairSorter{\n\t\tpairs: b.pairs,\n\t\torders: orders,\n\t\thasNonUnique: false,\n\t}\n\tb.pairs = nil\n\tsort.Sort(&sorter)\n\tsorter.unique()\n\treturn jsonObject(sorter.pairs)\n}", "func DBCreateSearchJSON(term string, limit int, baseUrl string) ([]projectModels.StructureSearchJSON, error) {\n\t//r.db(\"animedom\")\n\t//.table(\"animes\")\n\t//.filter(function(doc){\n\t//\treturn doc('MALTitle').downcase().match(\"attack\")\n\t//\t.or(doc('MALEnglish').downcase().match(\"attack\"))\n\t//}).pluck(\"MALTitle\", \"id\", \"Slug\").orderBy(\"MALTitle\")\n\tres, err := r.Table(\"animes\").Filter(func(anime r.Term) r.Term {\n\t\treturn anime.Field(\"MALTitle\").Downcase().Match(strings.ToLower(term)).\n\t\t\tOr(anime.Field(\"MALEnglish\").Downcase().Match(strings.ToLower(term)))\n\t}).OrderBy(\"MALTitle\").Pluck(\"MALTitle\", \"MALEnglish\", \"id\", \"Slug\", \"Genre\", \"Type\", \"WikiFNVHash\").Limit(limit).Run(dbSession)\n\tif err != nil {\n\t\treturn []projectModels.StructureSearchJSON{}, err\n\t}\n\n\tif res.IsNil() {\n\t\treturn []projectModels.StructureSearchJSON{}, errors.New(\"Empty Result\")\n\t}\n\n\tvar mappedData []projectModels.StructureAnime\n\n\terr = res.All(&mappedData)\n\tif err != nil {\n\t\treturn []projectModels.StructureSearchJSON{}, err\n\t}\n\n\terr = res.Close()\n\tcommon.CheckErrorAndPanic(err)\n\n\tvar readyData []projectModels.StructureSearchJSON\n\n\timgCDNUrl := baseUrl\n\tif imgCDNUrl != \"http://localhost:1993\" {\n\t\tif imgCDNUrl[0:5] == \"https\" {\n\t\t\timgCDNUrl = imgCDNUrl[0:5] + \"://cdn.\" + imgCDNUrl[8:]\n\t\t} else {\n\t\t\timgCDNUrl = imgCDNUrl[0:4] + \"://cdn.\" + imgCDNUrl[7:]\n\t\t}\n\t}\n\n\tfor i := 0; i < len(mappedData); i++ {\n\t\treadyDataInstance := projectModels.StructureSearchJSON{}\n\n\t\t// Assign Name\n\t\tif mappedData[i].MALEnglish != \"\" {\n\t\t\treadyDataInstance.AnimeName = mappedData[i].MALEnglish\n\t\t} else {\n\t\t\treadyDataInstance.AnimeName = mappedData[i].MALTitle\n\t\t}\n\n\t\t// Assign URL\n\t\treadyDataInstance.AnimeUrl = baseUrl + \"/wiki/\" +\n\t\t\tmappedData[i].MALID + \"/\" + mappedData[i].WikiHash\n\n\t\t// Assign Genres\n\t\tfor index := 0; index < len(mappedData[i].Genre); index++ {\n\t\t\tif index == len(mappedData[i].Genre)-1 {\n\t\t\t\treadyDataInstance.AnimeGenre += mappedData[i].Genre[index]\n\t\t\t} else {\n\t\t\t\treadyDataInstance.AnimeGenre += mappedData[i].Genre[index] + \", \"\n\t\t\t}\n\t\t}\n\n\t\t// Assign Thumbnail\n\t\treadyDataInstance.AnimeThumb = imgCDNUrl + \"/assets/img/smallestanime/\" + mappedData[i].MALID + \".jpg\"\n\n\t\t// Assign Type\n\t\treadyDataInstance.AnimeType = mappedData[i].Type\n\n\t\t// IsLast?\n\t\treadyDataInstance.IsLast = false\n\n\t\treadyData = append(readyData, readyDataInstance)\n\t}\n\n\treadyData = append(readyData, projectModels.StructureSearchJSON{IsLast: true, SearchTerm: term})\n\n\treturn readyData, nil\n}", "func (rb *ShardStoreBuilder) Build() ShardStore {\n\treturn *rb.v\n}", "func GetConsoleBuilds(\n\tc context.Context, builders []resp.BuilderRef, commits []string) (\n\t[][]*resp.ConsoleBuild, error) {\n\n\tresults := make([][]*resp.ConsoleBuild, len(commits))\n\tfor i := range results {\n\t\tresults[i] = make([]*resp.ConsoleBuild, len(builders))\n\t}\n\t// HACK(hinoka): This fetches 25 full builds and then filters them. Replace this\n\t// with something more reasonable.\n\t// This is kind of a hack but it's okay for now.\n\terr := parallel.FanOutIn(func(taskC chan<- func() error) {\n\t\tfor i, builder := range builders {\n\t\t\ti := i\n\t\t\tbuilder := builder\n\t\t\tbuilderComponents := strings.SplitN(builder.Name, \"/\", 2)\n\t\t\tif len(builderComponents) != 2 {\n\t\t\t\ttaskC <- func() error {\n\t\t\t\t\treturn fmt.Errorf(\"%s is an invalid builder name\", builder.Name)\n\t\t\t\t}\n\t\t\t\treturn\n\t\t\t}\n\t\t\tmaster := builderComponents[0]\n\t\t\tbuilderName := builderComponents[1]\n\t\t\ttaskC <- func() error {\n\t\t\t\tt1 := clock.Now(c)\n\t\t\t\tbuilds, err := getFullBuilds(c, master, builderName, true)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tt2 := clock.Now(c)\n\t\t\t\tvar currentStatus *model.Status\n\t\t\t\tfor j, commit := range commits {\n\t\t\t\t\tfor _, build := range builds {\n\t\t\t\t\t\tif build.Sourcestamp.Revision == commit {\n\t\t\t\t\t\t\tresults[j][i] = &resp.ConsoleBuild{\n\t\t\t\t\t\t\t\tLink: resp.NewLink(\n\t\t\t\t\t\t\t\t\tstrings.Join(build.Text, \" \"),\n\t\t\t\t\t\t\t\t\tfmt.Sprintf(\"/buildbot/%s/%s/%d\", master, builderName, build.Number),\n\t\t\t\t\t\t\t\t),\n\t\t\t\t\t\t\t\tStatus: build.toStatus(),\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\tcurrentStatus = &results[j][i].Status\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t\tif currentStatus != nil && results[j][i] == nil {\n\t\t\t\t\t\tresults[j][i] = &resp.ConsoleBuild{Status: *currentStatus}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tlog.Debugf(c,\n\t\t\t\t\t\"Builder %s took %s to query, %s to compute.\", builderName,\n\t\t\t\t\tt2.Sub(t1), clock.Since(c, t2))\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t})\n\treturn results, err\n}", "func Buildings(mods ...qm.QueryMod) buildingQuery {\n\tmods = append(mods, qm.From(\"\\\"buildings\\\"\"))\n\treturn buildingQuery{NewQuery(mods...)}\n}", "func New(meta Metadata, data []byte) (*Snapshot, error) {\n\n\tvar b bytes.Buffer\n\tgw, err := gzip.NewWriterLevel(&b, gzip.BestSpeed)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error building gzip writer: %w\", err)\n\t}\n\tgw.Write(data)\n\tgw.Close()\n\n\treturn &Snapshot{meta: meta, data: b.Bytes()}, nil\n}", "func (cli *Client) SnapshotList(filter filters.Args) (types.SnapshotsListResponse, error) {\n\tvar snapshots types.SnapshotsListResponse\n\tquery := url.Values{}\n\n\tif filter.Len() > 0 {\n\t\tfilterJSON, err := filters.ToParam(filter)\n\t\tif err != nil {\n\t\t\treturn snapshots, err\n\t\t}\n\t\tquery.Set(\"filters\", filterJSON)\n\t}\n\tresp, err := cli.get(\"/snapshots\", query, nil)\n\tif err != nil {\n\t\treturn snapshots, err\n\t}\n\n\terr = json.NewDecoder(resp.body).Decode(&snapshots)\n\tensureReaderClosed(resp)\n\treturn snapshots, err\n}", "func (b *Executor) Build(ctx context.Context, stages imagebuilder.Stages) (string, reference.Canonical, error) {\n\tif len(stages) == 0 {\n\t\terrors.New(\"error building: no stages to build\")\n\t}\n\tvar (\n\t\tstageExecutor *Executor\n\t\tlastErr error\n\t)\n\tb.imageMap = make(map[string]string)\n\tstageCount := 0\n\tfor _, stage := range stages {\n\t\tib := stage.Builder\n\t\tnode := stage.Node\n\t\tbase, err := ib.From(node)\n\t\tif err != nil {\n\t\t\tlogrus.Debugf(\"Build(node.Children=%#v)\", node.Children)\n\t\t\treturn \"\", nil, err\n\t\t}\n\n\t\tstageExecutor = b.withName(stage.Name, stage.Position, base)\n\t\tif err := stageExecutor.Prepare(ctx, stage, base); err != nil {\n\t\t\treturn \"\", nil, err\n\t\t}\n\t\t// Always remove the intermediate/build containers, even if the build was unsuccessful.\n\t\t// If building with layers, remove all intermediate/build containers if b.forceRmIntermediateCtrs\n\t\t// is true.\n\t\tif b.forceRmIntermediateCtrs || (!b.layers && !b.noCache) {\n\t\t\tdefer stageExecutor.Delete()\n\t\t}\n\t\tif err := stageExecutor.Execute(ctx, stage); err != nil {\n\t\t\tlastErr = err\n\t\t}\n\n\t\t// Delete the successful intermediate containers if an error in the build\n\t\t// process occurs and b.removeIntermediateCtrs is true.\n\t\tif lastErr != nil {\n\t\t\tif b.removeIntermediateCtrs {\n\t\t\t\tstageExecutor.deleteSuccessfulIntermediateCtrs()\n\t\t\t}\n\t\t\treturn \"\", nil, lastErr\n\t\t}\n\t\tb.containerIDs = append(b.containerIDs, stageExecutor.containerIDs...)\n\t\t// If we've a stage.Name with alpha and not numeric, we've an\n\t\t// AS clause in play. Create an intermediate image for this\n\t\t// stage to be used by other FROM statements that will want\n\t\t// to use it later in the Dockerfile. Note the id in our map.\n\t\tif _, err := strconv.Atoi(stage.Name); err != nil {\n\t\t\timgID, _, err := stageExecutor.Commit(ctx, stages[stageCount].Builder, \"\")\n\t\t\tif err != nil {\n\t\t\t\treturn \"\", nil, err\n\t\t\t}\n\t\t\tb.imageMap[stage.Name] = imgID\n\t\t}\n\t\tstageCount++\n\t}\n\n\tvar imageRef reference.Canonical\n\timageID := \"\"\n\n\t// Check if we have a one line Dockerfile making layers irrelevant\n\t// or the user told us to ignore layers.\n\tsingleLineDockerfile := (len(stages) < 2 && len(stages[0].Node.Children) < 1)\n\tignoreLayers := singleLineDockerfile || !b.layers && !b.noCache\n\n\tif ignoreLayers {\n\t\timgID, ref, err := stageExecutor.Commit(ctx, stages[len(stages)-1].Builder, \"\")\n\t\tif err != nil {\n\t\t\treturn \"\", nil, err\n\t\t}\n\t\tif singleLineDockerfile {\n\t\t\tb.log(\"COMMIT %s\", ref)\n\t\t}\n\t\timageID = imgID\n\t\timageRef = ref\n\t}\n\t// If building with layers and b.removeIntermediateCtrs is true\n\t// only remove intermediate container for each step if an error\n\t// during the build process doesn't occur.\n\t// If the build is unsuccessful, the container created at the step\n\t// the failure happened will persist in the container store.\n\t// This if condition will be false if not building with layers and\n\t// the removal of intermediate/build containers will be handled by the\n\t// defer statement above.\n\tif b.removeIntermediateCtrs && (b.layers || b.noCache) {\n\t\tif err := b.deleteSuccessfulIntermediateCtrs(); err != nil {\n\t\t\treturn \"\", nil, errors.Errorf(\"Failed to cleanup intermediate containers\")\n\t\t}\n\t}\n\t// Remove intermediate images that we created for AS clause handling\n\tfor _, value := range b.imageMap {\n\t\tif _, err := b.store.DeleteImage(value, true); err != nil {\n\t\t\tlogrus.Debugf(\"unable to remove intermediate image %q: %v\", value, err)\n\t\t}\n\t}\n\treturn imageID, imageRef, nil\n}", "func (u *TrackHitsBuilder) Build() TrackHits {\n\treturn u.v\n}", "func (o *Ordered) Pull(pipeline_name string, pipeline_version string) jenkins_types.Pipeline {\n\turlWithFilter := fmt.Sprintf(o.URL + o.filter())\n\n\tfmt.Printf(\"Pulling Ordered Job Data from latest build: %s\\n\", urlWithFilter)\n\tbody := o.Get(urlWithFilter)\n\n\tjson.Unmarshal(body, &o)\n\n\t// This needs to be processed by a job aggregator.\n\n\tvar jd jenkins_types.BuildsAndJobs\n\n\tlastBuild := o.LastBuild\n\tlastBuild.Fetch()\n\n\t/*\n\t * This gets all the builds and jobs from the Ordered Job, for example:\n\t * https://jenkins.example.com/job/project_2016.4.x\n\t */\n\tjd.Jobs, jd.Builds = jenkins_types.OrderedJobsAndBuildsFromDownstreamProjects(lastBuild, o.DownstreamProjects)\n\t// jd.Builds = BuildsFromAllJobs(o.LastBuild, jd.Jobs)\n\n\tlastBuild.TopLevelBuild = true\n\tjd.Builds = append(jd.Builds, lastBuild)\n\tfmt.Printf(\"======= ProcessTopLevelBuilds =====\")\n\t// subJobsBuilds := o.ProcessTopLevelBuilds(jd)\n\n\tsubJobsBuilds := jenkins_types.Builds{List: jd.Builds}\n\n\tjobData, trainData := subJobsBuilds.GetJobData(pipeline_name, pipeline_version)\n\n\t/*\n\t * Here is where I left off, I need to process all the builds for the sub jobs and turn it into job data.\n\t */\n\n\treturn jenkins_types.Pipeline{JobData: jobData, URL: o.URL, BuildNumber: lastBuild.Number, TrainData: trainData}\n}", "func BuildMerkleRoot(hashs []*crypto.HashType) []*crypto.HashType {\n\n\tleafSize := calcLowestHierarchyCount(len(hashs))\n\tarraySize := leafSize*2 - 1\n\tmerkles := make([]*crypto.HashType, arraySize)\n\tfor i, hash := range hashs {\n\t\tmerkles[i] = hash\n\t}\n\n\toffset := leafSize\n\tfor i := 0; i < arraySize-1; i += 2 {\n\t\tswitch {\n\t\tcase merkles[i] == nil:\n\t\t\tmerkles[offset] = nil\n\t\tcase merkles[i+1] == nil:\n\t\t\tnewHash := combineHash(merkles[i], merkles[i])\n\t\t\tmerkles[offset] = newHash\n\t\tdefault:\n\t\t\tnewHash := combineHash(merkles[i], merkles[i+1])\n\t\t\tmerkles[offset] = newHash\n\t\t}\n\t\toffset++\n\t}\n\n\treturn merkles\n}", "func (builder *OnDiskBuilder) Build() {\n\tdefer builder.parentWaitGroup.Done()\n\t// Fill the index with the documents the reader sends\n\tvar wg sync.WaitGroup\n\twg.Add(1)\n\tgo indexbuilders.FillIndex(builder.index, builder.readingChannel, &wg)\n\t// Start the disk writer\n\tgo indexbuilders.WritePostingLists(builder.writingChannel, &wg) // the block is copied to allow continuing operations on the builder\n\t// Wait for the index filling to be done before finishing the index\n\twg.Wait()\n\twg.Add(1)\n\t// Finish the index\n\tbuilder.finish()\n\t// Wait for the disk writing to be done\n\twg.Wait()\n\tlog.Printf(\"Done filling with %d documents\", builder.index.GetDocCounter())\n}", "func Build_Iot() PubsubHandlerFunc {\n\t// This is the pubsub handler\n\treturn func(ctx context.Context, m *pubsub.Message) error {\n\t\tdefer LogTiming(time.Now(), \"Collect\")\n\n\t\t// data:{\"name\":\"10.16.118.74\",\"privateIPv4\":\"10.16.118.74\",\"privateIPv6\":\"\"}\n\n\t\tbucketName := m.Attributes[\"deviceId\"]\n\t\tif !strings.HasPrefix(bucketName, \"source-\") {\n\t\t\treturn fmt.Errorf(\"message must be a bucket name starting with 'source-': %v\", bucketName)\n\t\t}\n\n\t\t// Create GCP client and get a handle on the bucket\n\t\tclient, err := storage.NewClient(context.Background())\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tbucket := client.Bucket(bucketName)\n\n\t\t// Load source config from the bucket\n\t\tvar config map[string]string\n\t\terr = unmarshalFromBucket(bucket, \"config.json\", &config)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Cannot read config.json: %v\", err)\n\t\t}\n\n\t\t// Load state from previous runs\n\t\tvar checksums map[string]string\n\t\terr = unmarshalFromBucket(bucket, \"state.json\", &checksums)\n\t\tif err != nil {\n\t\t\tif err != storage.ErrObjectNotExist {\n\t\t\t\treturn fmt.Errorf(\"Cannot read state.json: %v\", err)\n\t\t\t}\n\t\t\t// Use a default empty value instead\n\t\t\tchecksums = map[string]string{}\n\t\t}\n\n\t\tvar hostData = make(map[string]string)\n\t\terr = json.Unmarshal(m.Data, &hostData)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// Write items to the bucket\n\t\tvar item = MimosaData{\n\t\t\tVersion: \"1.0\",\n\t\t\tTyp: \"netscan-instance\",\n\t\t\tData: m.Data,\n\t\t}\n\t\tstart := time.Now()\n\t\tid := hostData[\"name\"]\n\t\tpreviousChecksum, present := checksums[id]\n\t\tsha := sha1.New()\n\t\t_, err = sha.Write(m.Data)\n\t\tif err != nil {\n\t\t\tlog.Printf(\"failed to compute SHA: %v\", err)\n\t\t\treturn err\n\t\t}\n\t\tchecksum := hex.EncodeToString(sha.Sum(nil))\n\t\tif !present || checksum != previousChecksum {\n\t\t\terr = writeToBucket(bucket, id, item.Typ, item.Version, item.Data)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tchecksums[id] = checksum\n\t\t\tlog.Printf(\"Change: %s\", id)\n\t\t\tlog.Printf(\"Timing: Write: %dms\", uint(time.Since(start).Seconds()*1000)) // Milliseconds not supported in Go 1.11\n\t\t} else {\n\t\t\tlog.Printf(\"No change found: %s\", id)\n\t\t}\n\n\t\t// Write state back to the bucket\n\t\tdata, err := json.Marshal(checksums)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Cannot marshal the value: %v\", err)\n\t\t}\n\t\terr = writeToBucket(bucket, \"state.json\", \"\", \"\", data)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}\n}", "func (b *Builder) Build() []*Collector {\n\n\tcollectors := []*Collector{}\n\tactiveCollectorNames := []string{}\n\n\tfor c := range b.enabledCollectors {\n\t\tconstructor, ok := availableCollectors[c]\n\t\tif ok {\n\t\t\tcollector := constructor(b)\n\t\t\tactiveCollectorNames = append(activeCollectorNames, c)\n\t\t\tcollectors = append(collectors, collector)\n\t\t}\n\t\t// TODO: What if not ok?\n\t}\n\n\tglog.Infof(\"Active collectors: %s\", strings.Join(activeCollectorNames, \",\"))\n\n\treturn collectors\n}", "func (rb *StringStatsAggregationBuilder) Build() StringStatsAggregation {\n\treturn *rb.v\n}", "func getNewSnapshotRequests(config types.AppInstanceConfig, status *types.AppInstanceStatus) (snapRequests []types.SnapshotDesc) {\n\tif config.Snapshot.Snapshots != nil {\n\t\tfor _, snap := range config.Snapshot.Snapshots {\n\t\t\tif isNewSnapshotRequest(snap.SnapshotID, status) {\n\t\t\t\tlog.Noticef(\"A new snapshot %s is requested\", snap.SnapshotID)\n\t\t\t\tsnapRequests = append(snapRequests, snap)\n\t\t\t}\n\t\t}\n\t}\n\treturn snapRequests\n}", "func (a *SnapshotApiService) SnapshotsGet(ctx _context.Context, optionals *SnapshotsGetOpts) (Snapshots, *APIResponse, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodGet\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue Snapshots\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/snapshots\"\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tif optionals != nil && optionals.Pretty != nil {\n\t\tlocalVarQueryParams.Add(\"pretty\", parameterToString(*optionals.Pretty, \"\"))\n\t}\n\tif optionals != nil && optionals.Depth != nil {\n\t\tlocalVarQueryParams.Add(\"depth\", parameterToString(*optionals.Depth, \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tif optionals != nil && optionals.XContractNumber != nil {\n\t\tlocalVarHeaderParams[\"X-Contract-Number\"] = parameterToString(*optionals.XContractNumber, \"\")\n\t}\n\tif ctx != nil {\n\t\t// API Key Authentication\n\t\tif auth, ok := ctx.Value(ContextAPIKey).(APIKey); ok {\n\t\t\tvar key string\n\t\t\tif auth.Prefix != \"\" {\n\t\t\t\tkey = auth.Prefix + \" \" + auth.Key\n\t\t\t} else {\n\t\t\t\tkey = auth.Key\n\t\t\t}\n\t\t\tlocalVarHeaderParams[\"Authorization\"] = key\n\t\t}\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tlocalVarAPIResponse := &APIResponse {\n\t\tResponse: localVarHTTPResponse,\n\t\tMethod: localVarHTTPMethod,\n\t\tRequestURL: localVarPath,\n\t\tOperation: \"SnapshotsGet\",\n\t}\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarAPIResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tlocalVarAPIResponse.Payload = localVarBody\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarAPIResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarAPIResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\treturn localVarReturnValue, localVarAPIResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarAPIResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarAPIResponse, nil\n}", "func Snapshot(player Player) Player {\n\tsrcProfile := player.Profile()\n\tsrcWpq := player.WaypointQueue()\n\n\tsnapshot := &PlayerSnapshot{\n\t\tindex: player.Index(),\n\t\tflags: player.Flags(),\n\t\tloadedRegion: player.LoadedRegion(),\n\t\tprofile: &ProfileSnapshot{\n\t\t\tusername: srcProfile.Username(),\n\t\t\tpassword: srcProfile.Password(),\n\t\t\trights: srcProfile.Rights(),\n\t\t\tpos: srcProfile.Position(),\n\t\t},\n\t}\n\n\tcurrentDirection, lastDirection := srcWpq.WalkDirection()\n\tsnapshot.waypointQueue = &WaypointQueueSnapshot{\n\t\tcurrentDirection: currentDirection,\n\t\tlastDirection: lastDirection,\n\t}\n\n\tskills := &SkillsSnapshot{\n\t\tcombatLevel: srcProfile.Skills().CombatLevel(),\n\t}\n\tsnapshot.profile.(*ProfileSnapshot).skills = skills\n\n\tsrcAppearance := srcProfile.Appearance()\n\tappearance := &AppearanceSnapshot{\n\t\tgender: srcAppearance.Gender(),\n\t\theadIcon: srcAppearance.HeadIcon(),\n\t\tmodels: map[BodyPart]int{\n\t\t\tTorso: srcAppearance.Model(Torso),\n\t\t\tArms: srcAppearance.Model(Arms),\n\t\t\tLegs: srcAppearance.Model(Legs),\n\t\t\tHead: srcAppearance.Model(Head),\n\t\t\tHands: srcAppearance.Model(Hands),\n\t\t\tFeet: srcAppearance.Model(Feet),\n\t\t\tBeard: srcAppearance.Model(Beard),\n\t\t},\n\t\tcolors: map[BodyPart]int{\n\t\t\tTorso: srcAppearance.Color(Torso),\n\t\t\tHair: srcAppearance.Color(Hair),\n\t\t\tLegs: srcAppearance.Color(Legs),\n\t\t\tFeet: srcAppearance.Color(Feet),\n\t\t\tSkin: srcAppearance.Color(Skin),\n\t\t},\n\t}\n\tsnapshot.profile.(*ProfileSnapshot).appearance = appearance\n\n\tsrcAnimations := player.Animations()\n\tanimations := &AnimationsSnapshot{\n\t\tanims: map[Anim]int{\n\t\t\tAnimIdle: srcAnimations.Animation(AnimIdle),\n\t\t\tAnimSpotRotate: srcAnimations.Animation(AnimSpotRotate),\n\t\t\tAnimWalk: srcAnimations.Animation(AnimWalk),\n\t\t\tAnimRotate180: srcAnimations.Animation(AnimRotate180),\n\t\t\tAnimRotateCCW: srcAnimations.Animation(AnimRotateCCW),\n\t\t\tAnimRotateCW: srcAnimations.Animation(AnimRotateCW),\n\t\t\tAnimRun: srcAnimations.Animation(AnimRun),\n\t\t},\n\t}\n\tsnapshot.animations = animations\n\n\treturn snapshot\n}", "func (b *Builder) Build() *monitoringv1.ServiceMonitor { return b.sm }", "func (s *OsdCsiServer) listMultipleSnapshots(\n\tctx context.Context,\n\treq *csi.ListSnapshotsRequest,\n) (*csi.ListSnapshotsResponse, error) {\n\tsourceVolumeId := req.GetSourceVolumeId()\n\tstartingToken := req.GetStartingToken()\n\tmaxEntries := req.GetMaxEntries()\n\n\tclogger.WithContext(ctx).Infof(\"ListSnapshots for multiple snapshots received. sourceVolumeId: %s, startingToken: %s, maxEntries: %v\",\n\t\tsourceVolumeId,\n\t\tstartingToken,\n\t\tmaxEntries,\n\t)\n\n\t// Get grpc connection\n\tconn, err := s.getConn()\n\tif err != nil {\n\t\treturn nil, status.Errorf(\n\t\t\tcodes.Unavailable,\n\t\t\t\"Unable to connect to SDK server: %v\", err)\n\t}\n\n\t// Get secret if any was passed\n\tctx = s.setupContext(ctx, req.GetSecrets())\n\tctx, cancel := grpcutil.WithDefaultTimeout(ctx)\n\tdefer cancel()\n\tvolumes := api.NewOpenStorageVolumeClient(conn)\n\n\t// Get all SnapshotIDs. Filter by source ID if provided.\n\tsnapshotsReq := &api.SdkVolumeSnapshotEnumerateWithFiltersRequest{\n\t\tVolumeId: sourceVolumeId,\n\t}\n\tsnapshotsResp, err := volumes.SnapshotEnumerateWithFilters(ctx, snapshotsReq)\n\tif err != nil {\n\t\terrStatus, ok := status.FromError(err)\n\t\tif ok && errStatus.Code() == codes.NotFound {\n\t\t\treturn &csi.ListSnapshotsResponse{}, nil\n\t\t}\n\t\treturn nil, status.Errorf(codes.Internal, \"Unable to get all snapshots: %v\", err)\n\t}\n\n\t// InspectWithFilters for all volumes\n\tvolumesResp, err := volumes.InspectWithFilters(ctx, &api.SdkVolumeInspectWithFiltersRequest{})\n\tif err != nil {\n\t\treturn nil, status.Errorf(codes.Internal, \"Unable to get all volumes: %v\", err)\n\t}\n\n\t// Sort snapshot IDs for repeatable results\n\tsortedSnapshotIds := sort.StringSlice(snapshotsResp.VolumeSnapshotIds)\n\tsort.Sort(sortedSnapshotIds)\n\n\t// Keep track of which volumes are snapshots\n\tvolumeForSnapId := make(map[string]*api.Volume)\n\tfor _, volResp := range volumesResp.Volumes {\n\t\tfor _, snapId := range sortedSnapshotIds {\n\t\t\tif volResp.Volume.Id == snapId {\n\t\t\t\tvolumeForSnapId[snapId] = volResp.Volume\n\t\t\t}\n\t\t}\n\t}\n\n\t// Generate response for all snapshots.\n\tlistSnapshotsResp := &csi.ListSnapshotsResponse{}\n\n\t// If starting token is provided, start skipping entries\n\t// until we hit the starting token.\n\tvar skipEntries bool\n\tif len(startingToken) > 0 {\n\t\tskipEntries = true\n\t}\n\tfor _, snapId := range sortedSnapshotIds {\n\t\t// Skip entries until we hit the starting token.\n\t\tif skipEntries && startingToken != snapId {\n\t\t\tcontinue\n\t\t}\n\t\tskipEntries = false\n\n\t\t// Before adding new object to response, check if we're at the max entries.\n\t\t// If we are at max entries, return with current iteration as NextToken.\n\t\t// This allows for calls to ListSnapshots to begin where we left off.\n\t\tvol := volumeForSnapId[snapId]\n\t\tif maxEntries > 0 && len(listSnapshotsResp.Entries) >= int(maxEntries) {\n\t\t\tlistSnapshotsResp.NextToken = vol.Id\n\t\t\treturn listSnapshotsResp, nil\n\t\t}\n\n\t\t// Populate entry with volume info\n\t\tentry := &csi.ListSnapshotsResponse_Entry{\n\t\t\tSnapshot: &csi.Snapshot{\n\t\t\t\tSizeBytes: int64(vol.GetSpec().GetSize()),\n\t\t\t\tSnapshotId: vol.Id,\n\t\t\t\tSourceVolumeId: vol.GetSource().Parent,\n\t\t\t\tCreationTime: vol.Ctime,\n\t\t\t\tReadyToUse: isSnapshotReady(vol),\n\t\t\t},\n\t\t}\n\n\t\tlistSnapshotsResp.Entries = append(listSnapshotsResp.Entries, entry)\n\t}\n\n\treturn listSnapshotsResp, nil\n}" ]
[ "0.5809499", "0.4765866", "0.47172648", "0.47149497", "0.46762368", "0.4637176", "0.46174473", "0.45891976", "0.45740902", "0.45685527", "0.45549625", "0.45362622", "0.4486892", "0.4464805", "0.44427574", "0.4436684", "0.44351923", "0.4427463", "0.44207242", "0.44114864", "0.43942273", "0.43871215", "0.43843183", "0.4375569", "0.43647528", "0.43596646", "0.43569028", "0.43469152", "0.434577", "0.43325466", "0.4331275", "0.43226376", "0.43225035", "0.43117398", "0.43005168", "0.4293849", "0.42796203", "0.42723155", "0.42717955", "0.4267626", "0.4267619", "0.42603558", "0.42592654", "0.42536578", "0.425174", "0.42448595", "0.4239428", "0.42352846", "0.4234699", "0.42341447", "0.42231992", "0.42222893", "0.422105", "0.42187977", "0.42077276", "0.42068985", "0.41979495", "0.4197738", "0.41940024", "0.4191344", "0.41891065", "0.41887793", "0.41878378", "0.4150488", "0.4150474", "0.41477096", "0.41377977", "0.4135796", "0.4134719", "0.41329268", "0.412902", "0.4125924", "0.41195545", "0.41187644", "0.4117512", "0.41100523", "0.41086787", "0.41013518", "0.40998375", "0.40996376", "0.40985987", "0.4098525", "0.40968996", "0.40948182", "0.4092941", "0.4090701", "0.4087908", "0.4085971", "0.4083248", "0.40781274", "0.40744707", "0.40731335", "0.407137", "0.4070619", "0.40680563", "0.40659776", "0.40645126", "0.40625942", "0.40565318", "0.40546712" ]
0.67760396
0
ProtocolVersion returns transport protocol version.
func (t *grpcTransport) ProtocolVersion() centrifuge.ProtocolVersion { return centrifuge.ProtocolVersion2 }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (peer *peerImp) GetProtocolVersion() uint8 {\n\treturn peer.protocolVersion\n}", "func (m *_KnxNetIpMessage) GetProtocolVersion() uint8 {\n\treturn KnxNetIpMessage_PROTOCOLVERSION\n}", "func (p *Peer) ProtocolVersion() uint32 {\n\tp.flagsMtx.Lock()\n\tprotocolVersion := p.protocolVersion\n\tp.flagsMtx.Unlock()\n\n\treturn protocolVersion\n}", "func GetEthProtocolVersion() (version string, err error) {\n\tversion, err = EthRpcClient.EthProtocolVersion()\n\tif err != nil {\n\t\tlogs.Error(\"eth_protocolVersion failed, %v\", err)\n\t}\n\treturn\n}", "func (s *PublicHarmonyAPI) ProtocolVersion() hexutil.Uint {\n\treturn hexutil.Uint(proto.ProtocolVersion)\n}", "func (api *PublicEthereumAPI) ProtocolVersion() hexutil.Uint {\n\tapi.logger.Debug(\"eth_protocolVersion\")\n\treturn hexutil.Uint(ethermint.ProtocolVersion)\n}", "func (e *Ethereum) ProtocolVersion() (uint64, error) {\n\tvar resProtocolVer string\n\terr := e.rpcClient.CallContext(e.ctx, &resProtocolVer, \"eth_protocolVersion\")\n\tif err != nil {\n\t\treturn 0, errors.Wrapf(err, \"fail to call rpc.CallContext(eth_protocolVersion) error: %s\", err)\n\t}\n\th, err := e.DecodeBig(resProtocolVer)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn h.Uint64(), err\n}", "func CurrentProtocolVersion() ProtocolVersion {\n\treturn ProtocolVersion(currentRPCProtocolVersion)\n}", "func (t *websocketTransport) ProtocolVersion() ProtocolVersion {\n\treturn ProtocolVersion2\n}", "func (c *Client) ProtocolVersion() (*NumberResponse, error) {\n\trequest := c.newRequest(EthProtocolVersion)\n\n\tresponse := &NumberResponse{}\n\n\treturn response, c.send(request, response)\n}", "func GetTLSVersion(tr *http.Transport) string {\n switch tr.TLSClientConfig.MinVersion {\n case tls.VersionTLS10:\n return \"TLS 1.0\"\n case tls.VersionTLS11:\n return \"TLS 1.1\"\n case tls.VersionTLS12:\n return \"TLS 1.2\"\n case tls.VersionTLS13:\n return \"TLS 1.3\"\n }\n\n return \"Unknown\"\n}", "func (m *Memberlist) ProtocolVersion() uint8 {\n\t// NOTE: This method exists so that in the future we can control\n\t// any locking if necessary, if we change the protocol version at\n\t// runtime, etc.\n\treturn m.config.ProtocolVersion\n}", "func ProtocolVersion(major int, minor int) int {\n\tif minor > 999 {\n\t\tminor = 999\n\t}\n\treturn major*1000 + minor\n}", "func (o TargetGroupOutput) ProtocolVersion() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v TargetGroup) *string { return v.ProtocolVersion }).(pulumi.StringPtrOutput)\n}", "func (b IPv4Header) Protocol() int {\n\treturn int(b[9])\n}", "func (ds *Dsync) ProtocolVersion() (protocol.ID, error) {\n\treturn DsyncProtocolID, nil\n}", "func (c ClientProperties) ProtocolVersion() uint8 {\n\treturn c[4]\n}", "func TestServiceProtocolVersion(t *testing.T) {\n\ts := res.NewService(\"test\")\n\trestest.AssertEqualJSON(t, \"ProtocolVersion()\", s.ProtocolVersion(), \"1.2.2\")\n}", "func (peer *peerImp) SetProtocolVersion(version uint8) {\n\tpeer.protocolVersion = version\n}", "func (x *InitializeErrorUnknownProtocolVersion) GetUnknownProtocolVersion() int32 {\n\tif x != nil {\n\t\treturn x.UnknownProtocolVersion\n\t}\n\treturn 0\n}", "func (e E_OpenconfigSystem_System_SshServer_Config_ProtocolVersion) String() string {\n\treturn ygot.EnumLogString(e, int64(e), \"E_OpenconfigSystem_System_SshServer_Config_ProtocolVersion\")\n}", "func (*protocol) Number() tcpip.TransportProtocolNumber {\n\treturn ProtocolNumber\n}", "func (c *Config) CheckProtocolVersion() error {\n\tif c.ProtocolVersion < ProtocolVersionMin {\n\t\treturn fmt.Errorf(\"Protocol version '%d' too low. Must be in range: [%d, %d]\", c.ProtocolVersion, ProtocolVersionMin, ProtocolVersionMax)\n\t}\n\tif c.ProtocolVersion > ProtocolVersionMax {\n\t\treturn fmt.Errorf(\"Protocol version '%d' too high. Must be in range: [%d, %d]\", c.ProtocolVersion, ProtocolVersionMin, ProtocolVersionMax)\n\t}\n\treturn nil\n}", "func CurrentRPCProtocolVersion() looprpc.ProtocolVersion {\n\treturn currentRPCProtocolVersion\n}", "func (t *websocketTransport) Protocol() ProtocolType {\n\treturn t.opts.protoType\n}", "func (o TargetGroupPtrOutput) ProtocolVersion() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *TargetGroup) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.ProtocolVersion\n\t}).(pulumi.StringPtrOutput)\n}", "func (p OpenFlow10Protocol) GetVersion() uint8 {\n\treturn goloxi.VERSION_1_0\n}", "func (b IPv4Header) Version() int {\n\treturn int(b[0] >> 4)\n}", "func (v *Client) Protocol() string {\n\tif v.protocol == \"\" {\n\t\tif v.req.TLS != nil {\n\t\t\tv.protocol = \"https\"\n\t\t} else {\n\t\t\tv.protocol = \"http\"\n\t\t}\n\t}\n\treturn v.protocol\n}", "func (m *_CipIdentity) GetEncapsulationProtocolVersion() uint16 {\n\treturn m.EncapsulationProtocolVersion\n}", "func (h *RequestHeader) Protocol() []byte {\n\tif len(h.proto) == 0 {\n\t\treturn strHTTP11\n\t}\n\treturn h.proto\n}", "func (c *RedfishClient) GetProtocol() string {\n\treturn constvalue.RedfishV1\n}", "func (c *RedfishClient) GetProtocol() string {\n\treturn constvalue.RedfishV1\n}", "func TLSVersion(ver uint16) string {\n\tswitch ver {\n\tcase tls.VersionTLS10:\n\t\treturn \"1.0\"\n\tcase tls.VersionTLS11:\n\t\treturn \"1.1\"\n\tcase tls.VersionTLS12:\n\t\treturn \"1.2\"\n\t}\n\treturn fmt.Sprintf(\"Unknown [%x]\", ver)\n}", "func (m *Heartbeat) GetVersion() int {\n\tif m.HasExtensionFieldValues {\n\t\treturn 2\n\t}\n\n\treturn 1\n}", "func (p *protocol) Number() tcpip.NetworkProtocolNumber {\r\n\treturn ProtocolNumber\r\n}", "func (d *Datagram) Version() int {\n\treturn int(d.Data[0] >> 4)\n}", "func (p *IPv4) Version() int {\n\treturn p.version\n}", "func (*NodeInfo_ProtocolVersion) Descriptor() ([]byte, []int) {\n\treturn file_manager_proto_rawDescGZIP(), []int{0, 0}\n}", "func (m MessageHeader) Version() uint8 {\n\treturn m.version\n}", "func (btc *ExchangeWallet) getVersion() (uint64, uint64, error) {\n\tr := &struct {\n\t\tVersion uint64 `json:\"version\"`\n\t\tProtocolVersion uint64 `json:\"protocolversion\"`\n\t}{}\n\terr := btc.wallet.call(methodGetNetworkInfo, nil, r)\n\tif err != nil {\n\t\treturn 0, 0, err\n\t}\n\treturn r.Version, r.ProtocolVersion, nil\n}", "func (h *ResponseHeader) Protocol() []byte {\n\tif len(h.protocol) > 0 {\n\t\treturn h.protocol\n\t}\n\treturn strHTTP11\n}", "func (p TCPPort) Protocol() Protocol {\n\treturn TCP\n}", "func (sc *ServerConn) negotiateVersion() (string, error) {\n\treqMsg, err := prepareRequest(sc.nextID(), \"server.version\", positional{\"Electrum\", \"1.4\"})\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treqMsg = append(reqMsg, newline)\n\n\tif err = sc.send(reqMsg); err != nil {\n\t\treturn \"\", err\n\t}\n\n\terr = sc.conn.SetReadDeadline(time.Now().Add(10 * time.Second))\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treader := bufio.NewReader(io.LimitReader(sc.conn, 1<<18))\n\tmsg, err := reader.ReadBytes(newline)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tvar jsonResp response\n\terr = json.Unmarshal(msg, &jsonResp)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tvar vers []string // [server_software_version, protocol_version]\n\terr = json.Unmarshal(jsonResp.Result, &vers)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif len(vers) != 2 {\n\t\treturn \"\", fmt.Errorf(\"unexpected version response: %v\", vers)\n\t}\n\treturn vers[1], nil\n}", "func (*NodeInfo_ProtocolVersion) Descriptor() ([]byte, []int) {\n\treturn file_resources_proto_rawDescGZIP(), []int{25, 0}\n}", "func (p *IPv4) Protocol() Protocol {\n\treturn p.protocol\n}", "func GetMetadataVersion(v *ProtocolVersion) byte {\n\tif v.Below(2, 3) {\n\t\treturn 0x02 // no support for optional attributes\n\t}\n\treturn 0x03 // current version\n}", "func protocolVersion(opts *ServeConfig) (int, Protocol, PluginSet) {\n\tprotoVersion := int(opts.ProtocolVersion)\n\tpluginSet := opts.Plugins\n\tprotoType := ProtocolNetRPC\n\t// Check if the client sent a list of acceptable versions\n\tvar clientVersions []int\n\tif vs := os.Getenv(\"PLUGIN_PROTOCOL_VERSIONS\"); vs != \"\" {\n\t\tfor _, s := range strings.Split(vs, \",\") {\n\t\t\tv, err := strconv.Atoi(s)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Fprintf(os.Stderr, \"server sent invalid plugin version %q\", s)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tclientVersions = append(clientVersions, v)\n\t\t}\n\t}\n\n\t// We want to iterate in reverse order, to ensure we match the newest\n\t// compatible plugin version.\n\tsort.Sort(sort.Reverse(sort.IntSlice(clientVersions)))\n\n\t// set the old un-versioned fields as if they were versioned plugins\n\tif opts.VersionedPlugins == nil {\n\t\topts.VersionedPlugins = make(map[int]PluginSet)\n\t}\n\n\tif pluginSet != nil {\n\t\topts.VersionedPlugins[protoVersion] = pluginSet\n\t}\n\n\t// Sort the version to make sure we match the latest first\n\tvar versions []int\n\tfor v := range opts.VersionedPlugins {\n\t\tversions = append(versions, v)\n\t}\n\n\tsort.Sort(sort.Reverse(sort.IntSlice(versions)))\n\n\t// See if we have multiple versions of Plugins to choose from\n\tfor _, version := range versions {\n\t\t// Record each version, since we guarantee that this returns valid\n\t\t// values even if they are not a protocol match.\n\t\tprotoVersion = version\n\t\tpluginSet = opts.VersionedPlugins[version]\n\n\t\t// If we have a configured gRPC server we should select a protocol\n\t\tif opts.GRPCServer != nil {\n\t\t\t// All plugins in a set must use the same transport, so check the first\n\t\t\t// for the protocol type\n\t\t\tfor _, p := range pluginSet {\n\t\t\t\tswitch p.(type) {\n\t\t\t\tcase GRPCPlugin:\n\t\t\t\t\tprotoType = ProtocolGRPC\n\t\t\t\tdefault:\n\t\t\t\t\tprotoType = ProtocolNetRPC\n\t\t\t\t}\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\n\t\tfor _, clientVersion := range clientVersions {\n\t\t\tif clientVersion == protoVersion {\n\t\t\t\treturn protoVersion, protoType, pluginSet\n\t\t\t}\n\t\t}\n\t}\n\n\t// Return the lowest version as the fallback.\n\t// Since we iterated over all the versions in reverse order above, these\n\t// values are from the lowest version number plugins (which may be from\n\t// a combination of the Handshake.ProtocolVersion and ServeConfig.Plugins\n\t// fields). This allows serving the oldest version of our plugins to a\n\t// legacy client that did not send a PLUGIN_PROTOCOL_VERSIONS list.\n\treturn protoVersion, protoType, pluginSet\n}", "func (networkClient NetworkClient) UsedProtocol() string {\n\treturn networkClient.usedProtocol\n}", "func (w *Whisper) Version() uint {\n\treturn w.protocol.Version\n}", "func (f *FastURL) GetProtocol() []byte {\n\treturn f.protocol\n}", "func (c *coreVersion) IsProtocolVersionEqualOrAbove(protocolVer int) bool {\n\tif c.ledgerProtocolVersion == 0 {\n\t\treturn false\n\t}\n\treturn c.ledgerProtocolVersion >= protocolVer\n}", "func (o StudioComponentInitializationScriptOutput) LaunchProfileProtocolVersion() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v StudioComponentInitializationScript) *string { return v.LaunchProfileProtocolVersion }).(pulumi.StringPtrOutput)\n}", "func (m *MockRPCConfig) TransportProtocol() (r transport.Protocol) {\n\tif m.TransportProtocolFunc != nil {\n\t\treturn m.TransportProtocolFunc()\n\t}\n\treturn\n}", "func Protocol(v protocol.ProtocolVersion) RequestOption {\n\treturn func(o *RequestOptions) {\n\t\to.ProtocolVersion = v\n\t}\n}", "func (*HeartbeatRequest) Version() int16 {\n\treturn 0\n}", "func (b IPv6Header) Version() int {\n\treturn int(b[0]) >> 4\n}", "func (http HTTPConfig) Protocol() string {\n\tif http.TLS.Enabled() {\n\t\treturn \"https\"\n\t}\n\treturn \"http\"\n}", "func (m *Message) Version() uint8 {\n\treturn m.version\n}", "func (conn *Conn) Version(t string) { conn.Ctcp(t, VERSION) }", "func (req *RequestMessage) GetVersion() int16 {\n\treturn req.version\n}", "func (p *PacketCryptAnn) GetVersion() uint {\n\treturn uint(p.Header[0])\n}", "func (rr *OPT) Version() uint8 {\n\treturn uint8(rr.Hdr.Ttl & 0x00FF0000 >> 16)\n}", "func (c *Client) Version() string {\n\treturn c.conn.Version()\n}", "func (d *Datagram) Protocol() string {\n\tprotocolHeaderNumber := BytesToIntForEndianness(d.Data[9:10], binary.BigEndian)\n\treturn ProtocolMap[protocolHeaderNumber]\n}", "func (pipe *PipeWS) LocalProtocol() uint16 {\n\treturn pipe.proto.Self\n}", "func getManagedTransportVer() string {\n\tconst transportVersion = \"1\"\n\tfor _, offered := range strings.Split(getenvRequired(\"TOR_PT_MANAGED_TRANSPORT_VER\"), \",\") {\n\t\tif offered == transportVersion {\n\t\t\treturn offered\n\t\t}\n\t}\n\treturn \"\"\n}", "func GetProtocol(route routev1.Route, ingress iextensionsv1.Ingress) string {\n\tif !reflect.DeepEqual(ingress, iextensionsv1.Ingress{}) && ingress.Spec.TLS != nil {\n\t\treturn \"https\"\n\t} else if !reflect.DeepEqual(route, routev1.Route{}) && route.Spec.TLS != nil {\n\t\treturn \"https\"\n\t}\n\treturn \"http\"\n}", "func protocolNameToNumber(protocolName string) uint16 {\n\tswitch strings.ToLower(protocolName) {\n\tcase \"tcp\":\n\t\treturn 6\n\tcase \"udp\":\n\t\treturn 17\n\tcase \"icmp\":\n\t\treturn 1\n\tcase \"icmpv6\":\n\t\treturn 58\n\tcase \"sctp\":\n\t\treturn 132\n\tcase \"udplite\":\n\t\treturn 136\n\tdefault:\n\t\treturn 256 // any (as understood by hns)\n\t}\n}", "func (g *Generic) Version() uint8 {\n\treturn g.Header.Version\n}", "func (p *Plugin) Protocol() string {\n\tif p.PluginObj.Config.Interface.ProtocolScheme != \"\" {\n\t\treturn p.PluginObj.Config.Interface.ProtocolScheme\n\t}\n\treturn plugins.ProtocolSchemeHTTPV1\n}", "func (SslParameters_ProtocolVersion) EnumDescriptor() ([]byte, []int) {\n\treturn file_github_com_solo_io_gloo_projects_gloo_api_v1_ssl_ssl_proto_rawDescGZIP(), []int{5, 0}\n}", "func getTransportProtocol(ctx context.Context, stype linux.SockType, protocol int) (tcpip.TransportProtocolNumber, bool, *syserr.Error) {\n\tswitch stype {\n\tcase linux.SOCK_STREAM:\n\t\tif protocol != 0 && protocol != unix.IPPROTO_TCP {\n\t\t\treturn 0, true, syserr.ErrInvalidArgument\n\t\t}\n\t\treturn tcp.ProtocolNumber, true, nil\n\n\tcase linux.SOCK_DGRAM:\n\t\tswitch protocol {\n\t\tcase 0, unix.IPPROTO_UDP:\n\t\t\treturn udp.ProtocolNumber, true, nil\n\t\tcase unix.IPPROTO_ICMP:\n\t\t\treturn header.ICMPv4ProtocolNumber, true, nil\n\t\tcase unix.IPPROTO_ICMPV6:\n\t\t\treturn header.ICMPv6ProtocolNumber, true, nil\n\t\t}\n\n\tcase linux.SOCK_RAW:\n\t\t// Raw sockets require CAP_NET_RAW.\n\t\tcreds := auth.CredentialsFromContext(ctx)\n\t\tif !creds.HasCapability(linux.CAP_NET_RAW) {\n\t\t\trawMissingLogger.Infof(\"A process tried to create a raw socket without CAP_NET_RAW. Should the container config enable CAP_NET_RAW?\")\n\t\t\treturn 0, true, syserr.ErrNotPermitted\n\t\t}\n\n\t\tswitch protocol {\n\t\tcase unix.IPPROTO_ICMP:\n\t\t\treturn header.ICMPv4ProtocolNumber, true, nil\n\t\tcase unix.IPPROTO_ICMPV6:\n\t\t\treturn header.ICMPv6ProtocolNumber, true, nil\n\t\tcase unix.IPPROTO_UDP:\n\t\t\treturn header.UDPProtocolNumber, true, nil\n\t\tcase unix.IPPROTO_TCP:\n\t\t\treturn header.TCPProtocolNumber, true, nil\n\t\t// IPPROTO_RAW signifies that the raw socket isn't assigned to\n\t\t// a transport protocol. Users will be able to write packets'\n\t\t// IP headers and won't receive anything.\n\t\tcase unix.IPPROTO_RAW:\n\t\t\treturn tcpip.TransportProtocolNumber(0), false, nil\n\t\t}\n\t}\n\treturn 0, true, syserr.ErrProtocolNotSupported\n}", "func (fs *FakeSession) GetVersion() gosnmp.SnmpVersion {\n\treturn gosnmp.Version3\n}", "func (E_OpenconfigSystem_System_SshServer_Config_ProtocolVersion) IsYANGGoEnum() {}", "func TransportProtocol_Values() []string {\n\treturn []string{\n\t\tTransportProtocolTcp,\n\t\tTransportProtocolUdp,\n\t}\n}", "func (c *Conn) Protocol() string {\n\treturn \"websocket\"\n}", "func TPMVersion() (string, error) {\n\t_, tpmVersion, err := tpmEnabledAndVersion()\n\treturn tpmVersion, err\n}", "func (t JoinGroupRequestProtocol11) Size(version int16) int32 {\n\tvar sz int32\n\tsz += sizeof.String(t.Name) // Name\n\tsz += sizeof.Bytes(t.Metadata) // Metadata\n\treturn sz\n}", "func (p IPPacket) Version() int {\n\tif len(p) == 0 {\n\t\treturn 0\n\t}\n\treturn int(p[0] >> 4)\n}", "func (E_OpenconfigOfficeAp_System_SshServer_Config_ProtocolVersion) IsYANGGoEnum() {}", "func (_m *MockBackend) ProtocolVersion() string {\n\tret := _m.Called()\n\n\tvar r0 string\n\tif rf, ok := ret.Get(0).(func() string); ok {\n\t\tr0 = rf()\n\t} else {\n\t\tr0 = ret.Get(0).(string)\n\t}\n\n\treturn r0\n}", "func (requestHeader *RequestHeader) Version() uint16 {\n\treturn requestHeader.APIVersion\n}", "func (m *msgHandler) Version() uint32 {\n\tm.owner.mu.Lock()\n\tdefer m.owner.mu.Unlock()\n\treturn m.owner.version\n}", "func (o *StoragePhysicalDisk) GetProtocol() string {\n\tif o == nil || o.Protocol == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Protocol\n}", "func (s *DatabaseServerV3) GetTeleportVersion() string {\n\treturn s.Spec.Version\n}", "func (h Header) Version() byte {\n\treturn h[1]\n}", "func (pipe *PipeWS) RemoteProtocol() uint16 {\n\treturn pipe.proto.Self\n}", "func (o *SSHAuthority) Version() int {\n\n\treturn 1\n}", "func (_TTFT20 *TTFT20Session) Version() (string, error) {\n\treturn _TTFT20.Contract.Version(&_TTFT20.CallOpts)\n}", "func TLSVersionName(version uint16) string {\n\tswitch version {\n\tcase tls.VersionSSL30:\n\t\treturn \"SSL 3.0\"\n\tcase tls.VersionTLS10:\n\t\treturn \"TLS 1.0\"\n\tcase tls.VersionTLS11:\n\t\treturn \"TLS 1.1\"\n\tcase tls.VersionTLS12:\n\t\treturn \"TLS 1.2\"\n\tcase tls.VersionTLS13:\n\t\treturn \"TLS 1.3\"\n\t}\n\n\treturn \"Unknown\"\n}", "func stringToProtocol(protocol string) uint16 {\n\tswitch strings.ToLower(protocol) {\n\tcase \"tcp\":\n\t\treturn uint16(unix.IPPROTO_TCP)\n\tcase \"udp\":\n\t\treturn uint16(unix.IPPROTO_UDP)\n\tcase \"sctp\":\n\t\treturn uint16(unix.IPPROTO_SCTP)\n\t}\n\treturn uint16(0)\n}", "func stringToProtocol(protocol string) uint16 {\n\tswitch strings.ToLower(protocol) {\n\tcase \"tcp\":\n\t\treturn uint16(unix.IPPROTO_TCP)\n\tcase \"udp\":\n\t\treturn uint16(unix.IPPROTO_UDP)\n\tcase \"sctp\":\n\t\treturn uint16(unix.IPPROTO_SCTP)\n\t}\n\treturn uint16(0)\n}", "func (r *Ricochet) negotiateVersion(conn net.Conn, outbound bool) (*OpenConnection, error) {\n\tversions := []byte{0x49, 0x4D, 0x01, 0x01}\n\n\t// Outbound side of the connection sends a list of supported versions\n\tif outbound {\n\t\tif n, err := conn.Write(versions); err != nil || n < len(versions) {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tres := make([]byte, 1)\n\t\tif _, err := io.ReadAtLeast(conn, res, len(res)); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tif res[0] != 0x01 {\n\t\t\treturn nil, errors.New(\"unsupported protocol version\")\n\t\t}\n\t} else {\n\t\t// Read version response header\n\t\theader := make([]byte, 3)\n\t\tif _, err := io.ReadAtLeast(conn, header, len(header)); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tif header[0] != versions[0] || header[1] != versions[1] || header[2] < 1 {\n\t\t\treturn nil, errors.New(\"invalid protocol response\")\n\t\t}\n\n\t\t// Read list of supported versions (which is header[2] bytes long)\n\t\tversionList := make([]byte, header[2])\n\t\tif _, err := io.ReadAtLeast(conn, versionList, len(versionList)); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tselectedVersion := byte(0xff)\n\t\tfor _, v := range versionList {\n\t\t\tif v == 0x01 {\n\t\t\t\tselectedVersion = v\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\n\t\tif n, err := conn.Write([]byte{selectedVersion}); err != nil || n < 1 {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tif selectedVersion == 0xff {\n\t\t\treturn nil, errors.New(\"no supported protocol version\")\n\t\t}\n\t}\n\n\toc := new(OpenConnection)\n\toc.Init(outbound, conn)\n\treturn oc, nil\n}", "func (r *Distribution) HttpVersion() pulumi.StringOutput {\n\treturn (pulumi.StringOutput)(r.s.State[\"httpVersion\"])\n}", "func (_Contract *ContractSession) Version() ([4]byte, error) {\n\treturn _Contract.Contract.Version(&_Contract.CallOpts)\n}", "func (s *DatabaseServerV3) GetVersion() string {\n\treturn s.Version\n}", "func (m *ServerContext) Protocol() protocol.Client {\n\treturn m.ProtocolClient\n}", "func (o *StoragePhysicalDisk) GetProtocolOk() (*string, bool) {\n\tif o == nil || o.Protocol == nil {\n\t\treturn nil, false\n\t}\n\treturn o.Protocol, true\n}", "func (net *NetAPI) Version() (string, error) {\n\treq := net.requestManager.newRequest(\"net_version\")\n\tresp, err := net.requestManager.send(req)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn resp.Get(\"result\").(string), nil\n}" ]
[ "0.73106056", "0.7196808", "0.66493696", "0.64935774", "0.64898866", "0.63805115", "0.6236101", "0.6215604", "0.6141964", "0.6042616", "0.597559", "0.59633476", "0.5883343", "0.58783805", "0.58353376", "0.5822402", "0.58108544", "0.5782251", "0.57608676", "0.572073", "0.56976837", "0.5690141", "0.5678664", "0.55862623", "0.5533395", "0.5531711", "0.54903305", "0.5388527", "0.53788996", "0.53723353", "0.5311066", "0.52918935", "0.52918935", "0.52214044", "0.52110565", "0.5202939", "0.5185434", "0.5185297", "0.5169815", "0.5166881", "0.51663214", "0.5144287", "0.51434433", "0.51429707", "0.51337147", "0.51269567", "0.5104845", "0.5095082", "0.5084592", "0.5027558", "0.5017131", "0.50036234", "0.50002795", "0.4985906", "0.49590725", "0.49551594", "0.4938731", "0.49383083", "0.49235827", "0.4922043", "0.4911511", "0.48848745", "0.48580685", "0.48463234", "0.48457375", "0.4841821", "0.4835039", "0.4830986", "0.48038372", "0.47884694", "0.47865143", "0.4778741", "0.47747862", "0.4774359", "0.47732392", "0.47603542", "0.47561124", "0.47549626", "0.4748194", "0.47456196", "0.4739182", "0.4737791", "0.47306016", "0.47293985", "0.4727109", "0.47255823", "0.4712661", "0.47092924", "0.46998286", "0.46894598", "0.46768713", "0.46719772", "0.46719772", "0.46686852", "0.46436217", "0.46358496", "0.46322647", "0.46311402", "0.46252495", "0.46175316" ]
0.64399546
5
Unidirectional returns whether transport is unidirectional.
func (t *grpcTransport) Unidirectional() bool { return true }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (t *websocketTransport) Unidirectional() bool {\n\treturn false\n}", "func (p *peer) IsOutbound() bool {\r\n\treturn p.peerConn.outbound\r\n}", "func (e *Event) IsDownstream() bool {\n\treturn int32(e.GetType())&int32(EVENT_TYPE_DOWNSTREAM) != 0\n}", "func (t *Transport) Able() bool {\n\treturn t.blob.Able() && t.pubsub.Able()\n}", "func (me TPubStatusUnion) IsReceived() bool { return me.String() == \"received\" }", "func (mts *StandardTurndownStrategy) IsReversible() bool {\n\treturn true\n}", "func (r *Route) local() bool {\n\treturn r.Loop() == PacketLoop || r.outgoingNIC.IsLoopback()\n}", "func (e *Edge) IsDirectional() bool {\n\treturn e.isDirectional\n}", "func (p *TBufferedReadTransport) IsOpen() bool {\n\treturn true\n}", "func (h Handler) isBidirectionalStream(req *http.Request, res *http.Response) bool {\n\t// We have to check the encoding here; only flush headers with identity encoding.\n\t// Non-identity encoding might combine with \"encode\" directive, and in that case,\n\t// if body size larger than enc.MinLength, upper level encode handle might have\n\t// Content-Encoding header to write.\n\t// (see https://github.com/caddyserver/caddy/issues/3606 for use case)\n\tae := req.Header.Get(\"Accept-Encoding\")\n\n\treturn req.ProtoMajor == 2 &&\n\t\tres.ProtoMajor == 2 &&\n\t\tres.ContentLength == -1 &&\n\t\t(ae == \"identity\" || ae == \"\")\n}", "func (r *Network) Internal() pulumi.BoolOutput {\n\treturn (pulumi.BoolOutput)(r.s.State[\"internal\"])\n}", "func (m MessageType) Direct() bool {\n\treturn !m.Broadcast()\n}", "func (mts *MasterlessTurndownStrategy) IsReversible() bool {\n\treturn false\n}", "func (d *EmulatedBTPeerDevice) SupportsTransportMethodLE() bool {\n\treturn d.SupportsTransportMethod(cbt.TransportMethodLE)\n}", "func (m *Myself) IsRelayed() bool {\n\tm.mutex.RLock()\n\tdefer m.mutex.RUnlock()\n\treturn m.status == Port0 && m.relayServer != nil\n}", "func (p *peer) IsPersistent() bool {\r\n\treturn p.peerConn.persistent\r\n}", "func (noopt *NoiseTransport) Compatable(routerInfo router_info.RouterInfo) bool {\n\t_, ok := noopt.peerConnections[routerInfo.IdentHash()]\n\treturn ok\n}", "func (t *T) Reversible() bool { return reversible(t.lhs.Binds(), t.rhs.Binds()) }", "func (c *SubChannel) Isolated() bool {\n\tc.RLock()\n\tdefer c.RUnlock()\n\treturn c.topChannel.Peers() != c.peers\n}", "func (p *Peer) Inbound() bool {\n\treturn p.inbound\n}", "func (socket *MockSocket) Tethering() *socket.TetheringProtocol {\n\treturn socket.tethering\n}", "func (o *FabricFlowControlPolicyAllOf) HasReceiveDirection() bool {\n\tif o != nil && o.ReceiveDirection != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (e *Event) IsUpstream() bool {\n\treturn int32(e.GetType())&int32(EVENT_TYPE_UPSTREAM) != 0\n}", "func (me TxsdComponentTransferFunctionAttributesType) IsLinear() bool { return me.String() == \"linear\" }", "func (rtr router) Sealed() bool {\n\treturn rtr.sealed\n}", "func (msg *Message) IsSync() bool {\n\treturn msg.Header.Sync\n}", "func (s *Server) IsLocal() bool {\n\treturn s.ForwardAddr != \"\"\n}", "func (domain Domain) IsTCP() bool {\n\treturn domain.RouterGroupType == constant.TCPRouterGroup\n}", "func (rt *recvTxOut) Change() bool {\n\treturn rt.change\n}", "func (pvc ProtocolVersionAndCommand) IsUnspec() bool {\n\treturn !(pvc.IsLocal() || pvc.IsProxy())\n}", "func (cp *CustomerOrder) IsReceived() bool {\n\treturn cp.state == RECEIVED\n}", "func (obj *httpTransport) Verify() bool {\n\treturn obj.verify\n}", "func (st *fakeConn) IsReadOnly() bool {\n\treturn st.readOnly\n}", "func (info *BaseEndpointInfo) GetIsLocal() bool {\n\treturn info.IsLocal\n}", "func (transporter *IPCTransporter) IsConnected() bool {\n\treturn true\n}", "func (c *Client) Uniswap() *uniswap.Client { return c.uc }", "func (d *DHCPv4) IsUnicast() bool {\n\treturn d.Flags&0x8000 == 0\n}", "func (f HeaderFlags) Unsynchronisation() bool {\n\treturn (f & 128) > 0\n}", "func (n UTSMode) IsPrivate() bool {\n\treturn !n.IsHost()\n}", "func UnidirectionalEdgeIsValid(edge H3Index) bool {\n\treturn C.h3UnidirectionalEdgeIsValid(edge) == 1\n}", "func UnidirectionalEdgeIsValid(edge H3Index) bool {\n\treturn C.h3UnidirectionalEdgeIsValid(edge) == 1\n}", "func (peer *Peer) IsOutBound() bool {\n\treturn peer.outBound.Load().(bool)\n}", "func (so *SocketOptions) GetReceiveTOS() bool {\n\treturn so.receiveTOSEnabled.Load() != 0\n}", "func (o *FabricFlowControlPolicyAllOf) GetReceiveDirectionOk() (*string, bool) {\n\tif o == nil || o.ReceiveDirection == nil {\n\t\treturn nil, false\n\t}\n\treturn o.ReceiveDirection, true\n}", "func (h HTTPTransport) TLSEnabled() bool {\n\treturn h.TLS != nil\n}", "func (me TxsdPresentationAttributesTextContentElementsUnicodeBidi) IsNormal() bool {\n\treturn me.String() == \"normal\"\n}", "func (pvc ProtocolVersionAndCommand) IsLocal() bool {\n\treturn 0x20 == pvc&0xF0 && 0x00 == pvc&0x0F\n}", "func (n NetworkMode) IsPrivate() bool {\n\treturn !(n.IsHost() || n.IsContainer())\n}", "func (d *Descriptor) Alternate() bool {\n\treturn d.alternate\n}", "func (transporter *Transporter) IsConnected() bool {\n\treturn transporter.isConnected && transporter.isHandling && !transporter.isConnecting\n}", "func (me TxsdPresentationAttributesTextContentElementsUnicodeBidi) IsBidiOverride() bool {\n\treturn me.String() == \"bidi-override\"\n}", "func (obj *content) IsTor() bool {\n\treturn obj.tor != nil\n}", "func (p *protocol) NonBlocking() bool {\n\treturn false\n}", "func (packet *ReliablePacket) IsOrdered() bool {\n\treturn packet.Reliability == UnreliableSequenced || packet.Reliability == ReliableSequenced || packet.Reliability == ReliableOrdered\n}", "func (d *Departure)IsForeignLine()bool{\n\treturn d.ForeignLine==TRUE\n}", "func (socket *MockSocket) Accessibility() *socket.AccessibilityProtocol {\n\treturn socket.accessibility\n}", "func _TransactionDirectionNoOp() {\n\tvar x [1]struct{}\n\t_ = x[Inbound-(1)]\n\t_ = x[Outbound-(2)]\n}", "func (h *Handler) IsTLS() bool {\n\treturn h.encrypted\n}", "func (s DDLStrategy) IsDirect() bool {\n\tswitch s {\n\tcase DDLStrategyVitess, DDLStrategyOnline, DDLStrategyGhost, DDLStrategyPTOSC, DDLStrategyMySQL:\n\t\treturn false\n\t}\n\treturn true\n}", "func (b *Blob) IsForeignLayer() bool {\n\treturn b.ContentType == schema2.MediaTypeForeignLayer\n}", "func (e *connectionedEndpoint) isBound() bool {\n\treturn e.path != \"\" && e.acceptedChan == nil\n}", "func (t *DNSOverTCPTransport) Network() string {\n\treturn t.network\n}", "func (packet *ReliablePacket) IsReliable() bool {\n\treturn packet.Reliability == Reliable || packet.Reliability == ReliableSequenced || packet.Reliability == ReliableOrdered\n}", "func (p *TSocket) IsOpen() bool {\n\tif p.conn == nil {\n\t\treturn false\n\t}\n\treturn true\n}", "func (me TNotificationTransport) IsSoap() bool { return me.String() == \"SOAP\" }", "func (n IpcMode) IsPrivate() bool {\n\treturn n == IPCModePrivate\n}", "func (me TxsdActuate) IsOnRequest() bool { return me == \"onRequest\" }", "func (E_OpenconfigInterfaces_Interfaces_Interface_Ethernet_State_NegotiatedDuplexMode) IsYANGGoEnum() {}", "func (E_OpenconfigInterfaces_Interfaces_Interface_Ethernet_State_NegotiatedDuplexMode) IsYANGGoEnum() {}", "func canReceive(inputType0 reflect.Type) bool {\n\treturn inputType0.ChanDir()&reflect.RecvDir == reflect.RecvDir\n}", "func (c *connAttrs) Legacy() bool { c.mu.RLock(); defer c.mu.RUnlock(); return c._legacy }", "func (op *ReverseReplicationDirectionOperation) Done() bool {\n\treturn op.lro.Done()\n}", "func (me TxsdPaymentMechanism) IsTr() bool { return me.String() == \"TR\" }", "func (me TxsdFeTurbulenceTypeType) IsTurbulence() bool { return me.String() == \"turbulence\" }", "func (p *Piece) IsUnspecified() bool {\n\tretVal := p.pcolor == 0 || p.ptype == 0\n\treturn (retVal)\n}", "func (this *IoHandlerImp) IsOutBound() bool {\n\treturn this.boundType&OutBound != 0\n}", "func (o *FabricFlowControlPolicyAllOf) GetReceiveDirection() string {\n\tif o == nil || o.ReceiveDirection == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.ReceiveDirection\n}", "func (s *State) IsRemote() bool {\n\tif s == nil {\n\t\treturn false\n\t}\n\ts.Lock()\n\tdefer s.Unlock()\n\n\tif s.Remote == nil {\n\t\treturn false\n\t}\n\tif s.Remote.Type == \"\" {\n\t\treturn false\n\t}\n\n\treturn true\n}", "func (m *InternalUpstreamTransport) Validate() error {\n\treturn m.validate(false)\n}", "func (so *SocketOptions) GetReceiveTTL() bool {\n\treturn so.receiveTTLEnabled.Load() != 0\n}", "func (m VolumeStoreMode) IsRemote() bool {\n\treturn (m & RemoteStore) != 0\n}", "func (c *Channel) CanTransfer() bool {\n\treturn channeltype.CanTransferMap[c.State]\n}", "func (ifce *Interface) IsTUN() bool {\n\treturn !ifce.isTAP\n}", "func (ifce *Interface) IsTUN() bool {\n\treturn !ifce.isTAP\n}", "func (l *RelayDriver) State() bool {\n\treturn l.high\n}", "func (this *Device) SupportsNetwork() bool {\n return this.Agent.SupportsNetwork()\n}", "func receive(relayName string) bool {\n\trelay := GetRelay(relayName)\n\tif relay != nil {\n\t\tresp, err := relay.receive()\n\t\tif err != nil {\n\t\t\tdlog.Println(err)\n\t\t\treturn false\n\t\t} else {\n\t\t\tdefer resp.Body.Close()\n\n\t\t\tbody, _ := ioutil.ReadAll(resp.Body)\n\t\t\t// dlog.Printf(\"%s\\n\", body)\n\n\t\t\treturn relay.processReceiveResult(body)\n\t\t}\n\t}\n\n\treturn false\n}", "func CanTransact(side gctorder.Side) bool {\n\treturn side.IsLong() || side.IsShort() || side == gctorder.ClosePosition\n}", "func (o *out) IsOpen() bool {\n\treturn o.stream != nil\n}", "func (v Validator) IsUnbonding() bool {\n\treturn v.GetStatus().Equal(sdk.Unbonding)\n}", "func (_UniswapV2Router02 *UniswapV2Router02TransactorSession) Receive() (*types.Transaction, error) {\r\n\treturn _UniswapV2Router02.Contract.Receive(&_UniswapV2Router02.TransactOpts)\r\n}", "func (m *Message) IsForwarded() bool {\n\treturn m.OriginalSender != nil || m.OriginalChat != nil\n}", "func (me TxsdPaymentMechanism) IsCh() bool { return me.String() == \"CH\" }", "func (me TactionType) IsBlockNetwork() bool { return me.String() == \"block-network\" }", "func (a Entity) Switched() bool {\n\treturn a.switched\n}", "func (peer *Peer) IsPersistent() bool {\n\tpeer.lock.RLock()\n\tdefer peer.lock.RUnlock()\n\treturn peer.persistent\n}", "func (me TAttlistSupplMeshNameType) IsProtocol() bool { return me.String() == \"Protocol\" }", "func (w *Writer) GetTransferSyntax() (binary.ByteOrder, bool) {\n\treturn w.bo, w.implicit\n}", "func (a CollateralAuction) IsReversePhase() bool {\n\treturn a.Bid.IsEqual(a.MaxBid)\n}", "func (pipe *PipeWS) LocalProtocol() uint16 {\n\treturn pipe.proto.Self\n}" ]
[ "0.8311506", "0.56543714", "0.5520934", "0.54517925", "0.5419014", "0.5353592", "0.5339711", "0.52754325", "0.52446", "0.51986665", "0.519469", "0.511227", "0.50982237", "0.50707126", "0.4984875", "0.4965163", "0.49610353", "0.49557832", "0.49073792", "0.48993012", "0.48911664", "0.48453683", "0.4829797", "0.48244095", "0.47962108", "0.4795552", "0.4793332", "0.47873685", "0.47822604", "0.47781205", "0.47716007", "0.4765645", "0.47602126", "0.47601613", "0.4758885", "0.4753785", "0.47505984", "0.47504288", "0.4740086", "0.47358325", "0.47358325", "0.47036612", "0.4703572", "0.47006032", "0.4691753", "0.46754563", "0.46751815", "0.46742976", "0.4663666", "0.46634385", "0.46579054", "0.46429226", "0.46248144", "0.46229106", "0.4622035", "0.46220136", "0.46202123", "0.4619664", "0.46187544", "0.46047318", "0.45934954", "0.4592085", "0.45793912", "0.4577883", "0.4553373", "0.45505023", "0.45493793", "0.45401376", "0.45401376", "0.45395797", "0.4538185", "0.45288044", "0.45251936", "0.45246327", "0.45200273", "0.4513311", "0.45039153", "0.4496184", "0.44951", "0.4490588", "0.44890243", "0.44886902", "0.44858846", "0.44858846", "0.44815707", "0.4478059", "0.4477125", "0.4476746", "0.4469794", "0.4461916", "0.4450007", "0.4449488", "0.44436544", "0.44433784", "0.44429642", "0.44415677", "0.44357443", "0.44326907", "0.44250605", "0.44226864" ]
0.81860495
1
Build factory a new Device App only with mongoClient
func Build(mongoClient *mongo.Client) *App { return &App{ mongoClient: mongoClient, } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func New(jwtMiddleware *jwt.JWTMiddleware, mongoClient *mongo.Client) *App {\n\tapp := new(App)\n\tapp.jwtMiddleware = jwtMiddleware\n\tapp.mongoClient = mongoClient\n\n\t_, err := caclient.GetDefaultCAClient()\n\tif err != nil {\n\t\tif err, ok := err.(*caclient.ClientError); ok {\n\t\t\tif err.Code != caclient.ErrorNotConfig {\n\t\t\t\tlog.Fatalf(\"Error loading caclient. Error Code: %d -- %s\", err.Code, err.Error())\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t}\n\n\tcollection := app.mongoClient.Database(utils.MongoDb).Collection(\"pantahub_devices\")\n\n\tCreateIndexesOptions := options.CreateIndexesOptions{}\n\tCreateIndexesOptions.SetMaxTime(CreateIndexTimeout)\n\n\tindexOptions := options.IndexOptions{}\n\tindexOptions.SetUnique(true)\n\tindexOptions.SetSparse(false)\n\tindexOptions.SetBackground(true)\n\n\tindex := mongo.IndexModel{\n\t\tKeys: bsonx.Doc{\n\t\t\t{Key: \"owner\", Value: bsonx.Int32(1)},\n\t\t\t{Key: \"nick\", Value: bsonx.Int32(1)},\n\t\t},\n\t\tOptions: &indexOptions,\n\t}\n\t_, err = collection.Indexes().CreateOne(context.Background(), index, &CreateIndexesOptions)\n\tif err != nil {\n\t\tlog.Fatalln(\"Error setting up index for pantahub_devices: \" + err.Error())\n\t\treturn nil\n\t}\n\n\tCreateIndexesOptions = options.CreateIndexesOptions{}\n\tCreateIndexesOptions.SetMaxTime(CreateIndexTimeout)\n\n\tindexOptions = options.IndexOptions{}\n\tindexOptions.SetUnique(false)\n\tindexOptions.SetSparse(false)\n\tindexOptions.SetBackground(true)\n\n\tindex = mongo.IndexModel{\n\t\tKeys: bsonx.Doc{\n\t\t\t{Key: \"timemodified\", Value: bsonx.Int32(1)},\n\t\t},\n\t\tOptions: &indexOptions,\n\t}\n\tcollection = app.mongoClient.Database(utils.MongoDb).Collection(\"pantahub_devices\")\n\t_, err = collection.Indexes().CreateOne(context.Background(), index, &CreateIndexesOptions)\n\tif err != nil {\n\t\tlog.Fatalln(\"Error setting up index for pantahub_devices: \" + err.Error())\n\t\treturn nil\n\t}\n\n\tCreateIndexesOptions = options.CreateIndexesOptions{}\n\tCreateIndexesOptions.SetMaxTime(CreateIndexTimeout)\n\n\tindexOptions = options.IndexOptions{}\n\tindexOptions.SetUnique(false)\n\tindexOptions.SetSparse(false)\n\tindexOptions.SetBackground(true)\n\n\tindex = mongo.IndexModel{\n\t\tKeys: bsonx.Doc{\n\t\t\t{Key: \"prn\", Value: bsonx.Int32(1)},\n\t\t},\n\t\tOptions: &indexOptions,\n\t}\n\tcollection = app.mongoClient.Database(utils.MongoDb).Collection(\"pantahub_devices\")\n\t_, err = collection.Indexes().CreateOne(context.Background(), index, &CreateIndexesOptions)\n\tif err != nil {\n\t\tlog.Fatalln(\"Error setting up index for pantahub_devices: \" + err.Error())\n\t\treturn nil\n\t}\n\t// Indexing for the owner,garbage fields\n\tCreateIndexesOptions = options.CreateIndexesOptions{}\n\tCreateIndexesOptions.SetMaxTime(CreateIndexTimeout)\n\n\tindexOptions = options.IndexOptions{}\n\tindexOptions.SetUnique(false)\n\tindexOptions.SetSparse(false)\n\tindexOptions.SetBackground(true)\n\n\tindex = mongo.IndexModel{\n\t\tKeys: bsonx.Doc{\n\t\t\t{Key: \"owner\", Value: bsonx.Int32(1)},\n\t\t\t{Key: \"garbage\", Value: bsonx.Int32(1)},\n\t\t},\n\t\tOptions: &indexOptions,\n\t}\n\tcollection = app.mongoClient.Database(utils.MongoDb).Collection(\"pantahub_devices\")\n\t_, err = collection.Indexes().CreateOne(context.Background(), index, &CreateIndexesOptions)\n\tif err != nil {\n\t\tlog.Fatalln(\"Error setting up index for pantahub_devices: \" + err.Error())\n\t\treturn nil\n\t}\n\t// Indexing for the device,garbage fields\n\tCreateIndexesOptions = options.CreateIndexesOptions{}\n\tCreateIndexesOptions.SetMaxTime(CreateIndexTimeout)\n\n\tindexOptions = options.IndexOptions{}\n\tindexOptions.SetUnique(false)\n\tindexOptions.SetSparse(false)\n\tindexOptions.SetBackground(true)\n\n\tindex = mongo.IndexModel{\n\t\tKeys: bsonx.Doc{\n\t\t\t{Key: \"device\", Value: bsonx.Int32(1)},\n\t\t\t{Key: \"garbage\", Value: bsonx.Int32(1)},\n\t\t},\n\t\tOptions: &indexOptions,\n\t}\n\tcollection = app.mongoClient.Database(utils.MongoDb).Collection(\"pantahub_devices\")\n\t_, err = collection.Indexes().CreateOne(context.Background(), index, &CreateIndexesOptions)\n\tif err != nil {\n\t\tlog.Fatalln(\"Error setting up index for pantahub_devices: \" + err.Error())\n\t\treturn nil\n\t}\n\n\terr = app.EnsureTokenIndices()\n\tif err != nil {\n\t\tlog.Println(\"Error creating indices for pantahub devices tokens: \" + err.Error())\n\t\treturn nil\n\t}\n\n\tapp.API = rest.NewApi()\n\t// we dont use default stack because we dont want content type enforcement\n\tapp.API.Use(&rest.AccessLogJsonMiddleware{Logger: log.New(os.Stdout,\n\t\t\"/devices:\", log.Lshortfile)})\n\tapp.API.Use(&utils.AccessLogFluentMiddleware{Prefix: \"devices\"})\n\tapp.API.Use(&rest.StatusMiddleware{})\n\tapp.API.Use(&rest.TimerMiddleware{})\n\tapp.API.Use(&metrics.Middleware{})\n\n\tapp.API.Use(rest.DefaultCommonStack...)\n\tapp.API.Use(&rest.CorsMiddleware{\n\t\tRejectNonCorsRequests: false,\n\t\tOriginValidator: func(origin string, request *rest.Request) bool {\n\t\t\treturn true\n\t\t},\n\t\tAllowedMethods: []string{\"GET\", \"POST\", \"PUT\", \"PATCH\", \"DELETE\", \"OPTIONS\"},\n\t\tAllowedHeaders: []string{\n\t\t\t\"Accept\", \"Content-Type\", \"X-Custom-Header\", \"Origin\", \"Authorization\"},\n\t\tAccessControlAllowCredentials: true,\n\t\tAccessControlMaxAge: 3600,\n\t})\n\n\tapp.API.Use(&rest.IfMiddleware{\n\t\tCondition: func(request *rest.Request) bool {\n\t\t\t// if call is coming with authorization attempt, ensure JWT middleware\n\t\t\t// is used... otherwise let through anonymous POST for registration\n\t\t\tauth := request.Header.Get(\"Authorization\")\n\t\t\tif auth != \"\" && strings.HasPrefix(strings.ToLower(strings.TrimSpace(auth)), \"bearer \") {\n\t\t\t\treturn true\n\t\t\t}\n\n\t\t\t// post new device means to register... allow this unauthenticated\n\t\t\treturn !((request.Method == \"POST\" && request.URL.Path == \"/\") ||\n\t\t\t\t(request.Method == \"POST\" && request.URL.Path == \"/register\"))\n\t\t},\n\t\tIfTrue: app.jwtMiddleware,\n\t})\n\tapp.API.Use(&rest.IfMiddleware{\n\t\tCondition: func(request *rest.Request) bool {\n\t\t\t// if call is coming with authorization attempt, ensure JWT middleware\n\t\t\t// is used... otherwise let through anonymous POST for registration\n\t\t\tauth := request.Header.Get(\"Authorization\")\n\t\t\tif auth != \"\" && strings.HasPrefix(strings.ToLower(strings.TrimSpace(auth)), \"bearer \") {\n\t\t\t\treturn true\n\t\t\t}\n\n\t\t\t// post new device means to register... allow this unauthenticated\n\t\t\treturn !((request.Method == \"POST\" && request.URL.Path == \"/\") ||\n\t\t\t\t(request.Method == \"POST\" && request.URL.Path == \"/register\"))\n\t\t},\n\t\tIfTrue: &utils.AuthMiddleware{},\n\t})\n\n\twriteDevicesScopes := []utils.Scope{\n\t\tutils.Scopes.API,\n\t\tutils.Scopes.Devices,\n\t\tutils.Scopes.WriteDevices,\n\t}\n\treadDevicesScopes := []utils.Scope{\n\t\tutils.Scopes.API,\n\t\tutils.Scopes.Devices,\n\t\tutils.Scopes.ReadDevices,\n\t}\n\tupdateDevicesScopes := []utils.Scope{\n\t\tutils.Scopes.API,\n\t\tutils.Scopes.Devices,\n\t\tutils.Scopes.UpdateDevices,\n\t}\n\n\t// /auth_status endpoints\n\tapiRouter, _ := rest.MakeRouter(\n\t\t// TPM auto enroll register\n\t\trest.Post(\"/register\", app.handleRegister),\n\n\t\t// token api\n\t\trest.Post(\"/tokens\", utils.ScopeFilter(readDevicesScopes, app.handlePostTokens)),\n\t\trest.Delete(\"/tokens/#id\", utils.ScopeFilter(updateDevicesScopes, app.handleDisableTokens)),\n\t\trest.Get(\"/tokens\", utils.ScopeFilter(readDevicesScopes, app.handleGetTokens)),\n\n\t\t// default api\n\t\trest.Get(\"/auth_status\", utils.ScopeFilter(readDevicesScopes, handleAuth)),\n\t\trest.Get(\"/\", utils.ScopeFilter(readDevicesScopes, app.handleGetDevices)),\n\t\trest.Post(\"/\", utils.ScopeFilter(writeDevicesScopes,\n\t\t\tfunc(writer rest.ResponseWriter, request *rest.Request) {\n\t\t\t\tuserAgent := request.Header.Get(\"User-Agent\")\n\t\t\t\tif userAgent == \"\" {\n\t\t\t\t\tutils.RestErrorWrapperUser(writer, \"No Access (DOS) - no UserAgent\", \"Incompatible Client; upgrade pantavisor\", http.StatusForbidden)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tapp.handlePostDevice(writer, request)\n\t\t\t})),\n\t\trest.Get(\"/#id\", utils.ScopeFilter(readDevicesScopes, app.handleGetDevice)),\n\t\trest.Put(\"/#id\", utils.ScopeFilter(writeDevicesScopes, app.handlePutDevice)),\n\t\trest.Patch(\"/#id\", utils.ScopeFilter(writeDevicesScopes, app.handlePatchDevice)),\n\t\trest.Put(\"/#id/public\", utils.ScopeFilter(writeDevicesScopes, app.handlePutPublic)),\n\t\trest.Delete(\"/#id/public\", utils.ScopeFilter(writeDevicesScopes, app.handleDeletePublic)),\n\t\trest.Put(\"/#id/user-meta\", utils.ScopeFilter(writeDevicesScopes, app.handlePutUserData)),\n\t\trest.Patch(\"/#id/user-meta\", utils.ScopeFilter(writeDevicesScopes, app.handlePatchUserData)),\n\t\trest.Put(\"/#id/device-meta\", utils.ScopeFilter(writeDevicesScopes, app.handlePutDeviceData)),\n\t\trest.Patch(\"/#id/device-meta\", utils.ScopeFilter(writeDevicesScopes, app.handlePatchDeviceData)),\n\t\trest.Delete(\"/#id\", utils.ScopeFilter(writeDevicesScopes, app.handleDeleteDevice)),\n\t\t// lookup by nick-path (np)\n\t\trest.Get(\"/np/#usernick/#devicenick\", utils.ScopeFilter(readDevicesScopes, app.handleGetUserDevice)),\n\t)\n\tapp.API.Use(&tracer.OtelMiddleware{\n\t\tServiceName: os.Getenv(\"OTEL_SERVICE_NAME\"),\n\t\tRouter: apiRouter,\n\t})\n\tapp.API.SetApp(apiRouter)\n\n\treturn app\n}", "func Factory(changeStore *store.ChangeStore, configStore *store.ConfigurationStore, deviceStore *topocache.DeviceStore,\n\ttopoChannel <-chan events.TopoEvent, opStateChan chan<- events.OperationalStateEvent,\n\terrChan chan<- events.DeviceResponse, dispatcher *dispatcher.Dispatcher,\n\treadOnlyPaths map[string]modelregistry.ReadOnlyPathMap, operationalStateCache map[topocache.ID]change.TypedValueMap) {\n\tfor topoEvent := range topoChannel {\n\t\tdeviceName := topocache.ID(events.Event(topoEvent).Subject())\n\t\tif !dispatcher.HasListener(deviceName) && topoEvent.Connect() {\n\t\t\tconfigChan, respChan, err := dispatcher.RegisterDevice(deviceName)\n\t\t\tif err != nil {\n\t\t\t\tlog.Error(err)\n\t\t\t}\n\t\t\tdevice := deviceStore.Store[topocache.ID(deviceName)]\n\t\t\tctx := context.Background()\n\t\t\tcompleteID := utils.ToConfigName(deviceName, device.SoftwareVersion)\n\t\t\tcfg := configStore.Store[store.ConfigName(completeID)]\n\t\t\tmodelName := utils.ToModelName(cfg.Type, device.SoftwareVersion)\n\t\t\tmReadOnlyPaths, ok := readOnlyPaths[modelName]\n\t\t\tif !ok {\n\t\t\t\tlog.Warningf(\"Cannot check for read only paths for target %s with %s because \"+\n\t\t\t\t\t\"Model Plugin not available - continuing\", deviceName, device.SoftwareVersion)\n\t\t\t}\n\t\t\toperationalStateCache[deviceName] = make(change.TypedValueMap)\n\t\t\tsync, err := New(ctx, changeStore, configStore, &device, configChan, opStateChan,\n\t\t\t\terrChan, operationalStateCache[deviceName], mReadOnlyPaths)\n\t\t\tif err != nil {\n\t\t\t\tlog.Error(\"Error in connecting to client: \", err)\n\t\t\t\terrChan <- events.CreateErrorEventNoChangeID(events.EventTypeErrorDeviceConnect,\n\t\t\t\t\tstring(deviceName), err)\n\t\t\t\t//unregistering the listener for changes to the device\n\t\t\t\tunregErr := dispatcher.UnregisterDevice(deviceName)\n\t\t\t\tif unregErr != nil {\n\t\t\t\t\terrChan <- events.CreateErrorEventNoChangeID(events.EventTypeErrorDeviceDisconnect,\n\t\t\t\t\t\tstring(deviceName), unregErr)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\t//spawning two go routines to propagate changes and to get operational state\n\t\t\t\tgo sync.syncConfigEventsToDevice(respChan)\n\t\t\t\tgo sync.syncOperationalState(errChan)\n\t\t\t\t//respChan <- events.CreateConnectedEvent(events.EventTypeDeviceConnected, string(deviceName))\n\t\t\t}\n\t\t} else if dispatcher.HasListener(deviceName) && !topoEvent.Connect() {\n\n\t\t\terr := dispatcher.UnregisterDevice(deviceName)\n\t\t\tif err != nil {\n\t\t\t\tlog.Error(err)\n\t\t\t\t//TODO evaluate if fall through without upstreaming\n\t\t\t\t//errChan <- events.CreateErrorEventNoChangeID(events.EventTypeErrorDeviceDisconnect,\n\t\t\t\t//\tstring(deviceName), err)\n\t\t\t}\n\t\t}\n\t}\n}", "func newApp() (*app, error) {\n\tapp := &app{}\n\n\tgd, err := newGDrive(\"\")\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Unable initializate Google Drive configuration: %v\", err)\n\t}\n\tapp.gDrive = gd\n\n\tsrv, err := app.gDrive.getService()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Unable to retrieve Drive client: %v\", err)\n\t}\n\tapp.srv = srv\n\n\treturn app, nil\n}", "func buildMongoClient(prefix string, value []byte) (*ClientAndOpts, error) {\n\tif len(value) == 0 {\n\t\treturn nil, errors.New(\"watch时读取json出错,prefix=\" + prefix)\n\t}\n\n\tvar mcfg mongoCfg\n\terr := json.Unmarshal(value, &mcfg)\n\tif err != nil {\n\t\treturn nil, errors.New(\"json unmarshal, error=\" + err.Error())\n\t}\n\tif mcfg.Uri == \"\" {\n\t\treturn nil, errors.New(\"未找到 uri 配置, \" + prefix + \"无效\")\n\t}\n\n\tnewOpts := options.Client().ApplyURI(mcfg.Uri).SetAuth(options.Credential{Username: mcfg.Username, Password: mcfg.Password})\n\tnewMongoClient, err := mongo.Connect(context.TODO(), newOpts)\n\tif err != nil {\n\t\treturn nil, errors.New(\"Mongo 连接异常:\" + err.Error())\n\t}\n\n\treturn newMongoClientAndOpts(newMongoClient, newOpts), nil\n}", "func makeApp(def pchannel.App, data pchannel.Data) perun.App {\n\treturn perun.App{\n\t\tDef: def,\n\t\tData: data,\n\t}\n}", "func New(c *conf.Config) (dao *Dao) {\n\tdao = &Dao{\n\t\tc: c,\n\t\tRoomApi: room_api.New(getConf(\"room\")),\n\t\tUserApi: user_api.New(getConf(\"user\")),\n\t\tRelationApi: relation_api.New(getConf(\"relation\")),\n\t\tFansMedalApi: fans_medal_api.New(getConf(\"fans_medal\")),\n\t\tHttpCli: bm.NewClient(c.HttpClient),\n\t}\n\tMemberCli, err := member_cli.NewClient(c.GrpcCli)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tdao.memberCli = MemberCli\n\n\tTitansCli, errTitans := resource_cli.NewClient(c.GrpcCli)\n\tif errTitans != nil {\n\t\tpanic(err)\n\t}\n\tdao.titansCli = TitansCli\n\treturn\n}", "func (c Client) createApp(body io.Reader) (*App, error) {\n\treq, err := http.NewRequest(\"POST\", c.getURL(\"/apps\"), body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar marshalled map[string]*App\n\terr = c.executeRequestAndMarshal(req, &marshalled)\n\treturn marshalled[\"app\"], err\n}", "func App() *buffalo.App {\n\t// The Seed string needs to be length 32\n\t// TODO: get the seed from env variable\n\tprivateKey = ed25519.NewKeyFromSeed([]byte(\"b{2'*&-kjECuLynMZaE7@f:yzD}$MND?\"))\n\tpublicKey = (privateKey.Public()).(ed25519.PublicKey)\n\t/*var err error\n\tpublicKey, privateKey, err = ed25519.GenerateKey(nil)\n\tif err != nil {\n\t\tlog.Fatal(\"Could not generate public- and privateKey: \", err.Error())\n\t}*/\n\tif app == nil {\n\t\tapp = buffalo.New(buffalo.Options{\n\t\t\tEnv: ENV,\n\t\t\tSessionName: \"_notification_server_session\",\n\t\t})\n\n\t\tvar connErr error\n\t\tnsdb, connErr = connDB()\n\t\tif connErr != nil {\n\t\t\tfmt.Println(\"Error: \" + connErr.Error())\n\t\t}\n\n\t\tvar bucketErr error\n\t\tnsBucket, bucketErr = getBucket(nsdb, \"NotificationServer\")\n\t\tif bucketErr != nil {\n\t\t\tfmt.Println(\"Error: \" + bucketErr.Error())\n\t\t}\n\n\t\t// Automatically redirect to SSL\n\t\tapp.Use(forceSSL())\n\n\t\t// Log request parameters (filters apply).\n\t\tapp.Use(paramlogger.ParameterLogger)\n\n\t\t// Protect against CSRF attacks. https://www.owasp.org/index.php/Cross-Site_Request_Forgery_(CSRF)\n\t\t// Remove to disable this.\n\t\t//app.Use(csrf.New)\n\n\t\t// Setup and use translations:\n\t\tapp.Use(translations())\n\n\t\t// API Version 1\n\t\tv1 := app.Group(\"/api/v1\")\n\n\t\t// Ressources\n\t\tusersr := &UsersResource{}\n\t\tudr := &UserDevicesResource{}\n\n\t\t// Services\n\t\tfor name, provider := range services.Providers {\n\t\t\tp, ok := provider.(services.Webhook)\n\t\t\tif ok {\n\t\t\t\tv1.GET(\"/services/\" + name + \"/{botToken}\", p.WebhookHandler)\n\t\t\t\tv1.POST(\"/services/\" + name + \"/{botToken}\", p.WebhookHandler)\n\t\t\t}\n\t\t}\n\n\t\tv1.GET(\"/\", HomeHandler)\n\n\t\tv1.POST(\"/auth\", authHandler)\n\n\t\tv1.GET(\"/users\", usersr.Show)\n\t\tv1.PUT(\"/users\", usersr.Update)\n\t\tv1.POST(\"/users\", usersr.Add)\n\t\tv1.DELETE(\"/users\", usersr.Delete)\n\t\t\n\t\tv1.GET(\"/user/{apikey:[a-z0-9]+}/devices\", udr.Show)\n\t\tv1.PUT(\"/user/{apikey:[a-z0-9]+}/devices\", udr.Update)\n\t\tv1.POST(\"/user/{apikey:[a-z0-9]+}/devices\", udr.Add)\n\t\tv1.DELETE(\"/user/{apikey:[a-z0-9]+}/devices\", udr.Delete)\n\n\t\tv1.POST(\"/user/{apikey:[a-z0-9]+}/sendMessage\", sendMessageHandler)\n\t\tv1.POST(\"/user/{apikey:[a-z0-9]+}/sendmessage\", sendMessageHandler)\n\n\t\tv1.ServeFiles(\"/\", assetsBox) // serve files from the public directory\n\t}\n\n\treturn app\n}", "func newApp(name string) (app *App, err error) {\n\tapp = &App{\n\t\tName: name,\n\t\tID: uuid.NewV5(namespace, \"org.homealone.\"+name).String(),\n\t\thandler: make(map[queue.Topic]message.Handler),\n\t\tdebug: *debug,\n\t\tfilterMessages: true,\n\t}\n\tapp.Log = log.NewLogger().With(log.Fields{\"app\": name, \"id\": app.ID})\n\treturn app, errors.Wrap(err, \"newApp failed\")\n}", "func CreateApplication() *Alpha {\n app := &Alpha{}\n app.Request = &Request{}\n app.Response = &Response{}\n app.init()\n return app\n}", "func NewApp(ctx context.Context, appCfg AppConfig) (*API, error) {\n\tappCfg.checkConfig()\n\n\tlog.Debug().Interface(\"api app config\", appCfg).Msg(\"starting initialize api application\")\n\n\te := echo.New()\n\n\ta := &API{\n\t\te: e,\n\t\taddr: appCfg.NetInterface,\n\t}\n\n\te.Use(func(next echo.HandlerFunc) echo.HandlerFunc {\n\t\treturn func(c echo.Context) error {\n\t\t\tcc := &Context{\n\t\t\t\tContext: c,\n\t\t\t\tCtx: ctx,\n\t\t\t}\n\t\t\treturn next(cc)\n\t\t}\n\t})\n\te.Validator = &Validator{validator: validator.New()}\n\te.Use(logMiddleware)\n\n\te.GET(\"/healthcheck\", a.handleHealthcheck)\n\n\tg := e.Group(\"/api\")\n\tg.GET(\"/:collection/documents\", a.handleSearch)\n\tg.POST(\"/:collection/documents\", a.handleAddDocuments)\n\n\tlog.Debug().Msg(\"endpoints registered\")\n\n\treturn a, nil\n}", "func Build() {\n\tonce.Do(func() {\n\t\tdb, client := connect()\n\t\tinstance = MongoConnection{\n\t\t\tDb: db,\n\t\t\tClient: client,\n\t\t}\n\t})\n}", "func NewApp(dbname, dburl, slackSetLocationToken, slackWhereIsToken, slackReqToken string) *App {\n\ta := App{}\n\n\ta.Router = gin.Default()\n\n\ta.SlackSetLocationToken = slackSetLocationToken\n\n\ta.SlackWhereIsToken = slackWhereIsToken\n\n\ta.SlackRequestToken = slackReqToken\n\n\ta.initialiseRoutes()\n\n\tdialled := false\n\tcount := 1\n\n\tfor dialled == false {\n\t\tfmt.Printf(\"Connecting to database, attempt %v\\n\", count)\n\t\ts, err := mgo.Dial(dburl)\n\n\t\tif err != nil {\n\t\t\tfmt.Println(\"Error: \" + err.Error())\n\t\t\tfmt.Println(\"Retrying...\")\n\t\t} else {\n\t\t\tfmt.Println(\"Connected!\")\n\t\t\tdialled = true\n\t\t\ta.DB = s.DB(dbname)\n\t\t}\n\n\t\ttime.Sleep(2 * time.Second)\n\n\t\tif count > 5 {\n\t\t\tfmt.Printf(\"Tried to connect %v times. Exiting program\\n\", count)\n\t\t\tos.Exit(1)\n\t\t}\n\t\tcount++\n\t}\n\n\treturn &a\n}", "func newDiscoveryApp(img string) *discovery.App {\n\tapp, err := discovery.NewAppFromString(img)\n\tif err != nil {\n\t\treturn nil\n\t}\n\tu, err := url.Parse(app.Name.String())\n\tif err != nil || u.Scheme != \"\" {\n\t\treturn nil\n\t}\n\tif _, ok := app.Labels[\"arch\"]; !ok {\n\t\tapp.Labels[\"arch\"] = defaultArch\n\t}\n\tif _, ok := app.Labels[\"os\"]; !ok {\n\t\tapp.Labels[\"os\"] = defaultOS\n\t}\n\treturn app\n}", "func AppCreate(ctx *Context, name string, typ DeviceType) (*Application, error) {\n\th := authHeader(ctx.Config.AuthToken)\n\turi := ctx.Config.APIEndpoint(\"application\")\n\tdata := make(map[string]interface{})\n\tdata[\"app_name\"] = name\n\tdata[\"device_type\"] = typ.String()\n\tbody, err := marhsalReader(data)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tb, err := doJSON(ctx, \"POST\", uri, h, nil, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\trst := &Application{}\n\terr = json.Unmarshal(b, rst)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn rst, nil\n}", "func NewStore(\n\tkubeconfig *rest.Config,\n\tclientset kubernetes.Interface,\n\tkatoClient katoversioned.Interface,\n\tdbmanager db.Manager,\n\tconf option.Config) Storer {\n\tctx, cancel := context.WithCancel(context.Background())\n\tstore := &appRuntimeStore{\n\t\tkubeconfig: kubeconfig,\n\t\tclientset: clientset,\n\t\tkatoClient: katoClient,\n\t\tctx: ctx,\n\t\tcancel: cancel,\n\t\tinformers: &Informer{CRS: make(map[string]cache.SharedIndexInformer)},\n\t\tlisters: &Lister{},\n\t\tappServices: sync.Map{},\n\t\tconf: conf,\n\t\tdbmanager: dbmanager,\n\t\tcrClients: make(map[string]interface{}),\n\t\tresourceCache: NewResourceCache(),\n\t\tpodUpdateListeners: make(map[string]chan<- *corev1.Pod, 1),\n\t\tvolumeTypeListeners: make(map[string]chan<- *model.TenantServiceVolumeType, 1),\n\t}\n\tcrdClient, err := internalclientset.NewForConfig(kubeconfig)\n\tif err != nil {\n\t\tlogrus.Errorf(\"create crd client failure %s\", err.Error())\n\t}\n\tif crdClient != nil {\n\t\tstore.crdClient = crdClient\n\t\tcrdFactory := internalinformers.NewSharedInformerFactory(crdClient, 5*time.Minute)\n\t\tstore.informers.CRD = crdFactory.Apiextensions().V1().CustomResourceDefinitions().Informer()\n\t\tstore.listers.CRD = crdFactory.Apiextensions().V1().CustomResourceDefinitions().Lister()\n\t}\n\n\t// create informers factory, enable and assign required informers\n\tinfFactory := informers.NewSharedInformerFactoryWithOptions(conf.KubeClient, 10*time.Second,\n\t\tinformers.WithNamespace(corev1.NamespaceAll))\n\n\tstore.informers.Namespace = infFactory.Core().V1().Namespaces().Informer()\n\n\tstore.informers.Deployment = infFactory.Apps().V1().Deployments().Informer()\n\tstore.listers.Deployment = infFactory.Apps().V1().Deployments().Lister()\n\n\tstore.informers.StatefulSet = infFactory.Apps().V1().StatefulSets().Informer()\n\tstore.listers.StatefulSet = infFactory.Apps().V1().StatefulSets().Lister()\n\n\tstore.informers.Service = infFactory.Core().V1().Services().Informer()\n\tstore.listers.Service = infFactory.Core().V1().Services().Lister()\n\n\tstore.informers.Pod = infFactory.Core().V1().Pods().Informer()\n\tstore.listers.Pod = infFactory.Core().V1().Pods().Lister()\n\n\tstore.informers.Secret = infFactory.Core().V1().Secrets().Informer()\n\tstore.listers.Secret = infFactory.Core().V1().Secrets().Lister()\n\n\tstore.informers.ConfigMap = infFactory.Core().V1().ConfigMaps().Informer()\n\tstore.listers.ConfigMap = infFactory.Core().V1().ConfigMaps().Lister()\n\n\tstore.informers.Ingress = infFactory.Extensions().V1beta1().Ingresses().Informer()\n\tstore.listers.Ingress = infFactory.Networking().V1().Ingresses().Lister()\n\n\tstore.informers.ReplicaSet = infFactory.Apps().V1().ReplicaSets().Informer()\n\tstore.listers.ReplicaSets = infFactory.Apps().V1().ReplicaSets().Lister()\n\n\tstore.informers.Endpoints = infFactory.Core().V1().Endpoints().Informer()\n\tstore.listers.Endpoints = infFactory.Core().V1().Endpoints().Lister()\n\n\tstore.informers.Nodes = infFactory.Core().V1().Nodes().Informer()\n\tstore.listers.Nodes = infFactory.Core().V1().Nodes().Lister()\n\n\tstore.informers.StorageClass = infFactory.Storage().V1().StorageClasses().Informer()\n\tstore.listers.StorageClass = infFactory.Storage().V1().StorageClasses().Lister()\n\n\tstore.informers.Claims = infFactory.Core().V1().PersistentVolumeClaims().Informer()\n\tstore.listers.Claims = infFactory.Core().V1().PersistentVolumeClaims().Lister()\n\n\tstore.informers.Events = infFactory.Core().V1().Events().Informer()\n\n\tstore.informers.HorizontalPodAutoscaler = infFactory.Autoscaling().V2beta2().HorizontalPodAutoscalers().Informer()\n\tstore.listers.HorizontalPodAutoscaler = infFactory.Autoscaling().V2beta2().HorizontalPodAutoscalers().Lister()\n\n\t// kato custom resource\n\tkatoInformer := externalversions.NewSharedInformerFactoryWithOptions(katoClient, 10*time.Second,\n\t\texternalversions.WithNamespace(corev1.NamespaceAll))\n\tstore.listers.HelmApp = katoInformer.Kato().V1alpha1().HelmApps().Lister()\n\tstore.informers.HelmApp = katoInformer.Kato().V1alpha1().HelmApps().Informer()\n\tstore.listers.ThirdComponent = katoInformer.Kato().V1alpha1().ThirdComponents().Lister()\n\tstore.informers.ThirdComponent = katoInformer.Kato().V1alpha1().ThirdComponents().Informer()\n\tstore.listers.ComponentDefinition = katoInformer.Kato().V1alpha1().ComponentDefinitions().Lister()\n\tstore.informers.ComponentDefinition = katoInformer.Kato().V1alpha1().ComponentDefinitions().Informer()\n\tstore.informers.ComponentDefinition.AddEventHandlerWithResyncPeriod(componentdefinition.GetComponentDefinitionBuilder(), time.Second*300)\n\n\t// Endpoint Event Handler\n\tepEventHandler := cache.ResourceEventHandlerFuncs{\n\t\tAddFunc: func(obj interface{}) {\n\t\t\tep := obj.(*corev1.Endpoints)\n\n\t\t\tserviceID := ep.Labels[\"service_id\"]\n\t\t\tversion := ep.Labels[\"version\"]\n\t\t\tcreaterID := ep.Labels[\"creater_id\"]\n\t\t\tif serviceID != \"\" && createrID != \"\" {\n\t\t\t\tappservice, err := store.getAppService(serviceID, version, createrID, true)\n\t\t\t\tif err == conversion.ErrServiceNotFound {\n\t\t\t\t\tlogrus.Debugf(\"ServiceID: %s; Action: AddFunc; service not found\", serviceID)\n\t\t\t\t}\n\t\t\t\tif appservice != nil {\n\t\t\t\t\tappservice.AddEndpoints(ep)\n\t\t\t\t}\n\t\t\t}\n\t\t},\n\t\tDeleteFunc: func(obj interface{}) {\n\t\t\tep := obj.(*corev1.Endpoints)\n\t\t\tserviceID := ep.Labels[\"service_id\"]\n\t\t\tversion := ep.Labels[\"version\"]\n\t\t\tcreaterID := ep.Labels[\"creater_id\"]\n\t\t\tif serviceID != \"\" && createrID != \"\" {\n\t\t\t\tappservice, _ := store.getAppService(serviceID, version, createrID, false)\n\t\t\t\tif appservice != nil {\n\t\t\t\t\tappservice.DelEndpoints(ep)\n\t\t\t\t\tif appservice.IsClosed() {\n\t\t\t\t\t\tlogrus.Debugf(\"ServiceID: %s; Action: DeleteFunc;service is closed\", serviceID)\n\t\t\t\t\t\tstore.DeleteAppService(appservice)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t},\n\t\tUpdateFunc: func(old, cur interface{}) {\n\t\t\tcep := cur.(*corev1.Endpoints)\n\n\t\t\tserviceID := cep.Labels[\"service_id\"]\n\t\t\tversion := cep.Labels[\"version\"]\n\t\t\tcreaterID := cep.Labels[\"creater_id\"]\n\t\t\tif serviceID != \"\" && createrID != \"\" {\n\t\t\t\tappservice, err := store.getAppService(serviceID, version, createrID, true)\n\t\t\t\tif err == conversion.ErrServiceNotFound {\n\t\t\t\t\tlogrus.Debugf(\"ServiceID: %s; Action: UpdateFunc; service not found\", serviceID)\n\t\t\t\t}\n\t\t\t\tif appservice != nil {\n\t\t\t\t\tappservice.AddEndpoints(cep)\n\t\t\t\t}\n\t\t\t}\n\t\t},\n\t}\n\n\tstore.informers.Namespace.AddEventHandler(store.nsEventHandler())\n\tstore.informers.Deployment.AddEventHandlerWithResyncPeriod(store, time.Second*10)\n\tstore.informers.StatefulSet.AddEventHandlerWithResyncPeriod(store, time.Second*10)\n\tstore.informers.Pod.AddEventHandlerWithResyncPeriod(store.podEventHandler(), time.Second*10)\n\tstore.informers.Secret.AddEventHandlerWithResyncPeriod(store, time.Second*10)\n\tstore.informers.Service.AddEventHandlerWithResyncPeriod(store, time.Second*10)\n\tstore.informers.Ingress.AddEventHandlerWithResyncPeriod(store, time.Second*10)\n\tstore.informers.ConfigMap.AddEventHandlerWithResyncPeriod(store, time.Second*10)\n\tstore.informers.ReplicaSet.AddEventHandlerWithResyncPeriod(store, time.Second*10)\n\tstore.informers.Endpoints.AddEventHandlerWithResyncPeriod(epEventHandler, time.Second*10)\n\tstore.informers.Nodes.AddEventHandlerWithResyncPeriod(store, time.Second*10)\n\tstore.informers.StorageClass.AddEventHandlerWithResyncPeriod(store, time.Second*300)\n\tstore.informers.Claims.AddEventHandlerWithResyncPeriod(store, time.Second*10)\n\tstore.informers.Events.AddEventHandlerWithResyncPeriod(store.evtEventHandler(), time.Second*10)\n\tstore.informers.HorizontalPodAutoscaler.AddEventHandlerWithResyncPeriod(store, time.Second*10)\n\tstore.informers.ThirdComponent.AddEventHandlerWithResyncPeriod(store, time.Second*10)\n\n\treturn store\n}", "func CreateApp(conf *config.Config) *TelegramBotApp {\n\n\t// Init the telegram bot\n\tbot, err := botFactory.BotFactory(conf)\n\tif err != nil {\n\t\tlog.Fatal(\"Can't create a telegram bot: \", err)\n\t}\n\n\t// Init the database\n\tdb, err := dbFactory.DatabaseFactory(conf)\n\tif err != nil {\n\t\tlog.Fatal(\"Couldn't create database: \", err)\n\t}\n\n\t// Init the database storage\n\tstorage, err := storages.CreateStorages(db)\n\tif err != nil {\n\t\tlog.Fatal(\"Couldn't setup database: \", err)\n\t}\n\n\t// Init the database Repository\n\trepository, err := repositories.CreateRepository(db)\n\tif err != nil {\n\t\tlog.Fatal(\"Couldn't setup database: \", err)\n\t}\n\n\treturn &TelegramBotApp{conf, bot, storage, repository}\n}", "func buildFakeClient(objs []runtime.Object) *ReconcilePerconaServerMongoDB {\n\ts := scheme.Scheme\n\n\ts.AddKnownTypes(api.SchemeGroupVersion, &api.PerconaServerMongoDB{})\n\n\tcl := fake.NewFakeClientWithScheme(s, objs...)\n\n\treturn &ReconcilePerconaServerMongoDB{client: cl, scheme: s}\n}", "func New(db *db.DB, gameService game.Service, userService user.Service) *App {\n\treturn &App{\n\t\tdb: db,\n\t\tgameService: gameService,\n\t\tuserService: userService,\n\t}\n}", "func NewAppCollection(repository *storage.Apps) (*AppCollection, error) {\n instance := new(AppCollection)\n instance.apps = make(map[string]IApp)\n instance.repository = repository\n\n var err error\n\n if instance.repository != nil {\n err = instance.loadAll(repository.All())\n } else {\n err = errors.New(collection.ERR_NO_REPOSITORY)\n }\n\n return instance, err\n}", "func New(config AppConfig, logger *logrus.Logger) App {\n\t// Open CSV File\n\tcsvFile, err := os.OpenFile(config.CSVFile, os.O_RDWR|os.O_CREATE, 0644)\n\tif err != nil {\n\t\tlogger.Fatal(err)\n\t}\n\n\t// Set client options\n\tclientOptions := options.Client().ApplyURI(config.MongoURI)\n\n\t// Context with timeout for establish connection with Mongo Atlas\n\tctx, cancel := context.WithTimeout(context.Background(), mongoTimeout)\n\tdefer cancel()\n\n\t// Connect to Mongo Atlas\n\tclient, err := mongo.Connect(ctx, clientOptions)\n\tif err != nil {\n\t\tlogger.Fatal(err)\n\t}\n\n\t// Init eventsRepository\n\teventsRepo := axiom.NewEventsRepo(config.MongoDB, client)\n\n\t// Init reservationsRepository\n\treservationsRepo := axiom.NewReservationRepo(config.MongoDB, client)\n\n\t// Init reservationsCache\n\treservationsCache := local.NewReservationRepo(csvFile)\n\n\t// Init usecase\n\tevents := usecase.NewEventUseCase(eventsRepo, reservationsRepo, reservationsCache)\n\n\t// Init controller\n\tcontroller := controller.NewEventController(events)\n\n\t// Setup router\n\trouter := router.New(controller)\n\n\treturn App{\n\t\tlogger: logger,\n\t\trouter: router.Router(),\n\t\tconfig: config,\n\t}\n}", "func New(c *deis.Client, appID string) (api.App, error) {\n\tbody := []byte{}\n\n\tif appID != \"\" {\n\t\treq := api.AppCreateRequest{ID: appID}\n\t\tb, err := json.Marshal(req)\n\n\t\tif err != nil {\n\t\t\treturn api.App{}, err\n\t\t}\n\t\tbody = b\n\t}\n\n\tres, reqErr := c.Request(\"POST\", \"/v2/apps/\", body)\n\tif reqErr != nil && !deis.IsErrAPIMismatch(reqErr) {\n\t\treturn api.App{}, reqErr\n\t}\n\tdefer res.Body.Close()\n\n\tapp := api.App{}\n\tif err := json.NewDecoder(res.Body).Decode(&app); err != nil {\n\t\treturn api.App{}, err\n\t}\n\n\treturn app, reqErr\n}", "func createApp() *menta.MentaApp {\n\t// runs tendermint init if needed\n\tmenta.InitTendermint(homeDir)\n\t// setup the app\n\tapp := menta.NewApp(\"counter-example\", homeDir)\n\t// Register the service\n\tapp.AddService(counter.Service{})\n\n\treturn app\n}", "func CreateApp(debug bool, config *Config, swaggerInfo *swag.Spec) *App {\n\tif !config.Static {\n\t\tf, err := os.Open(config.DocFile)\n\t\tif err != nil {\n\t\t\tlog.Warnf(\"open doc file error: %s\", err)\n\t\t\tlog.Infof(\"fallback to static mode\")\n\t\t\tconfig.Static = true\n\t\t}\n\t\tf.Close()\n\t}\n\n\tif config.Static {\n\t\tif swaggerInfo != nil {\n\t\t\tswaggerInfo.Host = config.Host\n\t\t}\n\t\tif _, err := swag.ReadDoc(); err != nil {\n\t\t\tpanic(\"no swagger registered, can't use static mode\")\n\t\t}\n\t}\n\n\tapp := &App{config, echo.New()}\n\n\tapp.e.Debug = debug\n\n\tapp.e.Use(middleware.Logger())\n\tapp.e.Use(middleware.Recover())\n\tapp.e.Use(middleware.CORS())\n\n\t// routers\n\tapp.e.GET(\"/\", func(c echo.Context) error {\n\t\treturn c.Redirect(http.StatusMovedPermanently, config.SwaggerPath+\"index.html\")\n\t})\n\n\tg := app.e.Group(\"\")\n\tRegister(g, config)\n\n\treturn app\n}", "func NewApp(\n\tlogger logur.Logger,\n\tpublisher message.Publisher,\n\terrorHandler emperror.Handler,\n) (http.Handler, func(*grpc.Server)) {\n\tcommonLogger := commonadapter.NewContextAwareLogger(logger, &correlation.ContextExtractor{})\n\n\tvar todoList tododriver.TodoList\n\t{\n\t\teventBus, _ := cqrs.NewEventBus(\n\t\t\tpublisher,\n\t\t\tfunc(eventName string) string { return todoTopic },\n\t\t\tcqrs.JSONMarshaler{GenerateName: cqrs.StructName},\n\t\t)\n\t\ttodoList = todo.NewList(\n\t\t\tulidgen.NewGenerator(),\n\t\t\ttodo.NewInmemoryStore(),\n\t\t\ttodoadapter.NewEventDispatcher(eventBus),\n\t\t)\n\t\tlogger := commonLogger.WithFields(map[string]interface{}{\"module\": \"todo\"})\n\t\ttodoList = tododriver.LoggingMiddleware(logger)(todoList)\n\t\ttodoList = tododriver.InstrumentationMiddleware()(todoList)\n\t}\n\n\ttodoListEndpoint := tododriver.MakeEndpoints(todoList)\n\n\trouter := mux.NewRouter()\n\trouter.Use(ocmux.Middleware())\n\trouter.Use(correlation.HTTPMiddleware(ulidgen.NewGenerator()))\n\n\trouter.Path(\"/\").Methods(\"GET\").Handler(landingdriver.NewHTTPHandler())\n\trouter.PathPrefix(\"/todos\").Handler(tododriver.MakeHTTPHandler(todoListEndpoint, errorHandler))\n\trouter.PathPrefix(\"/graphql\").Handler(tododriver.MakeGraphQLHandler(todoListEndpoint, errorHandler))\n\trouter.PathPrefix(\"/httpbin\").Handler(http.StripPrefix(\n\t\t\"/httpbin\",\n\t\thttpbin.MakeHTTPHandler(commonLogger.WithFields(map[string]interface{}{\"module\": \"httpbin\"})),\n\t))\n\n\treturn router, func(s *grpc.Server) {\n\t\ttodov1beta1.RegisterTodoListServer(s, tododriver.MakeGRPCServer(todoListEndpoint, errorHandler))\n\t}\n}", "func NewMgmtApp(config *Config) (*MgmtApp, error) {\n\n c := &MgmtApp{}\n c.userAuthenticator = config.UserAuthenticator\n c.handleCommand = config.HandleCommand\n c.notifyClientAuthenticated = config.NotifyClientAuthenticated\n c.notifyClientAuthenticatedFailed = config.NotifyClientAuthenticatedFailed\n c.welcomeUser = config.WelcomeUser\n c.unregisterUser = config.UnregisterUser\n c.defaultPrompt = config.DefaultPrompt\n\n c.clientInitializer = config.ClientInitializer\n\n if c.userAuthenticator == nil {\n c.userAuthenticator = c.defaultUserAuthenticator\n }\n if c.handleCommand == nil {\n c.handleCommand = HandleCommands(DefaultCommands)\n }\n if c.notifyClientAuthenticated == nil {\n c.notifyClientAuthenticated = c.defaultNotifyClientAuthenticated\n }\n if c.notifyClientAuthenticatedFailed == nil {\n c.notifyClientAuthenticatedFailed = c.defaultNotifyClientAuthenticatedFailed\n }\n if c.welcomeUser == nil {\n c.welcomeUser = c.defaultWelcomeUser\n }\n\n if c.unregisterUser == nil {\n c.unregisterUser = c.defaultUnregisterUser\n }\n\n if c.clientInitializer == nil {\n c.clientInitializer = c.defaultClientInitializer\n }\n\n if c.defaultPrompt == \"\" {\n c.defaultPrompt = \"$\"\n }\n\n c.hub = newHub()\n go c.hub.run()\n return c, nil\n}", "func New(c *conf.Config) (d *Dao) {\n\td = &Dao{\n\t\t// http client\n\t\tclient: bm.NewClient(c.HTTPAudio),\n\t\taudioByCidsURL: c.Host.APICo + _audioByCids,\n\t}\n\treturn\n}", "func (cf *couchdbFactory) Build(c container.Container, dsc *config.DataStoreConfig) (DataStoreInterface, error) {\n\tlogger.Log.Debug(\"couchdbFactory\")\n\tkey := dsc.Code\n\n\t//if it is already in container, return\n\tif value, found := c.Get(key); found {\n\t\tlogger.Log.Debug(\"found couchdb in container for key:\", key)\n\t\treturn value.(*kivik.DB), nil\n\t}\n\t// Don't know why needs adding the following line, because the driver is already registered in init() in couchdbKiv\n\t// however, not adding this, I got the error \"unknown driver \"couch\" (forgotten import?)\"\n\tkivik.Register(config.COUCHDB, &couchdbKivid.Couch{})\n\n\tclient, err := kivik.New(context.TODO(), dsc.Code, dsc.UrlAddress)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"\")\n\t}\n\tdb, err := client.DB(context.TODO(), dsc.DbName)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"\")\n\t}\n\tc.Put(key, db)\n\treturn db, nil\n\n}", "func NewApp(databaseName, password string) *App {\n\t// Setup db:\n\tdb, err := newDatabase(databaseName, password)\n\tif err != nil {\n\t\tif err, ok := err.(*mysql.MySQLError); ok && err.Number == 1049 { // unknown database\n\t\t\tfmt.Printf(\"Creating and reconnecting to %s\", databaseName)\n\t\t\t// Create and use `ElasticJury`\n\t\t\tdbRoot, err := newDatabase(\"\", password) // as root\n\t\t\tif err != nil {\n\t\t\t\tpanic(err)\n\t\t\t}\n\t\t\tdbRoot.mustExec(fmt.Sprintf(\"CREATE DATABASE IF NOT EXISTS %s DEFAULT CHARACTER SET utf8\", databaseName))\n\t\t\tdb, err = newDatabase(databaseName, password)\n\t\t\tif err != nil {\n\t\t\t\tpanic(err)\n\t\t\t}\n\t\t\tdb.mustExecScriptFile(InitTableScriptPath)\n\t\t} else {\n\t\t\tpanic(err) // unknown err\n\t\t}\n\t}\n\tprintln(\"[Info] Database initialized.\")\n\n\t// Setup router:\n\t// Disable Console Color\n\t// gin.DisableConsoleColor()\n\t// Release mode is faster\n\tgin.SetMode(gin.ReleaseMode)\n\trouter := gin.Default()\n\t{\n\t\t// Ping test\n\t\trouter.GET(\"/ping\", func(context *gin.Context) {\n\t\t\tcontext.String(http.StatusOK, \"pong\")\n\t\t})\n\t\t// Retrieve case id by tag, law, judge\n\t\trouter.POST(\"/search\", db.makeSearchHandler())\n\t\t// Retrieve case info by case id\n\t\trouter.POST(\"/info\", db.makeCaseInfoHandler())\n\t\t// Associate\n\t\trouter.GET(\"/associate/:field/:item\", natural.MakeAssociateHandler())\n\t\t// Retrieve case detail by one case id\n\t\trouter.GET(\"/detail/:id\", db.makeCaseDetailHandler())\n\t}\n\tprintln(\"[Info] Search engine initialized.\")\n\n\treturn &App{\n\t\tEngine: router,\n\t\tdb: db,\n\t}\n}", "func (s *server) Create(ctx context.Context, body *pb.RequestBody) (*pb.ResponseBody, error) {\n\tlanguage := body.GetLanguage()\n\tapp := &types.ApplicationConfig{}\n\n\terr := json.Unmarshal(body.GetData(), app)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tuser, err := mongo.FetchSingleUser(body.GetOwner())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tmaxCount := configs.ServiceConfig.AppMaker.AppLimit\n\trateCount := configs.ServiceConfig.RateLimit\n\ttimeInterval := configs.ServiceConfig.RateInterval\n\tif !user.IsAdmin() && maxCount >= 0 {\n\t\trateLimitCount := mongo.CountInstanceInTimeFrame(body.GetOwner(), mongo.AppInstance, timeInterval)\n\t\ttotalCount := mongo.CountInstancesByUser(body.GetOwner(), mongo.AppInstance)\n\t\tif totalCount < maxCount {\n\t\t\tif rateLimitCount >= rateCount && rateCount >= 0 {\n\t\t\t\treturn nil, fmt.Errorf(\"cannot deploy more than %d app instances in %d hours\", rateCount, timeInterval)\n\t\t\t}\n\t\t} else {\n\t\t\treturn nil, fmt.Errorf(\"cannot deploy more than %d app instances\", maxCount)\n\t\t}\n\t}\n\n\tapp.SetLanguage(language)\n\tapp.SetOwner(body.GetOwner())\n\tapp.SetInstanceType(mongo.AppInstance)\n\tapp.SetHostIP(utils.HostIP)\n\tapp.SetNameServers(configs.GasperConfig.DNSServers)\n\tapp.SetDateTime()\n\n\tgendnsNameServers, _ := redis.FetchServiceInstances(types.GenDNS)\n\tfor _, nameServer := range gendnsNameServers {\n\t\tif strings.Contains(nameServer, \":\") {\n\t\t\tapp.AddNameServers(strings.Split(nameServer, \":\")[0])\n\t\t} else {\n\t\t\tutils.LogError(\"AppMaker-Controller-1\", fmt.Errorf(\"GenDNS instance %s is of invalid format\", nameServer))\n\t\t}\n\t}\n\n\tif pipeline[language] == nil {\n\t\treturn nil, fmt.Errorf(\"language `%s` is not supported\", language)\n\t}\n\tresErr := pipeline[language].create(app)\n\tif resErr != nil {\n\t\tif resErr.Message() != \"repository already exists\" && resErr.Message() != \"container already exists\" {\n\t\t\tgo diskCleanup(app.GetName())\n\t\t}\n\t\treturn nil, fmt.Errorf(resErr.Error())\n\t}\n\n\tsshEntrypointIP := configs.ServiceConfig.GenSSH.EntrypointIP\n\tif len(sshEntrypointIP) == 0 {\n\t\tsshEntrypointIP = utils.HostIP\n\t}\n\tapp.SetSSHCmd(configs.ServiceConfig.GenSSH.Port, app.GetName(), sshEntrypointIP)\n\n\tapp.SetAppURL(fmt.Sprintf(\"%s.%s.%s\", app.GetName(), cloudflare.ApplicationInstance, configs.GasperConfig.Domain))\n\n\tif configs.CloudflareConfig.PlugIn {\n\t\tresp, err := cloudflare.CreateApplicationRecord(app.GetName())\n\t\tif err != nil {\n\t\t\tgo diskCleanup(app.GetName())\n\t\t\treturn nil, err\n\t\t}\n\t\tapp.SetCloudflareID(resp.Result.ID)\n\t\tapp.SetPublicIP(configs.CloudflareConfig.PublicIP)\n\t}\n\n\terr = mongo.UpsertInstance(\n\t\ttypes.M{\n\t\t\tmongo.NameKey: app.GetName(),\n\t\t\tmongo.InstanceTypeKey: mongo.AppInstance,\n\t\t}, app)\n\n\tif err != nil && err != mongo.ErrNoDocuments {\n\t\tgo diskCleanup(app.GetName())\n\t\tgo stateCleanup(app.GetName())\n\t\treturn nil, err\n\t}\n\n\terr = redis.RegisterApp(\n\t\tapp.GetName(),\n\t\tfmt.Sprintf(\"%s:%d\", utils.HostIP, configs.ServiceConfig.AppMaker.Port),\n\t\tfmt.Sprintf(\"%s:%d\", utils.HostIP, app.GetContainerPort()),\n\t)\n\n\tif err != nil {\n\t\tgo diskCleanup(app.GetName())\n\t\tgo stateCleanup(app.GetName())\n\t\treturn nil, err\n\t}\n\n\terr = redis.IncrementServiceLoad(\n\t\tServiceName,\n\t\tfmt.Sprintf(\"%s:%d\", utils.HostIP, configs.ServiceConfig.AppMaker.Port),\n\t)\n\n\tif err != nil {\n\t\tgo diskCleanup(app.GetName())\n\t\tgo stateCleanup(app.GetName())\n\t\treturn nil, err\n\t}\n\n\tapp.SetSuccess(true)\n\n\tresponse, err := json.Marshal(app)\n\treturn &pb.ResponseBody{Data: response}, err\n}", "func New(cfg *config.Giffy) (*web.App, error) {\n\n\tif cfg.EncryptionKey == \"\" {\n\t\treturn nil, fmt.Errorf(\"encryption key is unset, cannot continue\")\n\t}\n\n\tlog := logger.MustNew(\n\t\tlogger.OptConfig(cfg.Logger),\n\t)\n\tlog.Enable(core.FlagSearch, core.FlagModeration)\n\tlog.Disable(db.QueryStartFlag)\n\n\tconn, err := db.New(\n\t\tdb.OptConfig(cfg.DB),\n\t\tdb.OptLog(log),\n\t)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif err := conn.Open(); err != nil {\n\t\treturn nil, err\n\t}\n\n\tlog.Infof(\"using service env: %s\", cfg.ServiceEnvOrDefault())\n\tlog.Infof(\"using database: %s\", conn.Config.CreateLoggingDSN())\n\tlog.Infof(\"using admin user email: %s\", cfg.AdminUserEmail)\n\tif cfg.Aws.AccessKeyID != \"\" {\n\t\tlog.Infof(\"using aws access key: %s\", cfg.Aws.AccessKeyID)\n\t} else {\n\t\tlog.Warningf(\"aws access key unset, uploads will fail\")\n\t}\n\tlog.Infof(\"using aws region: %s\", cfg.Aws.RegionOrDefault())\n\tlog.Infof(\"using aws s3 bucket: %s\", cfg.S3Bucket)\n\tlog.Infof(\"using aws cloudfront dns: %s\", cfg.CloudFrontDNS)\n\n\tmgr := &model.Manager{BaseManager: dbutil.NewBaseManager(conn)}\n\n\toauthMgr, err := oauth.New(\n\t\toauth.OptConfig(cfg.GoogleAuth),\n\t)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tapp := web.MustNew(web.OptConfig(cfg.Web), web.OptLog(log))\n\tif cfg.Meta.Version != \"\" {\n\t\tapp.BaseHeaders.Add(\"X-Server-Version\", cfg.Meta.Version)\n\t} else if cfg.Meta.GitRef != \"\" {\n\t\tapp.BaseHeaders.Add(\"X-Server-Version\", cfg.Meta.GitRef)\n\t}\n\n\tlog.Listen(logger.Fatal, \"error-writer\", logger.NewErrorEventListener(func(_ context.Context, ev logger.ErrorEvent) {\n\t\tif req, isReq := ev.State.(*http.Request); isReq {\n\t\t\tmgr.Invoke(context.Background()).Create(model.NewError(ev.Err, req))\n\t\t} else {\n\t\t\tmgr.Invoke(context.Background()).Create(model.NewError(ev.Err, nil))\n\t\t}\n\t}))\n\tlog.Listen(logger.Error, \"error-writer\", logger.NewErrorEventListener(func(_ context.Context, ev logger.ErrorEvent) {\n\t\tif req, isReq := ev.State.(*http.Request); isReq {\n\t\t\tmgr.Invoke(context.Background()).Create(model.NewError(ev.Err, req))\n\t\t} else {\n\t\t\tmgr.Invoke(context.Background()).Create(model.NewError(ev.Err, nil))\n\t\t}\n\t}))\n\tlog.Listen(core.FlagSearch, \"event-writer\", func(_ context.Context, e logger.Event) {\n\t\tif typed, ok := e.(db.DatabaseMapped); ok {\n\t\t\tlogger.MaybeError(log, mgr.Invoke(context.Background()).Create(typed))\n\t\t}\n\t})\n\tlog.Listen(core.FlagModeration, \"event-writer\", func(_ context.Context, e logger.Event) {\n\t\tif typed, ok := e.(db.DatabaseMapped); ok {\n\t\t\tlogger.MaybeError(log, mgr.Invoke(context.Background()).Create(typed))\n\t\t}\n\t})\n\n\tapp.Views.AddPaths(ViewPaths...)\n\n\tfm := filemanager.New(cfg.S3Bucket, cfg.Aws)\n\n\tapp.Register(controller.Index{Log: log, Model: mgr, Config: cfg})\n\tapp.Register(controller.APIs{Log: log, Model: mgr, Config: cfg, Files: fm, OAuth: oauthMgr})\n\tapp.Register(controller.Integrations{Log: log, Model: mgr, Config: cfg})\n\tapp.Register(controller.Auth{Log: log, Model: mgr, Config: cfg, OAuth: oauthMgr})\n\tapp.Register(controller.UploadImage{Log: log, Model: mgr, Config: cfg, Files: fm})\n\tapp.Register(controller.Chart{Log: log, Model: mgr, Config: cfg})\n\n\tif model.Migrations(cfg).Apply(context.Background(), conn); err != nil {\n\t\treturn nil, err\n\t}\n\n\tcron.Default().Log = log.WithPath(\"jobs\")\n\tcron.Default().LoadJobs(jobs.DeleteOrphanedTags{Model: mgr})\n\tcron.Default().LoadJobs(jobs.CleanTagValues{Model: mgr})\n\tcron.Default().LoadJobs(jobs.FixContentRating{Model: mgr})\n\tcron.Default().StartAsync()\n\n\treturn app, nil\n}", "func (b BotApp) construct() BotAppClass { return &b }", "func newMongoDB(addr string, cred *mgo.Credential) (ProductDatabase, error) {\n\tconn, err := mgo.Dial(addr)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"mongo: could not dial: %v\", err)\n\t}\n\n\tif cred != nil {\n\t\tif err := conn.Login(cred); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn &mongoDB{\n\t\tconn: conn,\n\t\tc: conn.DB(\"productshelf\").C(\"products\"),\n\t}, nil\n}", "func (b BotAppNotModified) construct() BotAppClass { return &b }", "func MakeFactory(ctor Ctor, logger *zap.Logger) Factory {\n\treturn func(t *testing.T, r *TableRow) (controller.Reconciler, ActionRecorderList, EventList, *FakeStatsReporter) {\n\t\tls := NewListers(r.Objects)\n\n\t\tctx := context.Background()\n\t\tctx = logging.WithLogger(ctx, logger.Sugar())\n\n\t\tctx, kubeClient := fakekubeclient.With(ctx, ls.GetKubeObjects()...)\n\t\tctx, eventingClient := fakeeventingclient.With(ctx, ls.GetEventingObjects()...)\n\t\tctx, legacy := fakelegacyclient.With(ctx, ls.GetLegacyObjects()...)\n\t\tctx, client := fakeknativekafkaclient.With(ctx, ls.GetKafkaChannelObjects()...)\n\n\t\tdynamicScheme := runtime.NewScheme()\n\t\tfor _, addTo := range clientSetSchemes {\n\t\t\taddTo(dynamicScheme)\n\t\t}\n\n\t\tctx, dynamicClient := fakedynamicclient.With(ctx, dynamicScheme, ls.GetAllObjects()...)\n\n\t\teventRecorder := record.NewFakeRecorder(maxEventBufferSize)\n\t\tctx = controller.WithEventRecorder(ctx, eventRecorder)\n\t\tstatsReporter := &FakeStatsReporter{}\n\n\t\t// Set up our Controller from the fakes.\n\t\tc := ctor(ctx, &ls, configmap.NewStaticWatcher())\n\n\t\tfor _, reactor := range r.WithReactors {\n\t\t\tkubeClient.PrependReactor(\"*\", \"*\", reactor)\n\t\t\tclient.PrependReactor(\"*\", \"*\", reactor)\n\t\t\tlegacy.PrependReactor(\"*\", \"*\", reactor)\n\t\t\tdynamicClient.PrependReactor(\"*\", \"*\", reactor)\n\t\t\teventingClient.PrependReactor(\"*\", \"*\", reactor)\n\t\t}\n\n\t\t// Validate all Create operations through the eventing client.\n\t\tclient.PrependReactor(\"create\", \"*\", func(action clientgotesting.Action) (handled bool, ret runtime.Object, err error) {\n\t\t\treturn ValidateCreates(context.Background(), action)\n\t\t})\n\t\tclient.PrependReactor(\"update\", \"*\", func(action clientgotesting.Action) (handled bool, ret runtime.Object, err error) {\n\t\t\treturn ValidateUpdates(context.Background(), action)\n\t\t})\n\n\t\t// Validate all Create operations through the legacy client.\n\t\tlegacy.PrependReactor(\"create\", \"*\", func(action clientgotesting.Action) (handled bool, ret runtime.Object, err error) {\n\t\t\treturn ValidateCreates(ctx, action)\n\t\t})\n\t\tlegacy.PrependReactor(\"update\", \"*\", func(action clientgotesting.Action) (handled bool, ret runtime.Object, err error) {\n\t\t\treturn ValidateUpdates(ctx, action)\n\t\t})\n\n\t\tactionRecorderList := ActionRecorderList{dynamicClient, client, kubeClient, legacy}\n\t\teventList := EventList{Recorder: eventRecorder}\n\n\t\treturn c, actionRecorderList, eventList, statsReporter\n\t}\n}", "func createAppContext() {\n\t//init data access layer\n\trepo.InitRepo(\n\t\tos.Getenv(\"APP_CONNECTION_POOL_SIZE\"),\n\t\tos.Getenv(\"DB_PORT\"),\n\t\tos.Getenv(\"DB_HOST\"),\n\t\tos.Getenv(\"DB_USER\"),\n\t\tos.Getenv(\"DB_PASSWORD\"),\n\t\tos.Getenv(\"DB_NAME\"))\n\t//init services layer\n\tservice.InitService()\n\t//init HTTP API layer\n\tapi.InitHttp()\n}", "func NewApp(ctx context.Context, registryClient client.RegistryClient, dockerConfig *configuration.Configuration, extraConfig *registryconfig.Configuration, writeLimiter maxconnections.Limiter) http.Handler {\n\tapp := &App{\n\t\tctx: ctx,\n\t\tregistryClient: registryClient,\n\t\tconfig: extraConfig,\n\t\twriteLimiter: writeLimiter,\n\t\tquotaEnforcing: newQuotaEnforcingConfig(ctx, extraConfig.Quota),\n\t\tpaginationCache: kubecache.NewLRUExpireCache(defaultPaginationCacheSize),\n\t}\n\n\tif app.config.Metrics.Enabled {\n\t\tapp.metrics = metrics.NewMetrics(metrics.NewPrometheusSink())\n\t} else {\n\t\tapp.metrics = metrics.NewNoopMetrics()\n\t}\n\n\tcacheTTL := time.Duration(0)\n\tif !app.config.Cache.Disabled {\n\t\tcacheTTL = app.config.Cache.BlobRepositoryTTL\n\t}\n\n\tdigestCache, err := cache.NewBlobDigest(\n\t\tdefaultDescriptorCacheSize,\n\t\tdefaultDigestToRepositoryCacheSize,\n\t\tcacheTTL,\n\t\tapp.metrics,\n\t)\n\tif err != nil {\n\t\tdcontext.GetLogger(ctx).Fatalf(\"unable to create cache: %v\", err)\n\t}\n\tapp.cache = digestCache\n\n\tsuperapp := supermiddleware.App(app)\n\tif am := appMiddlewareFrom(ctx); am != nil {\n\t\tsuperapp = am.Apply(superapp)\n\t}\n\tdockerApp := supermiddleware.NewApp(ctx, dockerConfig, superapp)\n\n\tif app.driver == nil {\n\t\tdcontext.GetLogger(ctx).Fatalf(\"configuration error: the storage driver middleware %q is not activated\", supermiddleware.Name)\n\t}\n\tif app.registry == nil {\n\t\tdcontext.GetLogger(ctx).Fatalf(\"configuration error: the registry middleware %q is not activated\", supermiddleware.Name)\n\t}\n\n\t// Add a token handling endpoint\n\tif dockerConfig.Auth.Type() == supermiddleware.Name {\n\t\ttokenRealm, err := registryconfig.TokenRealm(extraConfig.Auth.TokenRealm)\n\t\tif err != nil {\n\t\t\tdcontext.GetLogger(dockerApp).Fatalf(\"error setting up token auth: %s\", err)\n\t\t}\n\t\terr = dockerApp.NewRoute().Methods(\"GET\").PathPrefix(tokenRealm.Path).Handler(NewTokenHandler(ctx, registryClient)).GetError()\n\t\tif err != nil {\n\t\t\tdcontext.GetLogger(dockerApp).Fatalf(\"error setting up token endpoint at %q: %v\", tokenRealm.Path, err)\n\t\t}\n\t\tdcontext.GetLogger(dockerApp).Debugf(\"configured token endpoint at %q\", tokenRealm.String())\n\t}\n\n\tapp.registerBlobHandler(dockerApp)\n\n\t// Registry extensions endpoint provides extra functionality to handle the image\n\t// signatures.\n\tisImageClient, err := registryClient.Client()\n\tif err != nil {\n\t\tdcontext.GetLogger(dockerApp).Fatalf(\"unable to get client for signatures: %v\", err)\n\t}\n\tRegisterSignatureHandler(dockerApp, isImageClient)\n\n\t// Advertise features supported by OpenShift\n\tif dockerApp.Config.HTTP.Headers == nil {\n\t\tdockerApp.Config.HTTP.Headers = http.Header{}\n\t}\n\tdockerApp.Config.HTTP.Headers.Set(\"X-Registry-Supports-Signatures\", \"1\")\n\n\tdockerApp.RegisterHealthChecks()\n\n\th := http.Handler(dockerApp)\n\n\t// Registry extensions endpoint provides prometheus metrics.\n\tif extraConfig.Metrics.Enabled {\n\t\tRegisterMetricHandler(dockerApp)\n\t\th = promhttp.InstrumentHandlerCounter(metrics.HTTPRequestsTotal, h)\n\t\th = promhttp.InstrumentHandlerDuration(metrics.HTTPRequestDurationSeconds, h)\n\t\th = promhttp.InstrumentHandlerInFlight(metrics.HTTPInFlightRequests, h)\n\t\th = promhttp.InstrumentHandlerRequestSize(metrics.HTTPRequestSizeBytes, h)\n\t\th = promhttp.InstrumentHandlerResponseSize(metrics.HTTPResponseSizeBytes, h)\n\t\th = promhttp.InstrumentHandlerTimeToWriteHeader(metrics.HTTPTimeToWriteHeaderSeconds, h)\n\t}\n\n\tdcontext.GetLogger(dockerApp).Infof(\"Using %q as Docker Registry URL\", extraConfig.Server.Addr)\n\n\treturn h\n}", "func NewApp() App {\n\n\tvar err error\n\tvar app *appmgr\n\n\tconst appname = \"app\"\n\n\tapp = &appmgr{\n\t\tstate: fsm.Undefined,\n\t\tname: appname,\n\t\tprops: make(map[string]map[string]interface{}),\n\t\tdflow: nil,\n\t\tstore: nil,\n\t\tmsg: newMsgStream(\n\t\t\tappname,\n\t\t\tLvlInfo,\n\t\t\t//LvlDebug,\n\t\t\t//LvlError,\n\t\t\tnil,\n\t\t),\n\t\tevtmax: -1,\n\t\tnprocs: -1,\n\t\tcomps: make(map[string]Component),\n\t\ttsks: make([]Task, 0),\n\t\tsvcs: make([]Svc, 0),\n\t}\n\n\tsvc, err := app.New(\"go-hep.org/x/hep/fwk.datastore\", \"evtstore\")\n\tif err != nil {\n\t\tapp.msg.Errorf(\"fwk.NewApp: could not create evtstore: %w\\n\", err)\n\t\treturn nil\n\t}\n\tapp.store = svc.(*datastore)\n\n\terr = app.AddSvc(app.store)\n\tif err != nil {\n\t\tapp.msg.Errorf(\"fwk.NewApp: could not create evtstore: %w\\n\", err)\n\t\treturn nil\n\t}\n\n\tsvc, err = app.New(\"go-hep.org/x/hep/fwk.dflowsvc\", \"dataflow\")\n\tif err != nil {\n\t\tapp.msg.Errorf(\"fwk.NewApp: could not create dataflow svc: %w\\n\", err)\n\t\treturn nil\n\t}\n\tapp.dflow = svc.(*dflowsvc)\n\n\terr = app.AddSvc(app.dflow)\n\tif err != nil {\n\t\tapp.msg.Errorf(\"fwk.NewApp: could not create dataflow svc: %w\\n\", err)\n\t\treturn nil\n\t}\n\n\terr = app.DeclProp(app, \"EvtMax\", &app.evtmax)\n\tif err != nil {\n\t\tapp.msg.Errorf(\"fwk.NewApp: could not declare property 'EvtMax': %w\\n\", err)\n\t\treturn nil\n\t}\n\n\terr = app.DeclProp(app, \"NProcs\", &app.nprocs)\n\tif err != nil {\n\t\tapp.msg.Errorf(\"fwk.NewApp: could not declare property 'NProcs': %w\\n\", err)\n\t\treturn nil\n\t}\n\n\terr = app.DeclProp(app, \"MsgLevel\", &app.msg.lvl)\n\tif err != nil {\n\t\tapp.msg.Errorf(\"fwk.NewApp: could not declare property 'MsgLevel': %w\\n\", err)\n\t\treturn nil\n\t}\n\n\treturn app\n}", "func (p ServiceFactory) New(c config.Config) (service.Service, error) {\n\t// check mongo db configuration\n\thosts := c.MustString(\"mongo\")\n\ttimeout := c.MustInt(\"connect_timeout\")\n\tsession, err := mgo.DialWithTimeout(hosts, time.Duration(timeout)*time.Second)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer session.Close()\n\n\treturn &metadataService{\n\t\tconfig: c,\n\t\twaitgroup: sync.WaitGroup{},\n\t\tquitChan: make(chan interface{}),\n\t\teventChan: make(chan event.Event),\n\t}, nil\n\n}", "func Factory(ctx context.Context, conf *logical.BackendConfig) (logical.Backend, error) {\n\tb := Backend()\n\tif err := b.Setup(ctx, conf); err != nil {\n\t\treturn nil, err\n\t}\n\treturn b, nil\n}", "func CreateDemoDeviceInstance() (*oc.Device, error) {\n\t// Initialize a device.\n\td := &oc.Device{\n\t\tSystem: &oc.System{\n\t\t\tHostname: ygot.String(\"rtr02.pop44\"),\n\t\t\tΛHostname: []ygot.Annotation{\n\t\t\t\t&ExampleAnnotation{ConfigSource: \"devicedemo\"},\n\t\t\t},\n\t\t},\n\t}\n\n\t// Create a new interface under the device. In this case /interfaces/interface\n\t// is the list that is being populated, but due to schema compression the\n\t// 'interfaces' container is not created, making the 'interface' list a top-level\n\t// entity. The New... helper methods are therefore mapped to device.\n\teth0, err := d.NewInterface(\"eth0\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Set some attributes of the interface.\n\teth0.Description = ygot.String(\"Link to rtr01.pop44\")\n\teth0.Type = oc.IETFInterfaces_InterfaceType_ethernetCsmacd\n\n\tif err := addNetworkInstance(d); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Add a component.\n\tc, err := d.NewComponent(\"os\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tc.Type = oc.PlatformTypes_OPENCONFIG_SOFTWARE_COMPONENT_OPERATING_SYSTEM\n\n\t// Create a second device instance, and populate the OS component under\n\t// it. This code demonstrates how ygot.MergeStructs can be used to combine\n\t// multiple instances of the same type of struct together, allowing each\n\t// subtree to be generated in its own context.\n\tsecondDev := &oc.Device{}\n\tsc, err := secondDev.NewComponent(\"os\")\n\tsc.Description = ygot.String(\"RouterOS 14.0\")\n\tmergedDev, err := ygot.MergeStructs(d, secondDev)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// Since ygot.MergeStructs returns an ygot.GoStruct interface, we\n\t// must type assert it back to *oc.Device.\n\treturn mergedDev.(*oc.Device), nil\n}", "func newClientMongoRepository() repository.ClientRepository {\n\tmongoAddr := os.Getenv(\"DATABASE_CONN\")\n\tfmt.Println(\"mongoAddr => \", mongoAddr)\n\tclient := repositoryimpl.Connect(mongoAddr)\n\treturn repositoryimpl.NewRepository(client)\n}", "func CreateGenericDevice(info IDeviceInfo) *Device {\n connectTimeout, _ := time.ParseDuration(agent.DEFAULT_CONNECT_TIMEOUT)\n\n instance := new(Device)\n agent.SetupAgent(&instance.Agent, info, 0, connectTimeout, agent.DEFAULT_BUFFER_SIZE, true)\n\n instance.SetOnMessageReceivedHandler(instance.onMessageReceived)\n instance.info = info\n instance.commandProcessor = func (deviceInfo map[string]string, command api.ICommand) (string, error) {\n return command.PropertyName() + command.PropertyValue(), nil\n }\n instance.queryProcessor = func (deviceInfo map[string]string, query api.IQuery) (string, error) {\n return query.PropertyName(), nil\n }\n\n return instance\n}", "func NewApp(name string, scopes []string, redirectURI string) (g *Gondole, err error) {\n\t// Load configuration, will register if none is found\n\tcnf, err := LoadConfig(name)\n\tif err != nil {\n\t\t// Nothing exist yet\n\t\tcnf := Config{\n\t\t\tDefault: name,\n\t\t}\n\t\terr = cnf.Write()\n\t\tif err != nil {\n\t\t\tlog.Fatalf(\"error: can not write config for %s\", name)\n\t\t}\n\n\t\t// Now register this through OAuth\n\t\tif scopes == nil {\n\t\t\tscopes = ourScopes\n\t\t}\n\n\t\tg, err = registerApplication(name, scopes, redirectURI)\n\n\t} else {\n\t\tg = &Gondole{\n\t\t\tName: cnf.Name,\n\t\t\tID: cnf.ID,\n\t\t\tSecret: cnf.BearerToken,\n\t\t}\n\t}\n\n\treturn\n}", "func NewApp(port, dataFileName string) *App {\n\ta := &App{Port: port, DataFileName: dataFileName}\n\tnoteRepository := repositories.NewNoteRepository(a.DataFileName)\n\ta.NoteController = controllers.NewNoteController(noteRepository)\n\ta.Router = routers.NewRouter(http.NewServeMux(), a.NoteController)\n\n\treturn a\n}", "func newGoFactory() *GOFactory {\n\tgologger.SLogger.Println(\"Init Game Object Factory Singleton\")\n\tfOnce.Do(func() {\n\t\tgofactory = &GOFactory{\n\t\t\tGoCreator: make(map[string]ICreator),\n\t\t}\n\t})\n\treturn gofactory\n}", "func Factory(ctx context.Context, conf *logical.BackendConfig) (logical.Backend, error) {\n\tb := Backend(conf)\n\tif err := b.Setup(ctx, conf); err != nil {\n\t\treturn nil, err\n\t}\n\treturn b, nil\n}", "func Factory(ctx context.Context, conf *logical.BackendConfig) (logical.Backend, error) {\n\tb := Backend(conf)\n\tif err := b.Setup(ctx, conf); err != nil {\n\t\treturn nil, err\n\t}\n\treturn b, nil\n}", "func Create(co ClientOptions) (Client, error) {\n\tclientOptions := options.Client().ApplyURI(co.URI)\n\tclient, err := mongo.Connect(context.TODO(), clientOptions)\n\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t\treturn Client{}, err\n\t}\n\terr = client.Ping(context.TODO(), nil)\n\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t\treturn Client{}, err\n\t}\n\tfmt.Println(\"Connected to MongoDB!\")\n\tdb := client.Database(co.DBName)\n\tcl := Client{DB: db}\n\treturn cl, nil\n}", "func newReconciler(mgr manager.Manager) reconcile.Reconciler {\n\timageClient, err := imagev1.NewForConfig(mgr.GetConfig())\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Error getting image client: %v\", err))\n\t}\n\tbuildClient, err := buildv1.NewForConfig(mgr.GetConfig())\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Error getting build client: %v\", err))\n\t}\n\n\tclient := &kogitocli.Client{\n\t\tControlCli: mgr.GetClient(),\n\t\tBuildCli: buildClient,\n\t\tImageCli: imageClient,\n\t}\n\n\treturn &ReconcileKogitoApp{\n\t\tclient: client,\n\t\tscheme: mgr.GetScheme(),\n\t\tcache: mgr.GetCache(),\n\t}\n}", "func (handler *applicationHandler) create(app *types.ApplicationConfig) types.ResponseError {\n\tapp.SetDockerImage(handler.image)\n\tapp.SetConfGenerator(handler.confGenerator)\n\treturn api.SetupApplication(app)\n}", "func (c *DeviceController) Create(w http.ResponseWriter, r *http.Request) {\n\tfmt.Println(json.NewDecoder(r.Body))\n\tres := db.PutItem(\"home-devices\", NewDevice(\"hello\", \"world\", \"indeed\"))\n\tc.SendJSON(\n\t\tw,\n\t\tr,\n\t\tres,\n\t\thttp.StatusOK,\n\t)\n}", "func New(service string, appConfig interface{}) (*App, error) {\n\t// Get config store\n\tconfigStore, err := config.NewStore(os.Getenv(\"CONFIG_URI\"))\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"config store error: %s\", err)\n\t}\n\n\t// Load config from store\n\tc := &config.Config{App: appConfig}\n\terr = configStore.Load(c)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"cannot load config: %s\", err)\n\t}\n\n\t// Convert potential environment variables\n\tc.Node = config.ValueOf(c.Node)\n\tc.Version = config.ValueOf(c.Version)\n\n\t// Create logger\n\tl, err := logger.New(service, &c.Log)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"logger error: %s\", err)\n\t}\n\n\t// Build stats\n\ts, err := statsAdapter.New(&c.Stats)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"stats error: %s\", err)\n\t}\n\ts.SetLogger(l)\n\n\t// Build app context\n\tctx := app.NewCtx(service, c, l, s)\n\n\t// Build ready cond flag\n\tlock := &sync.Mutex{}\n\tlock.Lock()\n\tready := sync.NewCond(lock)\n\n\t// Build app struct\n\tapp := &App{\n\t\tservice: service,\n\t\tready: ready,\n\t\tctx: ctx,\n\t\tconfig: c,\n\t\thandlers: handler.NewReg(ctx),\n\t\tdone: make(chan bool, 1),\n\t}\n\n\t// Start background services\n\tctx.BG().Dispatch(s)\n\tctx.BG().Dispatch(&hearbeat{app: app})\n\n\t// Trap OS signals\n\tgo trapSignals(app)\n\n\treturn app, nil\n}", "func buildDatastore() (autoupdate.Datastore, error) {\n\tvar f *faker\n\tvar url string\n\tdsService := getEnv(\"DATASTORE\", \"fake\")\n\tswitch dsService {\n\tcase \"fake\":\n\t\tfmt.Println(\"Fake Datastore\")\n\t\tf = newFaker(os.Stdin)\n\t\turl = f.ts.TS.URL\n\n\tcase \"service\":\n\t\thost := getEnv(\"DATASTORE_READER_HOST\", \"localhost\")\n\t\tport := getEnv(\"DATASTORE_READER_PORT\", \"9010\")\n\t\tprotocol := getEnv(\"DATASTORE_READER_PROTOCOL\", \"http\")\n\t\turl = protocol + \"://\" + host + \":\" + port\n\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"unknown datastore %s\", dsService)\n\t}\n\n\tfmt.Println(\"Datastore URL:\", url)\n\treceiver, err := buildReceiver(f)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"build receiver: %w\", err)\n\t}\n\treturn datastore.New(url, receiver), nil\n}", "func buildDatastore() autoupdate.Datastore {\n\tdsService := getEnv(\"DATASTORE\", \"fake\")\n\turl := getEnv(\"DATASTORE_URL\", \"http://localhost:8001\")\n\n\tvar f *faker\n\tif dsService == \"fake\" {\n\t\tfmt.Println(\"Fake Datastore\")\n\t\tf = newFaker(os.Stdin)\n\t\turl = f.ts.TS.URL\n\t}\n\tfmt.Println(\"Datastore URL:\", url)\n\treturn datastore.New(url, buildReceiver(f))\n}", "func New() datastore.Datastore {\n\tctx := context.Background()\n\n\tc, err := fc.New(ctx)\n\tif err != nil {\n\t\tlog.Fatalln(err)\n\t}\n\treturn &client{\n\t\tfirestoreClient: c,\n\t\tctx: ctx,\n\t}\n}", "func (*SpecFactory) Build(resource string) runtime.Object {\n\n\tswitch resource {\n\tcase \"services\":\n\t\treturn &v1.Service{}\n\tcase \"configmaps\":\n\t\treturn &v1.ConfigMap{}\n\t}\n\n\tpanic(fmt.Errorf(\"no resource mapped for %s\", resource))\n}", "func NewFactory() MongoFactory {\n\tfactory := MongoFactory{}\n\tfactory.GetClient()\n\tfactory.GetContext()\n\tfactory.GetDatabase()\n\treturn factory\n}", "func (controller AppsController) Create(c *gin.Context) {\n\tvar configE entities.App\n\terr := c.BindJSON(&configE)\n\n\tif err != nil {\n\t\tfmt.Printf(\"error %v\", err)\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"message\": \"Invalid Parameters\"})\n\t\treturn\n\t}\n\n\tconfig, err := mongodb.Create(controller.MongoDBClient, Collections[\"apps\"], &configE)\n\tif err != nil {\n\t\tfmt.Printf(\"error %v\", err)\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"message\": \"Unable to save data\"})\n\t\treturn\n\t}\n\n\tc.JSON(http.StatusOK, gin.H{\"apps\": config})\n}", "func init() {\n\tApp = New()\n}", "func createApp() *cli.App {\n\tapp := cli.NewApp()\n\tapp.EnableBashCompletion = true\n\tapp.Name = \"iSulad-img\"\n\tif gitCommit != \"\" {\n\t\tapp.Version = fmt.Sprintf(\"%s commit: %s\", Version, gitCommit)\n\t} else {\n\t\tapp.Version = Version\n\t}\n\tapp.Usage = \"Various operations with container images and container image registries\"\n\tapp.Flags = []cli.Flag{\n\t\tcli.BoolFlag{\n\t\t\tName: \"debug\",\n\t\t\tUsage: \"enable debug output\",\n\t\t},\n\t\tcli.StringFlag{\n\t\t\tName: \"log-level, l\",\n\t\t\tUsage: \"Set the logging level\",\n\t\t},\n\t\tcli.StringFlag{\n\t\t\tName: \"run-root\",\n\t\t\tValue: defaultRunRoot,\n\t\t\tUsage: \"use `PATH` as the root directory for execution state files\",\n\t\t},\n\t\tcli.StringFlag{\n\t\t\tName: \"graph-root\",\n\t\t\tValue: defaultGraphRoot,\n\t\t\tUsage: \"use `PATH` as the graph driver's root directory for execution state files\",\n\t\t},\n\t\tcli.StringFlag{\n\t\t\tName: \"driver-name\",\n\t\t\tValue: defaultGraphDriverName,\n\t\t\tUsage: \"use `NAME` as the graph driver\",\n\t\t},\n\t\tcli.StringSliceFlag{\n\t\t\tName: \"driver-options\",\n\t\t\tUsage: \"Options of the graph driver\",\n\t\t},\n\t\tcli.StringSliceFlag{\n\t\t\tName: \"storage-opt\",\n\t\t\tUsage: \"Options of the storage when mount container rootfs\",\n\t\t},\n\t\tcli.StringSliceFlag{\n\t\t\tName: \"insecure-registry\",\n\t\t\tUsage: \"whether to disable TLS verification for the given registry\",\n\t\t},\n\t\tcli.StringSliceFlag{\n\t\t\tName: \"registry\",\n\t\t\tUsage: \"registry to be prepended when pulling unqualified images, can be specified multiple times\",\n\t\t},\n\t\tcli.StringFlag{\n\t\t\tName: \"policy\",\n\t\t\tValue: \"\",\n\t\t\tUsage: \"Path to a trust policy file\",\n\t\t},\n\t\tcli.BoolFlag{\n\t\t\tName: \"insecure-policy\",\n\t\t\tUsage: \"run the tool without any policy check\",\n\t\t},\n\t\tcli.DurationFlag{\n\t\t\tName: \"command-timeout\",\n\t\t\tUsage: \"timeout for the command execution\",\n\t\t},\n\t}\n\tapp.Before = func(c *cli.Context) error {\n\t\tif c.GlobalBool(\"debug\") {\n\t\t\tlogrus.SetLevel(logrus.DebugLevel)\n\t\t} else {\n\t\t\tsetLogLevel(c.GlobalString(\"log-level\"))\n\t\t}\n\t\treturn nil\n\t}\n\tapp.Commands = []cli.Command{\n\t\tinfoCmd,\n\t\timagesCmd,\n\t\tdaemonCmd,\n\t}\n\treturn app\n}", "func CreateApp(appoptions *options.App) (*App, error) {\n\tvar err error\n\n\tctx := context.Background()\n\n\t// Merge default options\n\toptions.MergeDefaults(appoptions)\n\n\tdebug := IsDebug()\n\tdevtools := IsDevtoolsEnabled()\n\tctx = context.WithValue(ctx, \"debug\", debug)\n\tctx = context.WithValue(ctx, \"devtools\", devtools)\n\n\t// Set up logger\n\tmyLogger := logger.New(appoptions.Logger)\n\tif IsDebug() {\n\t\tmyLogger.SetLogLevel(appoptions.LogLevel)\n\t} else {\n\t\tmyLogger.SetLogLevel(appoptions.LogLevelProduction)\n\t}\n\tctx = context.WithValue(ctx, \"logger\", myLogger)\n\tctx = context.WithValue(ctx, \"obfuscated\", IsObfuscated())\n\n\t// Preflight Checks\n\terr = PreflightChecks(appoptions, myLogger)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Create the menu manager\n\tmenuManager := menumanager.NewManager()\n\n\t// Process the application menu\n\tif appoptions.Menu != nil {\n\t\terr = menuManager.SetApplicationMenu(appoptions.Menu)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// Create binding exemptions - Ugly hack. There must be a better way\n\tbindingExemptions := []interface{}{\n\t\tappoptions.OnStartup,\n\t\tappoptions.OnShutdown,\n\t\tappoptions.OnDomReady,\n\t\tappoptions.OnBeforeClose,\n\t}\n\tappBindings := binding.NewBindings(myLogger, appoptions.Bind, bindingExemptions, IsObfuscated())\n\teventHandler := runtime.NewEvents(myLogger)\n\tctx = context.WithValue(ctx, \"events\", eventHandler)\n\t// Attach logger to context\n\tif debug {\n\t\tctx = context.WithValue(ctx, \"buildtype\", \"debug\")\n\t} else {\n\t\tctx = context.WithValue(ctx, \"buildtype\", \"production\")\n\t}\n\n\tmessageDispatcher := dispatcher.NewDispatcher(ctx, myLogger, appBindings, eventHandler, appoptions.ErrorFormatter)\n\tappFrontend := desktop.NewFrontend(ctx, appoptions, myLogger, appBindings, messageDispatcher)\n\teventHandler.AddFrontend(appFrontend)\n\n\tctx = context.WithValue(ctx, \"frontend\", appFrontend)\n\tresult := &App{\n\t\tctx: ctx,\n\t\tfrontend: appFrontend,\n\t\tlogger: myLogger,\n\t\tmenuManager: menuManager,\n\t\tstartupCallback: appoptions.OnStartup,\n\t\tshutdownCallback: appoptions.OnShutdown,\n\t\tdebug: debug,\n\t\tdevtools: devtools,\n\t\toptions: appoptions,\n\t}\n\n\treturn result, nil\n\n}", "func getFactory(flags *pflag.FlagSet) *clientcmd.Factory {\n\tfactory, err := getFactoryFromCluster()\n\tif err != nil {\n\t\tglog.Infof(\"Seems like we are not running in an OpenShift environment (%s), falling back to building a std factory...\", err)\n\t\tfactory = clientcmd.New(flags)\n\t}\n\n\treturn factory\n}", "func New() (App, error) {\n\tcfg := config.New()\n\n\tdb, err := database.New(cfg)\n\tif err != nil {\n\t\treturn App{}, err\n\t}\n\n\tm, err := metrics.New(metrics.Config{\n\t\tEnvironment: cfg.Environment,\n\t\tHostname: cfg.Hostname,\n\t\tNamespace: \"pharos-api-server\",\n\t\tStatsdHost: cfg.StatsdHost,\n\t\tStatsdPort: cfg.StatsdPort,\n\t})\n\tif err != nil {\n\t\treturn App{}, errors.Wrap(err, \"application\")\n\t}\n\n\ts, err := sentry.New(cfg.SentryDSN)\n\tif err != nil {\n\t\treturn App{}, errors.Wrap(err, \"application\")\n\t}\n\n\tv := token.NewVerifier()\n\n\treturn App{cfg, db, m, s, v}, nil\n}", "func newApp(infile, outfile string) *App {\n\treturn &App{\n\t\tAddressFile: infile,\n\t\tGeoDecodeFile: outfile,\n\t\tClient: &http.Client{},\n\t}\n}", "func newTestApp() *App {\n\treturn &App{\n\t\tstorage: storage.NewMemory(),\n\t}\n}", "func New() *Client {\n return &Client{&API{}}\n}", "func createDevice() []*Device {\n\tattrs := createMessageAttribute()\n\tdevices := []*Device{}\n\tdevice := &Device{\n\t\tID: \"id1\",\n\t\tName: \"SensorTag\",\n\t\tDescription: \"Sensor\",\n\t\tState: \"ON\",\n\t\tLastOnline: \"TODAY\",\n\t\tAttributes: attrs,\n\t}\n\tdevices = append(devices, device)\n\treturn devices\n}", "func NewApp(host string, port int) (*App, error) {\n\ta := &App{\n\t\tHost: host,\n\t\tPort: port,\n\t}\n\t// setup DB\n\tdb, err := newDB(\"hades.db\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ta.DB = db\n\t// setup Sessions\n\ts, err := newSessions(a)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ta.Sessions = s\n\t// setup Hades\n\th, err := hades.NewHades(db)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ta.Hades = h\n\t// setup Listener\n\tln, err := newListener(a)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ta.Listener = ln\n\t// setup Templates\n\tt, err := newTemplates(\"../../templates\")\n\ta.Templates = t\n\t// setup Router\n\tr := mux.NewRouter().StrictSlash(true)\n\t// static file handler\n\tsbox := packr.NewBox(\"../../static\")\n\tfsHandler := http.StripPrefix(\"/static/\", http.FileServer(sbox))\n\tr.PathPrefix(\"/static/\").Handler(fsHandler).Methods(\"GET\")\n\t// application routes\n\tr.HandleFunc(\"/\", a.getIndexHandler).Methods(\"GET\")\n\tr.HandleFunc(\"/error\", a.getErrorHandler).Methods(\"GET\")\n\tr.HandleFunc(\"/login\", a.getLoginHandler).Methods(\"GET\")\n\tr.HandleFunc(\"/login\", a.postLoginHandler).Methods(\"POST\")\n\tr.HandleFunc(\"/logout\", a.getLogoutHandler).Methods(\"POST\")\n\tr.HandleFunc(\"/add\", a.getAddHandler).Methods(\"GET\")\n\tr.HandleFunc(\"/add\", a.postAddHandler).Methods(\"POST\")\n\tr.HandleFunc(\"/{id}/action\", a.postActionHandler).Methods(\"POST\")\n\ta.Router = r\n\treturn a, nil\n}", "func New(conf *liverpc.ClientConfig) *Client {\n\tif conf == nil {\n\t\tconf = &liverpc.ClientConfig{}\n\t}\n\tconf.AppID = DiscoveryAppId\n\tvar realCli = liverpc.NewClient(conf)\n\tcli := &Client{cli: realCli}\n\tcli.clientInit(realCli)\n\treturn cli\n}", "func NewApp(listenAddr string, ghClient *githubclient.Client, cache cache.Cache, objTTL time.Duration) App {\n\treturn App{\n\t\tlistenAddr: listenAddr,\n\t\tghClient: ghClient,\n\t\tcache: cache,\n\t\tcacheObjTTL: objTTL,\n\t}\n}", "func NewApp(root string) *App {\n\n CheckEnv()\n\n // Use negroni for middleware\n ne := negroni.New()\n\n // Use gorilla/mux for routing\n ro := mux.NewRouter()\n\n // Use Render for template. Pass in path to templates folder\n // as well as asset helper functions.\n re := render.New(render.Options{\n Directory: filepath.Join(root, \"templates\"),\n Layout: \"layouts/layout\",\n Extensions: []string{\".html\"},\n Funcs: []template.FuncMap{\n\t\t\tAssetHelpers(root),\n\t\t},\n })\n qre := render.New(render.Options{\n Directory: filepath.Join(root, \"templates\"),\n Layout: \"layouts/message\",\n Extensions: []string{\".html\"},\n Funcs: []template.FuncMap{\n\t\t\tAssetHelpers(root),\n\t\t},\n })\n\n // Establish connection to DB as specificed in database.go\n db := NewDB()\n\n // Add middleware to the stack\n ne.Use(negroni.NewRecovery())\n ne.Use(negroni.NewLogger())\n ne.Use(NewAssetHeaders())\n ne.Use(negroni.NewStatic(http.Dir(\"public\")))\n ne.UseHandler(ro)\n\n train.Config.SASS.DebugInfo = true\n train.Config.SASS.LineNumbers = true\n train.Config.Verbose = true\n train.Config.BundleAssets = true\n //ZZZtrain.ConfigureHttpHandler(ro)\n\n // Return a new App struct with all these things.\n return &App{ne, ro, re, qre, db}\n}", "func createAPIClient(ctx context.Context) (manipulate.Manipulator, error) {\n\t// Read the appcreds from a local file. The appcreds\n\t// have been created through apoctl, UI, or the API.\n\tdata, err := ioutil.ReadFile(credsPath)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"unable to read credential file: %s\", err)\n\t}\n\n\t// Unmarshall the data in the data structure.\n\tappCred, tlsConfig, err := midgardclient.ParseCredentials(data)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"unable to parse credential: %s\", err)\n\t}\n\n\treturn maniphttp.New(\n\t\tctx,\n\t\tappCred.APIURL,\n\t\tmaniphttp.OptionNamespace(appCred.Namespace),\n\t\tmaniphttp.OptionTLSConfig(tlsConfig),\n\t\tmaniphttp.OptionTokenManager(\n\t\t\ttokenmanager.NewX509TokenManager(\n\t\t\t\tappCred.APIURL,\n\t\t\t\t24*time.Hour,\n\t\t\t\ttlsConfig,\n\t\t\t),\n\t\t),\n\t)\n}", "func New(\n\tfactories config.Factories,\n\tappInfo ApplicationStartInfo,\n) (*Application, error) {\n\n\tif err := configcheck.ValidateConfigFromFactories(factories); err != nil {\n\t\treturn nil, err\n\t}\n\n\tapp := &Application{\n\t\tinfo: appInfo,\n\t\tv: viper.New(),\n\t\treadyChan: make(chan struct{}),\n\t\tfactories: factories,\n\t}\n\n\trootCmd := &cobra.Command{\n\t\tUse: appInfo.ExeName,\n\t\tLong: appInfo.LongName,\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\tapp.init()\n\t\t\tapp.execute()\n\t\t},\n\t}\n\n\t// TODO: coalesce this code and expose this information to other components.\n\tflagSet := new(flag.FlagSet)\n\taddFlagsFns := []func(*flag.FlagSet){\n\t\ttelemetryFlags,\n\t\tbuilder.Flags,\n\t\tloggerFlags,\n\t}\n\tfor _, addFlags := range addFlagsFns {\n\t\taddFlags(flagSet)\n\t}\n\trootCmd.Flags().AddGoFlagSet(flagSet)\n\n\tapp.rootCmd = rootCmd\n\n\treturn app, nil\n}", "func New(host, port, dbName string, password string) (storage.Service, error) {\n\turl := host + \":\" + port\n\tinfo := &mgo.DialInfo{\n\t\tAddrs: []string{url},\n\t\tUsername: \"mythreekit\",\n\t\tPassword: password,\n\t}\n\tdb, err := mgo.DialWithInfo(info)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\terr = db.Ping()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &mongo{db}, nil\n}", "func CreateNewClient(database string) MongoInstance {\n\tmongoClient, err := mongo.NewClient(options.Client().ApplyURI(\"mongodb://localhost:27045/\"))\n\tmongoContext, cancel := context.WithTimeout(context.Background(), 30*time.Second)\n\n\tdefer cancel()\n\n\terr = mongoClient.Connect(mongoContext)\n\tdb := mongoClient.Database(database)\n\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tif err := mongoClient.Ping(mongoContext, readpref.Primary()); err != nil {\n\t\tpanic(err)\n\t}\n\n\tfmt.Println(\"Successfully connected and pinged.\")\n\n\tmg := MongoInstance{\n\t\tClient: mongoClient,\n\t\tDb: db,\n\t}\n\n\treturn mg\n}", "func New() *App {\n\ta = App{\n\t\tEnv: Development,\n\t\tServer: new(http.Server),\n\t\tredis: NewRedis(\"localhost:6379\", \"\", 0), // TODO: move redis configuration to .env\n\t\tSummery: make(map[string]*prometheus.SummaryVec),\n\t\tCounter: make(map[string]prometheus.Counter),\n\t\tDefaultHandler: func(c *Ctx) error {\n\t\t\treturn c.NotFound()\n\t\t},\n\t}\n\treturn &a\n}", "func newDM(namespace, name string, conf fixture.DMConfig) *DM {\n\tdmMasterName := fmt.Sprintf(\"tipocket-dm-master-%s\", name)\n\tdmMasterLabels := map[string]string{\n\t\t\"app\": \"tipocket-dm\",\n\t\t\"instance\": dmMasterName,\n\t}\n\n\tdmWorkerName := fmt.Sprintf(\"tipocket-dm-worker-%s\", name)\n\tdmWorkerLabels := map[string]string{\n\t\t\"app\": \"tipocket-dm\",\n\t\t\"instance\": dmWorkerName,\n\t}\n\n\timage := fmt.Sprintf(\"pingcap/dm:%s\", conf.DMVersion)\n\n\tdmMasterInitCluster := \"--initial-cluster=\"\n\tfor i := 0; i < conf.MasterReplica; i++ {\n\t\tdmMasterInitCluster += fmt.Sprintf(\"%[1]s-%[3]d=http://%[1]s-%[3]d.%[1]s.%[2]s:8291\", dmMasterName, namespace, i)\n\t\tif i+1 != conf.MasterReplica {\n\t\t\tdmMasterInitCluster += \",\"\n\t\t}\n\t}\n\n\tdm := &DM{\n\t\tSvcMaster: &corev1.Service{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tName: dmMasterName,\n\t\t\t\tNamespace: namespace,\n\t\t\t\tLabels: dmMasterLabels,\n\t\t\t},\n\t\t\tSpec: corev1.ServiceSpec{\n\t\t\t\tType: corev1.ServiceTypeNodePort,\n\t\t\t\tPorts: []corev1.ServicePort{{\n\t\t\t\t\tName: \"dm-master\",\n\t\t\t\t\tPort: 8261,\n\t\t\t\t\tTargetPort: intstr.FromInt(8261),\n\t\t\t\t\tProtocol: corev1.ProtocolTCP,\n\t\t\t\t}, {\n\t\t\t\t\tName: \"dm-master-peer\",\n\t\t\t\t\tPort: 8291,\n\t\t\t\t\tTargetPort: intstr.FromInt(8291),\n\t\t\t\t\tProtocol: corev1.ProtocolTCP,\n\t\t\t\t}},\n\t\t\t\tSelector: dmMasterLabels,\n\t\t\t},\n\t\t},\n\t\tSvcWorker: &corev1.Service{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tName: dmWorkerName,\n\t\t\t\tNamespace: namespace,\n\t\t\t\tLabels: dmWorkerLabels,\n\t\t\t},\n\t\t\tSpec: corev1.ServiceSpec{\n\t\t\t\tType: corev1.ServiceTypeClusterIP,\n\t\t\t\tClusterIP: corev1.ClusterIPNone,\n\t\t\t\tPorts: []corev1.ServicePort{{\n\t\t\t\t\tName: \"dm-worker\",\n\t\t\t\t\tPort: 8262,\n\t\t\t\t\tTargetPort: intstr.FromInt(8262),\n\t\t\t\t\tProtocol: corev1.ProtocolTCP,\n\t\t\t\t}},\n\t\t\t\tSelector: dmWorkerLabels,\n\t\t\t},\n\t\t},\n\t\tStsMaster: &appsv1.StatefulSet{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tName: dmMasterName,\n\t\t\t\tNamespace: namespace,\n\t\t\t\tLabels: dmMasterLabels,\n\t\t\t},\n\t\t\tSpec: appsv1.StatefulSetSpec{\n\t\t\t\tServiceName: dmMasterName,\n\t\t\t\tPodManagementPolicy: appsv1.ParallelPodManagement,\n\t\t\t\tReplicas: pointer.Int32Ptr(int32(conf.MasterReplica)),\n\t\t\t\tSelector: &metav1.LabelSelector{MatchLabels: dmMasterLabels},\n\t\t\t\tTemplate: corev1.PodTemplateSpec{\n\t\t\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\t\t\tLabels: dmMasterLabels,\n\t\t\t\t\t\tAnnotations: map[string]string{\n\t\t\t\t\t\t\t\"prometheus.io/path\": \"/metrics\",\n\t\t\t\t\t\t\t\"prometheus.io/port\": \"8261\",\n\t\t\t\t\t\t\t\"prometheus.io/scrape\": \"true\",\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t\tSpec: corev1.PodSpec{\n\t\t\t\t\t\tContainers: []corev1.Container{{\n\t\t\t\t\t\t\tName: \"dm\",\n\t\t\t\t\t\t\tImage: image,\n\t\t\t\t\t\t\tImagePullPolicy: corev1.PullIfNotPresent,\n\t\t\t\t\t\t\tVolumeMounts: []corev1.VolumeMount{\n\t\t\t\t\t\t\t\t{\n\t\t\t\t\t\t\t\t\tName: name,\n\t\t\t\t\t\t\t\t\tMountPath: \"/data\",\n\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\tPorts: []corev1.ContainerPort{\n\t\t\t\t\t\t\t\t{\n\t\t\t\t\t\t\t\t\tName: \"dm-master\",\n\t\t\t\t\t\t\t\t\tContainerPort: 8261,\n\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t{\n\t\t\t\t\t\t\t\t\tName: \"dm-master-peer\",\n\t\t\t\t\t\t\t\t\tContainerPort: 8291,\n\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\tCommand: []string{\n\t\t\t\t\t\t\t\t\"/dm-master\",\n\t\t\t\t\t\t\t\t\"--data-dir=/data\",\n\t\t\t\t\t\t\t\t\"--name=$(MY_POD_NAME)\",\n\t\t\t\t\t\t\t\t\"--master-addr=:8261\",\n\t\t\t\t\t\t\t\tfmt.Sprintf(\"--advertise-addr=$(MY_POD_NAME).%s.%s:8261\", dmMasterName, namespace),\n\t\t\t\t\t\t\t\t\"--peer-urls=:8291\",\n\t\t\t\t\t\t\t\tfmt.Sprintf(\"--advertise-peer-urls=http://$(MY_POD_NAME).%s.%s:8291\", dmMasterName, namespace),\n\t\t\t\t\t\t\t\tdmMasterInitCluster,\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\tEnv: []corev1.EnvVar{\n\t\t\t\t\t\t\t\t{\n\t\t\t\t\t\t\t\t\tName: \"MY_POD_NAME\",\n\t\t\t\t\t\t\t\t\tValueFrom: &corev1.EnvVarSource{\n\t\t\t\t\t\t\t\t\t\tFieldRef: &corev1.ObjectFieldSelector{\n\t\t\t\t\t\t\t\t\t\t\tAPIVersion: \"v1\",\n\t\t\t\t\t\t\t\t\t\t\tFieldPath: \"metadata.name\",\n\t\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t}},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tVolumeClaimTemplates: []corev1.PersistentVolumeClaim{{\n\t\t\t\t\tObjectMeta: metav1.ObjectMeta{Name: name},\n\t\t\t\t\tSpec: corev1.PersistentVolumeClaimSpec{\n\t\t\t\t\t\tAccessModes: []corev1.PersistentVolumeAccessMode{\n\t\t\t\t\t\t\tcorev1.ReadWriteOnce,\n\t\t\t\t\t\t},\n\t\t\t\t\t\tStorageClassName: &fixture.Context.LocalVolumeStorageClass,\n\t\t\t\t\t\tResources: fixture.WithStorage(fixture.Small, \"1Gi\"),\n\t\t\t\t\t},\n\t\t\t\t}},\n\t\t\t},\n\t\t},\n\t\tStsWorker: &appsv1.StatefulSet{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tName: dmWorkerName,\n\t\t\t\tNamespace: namespace,\n\t\t\t\tLabels: dmWorkerLabels,\n\t\t\t},\n\t\t\tSpec: appsv1.StatefulSetSpec{\n\t\t\t\tServiceName: dmWorkerName,\n\t\t\t\tPodManagementPolicy: appsv1.ParallelPodManagement,\n\t\t\t\tReplicas: pointer.Int32Ptr(int32(conf.WorkerReplica)),\n\t\t\t\tSelector: &metav1.LabelSelector{MatchLabels: dmWorkerLabels},\n\t\t\t\tTemplate: corev1.PodTemplateSpec{\n\t\t\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\t\t\tLabels: dmWorkerLabels,\n\t\t\t\t\t\tAnnotations: map[string]string{\n\t\t\t\t\t\t\t\"prometheus.io/path\": \"/metrics\",\n\t\t\t\t\t\t\t\"prometheus.io/port\": \"8262\",\n\t\t\t\t\t\t\t\"prometheus.io/scrape\": \"true\",\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t\tSpec: corev1.PodSpec{\n\t\t\t\t\t\tContainers: []corev1.Container{{\n\t\t\t\t\t\t\tName: \"dm\",\n\t\t\t\t\t\t\tImage: image,\n\t\t\t\t\t\t\tImagePullPolicy: corev1.PullIfNotPresent,\n\t\t\t\t\t\t\tVolumeMounts: []corev1.VolumeMount{\n\t\t\t\t\t\t\t\t{\n\t\t\t\t\t\t\t\t\tName: name,\n\t\t\t\t\t\t\t\t\tMountPath: \"/data\",\n\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\tPorts: []corev1.ContainerPort{\n\t\t\t\t\t\t\t\t{\n\t\t\t\t\t\t\t\t\tName: \"client\",\n\t\t\t\t\t\t\t\t\tContainerPort: 8262,\n\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\tCommand: []string{\n\t\t\t\t\t\t\t\t\"/dm-worker\",\n\t\t\t\t\t\t\t\t\"--name=$(MY_POD_NAME)\",\n\t\t\t\t\t\t\t\t\"--worker-addr=:8262\",\n\t\t\t\t\t\t\t\tfmt.Sprintf(\"--advertise-addr=$(MY_POD_NAME).%s.%s:8262\", dmWorkerName, namespace),\n\t\t\t\t\t\t\t\tfmt.Sprintf(\"--join=%s.%s:8261\", dmMasterName, namespace),\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\tEnv: []corev1.EnvVar{\n\t\t\t\t\t\t\t\t{\n\t\t\t\t\t\t\t\t\tName: \"MY_POD_NAME\",\n\t\t\t\t\t\t\t\t\tValueFrom: &corev1.EnvVarSource{\n\t\t\t\t\t\t\t\t\t\tFieldRef: &corev1.ObjectFieldSelector{\n\t\t\t\t\t\t\t\t\t\t\tAPIVersion: \"v1\",\n\t\t\t\t\t\t\t\t\t\t\tFieldPath: \"metadata.name\",\n\t\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\tReadinessProbe: &corev1.Probe{\n\t\t\t\t\t\t\t\t// NOTE: no `ReadinessProbe` for DM-master now, because\n\t\t\t\t\t\t\t\t// - DM-master relies on endpoint to communicate with other members when starting\n\t\t\t\t\t\t\t\t// - but endpoint may no available if DM-master not started\n\t\t\t\t\t\t\t\t// then DM-master may never become ready.\n\t\t\t\t\t\t\t\t// so we only check the readiness for DM-worker.\n\t\t\t\t\t\t\t\tHandler: corev1.Handler{\n\t\t\t\t\t\t\t\t\tHTTPGet: &corev1.HTTPGetAction{\n\t\t\t\t\t\t\t\t\t\tPath: \"/status\",\n\t\t\t\t\t\t\t\t\t\tPort: intstr.FromInt(8262),\n\t\t\t\t\t\t\t\t\t\tScheme: \"HTTP\",\n\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\tInitialDelaySeconds: 5,\n\t\t\t\t\t\t\t\tPeriodSeconds: 5,\n\t\t\t\t\t\t\t\tFailureThreshold: 5,\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t}},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tVolumeClaimTemplates: []corev1.PersistentVolumeClaim{{\n\t\t\t\t\tObjectMeta: metav1.ObjectMeta{Name: name},\n\t\t\t\t\tSpec: corev1.PersistentVolumeClaimSpec{\n\t\t\t\t\t\tAccessModes: []corev1.PersistentVolumeAccessMode{\n\t\t\t\t\t\t\tcorev1.ReadWriteOnce,\n\t\t\t\t\t\t},\n\t\t\t\t\t\tStorageClassName: &fixture.Context.LocalVolumeStorageClass,\n\t\t\t\t\t\tResources: fixture.WithStorage(fixture.Small, \"1Gi\"),\n\t\t\t\t\t},\n\t\t\t\t}},\n\t\t\t},\n\t\t},\n\t}\n\n\treturn dm\n}", "func New() App {\n\treturn App{}\n}", "func NewApp(clientService ClientService, lobby Lobby) *App {\n\treturn &App{clientService, lobby}\n}", "func InitApp(keyFile string) (*firestore.Client, error) {\n\tctx := context.Background()\n\tsa := option.WithCredentialsFile(keyFile)\n\tapp, err := firebase.NewApp(ctx, nil, sa)\n\tif err != nil {\n\t\tlog.Fatalln(err)\n\t}\n\tclient, err := app.Firestore(ctx)\n\tif err != nil {\n\t\tlog.Fatalln(err)\n\t}\n\t\n\treturn client, err \n}", "func MakeController(datastore datastore.DataStore) (*Controller, error) {\n\n // Check that we can retrieve an inventory from the datastore\n _, error := datastore.GetInventory()\n if error != nil {\n return nil, error\n }\n\n controller := Controller{\n datastore: datastore,\n register: MakeRegister(),\n }\n return &controller, nil\n}", "func (f *extendedPodFactory) CreateClient(cfg *rest.Config) (interface{}, error) {\n\treturn f.client, nil\n}", "func buildReceiver(f *faker) datastore.KeysChangedReceiver {\n\tvar receiver datastore.KeysChangedReceiver\n\tvar serviceName string\n\tswitch getEnv(\"MESSAGIN_SERVICE\", \"fake\") {\n\tcase \"redis\":\n\t\tconn := redis.NewConnection(getEnv(\"REDIS_ADDR\", \"localhost:6379\"))\n\t\tif getEnv(\"REDIS_TEST_CONN\", \"true\") == \"true\" {\n\t\t\tif err := conn.TestConn(); err != nil {\n\t\t\t\tlog.Fatalf(\"Can not connect to redis: %v\", err)\n\t\t\t}\n\t\t}\n\t\treceiver = &redis.Service{Conn: conn}\n\t\tserviceName = \"redis\"\n\tdefault:\n\t\treceiver = f\n\t\tserviceName = \"fake\"\n\t\tif f == nil {\n\t\t\tfmt.Println(\"none\")\n\t\t}\n\t}\n\tfmt.Printf(\"Messagin Service: %s\", serviceName)\n\treturn receiver\n}", "func NewApp(mode string, cfg *Config) (app *App, err error) {\n\tif cfg == nil {\n\t\tcfg = DefaultConfig()\n\t}\n\tv := &App{\n\t\tobjs: make(map[int]*_object),\n\t\teventMux: newMux(),\n\t\ttrans: newFnMgr(mode),\n\t\tcfg: *cfg,\n\t}\n\tv.b = newBridge(mode, v.trans)\n\n\t// refer to 'ReadMostly' example in sync/atomic\n\tv.eventOut.Store(make(map[int]*_eventListener))\n\tv.eventMux.Handle(func(val interface{}, _ int) {\n\t\te := val.(*Event)\n\t\tm := v.eventOut.Load().(map[int]*_eventListener)\n\t\t// to the channel containing everythin errors\n\t\tfor _, eln := range m {\n\t\t\tif (eln.targets&e.Origin) == 0 || eln.level > e.Level {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// non-blocking channel sending\n\t\t\tselect {\n\t\t\tcase eln.events <- e:\n\t\t\tdefault:\n\t\t\t\t// drop this event\n\t\t\t\t// TODO: log it?\n\t\t\t}\n\t\t}\n\t})\n\n\tremain, err := v.eventMux.More(1)\n\tif err != nil || remain != 0 {\n\t\terr = fmt.Errorf(\"Unable to allocate mux routine: %v\", remain)\n\t}\n\n\t// init mappers\n\tif v.mappers, err = newMappers(v.trans, v.b.(exHooks)); err != nil {\n\t\treturn\n\t}\n\tif err = v.attachObject(v.mappers, ObjT.Mapper); err != nil {\n\t\treturn\n\t}\n\t// 'local' mode\n\tif err = v.allocateMappers(); err != nil {\n\t\treturn\n\t}\n\n\t// init workers\n\tif v.workers, err = newWorkers(v.trans, v.b.(exHooks)); err != nil {\n\t\treturn\n\t}\n\tif err = v.attachObject(v.workers, ObjT.Worker); err != nil {\n\t\treturn\n\t}\n\n\tapp = v\n\treturn\n}", "func NewDevice(info IDeviceInfo, commandProcessor CommandProcessor, eventProcessor EventProcessor, queryProcessor QueryProcessor) *Device {\n instance := new(Device)\n agent.SetupAgent(&instance.Agent, info, 0, 0, agent.DEFAULT_BUFFER_SIZE, true)\n instance.info = info\n instance.commandProcessor = commandProcessor\n instance.eventProcessor = eventProcessor\n instance.queryProcessor = queryProcessor\n return instance\n}", "func New(c *conf.Config) (d *Dao) {\n\td = &Dao{\n\t\t// http client\n\t\tclient: httpx.NewClient(c.HTTPAd),\n\t\tbce: c.Host.Ad + _bce,\n\t}\n\treturn\n}", "func (c *DeviceConfig) newDeviceFinder() (*device.DeviceFinder, error) {\n\tif c.deviceFinderPath == \"\" {\n\t\treturn nil, fmt.Errorf(\"--device-finder-path not specified\")\n\t}\n\n\treturn device.NewDeviceFinder(c.deviceFinderPath), nil\n}", "func newClient(conf config) (*storage.Client, error) {\n\tdb, err := storage.NewDBClient(conf.MongoURI, conf.DBName, conf.MongoMICol, conf.MongoAgCol)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error creating DB client: %q\", err)\n\t}\n\tdb.Collection(conf.MongoMICol)\n\tbc := storage.NewCloudClient(conf.SwiftUsername, conf.SwiftAPIKey, conf.SwiftAuthURL, conf.SwiftDomain, conf.SwiftContainer)\n\tclient, err := storage.NewClient(db, bc)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error creating storage.client: %q\", err)\n\t}\n\treturn client, nil\n}", "func Factory() (*ORM, error) {\n\tdsn := fmt.Sprintf(\"host=%s user=%s dbname=%s sslmode=%s password=%s\", hostDB, userDB, nameDB, sslDB, passwordDB) //Build connection string\n\n\tdb, err := gorm.Open(dialect, dsn)\n\tif err != nil {\n\t\tlog.Panic(\"[ORM] err: \", err)\n\t}\n\torm := &ORM{\n\t\tDB: db,\n\t}\n\t// Log every SQL command on dev, @prod: this should be disabled?\n\tdb.LogMode(logMode)\n\t// Automigrate tables\n\tif autoMigrate {\n\t\terr = migration.ServiceAutoMigration(orm.DB)\n\t}\n\tlog.Info(\"[ORM] Database connection initialized.\")\n\treturn orm, err\n}", "func (c Client) CreateApp() (*App, error) {\n\treturn c.createApp(nil)\n}", "func (o *DesktopApp) Create() (*restapi.SliceResponse, error) {\n\tvar queryArg = make(map[string]interface{})\n\n\tqueryArg[\"ID\"] = []string{o.TemplateName}\n\tLogD.Printf(\"Generated Map for Create(): %+v\", queryArg)\n\n\tresp, err := o.client.CallSliceAPI(o.apiCreate, queryArg)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif !resp.Success {\n\t\treturn nil, errors.New(resp.Message)\n\t}\n\n\treturn resp, nil\n}", "func (a *App) Generator(data makr.Data) (*makr.Generator, error) {\n\tg := makr.New()\n\tg.Add(makr.NewCommand(makr.GoGet(\"golang.org/x/tools/cmd/goimports\", \"-u\")))\n\tg.Add(makr.NewCommand(makr.GoGet(\"github.com/golang/dep/cmd/dep\", \"-u\")))\n\tg.Add(makr.NewCommand(makr.GoGet(\"github.com/motemen/gore\", \"-u\")))\n\n\tfiles, err := generators.Find(filepath.Join(generators.TemplatesPath, \"newapp\"))\n\tif err != nil {\n\t\treturn nil, errors.WithStack(err)\n\t}\n\n\tfor _, f := range files {\n\t\tg.Add(makr.NewFile(f.WritePath, f.Body))\n\t}\n\trr, err := refresh.New()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tg.Add(rr)\n\n\tif data[\"ciProvider\"] == \"travis\" {\n\t\tg.Add(makr.NewFile(\".travis.yml\", nTravis))\n\t} else if data[\"ciProvider\"] == \"gitlab-ci\" {\n\t\tif _, ok := data[\"withPop\"]; ok {\n\t\t\tif data[\"dbType\"] == \"postgres\" {\n\t\t\t\tdata[\"testDbUrl\"] = \"postgres://postgres:postgres@postgres:5432/\" + data[\"name\"].(string) + \"_test?sslmode=disable\"\n\t\t\t} else if data[\"dbType\"] == \"mysql\" {\n\t\t\t\tdata[\"testDbUrl\"] = \"mysql://root:root@mysql:3306/\" + data[\"name\"].(string) + \"_test\"\n\t\t\t} else {\n\t\t\t\tdata[\"testDbUrl\"] = \"\"\n\t\t\t}\n\t\t\tg.Add(makr.NewFile(\".gitlab-ci.yml\", nGitlabCi))\n\t\t} else {\n\t\t\tg.Add(makr.NewFile(\".gitlab-ci.yml\", nGitlabCiNoPop))\n\t\t}\n\t}\n\n\tif !a.API {\n\t\tif a.SkipWebpack {\n\t\t\twg, err := standard.New(data)\n\t\t\tif err != nil {\n\t\t\t\treturn g, errors.WithStack(err)\n\t\t\t}\n\t\t\tg.Add(wg)\n\t\t} else if a.WithReact {\n\t\t\twg, err := react.New(data)\n\t\t\tif err != nil {\n\t\t\t\treturn g, err\n\t\t\t}\n\t\t\tg.Add(wg)\n\t\t} else {\n\t\t\twg, err := webpack.New(data)\n\t\t\tif err != nil {\n\t\t\t\treturn g, errors.WithStack(err)\n\t\t\t}\n\t\t\tg.Add(wg)\n\t\t}\n\t}\n\tif !a.SkipPop {\n\t\tg.Add(newSodaGenerator())\n\t}\n\tif a.API {\n\t\tg.Add(makr.Func{\n\t\t\tRunner: func(path string, data makr.Data) error {\n\t\t\t\treturn os.RemoveAll(filepath.Join(path, \"templates\"))\n\t\t\t},\n\t\t})\n\t\tg.Add(makr.Func{\n\t\t\tRunner: func(path string, data makr.Data) error {\n\t\t\t\treturn os.RemoveAll(filepath.Join(path, \"locales\"))\n\t\t\t},\n\t\t})\n\t}\n\tif a.Docker != \"none\" {\n\t\tdg, err := docker.New()\n\t\tif err != nil {\n\t\t\treturn g, errors.WithStack(err)\n\t\t}\n\t\tg.Add(dg)\n\t}\n\tg.Add(makr.NewCommand(a.goGet()))\n\n\tif _, err := exec.LookPath(\"git\"); err == nil {\n\t\tg.Add(makr.NewCommand(exec.Command(\"git\", \"init\")))\n\t\tg.Add(makr.NewCommand(exec.Command(\"git\", \"add\", \".\")))\n\t\tg.Add(makr.NewCommand(exec.Command(\"git\", \"commit\", \"-m\", \"Initial Commit\")))\n\t}\n\n\treturn g, nil\n}", "func (g *FakeDatabaseClientFactory) New(context.Context, client.Reader, string, string) (dbdpb.DatabaseDaemonClient, func() error, error) {\n\tif g.Dbclient == nil {\n\t\tg.Reset()\n\t}\n\treturn g.Dbclient, func() error { return nil }, nil\n}", "func New(opts ...Option) *App {\n\tlogger := fxlog.New()\n\tlc := &lifecycleWrapper{lifecycle.New(logger)}\n\n\tapp := &App{\n\t\tcontainer: dig.New(),\n\t\tlifecycle: lc,\n\t\tlogger: logger,\n\t}\n\n\tfor _, opt := range opts {\n\t\topt.apply(app)\n\t}\n\n\tfor _, p := range app.provides {\n\t\tapp.provide(p)\n\t}\n\tapp.provide(func() Lifecycle { return app.lifecycle })\n\treturn app\n}", "func Factory(builder string, config map[string]interface{}) (Builder, error) {\n\tfactory, ok := SupportedBuilders[builder]\n\tif !ok {\n\t\terrors.Errorf(\"Unsupported builder '%s' valid builders are(%s)\", builder, GetBuilders())\n\t}\n\treturn factory(config)\n}", "func BuildFactory(appearance int) GUIFactory {\n\tswitch appearance {\n\tcase React:\n\t\treturn new(ReactFactory)\n\tcase Vue:\n\t\treturn new(VueFactory)\n\tcase Angular:\n\t\treturn new(AngularFactory)\n\t}\n\treturn nil\n}", "func NewIosDeviceType()(*IosDeviceType) {\n m := &IosDeviceType{\n }\n m.SetAdditionalData(make(map[string]interface{}));\n return m\n}", "func Open(\n\tctx context.Context,\n\taddr string,\n\tappname string,\n\tnames ...string,\n) (*MongoDB, error) {\n\tm := &MongoDB{\n\t\tctx: context.WithValue(\n\t\t\tctx,\n\t\t\tkeyPrincipalID,\n\t\t\tprimitive.NewObjectID().Hex(),\n\t\t),\n\t\tconnstr: fmt.Sprintf(\n\t\t\t\"mongodb://%s/?readPreference=primary&appname=%s\",\n\t\t\taddr, appname,\n\t\t),\n\t\tcoll: make(map[string]*mongo.Collection),\n\t}\n\n\topts := options.Client().ApplyURI(m.connstr)\n\tclient, err := mongo.NewClient(opts)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tm.client = client\n\n\tif err := m.Connect(); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif len(names) > 0 {\n\t\tm.SetDatabase(names[0])\n\t}\n\n\tif len(names) > 1 {\n\t\tm.SetCollections(names[1:]...)\n\t}\n\n\treturn m, nil\n}", "func (a *appHandler) CreateApp(w http.ResponseWriter, r *http.Request) {\n\tvar app model.App\n\terr := json.NewDecoder(r.Body).Decode(&app)\n\tif err != nil {\n\t\ta.httpUtil.WriteJSONBadRequestResponse(w, err)\n\t\treturn\n\t}\n\n\t// TODO : Create\n\n\tjsonR, err := json.Marshal(app)\n\tif err != nil {\n\t\ta.httpUtil.WriteJSONInternalServerErrorResponse(w, err)\n\t}\n\n\ta.httpUtil.WriteJSONSuccessResponse(w, jsonR)\n}" ]
[ "0.650424", "0.5821867", "0.5741955", "0.5705958", "0.5657533", "0.56301856", "0.556906", "0.5550002", "0.5528626", "0.55113107", "0.5511156", "0.5497224", "0.54902005", "0.5483873", "0.54688966", "0.5464506", "0.5460464", "0.5453676", "0.54433185", "0.5423936", "0.5401687", "0.5401557", "0.53918374", "0.5389848", "0.5389832", "0.5369868", "0.5352729", "0.53353316", "0.5322693", "0.531826", "0.53161454", "0.5296377", "0.5288534", "0.5286804", "0.5270835", "0.5261587", "0.5249951", "0.52347815", "0.523459", "0.5225561", "0.5222704", "0.5207701", "0.5181976", "0.5180927", "0.51804477", "0.5171227", "0.5163149", "0.5163149", "0.5161398", "0.51599336", "0.515668", "0.51560825", "0.51502424", "0.51412064", "0.5139308", "0.51252997", "0.51250196", "0.51220405", "0.512017", "0.5116108", "0.51158047", "0.5111288", "0.5107186", "0.51028794", "0.5097346", "0.50968605", "0.50873333", "0.5084568", "0.5083707", "0.5071088", "0.5059649", "0.5057221", "0.50467515", "0.5044178", "0.50440305", "0.50359577", "0.5032612", "0.502185", "0.5019326", "0.5017936", "0.5017053", "0.50128394", "0.50091684", "0.50065225", "0.4988821", "0.49856487", "0.4985166", "0.4983724", "0.49789703", "0.4978129", "0.4974737", "0.49747023", "0.4973961", "0.4969632", "0.49682432", "0.49595758", "0.49595615", "0.49581343", "0.49576864", "0.49536577" ]
0.6445273
1
New create devices web app
func New(jwtMiddleware *jwt.JWTMiddleware, mongoClient *mongo.Client) *App { app := new(App) app.jwtMiddleware = jwtMiddleware app.mongoClient = mongoClient _, err := caclient.GetDefaultCAClient() if err != nil { if err, ok := err.(*caclient.ClientError); ok { if err.Code != caclient.ErrorNotConfig { log.Fatalf("Error loading caclient. Error Code: %d -- %s", err.Code, err.Error()) return nil } } } collection := app.mongoClient.Database(utils.MongoDb).Collection("pantahub_devices") CreateIndexesOptions := options.CreateIndexesOptions{} CreateIndexesOptions.SetMaxTime(CreateIndexTimeout) indexOptions := options.IndexOptions{} indexOptions.SetUnique(true) indexOptions.SetSparse(false) indexOptions.SetBackground(true) index := mongo.IndexModel{ Keys: bsonx.Doc{ {Key: "owner", Value: bsonx.Int32(1)}, {Key: "nick", Value: bsonx.Int32(1)}, }, Options: &indexOptions, } _, err = collection.Indexes().CreateOne(context.Background(), index, &CreateIndexesOptions) if err != nil { log.Fatalln("Error setting up index for pantahub_devices: " + err.Error()) return nil } CreateIndexesOptions = options.CreateIndexesOptions{} CreateIndexesOptions.SetMaxTime(CreateIndexTimeout) indexOptions = options.IndexOptions{} indexOptions.SetUnique(false) indexOptions.SetSparse(false) indexOptions.SetBackground(true) index = mongo.IndexModel{ Keys: bsonx.Doc{ {Key: "timemodified", Value: bsonx.Int32(1)}, }, Options: &indexOptions, } collection = app.mongoClient.Database(utils.MongoDb).Collection("pantahub_devices") _, err = collection.Indexes().CreateOne(context.Background(), index, &CreateIndexesOptions) if err != nil { log.Fatalln("Error setting up index for pantahub_devices: " + err.Error()) return nil } CreateIndexesOptions = options.CreateIndexesOptions{} CreateIndexesOptions.SetMaxTime(CreateIndexTimeout) indexOptions = options.IndexOptions{} indexOptions.SetUnique(false) indexOptions.SetSparse(false) indexOptions.SetBackground(true) index = mongo.IndexModel{ Keys: bsonx.Doc{ {Key: "prn", Value: bsonx.Int32(1)}, }, Options: &indexOptions, } collection = app.mongoClient.Database(utils.MongoDb).Collection("pantahub_devices") _, err = collection.Indexes().CreateOne(context.Background(), index, &CreateIndexesOptions) if err != nil { log.Fatalln("Error setting up index for pantahub_devices: " + err.Error()) return nil } // Indexing for the owner,garbage fields CreateIndexesOptions = options.CreateIndexesOptions{} CreateIndexesOptions.SetMaxTime(CreateIndexTimeout) indexOptions = options.IndexOptions{} indexOptions.SetUnique(false) indexOptions.SetSparse(false) indexOptions.SetBackground(true) index = mongo.IndexModel{ Keys: bsonx.Doc{ {Key: "owner", Value: bsonx.Int32(1)}, {Key: "garbage", Value: bsonx.Int32(1)}, }, Options: &indexOptions, } collection = app.mongoClient.Database(utils.MongoDb).Collection("pantahub_devices") _, err = collection.Indexes().CreateOne(context.Background(), index, &CreateIndexesOptions) if err != nil { log.Fatalln("Error setting up index for pantahub_devices: " + err.Error()) return nil } // Indexing for the device,garbage fields CreateIndexesOptions = options.CreateIndexesOptions{} CreateIndexesOptions.SetMaxTime(CreateIndexTimeout) indexOptions = options.IndexOptions{} indexOptions.SetUnique(false) indexOptions.SetSparse(false) indexOptions.SetBackground(true) index = mongo.IndexModel{ Keys: bsonx.Doc{ {Key: "device", Value: bsonx.Int32(1)}, {Key: "garbage", Value: bsonx.Int32(1)}, }, Options: &indexOptions, } collection = app.mongoClient.Database(utils.MongoDb).Collection("pantahub_devices") _, err = collection.Indexes().CreateOne(context.Background(), index, &CreateIndexesOptions) if err != nil { log.Fatalln("Error setting up index for pantahub_devices: " + err.Error()) return nil } err = app.EnsureTokenIndices() if err != nil { log.Println("Error creating indices for pantahub devices tokens: " + err.Error()) return nil } app.API = rest.NewApi() // we dont use default stack because we dont want content type enforcement app.API.Use(&rest.AccessLogJsonMiddleware{Logger: log.New(os.Stdout, "/devices:", log.Lshortfile)}) app.API.Use(&utils.AccessLogFluentMiddleware{Prefix: "devices"}) app.API.Use(&rest.StatusMiddleware{}) app.API.Use(&rest.TimerMiddleware{}) app.API.Use(&metrics.Middleware{}) app.API.Use(rest.DefaultCommonStack...) app.API.Use(&rest.CorsMiddleware{ RejectNonCorsRequests: false, OriginValidator: func(origin string, request *rest.Request) bool { return true }, AllowedMethods: []string{"GET", "POST", "PUT", "PATCH", "DELETE", "OPTIONS"}, AllowedHeaders: []string{ "Accept", "Content-Type", "X-Custom-Header", "Origin", "Authorization"}, AccessControlAllowCredentials: true, AccessControlMaxAge: 3600, }) app.API.Use(&rest.IfMiddleware{ Condition: func(request *rest.Request) bool { // if call is coming with authorization attempt, ensure JWT middleware // is used... otherwise let through anonymous POST for registration auth := request.Header.Get("Authorization") if auth != "" && strings.HasPrefix(strings.ToLower(strings.TrimSpace(auth)), "bearer ") { return true } // post new device means to register... allow this unauthenticated return !((request.Method == "POST" && request.URL.Path == "/") || (request.Method == "POST" && request.URL.Path == "/register")) }, IfTrue: app.jwtMiddleware, }) app.API.Use(&rest.IfMiddleware{ Condition: func(request *rest.Request) bool { // if call is coming with authorization attempt, ensure JWT middleware // is used... otherwise let through anonymous POST for registration auth := request.Header.Get("Authorization") if auth != "" && strings.HasPrefix(strings.ToLower(strings.TrimSpace(auth)), "bearer ") { return true } // post new device means to register... allow this unauthenticated return !((request.Method == "POST" && request.URL.Path == "/") || (request.Method == "POST" && request.URL.Path == "/register")) }, IfTrue: &utils.AuthMiddleware{}, }) writeDevicesScopes := []utils.Scope{ utils.Scopes.API, utils.Scopes.Devices, utils.Scopes.WriteDevices, } readDevicesScopes := []utils.Scope{ utils.Scopes.API, utils.Scopes.Devices, utils.Scopes.ReadDevices, } updateDevicesScopes := []utils.Scope{ utils.Scopes.API, utils.Scopes.Devices, utils.Scopes.UpdateDevices, } // /auth_status endpoints apiRouter, _ := rest.MakeRouter( // TPM auto enroll register rest.Post("/register", app.handleRegister), // token api rest.Post("/tokens", utils.ScopeFilter(readDevicesScopes, app.handlePostTokens)), rest.Delete("/tokens/#id", utils.ScopeFilter(updateDevicesScopes, app.handleDisableTokens)), rest.Get("/tokens", utils.ScopeFilter(readDevicesScopes, app.handleGetTokens)), // default api rest.Get("/auth_status", utils.ScopeFilter(readDevicesScopes, handleAuth)), rest.Get("/", utils.ScopeFilter(readDevicesScopes, app.handleGetDevices)), rest.Post("/", utils.ScopeFilter(writeDevicesScopes, func(writer rest.ResponseWriter, request *rest.Request) { userAgent := request.Header.Get("User-Agent") if userAgent == "" { utils.RestErrorWrapperUser(writer, "No Access (DOS) - no UserAgent", "Incompatible Client; upgrade pantavisor", http.StatusForbidden) return } app.handlePostDevice(writer, request) })), rest.Get("/#id", utils.ScopeFilter(readDevicesScopes, app.handleGetDevice)), rest.Put("/#id", utils.ScopeFilter(writeDevicesScopes, app.handlePutDevice)), rest.Patch("/#id", utils.ScopeFilter(writeDevicesScopes, app.handlePatchDevice)), rest.Put("/#id/public", utils.ScopeFilter(writeDevicesScopes, app.handlePutPublic)), rest.Delete("/#id/public", utils.ScopeFilter(writeDevicesScopes, app.handleDeletePublic)), rest.Put("/#id/user-meta", utils.ScopeFilter(writeDevicesScopes, app.handlePutUserData)), rest.Patch("/#id/user-meta", utils.ScopeFilter(writeDevicesScopes, app.handlePatchUserData)), rest.Put("/#id/device-meta", utils.ScopeFilter(writeDevicesScopes, app.handlePutDeviceData)), rest.Patch("/#id/device-meta", utils.ScopeFilter(writeDevicesScopes, app.handlePatchDeviceData)), rest.Delete("/#id", utils.ScopeFilter(writeDevicesScopes, app.handleDeleteDevice)), // lookup by nick-path (np) rest.Get("/np/#usernick/#devicenick", utils.ScopeFilter(readDevicesScopes, app.handleGetUserDevice)), ) app.API.Use(&tracer.OtelMiddleware{ ServiceName: os.Getenv("OTEL_SERVICE_NAME"), Router: apiRouter, }) app.API.SetApp(apiRouter) return app }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *DeviceController) Create(w http.ResponseWriter, r *http.Request) {\n\tfmt.Println(json.NewDecoder(r.Body))\n\tres := db.PutItem(\"home-devices\", NewDevice(\"hello\", \"world\", \"indeed\"))\n\tc.SendJSON(\n\t\tw,\n\t\tr,\n\t\tres,\n\t\thttp.StatusOK,\n\t)\n}", "func RegisterDevice(w http.ResponseWriter, r *http.Request) {\n\tlog.Println(\"Received a POST. Registering a new Device!\")\n\n\tdata, err := simplejson.NewFromReader(r.Body)\n\n\tif err != nil {\n\t\tlog.Println(\"Failed to parse incoming JSON: \", err)\n\t\thttp.Error(w, fmt.Sprintf(\"{\\\"error\\\":\\\"JSON Parse Failure\\\",\\\"details\\\":\\\"%v\\\"}\", err), 500)\n\t\treturn\n\t}\n\n\tdevice_type := data.Get(\"deviceType\").MustString()\n\tdevice_id := data.Get(\"deviceId\").MustString()\n\tdevice_token := data.Get(\"deviceToken\").MustString()\n\n\tlog.Print(\"A new device registration will be created for an '\", device_type, \"'.\")\n\n\tendpoint, err := CreateEndpoint(device_type, device_id, device_token)\n\n\tif err != nil {\n\t\tlog.Println(\"Failed to create a new endpoint: \", err)\n\t\thttp.Error(w, fmt.Sprintf(\"{\\\"error\\\":\\\"Failed to create a new endpoint.\\\",\\\"details\\\":\\\"%v\\\"}\", err), 500)\n\t\treturn\n\t}\n\n\t// Off in the background we'll report a new user using this software.\n\t// How's that for dogfood?\n\tgo func(device_type string) {\n\t\tif *notificEndpoint != \"\" {\n\t\t\tlog.Println(\"Using Notific.io to report new registration: \", device_type)\n\t\t\t_, err := http.Post(fmt.Sprintf(\"http://api.io/\"),\n\t\t\t\t\"application/json\",\n\t\t\t\tnil)\n\n\t\t\tif err != nil {\n\t\t\t\tlog.Println(\"Failed to notify with io about a new user.\")\n\t\t\t}\n\t\t} else {\n\t\t\tlog.Println(\"No inbuilt notific token to dogfood a regisrtation with...\")\n\t\t}\n\t}(device_type)\n\n\t// 201 OK (but a little bit more)\n\tw.WriteHeader(http.StatusCreated)\n\tfmt.Fprintf(w, \"{ \\\"endpoint\\\":\\\"%v\\\", \\\"token\\\": \\\"%v\\\" }\", endpoint.Id, endpoint.Token)\n}", "func (a *DeviceAPI) Create(ctx context.Context, req *api.CreateDeviceRequest) (*empty.Empty, error) {\n\tvar response empty.Empty\n\tif req.Device == nil {\n\t\treturn nil, status.Errorf(codes.InvalidArgument, \"device must not be nil\")\n\t}\n\n\tvar devEUI lorawan.EUI64\n\tif err := devEUI.UnmarshalText([]byte(req.Device.DevEui)); err != nil {\n\t\treturn nil, status.Errorf(codes.InvalidArgument, err.Error())\n\t}\n\n\tdpID, err := uuid.FromString(req.Device.DeviceProfileId)\n\tif err != nil {\n\t\treturn nil, status.Errorf(codes.InvalidArgument, err.Error())\n\t}\n\n\t// if Name is \"\", set it to the DevEUI\n\tif req.Device.Name == \"\" {\n\t\treq.Device.Name = req.Device.DevEui\n\t}\n\n\t// Validate that application and device-profile are under the same\n\t// organization ID.\n\tapp, err := a.st.GetApplication(ctx, req.Device.ApplicationId)\n\tif err != nil {\n\t\treturn nil, helpers.ErrToRPCError(err)\n\t}\n\tdp, err := a.st.GetDeviceProfile(ctx, dpID, false)\n\tif err != nil {\n\t\treturn nil, helpers.ErrToRPCError(err)\n\t}\n\tif app.OrganizationID != dp.OrganizationID {\n\t\treturn nil, status.Errorf(codes.InvalidArgument, \"device-profile and application must be under the same organization\")\n\t}\n\n\tcred, err := a.auth.GetCredentials(ctx, auth.NewOptions().WithOrgID(app.OrganizationID))\n\tif err != nil {\n\t\treturn nil, status.Errorf(codes.Unauthenticated, \"authentication failed: %v\", err)\n\t}\n\tif !cred.IsGlobalAdmin && !cred.IsDeviceAdmin {\n\t\treturn nil, status.Errorf(codes.PermissionDenied, \"permission denied\")\n\t}\n\n\t// Set Device struct.\n\td := Device{\n\t\tDevEUI: devEUI,\n\t\tApplicationID: req.Device.ApplicationId,\n\t\tDeviceProfileID: dpID,\n\t\tName: req.Device.Name,\n\t\tDescription: req.Device.Description,\n\t\tSkipFCntCheck: req.Device.SkipFCntCheck,\n\t\tReferenceAltitude: req.Device.ReferenceAltitude,\n\t\tVariables: hstore.Hstore{\n\t\t\tMap: make(map[string]sql.NullString),\n\t\t},\n\t\tTags: hstore.Hstore{\n\t\t\tMap: make(map[string]sql.NullString),\n\t\t},\n\t}\n\n\tfor k, v := range req.Device.Variables {\n\t\td.Variables.Map[k] = sql.NullString{String: v, Valid: true}\n\t}\n\n\tfor k, v := range req.Device.Tags {\n\t\td.Tags.Map[k] = sql.NullString{String: v, Valid: true}\n\t}\n\n\tnsCli, err := a.nsCli.GetNetworkServerServiceClient(dp.NetworkServerID)\n\tif err != nil {\n\t\treturn nil, status.Errorf(codes.Internal, \"couldn't get network server client: %v\", err)\n\t}\n\n\tif err := device.CreateDevice(ctx, a.st, &d, &app, a.ApplicationServerID, a.mxpCli, nsCli); err != nil {\n\t\treturn nil, status.Errorf(codes.Internal, err.Error())\n\t}\n\n\treturn &response, nil\n}", "func (o *Oispapi) CreateDevice(device *Device) error {\n\tclient := &http.Client{}\n\n\treplacements := map[string]string{\"accountId\": o.accounts[o.activAccount].ID}\n\turl := makeURL(o.url, getDevicesPath, replacements)\n\tjsonBody, _ := json.Marshal(device)\n\treq, _ := http.NewRequest(\"POST\", url, bytes.NewBuffer(jsonBody))\n\tsetHeaders(req, o.token)\n\tresponse, err := client.Do(req)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer response.Body.Close()\n\tresponseData, err := ioutil.ReadAll(response.Body)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif response.StatusCode != 201 {\n\t\treturn errors.New(string(responseData))\n\t}\n\tresponseObject := []Device{}\n\tjson.Unmarshal(responseData, &responseObject)\n\treturn nil\n}", "func (u *TenantUser) CreateDevice(device *Device) error {\n\turl := u.ServerHost + saveDeviceEnding\n\n\tsaveDeviceRequest := SaveDeviceRequest{device.Name, device.Type}\n\tdata, err := json.Marshal(saveDeviceRequest)\n\tif err != nil {\n\t\treturn err\n\t}\n\trequest, err := http.NewRequest(http.MethodPost, url, bytes.NewReader(data))\n\tif err != nil {\n\t\treturn err\n\t}\n\trequest.Header = map[string][]string{\"Content-Type\": {applicationJson}, \"X-Authorization\": {u.Jwt}}\n\n\tresponse, err := http.DefaultClient.Do(request)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer func() { _ = response.Body.Close() }()\n\n\tall, err := ioutil.ReadAll(response.Body)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t//log.Printf(\"response = %s\\n\" , string(all))\n\tvar tbDevice TBDevice\n\tif err = json.Unmarshal(all, &tbDevice); err != nil {\n\t\treturn err\n\t}\n\t//log.Printf(\"tbDevice = %+v\\n\", tbDevice)\n\n\tif len(tbDevice.Id.Id) == 0 {\n\t\treturn fmt.Errorf(\"create fail|msg=%s\", string(all))\n\t}\n\n\tdevice.Id = tbDevice.Id.Id\n\tdevice.AdditionInfo = tbDevice.AdditionalInfo\n\tdevice.Label = tbDevice.Label\n\n\treturn nil\n}", "func CreateDevice(projectID, hostname, plan, facility, operatingSystem, billingCycle, userData, ipxeScriptURL string, tags []string, spotInstance, alwaysPXE bool, spotPriceMax float64, terminationTime *time.Time, silent bool) error {\n\tclient, err := NewPacketClient()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treq := packngo.DeviceCreateRequest{\n\t\tHostname: hostname,\n\t\tPlan: plan,\n\t\tFacility: facility,\n\t\tOS: operatingSystem,\n\t\tBillingCycle: billingCycle,\n\t\tProjectID: projectID,\n\t\tUserData: userData,\n\t\tSpotInstance: spotInstance,\n\t\tSpotPriceMax: spotPriceMax,\n\t\tTags: tags,\n\t\tAlwaysPXE: alwaysPXE,\n\t\tIPXEScriptURL: ipxeScriptURL,\n\t}\n\tif terminationTime != nil {\n\t\treq.TerminationTime = &packngo.Timestamp{*terminationTime}\n\t}\n\n\td, _, err := client.Devices.Create(&req)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif silent {\n\t\te := MarshallAndPrint(d)\n\t\treturn e\n\t}\n\n\t// print events till device is provisionned\n\tfinalEvent := \"Provision complete! Your device is ready to go.\"\n\tlastEvent := \"\"\n\n\textclient, err := NewExtPacketClient()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfmt.Fprint(os.Stderr, \"\\n\")\n\tfmt.Fprint(os.Stderr, \"Provisioning of device successfully started...\")\n\n\tfor {\n\t\tevents, _, err := extclient.Events.ListDeviceEvents(d.ID)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tcurrentEventO := events[0]\n\n\t\tif currentEventO.Body != lastEvent {\n\t\t\tfmt.Fprintf(os.Stderr, \" [ %s ] %s\\n\", currentEventO.Create, currentEventO.Body)\n\t\t\tlastEvent = currentEventO.Body\n\t\t}\n\n\t\tif currentEventO.Body == finalEvent {\n\t\t\tfmt.Println()\n\t\t\tbreak\n\t\t}\n\n\t\ttime.Sleep(10 * time.Second)\n\t}\n\n\treturn ListDevice(d.ID)\n}", "func createDevice() []*Device {\n\tattrs := createMessageAttribute()\n\tdevices := []*Device{}\n\tdevice := &Device{\n\t\tID: \"id1\",\n\t\tName: \"SensorTag\",\n\t\tDescription: \"Sensor\",\n\t\tState: \"ON\",\n\t\tLastOnline: \"TODAY\",\n\t\tAttributes: attrs,\n\t}\n\tdevices = append(devices, device)\n\treturn devices\n}", "func (c *Controller) createPlatform(r *web.Request) (*web.Response, error) {\n\tctx := r.Context()\n\tlogger := log.C(ctx)\n\tlogger.Debug(\"Creating new platform\")\n\n\tplatform := &types.Platform{}\n\tif err := util.BytesToObject(r.Body, platform); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif platform.ID == \"\" {\n\t\tUUID, err := uuid.NewV4()\n\t\tif err != nil {\n\t\t\tlogger.Error(\"Could not generate GUID\")\n\t\t\treturn nil, err\n\t\t}\n\t\tplatform.ID = UUID.String()\n\t}\n\tcurrentTime := time.Now().UTC()\n\tplatform.CreatedAt = currentTime\n\tplatform.UpdatedAt = currentTime\n\n\tcredentials, err := types.GenerateCredentials()\n\tif err != nil {\n\t\tlogger.Error(\"Could not generate credentials for platform\")\n\t\treturn nil, err\n\t}\n\tplainPassword := credentials.Basic.Password\n\ttransformedPassword, err := c.Encrypter.Encrypt(ctx, []byte(plainPassword))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tcredentials.Basic.Password = string(transformedPassword)\n\tplatform.Credentials = credentials\n\n\tif _, err := c.PlatformStorage.Create(ctx, platform); err != nil {\n\t\treturn nil, util.HandleStorageError(err, \"platform\")\n\t}\n\tplatform.Credentials.Basic.Password = plainPassword\n\treturn util.NewJSONResponse(http.StatusCreated, platform)\n}", "func (o *DesktopApp) Create() (*restapi.SliceResponse, error) {\n\tvar queryArg = make(map[string]interface{})\n\n\tqueryArg[\"ID\"] = []string{o.TemplateName}\n\tLogD.Printf(\"Generated Map for Create(): %+v\", queryArg)\n\n\tresp, err := o.client.CallSliceAPI(o.apiCreate, queryArg)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif !resp.Success {\n\t\treturn nil, errors.New(resp.Message)\n\t}\n\n\treturn resp, nil\n}", "func AppCreate(ctx *Context, name string, typ DeviceType) (*Application, error) {\n\th := authHeader(ctx.Config.AuthToken)\n\turi := ctx.Config.APIEndpoint(\"application\")\n\tdata := make(map[string]interface{})\n\tdata[\"app_name\"] = name\n\tdata[\"device_type\"] = typ.String()\n\tbody, err := marhsalReader(data)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tb, err := doJSON(ctx, \"POST\", uri, h, nil, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\trst := &Application{}\n\terr = json.Unmarshal(b, rst)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn rst, nil\n}", "func PostDevice(c *gin.Context) {\n\tvar device database.Device\n\tc.BindJSON(&device)\n\terr := database.InsertDevice(c, device)\n\tif err != nil {\n\t\tc.Status(http.StatusForbidden)\n\t} else {\n\t\tc.Status(http.StatusOK)\n\t}\n\n}", "func AddDevices() {\n\tcfg := config.Config()\n\tcameraAddr := cfg.GetString(\"camera-addr\")\n\trequestURL := fmt.Sprintf(\"%s\", cameraAddr)\n\n\tcameras, err := device.GetCameras()\n if err != nil {\n log.Info(\"Device Post Failed: \", err)\n return\n }\n\n for _, c := range cameras {\n pc := &schema.CaptureCamera{\n Device: c.DeviceName,\n Rtmp: c.Rtmp,\n }\n jsonValue, _ := json.Marshal(pc)\n\t resp, err := http.Post(requestURL,\"application/json\", bytes.NewBuffer(jsonValue))\n log.Info(\"Device Post: \", pc, \" result: \", resp, \" error: \", err)\n }\n}", "func (c *connections) CreateDevice(g *gin.Context) {\n\t// validate device\n\tvar dp DeviceProfile\n\terr := g.Bind(&dp)\n\tif err != nil {\n\t\tg.IndentedJSON(http.StatusBadRequest,\n\t\t\tErrorResponse{\n\t\t\t\tErrorType: \"PayloadDecodeError\",\n\t\t\t\tErrorMessage: err.Error()})\n\t\treturn\n\t}\n\n\t// Validate description data.\n\tif (dp.lat < -90.0) || (dp.lat > 90.0) || (dp.lon < -180.0) || (dp.lon > 180.0) {\n\t\tg.IndentedJSON(http.StatusBadRequest,\n\t\t\tErrorResponse{\n\t\t\t\tErrorType: \"LocationError\",\n\t\t\t\tErrorMessage: \"Missing or invalid latitude/longitude\"})\n\t\treturn\n\t}\n\n\tcdp := cassandra.DeviceProfile{\n\t\tName: dp.name,\n\t\tDeviceID: dp.deviceID,\n\t\tDeviceType: dp.deviceType,\n\t\tModelNumber: dp.modelNumber,\n\t\tLat: dp.lat,\n\t\tLon: dp.lon,\n\t}\n\t// Generate a random API Key for the device\n\tk := randomKey(8)\n\n\terr = c.csndra.AddDeviceData(&cdp)\n\tif err != nil {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"Error\": err,\n\t\t}).Error(\"Error adding Device\")\n\t\tg.IndentedJSON(http.StatusInternalServerError,\n\t\t\tErrorResponse{\n\t\t\t\tErrorType: \"DeviceCreationError\",\n\t\t\t\tErrorMessage: err.Error()})\n\t} else {\n\t\tg.IndentedJSON(http.StatusCreated,\n\t\t\tCreateSuccessResponse{\n\t\t\t\tDeviceID: cdp.DeviceName,\n\t\t\t\tAPIKey: k})\n\t}\n\n}", "func (a *appHandler) CreateApp(w http.ResponseWriter, r *http.Request) {\n\tvar app model.App\n\terr := json.NewDecoder(r.Body).Decode(&app)\n\tif err != nil {\n\t\ta.httpUtil.WriteJSONBadRequestResponse(w, err)\n\t\treturn\n\t}\n\n\t// TODO : Create\n\n\tjsonR, err := json.Marshal(app)\n\tif err != nil {\n\t\ta.httpUtil.WriteJSONInternalServerErrorResponse(w, err)\n\t}\n\n\ta.httpUtil.WriteJSONSuccessResponse(w, jsonR)\n}", "func (s *Service) registerDeviceHandler(w http.ResponseWriter, r *http.Request) {\n\t// Get the authenticated user from the request context\n\tauthenticatedUser, err := accounts.GetAuthenticatedUser(r)\n\tif err != nil {\n\t\tresponse.UnauthorizedError(w, err.Error())\n\t\treturn\n\t}\n\n\t// Request body cannot be nil\n\tif r.Body == nil {\n\t\tresponse.Error(w, \"Request body cannot be nil\", http.StatusBadRequest)\n\t\treturn\n\t}\n\n\t// Read the request body\n\tpayload, err := ioutil.ReadAll(r.Body)\n\tif err != nil {\n\t\tresponse.Error(w, \"Error reading request body\", http.StatusBadRequest)\n\t\treturn\n\t}\n\n\t// Unmarshal the request body into the request prototype\n\tdeviceRequest := new(DeviceRequest)\n\tif err := json.Unmarshal(payload, deviceRequest); err != nil {\n\t\tlogger.ERROR.Printf(\"Failed to unmarshal device request: %s\", payload)\n\t\tresponse.Error(w, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\t// Get the correct platform application ARN\n\tapplicationARN, ok := map[string]string{\n\t\tPlatformIOS: s.cnf.AWS.APNSPlatformApplicationARN,\n\t\tPlatformAndroid: s.cnf.AWS.GCMPlatformApplicationARN,\n\t}[deviceRequest.Platform]\n\tif !ok {\n\t\tresponse.Error(w, ErrPlatformNotSupported.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\t// Register a new endpoint for this device\n\t_, err = s.createOrUpdateEndpoint(\n\t\tauthenticatedUser,\n\t\tapplicationARN,\n\t\tdeviceRequest.Token,\n\t)\n\tif err != nil {\n\t\tresponse.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\t// 204 no content response\n\tresponse.NoContent(w)\n}", "func (h *Handler) serveCreateDBUser(w http.ResponseWriter, r *http.Request) {}", "func DeviceCreateHandler(\n\tparams wireguard.DeviceCreateParams,\n\tprincipal interface{},\n) middleware.Responder {\n\tname := *params.Device.Name\n\twgConfPath := getWgConfPath(name)\n\n\t// check if interface already exist\n\t_, err := os.Stat(wgConfPath)\n\tif err != nil && !os.IsNotExist(err) {\n\t\tmsg := fmt.Sprintf(\"os err: %s\", err.Error())\n\t\twgrest.Logger.Println(msg)\n\n\t\treturn wireguard.NewDeviceCreateDefault(http.StatusInternalServerError).WithPayload(\n\t\t\t&models.Error{Detail: msg},\n\t\t)\n\t} else if err == nil {\n\t\tmsg := fmt.Sprintf(\"file %s exists\", wgConfPath)\n\t\treturn wireguard.NewDeviceCreateConflict().WithPayload(\n\t\t\t&models.Error{Detail: msg},\n\t\t)\n\t}\n\n\tclient, err := wgctrl.New()\n\tif err != nil {\n\t\tmsg := fmt.Sprintf(\"wgctrl err: %s\", err.Error())\n\t\twgrest.Logger.Println(msg)\n\n\t\treturn wireguard.NewDeviceCreateDefault(http.StatusInternalServerError).WithPayload(\n\t\t\t&models.Error{Detail: msg},\n\t\t)\n\t}\n\n\tprivateKey, err := wgtypes.ParseKey(*params.Device.PrivateKey)\n\tif err != nil {\n\t\tmsg := fmt.Sprintf(\"wgctrl err: %s\", err.Error())\n\t\twgrest.Logger.Println(msg)\n\n\t\treturn wireguard.NewDeviceCreateDefault(http.StatusInternalServerError).WithPayload(\n\t\t\t&models.Error{Detail: msg},\n\t\t)\n\t}\n\n\tlistenPort := int(*params.Device.ListenPort)\n\n\tla := netlink.NewLinkAttrs()\n\tla.Name = name\n\n\twgDev := &netlink.GenericLink{\n\t\tLinkAttrs: la,\n\t\tLinkType: \"wireguard\",\n\t}\n\n\terr = netlink.LinkAdd(wgDev)\n\tif err != nil {\n\t\tmsg := fmt.Sprintf(\"netlink err: %s\", err.Error())\n\t\twgrest.Logger.Println(msg)\n\n\t\treturn wireguard.NewDeviceCreateDefault(http.StatusInternalServerError).WithPayload(\n\t\t\t&models.Error{Detail: msg},\n\t\t)\n\t}\n\n\tnet, err := netlink.ParseAddr(*params.Device.Network)\n\tif err != nil {\n\t\tmsg := fmt.Sprintf(\"netlink err: %s\", err.Error())\n\t\twgrest.Logger.Println(msg)\n\n\t\treturn wireguard.NewDeviceCreateDefault(http.StatusInternalServerError).WithPayload(\n\t\t\t&models.Error{Detail: msg},\n\t\t)\n\t}\n\n\terr = netlink.AddrAdd(wgDev, net)\n\tif err != nil {\n\t\tmsg := fmt.Sprintf(\"netlink err: %s\", err.Error())\n\t\twgrest.Logger.Println(msg)\n\n\t\treturn wireguard.NewDeviceCreateDefault(http.StatusInternalServerError).WithPayload(\n\t\t\t&models.Error{Detail: msg},\n\t\t)\n\t}\n\n\tcfg := wgtypes.Config{\n\t\tPrivateKey: &privateKey,\n\t\tListenPort: &listenPort,\n\t}\n\n\terr = client.ConfigureDevice(name, cfg)\n\tif err != nil {\n\t\tmsg := fmt.Sprintf(\"wgctrl err: %s\", err.Error())\n\t\twgrest.Logger.Println(msg)\n\n\t\treturn wireguard.NewDeviceCreateDefault(http.StatusInternalServerError).WithPayload(\n\t\t\t&models.Error{Detail: msg},\n\t\t)\n\t}\n\n\terr = netlink.LinkSetUp(wgDev)\n\tif err != nil {\n\t\tmsg := fmt.Sprintf(\"netlink err: %s\", err.Error())\n\t\twgrest.Logger.Println(msg)\n\n\t\treturn wireguard.NewDeviceCreateDefault(http.StatusInternalServerError).WithPayload(\n\t\t\t&models.Error{Detail: msg},\n\t\t)\n\t}\n\n\tst := storage.NewStorage(storage.DiskStorage)\n\trwc, err := st.Open(wgConfPath)\n\tif err != nil {\n\t\tmsg := fmt.Sprintf(\"storage err: %s\", err.Error())\n\t\twgrest.Logger.Println(msg)\n\n\t\treturn wireguard.NewDeviceCreateDefault(http.StatusInternalServerError).WithPayload(\n\t\t\t&models.Error{Detail: msg},\n\t\t)\n\t}\n\tdefer rwc.Close()\n\n\tfmt.Fprintf(rwc, \"[Interface]\\n\")\n\tfmt.Fprintf(rwc, \"Address = %s\\n\", *params.Device.Network)\n\tfmt.Fprintf(rwc, \"ListenPort = %v\\n\", *params.Device.ListenPort)\n\tfmt.Fprintf(rwc, \"PrivateKey = %s\\n\", *params.Device.PrivateKey)\n\tfmt.Fprintf(rwc, \"SaveConfig = true\\n\\n\")\n\n\tscheme := \"https\"\n\tif params.HTTPRequest.URL.Scheme != \"\" {\n\t\tscheme = params.HTTPRequest.URL.Scheme\n\t}\n\n\tlocation := fmt.Sprintf(\"%s://%s%s%s\", scheme, params.HTTPRequest.Host, params.HTTPRequest.RequestURI, name)\n\treturn wireguard.NewDeviceCreateCreated().WithLocation(location)\n}", "func (h *DeviceHandler) DeviceAdd(c context.Context, sessionID int) error {\n\tuis := libkb.UIs{\n\t\tProvisionUI: h.getProvisionUI(sessionID),\n\t\tSecretUI: h.getSecretUI(sessionID, h.G()),\n\t\tSessionID: sessionID,\n\t}\n\tm := libkb.NewMetaContext(c, h.G()).WithUIs(uis)\n\teng := engine.NewDeviceAdd(h.G())\n\treturn engine.RunEngine2(m, eng)\n}", "func CreateType(e echo.Context) error {\n\ttipe := db.Types{}\n\te.Bind(&tipe)\n\n\tif err := config.DB.Save(&tipe).Error; err != nil {\n\t\treturn echo.NewHTTPError(http.StatusBadRequest, err.Error())\n\t}\n\treturn e.JSON(http.StatusOK, map[string]interface{}{\n\t\t\"message\": \"Berhasil menambahkan tipe smartphone\",\n\t\t\"tipe\": tipe,\n\t})\n}", "func ApplicationCreate(w http.ResponseWriter, r *http.Request) {\n\tdb, err := database.Connect()\n\tdefer db.Close()\n\tif err != nil {\n\t\tlog.Printf(\"Database error: '%s'\\n\", err)\n\t\thttp.Error(w, \"there was an error when attempting to connect to the database\", http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tvar createForm struct {\n\t\tName string\n\t}\n\tdecoder := json.NewDecoder(r.Body)\n\terr = decoder.Decode(&createForm)\n\tdefer r.Body.Close()\n\tif err != nil {\n\t\tlog.Printf(\"decoding error: '%s'\\n\", err)\n\t\thttp.Error(w, \"there was an error when attempting to parse the form\", http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tapp := resources.Application{\n\t\tName: createForm.Name,\n\t}\n\t_, err = resources.CreateApplication(db, &app)\n\t// @todo handle failed save w/out error?\n\tif err != nil {\n\t\tlog.Printf(\"Error when retrieving application: '%s'\\n\", err)\n\t\thttp.Error(w, \"there was an error when retrieving application\", http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\t// @todo return some sort of content?\n\tw.WriteHeader(http.StatusCreated)\n\treturn\n}", "func CreateDevice(tab *DeviceTable, logger hasPrintf, modelName, id, hostPort, transports, user, pass, enable string, debug bool, change *conf.Change) error {\n\tlogger.Printf(\"CreateDevice: %s %s %s %s\", modelName, id, hostPort, transports)\n\n\tmod, getErr := tab.GetModel(modelName)\n\tif getErr != nil {\n\t\terr := fmt.Errorf(\"CreateDevice: could not find model '%s': %v\", modelName, getErr)\n\t\tlogger.Printf(err.Error())\n\t\treturn err\n\t}\n\n\td := NewDevice(logger, mod, id, hostPort, transports, user, pass, enable, debug)\n\n\tif change != nil {\n\t\td.LastChange = *change\n\t}\n\n\tif newDevErr := tab.SetDevice(d); newDevErr != nil {\n\t\terr := fmt.Errorf(\"CreateDevice: could not add device '%s': %v\", id, newDevErr)\n\t\tlogger.Printf(err.Error())\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *Controller) renderNew(ctx context.Context, w http.ResponseWriter, app *database.MobileApp) {\n\tm := templateMap(ctx)\n\tm.Title(\"New mobile app\")\n\tm[\"app\"] = app\n\tc.h.RenderHTML(w, \"mobileapps/new\", m)\n}", "func (db *MongoDB) CreateDevice(c echo.Context) (err error) {\n\t// validate data\n\n\tfile, src, _ := openFile(c)\n\timgName, filePath := genImgID(file.Filename)\n\tcreateFile(filePath, src)\n\n\tm, err := db.insertDeviceDB(c, imgName)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn c.JSON(http.StatusCreated, m)\n}", "func (d *CreateCore) CreateDevice(entity models.Device) error {\n\tdevice, _ := dynamodbattribute.MarshalMap(entity)\n\tinput := &dynamodb.PutItemInput{\n\t\tItem: device,\n\t\tTableName: aws.String(os.Getenv(\"TABLE_NAME\")),\n\t}\n\t_, err := d.db.PutItem(input)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn errors.New(\"server error\")\n\t}\n\treturn nil\n}", "func TestCreateSuccess(t *testing.T) {\n\tisTesting = true\n\tvar request = Request{\n\t\tPath: \"/api/devices\",\n\t\tHTTPMethod: \"POST\",\n\t\tBody: `{\n\t\t\t\"id\": \"valid-id\",\n\t\t\t\"deviceModel\": \"valid-device-model\",\n\t\t\t\"model\": \"valid-model\",\n\t\t\t\"name\": \"valid-name\",\n\t\t\t\"serial\": \"valid-serial\"\n\t\t}`,\n\t}\n\tvar response, _ = Handler(request)\n\tif response.StatusCode != 201 {\n\t\tt.Errorf(\"response status code has to be 201\")\n\t}\n}", "func registerNewDevice(deviceType string, deviceId string, deviceName string, email string) bool {\n\tdb := connect()\n\toutput := true\n\tvar res []mysql.Row\n\tif strings.Contains(deviceType, \"'\") || strings.Contains(deviceId, \"'\") || strings.Contains(deviceName, \"'\") || strings.Contains(email, \"'\") {\n\t\tdeviceType = strings.Replace(deviceType, \"'\", \"\\\\'\", -1)\n\t\tdeviceId = strings.Replace(deviceId, \"'\", \"\\\\'\", -1)\n\t\tdeviceName = strings.Replace(deviceName, \"'\", \"\\\\'\", -1)\n\t\temail = strings.Replace(email, \"'\", \"\\\\'\", -1)\n\t}\n\tif deviceType == \"gps\" {\n\t\tres, _, _ = db.Query(\"SELECT * FROM gpsDevice WHERE deviceId = '\" + deviceId + \"'\")\n\t} else if deviceType == \"laptop\" {\n\t\tres, _, _ = db.Query(\"SELECT * FROM laptopDevice WHERE deviceId = '\" + deviceId + \"'\")\n\t}\n\n\tif len(res) != 0 {\n\t\toutput = false\n\t\tfmt.Println(\"check\")\n\t} else {\n\t\tif deviceType != \"gps\" && deviceType != \"laptop\" {\n\t\t\tprint(\"invalid device type\")\n\t\t} else {\n\t\t\tif deviceType == \"gps\" {\n\t\t\t\tfmt.Println(\"Writing to gpsDevice...\")\n\t\t\t\tdb.Query(\"INSERT INTO gpsDevice (deviceName, deviceId, customerId) SELECT '\" + deviceName + \"', '\" + deviceId + \"', id FROM customer WHERE email='\" + email + \"'\")\n\t\t\t} else if deviceType == \"laptop\" {\n\t\t\t\tfmt.Println(\"Writing to laptopDevice...\")\n\t\t\t\tdb.Query(\"INSERT INTO laptopDevice (deviceName, deviceId, customerId) SELECT '\" + deviceName + \"', '\" + deviceId + \"', id FROM customer WHERE email='\" + email + \"'\")\n\t\t\t}\n\t\t}\n\t}\n\tdisconnect(db)\n\n\treturn output\n}", "func (ctx *HandlerContext) DevicesHandler(w http.ResponseWriter, r *http.Request) {\n\tif r.Method != \"POST\" && r.Method != \"PATCH\" {\n\t\thttp.Error(w, \"method must be Post or PATCH\", http.StatusMethodNotAllowed)\n\t\treturn\n\t}\n\n\tif r.Header.Get(headerContentType) != contentTypeJSON {\n\t\thttp.Error(w, \"content type must be application/json\", http.StatusUnsupportedMediaType)\n\t\treturn\n\t}\n\n\tif r.Method == \"POST\" {\n\t\tnewDevice := &devices.NewDevice{}\n\t\tif err := json.NewDecoder(r.Body).Decode(newDevice); err != nil {\n\t\t\thttp.Error(w, fmt.Sprintf(\"error decoding JSON: %v\", err), http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\n\t\tdevice, err := newDevice.ToDevice()\n\t\tif err != nil {\n\t\t\thttp.Error(w, fmt.Sprintf(\"error creating new device: %v\", err), http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\n\t\tif _, err := ctx.deviceStore.GetByName(device.Name); err == nil {\n\t\t\thttp.Error(w, fmt.Sprintf(\"device name already exists\"), http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\n\t\tdevice, err = ctx.deviceStore.Insert(device)\n\t\tif err != nil {\n\t\t\thttp.Error(w, fmt.Sprintf(\"error adding device: %v\", err), http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\n\t\tnewSession := &SessionState{StartTime: time.Now(), Device: device}\n\t\tif _, err := sessions.BeginSession(ctx.SigningKey, ctx.SessStore, newSession, w); err != nil {\n\t\t\thttp.Error(w, fmt.Sprintf(\"error creating new session: %v\", err), http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\tVerify(\"+1\"+device.Phone, trialNum, \"sup\")\n\t\trespond(w, device, http.StatusCreated, ctx.PubVapid)\n\t} else {\n\t\ttype VerificationCheck struct {\n\t\t\tCode string `json:\"code\"`\n\t\t\tPhone string `json:\"phone\"`\n\t\t}\n\n\t\tnewVerificationCheck := &VerificationCheck{}\n\t\tif err := json.NewDecoder(r.Body).Decode(newVerificationCheck); err != nil {\n\t\t\thttp.Error(w, fmt.Sprintf(\"error decoding JSON into VerificationCheck: %v\", err), http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\n\t\terr := CheckVerification(newVerificationCheck.Code, \"+1\"+newVerificationCheck.Phone)\n\t\tif err != nil {\n\t\t\thttp.Error(w, fmt.Sprintf(\"error sending VerificationCheck: %v\", err), 500)\n\t\t}\n\n\t}\n\n}", "func Create(w http.ResponseWriter, r *http.Request, _ httprouter.Params) {\n\tfmt.Fprint(w, \"Welcome!\\n\")\n}", "func createServer(w http.ResponseWriter, r *http.Request) {\n w.Header().Set(\"Content-Type\", \"application/json\")\n\n\tvar newServer server\n\treqBody, err := ioutil.ReadAll(r.Body)\n\tif err != nil {\n\t fmt.Println(err)\n\t\tfmt.Fprintf(w, \"Kindly enter data with the server's address, MSA and MTA network addresses only in order to create new server\")\n\t w.WriteHeader(http.StatusInternalServerError)\n\t return\n\t}\n\tnewServer.ID = strconv.Itoa(len(servers)+1)\n\n\tjson.Unmarshal(reqBody, &newServer)\n\tservers = append(servers, newServer)\n\tw.WriteHeader(http.StatusCreated)\n\n\tjson.NewEncoder(w).Encode(newServer)\n}", "func(db *Persistence) CreateNewApplication(appName, description, redirect string,\n trimName bool) error {\n\n log.Debug(\"adding new application to datbase...\")\n appId := uuid.New()\n\n query := `INSERT INTO applications(application_id,application_name,description,\n redirect_url,trim_app_name) VALUES($1,$2,$3,$4,$5)`\n _, err := db.Session.Exec(context.Background(), query, appId, appName, description,\n redirect, trimName)\n return err\n}", "func CreateDesains(e echo.Context) error {\n\tdesain := db.Desains{}\n\te.Bind(&desain)\n\n\tif err := config.DB.Save(&desain).Error; err != nil {\n\t\treturn echo.NewHTTPError(http.StatusBadRequest, err.Error())\n\t}\n\treturn e.JSON(http.StatusOK, map[string]interface{}{\n\t\t\"message\": \"Berhasil menambahkan desain smartphone\",\n\t\t\"desain\": desain,\n\t})\n}", "func (b *Bridge) create() error {\n\tcmd := exec.Command(\"ifconfig\", b.Device, \"create\")\n\tfmt.Printf(\"cmd: %s\\n\", strings.Join(cmd.Args, \" \"))\n\treturn cmd.Run()\n}", "func Create(writer http.ResponseWriter, request *http.Request) {\n\ttemplate_html.ExecuteTemplate(writer, \"Create\", nil)\n}", "func (c *RestClient) CreateApp(name string, memory int) (string, error) {\n\t// Ensure that app name is unique for this user. We do this as\n\t// unfortunately the server doesn't enforce it.\n\tapps, err := c.ListApps()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tfor _, app := range apps {\n\t\tif app.Name == name {\n\t\t\treturn \"\", fmt.Errorf(\"App by that name (%s) already exists\", name)\n\t\t}\n\t}\n\n\t// The CC requires that a POST on /apps sends, at minimum, these\n\t// fields. The values for framework/runtime doesn't matter for our\n\t// purpose (they will get overwritten by a subsequent app push).\n\tcreateArgs := map[string]interface{}{\n\t\t\"name\": name,\n\t\t\"space_guid\": c.Space,\n\t\t\"memory\": memory,\n\t}\n\n\tvar resp struct {\n\t\tMetadata struct {\n\t\t\tGUID string\n\t\t}\n\t}\n\terr = c.MakeRequest(\"POST\", \"/v2/apps\", createArgs, &resp)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif resp.Metadata.GUID == \"\" {\n\t\treturn \"\", fmt.Errorf(\"Missing App GUID from CC\")\n\t}\n\n\treturn resp.Metadata.GUID, nil\n}", "func (r *app) Create(appPayload AppRequest, opts ...bool) (*AppFields, error) {\n\tasync := true\n\tif len(opts) > 0 {\n\t\tasync = opts[0]\n\t}\n\trawURL := fmt.Sprintf(\"/v2/apps?async=%t\", async)\n\tappFields := AppFields{}\n\t_, err := r.client.Post(rawURL, appPayload, &appFields)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &appFields, nil\n}", "func AdminCreate(w http.ResponseWriter, data interface{}) {\n\trender(tpAdminCreate, w, data)\n}", "func AdminCreate(w http.ResponseWriter, data interface{}) {\n\trender(tpAdminCreate, w, data)\n}", "func (c Client) createApp(body io.Reader) (*App, error) {\n\treq, err := http.NewRequest(\"POST\", c.getURL(\"/apps\"), body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar marshalled map[string]*App\n\terr = c.executeRequestAndMarshal(req, &marshalled)\n\treturn marshalled[\"app\"], err\n}", "func create_app() (string) {\n\t// Create a new LUIS app\n\t// Get the context, which is required by the SDK methods.\n\tctx := context.Background()\n\n\tclient := authoring.NewAppsClient(authoring_endpoint)\n\t// Set the subscription key on the client.\n\tclient.Authorizer = autorest.NewCognitiveServicesAuthorizer(authoring_key)\n\n\tdomain_name := \"HomeAutomation\"\n\tculture := \"en-us\"\n\n\tcreate_app_payload := authoring.PrebuiltDomainCreateObject { DomainName: &domain_name, Culture: &culture }\n\n\tresult, err := client.AddCustomPrebuiltDomain(ctx, create_app_payload)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tfmt.Printf(\"Created LUIS app %s with ID %s\\n\", domain_name, (*result.Value).String())\n\treturn (*result.Value).String()\n}", "func (f *DMAClient) Create(obj *dm.M4DApplication) (*dm.M4DApplication, error) {\n\tvar result dm.M4DApplication\n\terr := f.client.Post().\n\t\tNamespace(f.namespace).Resource(f.plural).\n\t\tBody(obj).Do(context.Background()).Into(&result)\n\treturn &result, err\n}", "func PostDevice(context *gin.Context) {\n\tapplication.DeviceRepository = infrastructure.NewDeviceRepository()\n\tvar dto DeviceDto\n\tvar device *domain.Device\n\n\terr := context.BindJSON(&dto)\n\tif err == nil {\n\t\tdevice, err = application.CreateDevice(dto.Id, dto.Os, &dto.SenderId)\n\t}\n\tif err != nil {\n\t\tapiError(context, err)\n\t\treturn\n\t}\n\tapiSuccess(context, http.StatusCreated, device)\n}", "func (a *Api) registerDevice(c *gin.Context) {\n\tdecoder := json.NewDecoder(c.Request.Body)\n\tdefer c.Request.Body.Close()\n\tvar postDevice PostDevice\n\terr := decoder.Decode(&postDevice)\n\tif err != nil {\n\t\tinternalError(c, \"marshalling error\",\n\t\t\t\"marshalling error\"+err.Error())\n\t}\n\tif err = a.ms.RegisterDevice(postDevice.DeviceNumber, time.Now()); err != nil {\n\t\tinternalError(c, \"register err\",\n\t\t\t\"register err\"+err.Error())\n\t}\n\tc.JSON(http.StatusOK, gin.H{\"message\": \"Device registered\"})\n}", "func CreateFromDevice(d *v1alpha2.Device, projectID string) *packngo.DeviceCreateRequest {\n\tips := []packngo.IPAddressCreateRequest{}\n\tfor _, ip := range d.Spec.ForProvider.IPAddresses {\n\t\tips = append(ips, packngo.IPAddressCreateRequest{\n\t\t\tAddressFamily: ip.AddressFamily,\n\t\t\tPublic: ip.Public,\n\t\t\tCIDR: ip.CIDR,\n\t\t\tReservations: ip.Reservations,\n\t\t})\n\t}\n\n\tr := &packngo.DeviceCreateRequest{\n\t\tHostname: emptyIfNil(d.Spec.ForProvider.Hostname),\n\t\tPlan: d.Spec.ForProvider.Plan,\n\t\tFacility: []string{d.Spec.ForProvider.Facility},\n\t\tMetro: d.Spec.ForProvider.Metro,\n\t\tOS: d.Spec.ForProvider.OS,\n\t\tBillingCycle: emptyIfNil(d.Spec.ForProvider.BillingCycle),\n\t\tProjectID: projectID,\n\t\tUserData: emptyIfNil(d.Spec.ForProvider.UserData),\n\t\tTags: d.Spec.ForProvider.Tags,\n\t\tIPAddresses: ips,\n\t\tCustomData: emptyIfNil(d.Spec.ForProvider.CustomData),\n\t\tIPXEScriptURL: emptyIfNil(d.Spec.ForProvider.IPXEScriptURL),\n\t\tPublicIPv4SubnetSize: zeroIfNil(d.Spec.ForProvider.PublicIPv4SubnetSize),\n\t\tAlwaysPXE: falseIfNil(d.Spec.ForProvider.AlwaysPXE),\n\t\tHardwareReservationID: emptyIfNil(d.Spec.ForProvider.HardwareReservationID),\n\t\tFeatures: d.Spec.ForProvider.Features,\n\t\tUserSSHKeys: d.Spec.ForProvider.UserSSHKeys,\n\t\tProjectSSHKeys: d.Spec.ForProvider.ProjectSSHKeys,\n\n\t\t// TODO:\n\t\t// Storage\n\t\t// SpotInstance\n\t\t// SpotPriceMax\n\t\t// TerminationTime\n\t}\n\n\treturn r\n}", "func (c *SmartThingsClient) DeviceCreateEvent(deviceID string, command ...DeviceEvent) error {\n\ts := make([]DeviceEvent, 0)\n\tfor _, c := range command {\n\t\ts = append(s, c)\n\t}\n\tb := &DeviceEventList{\n\t\tDeviceEvents: s,\n\t}\n\treq, err := c.newRequest(http.MethodPost, fmt.Sprintf(\"/v1/devices/%s/events\", deviceID), b)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = c.do(req, nil)\n\treturn err\n}", "func (webType WebType) Create() error {\n\tlog.WithFields(log.Fields{}).Debug(\"Creating client app\")\n\n\terr := webType.Client.Create()\n\n\tif err != nil {\n\t\treturn err\n\t}\n\tif !webType.includeBackend {\n\t\treturn nil\n\t}\n\tlog.WithFields(log.Fields{}).Debug(\"Creating backend app\")\n\n\treturn webType.Server.Create()\n}", "func HandleInstanceCreate(w rest.ResponseWriter, r *rest.Request) {\n\t// get ima\n\tima := Ima{}\n\terr := r.DecodeJsonPayload(&ima)\n\tif err != nil {\n\t\trest.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\tif ima.Ima == \"\" {\n\t\trest.Error(w, \"ima required\", 400)\n\t\treturn\n\t}\n\tif ima.Mem == 0 {\n\t\trest.Error(w, \"memory required\", 400)\n\t\treturn\n\t}\n\tif ima.Cpu == 0 {\n\t\trest.Error(w, \"cpu required\", 400)\n\t\treturn\n\t}\n\n\t// start the instance\n\tos := getImaOs(ima.Ima)\n\tswitch os {\n\tcase \"freebsd\":\n\t\t// clone ima to instance\n\t\tinstanceid := allocateInstanceId()\n\t\tcloneIma(ima.Ima, instanceid)\n\n\t\t// create network interface and bring up\n\t\ttap := allocateTap()\n\t\tif tap == \"\" {\n\t\t\treturn\n\t\t}\n\t\tsaveTap(tap, instanceid)\n\t\tbridge := findBridge()\n\t\taddTapToBridge(tap, bridge)\n\t\tbridgeUp(bridge)\n\n\t\tnmdm := \"/dev/nmdm-\" + instanceid + \"-A\"\n\t\tsaveCpu(ima.Cpu, instanceid)\n\t\tsaveMem(ima.Mem, instanceid)\n\t\tgo startFreeBSDVM(nmdm, ima.Cpu, ima.Mem, tap, instanceid)\n\t\tw.WriteJson(&instanceid)\n\tcase \"linux\":\n\t\t// clone ima to instance\n\t\tinstanceid := allocateInstanceId()\n\t\tcloneIma(ima.Ima, instanceid)\n\n\t\t// create network interface and bring up\n\t\ttap := allocateTap()\n\t\tif tap == \"\" {\n\t\t\treturn\n\t\t}\n\t\tsaveTap(tap, instanceid)\n\t\tbridge := findBridge()\n\t\taddTapToBridge(tap, bridge)\n\t\tbridgeUp(bridge)\n\n\t\t//nmdm := \"/dev/nmdm-\" + instanceid + \"-A\"\n\t\tsaveCpu(ima.Cpu, instanceid)\n\t\tsaveMem(ima.Mem, instanceid)\n\t\tbhyveDestroy(instanceid)\n\t\tnmdm := \"/dev/nmdm-\" + instanceid + \"-A\"\n\t\tgo startLinuxVM(nmdm, ima.Cpu, ima.Mem, tap, instanceid)\n\t\tw.WriteJson(&instanceid)\n\tdefault:\n\t\trest.Error(w, \"unknown OS\", 400)\n\t}\n}", "func (f *FUOTADeploymentAPI) CreateForDevice(ctx context.Context, req *pb.CreateFUOTADeploymentForDeviceRequest) (*pb.CreateFUOTADeploymentForDeviceResponse, error) {\n\tif req.FuotaDeployment == nil {\n\t\treturn nil, status.Errorf(codes.InvalidArgument, \"fuota_deployment must not be nil\")\n\t}\n\n\tvar devEUI lorawan.EUI64\n\tif err := devEUI.UnmarshalText([]byte(req.DevEui)); err != nil {\n\t\treturn nil, status.Errorf(codes.InvalidArgument, err.Error())\n\t}\n\n\tif valid, err := fuotaCred.NewValidator().ValidateFUOTADeploymentsAccess(ctx, auth.Create, 0, devEUI); !valid || err != nil {\n\t\treturn nil, status.Errorf(codes.Unauthenticated, \"authentication failed: %s\", err)\n\t}\n\n\tn, err := f.st.GetNetworkServerForDevEUI(ctx, devEUI)\n\tif err != nil {\n\t\treturn nil, helpers.ErrToRPCError(err)\n\t}\n\n\tnsClient, err := networkserver.GetPool().Get(n.Server, []byte(n.CACert), []byte(n.TLSCert), []byte(n.TLSKey))\n\tif err != nil {\n\t\treturn nil, helpers.ErrToRPCError(err)\n\t}\n\n\tversionResp, err := nsClient.GetVersion(ctx, &empty.Empty{})\n\tif err != nil {\n\t\treturn nil, helpers.ErrToRPCError(err)\n\t}\n\n\tvar b band.Band\n\n\tswitch versionResp.Region {\n\tcase common.Region_EU868:\n\t\tb, err = band.GetConfig(band.EU868, false, lorawan.DwellTimeNoLimit)\n\t\tif err != nil {\n\t\t\treturn nil, helpers.ErrToRPCError(err)\n\t\t}\n\tcase common.Region_US915:\n\t\tb, err = band.GetConfig(band.US915, false, lorawan.DwellTimeNoLimit)\n\t\tif err != nil {\n\t\t\treturn nil, helpers.ErrToRPCError(err)\n\t\t}\n\tcase common.Region_CN779:\n\t\tb, err = band.GetConfig(band.CN779, false, lorawan.DwellTimeNoLimit)\n\t\tif err != nil {\n\t\t\treturn nil, helpers.ErrToRPCError(err)\n\t\t}\n\tcase common.Region_EU433:\n\t\tb, err = band.GetConfig(band.EU433, false, lorawan.DwellTimeNoLimit)\n\t\tif err != nil {\n\t\t\treturn nil, helpers.ErrToRPCError(err)\n\t\t}\n\tcase common.Region_AU915:\n\t\tb, err = band.GetConfig(band.AU915, false, lorawan.DwellTimeNoLimit)\n\t\tif err != nil {\n\t\t\treturn nil, helpers.ErrToRPCError(err)\n\t\t}\n\tcase common.Region_CN470:\n\t\tb, err = band.GetConfig(band.CN470, false, lorawan.DwellTimeNoLimit)\n\t\tif err != nil {\n\t\t\treturn nil, helpers.ErrToRPCError(err)\n\t\t}\n\tcase common.Region_AS923:\n\t\tb, err = band.GetConfig(band.AS923, false, lorawan.DwellTimeNoLimit)\n\t\tif err != nil {\n\t\t\treturn nil, helpers.ErrToRPCError(err)\n\t\t}\n\tcase common.Region_KR920:\n\t\tb, err = band.GetConfig(band.KR920, false, lorawan.DwellTimeNoLimit)\n\t\tif err != nil {\n\t\t\treturn nil, helpers.ErrToRPCError(err)\n\t\t}\n\tcase common.Region_IN865:\n\t\tb, err = band.GetConfig(band.IN865, false, lorawan.DwellTimeNoLimit)\n\t\tif err != nil {\n\t\t\treturn nil, helpers.ErrToRPCError(err)\n\t\t}\n\tcase common.Region_RU864:\n\t\tb, err = band.GetConfig(band.RU864, false, lorawan.DwellTimeNoLimit)\n\t\tif err != nil {\n\t\t\treturn nil, helpers.ErrToRPCError(err)\n\t\t}\n\tdefault:\n\t\treturn nil, status.Errorf(codes.Internal, \"region %s is not implemented\", versionResp.Region)\n\t}\n\n\tmaxPLSize, err := b.GetMaxPayloadSizeForDataRateIndex(\"\", \"\", int(req.FuotaDeployment.Dr))\n\tif err != nil {\n\t\treturn nil, helpers.ErrToRPCError(err)\n\t}\n\n\tfd := FUOTADeployment{\n\t\tName: req.FuotaDeployment.Name,\n\t\tDR: int(req.FuotaDeployment.Dr),\n\t\tFrequency: int(req.FuotaDeployment.Frequency),\n\t\tPayload: req.FuotaDeployment.Payload,\n\t\tFragSize: maxPLSize.N - 3,\n\t\tRedundancy: int(req.FuotaDeployment.Redundancy),\n\t\tMulticastTimeout: int(req.FuotaDeployment.MulticastTimeout),\n\t}\n\n\tswitch req.FuotaDeployment.GroupType {\n\tcase pb.MulticastGroupType_CLASS_C:\n\t\tfd.GroupType = FUOTADeploymentGroupTypeC\n\tdefault:\n\t\treturn nil, status.Errorf(codes.InvalidArgument, \"group_type %s is not supported\", req.FuotaDeployment.GroupType)\n\t}\n\n\tif err := req.FuotaDeployment.UnicastTimeout.CheckValid(); err != nil {\n\t\treturn nil, status.Errorf(codes.InvalidArgument, \"unicast_timeout: %s\", err)\n\t}\n\tfd.UnicastTimeout = req.FuotaDeployment.UnicastTimeout.AsDuration()\n\n\terr = f.st.Tx(ctx, func(ctx context.Context, handler *store.Handler) error {\n\t\treturn handler.CreateFUOTADeploymentForDevice(ctx, &fd, devEUI)\n\t})\n\tif err != nil {\n\t\treturn nil, helpers.ErrToRPCError(err)\n\t}\n\n\treturn &pb.CreateFUOTADeploymentForDeviceResponse{\n\t\tId: fd.ID.String(),\n\t}, nil\n}", "func createHandler(w http.ResponseWriter, r *http.Request) *appError {\n\tvideo, err := videoFromForm(r)\n\tif err != nil {\n\t\treturn appErrorf(err, \"could not parse video from form: %v\", err)\n\t}\n\tid, err := bittube.DB.AddVideo(video)\n\tif err != nil {\n\t\treturn appErrorf(err, \"could not save video: %v\", err)\n\t}\n\t// go publishUpdate(id)\n\thttp.Redirect(w, r, fmt.Sprintf(\"/videos/%d\", id), http.StatusFound)\n\treturn nil\n}", "func (ac *AdminCtl) Create() {\n\tif ac.Ctx.Request().Method == \"POST\" {\n\t\tnumber, _ := strconv.Atoi(ac.Ctx.Request().FormValue(\"number\"))\n\t\tname := ac.Ctx.Request().FormValue(\"name\")\n\t\tmtype := ac.Ctx.Request().FormValue(\"type\")\n\t\tcount, _ := strconv.Atoi(strings.Trim(ac.Ctx.Request().FormValue(\"count\"), \" \")) // 去除空白字符\n\t\tprice, _ := strconv.Atoi(ac.Ctx.Request().FormValue(\"price\"))\n\t\thref := ac.Ctx.Request().FormValue(\"href\")\n\t\turl := ac.Ctx.Request().FormValue(\"url\")\n\t\tsnumber := ac.Ctx.Request().FormValue(\"number\")\n\n\t\tproduct := &models.Product{\n\t\t\tNumber: number,\n\t\t\tName: name,\n\t\t\tType: mtype,\n\t\t\tCount: count,\n\t\t\tPrice: price,\n\t\t\tHref: href,\n\t\t\tURL: url,\n\t\t\tBrief: \"/data/\" + snumber + \"/brief\",\n\t\t\tDetail: \"/data/\" + snumber + \"/detail\",\n\t\t}\n\t\tac.Ctx.DB.Create(&product)\n\t\tac.Ctx.Redirect(\"/admin\", http.StatusFound)\n\t} else {\n\t\tac.Ctx.Data[\"AddPage\"] = true\n\t\tac.Ctx.Template = \"admin-add\"\n\t\tac.HTML(http.StatusOK)\n\t}\n}", "func (app *application) listingCreate(w http.ResponseWriter, r *http.Request) {\n\tsession, err := app.sessionStore.Get(r, \"session-name\")\n\tif err != nil {\n\t\tapp.serverError(w, err)\n\t\treturn\n\t}\n\terr = r.ParseForm()\n\tif err != nil {\n\t\tapp.clientError(w, http.StatusBadRequest)\n\t\treturn\n\t}\n\tform := forms.New(r.PostForm)\n\tform.Required(\"name\", \"description\", \"price\")\n\tif !form.Valid() {\n\t\tapp.render(w, r, \"listingcreate.page.tmpl\", &templateData{Form: form})\n\t\treturn\n\t}\n\tvendorID := app.authenticatedVendor(r)\n\tprice, err := strconv.Atoi(form.Get(\"price\"))\n\tif err != nil {\n\t\tform.Errors.Add(\"price\", \"enter valid integer\")\n\t}\n\terr = app.listings.Insert(\n\t\tvendorID,\n\t\tprice,\n\t\tform.Get(\"description\"),\n\t\tform.Get(\"name\"),\n\t)\n\tif err != nil {\n\t\tapp.serverError(w, err)\n\t}\n\n\tsession.AddFlash(\"Succesful Listed\")\n\terr = session.Save(r, w)\n\tif err != nil {\n\t\tapp.serverError(w, err)\n\t}\n\thttp.Redirect(w, r, \"/vendor/listings\", http.StatusSeeOther)\n}", "func (c *Client) CreateApp(user, name string) (*App, error) {\n\tlog.Printf(\"[INFO] creating application %s/%s\", user, name)\n\n\tbody, err := json.Marshal(&appWrapper{&App{\n\t\tUser: user,\n\t\tName: name,\n\t}})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tendpoint := \"/api/v1/vagrant/applications\"\n\trequest, err := c.Request(\"POST\", endpoint, &RequestOptions{\n\t\tBody: bytes.NewReader(body),\n\t\tHeaders: map[string]string{\n\t\t\t\"Content-Type\": \"application/json\",\n\t\t},\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresponse, err := checkResp(c.HTTPClient.Do(request))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar app App\n\tif err := decodeJSON(response, &app); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &app, nil\n}", "func createHandler (w http.ResponseWriter, r *http.Request) {\n\terr := create.ExecuteTemplate(w, \"create.html\", nil)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t}\n}", "func (drc *DeviceRequestCreate) Exec(ctx context.Context) error {\n\t_, err := drc.Save(ctx)\n\treturn err\n}", "func (a *DeviceAPI) CreateKeys(ctx context.Context, req *api.CreateDeviceKeysRequest) (*empty.Empty, error) {\n\tvar response empty.Empty\n\tif req.DeviceKeys == nil {\n\t\treturn nil, status.Errorf(codes.InvalidArgument, \"device_keys must not be nil\")\n\t}\n\n\t// appKey is not used for LoRaWAN 1.0\n\tvar appKey lorawan.AES128Key\n\tif req.DeviceKeys.AppKey != \"\" {\n\t\tif err := appKey.UnmarshalText([]byte(req.DeviceKeys.AppKey)); err != nil {\n\t\t\treturn nil, status.Errorf(codes.InvalidArgument, err.Error())\n\t\t}\n\t}\n\n\t// genAppKey is only for LoRaWAN 1.0 devices that implement the\n\t// remote multicast setup specification.\n\tvar genAppKey lorawan.AES128Key\n\tif req.DeviceKeys.GenAppKey != \"\" {\n\t\tif err := genAppKey.UnmarshalText([]byte(req.DeviceKeys.GenAppKey)); err != nil {\n\t\t\treturn nil, status.Errorf(codes.InvalidArgument, err.Error())\n\t\t}\n\t}\n\n\t// nwkKey\n\tvar nwkKey lorawan.AES128Key\n\tif err := nwkKey.UnmarshalText([]byte(req.DeviceKeys.NwkKey)); err != nil {\n\t\treturn nil, status.Errorf(codes.InvalidArgument, err.Error())\n\t}\n\n\t// devEUI\n\tvar eui lorawan.EUI64\n\tif err := eui.UnmarshalText([]byte(req.DeviceKeys.DevEui)); err != nil {\n\t\treturn nil, status.Errorf(codes.InvalidArgument, err.Error())\n\t}\n\n\tif valid, err := devmod.NewValidator(a.st).ValidateNodeAccess(ctx, authcus.Update, eui); !valid || err != nil {\n\t\treturn nil, status.Errorf(codes.Unauthenticated, \"authentication failed: %s\", err)\n\t}\n\n\terr := a.st.CreateDeviceKeys(ctx, &DeviceKeys{\n\t\tDevEUI: eui,\n\t\tNwkKey: nwkKey,\n\t\tAppKey: appKey,\n\t\tGenAppKey: genAppKey,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &response, nil\n}", "func (c *DeviceController) Index(w http.ResponseWriter, r *http.Request) {\n\tc.SendJSON(\n\t\tw,\n\t\tr,\n\t\t[]*Device{NewDevice(\"host\", \"192.168.0.1\", \"80.80.80.80\")},\n\t\thttp.StatusOK,\n\t)\n}", "func (s *server) Create(ctx context.Context, body *pb.RequestBody) (*pb.ResponseBody, error) {\n\tlanguage := body.GetLanguage()\n\tapp := &types.ApplicationConfig{}\n\n\terr := json.Unmarshal(body.GetData(), app)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tuser, err := mongo.FetchSingleUser(body.GetOwner())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tmaxCount := configs.ServiceConfig.AppMaker.AppLimit\n\trateCount := configs.ServiceConfig.RateLimit\n\ttimeInterval := configs.ServiceConfig.RateInterval\n\tif !user.IsAdmin() && maxCount >= 0 {\n\t\trateLimitCount := mongo.CountInstanceInTimeFrame(body.GetOwner(), mongo.AppInstance, timeInterval)\n\t\ttotalCount := mongo.CountInstancesByUser(body.GetOwner(), mongo.AppInstance)\n\t\tif totalCount < maxCount {\n\t\t\tif rateLimitCount >= rateCount && rateCount >= 0 {\n\t\t\t\treturn nil, fmt.Errorf(\"cannot deploy more than %d app instances in %d hours\", rateCount, timeInterval)\n\t\t\t}\n\t\t} else {\n\t\t\treturn nil, fmt.Errorf(\"cannot deploy more than %d app instances\", maxCount)\n\t\t}\n\t}\n\n\tapp.SetLanguage(language)\n\tapp.SetOwner(body.GetOwner())\n\tapp.SetInstanceType(mongo.AppInstance)\n\tapp.SetHostIP(utils.HostIP)\n\tapp.SetNameServers(configs.GasperConfig.DNSServers)\n\tapp.SetDateTime()\n\n\tgendnsNameServers, _ := redis.FetchServiceInstances(types.GenDNS)\n\tfor _, nameServer := range gendnsNameServers {\n\t\tif strings.Contains(nameServer, \":\") {\n\t\t\tapp.AddNameServers(strings.Split(nameServer, \":\")[0])\n\t\t} else {\n\t\t\tutils.LogError(\"AppMaker-Controller-1\", fmt.Errorf(\"GenDNS instance %s is of invalid format\", nameServer))\n\t\t}\n\t}\n\n\tif pipeline[language] == nil {\n\t\treturn nil, fmt.Errorf(\"language `%s` is not supported\", language)\n\t}\n\tresErr := pipeline[language].create(app)\n\tif resErr != nil {\n\t\tif resErr.Message() != \"repository already exists\" && resErr.Message() != \"container already exists\" {\n\t\t\tgo diskCleanup(app.GetName())\n\t\t}\n\t\treturn nil, fmt.Errorf(resErr.Error())\n\t}\n\n\tsshEntrypointIP := configs.ServiceConfig.GenSSH.EntrypointIP\n\tif len(sshEntrypointIP) == 0 {\n\t\tsshEntrypointIP = utils.HostIP\n\t}\n\tapp.SetSSHCmd(configs.ServiceConfig.GenSSH.Port, app.GetName(), sshEntrypointIP)\n\n\tapp.SetAppURL(fmt.Sprintf(\"%s.%s.%s\", app.GetName(), cloudflare.ApplicationInstance, configs.GasperConfig.Domain))\n\n\tif configs.CloudflareConfig.PlugIn {\n\t\tresp, err := cloudflare.CreateApplicationRecord(app.GetName())\n\t\tif err != nil {\n\t\t\tgo diskCleanup(app.GetName())\n\t\t\treturn nil, err\n\t\t}\n\t\tapp.SetCloudflareID(resp.Result.ID)\n\t\tapp.SetPublicIP(configs.CloudflareConfig.PublicIP)\n\t}\n\n\terr = mongo.UpsertInstance(\n\t\ttypes.M{\n\t\t\tmongo.NameKey: app.GetName(),\n\t\t\tmongo.InstanceTypeKey: mongo.AppInstance,\n\t\t}, app)\n\n\tif err != nil && err != mongo.ErrNoDocuments {\n\t\tgo diskCleanup(app.GetName())\n\t\tgo stateCleanup(app.GetName())\n\t\treturn nil, err\n\t}\n\n\terr = redis.RegisterApp(\n\t\tapp.GetName(),\n\t\tfmt.Sprintf(\"%s:%d\", utils.HostIP, configs.ServiceConfig.AppMaker.Port),\n\t\tfmt.Sprintf(\"%s:%d\", utils.HostIP, app.GetContainerPort()),\n\t)\n\n\tif err != nil {\n\t\tgo diskCleanup(app.GetName())\n\t\tgo stateCleanup(app.GetName())\n\t\treturn nil, err\n\t}\n\n\terr = redis.IncrementServiceLoad(\n\t\tServiceName,\n\t\tfmt.Sprintf(\"%s:%d\", utils.HostIP, configs.ServiceConfig.AppMaker.Port),\n\t)\n\n\tif err != nil {\n\t\tgo diskCleanup(app.GetName())\n\t\tgo stateCleanup(app.GetName())\n\t\treturn nil, err\n\t}\n\n\tapp.SetSuccess(true)\n\n\tresponse, err := json.Marshal(app)\n\treturn &pb.ResponseBody{Data: response}, err\n}", "func CreateGenericDevice(info IDeviceInfo) *Device {\n connectTimeout, _ := time.ParseDuration(agent.DEFAULT_CONNECT_TIMEOUT)\n\n instance := new(Device)\n agent.SetupAgent(&instance.Agent, info, 0, connectTimeout, agent.DEFAULT_BUFFER_SIZE, true)\n\n instance.SetOnMessageReceivedHandler(instance.onMessageReceived)\n instance.info = info\n instance.commandProcessor = func (deviceInfo map[string]string, command api.ICommand) (string, error) {\n return command.PropertyName() + command.PropertyValue(), nil\n }\n instance.queryProcessor = func (deviceInfo map[string]string, query api.IQuery) (string, error) {\n return query.PropertyName(), nil\n }\n\n return instance\n}", "func AllocateHandler(w http.ResponseWriter, r *http.Request) {\n\n\tpayloadStr := r.FormValue(\"payload\")\n\tif payloadStr != \"\" {\n\n\t\tvar payload models.Payload\n\t\tjson.Unmarshal([]byte(payloadStr), &payload)\n\n\t\tfmt.Println(payload)\n\t\tfmt.Println(payloadStr)\n\n\t\tif len(payload.Images) > 0 {\n\n\t\t\tstateIDPrefix := \"periscope-mock\"\n\t\t\tstateIDStr := payload.StateID\n\t\t\tappName := fmt.Sprintf(\"%s-%s\", stateIDPrefix, stateIDStr)\n\n\t\t\tcontainers := []apiv1.Container{}\n\n\t\t\tfor _, elem := range payload.Images {\n\t\t\t\tcontainer := apiv1.Container{\n\t\t\t\t\tName: fmt.Sprintf(\"%s%s\", appName, tools.RandStr(10)),\n\t\t\t\t\tImage: elem.Image,\n\t\t\t\t\tPorts: []apiv1.ContainerPort{\n\t\t\t\t\t\t{\n\t\t\t\t\t\t\tName: \"http\",\n\t\t\t\t\t\t\tProtocol: apiv1.ProtocolTCP,\n\t\t\t\t\t\t\tContainerPort: int32(elem.Port),\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t}\n\t\t\t\tcontainers = append(containers, container)\n\t\t\t}\n\n\t\t\tservices.CreateDeployment(appName, \"periscope\", containers, \"gcr-json-key\")\n\t\t\tservices.CreateService(appName, \"periscope\", int32(payload.SvcPort))\n\t\t\tservices.CreateVirtualService(appName, \"periscope\", stateIDStr, payload.SvcPort, \"v1\")\n\t\t\tfmt.Fprintf(w, \"Done.\")\n\n\t\t} else {\n\t\t\tfmt.Println(\"Invalid payload\")\n\t\t\tfmt.Fprintf(w, \"Invalid Payload\")\n\t\t}\n\t} else {\n\t\tfmt.Fprintf(w, \"No Payload\")\n\t}\n}", "func (handler *applicationHandler) create(app *types.ApplicationConfig) types.ResponseError {\n\tapp.SetDockerImage(handler.image)\n\tapp.SetConfGenerator(handler.confGenerator)\n\treturn api.SetupApplication(app)\n}", "func (client IdentityClient) createMfaTotpDevice(ctx context.Context, request common.OCIRequest, binaryReqBody *common.OCIReadSeekCloser, extraHeaders map[string]string) (common.OCIResponse, error) {\n\n\thttpRequest, err := request.HTTPRequest(http.MethodPost, \"/users/{userId}/mfaTotpDevices\", binaryReqBody, extraHeaders)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar response CreateMfaTotpDeviceResponse\n\tvar httpResponse *http.Response\n\thttpResponse, err = client.Call(ctx, &httpRequest)\n\tdefer common.CloseBodyIfValid(httpResponse)\n\tresponse.RawResponse = httpResponse\n\tif err != nil {\n\t\treturn response, err\n\t}\n\n\terr = common.UnmarshalResponse(httpResponse, &response)\n\treturn response, err\n}", "func InitDevices(parentRoute *echo.Group) {\n\troute := parentRoute.Group(\"/devices\")\n\t//\troute.Use(middleware.JWT([]byte(config.AuthTokenKey)))\n\n\troute.POST(\"\", permission.AuthRequired(createDevice))\n\troute.GET(\"/:id\", permission.AuthRequired(readDevice))\n\troute.PUT(\"/:id\", permission.AuthRequired(updateDevice))\n\troute.DELETE(\"/:id\", permission.AuthRequired(deleteDevice))\n\n\troute.GET(\"\", permission.AuthRequired(readDevices))\n\n\tdeviceService.InitService()\n}", "func create(w http.ResponseWriter, req *http.Request) {\n\tresponse := \"\"\n\tswitch req.RequestURI {\n\tcase \"/get/accounts\":\n\t\tmapD := map[string]int{\"apple\": 5, \"lettuce\": 7}\n\t\tmapB, _ := json.Marshal(mapD)\n\t\tresponse = string(mapB)\n\t\tbreak\n\tdefault:\n\t\tr, _ := json.Marshal(\"Request not found\")\n\t\tresponse = string(r)\n\t\tbreak\n\t}\n\n\tcontext := Context{Title: response}\n\trender(w, \"api\", context)\n}", "func (h *Handler) serveCreateDatabase(w http.ResponseWriter, r *http.Request) {\n\tvar req struct {\n\t\tName string `json:\"name\"`\n\t}\n\n\t// TODO: Authentication\n\n\t// Decode the request from the body.\n\tif err := json.NewDecoder(r.Body).Decode(&req); err != nil {\n\t\th.error(w, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\t// Create the database.\n\tif err := h.server.CreateDatabase(req.Name); err == ErrDatabaseExists {\n\t\th.error(w, err.Error(), http.StatusConflict)\n\t\treturn\n\t} else if err != nil {\n\t\th.error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\tw.WriteHeader(http.StatusCreated)\n}", "func AddDevice(m *Device) (id int64, err error) {\n\to := orm.NewOrm()\n\tid, err = o.Insert(m)\n\treturn\n}", "func (c Client) CreateApplication(name string, email string, accounts string, cloudProviders string, instancePort string, description string) (types.TaskRef, error) {\n var taskRef types.TaskRef\n var jobs []types.Job\n\n a := strings.Split(accounts, \",\")\n\n for _, account := range a {\n jobs = append(jobs, types.Job {\n Type: \"createApplication\",\n Account: account,\n User: \"\",\n Application: types.CreateApplication {\n Name: name,\n Description: description,\n Accounts: accounts,\n CloudProviders: cloudProviders,\n Email: email,\n InstancePort: instancePort,\n },\n })\n }\n\n task := types.Task {\n Application: name,\n Description: \"Create Application: \" + name,\n Job : jobs,\n }\n\n resp, err := c.post(\"/applications/\" + name + \"/tasks\", task)\n defer ensureReaderClosed(resp)\n if err != nil {\n return taskRef, err\n }\n\n err = json.NewDecoder(resp.body).Decode(&taskRef)\n return taskRef, err\n}", "func (c *AppsServices) Create(opt *AppsOpt) (interface{}, error) {\n\tu, err := url.Parse(c.client.BaseURL.String() + \"apps\")\n\tif err != nil {\n\t\treturn Success{}, err\n\t}\n\n\tb, err := EncodeBody(opt)\n\tif err != nil {\n\t\treturn Success{}, err\n\t}\n\n\tc.client.UseAuthKey = true\n\tresp, err := POST(u.String(), b, c.client)\n\tif err != nil {\n\t\treturn Success{}, err\n\t}\n\n\treturn resp, nil\n}", "func Create (appName string) {\n\n checkGopath ()\n checkContainer (appName)\n\n app := Application { Name: appName }\n\n app.createContainer ()\n\n err := app.copyFileTree (\n GOPATH + slash + applicationTemplatesPath,\n GOPATH_SRC + app.Name,\n )\n\n if err != nil {\n log.Fatal (err)\n }\n}", "func (srv *Server) Create(w http.ResponseWriter, r *http.Request) {\n\tif r.Method != \"PUT\" {\n\t\thttp.Error(w, \"Please use a PUT request to create an application.\", http.StatusBadRequest)\n\t\treturn\n\t}\n\t// Read in body\n\tbody, err := ioutil.ReadAll(r.Body)\n\tif err != nil {\n\t\thttp.Error(w, \"Failed to read body of request\", http.StatusInternalServerError)\n\t}\n\t// Try to parse the metadata content\n\tmetadata := &types.ApplicationMetadata{}\n\terr = yaml.Unmarshal(body, metadata)\n\tif err != nil {\n\t\thttp.Error(w, \"Failed to parse YAML input. This likely indicates malformed request body. Verify the payload fields and parameter types are correct.\", http.StatusBadRequest)\n\t\tlog.Info(\"YAML parse error\")\n\t\treturn\n\t}\n\n\t// Validate input\n\terr = srv.Validate.Struct(metadata)\n\tif err != nil {\n\t\t// If we fail to validate, automatically return 400\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\tw.Write([]byte(\"Failed to validate input of the following parameters:\\n\"))\n\n\t\t// Be helpful and tell users what fails in their request\n\t\tfor _, err := range err.(validator.ValidationErrors) {\n\t\t\tfmt.Fprintf(w, \"%s has invalid value %s\\n\", err.Namespace(), err.Value())\n\t\t}\n\t\tlog.Info(\"Rejected invalid input.\")\n\t\treturn\n\t}\n\n\t// Check if a conflicting application already exists\n\tif util.CheckTitle(srv.Applications, metadata.Title) {\n\t\tw.WriteHeader(http.StatusConflict)\n\t\tfmt.Fprintf(w, \"An application with title %s already exists, please use a unique title.\", metadata.Title)\n\t\treturn\n\t}\n\n\tw.WriteHeader(http.StatusCreated)\n\tsrv.Applications = append(srv.Applications, metadata)\n\tlog.WithFields(log.Fields{\"name\": metadata.Title}).Info(\"Object added\")\n\treturn\n}", "func CreateFromDevice(d *v1alpha1.Device) *packngo.DeviceCreateRequest {\n\treturn &packngo.DeviceCreateRequest{\n\t\tHostname: d.Spec.ForProvider.Hostname,\n\t\tPlan: d.Spec.ForProvider.Plan,\n\t\tFacility: []string{d.Spec.ForProvider.Facility},\n\t\tOS: d.Spec.ForProvider.OS,\n\t\tBillingCycle: d.Spec.ForProvider.BillingCycle,\n\t\tProjectID: d.Spec.ForProvider.ProjectID,\n\t\tUserData: d.Spec.ForProvider.UserData,\n\t\tTags: d.Spec.ForProvider.Tags,\n\t}\n}", "func main() {\n\t// choose dcgm hostengine running mode\n\t// 1. dcgm.Embedded\n\t// 2. dcgm.Standalone -connect \"addr\", -socket \"isSocket\"\n\t// 3. dcgm.StartHostengine\n\tflag.Parse()\n\tif err := dcgm.Init(dcgm.Standalone, *connectAddr, *isSocket); err != nil {\n\t\tlog.Panicln(err)\n\t}\n\n\tdefer func() {\n\t\tif err := dcgm.Shutdown(); err != nil {\n\t\t\tlog.Panicln(err)\n\t\t}\n\t}()\n\n\tcount, err := dcgm.GetAllDeviceCount()\n\tif err != nil {\n\t\tlog.Panicln(err)\n\t}\n\n\tt := template.Must(template.New(\"Device\").Parse(deviceInfo))\n\n\tfor i := uint(0); i < count; i++ {\n\t\tdeviceInfo, err := dcgm.GetDeviceInfo(i)\n\t\tif err != nil {\n\t\t\tlog.Panicln(err)\n\t\t}\n\n\t\tif err = t.Execute(os.Stdout, deviceInfo); err != nil {\n\t\t\tlog.Panicln(\"Template error:\", err)\n\t\t}\n\t}\n}", "func (u *userController) Create() {\n\treturn\n}", "func Create() http.Handler {\r\n\r\n\tmux := http.NewServeMux()\r\n\tdb, err := config.NewDB(\"postgres://postgres:postgres@localhost/powerdata\")\r\n\tif err != nil {\r\n\t\tlog.Panic(err)\r\n\t}\r\n\toauthConfig := config.GetStripeOauthConfig()\r\n\r\n\tenv := &config.Env{DB: db, OauthConfig: oauthConfig}\r\n\r\n\tmux.Handle(\"/\", oauth.MainPageHandler(`<html><body>\r\n\t\t<a href=\"/oauth/stripeLogin\">ConnectWithStripe</a>\r\n\t\t</body></html>`))\r\n\tmux.Handle(\"/oauth/stripeLogin\", oauth.LoginHandler(env))\r\n\tmux.Handle(\"/oauth/stripe_callback\", oauth.CallbackHandler(env))\r\n\tmux.Handle(\"/oauth/deauthorize\", oauth.DeauthorizeHandler(env))\r\n\tmux.Handle(\"/account/\", account.Handler(env))\r\n\r\n\treturn mux\r\n}", "func CreateWare(c *server.Context) error {\n\tvar (\n\t\terr error\n\t\taddReq ware.Ware\n\t\tconn orm.Connection\n\t)\n\n\tisAdmin := c.Request().Context().Value(\"user\").(jwtgo.MapClaims)[util.IsAdmin].(bool)\n\tif !isAdmin {\n\t\tlogger.Error(\"You don't have access\")\n\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrToken, nil)\n\t}\n\n\terr = c.JSONBody(&addReq)\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrInvalidParam, nil)\n\t}\n\n\terr = c.Validate(addReq)\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrInvalidParam, nil)\n\t}\n\n\tif len(addReq.Avatar) > 0 {\n\t\taddReq.Avatar, err = util.SavePicture(addReq.Avatar, \"ware/\")\n\t\tif err != nil {\n\t\t\tlogger.Error(err)\n\t\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrInternalServerError, nil)\n\t\t}\n\t}\n\tif len(addReq.Image) > 0 {\n\t\taddReq.Image, err = util.SavePicture(addReq.Image, \"ware/\")\n\t\tif err != nil {\n\t\t\tlogger.Error(err)\n\t\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrInternalServerError, nil)\n\t\t}\n\t}\n\tif len(addReq.DetailPic) > 0 {\n\t\taddReq.DetailPic, err = util.SavePicture(addReq.DetailPic, \"wareIntro/\")\n\t\tif err != nil {\n\t\t\tlogger.Error(err)\n\t\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrInternalServerError, nil)\n\t\t}\n\t}\n\n\tconn, err = mysql.Pool.Get()\n\tdefer mysql.Pool.Release(conn)\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrMysql, nil)\n\t}\n\n\terr = ware.Service.CreateWare(conn, &addReq)\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\tif (len(addReq.Avatar) > 0 && !util.DeletePicture(addReq.Avatar)) ||\n\t\t\t(len(addReq.Image) > 0 && !util.DeletePicture(addReq.Image)) ||\n\t\t\t(len(addReq.DetailPic) > 0 && !util.DeletePicture(addReq.DetailPic)) {\n\t\t\tlogger.Error(errors.New(\"create ware failed and delete it's pictures go wrong, please delete picture manually\"))\n\t\t}\n\t\treturn core.WriteStatusAndDataJSON(c, constants.ErrMysql, nil)\n\t}\n\n\tlogger.Info(\"create ware\", addReq.Name, \"success\")\n\treturn core.WriteStatusAndDataJSON(c, constants.ErrSucceed, nil)\n}", "func (a *App) createSubscriber(w http.ResponseWriter, r *http.Request) {\n\tvar resp = map[string]interface{}{\"message\": \"sip device successfully created\", \"status\": \"success\"}\n\n\tsubscriber := &models.Subscribers{}\n\tbody, err := ioutil.ReadAll(r.Body)\n\tif err != nil {\n\t\trespond.ERROR(w, http.StatusBadRequest, err)\n\t\treturn\n\t}\n\n\tif err = json.Unmarshal(body, &subscriber); err != nil {\n\t\trespond.ERROR(w, http.StatusBadRequest, err)\n\t\treturn\n\t}\n\n\tsubscriber.Prepare()\n\n\tif sbname := models.GetSubscriberByUserName(a.Ctx, a.DB, subscriber.Username); sbname != nil {\n\t\trespond.ERROR(w, http.StatusBadRequest, sbname)\n\t\treturn\n\t}\n\n\tif err = subscriber.Validate(); err != nil {\n\t\trespond.ERROR(w, http.StatusBadRequest, err)\n\t\treturn\n\t}\n\n\tif err = subscriber.Save(a.Ctx, a.DB); err != nil {\n\t\trespond.ERROR(w, http.StatusBadRequest, err)\n\t\treturn\n\t}\n\n\trespond.JSON(w, http.StatusCreated, resp)\n\treturn\n}", "func (h *HandlersApp01sqVendor) TableCreate(w http.ResponseWriter, r *http.Request) {\n\tvar err error\n\n\tlog.Printf(\"hndlrVendor.TableCreate(%s)\\n\", r.Method)\n\n\tif r.Method != \"GET\" {\n\t\thttp.Error(w, http.StatusText(405), http.StatusMethodNotAllowed)\n\t\treturn\n\t}\n\n\t// Create the table.\n\terr = h.db.TableCreate()\n\tif err == nil {\n\t\t//h.ListShow(w, 0, \"Table was created\")\n\t\tw.Write([]byte(\"Table was created\"))\n\t} else {\n\t\tw.Write([]byte(\"Table creation had an error of:\" + err.Error()))\n\t}\n\n\tlog.Printf(\"...end hndlrVendor.TableCreate(%s)\\n\", util.ErrorString(err))\n\n}", "func CreateDeleteApDeviceRequest() (request *DeleteApDeviceRequest) {\n\trequest = &DeleteApDeviceRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"cloudesl\", \"2020-02-01\", \"DeleteApDevice\", \"\", \"\")\n\trequest.Method = requests.POST\n\treturn\n}", "func createDeviceTwin(devTwin dtclient.DeviceTwin) []dtclient.DeviceTwin {\n\tdeviceTwin := []dtclient.DeviceTwin{}\n\tdeviceTwin = append(deviceTwin, devTwin)\n\treturn deviceTwin\n}", "func InitNewWebsite(c *ishell.Context) {\n\n\t// Demande pour le nom du site web\n\tc.Print(\"Domain name of the website (ex : wafg.ca) : \")\n\n\tc.Print(\"Enter your full name : \")\n\n\t// Demande qu'elle page de base je veux d'enable\n\n\tc.Print(\"Admin account create with success\")\n\n}", "func CreateApplication() *Alpha {\n app := &Alpha{}\n app.Request = &Request{}\n app.Response = &Response{}\n app.init()\n return app\n}", "func CreateProductEndPoint(w http.ResponseWriter, r *http.Request) {\n\tfmt.Fprintln(w, \"not implemented yet !\")\n}", "func createAppUser(model *AppUserModel) api.Response {\n\tuser, err := auth.CreateAppUser(model.Email, model.Password, model.Roles)\n\tif err != nil {\n\t\treturn api.InternalServerError(err)\n\t}\n\n\treturn api.JSONResponse(http.StatusOK, user)\n}", "func newUssdApp(hostName string, services ServiceManager, sessionStore phada.SessionStore) *UssdApp {\n\treturn &UssdApp{\n\t\thost: hostName,\n\t\tsessionStore: sessionStore,\n\t\tservicectl: services,\n\t}\n}", "func webTaskMakeup(c *bm.Context) {\n\tparams := c.Request.Form\n\t// check user\n\tmidStr, ok := c.Get(\"mid\")\n\tif !ok {\n\t\tc.JSON(nil, ecode.CreativeNotLogin)\n\t\treturn\n\t}\n\tmid := midStr.(int64)\n\tif !dataSvc.IsWhite(mid) {\n\t\tc.JSON(nil, ecode.CreativeNotLogin)\n\t\treturn\n\t}\n\n\ttmid := params.Get(\"tmid\")\n\ttid, err := toInt64(tmid)\n\tif err != nil {\n\t\tc.JSON(nil, ecode.RequestErr)\n\t\treturn\n\t}\n\t// check white list\n\tif task := whiteSvc.TaskWhiteList(tid); task != 1 {\n\t\tc.JSON(nil, ecode.RequestErr)\n\t\treturn\n\t}\n\n\tif err := newcomerSvc.TaskMakeup(c, tid); err != nil {\n\t\tc.JSON(nil, err)\n\t\treturn\n\t}\n\tc.JSON(\"ok\", nil)\n}", "func CreateApp(a *App, units uint) error {\n\tif units == 0 {\n\t\treturn &ValidationError{Message: \"Cannot create app with 0 units.\"}\n\t}\n\tif !a.isValid() {\n\t\tmsg := \"Invalid app name, your app should have at most 63 \" +\n\t\t\t\"characters, containing only lower case letters or numbers, \" +\n\t\t\t\"starting with a letter.\"\n\t\treturn &ValidationError{Message: msg}\n\t}\n\tactions := []action{\n\t\tnew(insertApp),\n\t\tnew(createBucketIam),\n\t\tnew(createRepository),\n\t\tnew(provisionApp),\n\t}\n\treturn execute(a, actions, units)\n}", "func (controller *ActivityController) Create(res http.ResponseWriter, req *http.Request) {\n\tif !controller.auth.IsLogin(res, req) {\n\t\thttp.Redirect(res, req, \"/login\", http.StatusSeeOther)\n\t\treturn\n\t}\n\tcontroller.render(res, req, \"activity_create.gohtml\", nil, 2)\n}", "func (c *Client) CreateDevice(collectionID string, device Device) (Device, error) {\n\terr := c.create(fmt.Sprintf(\"/collections/%s/devices\", collectionID), &device)\n\treturn device, err\n}", "func (c *DeviceClient) Create() *DeviceCreate {\n\tmutation := newDeviceMutation(c.config, OpCreate)\n\treturn &DeviceCreate{config: c.config, hooks: c.Hooks(), mutation: mutation}\n}", "func CreateDemoDeviceInstance() (*oc.Device, error) {\n\t// Initialize a device.\n\td := &oc.Device{\n\t\tSystem: &oc.System{\n\t\t\tHostname: ygot.String(\"rtr02.pop44\"),\n\t\t\tΛHostname: []ygot.Annotation{\n\t\t\t\t&ExampleAnnotation{ConfigSource: \"devicedemo\"},\n\t\t\t},\n\t\t},\n\t}\n\n\t// Create a new interface under the device. In this case /interfaces/interface\n\t// is the list that is being populated, but due to schema compression the\n\t// 'interfaces' container is not created, making the 'interface' list a top-level\n\t// entity. The New... helper methods are therefore mapped to device.\n\teth0, err := d.NewInterface(\"eth0\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Set some attributes of the interface.\n\teth0.Description = ygot.String(\"Link to rtr01.pop44\")\n\teth0.Type = oc.IETFInterfaces_InterfaceType_ethernetCsmacd\n\n\tif err := addNetworkInstance(d); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Add a component.\n\tc, err := d.NewComponent(\"os\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tc.Type = oc.PlatformTypes_OPENCONFIG_SOFTWARE_COMPONENT_OPERATING_SYSTEM\n\n\t// Create a second device instance, and populate the OS component under\n\t// it. This code demonstrates how ygot.MergeStructs can be used to combine\n\t// multiple instances of the same type of struct together, allowing each\n\t// subtree to be generated in its own context.\n\tsecondDev := &oc.Device{}\n\tsc, err := secondDev.NewComponent(\"os\")\n\tsc.Description = ygot.String(\"RouterOS 14.0\")\n\tmergedDev, err := ygot.MergeStructs(d, secondDev)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// Since ygot.MergeStructs returns an ygot.GoStruct interface, we\n\t// must type assert it back to *oc.Device.\n\treturn mergedDev.(*oc.Device), nil\n}", "func (controller AppsController) Create(c *gin.Context) {\n\tvar configE entities.App\n\terr := c.BindJSON(&configE)\n\n\tif err != nil {\n\t\tfmt.Printf(\"error %v\", err)\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"message\": \"Invalid Parameters\"})\n\t\treturn\n\t}\n\n\tconfig, err := mongodb.Create(controller.MongoDBClient, Collections[\"apps\"], &configE)\n\tif err != nil {\n\t\tfmt.Printf(\"error %v\", err)\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"message\": \"Unable to save data\"})\n\t\treturn\n\t}\n\n\tc.JSON(http.StatusOK, gin.H{\"apps\": config})\n}", "func (c *Client) AddDevice(ctx context.Context, path string, payload *AddDeviceSourcePayload) (*http.Response, error) {\n\treq, err := c.NewAddDeviceRequest(ctx, path, payload)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn c.Client.Do(ctx, req)\n}", "func (api API) CreateApp(name, engine string) (app AppBundle, err error) {\n\n\tbearer, err := api.Authenticator.GetToken(\"code:all\")\n\tif err != nil {\n\t\treturn\n\t}\n\tpath := api.Authenticator.GetHostPath() + api.DesignAutomationPath\n\tapp, err = createApp(path, name, engine, bearer.AccessToken)\n\n\tapp.authenticator = api.Authenticator\n\tapp.path = path\n\tapp.name = name\n\tapp.uploadURL = api.UploadAppURL\n\n\t//WARNING: when an AppBundle is created, it is assigned an '$LATEST' alias\n\t// but this alias is not usable and if no other alias is created for this\n\t// appBundle, then the alias listing will fail.\n\t// Thus I decided to autoasign a \"default\" alias upon app creation\n\tgo app.CreateAlias(\"default\", 1)\n\n\treturn\n}", "func (c *Controller) renderNew(ctx context.Context, w http.ResponseWriter, authApp *database.AuthorizedApp) {\n\tm := controller.TemplateMapFromContext(ctx)\n\tm.Title(\"New API key\")\n\tm[\"authApp\"] = authApp\n\tm[\"typeAdmin\"] = database.APIKeyTypeAdmin\n\tm[\"typeDevice\"] = database.APIKeyTypeDevice\n\tm[\"typeStats\"] = database.APIKeyTypeStats\n\tc.h.RenderHTML(w, \"apikeys/new\", m)\n}", "func newDevice(cdevice C.nvmlDevice_t) (*Device, error) {\n\tdevice := Device{\n\t\tnvmldevice: cdevice,\n\t}\n\n\tuuid, err := device.UUID()\n\tif err != nil {\n\t\treturn nil, errors.New(\"Cannot retrieve UUID property\")\n\t}\n\tdevice.uuid = uuid\n\n\tname, err := device.Name()\n\tif err != nil {\n\t\treturn nil, errors.New(\"Cannot retrieve Name property\")\n\t}\n\tdevice.name = name\n\n\tindex, err := device.Index()\n\tif err != nil {\n\t\treturn nil, errors.New(\"Cannot retrieve Index property\")\n\t}\n\tdevice.index = index\n\n\tdevice.bus = device.BusID()\n\n\treturn &device, nil\n}", "func (r *SoftwareResource) Create(item SoftwareConfig) error {\n\tif err := r.c.ModQuery(\"POST\", BasePath+SoftwareEndpoint, item); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (env *Env) Create(res http.ResponseWriter, req *http.Request) {\n\tswitch req.Method {\n\tcase \"GET\":\n\t\tenv.Log.V(1, \"beginning handling of GET request for Create endnode.\")\n\t\tp := &webAppGo.Page{}\n\t\tenv.Log.V(1, \"rendering the create template.\")\n\t\tenv.Render(res, \"create\", p)\n\tcase \"POST\":\n\t\tenv.Log.V(1, \"beginning handling of POST request for Create endnode.\")\n\t\ttitle := strings.Title(req.FormValue(\"title\"))\n\t\tif strings.Contains(title, \" \") {\n\t\t\ttitle = strings.Replace(title, \" \", \"_\", -1)\n\t\t}\n\t\tbody := req.FormValue(\"body\")\n\t\tp := &webAppGo.Page{Title: strings.Title(title), Body: []byte(body)}\n\t\terr := env.Cache.SaveToCache(p)\n\t\tif err != nil {\n\t\t\tenv.Log.V(1, \"notifying client that an internal error occured. Error assocaited with Cache.SaveToCache.\")\n\t\t\thttp.Error(res, err.Error(), http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\terr = env.DB.SavePage(p)\n\t\tif err != nil {\n\t\t\tenv.Log.V(1, \"notifying client that an internal error occured. Error assocaited with DB.SavePage\")\n\t\t\thttp.Error(res, err.Error(), http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\tenv.Log.V(1, \"page successfully saved, redirecting the client to /view/PageTitle.\")\n\t\thttp.Redirect(res, req, \"/view/\"+title, 302)\n\t\treturn\n\t}\n}", "func CreateDevice(c context.Context, record *models.Device) error {\n\treturn FromContext(c).CreateDevice(record, Current(c))\n}", "func ServeNew(w http.ResponseWriter, r *http.Request) {\n\tvar data newReq\n\n\tID, err := ulid.New(ulid.Now(), entropy)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn\n\t}\n\n\terr = json.NewDecoder(r.Body).Decode(&data)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn\n\t}\n\tnewH := NewHMST(data.Resolution, data.MaxTime, data.Keys)\n\tregistry[ID.String()] = newH\n\tlog.Println(\"/new\", ID.String(), data, len(newH.Registers))\n\tfmt.Fprintf(w, \"%v\", ID)\n}", "func createIDEController(l *object.VirtualDeviceList) ([]types.BaseVirtualDeviceConfigSpec, error) {\n\tide, _ := l.CreateIDEController()\n\tcspec, err := object.VirtualDeviceList{ide}.ConfigSpec(types.VirtualDeviceConfigSpecOperationAdd)\n\t*l = applyDeviceChange(*l, cspec)\n\treturn cspec, err\n}", "func CreateCar(w http.ResponseWriter, r *http.Request) {\n\t// Set the way we will serve data between frontend and backend\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t// Allow cross origin connections making the routes accessible for everyone\n\tw.Header().Set(\"Access-Control-Allow-Origin\", \"*\")\n\t// Allow the server to perform post operation\n\tw.Header().Set(\"Access-Control-Allow-Methods\", \"POST\")\n\t// Allow the content type that is specified by client to be processed on server\n\tw.Header().Set(\"Access-Control-Allow-Headers\", \"Content-Type\")\n\t// Declare an empty car\n\tvar car models.Car\n\t// Take the car json from the client and decode it into car struct\n\t_ = json.NewDecoder(r.Body).Decode(&car)\n\tpayload := createCar(car)\n\tjson.NewEncoder(w).Encode(payload)\n}", "func (a *App) create(c *echo.Context) error {\n\ttask := &model.Task{}\n\tvar err error\n\n\terr = c.Bind(task)\n\tif err != nil {\n\t\tc.JSON(http.StatusInternalServerError, ErrorMsg{Msg: err.Error()})\n\t\treturn err\n\t}\n\tstatus, err := task.Create(a.GetDB())\n\tc.JSON(status, ErrorMsg{Msg: err.Error()})\n\treturn err\n}", "func NewDevice(userid string, devicename string, ipaddr net.IP) (device, error) {\n\tconst invalidusername string = \"invalid username string\"\n\tvar retdevice device\n\tvar err error\n\tvar devicejson []byte\n\tif !EvalName(userid) {\n\t\terr = errors.New(invalidusername)\n\t\treturn retdevice, err\n\t}\n\tdevicejson, err = CheckForDevice(userid, devicename)\n\tErrhandle_Log(err, ERRMSG_DEVICECHECK)\n\n\tif err != nil || devicejson == nil {\n\t\tcl := make(map[string]contentinfo)\n\t\tdeviceid := uuid.New()\n\t\tretdevice = device{Userid: userid, Deviceid: deviceid, Devicename: devicename, Ipaddr: ipaddr, Consentlist: cl, Online: false, DeviceType: \"cgo\"}\n\t\tAddDevice(retdevice)\n\t} else {\n\t\terr = json.Unmarshal(devicejson, &retdevice)\n\t\tErrhandle_Log(err, ERRMSG_JSON_UNMARSHALL)\n\t\tif err != nil {\n\t\t\treturn retdevice, errors.New(ERRMSG_JSON_UNMARSHALL)\n\t\t}\n\t}\n\n\tin := make(chan byte, 10*1024)\n\tinreq := make(chan contentinfo, 10)\n\tout := make(chan byte, 10)\n\n\tretdevice.Indata = in\n\tretdevice.Outdata = out\n\tretdevice.Inrequests = inreq\n\n\tretdevice.DeviceType = \"cgo\"\n\n\treturn retdevice, nil\n}" ]
[ "0.6723208", "0.61843246", "0.6101709", "0.6098106", "0.6088275", "0.58678204", "0.5845221", "0.5822698", "0.5819041", "0.5807063", "0.57806623", "0.5720011", "0.57058775", "0.5693602", "0.56549305", "0.5613098", "0.5608295", "0.5604464", "0.55713624", "0.55530673", "0.5523906", "0.5520271", "0.5443575", "0.54141366", "0.53919446", "0.53741723", "0.53708524", "0.5370824", "0.53555113", "0.53357524", "0.5332189", "0.5323778", "0.53053284", "0.5285255", "0.52729803", "0.5254688", "0.5254688", "0.5241343", "0.52297163", "0.5226327", "0.5214764", "0.5209499", "0.5197343", "0.5192487", "0.51907986", "0.5161723", "0.51577", "0.5153529", "0.5132599", "0.5128327", "0.5113712", "0.51130474", "0.511266", "0.5097329", "0.5089697", "0.50658", "0.5060127", "0.5040682", "0.5034757", "0.50259167", "0.50258976", "0.4992754", "0.4984287", "0.49761567", "0.4974277", "0.4970931", "0.49665928", "0.49629405", "0.49599192", "0.49593118", "0.49471563", "0.49418873", "0.49413905", "0.4925778", "0.4919799", "0.4915705", "0.49031824", "0.4901909", "0.48908854", "0.48875123", "0.48874527", "0.48847175", "0.48824543", "0.48750302", "0.487325", "0.48729628", "0.48640195", "0.4861265", "0.48499346", "0.48488745", "0.4838125", "0.48359492", "0.48343495", "0.48333192", "0.48320234", "0.48313025", "0.4818797", "0.48102257", "0.4807094", "0.4806251", "0.48049223" ]
0.0
-1
NewConnector returns new driver.Connector.
func NewConnector(connContext *bytehouse.ConnectionContext, dsn string) (driver.Connector, error) { return &connector{ dsn: dsn, connContext: connContext, }, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewConnector(name string, onInit func(driver.Conn) error) (driver.Connector, error) {\r\n\tcxr, err := defaultDrv.OpenConnector(name)\r\n\tif err != nil {\r\n\t\treturn nil, err\r\n\t}\r\n\tcx := cxr.(connector)\r\n\tcx.onInit = onInit\r\n\treturn cx, err\r\n}", "func NewConnector(dsn string) driver.Connector {\n\treturn &Connector{\n\t\tInfo: make(map[string]string),\n\t\tdsn: dsn,\n\t}\n}", "func NewConnector(name string, config interface{}) (err error, conn dbconnector.DatabaseConnector) {\n\tswitch name {\n\tcase \"janusgraph\":\n\t\terr, conn = janusgraph.New(config)\n\tcase \"foobar\":\n\t\terr, conn = foobar.New(config)\n\tdefault:\n\t\terr = fmt.Errorf(\"No connector with the name '%s' exists!\", name)\n\t}\n\treturn\n}", "func NewConnector(c driver.Connector, opts ...Option) driver.Connector {\n\tvar cfg config\n\tfor _, opt := range opts {\n\t\topt(&cfg)\n\t}\n\td := &driverDriver{\n\t\tDriver: c.Driver(),\n\t\tconfig: cfg,\n\t}\n\n\treturn &driverConnector{\n\t\tConnector: c,\n\t\tdriver: d,\n\t}\n}", "func NewConnector() *Connector {\n\treturn &Connector{\n\t\tdie: make(chan struct{}),\n\t\tcodec: codec.NewDecoder(),\n\t\tchSend: make(chan []byte, 64),\n\t\tmid: 1,\n\t\tevents: map[string]Callback{},\n\t\tresponses: map[uint64]Callback{},\n\t}\n}", "func NewConnector(dsn string) (*Connector, error) {\n\treturn nil, nil\n}", "func (d *drv) NewConnector(params dsn.ConnectionParams) driver.Connector {\n\treturn connector{drv: d, ConnectionParams: params}\n}", "func NewConnector(config cfg.Connection) IConnector {\n\tswitch config.Protocol {\n\tcase protocolHttp:\n\t\treturn &HttpConnector{config: config}\n\tcase protocolAmqp:\n\t\treturn &AmqpConnector{config: config}\n\t}\n\n\tlog.Fatalln(\"unknown connection protocol\", config.Protocol)\n\treturn nil\n}", "func NewConnector(verbose bool, trust *x509.CertPool) *Connector {\n\tc := Connector{trust: trust, verbose: verbose}\n\treturn &c\n}", "func NewConnector(cfg datastore.Config, launchID, keyID string) (*connector.Connector, error) {\n\treturn connector.New(cfg, launchID, keyID)\n}", "func NewConnector(conRes connector.Resources) tcp.Connector {\n\tconnect := func(\n\t\tclientConn net.Conn,\n\t\tcredentialValuesByID connector.CredentialValuesByID,\n\t) (backendConn net.Conn, err error) {\n\t\t// singleUseConnector is responsible for generating the authenticated connection\n\t\t// to the target service for each incoming client connection\n\t\tsingleUseConnector := &SingleUseConnector{\n\t\t\tlogger: conRes.Logger(),\n\t\t}\n\n\t\treturn singleUseConnector.Connect(clientConn, credentialValuesByID)\n\t}\n\n\treturn tcp.ConnectorFunc(connect)\n}", "func NewConnector(params dsn.ConnectionParams) driver.Connector {\n\treturn defaultDrv.NewConnector(params)\n}", "func NewConnector(conRes connector.Resources) tcp.Connector {\n\tconnectorFunc := func(\n\t\tclientConn net.Conn,\n\t\tcredentialValuesByID connector.CredentialValuesByID,\n\t) (backendConn net.Conn, err error) {\n\t\t// singleUseConnector is responsible for generating the authenticated connection\n\t\t// to the target service for each incoming client connection\n\t\tsingleUseConnector := &SingleUseConnector{\n\t\t\tlogger: conRes.Logger(),\n\t\t}\n\n\t\treturn singleUseConnector.Connect(clientConn, credentialValuesByID)\n\t}\n\n\treturn tcp.ConnectorFunc(connectorFunc)\n}", "func NewConnector(config Config) (driver.Connector, error) {\n\t// Set some mandatory parameters\n\tconfig.Params[\"parseTime\"] = \"true\"\n\tconfig.Params[\"rejectReadOnly\"] = \"true\"\n\n\t// TODO: fill in the config instead of playing with DSN\n\tconf, err := mysql.ParseDSN(config.DSN())\n\tif err != nil {\n\t\treturn nil, errors.WithStack(err)\n\t}\n\n\tconnector, err := mysql.NewConnector(conf)\n\tif err != nil {\n\t\treturn nil, errors.WithStack(err)\n\t}\n\n\treturn ocsql.WrapConnector(\n\t\tconnector,\n\t\tocsql.WithOptions(ocsql.TraceOptions{\n\t\t\tAllowRoot: false,\n\t\t\tPing: true,\n\t\t\tRowsNext: true,\n\t\t\tRowsClose: true,\n\t\t\tRowsAffected: true,\n\t\t\tLastInsertID: true,\n\t\t\tQuery: true,\n\t\t\tQueryParams: false,\n\t\t}),\n\t), nil\n}", "func NewConnector(hosts ...string) driver.Connector {\n\treturn &OCI8Connector{\n\t\tLogger: log.New(ioutil.Discard, \"\", 0),\n\t}\n}", "func NewConnector(ctx *pulumi.Context,\n\tname string, args *ConnectorArgs, opts ...pulumi.ResourceOption) (*Connector, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.PrincipalInfo == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'PrincipalInfo'\")\n\t}\n\treplaceOnChanges := pulumi.ReplaceOnChanges([]string{\n\t\t\"location\",\n\t\t\"project\",\n\t})\n\topts = append(opts, replaceOnChanges)\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Connector\n\terr := ctx.RegisterResource(\"google-native:beyondcorp/v1alpha:Connector\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func NewConnector(cfg *Config) (driver.Connector, error) {\n\topts := append(cfg.ClientOptions, option.WithUserAgent(userAgent))\n\tclient, err := spanner.NewClientWithConfig(\n\t\tcontext.Background(),\n\t\tcfg.Database,\n\t\tcfg.ClientConfig,\n\t\topts...,\n\t)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &SpannerConnector{client: client}, nil\n}", "func NewConnector(conn *net.TCPConn, connID uint32, userid int, callback mintinterfaces.MintHandler) *Connector {\n\tc := &Connector{Conn: conn, ConnID: connID, UserID: userid, funcHanlder: callback,\n\t\tisClosed: false, ExitChan: make(chan bool, 1), msgChan: make(chan []byte)}\n\treturn c\n}", "func NewConnector(brokerURL string) (*Connector, error) {\n\topts, topicPrefix, err := ClientOptionsFromURL(brokerURL)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Connector{\n\t\tDiscoverTimeout: DefaultDiscoverTimeout,\n\t\toptions: opts,\n\t\ttopicPrefix: topicPrefix,\n\t}, nil\n}", "func Connect(cmd *cobra.Command) (radio.Connector, error) {\n\tif cmd == nil {\n\t\treturn nil, errors.New(\"no cobra command given\")\n\t}\n\tconnector := connector{\n\t\tcmd: cmd,\n\t}\n\n\treturn &connector, nil\n}", "func (c Client) CreateConnector(req CreateConnectorRequest, sync bool) (ConnectorResponse, error) {\n\tresp := ConnectorResponse{}\n\n\tstatusCode, err := c.Request(http.MethodPost, \"connectors\", req, &resp)\n\tif err != nil {\n\t\treturn ConnectorResponse{}, err\n\t}\n\tif resp.ErrorCode != 0 {\n\t\treturn ConnectorResponse{}, resp.ErrorResponse\n\t}\n\n\tresp.Code = statusCode\n\n\tif sync {\n\t\tif !TryUntil(\n\t\t\tfunc() bool {\n\t\t\t\tresp, err := c.GetConnector(req.ConnectorRequest)\n\t\t\t\treturn err == nil && resp.Code == 200\n\t\t\t},\n\t\t\t2*time.Minute,\n\t\t) {\n\t\t\treturn resp, errors.New(\"timeout on creating connector sync\")\n\t\t}\n\t}\n\n\treturn resp, nil\n}", "func NewSimpleConnector(target string, options map[string]string) Connector {\n\treturn &simpleConnector{backendAddr: target, opts: options}\n}", "func (d *drv) OpenConnector(name string) (driver.Connector, error) {\r\n\tP, err := ParseConnString(name)\r\n\tif err != nil {\r\n\t\treturn nil, err\r\n\t}\r\n\r\n\treturn connector{ConnectionParams: P, drv: d}, nil\r\n}", "func (d *nrDriver) OpenConnector(name string) (driver.Connector, error) {\n\tc, err := d.original.(driver.DriverContext).OpenConnector(name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn wrapConnector(c, d, d.segmenter), nil\n}", "func MakeConnector(conf Config, maxConns int, tm *transport.Manager, httpC *http.Client, clientPublicIP string,\n\tlog *logging.Logger, mLog *logging.MasterLogger) Autoconnector {\n\tconnector := &autoconnector{}\n\tconnector.client = NewClient(log, mLog, conf, httpC, clientPublicIP)\n\tconnector.maxConns = maxConns\n\tconnector.log = log\n\tconnector.tm = tm\n\treturn connector\n}", "func LocalConnector(network, address string) Connector {\n\treturn &localConn{network, address}\n}", "func NewPrintConnector()(*PrintConnector) {\n m := &PrintConnector{\n Entity: *NewEntity(),\n }\n return m\n}", "func New(name string, connector interface{}) *Catalog {\n\treturn &Catalog{\n\t\tName: name,\n\t\tOrigins: make(map[string]*Origin),\n\t\tConnector: connector,\n\t}\n}", "func (d *drv) OpenConnector(name string) (driver.Connector, error) {\n\n\t// parse connection string\n\tP, err := dsn.Parse(name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn NewConnector(P), nil\n}", "func NewPGConnector() *PGConnector { return &PGConnector{} }", "func NewPGConnector() *PGConnector { return &PGConnector{} }", "func New(config config.ConnectorConfig) (http.Handler, error) {\n\tif config.URI == \"\" {\n\t\treturn nil, errors.Errorf(\"URI parameter missing in connector: %#v\", config)\n\t}\n\tif !strings.HasPrefix(config.URI, prefixURI) {\n\t\treturn nil, errors.Errorf(\"Invalid URI parameter in connector, expected '%s': %#v\", prefixURI, config)\n\t}\n\n\tdirectoryPath := strings.Replace(config.URI, prefixURI, \"\", 1)\n\tif directoryPath == \"\" {\n\t\treturn nil, errors.Errorf(\"Directory path is missing in URI parameter in connector: %#v\", config)\n\t}\n\n\tstat, err := os.Stat(directoryPath)\n\tif os.IsNotExist(err) {\n\t\treturn nil, errors.Errorf(\"Directory passed in URI parameter doesn't exists, connector: %#v\", config)\n\t}\n\tif !stat.Mode().IsDir() {\n\t\treturn nil, errors.Errorf(\"URI parameter doesn't point to directory, connector: %#v\", config)\n\t}\n\n\thandler := &directoryHandler{\n\t\tpathPrefix: config.PathPrefix,\n\t\tdirectoryPath: directoryPath,\n\t}\n\n\treturn handler, nil\n}", "func (d *hdbDriver) OpenConnector(dsn string) (driver.Connector, error) { return NewDSNConnector(dsn) }", "func WrapConnector(dc driver.Connector, t *zipkin.Tracer, options ...TraceOption) driver.Connector {\n\topts := TraceOptions{}\n\tfor _, o := range options {\n\t\to(&opts)\n\t}\n\n\treturn &zDriver{\n\t\tparent: dc.Driver(),\n\t\tconnector: dc,\n\t\ttracer: t,\n\t\toptions: opts,\n\t}\n}", "func GetConnector(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *ConnectorState, opts ...pulumi.ResourceOption) (*Connector, error) {\n\tvar resource Connector\n\terr := ctx.ReadResource(\"google-native:beyondcorp/v1alpha:Connector\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (cfg *Config) newClient(args []interface{}) (connector endpoint.Connector, err error) {\n\n\tvar clientArgs *newClientArgs\n\tclientArgs, err = getNewClientArguments(args)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar connectionTrustBundle *x509.CertPool\n\n\tif cfg.ConnectionTrust != \"\" {\n\t\tlog.Println(\"You specified a trust bundle.\")\n\t\tconnectionTrustBundle = x509.NewCertPool()\n\t\tif !connectionTrustBundle.AppendCertsFromPEM([]byte(cfg.ConnectionTrust)) {\n\t\t\treturn nil, fmt.Errorf(\"%w: failed to parse PEM trust bundle\", verror.UserDataError)\n\t\t}\n\t}\n\n\tswitch cfg.ConnectorType {\n\tcase endpoint.ConnectorTypeCloud:\n\t\tconnector, err = cloud.NewConnector(cfg.BaseUrl, cfg.Zone, cfg.LogVerbose, connectionTrustBundle)\n\tcase endpoint.ConnectorTypeTPP:\n\t\tconnector, err = tpp.NewConnector(cfg.BaseUrl, cfg.Zone, cfg.LogVerbose, connectionTrustBundle)\n\tcase endpoint.ConnectorTypeFirefly:\n\t\tconnector, err = firefly.NewConnector(cfg.BaseUrl, cfg.Zone, cfg.LogVerbose, connectionTrustBundle)\n\tcase endpoint.ConnectorTypeFake:\n\t\tconnector = fake.NewConnector(cfg.LogVerbose, connectionTrustBundle)\n\tdefault:\n\t\terr = fmt.Errorf(\"%w: ConnectorType is not defined\", verror.UserDataError)\n\t}\n\tif err != nil {\n\t\treturn\n\t}\n\n\tconnector.SetZone(cfg.Zone)\n\tconnector.SetHTTPClient(cfg.Client)\n\n\tif clientArgs.authenticate {\n\t\terr = connector.Authenticate(cfg.Credentials)\n\t}\n\n\treturn\n}", "func (c connector) Connect(context.Context) (driver.Conn, error) {\r\n\tconn, err := c.drv.openConn(c.ConnectionParams)\r\n\tif err != nil || c.onInit == nil || !conn.newSession {\r\n\t\treturn conn, err\r\n\t}\r\n\tif err = c.onInit(conn); err != nil {\r\n\t\tconn.close(true)\r\n\t\treturn nil, err\r\n\t}\r\n\treturn conn, nil\r\n}", "func NewFileDataConnector()(*FileDataConnector) {\n m := &FileDataConnector{\n IndustryDataConnector: *NewIndustryDataConnector(),\n }\n odataTypeValue := \"#microsoft.graph.industryData.fileDataConnector\"\n m.SetOdataType(&odataTypeValue)\n return m\n}", "func (c *nrConnector) Connect(ctx context.Context) (driver.Conn, error) {\n\tconn, err := c.original.Connect(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn wrapConn(conn, c.segmenter), nil\n}", "func NewExternalConnection()(*ExternalConnection) {\n m := &ExternalConnection{\n Entity: *iadcd81124412c61e647227ecfc4449d8bba17de0380ddda76f641a29edf2b242.NewEntity(),\n }\n return m\n}", "func newLocalSiteConnector(hostname string, port uint16, recvFunc workQueueFunc) *localSiteConnector {\n\tc := &localSiteConnector{}\n\n\tc.connMgr = newConnectionManager(\"local sites\", func() (interface{}, error) {\n\t\tservAddr := fmt.Sprintf(\"%s:%d\", hostname, port)\n\t\ttcpAddr, err := net.ResolveTCPAddr(\"tcp\", servAddr)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\treturn net.DialTCP(\"tcp\", nil, tcpAddr)\n\t})\n\n\tc.recvQueue = newWorkQueue(recvFunc)\n\tc.sendQueue = newWorkQueue(c.sendMessage)\n\n\tgo func() {\n\t\tc.connMgr.connect()\n\t\tc.connMgr.startReconnectLoop()\n\t\tc.startReadLoop()\n\t\tc.sendQueue.start()\n\t\tc.recvQueue.start()\n\t}()\n\n\treturn c\n}", "func (c Client) GetConnector(req ConnectorRequest) (ConnectorResponse, error) {\n\tresp := ConnectorResponse{}\n\n\tstatusCode, err := c.Request(http.MethodGet, fmt.Sprintf(\"connectors/%s\", req.Name), nil, &resp)\n\tif err != nil {\n\t\treturn ConnectorResponse{}, err\n\t}\n\n\t// because connector missing is not an error\n\tif resp.ErrorCode != 0 && resp.ErrorCode != 404 {\n\t\treturn ConnectorResponse{}, resp.ErrorResponse\n\t}\n\n\tresp.Code = statusCode\n\treturn resp, nil\n}", "func RegisterConnector(name string, connstarter func(Handler, *log.Logger) Connector) {\n\tif stopRegistrations {\n\t\treturn\n\t}\n\tif connectors[name] != nil {\n\t\tlog.Fatal(\"Attempted registration of duplicate connector:\", name)\n\t}\n\tconnectors[name] = connstarter\n}", "func (nc *NetClient) SetConnector(connector Connector) {\n\tnc.connectFunc = connector\n}", "func NewTestServerFromConnector(port int, sc data.Connector) (*httptest.Server, error) {\n\troot := mux.NewRouter()\n\troute.GetHandler(root, sc)\n\tn := negroni.New()\n\tn.UseHandler(root)\n\n\tserver := httptest.NewUnstartedServer(n)\n\taddr := fmt.Sprintf(\":%d\", port)\n\tl, err := net.Listen(\"tcp\", addr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsc.SetURL(fmt.Sprintf(\"http://localhost:%d\", port))\n\tserver.Listener = l\n\tserver.Start()\n\n\tgrip.Infoln(\"started server:\", sc.GetURL())\n\n\treturn server, nil\n}", "func (c Client) UpdateConnector(req CreateConnectorRequest, sync bool) (ConnectorResponse, error) {\n\tresp := ConnectorResponse{}\n\n\tstatusCode, err := c.Request(http.MethodPut, fmt.Sprintf(\"connectors/%s/config\", req.Name), req.Config, &resp)\n\tif err != nil {\n\t\treturn ConnectorResponse{}, err\n\t}\n\tif resp.ErrorCode != 0 {\n\t\treturn ConnectorResponse{}, resp.ErrorResponse\n\t}\n\n\tresp.Code = statusCode\n\n\tif sync {\n\t\tif !TryUntil(\n\t\t\tfunc() bool {\n\t\t\t\tuptodate, err := c.IsUpToDate(req.Name, req.Config)\n\t\t\t\treturn err == nil && uptodate\n\t\t\t},\n\t\t\t2*time.Minute,\n\t\t) {\n\t\t\treturn resp, errors.New(\"timeout on creating connector sync\")\n\t\t}\n\t}\n\n\treturn resp, nil\n}", "func (c connector) Driver() driver.Driver { return c.drv }", "func (c connector) Driver() driver.Driver { return c.drv }", "func New(config *config.Config, connector Connector) Bot {\n\tlogIn := make(chan msg.Message)\n\tlogOut := make(chan msg.Messages)\n\n\tmsglog.RunNew(logIn, logOut)\n\n\thistorySz := config.GetInt(\"bot.historysz\", 100)\n\n\tusers := []user.User{\n\t\t{\n\t\t\tName: config.Get(\"Nick\", \"bot\"),\n\t\t},\n\t}\n\n\tbot := &bot{\n\t\tconfig: config,\n\t\tplugins: make(map[string]Plugin),\n\t\tpluginOrdering: make([]string, 0),\n\t\tpluginBlacklist: make(map[string]bool),\n\t\tpluginWhitelist: make(map[string]bool),\n\t\tconn: connector,\n\t\tusers: users,\n\t\tme: users[0],\n\t\tlogIn: logIn,\n\t\tlogOut: logOut,\n\t\thttpEndPoints: make([]EndPoint, 0),\n\t\tfilters: make(map[string]func(string) string),\n\t\tcallbacks: make(CallbackMap),\n\t\trouter: chi.NewRouter(),\n\t\thistory: history.New(historySz),\n\t}\n\n\tbot.migrateDB()\n\n\tbot.RefreshPluginBlacklist()\n\tbot.RefreshPluginWhitelist()\n\n\tlog.Debug().Msgf(\"created web router\")\n\n\tbot.setupHTTP()\n\n\tconnector.RegisterEvent(bot.Receive)\n\n\treturn bot\n}", "func (b *bot) DefaultConnector() Connector {\n\treturn b.conn\n}", "func (c *Connector) Driver() driver.Driver {\n\treturn &Driver{}\n}", "func (c Connector) Connect(name string) (*grpc.ClientConn, error) {\n\tconnection, err := c.Get(name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tcnn, err := connection.Connect()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn cnn.Conn, nil\n}", "func (oci8Connector *OCI8Connector) Connect(ctx context.Context) (driver.Conn, error) {\n\toci8Conn := &OCI8Conn{\n\t\tlogger: oci8Connector.Logger,\n\t}\n\tif oci8Conn.logger == nil {\n\t\toci8Conn.logger = log.New(ioutil.Discard, \"\", 0)\n\t}\n\n\treturn oci8Conn, nil\n}", "func NewCassandraConnector(\n\tconfig *Config,\n\tscope tally.Scope,\n) (orm.Connector, error) {\n\tsession, err := CreateStoreSession(\n\t\tconfig.CassandraConn, config.StoreName)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// create a storeScope for the keyspace StoreName\n\tstoreScope := scope.SubScope(\"cql\").Tagged(\n\t\tmap[string]string{\"store\": config.StoreName})\n\n\treturn &cassandraConnector{\n\t\tSession: session,\n\t\tscope: storeScope,\n\t\texecuteSuccessScope: storeScope.Tagged(\n\t\t\tmap[string]string{\"result\": \"success\"}),\n\t\texecuteFailScope: storeScope.Tagged(\n\t\t\tmap[string]string{\"result\": \"fail\"}),\n\t\tConf: config,\n\t}, nil\n}", "func NewRemote() (Catalog, error) {\n\treturn newRemoteFunc()\n}", "func NewFactory() connector.Factory {\n\treturn connector.NewFactory(\n\t\ttypeStr,\n\t\tcreateDefaultConfig,\n\t\tconnector.WithTracesToMetrics(createTracesToMetricsConnector, stability),\n\t)\n}", "func NewConnectCmd(f factory.Factory) *cobra.Command {\n\tconnectCmd := &cobra.Command{\n\t\tUse: \"connect\",\n\t\tShort: \"Connect an external cluster to devspace cloud\",\n\t\tLong: `\n#######################################################\n################# devspace connect ####################\n#######################################################\n\t`,\n\t\tArgs: cobra.NoArgs,\n\t}\n\n\tconnectCmd.AddCommand(newClusterCmd(f))\n\n\treturn connectCmd\n}", "func CreateTraceCloningFanOutConnector(tcs []consumer.TraceConsumerBase) consumer.TraceConsumerBase {\n\tif len(tcs) == 1 {\n\t\t// Don't wrap if no need to do it.\n\t\treturn tcs[0]\n\t}\n\ttraceConsumersOld := make([]consumer.TraceConsumerOld, 0, len(tcs))\n\ttraceConsumers := make([]consumer.TraceConsumer, 0, len(tcs))\n\tallTraceConsumersOld := true\n\tfor _, tc := range tcs {\n\t\tif traceConsumer, ok := tc.(consumer.TraceConsumer); ok {\n\t\t\tallTraceConsumersOld = false\n\t\t\ttraceConsumers = append(traceConsumers, traceConsumer)\n\t\t} else {\n\t\t\ttraceConsumerOld := tc.(consumer.TraceConsumerOld)\n\t\t\ttraceConsumersOld = append(traceConsumersOld, traceConsumerOld)\n\t\t\ttraceConsumers = append(traceConsumers, converter.NewInternalToOCTraceConverter(traceConsumerOld))\n\t\t}\n\t}\n\n\tif allTraceConsumersOld {\n\t\treturn newTraceCloningFanOutConnectorOld(traceConsumersOld)\n\t}\n\treturn newTraceCloningFanOutConnector(traceConsumers)\n}", "func (srv *Server) newConn(rwc net.Conn) *conn {\n\treturn &conn{\n\t\tserver: srv,\n\t\trwc: rwc,\n\t}\n\n}", "func (dsf *testConnSource) newConn() (net.Conn, error) {\n\t// If we get called again after already opening a connection\n\t// (during the event subscription test for instance)\n\t// then make sure to close the existing connection (if there\n\t// is one) so we don't leak connections\n\tif server != nil {\n\t\tserver.Close()\n\t}\n\n\tserver, client = net.Pipe()\n\treturn client, nil\n}", "func newConn(params planetscalev2.VitessLockserverParams) *Conn {\n\tnow := time.Now()\n\tc := &Conn{\n\t\tparams: params,\n\t\tconnectDone: make(chan struct{}),\n\t\tlastOpened: now,\n\t\tlastChecked: now,\n\t}\n\n\tgo func() {\n\t\tconnLog := log.WithFields(logrus.Fields{\n\t\t\t\"implementation\": params.Implementation,\n\t\t\t\"address\": params.Address,\n\t\t\t\"rootPath\": params.RootPath,\n\t\t})\n\t\tconnLog.Info(\"connecting to Vitess topology server\")\n\n\t\tstartTime := time.Now()\n\t\tdefer func() {\n\t\t\tconnectLatency.Observe(time.Since(startTime).Seconds())\n\t\t}()\n\n\t\t// OpenServer has a built-in timeout that's not configurable.\n\t\t// TODO(enisoc): Upstream a change to make the timeout configurable.\n\t\tc.Server, c.connectErr = topo.OpenServer(params.Implementation, params.Address, params.RootPath)\n\t\tif c.connectErr == nil {\n\t\t\tconnLog.Info(\"successfully connected to Vitess topology server\")\n\t\t\tconnectSuccesses.Inc()\n\t\t} else {\n\t\t\tconnLog.WithField(\"err\", c.connectErr).Warning(\"failed to connect to Vitess topology server\")\n\t\t\tconnectErrors.Inc()\n\t\t}\n\t\tclose(c.connectDone)\n\t}()\n\n\treturn c\n}", "func NewExternal(stype linux.SockType, uid uniqueid.Provider, queue *waiter.Queue, receiver Receiver, connected ConnectedEndpoint) Endpoint {\n\tep := &connectionedEndpoint{\n\t\tbaseEndpoint: baseEndpoint{Queue: queue, receiver: receiver, connected: connected},\n\t\tid: uid.UniqueID(),\n\t\tidGenerator: uid,\n\t\tstype: stype,\n\t}\n\tep.ops.InitHandler(ep, &stackHandler{}, getSendBufferLimits, getReceiveBufferLimits)\n\tep.ops.SetSendBufferSize(connected.SendMaxQueueSize(), false /* notify */)\n\tep.ops.SetReceiveBufferSize(defaultBufferSize, false /* notify */)\n\treturn ep\n}", "func newConn(w http.ResponseWriter, r *http.Request, p url.Values) Conn {\n\treturn Conn{Writer: w, Request: r, Params: p, halted: false}\n}", "func New(busNumber int) (ConnCloser, error) {\n\topener, err := find(busNumber)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn opener()\n}", "func New(\n\trtspAddress string,\n\treadTimeout time.Duration,\n\trunOnConnect string,\n\trunOnConnectRestart bool,\n\tpathMan PathMan,\n\tstats *stats.Stats,\n\tconn *gortsplib.ServerConn,\n\tparent Parent) *Conn {\n\tc := &Conn{\n\t\trtspAddress: rtspAddress,\n\t\treadTimeout: readTimeout,\n\t\trunOnConnect: runOnConnect,\n\t\trunOnConnectRestart: runOnConnectRestart,\n\t\tpathMan: pathMan,\n\t\tstats: stats,\n\t\tconn: conn,\n\t\tparent: parent,\n\t}\n\n\tc.log(logger.Info, \"opened\")\n\n\tif c.runOnConnect != \"\" {\n\t\t_, port, _ := net.SplitHostPort(c.rtspAddress)\n\t\tc.onConnectCmd = externalcmd.New(c.runOnConnect, c.runOnConnectRestart, externalcmd.Environment{\n\t\t\tPath: \"\",\n\t\t\tPort: port,\n\t\t})\n\t}\n\n\treturn c\n}", "func newRawProxyConn(t *testing.T) *mysql.MySQLConn {\n\tnewTestServer()\n\n\td := mysql.MySQLDriver{}\n\n\tif conn, err := d.Open(testProxyDSN); err != nil {\n\t\tt.Fatal(err)\n\t} else if c, ok := conn.(*mysql.MySQLConn); !ok {\n\t\tt.Fatal(\"connection is not MySQLConn type\")\n\t} else {\n\t\treturn c\n\t}\n\n\treturn nil\n}", "func newConn(c conn) *Conn {\n\trtc := &Conn{\n\t\tc: c,\n\t}\n\n\trtc.Link = &LinkService{c: rtc}\n\trtc.Address = &AddressService{c: rtc}\n\trtc.Route = &RouteService{c: rtc}\n\trtc.Neigh = &NeighService{c: rtc}\n\n\treturn rtc\n}", "func New(v cty.Value) (*Communicator, error) {\n\tconnInfo, err := parseConnectionInfo(v)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tendpoint := &winrm.Endpoint{\n\t\tHost: connInfo.Host,\n\t\tPort: int(connInfo.Port),\n\t\tHTTPS: connInfo.HTTPS,\n\t\tInsecure: connInfo.Insecure,\n\t\tTimeout: connInfo.TimeoutVal,\n\t}\n\tif len(connInfo.CACert) > 0 {\n\t\tendpoint.CACert = []byte(connInfo.CACert)\n\t}\n\n\tcomm := &Communicator{\n\t\tconnInfo: connInfo,\n\t\tendpoint: endpoint,\n\t\t// Seed our own rand source so that script paths are not deterministic\n\t\trand: rand.New(rand.NewSource(time.Now().UnixNano())),\n\t}\n\n\treturn comm, nil\n}", "func NewBackend(url string) Backend {\n\tcli, err := ethclient.Dial(url)\n\tif err != nil {\n\t\tlog.Fatal(err.Error())\n\t}\n\treturn &backend{connect: cli}\n}", "func NewProtocol(bc blockchain.Blockchain) *Protocol { return &Protocol{bc} }", "func (db *Database) InsertConnector(connector *models.ConnectorBase) error {\n\tif !open {\n\t\treturn fmt.Errorf(\"db must be opened before saving!\")\n\t}\n\terr := db.bolt.Update(func(tx *bolt.Tx) error {\n\t\tb := tx.Bucket([]byte(connectorBucketName))\n\t\tenc, err := json.Marshal(connector)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"could not encode module %s: %s\", connector.GetName(), err)\n\t\t}\n\n\t\terr = b.Put([]byte(connector.GetID()), enc)\n\t\treturn err\n\t})\n\treturn err\n}", "func (r *ProjectsLocationsConnectorsService) Create(parent string, connector *Connector) *ProjectsLocationsConnectorsCreateCall {\n\tc := &ProjectsLocationsConnectorsCreateCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.parent = parent\n\tc.connector = connector\n\treturn c\n}", "func (q *Qsign) SetConnector(s string) {\n\tq.connector = s\n}", "func (s *Server) newConn(rwc net.Conn) *conn {\n\treturn &conn{\n\t\tserver: s,\n\t\trwc: rwc,\n\t\tbrc: bufio.NewReader(rwc),\n\t}\n}", "func newConn(ctx context.Context, t *Transport, remoteMa ma.Multiaddr,\n\tremotePID peer.ID, inbound bool) (tpt.CapableConn, error) {\n\t// Creates a manet.Conn\n\tpr, pw := io.Pipe()\n\tconnCtx, cancel := context.WithCancel(gListener.ctx)\n\n\tmaconn := &Conn{\n\t\treadIn: pw,\n\t\treadOut: pr,\n\t\tlocalMa: gListener.localMa,\n\t\tremoteMa: remoteMa,\n\t\tctx: connCtx,\n\t\tcancel: cancel,\n\t}\n\n\t// Unlock gListener locked from discovery.go (HandlePeerFound)\n\tgListener.inUse.Done()\n\n\t// Stores the conn in connMap, will be deleted during conn.Close()\n\tconnMap.Store(maconn.RemoteAddr().String(), maconn)\n\n\t// Returns an upgraded CapableConn (muxed, addr filtered, secured, etc...)\n\tif inbound {\n\t\treturn t.upgrader.UpgradeInbound(ctx, t, maconn)\n\t}\n\treturn t.upgrader.UpgradeOutbound(ctx, t, maconn, remotePID)\n}", "func newPeerClientConnection(block bool) (*grpc.ClientConn, error) {\n\treturn newPeerClientConnectionWithAddress(block, viper.GetString(\"service.cliaddress\"))\n}", "func newDriver() *driver {\n\treturn &driver{\n\t\tnetworks: map[string]*bridgeNetwork{},\n\t\tportAllocator: portallocator.Get(),\n\t}\n}", "func NewConn(protoName string, target string, id int) (*Conn, error) {\n if target == \"\" {\n return nil, fmt.Errorf(\"No target given\")\n }\n\n c := &Conn {\n\n }\n\n if proto, exists := protocols[protoName]; !exists {\n return nil, fmt.Errorf(\"protocol not registered: %v\", protoName)\n } else {\n c.proto = proto\n }\n\n if ipAddr, err := net.ResolveIPAddr(c.proto.resolveStr, target); err != nil {\n return nil, fmt.Errorf(\"net.ResolveIPAddr %v:%v: %v\", c.proto.resolveStr, target, err)\n } else {\n // unprivileged icmp mode uses SOCK_DGRAM\n c.targetAddr = &net.UDPAddr{IP: ipAddr.IP, Zone: ipAddr.Zone}\n }\n\n if icmpConn, err := icmp.ListenPacket(c.proto.listenStr, \"\"); err != nil {\n return nil, fmt.Errorf(\"icmp.ListenPacket %v: %v\", c.proto.listenStr, err)\n } else {\n c.icmpConn = icmpConn\n }\n\n // store local address\n if _, id, err := decodeAddr(c.icmpConn.LocalAddr()); err != nil {\n return nil, fmt.Errorf(\"Unkonwn icmpConn.LocalAddr(): %v\", err)\n } else {\n c.id = id\n }\n\n return c, nil\n}", "func (d *drv) Open(s string) (driver.Conn, error) {\n\tc, err := d.OpenConnector(s)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn d.createConnFromParams(context.Background(), c.(connector).ConnectionParams)\n}", "func TestAllocateConnector(t *testing.T) {\n\taddr, err := allocateConnector([]string{\"http://localhost:4001\"})\n\tif err != nil {\n\t\tt.Errorf(\"allocateConnector error (%v)\", err)\n\t}\n\tfmt.Printf(\"allocated %s\\n\", addr)\n}", "func New(api restapi.Connector) *Engine {\n\treturn &Engine{api: api}\n}", "func NewOSRMConnector(osrmEndpoint string) *OSRMConnector {\n\tosrm := &OSRMConnector{\n\t\tosrmClient: newOsrmHTTPClient(osrmEndpoint),\n\t}\n\tgo osrm.osrmClient.start()\n\treturn osrm\n}", "func newClient(addr string, max int, discardClientTimeout time.Duration, fn connectRPCFn) (Client, error) {\n\n\trpcClientFactory := func() (interface{}, error) {\n\t\treturn fn(addr)\n\t}\n\trpcPool, err := pool.NewPool(max, rpcClientFactory)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\trc := &reconnectingClient{addr: addr, pool: rpcPool, discardClientTimeout: discardClientTimeout}\n\treturn rc, nil\n}", "func (c Client) DeleteConnector(req ConnectorRequest, sync bool) (EmptyResponse, error) {\n\tresp := EmptyResponse{}\n\n\tstatusCode, err := c.Request(http.MethodDelete, fmt.Sprintf(\"connectors/%s\", req.Name), nil, &resp)\n\tif err != nil {\n\t\treturn EmptyResponse{}, err\n\t}\n\tif resp.ErrorCode != 0 {\n\t\treturn EmptyResponse{}, resp.ErrorResponse\n\t}\n\n\tresp.Code = statusCode\n\n\tif sync {\n\t\tif !TryUntil(\n\t\t\tfunc() bool {\n\t\t\t\tr, e := c.GetConnector(req)\n\t\t\t\treturn e == nil && r.Code == 404\n\t\t\t},\n\t\t\t2*time.Minute,\n\t\t) {\n\t\t\treturn resp, errors.New(\"timeout on deleting connector sync\")\n\t\t}\n\t}\n\n\treturn resp, nil\n}", "func (m *MockChoriaProvider) Connector() inter.Connector {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"Connector\")\n\tret0, _ := ret[0].(inter.Connector)\n\treturn ret0\n}", "func SSHCommNew(address string, config *SSHCommConfig) (result *comm, err error) {\n\t// Establish an initial connection and connect\n\tresult = &comm{\n\t\tconfig: config,\n\t\taddress: address,\n\t}\n\n\tif err = result.reconnect(); err != nil {\n\t\tresult = nil\n\t\treturn\n\t}\n\n\treturn\n}", "func newPair() (pcOffer *PeerConnection, pcAnswer *PeerConnection, err error) {\n\tpca, err := NewPeerConnection(Configuration{})\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tpcb, err := NewPeerConnection(Configuration{})\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treturn pca, pcb, nil\n}", "func newCloudConnection(config io.Reader) (cloudprovider.Interface, error) {\n\tklog.V(4).Infof(\"newCloudConnection called with %+v\", config)\n\tif config != nil {\n\t\tklog.Warningf(\"supplied config is not read by this version. Using environment\")\n\t}\n\tnewCloud := &cloud{}\n\t_, err := newCloud.cloudClient()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn newCloud, nil\n}", "func (c *Connector) Connect(\n\tr *gohttp.Request,\n\tcredentialsByID connector.CredentialValuesByID,\n) error {\n\t// TODO: add logic according to\n\t// https://github.com/cyberark/secretless-broker/blob/master/pkg/secretless/plugin/connector/README.md#http-connector\n\n\tvar err error\n\treturn err\n}", "func NewDisconnect() *Disconnect {\n\treturn &Disconnect{}\n}", "func newConnection(c net.Conn) *connection {\n\treturn &connection{\n\t\tc: c,\n\t}\n}", "func Connect(username string, password string) *connection {\n\tconnection_new := new(connection)\n\tconnection_new.username = username\n\tconnection_new.pass = password\n\tconnection_new.getSessionId()\n\tconnection_new.time = time.Now()\n\treturn connection_new\n}", "func (c *driverConnector) Close() error {\n\tif c, ok := c.Connector.(io.Closer); ok {\n\t\treturn c.Close()\n\t}\n\treturn nil\n}", "func New(chainID, remote string) (provider.Provider, error) {\n\thttpClient, err := rpcclient.NewHTTP(remote, \"/websocket\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn NewWithClient(chainID, httpClient), nil\n}", "func (c *connector) Connect(ctx context.Context) (driver.Conn, error) {\n\tif bhConnCtx, ok := ctx.(*bytehouse.ConnectionContext); ok {\n\t\tc.connContext = bhConnCtx\n\t} else {\n\t\tc.connContext.Context = ctx\n\t}\n\n\tgateway, err := sdk.Open(c.connContext, c.dsn)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tcn := &CHConn{\n\t\tGateway: gateway,\n\t}\n\treturn cn, nil\n}", "func CreateMetricsCloningFanOutConnector(mcs []consumer.MetricsConsumerBase) consumer.MetricsConsumerBase {\n\tif len(mcs) == 1 {\n\t\t// Don't wrap if no need to do it.\n\t\treturn mcs[0]\n\t}\n\tmetricsConsumersOld := make([]consumer.MetricsConsumerOld, 0, len(mcs))\n\tmetricsConsumers := make([]consumer.MetricsConsumer, 0, len(mcs))\n\tallMetricsConsumersOld := true\n\tfor _, mc := range mcs {\n\t\tif metricsConsumer, ok := mc.(consumer.MetricsConsumer); ok {\n\t\t\tallMetricsConsumersOld = false\n\t\t\tmetricsConsumers = append(metricsConsumers, metricsConsumer)\n\t\t} else {\n\t\t\tmetricsConsumerOld := mc.(consumer.MetricsConsumerOld)\n\t\t\tmetricsConsumersOld = append(metricsConsumersOld, metricsConsumerOld)\n\t\t\tmetricsConsumers = append(metricsConsumers, converter.NewInternalToOCMetricsConverter(metricsConsumerOld))\n\t\t}\n\t}\n\n\tif allMetricsConsumersOld {\n\t\treturn newMetricsCloningFanOutConnectorOld(metricsConsumersOld)\n\t}\n\treturn newMetricsCloningFanOutConnector(metricsConsumers)\n}", "func (srv *Server) newConn(rwc net.Conn) *conn {\n\treturn &conn{\n\t\tserver: srv,\n\t\tmuc: newMuxConn(rwc),\n\t}\n}", "func setConnector(c Connector) {\n\trobot.Lock()\n\trobot.Connector = c\n\trobot.Unlock()\n}", "func NewConn() (*Conn, error) {\n\tc := new(Conn)\n\n\tconn, err := share.GetSystemBusPrivateConn()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to connect to system bus: %v\", err)\n\t}\n\n\tc.conn = conn\n\tc.object = conn.Object(dbusInterface, dbus.ObjectPath(dbusPath))\n\n\treturn c, nil\n}", "func newConn(dial func(string) (net.Conn, error)) func(string) (net.Conn, error) {\n\treturn func(addr string) (net.Conn, error) {\n\t\t// get the proxy url\n\t\tproxyURL, err := getURL(addr)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\t// set to addr\n\t\tcallAddr := addr\n\n\t\t// got proxy\n\t\tif proxyURL != nil {\n\t\t\tcallAddr = proxyURL.Host\n\t\t}\n\n\t\t// dial the addr\n\t\tc, err := dial(callAddr)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\t// do proxy connect if we have proxy url\n\t\tif proxyURL != nil {\n\t\t\tc, err = proxyDial(c, addr, proxyURL)\n\t\t}\n\n\t\treturn c, err\n\t}\n}" ]
[ "0.7714393", "0.74868536", "0.74652475", "0.7335159", "0.7311851", "0.72668517", "0.72528714", "0.72027683", "0.7158545", "0.7097409", "0.7038028", "0.693401", "0.6924261", "0.6905243", "0.6848552", "0.66566956", "0.64212936", "0.63512844", "0.63188505", "0.62497795", "0.61543846", "0.6045318", "0.58601207", "0.5831694", "0.56931114", "0.5638064", "0.56089294", "0.5582439", "0.55665565", "0.5526995", "0.5526995", "0.55001646", "0.543472", "0.54247177", "0.53994757", "0.53734094", "0.536235", "0.52267915", "0.5212055", "0.5206879", "0.51958704", "0.51670206", "0.51327103", "0.5129735", "0.5115042", "0.51068133", "0.5099355", "0.5099355", "0.5098336", "0.50880414", "0.5086298", "0.50777006", "0.5075933", "0.5051742", "0.504507", "0.5044768", "0.5042518", "0.50199676", "0.5015949", "0.5014612", "0.50090784", "0.5002636", "0.50001425", "0.49635074", "0.4962565", "0.49500558", "0.4927315", "0.49043763", "0.49018064", "0.48947102", "0.4893113", "0.48895723", "0.4888255", "0.4879625", "0.4856633", "0.48498777", "0.4845157", "0.48420817", "0.48349726", "0.4831886", "0.48281083", "0.48168385", "0.48080525", "0.4807885", "0.48013753", "0.48008728", "0.4790231", "0.47829872", "0.47791892", "0.476062", "0.47602993", "0.47562358", "0.475242", "0.47509715", "0.4738617", "0.47328398", "0.47236246", "0.47213343", "0.47168404", "0.4710961" ]
0.71470785
9
Connect returns a connection to the database. Connect may return a cached connection (one previously closed), but doing so is unnecessary; the sql package maintains a pool of idle connections for efficient reuse. The provided context.Context is for dialing purposes only (see net.DialContext) and should not be stored or used for other purposes. A default timeout should still be used when dialing as a connection pool may call Connect asynchronously to any query. The returned connection is only used by one goroutine at a time.
func (c *connector) Connect(ctx context.Context) (driver.Conn, error) { if bhConnCtx, ok := ctx.(*bytehouse.ConnectionContext); ok { c.connContext = bhConnCtx } else { c.connContext.Context = ctx } gateway, err := sdk.Open(c.connContext, c.dsn) if err != nil { return nil, err } cn := &CHConn{ Gateway: gateway, } return cn, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Connect(ctx context.Context, driver, dsn string, connOpts *ConnectOptions) (*sqlx.DB, error) {\n\topts := connOpts\n\tif opts == nil {\n\t\topts = &ConnectOptions{}\n\t}\n\n\tdb, err := connectWithRetry(ctx, driver, dsn, opts.Retry)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdb.SetMaxOpenConns(opts.MaxOpenConnections)\n\tdb.SetMaxIdleConns(opts.MaxIdleConnections)\n\tdb.SetConnMaxLifetime(opts.ConnectionMaxLifetime)\n\treturn db, nil\n}", "func (c *Client) Connect() (*DBConnection, error) {\n\tdbRegistryLock.Lock()\n\tdefer dbRegistryLock.Unlock()\n\n\tdsn := c.config.connStr(c.databaseName)\n\tconn, found := dbRegistry[dsn]\n\tif !found {\n\t\tdb, err := sql.Open(proxyDriverName, dsn)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"Error connecting to PostgreSQL server %s: %w\", c.config.Host, err)\n\t\t}\n\n\t\t// We don't want to retain connection\n\t\t// So when we connect on a specific database which might be managed by terraform,\n\t\t// we don't keep opened connection in case of the db has to be dopped in the plan.\n\t\tdb.SetMaxIdleConns(0)\n\t\tdb.SetMaxOpenConns(c.config.MaxConns)\n\n\t\tconn = &DBConnection{\n\t\t\tdb,\n\t\t\tc,\n\t\t}\n\t\tdbRegistry[dsn] = conn\n\t}\n\n\treturn conn, nil\n}", "func Connect(ctx context.Context, host string, port int, dbName, user, password string) (*DB, error) {\n\tconfig := mysql.Config{\n\t\tAddr: fmt.Sprintf(\"%s:%d\", host, port),\n\t\tNet: \"tcp\",\n\t\tUser: user,\n\t\tPasswd: password,\n\t\tDBName: dbName,\n\t\tMultiStatements: true,\n\t}\n\tctxLogger := logger.FromContext(ctx)\n\tctx = logger.NewContext(ctx, ctxLogger.(logger.WithLogger).With(\"host\", host, \"dbName\", dbName, \"user\", user, \"port\", port))\n\n\tdb := &DB{\n\t\tCtx: ctx,\n\t\tLogger: logger.FromContext(ctx),\n\t}\n\tdb.Logger.Info(\"dsn\", config.FormatDSN(), \"msg\", \"Connecting\")\n\tif myLogger, ok := db.Logger.(logger.PrintLogger); ok {\n\t\tif myWithLogger, okWith := db.Logger.(logger.WithLogger); okWith {\n\t\t\tmyLogger = myWithLogger.With(\"package\", \"mysql\").(logger.PrintLogger)\n\t\t}\n\t\tmysql.SetLogger(myLogger)\n\t}\n\tcon, err := sql.Open(\"mysql\", config.FormatDSN())\n\tif err != nil {\n\t\treturn db, err\n\t}\n\terr = con.PingContext(ctx)\n\tif err != nil {\n\t\treturn db, err\n\t}\n\tdb.Database = goqu.New(\"mysql\", con)\n\treturn db, nil\n}", "func (conn *Conn) ConnectContext(ctx context.Context) error {\n\t// We don't want to hold conn.mu while firing the REGISTER event,\n\t// and it's much easier and less error prone to defer the unlock,\n\t// so the connect mechanics have been delegated to internalConnect.\n\terr := conn.internalConnect(ctx)\n\tif err == nil {\n\t\tconn.dispatch(&Line{Cmd: REGISTER, Time: time.Now()})\n\t}\n\treturn err\n}", "func Connect(driver, DSN string, timeout time.Duration) (*sql.DB, error) {\n\tticker := time.NewTicker(1 * time.Second)\n\tdefer ticker.Stop()\n\n\ttimeoutExceeded := time.After(timeout)\n\tfor {\n\t\tselect {\n\t\tcase <-timeoutExceeded:\n\t\t\treturn nil, fmt.Errorf(\"db connection failed after %s timeout\", timeout)\n\n\t\tcase <-ticker.C:\n\t\t\tdb, err := sql.Open(driver, DSN)\n\t\t\tif err == nil {\n\t\t\t\treturn db, nil\n\t\t\t}\n\t\t\tlog.Println(errors.Wrapf(err, \"failed to connect to db %s\", DSN))\n\t\t}\n\t}\n}", "func (c connector) Connect(ctx context.Context) (driver.Conn, error) {\n\tparams := c.ConnectionParams\n\tlogger := c.CommonParams.Logger\n\tif ctxValue := ctx.Value(paramsCtxKey{}); ctxValue != nil {\n\t\tif cc, ok := ctxValue.(commonAndConnParams); ok {\n\t\t\t// ContextWithUserPassw does not fill ConnParam.ConnectString\n\t\t\tif cc.ConnectString == \"\" {\n\t\t\t\tcc.ConnectString = params.ConnectString\n\t\t\t}\n\t\t\tlogger = cc.Logger\n\t\t\tif logger != nil {\n\t\t\t\tlogger.Debug(\"connect with params from context\", \"poolParams\", params.PoolParams, \"connParams\", cc, \"common\", cc.CommonParams)\n\t\t\t}\n\t\t\treturn c.drv.createConnFromParams(ctx, dsn.ConnectionParams{\n\t\t\t\tCommonParams: cc.CommonParams, ConnParams: cc.ConnParams,\n\t\t\t\tPoolParams: params.PoolParams,\n\t\t\t})\n\t\t}\n\t}\n\n\tif ctxValue := ctx.Value(userPasswCtxKey{}); ctxValue != nil {\n\t\tif up, ok := ctxValue.(UserPasswdConnClassTag); ok {\n\t\t\tparams.CommonParams.Username = up.Username\n\t\t\tparams.CommonParams.Password = up.Password\n\t\t\tparams.ConnParams.ConnClass = up.ConnClass\n\t\t}\n\t}\n\n\tif logger != nil {\n\t\tlogger.Debug(\"connect\", \"poolParams\", params.PoolParams, \"connParams\", params.ConnParams, \"common\", params.CommonParams)\n\t}\n\treturn c.drv.createConnFromParams(ctx, params)\n}", "func Connect(ctx context.Context, cfg Config) (*DB, error) {\n\tpool, err := pgxpool.New(ctx, ConnString(cfg))\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"database connection error: %w\", err)\n\t}\n\tdb := DB{pool}\n\n\treturn &db, nil\n}", "func (self *CSql) Connect(driverName string, dial string) (*sql.DB, error) {\r\n\tvar err error = nil\r\n\tself.db, err = sql.Open(driverName, dial)\r\n\tself.db.SetMaxOpenConns(20)\r\n\tself.db.SetMaxIdleConns(10)\r\n\tself.db.SetConnMaxLifetime(60 * time.Second)\r\n\treturn self.db, err\r\n}", "func Connect() (*sql.DB, error) {\n\treturn db.Connect(*dbURI)\n}", "func ConnectMockContext(ctx context.Context, t *testing.T) (*DB, sqlmock.Sqlmock, error) {\n\tcon, mock, err := sqlmock.New()\n\tif err != nil {\n\t\tt.Errorf(\"an error '%s' was not expected when opening a stub database connection\", err)\n\t\tt.Fail()\n\t}\n\tdb := &DB{\n\t\tCtx: ctx,\n\t\tDatabase: goqu.New(\"mysql\", con),\n\t\tLogger: logger.FromContext(ctx),\n\t}\n\treturn db, mock, err\n}", "func (c connector) Connect(context.Context) (driver.Conn, error) {\r\n\tconn, err := c.drv.openConn(c.ConnectionParams)\r\n\tif err != nil || c.onInit == nil || !conn.newSession {\r\n\t\treturn conn, err\r\n\t}\r\n\tif err = c.onInit(conn); err != nil {\r\n\t\tconn.close(true)\r\n\t\treturn nil, err\r\n\t}\r\n\treturn conn, nil\r\n}", "func Connect(c *Config) (*sql.DB, error) {\n\n\tdb, err := sql.Open(\"sqlserver\", generateConnectionString(c))\n\tif err != nil {\n\t\tlog.Fatal(\"Error creating connection pool: \" + err.Error())\n\t}\n\treturn db, nil\n}", "func (conn *Conn) Connect() error {\n\treturn conn.ConnectContext(context.Background())\n}", "func Connect(ctx context.Context, cfg config.Database) (*sqlx.DB, error) {\n\tpoolCfg, err := cfg.PoolConfig()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to parse DB connection string: %w\", err)\n\t}\n\n\tconn, err := sqlx.ConnectContext(ctx, \"pgx\", cfg.Address)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to connect to database: %w\", err)\n\t}\n\n\t// stdlib package doesn't have a compat layer for pgxpool\n\t// so had to use standard sql api for pool configuration.\n\tconn.SetConnMaxIdleTime(poolCfg.MaxConnIdleTime)\n\tconn.SetConnMaxLifetime(poolCfg.MaxConnLifetime)\n\tconn.SetMaxOpenConns(int(poolCfg.MaxConns))\n\n\tif err = conn.PingContext(ctx); err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to ping database: %w\", err)\n\t}\n\n\tif cfg.SkipMigration {\n\t\tzap.L().Info(\"database migration skipped\")\n\t\treturn conn, nil\n\t}\n\n\tmp := migrationParams{\n\t\tdbName: poolCfg.ConnConfig.Database,\n\t\tversionTable: cfg.VersionTable,\n\t\tmigrationsDir: cfg.MigrationsDirectory,\n\t\ttargetVersion: cfg.SchemaVersion,\n\t}\n\tif err = runMigration(conn, mp); err != nil {\n\t\tconn.Close()\n\t\treturn nil, err\n\t}\n\n\treturn conn, nil\n}", "func (fn MSSQLConnectorFunc) Connect(ctx context.Context) (net.Conn, error) {\n\treturn fn(ctx)\n}", "func (graphite *Graphite) Connect() error {\n\tif !graphite.IsNop() {\n\t\taddress := fmt.Sprintf(\"%s:%d\", graphite.Host, graphite.Port)\n\n\t\tif graphite.Timeout == 0 {\n\t\t\tgraphite.Timeout = defaultTimeout * time.Second\n\t\t}\n\n\t\tconn, err := net.DialTimeout(\"tcp\", address, graphite.Timeout)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tgraphite.conn = conn\n\t}\n\n\treturn nil\n}", "func Connect(cfg *Config) (*Connection, error) {\n\td, err := gorm.Open(cfg.Driver, cfg.Args)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tc := &Connection{\n\t\tC: d,\n\t\tlog: logrus.WithField(\"context\", \"db\"),\n\t}\n\tc.log.Info(\"connected to database\")\n\treturn c, nil\n}", "func Connect(c Configer) (*gorm.DB, error) {\n\tdb, err := gorm.Open(c.ConnectionStringWithDriver())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdb.LogMode(c.Logging())\n\tdb.DB().SetConnMaxLifetime(c.ConnectionMaxLifetime())\n\tdb.DB().SetMaxOpenConns(c.MaxOpenConnections())\n\tdb.DB().SetMaxIdleConns(c.MaxIdleConnections())\n\tif err = db.DB().Ping(); err != nil {\n\t\tdb.Close()\n\t\treturn nil, err\n\t}\n\treturn db, nil\n}", "func (d *Database) Connect() error {\n\tif d.connection != nil {\n\t\td.log.WithContext().Info(\"Connection to database is already initialized\")\n\t\treturn nil\n\t}\n\n\td.connection = redis.NewClient(&redis.Options{\n\t\tAddr: d.address,\n\t\tPassword: d.password,\n\t\tDB: 0,\n\t})\n\n\tif err := d.connection.Ping(d.ctx).Err(); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (r *Resolver) ConnectDB(ctx context.Context) {\n\tif r.db != nil {\n\t\tfmt.Printf(\"ConnectDB connection already open\\n\")\n\t\treturn\n\t}\n\n\tfmt.Printf(\"ConnectDB connecting...\\n\")\n\tconn, err := pgx.Connect(ctx, os.Getenv(\"DATABASE_URL\"))\n\t// defer conn.Close(ctx) // can't do this, keep the connection open then?\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tr.db = conn\n}", "func (oci8Connector *OCI8Connector) Connect(ctx context.Context) (driver.Conn, error) {\n\toci8Conn := &OCI8Conn{\n\t\tlogger: oci8Connector.Logger,\n\t}\n\tif oci8Conn.logger == nil {\n\t\toci8Conn.logger = log.New(ioutil.Discard, \"\", 0)\n\t}\n\n\treturn oci8Conn, nil\n}", "func Connect() (*sqlx.DB, error) {\n\tpsqlInfo := fmt.Sprintf(\"host=%s port=%d user=%s \"+\n\t\t\"password=%s dbname=%s sslmode=disable\",\n\t\thost, port, user, password, dbname)\n\n\tdb, err := sqlx.Connect(\"postgres\", psqlInfo)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\terr = db.Ping()\n\tif err != nil {\n\t\tdb.Close()\n\t\treturn nil, err\n\t}\n\tlogrus.Debugf(\"Successfully connected to database %s\", psqlInfo)\n\treturn db, nil\n}", "func Connect(configuration *config.Database) (*gorm.DB, error) {\n\tdsn := \"tcp://\" + configuration.Host + \":\" + configuration.Port + \"?database=\" + configuration.DB + \"&read_timeout=10\"\n\tdb, err := gorm.Open(clickhouse.Open(dsn), &gorm.Config{})\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn db, nil\n}", "func Connect(host string, port int, user, password, dbname string, sslmode bool) (*sql.DB, error) {\n\n\tconnStr := fmt.Sprintf(\"host=%s port=%d user=%s password=%s dbname=%s\", host, port, user, password, dbname)\n\tif sslmode {\n\t\tconnStr += \" sslmode=require\"\n\t}\n\n\tvar err error\n\tdb, err = sql.Open(\"postgres\", connStr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = db.Ping()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn db, nil\n}", "func Connect(p *DBConfig) (*sql.DB, error) {\n\tconnStr := fmt.Sprintf(connFmt, p.Host, p.Port, p.User, p.Pass, p.DB)\n\tdb, err := sql.Open(\"postgres\", connStr)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err = db.Ping(); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn db, nil\n}", "func (c *nrConnector) Connect(ctx context.Context) (driver.Conn, error) {\n\tconn, err := c.original.Connect(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn wrapConn(conn, c.segmenter), nil\n}", "func (c *SpannerConnector) Connect(ctx context.Context) (driver.Conn, error) {\n\tconn := &spannerConn{\n\t\tclient: c.client,\n\t\tclosech: make(chan struct{}),\n\t}\n\tconn.startWatcher()\n\tif err := conn.watchCancel(ctx); err != nil {\n\t\tconn.cleanup()\n\t\treturn nil, err\n\t}\n\tdefer conn.finish()\n\n\treturn conn, nil\n}", "func Connect(ctx context.Context, dbName string) (*sql.DB, error) {\n\tdbusername := os.Getenv(\"MARIA_USERNAME\")\n\tdbpassword := os.Getenv(\"MARIA_PASSWORD\")\n\n\tdb, err := sql.Open(\"mysql\", dbusername+\":\"+dbpassword+\"@tcp(127.0.0.1:3306)/\"+dbName)\n\tif err != nil {\n\t\tlogger.Error.Println(logger.GetCallInfo(), err.Error())\n\t\treturn nil, err\n\t}\n\n\terr = db.Ping()\n\tif err != nil {\n\t\tlogger.Error.Println(\"Error:\", err.Error())\n\t\treturn nil, err\n\t}\n\n\treturn db, nil\n}", "func Connect() (MongoConn, error) {\n\tclient, err := mongo.NewClient(clientOptions)\n\n\tctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)\n\n\tdefer cancel()\n\n\terr = client.Connect(ctx)\n\tdb := client.Database(dbName)\n\n\tif err != nil {\n\t\treturn conn, err\n\t}\n\n\tconn = MongoConn{\n\t\tClient: client,\n\t\tDB: db,\n\t}\n\n\tlog.Println(\"[x] - connected to database\")\n\n\treturn conn, nil\n}", "func Connect() (redis.Conn, error) {\n\tconn := pool.Get()\n\n\tif err := conn.Err(); err != nil {\n\t\treturn conn, err\n\t}\n\n\treturn conn, nil\n}", "func Connect(config *cfg.Config) (*sql.DB, error) {\n\tvar dbUri string\n\n\t// defaults or from file\n\tc := *config\n\thost := c.TransactionDB_Host\n\tport := c.TransactionDB_Port\n\tname := c.TransactionDB_Name\n\tuser := c.TransactionDB_User\n\tpassword := c.TransactionDB_Password\n\n\t// Is env variable provided?\n\tenvDbUri := os.Getenv(\"NPROXY_DB_URI\")\n\tif envDbUri != \"\" {\n\t\tdbUri = envDbUri\n\t} else {\n\t\tdbUri = fmt.Sprintf(\"postgres://%s:%s@%s:%d/%s\", user, password, host, port, name)\n\t}\n\n\tlog.Printf(\"Connecting to DB: %s\", dbUri)\n\tdb, err := sql.Open(\"postgres\", dbUri)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t} else {\n\t\t// BUG(m): It goes here even if network connection is unavailable.\n\t\t// Maybe we should detect it calling DNS Resolver?\n\t\tlog.Println(\"Connected to DB\")\n\t}\n\treturn db, err\n}", "func Connect(cf *SQLConfig) (*DB, error) {\n\tconnectionString := fmt.Sprintf(\"postgres://%v:%v@%v:%v/%v?sslmode=%v\", cf.Username, cf.Password, cf.Host, cf.Post, cf.Database, cf.SSLMode)\n\tdb, err := sql.Open(\"postgres\", connectionString)\n\tif err != nil {\n\t\treturn nil, err\n\t} else {\n\t\tpingErr := db.Ping()\n\t\tif pingErr != nil {\n\t\t\treturn nil, errors.New(fmt.Sprintf(\"Cannot connect to database. Error: %s\", pingErr.Error()))\n\t\t} else {\n\t\t\treturn &DB{db}, nil\n\t\t}\n\t}\n}", "func (c Connector) Connect() (db *sql.DB) {\n\tif c.db != nil {\n\t\tlevel.Info(c.log).Log(\"msg\", \"returning mysql\")\n\t\treturn db\n\t}\n\tlevel.Info(c.log).Log(\"msg\", \"connecting to mysql\")\n\tlevel.Info(c.log).Log(\"msg\", \"returning connection\")\n\t_ = c.config.dbdriver\n\tdbUser := c.config.dbUser\n\tdbHost := c.config.dbHost\n\tdbPass := c.config.dbPassword\n\tdbName := c.config.dbDatabase\n\tdbPort := c.config.dbPort\n\n\tdb, err := sql.Open(\"mysql\", dbUser+\":\"+dbPass+\"@(\"+dbHost+\":\"+dbPort+\")/\"+dbName)\n\tif err != nil {\n\t\tpanic(err.Error())\n\t}\n\tc.db = db\n\tc.db.SetMaxIdleConns(10)\n\tc.db.SetMaxOpenConns(10)\n\tc.db.SetConnMaxLifetime(time.Duration(360))\n\treturn c.db\n\n}", "func (s *Sqlite) Connect() error {\n\tvar err error\n\tlog.Debugf(\"Creating SQLite database (%s) if it does not exist...\", s.datasource)\n\tsqlxDB, err := sqlx.Connect(\"sqlite3\", s.datasource+\"?_busy_timeout=5000\")\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"Failed to open sqlite3 DB\")\n\t}\n\ts.SqlxDB = db.New(sqlxDB, s.CAName, s.Metrics)\n\treturn nil\n}", "func Connect() (db *sql.DB, err error) {\n\treturn sql.Open(DBDriver, DBSource+DBParams)\n}", "func (d *Dialer) DialContext(ctx context.Context, addr string) (net.Conn, error) {\n\tconn, err := (&net.Dialer{}).DialContext(ctx, \"tcp\", d.Address)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\terr = d.connect(ctx, conn, addr)\n\tif err != nil {\n\t\tconn.Close()\n\t\treturn nil, err\n\t}\n\treturn conn, nil\n}", "func Connect() (*sql.DB, error) {\n\tdsn := buildDSN()\n\n\tdb, err := sql.Open(\"mysql\", dsn)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tutils.Logger.Info(\"Connecting to database...\", map[string]interface{}{\"DSN\": dsn})\n\tvar dbErr error\n\tfor i := 1; i <= 3; i++ {\n\t\tdbErr = db.Ping()\n\t\tif dbErr != nil {\n\t\t\tutils.Logger.Info(fmt.Sprintf(\"Attempt #%d failed, will retry in 10 seconds\", i), map[string]interface{}{\"Error\": dbErr})\n\t\t\tif i < 3 {\n\t\t\t\ttime.Sleep(10 * time.Second)\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\n\t\tbreak\n\t}\n\n\tif dbErr != nil {\n\t\treturn nil, errors.New(\"can't connect to database after 3 attempts\")\n\t}\n\n\treturn db, nil\n}", "func Connect(ctx context.Context, connString string) (*Conn, error) {\n\tconnConfig, err := ParseConfig(connString)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn connect(ctx, connConfig)\n}", "func Connect() (*sql.DB, error) {\n\tpsqlInfo := fmt.Sprintf(\"host=%s port=%d user=%s password=%s dbname=%s sslmode=disable\", host, port, user, password, dbname)\n\tdb, err := sql.Open(\"postgres\", psqlInfo)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// defer db.Close()\n\n\terr = db.Ping()\n\tif err != nil {\n\t\treturn nil, err\n\n\t}\n\tfmt.Println(\"Connection to database was successful\")\n\treturn db, err\n}", "func Connect() (db *pg.DB) {\n\tdb = pg.Connect(&pg.Options{\n\t\tUser: info.User,\n\t\tDatabase: info.Database,\n\t})\n\n\treturn db\n}", "func (c *PostgresClient) Connect() error {\n\tvar err error\n\turl := fmt.Sprintf(dbConnection, c.DBName, c.User, c.Pass, c.Port, c.Host)\n\tlogURL := fmt.Sprintf(dbConnectionLog, c.DBName, c.User, c.Port, c.Host)\n\tlog.Infof(logURL)\n\tc.DB, err = gorm.Open(\"postgres\", url)\n\tif c.RetryOnFailure {\n\t\tfor err != nil && c.connectionRetries < c.TimeOut {\n\t\t\tc.connectionRetries++\n\t\t\tlog.Warningf(\"unable to connect to %s, error: %s\", logURL, err.Error())\n\t\t\ttime.Sleep(500 * time.Millisecond)\n\t\t\tc.DB, err = gorm.Open(\"postgres\", url)\n\t\t}\n\t}\n\treturn err\n}", "func (d *Database) Connect() (*sqlx.DB, error) {\n\tconnectParams := fmt.Sprintf(\"user=%s host=localhost port=%s dbname=%s sslmode=%s password=%s\", d.Cf.GetDBUser(), d.Cf.DbPort, d.Cf.GetDBName(), d.Cf.GetDBSSL(), d.Cf.DbPassword)\n\tdb, err := sqlx.Connect(\"postgres\", connectParams)\n\tutil.Handle(\"Error creating a DB connection: \", err)\n\treturn db, err\n}", "func Connect(ctx context.Context, drCSIAddress string, metricsManager metrics.CSIMetricsManager) (conn *grpc.ClientConn, err error) {\n\tvar m sync.Mutex\n\tvar canceled bool\n\tready := make(chan bool)\n\tgo func() {\n\t\tconn, err = connection.Connect(drCSIAddress, metricsManager)\n\n\t\tm.Lock()\n\t\tdefer m.Unlock()\n\t\tif err != nil && canceled {\n\t\t\t_ = conn.Close()\n\t\t}\n\n\t\tclose(ready)\n\t}()\n\n\tselect {\n\tcase <-ctx.Done():\n\t\tm.Lock()\n\t\tdefer m.Unlock()\n\t\tcanceled = true\n\t\treturn nil, ctx.Err()\n\n\tcase <-ready:\n\t\treturn conn, err\n\t}\n}", "func Connect(configServerDB ServerDB) (*DAL, error) {\n\tif configServerDB.DSN == \"\" {\n\t\tconfigServerDB.DSN = createDSN(configServerDB)\n\t}\n\tcon, err := sql.Open(configServerDB.Engine, configServerDB.DSN)\n\tengine := &DAL{\n\t\tdb: con,\n\t}\n\tengine.Connected = isConnected(engine.db)\n\treturn engine, err\n}", "func Connect(host, port, password string, db int) *redis.Client {\n\tredisClient := redis.NewClient(&redis.Options{\n\t\tAddr: host + \":\" + port,\n\t\tPassword: password,\n\t\tDB: db,\n\t})\n\n\treturn redisClient\n}", "func ConnectDb(DSN string, timeout time.Duration) (*sqlx.DB, error) {\n\tticker := time.NewTicker(1 * time.Second)\n\tdefer ticker.Stop()\n\n\ttimeoutExceeded := time.After(timeout)\n\tfor {\n\t\tselect {\n\t\tcase <-timeoutExceeded:\n\t\t\treturn nil, fmt.Errorf(\"db connection failed after %s timeout\", timeout)\n\n\t\tcase <-ticker.C:\n\t\t\tdb, err := sqlx.Connect(\"postgres\", DSN)\n\t\t\tif err == nil {\n\t\t\t\tdb.SetMaxIdleConns(maxIdleConns)\n\t\t\t\tdb.SetMaxOpenConns(maxOpenConns)\n\t\t\t\treturn db, nil\n\t\t\t}\n\t\t\tlog.Errorf(\"failed to connect to db %s %s\", DSN, err)\n\t\t}\n\t}\n}", "func Connect(info *mgo.DialInfo) (DB, error) {\n\tsession, err := mgo.DialWithInfo(info)\n\n\treturn Store{session: session}, err\n}", "func Connect(dbURL string) *DB {\n\tconn, err := pgx.Connect(context.Background(), dbURL)\n\tif err != nil {\n\t\tlog.Fatalf(\"unable to connect to db err: %v\", err)\n\t\tos.Exit(1)\n\t}\n\treturn &DB{\n\t\tconn: conn,\n\t}\n}", "func ConnectMock(t *testing.T) (*DB, sqlmock.Sqlmock, error) {\n\treturn ConnectMockContext(context.Background(), t)\n}", "func (m *CDatabase) Connect() {\n\tsession, err := mgo.Dial(m.Server)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdb = session.DB(m.Database)\n}", "func (database *Database) Connect(environment string) *gorm.DB {\n\tdatabase.setConfigs(environment)\n\n\tdb, err := gorm.Open(\"mysql\", fmt.Sprintf(\"%s:%s@tcp(%s:%s)/%s?charset=utf8&parseTime=True\",\n\t\tdatabase.username, database.password, database.host, database.port, database.name))\n\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\terr = db.DB().Ping()\n\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tdb = database.SetLogger(db)\n\tdb = database.SetConnection(db, 0, 100)\n\n\treturn db\n}", "func (ms *SessionManager) Connect() (*mongo.Database, error) {\n\tctx, cancel := context.WithTimeout(context.Background(), 20*time.Second)\n\tdefer cancel()\n\tclient, err := mongo.Connect(ctx, options.Client().ApplyURI(\"mongodb://\"+ms.Host+\":\"+ms.Port))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdb := client.Database(ms.Database)\n\treturn db, nil\n}", "func OpenConnection(ctx context.Context, logContext, dsn string, maxConns, maxIdleConns int, maxConnLifetime time.Duration) (*sql.DB, error) {\n\tvar (\n\t\turl *dburl.URL\n\t\tconn *sql.DB\n\t\terr error\n\t\tch = make(chan error)\n\t)\n\n\turl, err = safeParse(dsn)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdriver := url.Driver\n\tif url.GoDriver != \"\" {\n\t\tdriver = url.GoDriver\n\t}\n\n\t// Open the DB handle in a separate goroutine so we can terminate early if the context closes.\n\tgo func() {\n\t\tconn, err = sql.Open(driver, url.DSN)\n\t\tclose(ch)\n\t}()\n\n\tselect {\n\tcase <-ctx.Done():\n\t\treturn nil, ctx.Err()\n\tcase <-ch:\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tconn.SetMaxIdleConns(maxIdleConns)\n\tconn.SetMaxOpenConns(maxConns)\n\tconn.SetConnMaxLifetime(maxConnLifetime)\n\n\tif klog.V(1).Enabled() {\n\t\tif len(logContext) > 0 {\n\t\t\tlogContext = fmt.Sprintf(\"[%s] \", logContext)\n\t\t}\n\t\tklog.Infof(\"%sDatabase handle successfully opened with '%s' driver\", logContext, driver)\n\t}\n\treturn conn, nil\n}", "func (m *DAOUtil) Connect() *mgo.Database {\n\tsession, err := mgo.Dial(m.Server)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\treturn session.DB(m.Database)\n}", "func Connect() *sql.DB {\n\tdbinfo := fmt.Sprintf(\"user=%s password=%s dbname=%s sslmode=disable\", DbUser, DbPassword, DbName)\n\n\tdb, _ := sql.Open(\"postgres\", dbinfo)\n\n\treturn db\n}", "func (d DbConfig) Connect() (*gorm.DB, error) {\n\t/*\n\t * We will build the connection string\n\t * Then will connect to the database\n\t */\n\tcStr := fmt.Sprintf(\"host=%s port=%s dbname=%s user=%s password=%s sslmode=disable\",\n\t\td.Host, d.Port, d.Database, d.Username, d.Password)\n\n\treturn gorm.Open(\"postgres\", cStr)\n}", "func Connect() (*sqlx.DB, error) {\n\tvar err error\n\tConnectOnce.Do(func() {\n\t\tdb, err = sqlx.Open(\"mysql\", GetEnv()+\"?parseTime=true\")\n\t})\n\treturn db, err\n}", "func Connect() redis.Conn {\n\tvar err error\n\tconnect, err = redis.Dial(\"tcp\", getDial())\n\tif err != nil {\n\t\treturn nil\n\t}\n\treturn connect\n}", "func (s *ProxyServer) Connect(ctx context.Context, proxyCtx *common.ProxyContext) (net.Conn, error) {\n\t// There may be multiple database servers proxying the same database. If\n\t// we get a connection problem error trying to dial one of them, likely\n\t// the database server is down so try the next one.\n\tfor _, server := range getShuffleFunc()(proxyCtx.Servers) {\n\t\ts.log.Debugf(\"Dialing to %v.\", server)\n\t\ttlsConfig, err := s.getConfigForServer(ctx, proxyCtx.Identity, server)\n\t\tif err != nil {\n\t\t\treturn nil, trace.Wrap(err)\n\t\t}\n\t\tserviceConn, err := proxyCtx.Cluster.Dial(reversetunnel.DialParams{\n\t\t\tFrom: &utils.NetAddr{AddrNetwork: \"tcp\", Addr: \"@db-proxy\"},\n\t\t\tTo: &utils.NetAddr{AddrNetwork: \"tcp\", Addr: reversetunnel.LocalNode},\n\t\t\tServerID: fmt.Sprintf(\"%v.%v\", server.GetHostID(), proxyCtx.Cluster.GetName()),\n\t\t\tConnType: types.DatabaseTunnel,\n\t\t\tProxyIDs: server.GetProxyIDs(),\n\t\t})\n\t\tif err != nil {\n\t\t\t// If an agent is down, we'll retry on the next one (if available).\n\t\t\tif isReverseTunnelDownError(err) {\n\t\t\t\ts.log.WithError(err).Warnf(\"Failed to dial database %v.\", server)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\treturn nil, trace.Wrap(err)\n\t\t}\n\t\t// Upgrade the connection so the client identity can be passed to the\n\t\t// remote server during TLS handshake. On the remote side, the connection\n\t\t// received from the reverse tunnel will be handled by tls.Server.\n\t\tserviceConn = tls.Client(serviceConn, tlsConfig)\n\t\treturn serviceConn, nil\n\t}\n\treturn nil, trace.BadParameter(\"failed to connect to any of the database servers\")\n}", "func Connect(cfg *Config) *redis.Client {\n\treturn redis.NewClient(&redis.Options{\n\t\tAddr: fmt.Sprintf(\"%s:%s\", cfg.Host, cfg.Port),\n\t\tPassword: cfg.Password,\n\t\tDB: cfg.DB,\n\t\tPoolSize: cfg.PoolSize,\n\t\tDialTimeout: cfg.DialTimeout,\n\t\tReadTimeout: cfg.ReadWriteTimeout,\n\t\tWriteTimeout: cfg.ReadWriteTimeout,\n\t\tIdleTimeout: cfg.IdleTimeout,\n\t\tIdleCheckFrequency: cfg.IdleCheckFrequency,\n\t\tMaxConnAge: cfg.MaxConnAge,\n\t})\n}", "func Connect(driver, db string) *sql.DB {\n\tDB, err := sql.Open(driver, db)\n\tcheckErr(err)\n\treturn DB\n}", "func connect() (*mongo.Client, error) {\n\tvar err error\n\tclient, err := mongo.NewClient(options.Client().ApplyURI(config.DbParams.Host))\n\tif err != nil {\n\t\tlog.Println(\"Err db.connect->mongo.NewClient\", err)\n\t\treturn nil, err\n\t}\n\n\tctx, cancel := context.WithTimeout(context.Background(), config.DbParams.TimeOut*time.Second)\n\tdefer cancel()\n\terr = client.Connect(ctx)\n\tif err != nil {\n\t\tlog.Println(\"Err db.connect->context.WithTimeout\", err)\n\t\treturn nil, err\n\t}\n\n\treturn client, nil\n}", "func (p *ProviderDAO) Connect() {\n\tsession, err := mgo.Dial(p.Server)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdb = session.DB(p.Database)\n}", "func Connect(uri string) *mgo.Session {\n\tif uri == \"\" {\n\t\tlog.Fatalln(\"Error empty Connection String to mongo Db servers\")\n\t}\n\tdb, err := mgo.Dial(uri)\n\tif err != nil {\n\t\tlog.Fatalln(\"cannot dial mongo:\", uri, \"\\n with error:\", err)\n\t}\n\n\treturn db\n}", "func Connect(user string, password string, host string, port int, schema string, dsn string) (*sql.DB, error) {\n\tvar err error\n\tvar connString bytes.Buffer\n\n\tpara := map[string]interface{}{}\n\tpara[\"User\"] = user\n\tpara[\"Pass\"] = password\n\tpara[\"Host\"] = host\n\tpara[\"Port\"] = port\n\tpara[\"Schema\"] = schema\n\n\ttmpl, err := template.New(\"dbconn\").Option(\"missingkey=zero\").Parse(dsn)\n\tif err != nil {\n\t\tlog.Error().Err(err).Msg(\"tmpl parse\")\n\t\treturn nil, err\n\t}\n\n\terr = tmpl.Execute(&connString, para)\n\tif err != nil {\n\t\tlog.Error().Err(err).Msg(\"tmpl execute\")\n\t\treturn nil, err\n\t}\n\n\tlog.Debug().Str(\"dsn\", connString.String()).Msg(\"connect to db\")\n\tdb, err := sql.Open(\"mysql\", connString.String())\n\tif err != nil {\n\t\tlog.Error().Err(err).Msg(\"mysql connect\")\n\t\treturn nil, err\n\t}\n\n\treturn db, nil\n}", "func (cfg Config) Connect(ctx context.Context) (*mongo.Database, error) {\n\turi := cfg.makeURI()\n\n\tclient, err := mongo.NewClient(options.Client().ApplyURI(uri))\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"couldn't connect to mongo: %v\", err)\n\t}\n\terr = client.Connect(ctx)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"mongo client couldn't connect with background context: %v\", err)\n\t}\n\n\tstarWarsDB := client.Database(cfg.DBName)\n\n\treturn starWarsDB, nil\n}", "func Connect() *sql.DB {\n\tfmtStr := \"host=%s port=%s user=%s \" +\n\t\t\"password=%s dbname=%s sslmode=disable\"\n\tpsqlInfo := fmt.Sprintf(\n\t\tfmtStr,\n\t\tos.Getenv(\"PG_HOST\"),\n\t\tos.Getenv(\"PG_PORT\"),\n\t\tos.Getenv(\"PG_USER\"),\n\t\tos.Getenv(\"PG_PASSWORD\"),\n\t\tos.Getenv(\"PG_DBNAME\"),\n\t)\n\tdb, err := sql.Open(\"postgres\", psqlInfo)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\treturn db\n}", "func DialContext(ctx context.Context, endpoint string) (Conn, error) {\n\tvar d net.Dialer\n\tc, err := d.DialContext(ctx, \"tcp\", endpoint)\n\treturn &conn{c}, err\n}", "func Connect(dbCfg config.DBConfig) (*sql.DB, error) {\n\t// Assemble database connection string\n\tsqlConnStr := fmt.Sprintf(\"host=%s port=%d dbname=%s user=%s \"+\n\t\t\"sslmode=disable\", dbCfg.DBHost, dbCfg.DBPort,\n\t\tdbCfg.DBName, dbCfg.DBUsername)\n\n\tif len(dbCfg.DBPassword) > 0 {\n\t\tsqlConnStr += fmt.Sprintf(\" password=%s\", dbCfg.DBPassword)\n\t}\n\n\t// Connect to database\n\tdb, err := sql.Open(\"postgres\", sqlConnStr)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error opening connection to database: %s\",\n\t\t\terr.Error())\n\t}\n\n\t// Check connection\n\terr = db.Ping()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error checking connection: %s\", err.Error())\n\t}\n\n\treturn db, nil\n}", "func (c *Node) Connect(ctx context.Context, out io.Writer, dbName string) error {\n\tif c.pid < 1 {\n\t\treturn fmt.Errorf(\"can't connect to unstarted cockroach\")\n\t}\n\n\tconn, err := pgx.Connect(ctx, c.ConnectionString(dbName))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tc.conn = conn\n\treturn nil\n}", "func Connect() *gorm.DB {\n\tdsn := fmt.Sprintf(\n\t\t\"host=%s user=%s password=%s dbname=%s port=%s sslmode=disable TimeZone=America/Bogota\",\n\t\tconfig.Load(\"DB_HOST\"), config.Load(\"DB_USER\"), config.Load(\"DB_PWD\"), config.Load(\"DB_NAME\"),\n\t\tconfig.Load(\"DB_PORT\"),\n\t)\n\n\tdb, err := gorm.Open(postgres.Open(dsn), &gorm.Config{\n\t\tQueryFields: true,\n\t})\n\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\treturn db\n}", "func Connect(\n\tctx context.Context,\n\tconfig *Config,\n) (Handle, error) {\n\tlogger.Infof(\"connecting remote Bitcoin chain\")\n\n\tconnCfg := &rpcclient.ConnConfig{\n\t\tUser: config.Username,\n\t\tPass: config.Password,\n\t\tHost: config.URL,\n\t\tHTTPPostMode: true, // Bitcoin core only supports HTTP POST mode\n\t\tDisableTLS: true, // Bitcoin core does not provide TLS by default\n\t}\n\n\tclient, err := rpcclient.New(connCfg, nil)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\n\t\t\t\"failed to create rpc client at [%s]: [%v]\",\n\t\t\tconfig.URL,\n\t\t\terr,\n\t\t)\n\t}\n\n\terr = testConnection(client, connectionTimeout)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\n\t\t\t\"error while connecting to [%s]: [%v]; check if the Bitcoin node \"+\n\t\t\t\t\"is running and you provided correct credentials and url\",\n\t\t\tconfig.URL,\n\t\t\terr,\n\t\t)\n\t}\n\n\t// When the context is done, cancel all requests from the RPC client\n\t// and disconnect it.\n\tgo func() {\n\t\t<-ctx.Done()\n\t\tlogger.Info(\"disconnecting from remote Bitcoin chain\")\n\t\tclient.Shutdown()\n\t}()\n\n\treturn &remoteChain{client: client}, nil\n}", "func Connect() (*DB, error) {\n\tpsqlInfo := fmt.Sprintf(\"host=db port=5432 user=ggp-user password=password dbname=development_db sslmode=disable\")\n\tdb, err := gorm.Open(\"postgres\", psqlInfo)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = db.DB().Ping()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &DB{DB: db}, nil\n}", "func Connect(user, pwd, host, database string) (*Conn, error) {\n\treturn connectWithCredentials(&credentials{user: user, pwd: pwd, host: host, database: database})\n}", "func Connect() (redis.Conn, error) {\n\tconn, err := sockfile.Dial(\"redisd\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn redis.NewConn(conn, Timeout, Timeout), nil\n}", "func (p PostgresProvider) Connect(config *gormx.DatabaseConfig) (*gorm.DB, error) {\n\tif config.Dialect == gormx.DriverPostgres {\n\t\tif db, err := gorm.Open(pg.New(pg.Config{DSN: config.DSN}), &gorm.Config{\n\t\t\tLogger: gormx.DefaultLogger(&config.Logger),\n\t\t}); err == nil {\n\t\t\tif sqlDB, err := db.DB(); err == nil {\n\t\t\t\tif config.MaxIdle > 0 {\n\t\t\t\t\tsqlDB.SetMaxIdleConns(config.MaxIdle)\n\t\t\t\t}\n\t\t\t\tif config.MaxOpen > 0 && config.MaxOpen > config.MaxIdle {\n\t\t\t\t\tsqlDB.SetMaxOpenConns(100)\n\t\t\t\t}\n\t\t\t\tif config.MaxLifetime > 0 {\n\t\t\t\t\tsqlDB.SetConnMaxLifetime(time.Duration(config.MaxLifetime) * time.Second)\n\t\t\t\t}\n\t\t\t\treturn db, nil\n\t\t\t} else {\n\t\t\t\treturn nil, errors.New(\"open DB failed\")\n\t\t\t}\n\t\t} else {\n\t\t\tlog.Errorf(\"connect db failed: error=%s\", err.Error())\n\t\t}\n\t\treturn nil, errors.New(\"connect db failed\")\n\t}\n\treturn nil, errors.New(\"driver is not postgres\")\n}", "func (d *DBGenerator) connect(ctx context.Context) error {\n\tif d.Conn == nil {\n\t\tconnStr := fmt.Sprintf(\"vertica://%s:%s@%s:%d/%s?tlsmode=%s\",\n\t\t\td.Opts.User, d.Opts.Password, d.Opts.Host, d.Opts.Port, d.Opts.DBName, d.Opts.TLSMode,\n\t\t)\n\t\tconn, err := sql.Open(\"vertica\", connStr)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\td.Conn = conn\n\t}\n\n\treturn d.Conn.PingContext(ctx)\n}", "func Connect() *sql.DB {\n\tURL := fmt.Sprintf(\"user=%s password=%s dbname=%s sslmode=%s\", configs.Database.User, configs.Database.Pass, configs.Database.Name, \"disable\")\n\tdb, err := sql.Open(\"postgres\", URL)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t\treturn nil\n\t}\n\treturn db\n}", "func Connect() error {\n\tcfg := Config{}\n\tif err := env.Parse(&cfg); err != nil {\n\t\treturn fmt.Errorf(\"%+v\", err)\n\t}\n\tdsn := cfg.DbUser + \":\" + cfg.DbPassword + \"@\" + cfg.DbHost + \"/\" + cfg.\n\t\tDbName + \"?parseTime=true&charset=utf8\"\n\n\tdb, err := sql.Open(\"mysql\", dsn)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar dbErr error\n\tfor i := 1; i <= 8; i++ {\n\t\tdbErr = db.Ping()\n\t\tif dbErr != nil {\n\t\t\tif i < 8 {\n\t\t\t\tlog.Printf(\"db connection failed, %d retry : %v\", i, dbErr)\n\t\t\t\ttime.Sleep(10 * time.Second)\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\n\t\tbreak\n\t}\n\n\tif dbErr != nil {\n\t\treturn errors.New(\"can't connect to database after 3 attempts\")\n\t}\n\n\tDbConn = db\n\n\treturn nil\n}", "func (s *orm) Connect(ctx context.Context, config skelego.Config) {\n\t//databaseType := config.Get(\"storage.engine\").(string)\n\tlogger := skelego.Logger()\n\thost := config.Get(\"storage.host\")\n\tport := config.Get(\"storage.port\")\n\tdatabase := config.Get(\"storage.name\")\n\tusername := config.Get(\"storage.username\")\n\tpassword := config.Get(\"storage.password\")\n\n\tlogger.LogEvent(\"Connecting to %s\", s.dbURI(host, port, username, database, password))\n\n\tdb, err := gorm.Open(\"postgres\", s.dbURI(host, port, username, database, password))\n\tif err != nil {\n\t\tprint(err.Error())\n\t}\n\tlogger.LogEvent(\"Successfully connected!\", db)\n\ts.client = db\n\tif err := db.DB().Ping(); err != nil {\n\t\tlogger.LogFatal(\"Error in database connection; restart app; error: %s\", err.Error())\n\t}\n}", "func Connect(c *config.Config) (*sql.DB, error) {\n\tstr := fmt.Sprintf(\"postgres://%s:%s@%s:%s/%s?sslmode=disable\", c.DatabaseUser, c.DatabasePassword, c.DatabaseHost, c.DatabasePort, c.DatabaseName)\n\tdb, err := sql.Open(\"postgres\", str)\n\n\tif err != nil {\n\t\tlog.Print(err)\n\t}\n\n\treturn db, err\n}", "func Connect(connectionUrl string) {\n\tinfo := &mgo.DialInfo{\n\t\tAddrs: []string{connectionUrl},\n\t\tTimeout: 5 * time.Second,\n\t\tDatabase: DB,\n\t\tUsername: \"\",\n\t\tPassword: \"\",\n\t}\n\tsession, err := mgo.DialWithInfo(info)\n\tif err != nil {\n\t\tfmt.Println(err.Error())\n\t}\n\tdb = session.DB(DB)\n}", "func (c *Ctx) WithTimeout(timeout time.Duration) (cf context.CancelFunc) {\n\tc.netContext, cf = context.WithTimeout(c.netContext, timeout)\n\treturn\n}", "func (node *PostgresNode) Connect(dbname string) *sql.DB {\n\tconninfo := fmt.Sprintf(\"postgres://%s@%s:%d/%s?sslmode=disable\",\n\t\tnode.user, node.host, node.Port, dbname)\n\n\tdb, err := sql.Open(\"postgres\", conninfo)\n\tif err != nil {\n\t\tlog.Panic(\"Can't connect to database: \", err)\n\t}\n\n\tnode.connections = append(node.connections, db)\n\treturn db\n}", "func Connect(migrate bool) (*gorm.DB, *redis.Client, error) {\n\tvar err error\n\tdb, err := connectPostgres(migrate)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\trdb, err := connectRedis()\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treturn db, rdb, nil\n}", "func connect(ctx context.Context, config string) (*sql.DB, error) {\n\tdb, err := sql.Open(\"postgres\", config)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"connect open: %v\", err)\n\t}\n\tif _, err := db.ExecContext(ctx, `SELECT 1`); err != nil {\n\t\treturn nil, fmt.Errorf(\"connect select: %v\", err)\n\t}\n\treturn db, nil\n}", "func (w *Wrapper) Connect() (err error) {\n\terr = w.db.Connect()\n\treturn\n}", "func NewCustomDialContext(dialTimeout time.Duration) DialContext {\n\treturn func(ctx context.Context, network, addr string) (net.Conn, error) {\n\t\tdialer := &net.Dialer{\n\t\t\tTimeout: dialTimeout,\n\t\t\tControl: func(network, address string, c syscall.RawConn) error {\n\t\t\t\treturn c.Control(func(fdPtr uintptr) {\n\t\t\t\t\t// got socket file descriptor to set parameters.\n\t\t\t\t\tfd := int(fdPtr)\n\n\t\t\t\t\t// Enable TCP fast connect\n\t\t\t\t\t// TCPFastOpenConnect sets the underlying socket to use\n\t\t\t\t\t// the TCP fast open connect. This feature is supported\n\t\t\t\t\t// since Linux 4.11.\n\t\t\t\t\t_ = syscall.SetsockoptInt(fd, syscall.IPPROTO_TCP, unix.TCP_FASTOPEN_CONNECT, 1)\n\t\t\t\t})\n\t\t\t},\n\t\t}\n\t\treturn dialer.DialContext(ctx, network, addr)\n\t}\n}", "func Connect() (err error) {\n\tdefer logging.Logger.Sync()\n\tswitch config.Database.Driver {\n\tcase \"mysql\":\n\t\terr = mysqlConnect()\n\tdefault:\n\t\tlogging.Logger.Fatal(\"[database.connect] unsupported database driver.\")\n\t}\n\treturn err\n}", "func (this *DatabaseConnection) Connect(connectionName string) error {\n\tvar db *sql.DB\n\tvar err error\n\n\tswitch this.Engine {\n\tcase SQLITE:\n\t\tdb, err = ConnectSQLite(this)\n\n\tcase MYSQL:\n\t\tdb, err = ConnectMySQL(this)\n\n\tcase MSSQL:\n\t\tdb, err = ConnectMSSQL(this)\n\n\tcase TESTDB:\n\t\tdb, err = ConnectTestDB(this)\n\t}\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tDb[connectionName] = db\n\treturn nil\n}", "func connect(DSN string, timeout time.Duration) *sql.DB {\n\tticker := time.NewTicker(1 * time.Second)\n\ttimer := time.After(timeout)\n\tconn, _ := sql.Open(\"mysql\", DSN)\n\n\tfor {\n\t\tselect {\n\t\tcase <-ticker.C:\n\t\t\terr := conn.Ping()\n\t\t\tif err == nil {\n\t\t\t\treturn conn\n\t\t\t}\n\t\tcase <-timer:\n\t\t\tlog.Fatalf(\"Timeout trying To connect To %s after %d seconds. Forgot To run `make db_up`?\", DSN, timeout/1e9)\n\t\t\tos.Exit(1)\n\t\t}\n\t}\n}", "func Connect() (*mongo.Database, context.Context, context.CancelFunc) {\n\tdbUser := \"nahid_Anexa\"\n\tdbPass := \"nahidForAnexa\"\n\tdbName := \"anexa_test\"\n\n\t//this is mongoDB atlas connection string\n\tconnectionString := \"mongodb+srv://\" + dbUser + \":\" + dbPass + \"@anexacluster.73mtc.mongodb.net/\" + dbName + \"?retryWrites=true&w=majority\"\n\n\tdbClient, err := mongo.NewClient(options.Client().ApplyURI(connectionString))\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\n\tctx, cancel := context.WithTimeout(context.Background(), 10*time.Second)\n\terr = dbClient.Connect(ctx)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\n\terr = dbClient.Ping(ctx, readpref.Primary())\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\n\t//return db\n\treturn dbClient.Database(dbName), ctx, cancel\n}", "func (c Conn) WithContext(cx context.Context) *Conn {\n\treturn &Conn{\n\t\tlocal: c.local,\n\t\tremote: c.remote,\n\t\tConn: connCtxOverride{c: cx, Conn: c.Conn},\n\t}\n}", "func Connect() (*pg.DB, error) {\n\tvar (\n\t\tconn *pg.DB\n\t\tn int\n\t)\n\taddr := fmt.Sprintf(\"%s:%v\", \"192.168.0.103\", 6001)\n\tconn = pg.Connect(&pg.Options{\n\t\tAddr: addr,\n\t\tUser: User,\n\t\tPassword: Password,\n\t\tDatabase: DatabaseName,\n\t})\n\t_, err := conn.QueryOne(pg.Scan(&n), \"SELECT 1\")\n\tif err != nil {\n\t\treturn conn, fmt.Errorf(\"Error conecting to DB. addr: %s, user: %s, db: %s,%v\", addr, User, DatabaseName, err)\n\t}\n\n\tif err := createSchema(conn); err != nil {\n\t\treturn conn, fmt.Errorf(\"Error creating DB schemas. %v\", err)\n\t}\n\treturn conn, nil\n}", "func ConnectDB() (*DB, error) {\n\t// ?parseTime=true\n\t// https://stackoverflow.com/questions/45040319/unsupported-scan-storing-driver-value-type-uint8-into-type-time-time\n\tdb, err := sql.Open(\"mysql\", \"api:keepgrowth$@/book_report?parseTime=true\")\n\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\t// https://github.com/go-sql-driver/mysql/#important-settings\n\tdb.SetConnMaxLifetime(time.Minute * 3)\n\tdb.SetMaxOpenConns(10)\n\tdb.SetConnMaxIdleTime(10)\n\n\terrPing := db.Ping()\n\tif errPing != nil {\n\t\tpanic(err.Error())\n\t}\n\n\tqb := goqu.New(\"mysql\", db)\n\n\treturn &DB{qb}, nil\n}", "func (rt *RethinkDB) Connect(connString string) error {\n\tvar err error\n\trt.session, err = r.Connect(r.ConnectOpts{\n\t\tAddress: connString,\n\t\tDiscoverHosts: true,\n\t})\n\tif err != nil {\n\t\tlog.Fatalln(err.Error())\n\t\treturn err\n\t}\n\tlog.Printf(\"Session connected?: %t\", rt.session.IsConnected())\n\treturn nil\n}", "func Connect() *sql.DB {\n\n\tvar connStr string\n\n\tif os.Getenv(\"mode\") == \"dev\" {\n\t\tconnStr = \"root\" + \"@tcp(\" + \"127.0.0.1:3306\" + \")/\" + \"analasia\"\n\t} else {\n\t\tconnStr = os.Getenv(\"DATABASE_USER\") + \":\" + os.Getenv(\"DATABASE_PASSWORD\") + \"@tcp(\" + os.Getenv(\"DATABASE_HOST\") + \")/\" + os.Getenv(\"DATABASE_NAME\")\n\t}\n\tconn, err := sql.Open(\"mysql\", connStr)\n\tif err != nil {\n\t\tpanic(err.Error())\n\t}\n\treturn conn\n\n}", "func Connect(pool ServerPool, spec db.Specifier) (db.DB, error) {\n\tpool.RLock()\n\n\tif db, ok := pool.Get(spec); ok {\n\t\t// The connection was found. Do we need to update it?\n\t\tif db.Spec().NeedsUpdate(spec) {\n\t\t\tpool.RUnlock()\n\t\t\tpool.Lock()\n\t\t\t// By now, someone else may have updated it already.\n\t\t\tif db.Spec().NeedsUpdate(spec) {\n\t\t\t\tif err := db.Update(spec); err != nil {\n\t\t\t\t\tpool.Unlock()\n\t\t\t\t\treturn nil, err\n\t\t\t\t}\n\t\t\t}\n\t\t\tpool.Unlock()\n\t\t\tpool.RLock()\n\t\t}\n\t\tpool.RUnlock()\n\t\treturn db, nil\n\t}\n\tpool.RUnlock()\n\treturn insertNewDB(pool, spec)\n}", "func Connect() (*db.DB, error) {\n\tvar (\n\t\tdriver = \"sqlite3\"\n\t\tconfig = \":memory:?_foreign_keys=1\"\n\t)\n\tif os.Getenv(\"DRONE_DATABASE_DRIVER\") != \"\" {\n\t\tdriver = os.Getenv(\"DRONE_DATABASE_DRIVER\")\n\t\tconfig = os.Getenv(\"DRONE_DATABASE_DATASOURCE\")\n\t}\n\treturn db.Connect(driver, config)\n}", "func (m *TicketDAO) Connect() *mgo.Database {\n\tfmt.Println(\" From from createTicket\")\n\tsession, err := mgo.Dial(m.Server)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdb = session.DB(m.Database)\n\treturn db\n}" ]
[ "0.6372821", "0.61047596", "0.61021847", "0.6073866", "0.60574955", "0.60322917", "0.5914133", "0.58926475", "0.5844441", "0.5768442", "0.5765604", "0.5659135", "0.56569", "0.5654965", "0.5651231", "0.5647198", "0.56029594", "0.5578701", "0.55661476", "0.5515703", "0.5497718", "0.54855376", "0.54830027", "0.5479636", "0.543464", "0.5430881", "0.5430719", "0.5422984", "0.54198277", "0.54134923", "0.54133373", "0.54027027", "0.53867936", "0.53799474", "0.5376529", "0.53689456", "0.5361867", "0.5356621", "0.53488433", "0.534002", "0.5337409", "0.5335979", "0.53259104", "0.5306464", "0.52981293", "0.52947336", "0.52914417", "0.5290573", "0.5274611", "0.5274216", "0.5269511", "0.52635413", "0.5262711", "0.5261639", "0.52443945", "0.52388924", "0.52337646", "0.52258533", "0.52154005", "0.5209619", "0.5209292", "0.5207873", "0.5205637", "0.5203735", "0.5199594", "0.51929593", "0.5190131", "0.5185394", "0.518253", "0.517661", "0.5169978", "0.516665", "0.51555693", "0.515466", "0.5154266", "0.5150857", "0.5148587", "0.5145645", "0.5137638", "0.5135267", "0.51348037", "0.5128526", "0.5114501", "0.5111321", "0.5110657", "0.5105459", "0.5104291", "0.5102436", "0.5101972", "0.50925297", "0.5087254", "0.5081182", "0.5068915", "0.5068189", "0.50638205", "0.50595766", "0.50573", "0.50453925", "0.5044911", "0.50419587" ]
0.58671236
8
Driver returns the underlying Driver of the Connector, mainly to maintain compatibility with the Driver method on sql.DB.
func (c *connector) Driver() driver.Driver { return &GatewayDriver{} }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c connector) Driver() driver.Driver { return c.drv }", "func (c connector) Driver() driver.Driver { return c.drv }", "func (c *Connector) Driver() driver.Driver {\n\treturn &Driver{}\n}", "func (c *Conn) Driver() driver.Driver {\n\treturn c.driver\n}", "func (c *nrConnector) Driver() driver.Driver {\n\treturn c.driver\n}", "func (oci8Connector *OCI8Connector) Driver() driver.Driver {\n\treturn OCI8Driver\n}", "func (db *DB) Driver() string {\n\treturn driver\n}", "func (db *DB) Driver() string {\n\treturn driver\n}", "func (db *DB) Driver() driver.Driver {\n\treturn db.Master().Driver()\n}", "func (c *SpannerConnector) Driver() driver.Driver {\n\treturn &SpannerDriver{}\n}", "func (s *DbRecorder) Driver() string {\n\treturn s.flavor\n}", "func (drv StorageDriver) Driver() string {\n\treturn string(drv)\n}", "func (db *DB) DriverName() string {\n return db.driverName\n}", "func (server *SingleInstance) Driver() (neo4j.Driver, error) {\n\treturn neo4j.NewDriver(server.boltURI, server.authToken, server.config)\n}", "func (factory DriverFactory) Driver() (*strata.Driver, error) {\n\toptions := factory.GetOptions().(*Options)\n\n\taccountName := os.Getenv(\"AZURE_ACCOUNT_NAME\")\n\taccountSecret := os.Getenv(\"AZURE_ACCOUNT_SECRET\")\n\tif accountName == \"\" || accountSecret == \"\" {\n\t\treturn nil, errors.New(\"Environment variables AZURE_ACCOUNT_NAME and AZURE_ACCOUNT_SECRET must be set\")\n\t}\n\n\tazureBloblStorage, err := azureblobstorage.NewAzureBlobStorage(\n\t\taccountName,\n\t\taccountSecret,\n\t\toptions.AzureBlobOptions.Container,\n\t\toptions.AzureBlobOptions.BlobPrefix)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treplica, err := lreplica.NewLocalReplica(\n\t\toptions.Replica.MaxBackgroundCopies,\n\t\tstrconv.Itoa(options.Replica.Port),\n\t\toptions.Replica.Username,\n\t\toptions.Replica.Password,\n\t)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tmanager, err := strata.NewSnapshotManager(replica, azureBloblStorage)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &strata.Driver{Manager: manager}, err\n}", "func (d *DriverWrapper) Open(name string) (driver.Conn, error) {\n\tconn, err := d.Driver.Open(name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tswitch connType := reflect.TypeOf(conn).String(); connType {\n\tcase \"*mysql.mysqlConn\":\n\t\treturn &ConnWrapper{Conn: conn, dsn: name, integration: &mysqlIntegration{}}, nil\n\n\tcase \"*pq.conn\":\n\t\treturn &ConnWrapper{Conn: conn, dsn: name, integration: &postgresqlIntegration{}}, nil\n\t}\n\treturn conn, nil\n}", "func (o FioSpecVolumeVolumeSourceCsiPtrOutput) Driver() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *FioSpecVolumeVolumeSourceCsi) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Driver\n\t}).(pulumi.StringPtrOutput)\n}", "func (d *portworx) GetDriver() (*v1.StorageCluster, error) {\n\t// TODO: Need to implement it for Daemonset deployment as well, right now its only for StorageCluster\n\tstcList, err := pxOperator.ListStorageClusters(d.namespace)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed get StorageCluster list from namespace [%s], Err: %v\", d.namespace, err)\n\t}\n\n\tstc, err := pxOperator.GetStorageCluster(stcList.Items[0].Name, stcList.Items[0].Namespace)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get StorageCluster [%s] from namespace [%s], Err: %v\", stcList.Items[0].Name, stcList.Items[0].Namespace, err.Error())\n\t}\n\n\treturn stc, nil\n}", "func (a *MockApp) Driver() fyne.Driver {\n\targs := a.Called()\n\treturn args.Get(0).(fyne.Driver)\n}", "func Wrap(d driver.Driver) driver.Driver {\n\treturn &DriverWrapper{d}\n}", "func (o IopingSpecVolumeVolumeSourceCsiPtrOutput) Driver() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *IopingSpecVolumeVolumeSourceCsi) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Driver\n\t}).(pulumi.StringPtrOutput)\n}", "func (v *Volume) Driver() string {\n\treturn v.config.Driver\n}", "func (o *CreateOptions) GetDriver() string {\n\tif o.Driver == nil {\n\t\tvar z string\n\t\treturn z\n\t}\n\treturn *o.Driver\n}", "func (r *DrvRegistry) Driver(name string) (driverapi.Driver, *driverapi.Capability) {\n\td, c := r.Networks.Driver(name)\n\n\tif c == (driverapi.Capability{}) {\n\t\treturn d, nil\n\t}\n\treturn d, &c\n}", "func (db *DB) DriverName() string {\n\treturn db.DB.DriverName()\n}", "func Open(name string, settings Settings) (Database, error) {\n\n\tdriver, ok := wrappers[name]\n\tif ok == false {\n\t\t// Using panic instead of returning error because attemping to use an\n\t\t// nonexistent adapter will never result in a successful connection,\n\t\t// therefore should be considered a developer's mistake and must be catched\n\t\t// at compilation time.\n\t\tpanic(fmt.Sprintf(\"Open: Unknown adapter %s. (see: https://upper.io/db#database-adapters)\", name))\n\t}\n\n\t// Creating a new connection everytime Open() is called.\n\tdriverType := reflect.ValueOf(driver).Elem().Type()\n\tnewAdapter := reflect.New(driverType).Interface().(Database)\n\n\t// Setting up the connection.\n\terr := newAdapter.Setup(settings)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn newAdapter, nil\n}", "func Driver() *usart.Driver {\n\tif driver == nil {\n\t\tsetupDriver()\n\t}\n\treturn driver\n}", "func NewDriver() Driver {\n\treturn &boltDriver{}\n}", "func (o FioSpecVolumeVolumeSourceFlexVolumePtrOutput) Driver() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *FioSpecVolumeVolumeSourceFlexVolume) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Driver\n\t}).(pulumi.StringPtrOutput)\n}", "func (r *Network) Driver() pulumi.StringOutput {\n\treturn (pulumi.StringOutput)(r.s.State[\"driver\"])\n}", "func GetDriver() (driver EncryptDriver) {\n\tvar found bool\n\n\tif driverSelect != \"\" {\n\t\tif driver, found = driverMap[driverSelect]; found {\n\t\t\treturn\n\t\t}\n\t} else {\n\t\tif len(driverMap) == 1 {\n\t\t\tfor driverSelect, driver = range driverMap {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}\n\tpanic(driver_name + \" driver: Nothing registered\")\n\n}", "func (o FioSpecVolumeVolumeSourceCsiOutput) Driver() pulumi.StringOutput {\n\treturn o.ApplyT(func(v FioSpecVolumeVolumeSourceCsi) string { return v.Driver }).(pulumi.StringOutput)\n}", "func (d *RedisDriver) GetNativeConnector() interface{} {\n\treturn d.connector\n}", "func (o IopingSpecVolumeVolumeSourceFlexVolumePtrOutput) Driver() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *IopingSpecVolumeVolumeSourceFlexVolume) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Driver\n\t}).(pulumi.StringPtrOutput)\n}", "func GetDriver(name string) Driver {\n\treturn drivers[name]\n}", "func (o IopingSpecVolumeVolumeSourceCsiOutput) Driver() pulumi.StringOutput {\n\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceCsi) string { return v.Driver }).(pulumi.StringOutput)\n}", "func DB() *Connection {\n\treturn defaultConnection\n}", "func (d Display) GetDriver() string {\n\treturn C.GoString(C.caca_get_display_driver(d.Dp))\n}", "func GetDriver(name string) (Driver, error) {\n\tif val, ok := drivers[name]; ok {\n\t\treturn val, nil\n\t}\n\n\treturn nil, fmt.Errorf(\"unsupported driver: %s\", name)\n}", "func (d *hdbDriver) Name() string { return DriverName }", "func GetDriver(name string) Driver {\n\tdriversMu.Lock()\n\tdefer driversMu.Unlock()\n\tdriver := drivers[name]\n\treturn driver\n}", "func (c CredentialDatabase) GetAdapter() string {\n\treturn c.Adapter\n}", "func Driver(ctx context.Context, reference string) (driver.Driver, error) {\n\td, ok := driver.Get(reference)\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"quota not support for %s\", reference)\n\t}\n\n\treturn d, nil\n}", "func (d Driver) DB() *sql.DB {\n\treturn d.ExecQuerier.(*sql.DB)\n}", "func Connect(driver, db string) *sql.DB {\n\tDB, err := sql.Open(driver, db)\n\tcheckErr(err)\n\treturn DB\n}", "func (d *Driver) DriverName() string {\n\treturn DriverName\n}", "func GetDriverName(driver driver.Driver) string {\n\tif driver == nil {\n\t\treturn \"\"\n\t}\n\tdriverType := fmt.Sprintf(\"%s\", reflect.TypeOf(driver))\n\tswitch driverType {\n\tcase \"*mysql.MySQLDriver\":\n\t\treturn \"mysql\"\n\tcase \"*pq.Driver\":\n\t\treturn \"postgres\"\n\tdefault:\n\t\treturn \"\"\n\t}\n}", "func NewDriver(dialect string, c Conn) *Driver {\n\treturn &Driver{dialect: dialect, Conn: c}\n}", "func GetDialectByDriver(driver string) Dialect {\n\tswitch driver {\n\tcase \"mysql\":\n\t\treturn mysql{\n\t\t\tcommonDialect: commonDialect{delimiter: \"`\", delimiter2: \"`\"},\n\t\t}\n\tcase \"mssql\":\n\t\treturn mssql{\n\t\t\tcommonDialect: commonDialect{delimiter: \"[\", delimiter2: \"]\"},\n\t\t}\n\tcase \"postgresql\":\n\t\treturn postgresql{\n\t\t\tcommonDialect: commonDialect{delimiter: `\"`, delimiter2: `\"`},\n\t\t}\n\tcase \"sqlite\":\n\t\treturn sqlite{\n\t\t\tcommonDialect: commonDialect{delimiter: \"`\", delimiter2: \"`\"},\n\t\t}\n\tdefault:\n\t\treturn commonDialect{delimiter: \"`\", delimiter2: \"`\"}\n\t}\n}", "func GetDriverOpen(u *url.URL) (*sql.DB, error) {\n\tdrv, err := GetDriver(u.Scheme)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn drv.Open(u)\n}", "func DriverName() string {\n\tdriver, err := DriverNameByKey(DefaultConnectionName)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn driver\n}", "func (d *Driver) DriverName() string {\n\treturn driverName\n}", "func (d *Driver) Connection() gobot.Connection { return d.connection }", "func (db *Database) GetDriverTable() ([]Driver, error) {\n result := []Driver{}\n row, err := db.Query(\"SELECT * FROM Driver\")\n if err != nil {\n return result, err\n }\n defer row.Close()\n result = RowToDrivers(row)\n return result, nil\n}", "func (d *DB) Conn() *sql.DB { return d.conn }", "func (o FioSpecVolumeVolumeSourceFlexVolumeOutput) Driver() pulumi.StringOutput {\n\treturn o.ApplyT(func(v FioSpecVolumeVolumeSourceFlexVolume) string { return v.Driver }).(pulumi.StringOutput)\n}", "func ForDriver(driver string) reform.Dialect {\n\t// for sqlite3_with_sleep\n\tif strings.HasPrefix(driver, \"sqlite3\") {\n\t\treturn sqlite3.Dialect\n\t}\n\n\tswitch driver {\n\tcase \"postgres\", \"pgx\":\n\t\treturn postgresql.Dialect\n\tcase \"mysql\":\n\t\treturn mysql.Dialect\n\tcase \"mssql\":\n\t\treturn mssql.Dialect //nolint:staticcheck\n\tcase \"sqlserver\":\n\t\treturn sqlserver.Dialect\n\tdefault:\n\t\treturn nil\n\t}\n}", "func GetConnection() SQLConnector {\n\n\tconn, err := GetConnectionWithKey(DefaultConnectionName)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn conn\n}", "func (o IopingSpecVolumeVolumeSourceFlexVolumeOutput) Driver() pulumi.StringOutput {\n\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceFlexVolume) string { return v.Driver }).(pulumi.StringOutput)\n}", "func (f *Factor) driver() mfaDriver {\n\tif f.drv == nil {\n\t\tf.drv = newDriver(f)\n\t}\n\treturn f.drv\n}", "func (impl *Server) GetConnection() *gorm.DB {\n\treturn base.GetConnection()\n}", "func (db *mongoDataBase) GetConnection() interface{} {\n\treturn db.databaseConnection\n}", "func (t *driverTable) Get(name string) (Driver, error) {\n\tt.Lock()\n\tv, ok := t.drivers[name]\n\tif ok {\n\t\tt.Unlock()\n\t\treturn v, nil\n\t}\n\tt.Unlock()\n\n\tplugin, err := plugins.Get(volumePluginType, name)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"%s driver not found: %v\", name, err)\n\t}\n\n\tdriver := NewRemoteDriverWrapper(name, plugin)\n\n\tt.Lock()\n\tdefer t.Unlock()\n\n\tv, ok = t.drivers[name]\n\tif !ok {\n\t\tv = driver\n\t\tt.drivers[name] = v\n\t}\n\n\treturn v, nil\n}", "func (r *Redis) GetDriver(ctx context.Context, driverID string) (Driver, error) {\n\t// simulate RPC delay\n\tdriver, err := r.Get(driverID).Result()\n\tif err != nil {\n\t\tlog.WithField(\"driver_id\", driverID).WithError(err).Error(\"failed to get driver\")\n\t\treturn Driver{}, err\n\t}\n\n\treturn Driver{\n\t\tDriverID: driverID,\n\t\tLocation: driver,\n\t}, nil\n}", "func (extDb *Database) GetConnection() (*sql.DB, error) {\n\tif db := extDb.getDb(); db != nil {\n\t\treturn db, nil\n\t}\n\treturn nil, ErrNotInitialized\n}", "func Get(name string) (Driver, error) {\n\treturn backendDrivers.Get(name)\n}", "func GetConn() *sql.DB {\n\treturn dbConn()\n}", "func (p Driver) GetConnection() *sql.DB {\n\tdb, err := sql.Open(\"postgres\", p.Connection)\n\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tp.db = db\n\n\tfmt.Println(\"Opening postgresql connection.....\")\n\treturn db\n}", "func DBClient() driver.Client {\n\tclient := GetState(\"dbClient\")\n\treturn client.(driver.Client)\n}", "func GetDbConnection(driver string, connectionString string) *sqlx.DB {\n\tdb, err := sqlx.Connect(driver, connectionString)\n\tif err != nil {\n\t\tlog.Fatalf(\"Unable to establish connection: %v\\n\", err)\n\t}\n\treturn db\n}", "func (r *resolver) DB() driver.Database {\n\treturn r.db\n}", "func (d *driver) Name() string {\n\treturn driverName\n}", "func (d *driver) Name() string {\n\treturn driverName\n}", "func NewDriver(config dbmate.DriverConfig) dbmate.Driver {\n\treturn &Driver{\n\t\tmigrationsTableName: config.MigrationsTableName,\n\t\tdatabaseURL: config.DatabaseURL,\n\t\tlog: config.Log,\n\t}\n}", "func (d *driver) Name() string {\n\treturn driverName\n}", "func GetDBConnection(cfg *DBConfig) (conn ITransactionalDB, err error) {\n\tDSN := getDSN(cfg)\n\tdriver := cfg.Driver\n\n\tswitch driver {\n\tcase \"mysql\":\n\t\tconn, err = NewMySQLConn(DSN, cfg)\n\tcase \"postgres\":\n\t\tconn, err = NewPostgreSQLConn(\"postgres://\"+DSN, cfg)\n\tdefault:\n\t\terr = fmt.Errorf(\"Unsupported driver: %s\", driver)\n\t}\n\treturn\n}", "func (s *AdapterConfig) DB() base.DBInterface {\n\treturn adapterConfigDB\n}", "func (d *hdbDriver) Version() string { return DriverVersion }", "func Connect(settings *Settings) (Database, error) {\n\tconnector, ok := connectors[settings.Driver]\n\tif !ok {\n\t\treturn nil, errors.New(fmt.Sprintf(\"Connector not found for %s\", settings.Driver))\n\t}\n\n\t// Attempt to connect to the given db, if something goes wrong, we'll re-attempt a few times\n\t// before quitting. Mostly this helps in docker setups / test situations where a\n\t// container may not have spun up yet.\n\tattempts := 0\n\tfor {\n\t\tdb, err := connector(settings)\n\t\tif err != nil {\n\t\t\tattempts += 1\n\t\t\tif attempts >= maxAttempts {\n\t\t\t\treturn db, err\n\t\t\t}\n\t\t\ttime.Sleep(1 * time.Second)\n\t\t\tcontinue\n\t\t}\n\t\treturn db, err\n\t}\n}", "func GetConnection() *gorm.DB {\n\treturn db\n}", "func GetConnection() (database *sql.DB) {\n\tdatabaseDriver := \"mysql\"\n\tdatabaseUser := \"newuser\"\n\tdatabasePass := \"newuser\"\n\tdatabaseName := \"crm\"\n\tdatabase, err := sql.Open(databaseDriver, databaseUser+\":\"+databasePass+\"@/\"+databaseName)\n\tif err != nil {\n\t\tpanic(err.Error())\n\t}\n\treturn database\n}", "func (oci8Connector *OCI8Connector) Connect(ctx context.Context) (driver.Conn, error) {\n\toci8Conn := &OCI8Conn{\n\t\tlogger: oci8Connector.Logger,\n\t}\n\tif oci8Conn.logger == nil {\n\t\toci8Conn.logger = log.New(ioutil.Discard, \"\", 0)\n\t}\n\n\treturn oci8Conn, nil\n}", "func GetConnection() (database *sql.DB) {\n\tdatabaseDriver := \"mysql\"\n\tdatabaseUser := \"newuser\"\n\tdatabasePass := \"newuser\"\n\tdatabaseName := \"crm\"\n\tdatabase, error := sql.Open(databaseDriver, databaseUser+\":\"+databasePass+\"@/\"+databaseName)\n\t// stops the execution if there is any error in connecting\n\tif error != nil {\n\t\tpanic(error.Error())\n\t}\n\treturn database\n}", "func NewDBDriver(log *logrus.Logger) Driver {\n\treturn &DBDriver{log: log}\n}", "func (adapter *GORMAdapter) GetUnderlyingORM() interface{} {\n\treturn adapter.db\n}", "func (b *TestDriver) Connection() gobot.Connection { return b.connection }", "func DB() *sql.DB {\n\tdb, err := sql.Open(driver, connStr)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn db\n}", "func (c *Conn) GetConnection() *sql.DB {\n\treturn c.conn\n}", "func (impl *ServerServerGroup) GetConnection() *gorm.DB {\n\treturn base.GetConnection()\n}", "func (p *ProxySQL) Conn() *sql.DB {\n\treturn p.conn\n}", "func (impl *IPv4Pool) GetConnection() *gorm.DB {\n\treturn base.GetConnection()\n}", "func (c *DriversController) GetDriver() {\n\tdriver, err := models.GetDriver(c.Ctx.Input.Param(ID_PARAMETER))\n\tbuildResponse(err, c.Ctx, driver)\n}", "func (c Connector) Connect() (db *sql.DB) {\n\tif c.db != nil {\n\t\tlevel.Info(c.log).Log(\"msg\", \"returning mysql\")\n\t\treturn db\n\t}\n\tlevel.Info(c.log).Log(\"msg\", \"connecting to mysql\")\n\tlevel.Info(c.log).Log(\"msg\", \"returning connection\")\n\t_ = c.config.dbdriver\n\tdbUser := c.config.dbUser\n\tdbHost := c.config.dbHost\n\tdbPass := c.config.dbPassword\n\tdbName := c.config.dbDatabase\n\tdbPort := c.config.dbPort\n\n\tdb, err := sql.Open(\"mysql\", dbUser+\":\"+dbPass+\"@(\"+dbHost+\":\"+dbPort+\")/\"+dbName)\n\tif err != nil {\n\t\tpanic(err.Error())\n\t}\n\tc.db = db\n\tc.db.SetMaxIdleConns(10)\n\tc.db.SetMaxOpenConns(10)\n\tc.db.SetConnMaxLifetime(time.Duration(360))\n\treturn c.db\n\n}", "func (c connector) Close() error {\n\tif c.drv == nil || c.drv == defaultDrv {\n\t\treturn nil\n\t}\n\treturn c.drv.Close()\n}", "func WithDriver(driver string) EngineOpt {\n\treturn func(e *engine) error {\n\t\t// set the database type to interact with in the database engine\n\t\te.config.Driver = driver\n\n\t\treturn nil\n\t}\n}", "func EntDriver(name ...string) *entsql.Driver {\n\tif len(name) == 0 || name[0] == defaultConn {\n\t\tif defaultEntDriver == nil {\n\t\t\tlogger.Panic(fmt.Sprintf(\"yiigo: unknown db.%s (forgotten configure?)\", defaultConn))\n\t\t}\n\n\t\treturn defaultEntDriver\n\t}\n\n\tv, ok := entmap.Load(name[0])\n\n\tif !ok {\n\t\tlogger.Panic(fmt.Sprintf(\"yiigo: unknown db.%s (forgotten configure?)\", name[0]))\n\t}\n\n\treturn v.(*entsql.Driver)\n}", "func Client() (*gorm.DB, error) {\n\tmutex.Lock()\n\tdefer mutex.Unlock()\n\n\tif client != nil {\n\t\treturn client, nil\n\t}\n\n\tif err := connect(); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn client, nil\n}", "func (d *DefaultDriver) GetDriverVersion() (string, error) {\n\treturn \"\", &errors.ErrNotSupported{\n\t\tType: \"Function\",\n\t\tOperation: \"GetDriverVersion()\",\n\t}\n}", "func WrapDriver(driver driver.Driver, opts ...Opt) driver.Driver {\n\td := wrappedDriver{parent: driver}\n\n\tfor _, opt := range opts {\n\t\topt(&d.opts)\n\t}\n\n\tif d.Logger == nil {\n\t\td.Logger = nullLogger{}\n\t}\n\tif d.Tracer == nil {\n\t\td.Tracer = nullTracer{}\n\t}\n\n\treturn d\n}", "func WrapDriver(driver driver.Driver, opts ...Opt) driver.Driver {\n\td := wrappedDriver{parent: driver}\n\n\tfor _, opt := range opts {\n\t\topt(&d)\n\t}\n\n\tif d.Logger == nil {\n\t\td.Logger = nullLogger{}\n\t}\n\tif d.Tracer == nil {\n\t\td.Tracer = nullTracer{}\n\t}\n\n\treturn d\n}" ]
[ "0.8215049", "0.8215049", "0.7857132", "0.7591234", "0.7491992", "0.7387193", "0.73785657", "0.73785657", "0.70860475", "0.69730884", "0.6552133", "0.63778603", "0.6340669", "0.6295574", "0.61365694", "0.6051169", "0.60453224", "0.60428107", "0.60121566", "0.6011682", "0.597462", "0.59478855", "0.58996135", "0.58651245", "0.5837975", "0.5836982", "0.5815155", "0.5811048", "0.5742523", "0.57411975", "0.57358605", "0.5735762", "0.57300186", "0.57286805", "0.57185245", "0.56776285", "0.5668559", "0.56446254", "0.55726707", "0.5566929", "0.55551046", "0.55463356", "0.55379164", "0.5534458", "0.5521908", "0.5519408", "0.551901", "0.5510591", "0.5497619", "0.54802966", "0.5480182", "0.5479963", "0.5477734", "0.5458818", "0.5457388", "0.54541165", "0.54536474", "0.545347", "0.5426193", "0.54047245", "0.540035", "0.53947806", "0.5383319", "0.536889", "0.5365988", "0.53185517", "0.53159183", "0.5303305", "0.53019017", "0.5276831", "0.52766156", "0.5276488", "0.5276488", "0.52529544", "0.5237824", "0.52192765", "0.52153754", "0.5214351", "0.52015835", "0.51768655", "0.5176547", "0.5169682", "0.51656437", "0.51634735", "0.51618856", "0.51590085", "0.51574725", "0.51564765", "0.5152698", "0.5128473", "0.51133144", "0.5112047", "0.51066726", "0.5106267", "0.51020604", "0.5098282", "0.50969625", "0.50921637", "0.50819075", "0.5081689" ]
0.7141729
8
New creats a new account.
func (account *Account) New() { params, _ := pbc.NewParamsFromString(crypto.Params) pairing := params.NewPairing() A := pairing.NewG1() AT := pairing.NewG1() a := pairing.NewZr() a.Rand() G := pairing.NewG1() T := pairing.NewG1() G.SetString(crypto.G, 10) T.SetString(crypto.T, 10) A.PowZn(G, a) AT.PowZn(T, a) account.a = a.String() account.A = A.String() account.AT = AT.String() }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Create(w http.ResponseWriter, r *http.Request) {\n\n\taccountDecoder := json.NewDecoder(r.Body)\n\tvar accData Account\n\terr := accountDecoder.Decode(&accData)\n\tif err != nil {\n\t\tlog.Fatalln(\"error:\", err)\n\t}\n\taccData.CreateAccount()\n\tfmt.Fprintf(w, \"Account added successfully\")\n}", "func (controller *AccountController) NewAccount(ctx *gin.Context) {\n\tname, ok := ctx.GetPostForm(\"name\")\n\tif !ok {\n\t\tlog.WithFields(log.Fields{\"URL\": ctx.Request.URL.String()}).Warn(\"No name found in postform\")\n\n\t\terrResp, _ := restapi.NewErrorResponse(\"No name given\").Marshal()\n\t\tfmt.Fprint(ctx.Writer, string(errResp))\n\t\tctx.Abort()\n\t\treturn\n\t}\n\n\tinfo, err := authStuff.GetLoginInfoFromCtx(ctx)\n\tif err != nil {\n\t\tresponse, _ := restapi.NewErrorResponse(err.Error()).Marshal()\n\t\tfmt.Fprint(ctx.Writer, string(response))\n\t\tctx.Abort()\n\t\treturn\n\t}\n\n\tacc, err := controller.service.CreateAdd(name, info.Name, permissions.CRUD)\n\tif err != nil {\n\t\tlog.WithFields(log.Fields{\"user\": info.Name}).WithError(err).Error(\"Account Error New\")\n\n\t\tresponse, _ := restapi.NewErrorResponse(\"Couldn't create account: \" + err.Error()).Marshal()\n\t\tfmt.Fprint(ctx.Writer, string(response))\n\t\tctx.Abort()\n\t\treturn\n\t}\n\tresponse, _ := restapi.NewOkResponse(acc).Marshal()\n\tfmt.Fprint(ctx.Writer, string(response))\n\tctx.Next()\n}", "func (controller *Auth) CreateNewAccount() {\n\tpage := \"register-form-page\"\n\tcontroller.RegisterCaptchaAction(page)\n\n\tif !controller.IsCaptchaValid(page) {\n\t\tcontroller.DisplaySimpleError(\"Please enter a valid code!\")\n\t} else {\n\t\tcontroller.createNewAccount()\n\t}\n}", "func CreateAccount(w http.ResponseWriter, r *http.Request) {\n\n\tbodyR, erro := ioutil.ReadAll(r.Body)\n\tif erro != nil {\n\t\tresponses.Error(w, http.StatusUnprocessableEntity, erro)\n\t\treturn\n\t}\n\tvar account model.Account\n\tif erro = json.Unmarshal(bodyR, &account); erro != nil {\n\t\tresponses.Error(w, http.StatusBadRequest, erro)\n\t\treturn\n\t}\n\tfmt.Println(account)\n\tif erro = account.Prepare(\"cadastro\"); erro != nil {\n\t\tresponses.Error(w, http.StatusBadRequest, erro)\n\t\treturn\n\t}\n\n\tdb, erro := database.Connect()\n\tif erro != nil {\n\t\tresponses.Error(w, http.StatusInternalServerError, erro)\n\t\treturn\n\t}\n\tdefer db.Close()\n\n\trepository := repositories.NewAccountRepository(db)\n\t_, erro = repository.FindByCPF(account.Cpf)\n\tif erro != nil {\n\t\tresponses.Error(w, http.StatusConflict, erro)\n\t}\n\n\taccount.ID, erro = repository.Save(account)\n\tif erro != nil {\n\t\tresponses.Error(w, http.StatusInternalServerError, erro)\n\t\treturn\n\t}\n\n\tresponses.JSON(w, http.StatusCreated, account)\n}", "func (c *AccountController) Create(ctx *app.CreateAccountContext) error {\n\treturn nil\n}", "func (c *CoinTest) newAccount(stub shim.ChaincodeStubInterface, args []string) pb.Response {\r\n\tvar id string\r\n\tvar asset int\r\n\tvar err error\r\n\r\n\tif len(args) == 1{\r\n\t\tid = args[0]\r\n\t\tasset = 0\r\n\t}else if len(args) == 2{\r\n\t\tid = args[0]\r\n\t\t//asset, err = strconv.Atoi(args[1])\r\n\t\t//if err != nil{\r\n\t\t//\treturn shim.Error(\"Invalid asset amount, expecting a integer value\")\r\n\t\t//}\r\n\t\tasset = 0;\r\n\t}else{\r\n\t\treturn shim.Error(\"Incorrect number of arguments.\")\r\n\t}\r\n\r\n\t//deliver 0 number of TolakCoin to user\r\n\ttolakCoin.Token.Deliver(token.Address(id), int(asset))\r\n\r\n\t//write to ledger\r\n\terr = stub.PutState(id, []byte(strconv.Itoa(asset)))\r\n\tif err != nil{\r\n\t\treturn shim.Error(err.Error())\r\n\t}\r\n\r\n\treturn shim.Success(nil)\r\n}", "func (us *userService) CreateNewAccount(ctx *atreugo.RequestCtx) error {\n\tusername := string(ctx.FormValue(state.HandlerKeyUsername))\n\temail := string(ctx.FormValue(state.HandlerKeyEmail))\n\tpassword := string(ctx.FormValue(state.HandlerKeyPassword))\n\n\tif username == \"\" || email == \"\" || password == \"\" {\n\t\tctx.SetStatusCode(http.StatusBadRequest)\n\t\treturn errors.New(http.StatusText(http.StatusBadRequest))\n\t}\n\n\tus.hash.Write([]byte(password))\n\tpassword = fmt.Sprintf(\"%x\", us.hash.Sum(nil))\n\n\terr := us.usecase.Register(ctx.RequestCtx, &model.User{\n\t\tUsername: username,\n\t\tEmail: email,\n\t\tPassword: password,\n\t})\n\tif err != nil {\n\t\tctx.SetStatusCode(http.StatusInternalServerError)\n\t}\n\treturn err\n}", "func (as *Service) Create(name, iamRole, externalID string) (*Account, error) {\n\n\tbody := map[string]map[string]string{\n\t\t\"account\": {\"name\": name},\n\t}\n\n\tlog.Printf(\"Making request %v\\n\", body)\n\treq, err := as.httpClient.NewRequest(http.MethodPost, \"/setup/account\", &body)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar v common.Response\n\t_, err = as.httpClient.Do(req, &v)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif len(v.Response.Items) == 0 {\n\t\treturn nil, errors.New(\"Couldn't create account\")\n\t}\n\tvar account Account\n\n\tfmt.Println(string(v.Response.Items[0]))\n\n\terr = json.Unmarshal(v.Response.Items[0], &account)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttime.Sleep(time.Second * 5)\n\n\terr = as.setupCloudCredentials(account.ID, iamRole, externalID)\n\n\tif err != nil {\n\t\t_ = as.Delete(account.ID)\n\t\treturn nil, err\n\t}\n\n\treturn &account, nil\n}", "func NewAccount(id string, name string, primary bool, type_ string, currency Currency, balance AccountBalance, createdAt time.Time, updatedAt time.Time, resource string, resourcePath string, ready bool) *Account {\n\tthis := Account{}\n\tthis.Id = id\n\tthis.Name = name\n\tthis.Primary = primary\n\tthis.Type = type_\n\tthis.Currency = currency\n\tthis.Balance = balance\n\tthis.CreatedAt = createdAt\n\tthis.UpdatedAt = updatedAt\n\tthis.Resource = resource\n\tthis.ResourcePath = resourcePath\n\tthis.Ready = ready\n\treturn &this\n}", "func accountCreate(ctx *cli.Context) error {\n\n\tdataDir := ctx.GlobalString(utils.DataDirFlag.Name)\n\t//get keyStoreDir from KeyStoreDirFlag, if not use the default value\n\tkeyStoreDir := ctx.GlobalString(utils.KeyStoreDirFlag.Name)\n\tif keyStoreDir == \"\" {\n\t\tkeyStoreDir = keystore.KeyStoreScheme\n\t}\n\tkeyStoreDir = filepath.Join(dataDir, keyStoreDir)\n\tpassword := getPassPhrase(\"Your new account is locked with a password. Please give a password. Do not forget this password.\", true, 0, utils.MakePasswordList(ctx))\n\t_, err := utils.NewAccount(keyStoreDir, password)\n\n\treturn err\n}", "func CreateAccount(w http.ResponseWriter, r *http.Request) {\n\tvar acc models.Account\n\t_ = json.NewDecoder(r.Body).Decode(&acc)\n\n\tracc, err := models.CreateAccount(acc)\n\n\tif err != nil {\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\t_, _ = fmt.Fprintf(w, err.Error())\n\t} else {\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\t_ = json.NewEncoder(w).Encode(racc)\n\t}\n}", "func Create(w http.ResponseWriter, r *http.Request) {\n\tvar account Account\n\terr = r.ParseForm()\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\taccount.Email = r.Form[\"email\"][0]\n\taccount.Username = r.Form[\"username\"][0]\n\taccount.Password = r.Form[\"password\"][0]\n\tdb.Create(&account)\n}", "func (ws *WebServer) NewAccount(c *gin.Context) {\n\tnetwork := ws.nodeConfig.GetNetwork()\n\tif network == \"\" {\n\t\tReturnError(c, 500, ErrorNoNetwork.Error())\n\t\treturn\n\t}\n\tn := sdk.Testnet\n\tif network == \"bitmark\" {\n\t\tn = sdk.Livenet\n\t}\n\tseedFile := filepath.Join(ws.rootPath, \"bitmarkd\", network, \"proof.sign\")\n\tif _, err := os.Stat(seedFile); err == nil {\n\t\tReturnError(c, 500, ErrorNoSeedFile.Error())\n\t\treturn\n\t}\n\n\ta, err := sdk.NewAccount(n)\n\tif err != nil {\n\t\tReturnError(c, 500, ErrorCreateAccount.Error())\n\t\treturn\n\t}\n\tseed := a.Seed()\n\n\tf, err := os.OpenFile(seedFile, os.O_WRONLY|os.O_TRUNC|os.O_CREATE, 0600)\n\tif err != nil {\n\t\tReturnError(c, 500, ErrorOpenSeedFile.Error())\n\t\treturn\n\t}\n\tdefer f.Close()\n\n\t_, err = f.WriteString(fmt.Sprintf(\"SEED:%s\", seed))\n\tif err != nil {\n\t\tReturnError(c, 500, ErrorToWriteSeedFile.Error())\n\t\treturn\n\t}\n\tws.SetAccount(a.AccountNumber(), seed, network) // Record in AccountInfo in memory\n\terr = ws.saveAcct()\n\tif nil != err {\n\t\tReturnError(c, 500, ErrorAutoSaveAccount.Error())\n\t\treturn\n\n\t}\n\tc.JSON(200, map[string]interface{}{\n\t\t\"ok\": 1,\n\t})\n}", "func (c *AccountController) Create(ctx echo.Context) error {\n\tmodel := account.Account{}\n\terr := ctx.Bind(&model)\n\tif err != nil {\n\t\treturn ctx.JSON(http.StatusUnprocessableEntity, err.Error())\n\t}\n\n\tres, err := c.AccountUsecase.Create(&model)\n\tif err != nil {\n\t\treturn ctx.JSON(http.StatusInternalServerError, err.Error())\n\t}\n\n\treturn ctx.JSON(http.StatusCreated, res)\n}", "func CreateAccount(response http.ResponseWriter, request *http.Request) {\n\taccount := new(models.Account)\n\terr := utils.ExtractFromRequest(request, account)\n\tif err != nil {\n\t\tlog.Println(\"ERROR: Can't extract account from request:\", err)\n\t\thttp.Error(response, \"Error while parsing request body\", 500)\n\t\treturn\n\t}\n\n\tlog.Println(\"Create account with email:\", account.Payload.Email)\n\terr = persistence.InsertNewAccount(account)\n\tif err != nil {\n\t\tlog.Println(\"ERROR: Can't create new account:\", err)\n\t\tif err == persistence.ErrAccountAlreadyExists {\n\t\t\thttp.Error(response, \"User already exists\", 403)\n\t\t} else {\n\t\t\thttp.Error(response, \"Error while creating new account\", 500)\n\t\t}\n\t\treturn\n\t}\n\n\tresponse.WriteHeader(200)\n}", "func (s *Service) Create(newAccountDefinition *model.NewAccountDefinition) *CreateOp {\n\treturn &CreateOp{\n\t\tCredential: s.credential,\n\t\tMethod: \"POST\",\n\t\tPath: \"/v2/accounts\",\n\t\tPayload: newAccountDefinition,\n\t\tQueryOpts: make(url.Values),\n\t\tVersion: esign.APIv2,\n\t}\n}", "func (ak AccountKeeper) NewAccount(ctx sdk.Context, acc exported.Account) exported.Account {\n\t// FIXME: update account number\n\treturn acc\n}", "func (policy *ticketPolicy) OnCreateNewAccount(acc *types.Account) {\n}", "func (account *Account) create(accountType string) *Account {\n\tfmt.Println(\"account creation with type\")\n\taccount.accountType = accountType\n\treturn account\n}", "func Create(w http.ResponseWriter, r *http.Request) {\n\tauthUser, err := auth.GetUserFromJWT(w, r)\n\tif err != nil {\n\t\tresponse.FormatStandardResponse(false, \"error-auth\", \"\", err.Error(), w)\n\t\treturn\n\t}\n\n\tdefer r.Body.Close()\n\n\t// Decode the JSON body\n\tacct := datastore.Account{}\n\terr = json.NewDecoder(r.Body).Decode(&acct)\n\tswitch {\n\t// Check we have some data\n\tcase err == io.EOF:\n\t\tresponse.FormatStandardResponse(false, \"error-account-data\", \"\", \"No account data supplied.\", w)\n\t\treturn\n\t\t// Check for parsing errors\n\tcase err != nil:\n\t\tresponse.FormatStandardResponse(false, \"error-decode-json\", \"\", err.Error(), w)\n\t\treturn\n\t}\n\n\tcreateHandler(w, authUser, false, acct)\n}", "func CreateAccount(account models.Account) string {\n\t// generate the id for an account, with length of 15 digit\naccount.ID = genString(15)\n// store the values\nAccounts = append(Accounts, account)\n// Send account id in the response, if account stored successfully\nreturn account.ID\n}", "func CreateAccount(w http.ResponseWriter, r *http.Request) {\n\taccount := &model.Account{}\n\t// Decode the request body into struct and fail if any error occur\n\terr := json.NewDecoder(r.Body).Decode(account)\n\tif err != nil {\n\t\tresponse.RespondWithStatus(w,\n\t\t\tresponse.Message(false, \"Invalid request\"),\n\t\t\thttp.StatusBadRequest)\n\t\treturn\n\t}\n\n\tdata, status, ok := account.Create() // Create account\n\tif !ok {\n\t\tresponse.RespondWithStatus(w, data, status)\n\t\treturn\n\t}\n\tresponse.RespondWithStatus(w, data, status)\n}", "func CreateAccount(w http.ResponseWriter, r *http.Request) {\n\taccount := &models.Account{}\n\terr := json.NewDecoder(r.Body).Decode(account)\n\tif err != nil {\n\t\tutils.Respond(w, utils.Message(false, \"Invalid Request\"))\n\t\treturn\n\t}\n\n\tresponse := account.Create()\n\tutils.Respond(w, response)\n}", "func MakeNewAccount(name string) (*MyAccount, error) {\n\tkeys, err := NewKeypair()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &MyAccount{\n\t\tFields: make(map[string]string),\n\t\tKeys: keys,\n\t\tName: name,\n\t}, nil\n}", "func (s Service) CreateNewAccount(ctx context.Context, acc *account.Account, password string) (id, url, tmpToken string, err error) {\n\tspan := s.tracer.MakeSpan(ctx, \"CreateNewAccount\")\n\tdefer span.Finish()\n\n\t// pass data in context\n\ts.passContext(&ctx)\n\n\tacc.FirstName = strings.TrimSpace(acc.FirstName)\n\tacc.Lastname = strings.TrimSpace(acc.Lastname)\n\tacc.Emails[0].Email = strings.TrimSpace(acc.Emails[0].Email)\n\tacc.Emails[0].Email = strings.ToLower(acc.Emails[0].Email)\n\tacc.Username = strings.ToLower(acc.Username)\n\tacc.Username = strings.TrimSpace(acc.Username)\n\n\tyear, month, day := acc.Birthday.Birthday.Date()\n\tacc.BirthdayDate = account.Date{\n\t\tDay: day,\n\t\tMonth: int(month),\n\t\tYear: year,\n\t}\n\n\terr = emptyValidator(acc.FirstName, acc.Lastname, acc.Username)\n\tif err != nil {\n\t\treturn \"\", \"\", \"\", err\n\t}\n\terr = fromTwoToSixtyFour(acc.FirstName)\n\tif err != nil {\n\t\treturn \"\", \"\", \"\", err\n\t}\n\terr = fromTwoToHundredTwentyEight(acc.Lastname)\n\tif err != nil {\n\t\treturn \"\", \"\", \"\", err\n\t}\n\n\terr = userNameValidator(acc.Username)\n\tif err != nil {\n\t\treturn \"\", \"\", \"\", err\n\t}\n\tif len(acc.Emails) > 0 {\n\t\terr = emailValidator(acc.Emails[0].Email)\n\t\tif err != nil {\n\t\t\treturn \"\", \"\", \"\", err\n\t\t}\n\t} else {\n\t\treturn \"\", \"\", \"\", errors.New(\"Please Enter Email\")\n\t}\n\terr = validPassword(password)\n\tif err != nil {\n\t\treturn \"\", \"\", \"\", err\n\t}\n\n\t// TODO: trim data!\n\t// TODO: make first letter capital in some fields!\n\n\t// check if email is not busy\n\tinUse, err := s.repository.Users.IsEmailAlreadyInUse(ctx, acc.Emails[0].Email)\n\tif err != nil {\n\t\ts.tracer.LogError(span, err)\n\t}\n\tif inUse {\n\t\terr = errors.New(\"this_email_already_in_use\") // TODO: how it return as gRPC status?\n\t\treturn\n\t}\n\n\t// check if usernmae is not busy\n\tusernameInUse, err := s.repository.Users.IsUsernameBusy(ctx, acc.Username)\n\tif err != nil {\n\t\ts.tracer.LogError(span, err)\n\t}\n\tif usernameInUse {\n\t\terr = errors.New(\"this_username_already_in_use\") // TODO: how it return as gRPC status?\n\t\treturn\n\t}\n\n\t// TODO: check phone is not busy yet (in future)\n\n\t// define location by IP address\n\tvar ip string\n\tmd, ok := metadata.FromIncomingContext(ctx)\n\tif !ok {\n\t\ts.tracer.LogError(span, errors.New(\"coudn't resolve ip address\"))\n\t} else {\n\t\tstrArr := md.Get(\"ip\")\n\t\tif len(strArr) > 0 {\n\t\t\tip = strArr[0]\n\t\t}\n\t}\n\tcountry, err := s.repository.GeoIP.GetCountryISOCode(net.ParseIP(ip))\n\tif err != nil {\n\t\ts.tracer.LogError(span, err)\n\t}\n\tif country != \"\" {\n\t\tacc.Location = &account.UserLocation{\n\t\t\tLocation: location.Location{\n\t\t\t\tCountry: &location.Country{\n\t\t\t\t\tID: country,\n\t\t\t\t},\n\t\t\t},\n\t\t}\n\t}\n\n\tid = acc.GenerateID()\n\turl = acc.GenerateURL()\n\tacc.Status = status.UserStatusNotActivated // set not_activated status\n\tacc.CreatedAt = time.Now() // set date of registration\n\tacc.Emails[0].Primary = true // set email as primary\n\tacc.Emails[0].GenerateID()\n\n\t// encode password\n\tencryptedPass, err := bcrypt.GenerateFromPassword([]byte(password), bcrypt.DefaultCost)\n\tif err != nil {\n\t\ts.tracer.LogError(span, err)\n\t}\n\n\terr = s.repository.Users.SaveNewAccount(ctx, acc, string(encryptedPass))\n\tif err != nil {\n\t\ts.tracer.LogError(span, err)\n\t}\n\n\temptyString := \"\"\n\n\terr = s.repository.arrangoRepo.SaveUser(ctx, &arangorepo.User{\n\t\tID: acc.GetID(),\n\t\tCreatedAt: time.Now(),\n\t\tFirstname: acc.FirstName,\n\t\tLastname: acc.Lastname,\n\t\tStatus: \"ACTIVATED\",\n\t\tURL: acc.URL,\n\t\tPrimaryEmail: acc.Emails[0].Email,\n\t\tGender: arangorepo.Gender{\n\t\t\tGender: acc.Gender.Gender,\n\t\t\tType: &emptyString,\n\t\t},\n\t})\n\tif err != nil {\n\t\tlog.Println(\"arrangoRepo.SaveUser:\", err)\n\t}\n\n\tif acc.GetInvitedByID() != \"\" {\n\t\ts.AddGoldCoinsToWallet(ctx, acc.GetInvitedByID(), 1)\n\t}\n\n\t// generate tmp code for activation\n\ttmpCode, err := s.repository.Cache.CreateTemporaryCodeForEmailActivation(ctx, id, acc.Emails[0].Email)\n\tif err != nil {\n\t\ts.tracer.LogError(span, err)\n\t}\n\n\t// log.Println(\"activation code:\", tmpCode) // TODO: delete later\n\t// log.Println(\"user id:\", id) // TODO: delete later\n\n\t// send email\n\t// err = s.mailRPC.SendEmail(\n\t// \tctx,\n\t// \tacc.Emails[0].Email,\n\t// \tfmt.Sprint(\"<html><body><a target='_blank' href='https://\"+s.Host+\"/api/activate/user?token=\", tmpCode, \"&user_id=\", id, \"'>Activate</a></body></html>\")) // TODO: write template for message\n\t// if err != nil {\n\t// \ts.tracer.LogError(span, err)\n\t// }\n\t// // fmt.Println(fmt.Sprint(\"https://\"+s.Host+\"/api/activate/user?token=\", tmpCode, \"&user_id=\", id)) // TODO: delete later\n\n\t// emailMessage := fmt.Sprint(\"<html><body><a target='_blank' href='https://\"+s.Host+\"/api/activate/user?token=\", tmpCode, \"&user_id=\", id, \"'>Activate</a></body></html>\")\n\temailMessage := s.tpl.GetActivationMessage(fmt.Sprint( /*\"https://\"+s.Host+\"/api/activate/user?token=\", tmpCode, \"&user_id=\", id)*/ tmpCode))\n\t// log.Println(acc.Emails[0].Email, emailMessage)\n\n\terr = s.mq.SendEmail(acc.Emails[0].Email, \"Activation\", emailMessage)\n\tif err != nil {\n\t\ts.tracer.LogError(span, err)\n\t}\n\n\t// generate tmp token for not activated user\n\ttmpToken, err = s.repository.Cache.CreateTemporaryCodeForNotActivatedUser(id)\n\tif err != nil {\n\t\ts.tracer.LogError(span, err)\n\t}\n\n\terr = s.CreateWalletAccount(ctx, id)\n\tif err != nil {\n\t\ts.tracer.LogError(span, err)\n\t}\n\n\treturn id, url, tmpToken, nil\n}", "func newAccount() *Account {\n\treturn &Account{\n\t\tblocks: make(map[string]uint64),\n\t}\n}", "func generateNewAccount() string {\n\taccount := crypto.GenerateAccount()\n\tpassphrase, err := mnemonic.FromPrivateKey(account.PrivateKey)\n\tif err != nil {\n\t\tfmt.Printf(\"Error creating new account: %s\\n\", err)\n\t} else {\n\t\tfmt.Printf(\"Created new account: %s\\n\", account.Address)\n\t\tfmt.Printf(\"Generated mnemonic: \\\"%s\\\"\\n\", passphrase)\n\t}\n\treturn account.Address.String()\n}", "func CreateAccount(form form.Account) (model *Account, err error) {\n\tmodel = &Account{\n\t\tShareSize: \"\",\n\t\tShareExpires: 0,\n\t\tRetryLimit: 3,\n\t\tSyncStatus: AccountSyncStatusRefresh,\n\t}\n\n\terr = model.SaveForm(form)\n\n\treturn model, err\n}", "func NewAccount(newid AccountIDType, newemail string) *Account {\n\treturn &Account{\n\t\tid: newid,\n\t\temail: newemail,\n\t}\n}", "func NewAccount(user, apiKey string) *Account {\n\treturn &Account{user: user, apiKey: apiKey}\n}", "func NewAccount(\n\tconfig *accounts.AccountConfig,\n\taccountCoin *Coin,\n\thttpClient *http.Client,\n\tlog *logrus.Entry,\n) *Account {\n\tlog = log.WithField(\"group\", \"eth\").\n\t\tWithFields(logrus.Fields{\"coin\": accountCoin.String(), \"code\": config.Config.Code, \"name\": config.Config.Name})\n\tlog.Debug(\"Creating new account\")\n\n\taccount := &Account{\n\t\tBaseAccount: accounts.NewBaseAccount(config, accountCoin, log),\n\t\tcoin: accountCoin,\n\t\tdbSubfolder: \"\", // set in Initialize()\n\t\tsigningConfiguration: nil,\n\t\thttpClient: httpClient,\n\t\tbalance: coin.NewAmountFromInt64(0),\n\n\t\tenqueueUpdateCh: make(chan struct{}),\n\t\tquitChan: make(chan struct{}),\n\n\t\tlog: log,\n\t}\n\n\treturn account\n}", "func (e Account) EntNew() ent.Ent { return &Account{} }", "func (e Account) EntNew() ent.Ent { return &Account{} }", "func CreateAccount(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(http.StatusCreated)\n\n\tvar account entity.Account\n\tjson.NewDecoder(r.Body).Decode(&account)\n\tdb.DBCon.Create(&account)\n\n\tjson.NewEncoder(w).Encode(account)\n}", "func (service AccountsService) Create(a Account) (*Response, Account, error) {\n\treq, err := service.client.newRequest(\"POST\", \"accounts\", nil, a)\n\tif err != nil {\n\t\treturn nil, Account{}, err\n\t}\n\n\tvar dest Account\n\tres, err := service.client.do(req, &dest)\n\n\tdest.BillingInfo = nil\n\n\treturn res, dest, err\n}", "func CreateAccount(u usecase.UseCase) gin.HandlerFunc {\n\treturn func(c *gin.Context) {\n\t\tvar a presenter.Account\n\t\terr := c.BindJSON(&a)\n\t\tif err != nil {\n\t\t\tresponseFailure(c, http.StatusText(http.StatusInternalServerError),\n\t\t\t\t\"Account can't be created\",\n\t\t\t\t\"Error when converting the parameters sent to json\", \"\", http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\n\t\tacc, err := u.NewAccount(&account.Account{\n\t\t\tID: a.ID, DocumentNumber: a.DocumentNumber,\n\t\t\tAvailableCreditLimit: a.AvailableCreditLimit,\n\t\t})\n\t\tif err != nil {\n\t\t\tresponseFailure(c, http.StatusText(http.StatusInternalServerError),\n\t\t\t\t\"Account can't be created\",\n\t\t\t\tfmt.Sprintf(\"Internal server error when creating a new account - datails err: %s\", err.Error()), \"\", http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\tc.JSON(http.StatusCreated, acc)\n\t}\n}", "func NewAccount(txr *repository.Transactioner) repository.Account {\n\treturn &account{txr: txr}\n}", "func (a *Account) CreateAcct(password string) (*Account, *http.Response, []error) {\n\tk := kumoru.New()\n\n\tk.Put(fmt.Sprintf(\"%s/v1/accounts/%s\", k.EndPoint.Authorization, a.Email))\n\tk.Send(fmt.Sprintf(\"given_name=%s&surname=%s&password=%s\", a.GivenName, a.Surname, password))\n\n\tresp, body, errs := k.End()\n\n\tif len(errs) > 0 {\n\t\treturn a, resp, errs\n\t}\n\n\tif resp.StatusCode >= 400 {\n\t\terrs = append(errs, fmt.Errorf(\"%s\", resp.Status))\n\t}\n\n\terr := json.Unmarshal([]byte(body), &a)\n\n\tif err != nil {\n\t\terrs = append(errs, err)\n\t\treturn a, resp, errs\n\t}\n\n\treturn a, resp, nil\n}", "func NewAccount(firstName string) *Account {\n\tthis := Account{}\n\tthis.FirstName = firstName\n\treturn &this\n}", "func (u *accRepo) Create(account *models.Account) error {\n\treturn nil\n}", "func NewAccount(username string) (*AccountRow, error) {\n\tquery := `\n\t\tinsert into accounts (user_name)\n\t\t\tvalues ($1)\n\t\t\ton conflict (user_name)\n\t\t\t\tdo nothing\n\t\treturning\n\t\t\tid, user_name`\n\n\trowData := &AccountRow{}\n\trow := GlobalConn.QueryRow(query, username)\n\n\tif err := row.Scan(&rowData.ID, &rowData.Name); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn rowData, nil\n}", "func create_account_ (stub shim.ChaincodeStubInterface, account *Account) error {\n var old_account Account\n row_was_found,err := util.InsertTableRow(stub, ACCOUNT_TABLE, row_keys_of_Account(account), account, util.FAIL_BEFORE_OVERWRITE, &old_account)\n if err != nil {\n return err\n }\n if row_was_found {\n return fmt.Errorf(\"Could not create account %v because an account with that Name already exists\", *account)\n }\n return nil // success\n}", "func Create(r render.Render, params martini.Params, db database.Datastore, data Account) {\n\n\t// Create a bcrypt hash from the password as we don't want to store\n\t// plain-text passwords in the database\n\tpwdHash, bcryptError := bcrypt.GenerateFromPassword([]byte(data.Password), 0)\n\tif bcryptError != nil {\n\t\tresponses.Error(r, bcryptError.Error())\n\t}\n\n\t// Set the hashed password\n\tdata.Password = string(pwdHash)\n\n\t// Store the object in the database. In case the\n\t// database operation fails, an error response is sent back to the caller.\n\terr := db.GetDBMap().Insert(&data)\n\tif err != nil {\n\t\tresponses.Error(r, err.Error())\n\t\treturn\n\t}\n\tresponses.OKStatusPlusData(r, data, 1)\n}", "func (m *MegaCorp) createAccount(firstName, lastName string) (acct *Account, err error) {\n\tacct = &Account{ID: util.UUID4(), FirstName: firstName, LastName: lastName}\n\t_, err = Me.Put(Ledger, fmt.Sprintf(\"account.%s\", acct.ID), ToJSON(acct))\n\treturn\n}", "func (s *Service) AccountCreate(c *gin.Context, roleID model.AccessRole) error {\n\treturn s.IsLowerRole(c, roleID)\n}", "func (c *UsersController) AccountCreate(ctx *app.AccountCreateUsersContext) error {\n\t// UsersController_AccountCreate: start_implement\n\n\t// Put your logic here\n\t// 使用している端末の仮ユーザーが存在するか\n\tuserTerminalDB := models.NewUserTerminalDB(c.db)\n\tuserTerminal, err := userTerminalDB.GetByIdentifier(ctx.Context, ctx.Identifier)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"%v\", err)\n\t}\n\t// メールアドレスが既に登録されているか、登録されていれば既に登録されているユーザーに紐付ける\n\tuserDB := models.NewUserDB(c.db)\n\tcurrentUser, err := userDB.GetByEmail(ctx.Context, ctx.Email)\n\tif err == nil {\n\t\t// 既に存在しているユーザーと端末情報を紐付ける\n\t\tuserTerminal.UserID = currentUser.ID\n\t\tuserTerminalDB.Update(ctx.Context, userTerminal)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"%v\", err)\n\t\t}\n\t\tmessage := \"alreadyExists\"\n\t\treturn ctx.OK(&app.Message{&message})\n\t}\n\t// メールアドレスとユーザーを紐付ける\n\tnewUser := &models.User{}\n\tnewUser.ID = userTerminal.UserID\n\tnewUser.Email = ctx.Email\n\terr = userDB.Update(ctx.Context, newUser)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"%v\", err)\n\t}\n\t// UsersController_AccountCreate: end_implement\n\tmessage := \"ok\"\n\treturn ctx.OK(&app.Message{&message})\n}", "func (w *Wallet) NewAccount(index *uint32) (a *Account, err error) {\n\tindex2 := w.nextIndex\n\tif index != nil {\n\t\tindex2 = *index\n\t}\n\ta = &Account{w: w, index: index2}\n\tif err = w.impl.deriveAccount(a); err != nil {\n\t\treturn\n\t}\n\tpubkeyToAddress := util.PubkeyToAddress\n\tif w.isBanano {\n\t\tpubkeyToAddress = util.PubkeyToBananoAddress\n\t}\n\tif a.address, err = pubkeyToAddress(a.pubkey); err != nil {\n\t\treturn\n\t}\n\tif index == nil {\n\t\tw.nextIndex++\n\t}\n\tif _, ok := w.accounts[a.address]; !ok {\n\t\tw.accounts[a.address] = a\n\t} else if index == nil {\n\t\treturn w.NewAccount(nil)\n\t}\n\treturn\n}", "func (biz *BizAccountAccess) newAccountAccess(ctx context.Context, tx *sql.Tx,\n\tu coremodel.User, accType coremodel.AccountAccessType,\n) (*coremodel.AccountAccess, error) {\n\tac, err := coremodel.NewAccountAccess(u, accType)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif err = biz.dal.Insert(ctx, tx, ac); err != nil {\n\t\treturn nil, err\n\t}\n\treturn ac, nil\n}", "func CreateAccount(resp http.ResponseWriter, req *http.Request) {\n\tresp.Header().Add(\"Content-Type\", \"application/json\")\n\n\tvar account domain.Account\n\tdecoder := json.NewDecoder(req.Body)\n\n\terr := decoder.Decode(&account)\n\n\tif err != nil {\n\t\thttp.Error(resp, \"Bad Request\", http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tapiErr := service.CreateAccount(account)\n\n\tif apiErr != nil {\n\t\tresp.WriteHeader(http.StatusInternalServerError)\n\t\tencoder := json.NewEncoder(resp)\n\t\tencoder.Encode(apiErr)\n\t}\n\n\tresp.WriteHeader(http.StatusOK)\n\n\tjsonValue, _ := json.Marshal(utils.AppMsgs{http.StatusOK, \"Account was created\"})\n\n\tresp.Write(jsonValue)\n}", "func (h *handler) Create(w http.ResponseWriter, r *http.Request) {\n\tctx := r.Context()\n\n\tvar req AddRequest\n\tif err := json.NewDecoder(r.Body).Decode(&req); err != nil {\n\t\thttperror.BadRequest(w, err)\n\t\treturn\n\t}\n\n\taccount := new(object.Account)\n\taccount.Username = req.Username\n\tif err := account.SetPassword(req.Password); err != nil {\n\t\thttperror.InternalServerError(w, err)\n\t\treturn\n\t}\n\n\t// 既存のusernameを弾く、ここでやらないほうがいい気がするが、BadRequestを返すとなると、ここ?\n\talreadyAccount, err := h.app.Dao.Account().FindByUsername(ctx, req.Username)\n\tlog.Printf(\"already account: %v\", alreadyAccount)\n\tif err != nil {\n\t\thttperror.InternalServerError(w, err)\n\t\treturn\n\t}\n\tif alreadyAccount != nil {\n\t\thttperror.BadRequest(w, fmt.Errorf(\"%v\", \"Validation Error: This name exists.\"))\n\t\treturn\n\t}\n\n\taccount, err = h.app.Dao.Account().Create(ctx, *account)\n\tif err != nil {\n\t\thttperror.InternalServerError(w, err)\n\t\treturn\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tif err := json.NewEncoder(w).Encode(account); err != nil {\n\t\thttperror.InternalServerError(w, err)\n\t\treturn\n\t}\n}", "func CreateAccount(ee engine.Exchange) sknet.HandlerFunc {\n\treturn func(c *sknet.Context) error {\n\t\terrRlt := &pp.EmptyRes{}\n\t\tfor {\n\t\t\treq := pp.CreateAccountReq{}\n\t\t\tif err := c.BindJSON(&req); err != nil {\n\t\t\t\tlogger.Error(err.Error())\n\t\t\t\terrRlt = pp.MakeErrResWithCode(pp.ErrCode_WrongRequest)\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\t// validate pubkey.\n\t\t\tif err := validatePubkey(req.GetPubkey()); err != nil {\n\t\t\t\tlogger.Error(err.Error())\n\t\t\t\terrRlt = pp.MakeErrResWithCode(pp.ErrCode_WrongPubkey)\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\t// create account with pubkey.\n\t\t\tif _, err := ee.CreateAccountWithPubkey(req.GetPubkey()); err != nil {\n\t\t\t\tlogger.Error(err.Error())\n\t\t\t\terrRlt = pp.MakeErrResWithCode(pp.ErrCode_WrongRequest)\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tres := pp.CreateAccountRes{\n\t\t\t\tResult: pp.MakeResultWithCode(pp.ErrCode_Success),\n\t\t\t\tPubkey: req.Pubkey,\n\t\t\t\tCreatedAt: pp.PtrInt64(time.Now().Unix()),\n\t\t\t}\n\n\t\t\treturn c.SendJSON(&res)\n\t\t}\n\n\t\treturn c.Error(errRlt)\n\t}\n}", "func (pg *Postgres) CreateAccount(name string, isActive bool) (*dto.Account, error) {\n\tvar id int\n\trow := pg.db.QueryRow(\"INSERT INTO account (name, isActive) VALUES ($1, $2) RETURNING id\", name, isActive)\n\n\tif err := row.Scan(&id); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &dto.Account{\n\t\tID: id,\n\t\tName: name,\n\t\tIsActive: isActive,\n\t}, nil\n}", "func (a *AccountClient) Create(accountParams AccountParams) (*Resource, error) {\n\n\tresp, err := a.client.R().\n\t\tSetResult(&Resource{}).\n\t\tSetBody(map[string]AccountParams{\"data\": accountParams}).\n\t\tPost(\"/v1/organisation/accounts\")\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"create account failed: %s\", err)\n\t}\n\n\tif resp.Error() != nil {\n\t\treturn nil, getAPIError(resp)\n\t}\n\n\treturn resp.Result().(*Resource), nil\n}", "func NewAccount(opts ...AccountCreationOption) (*Account, error) {\n\taccount := &Account{\n\t\tBalance: big.NewInt(0),\n\t\tvotingWeight: big.NewInt(0),\n\t\taccountType: 1,\n\t}\n\tfor _, opt := range opts {\n\t\tif err := opt(account); err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"failed to apply account creation option\")\n\t\t}\n\t}\n\treturn account, nil\n}", "func (account *Account) Create() map[string]interface{} {\n\n\tif res, ok := account.Validate(); !ok {\n\t\treturn res\n\t}\n\n\t// 密码加密,并替换新输入的原密码(MinCost int = 4, MaxCost int = 32, DefaultCost int = 10)\n\thashedPassword, _ := bcrypt.GenerateFromPassword([]byte(account.Password), bcrypt.DefaultCost)\n\taccount.Password = string(hashedPassword)\n\n\tGetDB().Create(account)\n\n\tif account.ID <= 0 {\n\t\treturn u.Message(false, ERROR_DBCONNECTION)\n\t}\n\n\t// Create new JWT token from the new Registered account, 对account.ID 进行加密\n\ttk := &Token{UserId: account.ID}\n\n\ttokenString, err := HashContentWithHS256(tk)\n\tif err != nil {\n\t\treturn u.Message(false, \"Parse Error\")\n\t}\n\n\taccount.Token = tokenString\n\n\t// 生成返回数据给前端\n\taccount.Password = \"\"\n\tresponse := u.Message(true, \"Account has been Created\")\n\tresponse[\"account\"] = account\n\treturn response\n}", "func NewAccount(acc *types.Account) *Account {\n\treturn &Account{\n\t\tAccount: *acc,\n\t}\n}", "func (repository Accounts) Create(account models.Account) (uint64, error) {\n\tstatement, err := repository.db.Prepare(\n\t\t\"insert into accounts (name, cpf, secret, balance) values (?,?,?,?)\",\n\t)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tdefer statement.Close()\n\n\tresult, err := statement.Exec(account.Name, account.Cpf, account.Secret, account.Balance)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\tlastIDInserted, err := result.LastInsertId()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn uint64(lastIDInserted), nil\n}", "func NewAccount(email string) *Account {\n\treturn &Account{email: email}\n}", "func CreateAccount(name_optional ...string) (*types.Account, error) {\n\tname := \"\"\n\tif len(name_optional) > 0 {\n\t\tname = name_optional[0]\n\t}\n\tpublicKey, privateKey := crypto.GenerateKeyPair()\n\taddress := crypto.ToAddress(publicKey)\n\taccount := &types.Account{}\n\taccount.Address = hex.EncodeToString(address)\n\taccount.PrivateKey = hex.EncodeToString(privateKey)\n\taccount.Balance = big.NewInt(0)\n\taccount.Name = name\n\tnow := time.Now()\n\taccount.Created = now\n\taccount.Updated = now\n\n\treturn account, nil;\n}", "func (c *Client) CreateAccount(accountKey crypto.PrivateKey, email string) (string, error) {\n\tconst op errors.Op = \"acmeclient/client.CreateAccount\"\n\n\tuser := &User{\n\t\tEmail: email,\n\t\tPrivateKey: accountKey,\n\t}\n\tcfg := lego.NewConfig(user)\n\tcfg.CADirURL = c.DirectoryURL\n\tlegoClient, err := lego.NewClient(cfg)\n\tif err != nil {\n\t\treturn \"\", errors.New(op, \"create client for new ACME account\", err)\n\t}\n\terr = user.Register(legoClient)\n\tif err != nil {\n\t\treturn \"\", errors.New(op, \"register new ACME account\", err)\n\t}\n\treturn user.Registration.URI, nil\n}", "func (h *handler) Create(w http.ResponseWriter, r *http.Request) {\n\tctx := r.Context()\n\n\tvar req AddRequest\n\tif err := json.NewDecoder(r.Body).Decode(&req); err != nil {\n\t\thttperror.BadRequest(w, err)\n\t\treturn\n\t}\n\n\taccount := new(object.Account)\n\taccount.Username = req.Username\n\tif err := account.SetPassword(req.Password); err != nil {\n\t\thttperror.InternalServerError(w, err)\n\t\treturn\n\t}\n\n\trepoAccount := h.app.Dao.Account() // domain/repository の取得\n\taccount, err := repoAccount.Create(ctx, account)\n\tif err != nil {\n\t\t// Todo: これrepoにエラー出させたほうが良くない?\n\t\tif mysqlErr, ok := err.(*mysql.MySQLError); ok {\n\t\t\tif mysqlErr.Number == 1062 {\n\t\t\t\thttperror.BadRequest(w, errors.New(\"User already exists\"))\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\thttperror.InternalServerError(w, err)\n\t\treturn\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tif err := json.NewEncoder(w).Encode(account); err != nil {\n\t\thttperror.InternalServerError(w, err)\n\t\treturn\n\t}\n}", "func tNewUser(lbl string) *tUser {\n\tintBytes := make([]byte, 4)\n\tbinary.BigEndian.PutUint32(intBytes, acctCounter)\n\tacctID := account.AccountID{}\n\tcopy(acctID[:], acctTemplate[:])\n\tcopy(acctID[account.HashSize-4:], intBytes)\n\taddr := strconv.Itoa(int(acctCounter))\n\tsig := []byte{0xab} // Just to differentiate from the addr.\n\tsig = append(sig, intBytes...)\n\tsigHex := hex.EncodeToString(sig)\n\tacctCounter++\n\treturn &tUser{\n\t\tsig: sig,\n\t\tsigHex: sigHex,\n\t\tacct: acctID,\n\t\taddr: addr,\n\t\tlbl: lbl,\n\t}\n}", "func (act Account) Create(nr Application) (Resource, error) {\n\tvar r Resource\n\tif nr.Validate() != nil {\n\t\treturn r, nr.Validate()\n\t}\n\terr := common.SendPostRequest(fmt.Sprintf(applications.Create, act.AccountSid), nr, act, &r)\n\treturn r, err\n}", "func CreateAccount(info *Account) (*Account, error){\n\tgate := data.NewGateway()\n\tdto := info.ToDto()\n\tif acc, err := gate.Create(dto); err != nil {\n\t\tlog.Printf(err.Error())\n\t\treturn nil, err\n\t}else {\n\t\treturn NewAccountFromDto(acc), nil\n\t}\n}", "func NewAccount() *Account {\n\n\tcreatedAccount := Account{}\n\n\tnewPrivateKey, err := generatePrivateKey(4096)\n\tif err != nil {\n\t\tlog.Fatal(\"Error generating rsa private key \", err)\n\t}\n\tnewPublicKey, err := generatePublicKey(&newPrivateKey.PublicKey)\n\tif err != nil {\n\t\tlog.Fatal(\"Error generating rsa public key \", err)\n\t}\n\tcreatedAccount.privateKey = newPrivateKey\n\tcreatedAccount.Address = string(newPublicKey)\n\tcreatedAccount.Amount = 0\n\n\treturn &createdAccount\n}", "func NewAccount(owner string) *Account {\n\treturn &Account{owner: owner, balance: 0}\n}", "func (s *Service) CreateAccount(request commons.CreateAccountRequest) (*commons.CreateAccountResponse, error) {\n\tif ok, err := request.Validate(); !ok {\n\t\treturn nil, err\n\t}\n\treturn &commons.CreateAccountResponse{\n\t\tAccountNo: \"90010062653\",\n\t\tAccountType: commons.MainAccountType,\n\t}, nil\n}", "func NewAccount() *Account {\n\tacc := &Account{}\n\tpriv, pub := newKeyPair()\n\tacc.PriKey = priv\n\tacc.PubKey = pub\n\treturn acc\n}", "func (s *AuthServiceServer) CreateAccount(ctx context.Context, req *v1.CreateRequest) (*v1.CreateResponse, error) {\n\tif err := s.checkAPI(req.Api); err != nil {\n\t\treturn nil, err\n\t}\n\n\tc, err := s.connect(ctx)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer c.Close()\n\n\treminder, err := ptypes.Timestamp(req.Data.GetReminder())\n\tif err != nil {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"reminder field has invalid format-> \"+err.Error())\n\t}\n\tpassword , username := req.Data.GetPassword() , req.Data.GetEmail()\n\tres, err := c.ExecContext(ctx, \"INSERT INTO members(`password`, `username`, `Reminder`) VALUES(?, ?, ?)\", password, username , reminder)\n\tif err != nil {\n\t\treturn nil, status.Error(codes.Unknown, \"failed to insert into members-> \"+err.Error())\n\t}\n\tres , err = c.ExecContext(ctx, \"INSERT INTO logs(`action`, `reminder`, `user`) VALUES(? , ? , ? )\", \"CREATE\" , reminder , username )\n\tif err != nil {\n\t\treturn nil, status.Error(codes.Unknown, \"failed to insert into logs-> \"+err.Error())\n\t}\n\t_ , err = res.LastInsertId()\n\tif err != nil {\n\t\treturn nil, status.Error(codes.Unknown, \"failed to retrieve id for created logs-> \"+err.Error())\n\t}\n\treturn &v1.CreateResponse{\n\t\tApi: apiVersion,\n\t\tSuccess: true,\n\t\tError: \"El usuario fue exitosamente creado\" + username,\n\t\t}, nil\n}", "func CreateAccount (owner string) *Account {\n\taccount := Account{owner: owner, balance: 0}\n\treturn &account\n}", "func NewAccount(id string) esfazz.Aggregate {\n\tacc := AccountModel()\n\tacc.Id = id\n\treturn acc\n}", "func CreateAccount(t *testing.T) (priv *rsa.PrivateKey, address string) {\n\tvar err error\n\tpriv, err = rsa.GenerateKey(rand.Reader, 1024)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\taddress, err = PrivateKeyToAddress(priv)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\treturn\n}", "func New() *Accounts {\n\treturn &Accounts{\n\t\tdata: make(map[string]*Account),\n\t}\n}", "func (dc *DexClient) CreateAccount(user, pass, email string) (*UserResponse, error) {\n\treturn dc.CreateAccountContext(context.Background(), user, pass, email)\n}", "func (kb *Keybase) CreateAccount(name, mnemonic, bip39Passwd, encryptPasswd, hdPath string, algo keys.SigningAlgo) (keys.Info, error) {\n\tkb.mx.Lock()\n\tdefer kb.mx.Unlock()\n\treturn kb.kb.CreateAccount(name, mnemonic, bip39Passwd, encryptPasswd, hdPath, algo)\n}", "func (ba *BankAccount) New() Resource {\n\tvar obj = &BankAccount{}\n\treturn obj\n}", "func (m *Account) Create() error {\n\treturn Db().Create(m).Error\n}", "func NewAccount() *Account {\n\tpriv, _ := ecdsa.GenerateKey(crypto.S256(), cryptorand.Reader)\n\tpub := base64.URLEncoding.EncodeToString(\n\t\tcrypto.FromECDSAPub(&priv.PublicKey))\n\taddr := hex.EncodeToString(crypto.PubkeyToAddress(priv.PublicKey).Bytes())\n\treturn &Account{\n\t\tID: NewUUID(),\n\t\tEthAddr: addr,\n\t\tPublicKey: pub,\n\t\tPrivateKey: base64.URLEncoding.EncodeToString(crypto.FromECDSA(priv)),\n\t}\n}", "func NewAccount(addr, amount string) *Account {\n\tam := big.NewInt(0)\n\tam.SetString(amount, 0)\n\treturn &Account{Address: addr, Amount: am}\n}", "func NewAccount(id, entityID string) *Account {\n\treturn &Account{\n\t\tID: id,\n\t\tEntityID: entityID,\n\t\tWallets: make([]Wallet, 0),\n\t}\n}", "func (a *Client) CreateAccount(params *CreateAccountParams) (*CreateAccountCreated, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewCreateAccountParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"createAccount\",\n\t\tMethod: \"POST\",\n\t\tPathPattern: \"/AccountService/Accounts\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &CreateAccountReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*CreateAccountCreated), nil\n\n}", "func makeAccount(){\n\toperatorSecret, err := hedera.SecretKeyFromString(secret)\n\tif err != nil{\n\t\tpanic(err)\n\t}\n\n\tsecretKey, _ := hedera.GenerateSecretKey()\n\tpublic := secretKey.Public()\n\n\tfmt.Printf(\"secret = %v\\n\", secretKey)\n\tfmt.Printf(\"public = %v\\n\", public)\n\n\tclient, err := hedera.Dial(server)\n\tif err !=nil{\n\t\tpanic(err)\n\t}\n\tdefer client.Close()\n\n\tnodeAccountID := hedera.AccountID{Account: 3}\n\toperatorAccountID := hedera.AccountID{Account: 1001}\n\ttime.Sleep(2* time.Second)\n\tresponse, err := client.CreateAccount().Key(public).InitialBalance(0).Operator(operatorAccountID).Node(nodeAccountID).Memo(\"Test make Account\").Sign(operatorSecret).Execute()\n\tif err != nil{\n\t\tpanic(err)\n\t}\n\n\ttransactionID := response.ID\n\tfmt.Printf(\"Created account; transaction = %v\\n\", transactionID)\n\ttime.Sleep(2* time.Second)\n \n\treceipt,err := client.Transaction(*transactionID).Receipt().Get()\n\tif err != nil{\n\t\tpanic(err)\n\t}\n\tfmt.Printf(\"Account = %v\\n\", *receipt.AccountID)\n\n}", "func (a AccountManager) CreateAccount(username, password string) (success bool, err error) {\n\tsuccess = true\n\treturn\n}", "func (c Client) NewAccount(privateKey crypto.Signer, onlyReturnExisting, termsOfServiceAgreed bool, contact ...string) (Account, error) {\n\tvar opts []NewAccountOptionFunc\n\tif onlyReturnExisting {\n\t\topts = append(opts, NewAcctOptOnlyReturnExisting())\n\t}\n\tif termsOfServiceAgreed {\n\t\topts = append(opts, NewAcctOptAgreeTOS())\n\t}\n\tif contact != nil && len(contact) > 0 {\n\t\topts = append(opts, NewAcctOptWithContacts(contact...))\n\t}\n\n\treturn c.NewAccountOptions(privateKey, opts...)\n}", "func NewAccount(instanceID uuid.UUID, name, aud string) (*Account, error) {\n\tid, err := uuid.NewV4()\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"Error generating unique id\")\n\t}\n\n\taccount := &Account{\n\t\tInstanceID: instanceID,\n\t\tID: id,\n\t\tAud: aud,\n\t\tName: name,\n\t}\n\treturn account, nil\n}", "func (s *Service) CreateAccount(name, description, key, secret, redirectURI string) (*Account, error) {\n\t// Check uniqueness of the name\n\taccount, err := s.FindAccountByName(name)\n\tif account != nil && err == nil {\n\t\treturn nil, ErrAccountNameTaken\n\t}\n\n\t// Check uniqueness of the key (client ID)\n\tif s.GetOauthService().ClientExists(key) {\n\t\treturn nil, oauth.ErrClientIDTaken\n\t}\n\n\t// Begin a transaction\n\ttx := s.db.Begin()\n\n\t// Create a new oauth client\n\toauthClient, err := s.GetOauthService().CreateClientTx(\n\t\ttx,\n\t\tkey,\n\t\tsecret,\n\t\tredirectURI,\n\t)\n\tif err != nil {\n\t\ttx.Rollback() // rollback the transaction\n\t\treturn nil, err\n\t}\n\n\t// Create a new account\n\taccount = NewAccount(oauthClient, name, description)\n\n\t// Save the account to the database\n\tif err := tx.Create(account).Error; err != nil {\n\t\ttx.Rollback() // rollback the transaction\n\t\treturn nil, err\n\t}\n\n\t// Assign related object\n\taccount.OauthClient = oauthClient\n\n\t// Commit the transaction\n\tif err := tx.Commit().Error; err != nil {\n\t\ttx.Rollback() // rollback the transaction\n\t\treturn nil, err\n\t}\n\n\treturn account, nil\n}", "func (s *StateDB) CreateAccount(addr types.AddressHash) {\n\tnew, prev := s.createObject(addr)\n\tif prev != nil {\n\t\tnew.setBalance(prev.data.Balance)\n\t}\n}", "func New(db *sql.DB, name, email, passwordDigest string) error {\n\n\terr := validateName(db, name)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = validateEmail(db, email)\n\tif err != nil {\n\t\treturn err\n\t}\n\tpasswordDigest, err = digestPassword(passwordDigest)\n\tif err != nil {\n\t\treturn err\n\t}\n\tcreatedAt := time.Now().UTC()\n\tactivationDigest, err := activationCode(passwordDigest)\n\tif err != nil {\n\t\treturn errors.New(\"Internal Error AC_45\")\n\t}\n\tstmt, err := db.Prepare(\"INSERT INTO users (name, email, password_digest, created_at, updated_at, activation_digest, activated) VALUES(?, ?, ?, ?, ?, ?, ?)\")\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = stmt.Exec(name, email, passwordDigest, createdAt, createdAt, activationDigest, 0)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func Create(account *Account) (Data, error) {\n\tpayload, err := json.Marshal(Data{Account: *account})\n\tif err != nil {\n\t\treturn Data{}, err\n\t}\n\n\tresponseStatus, responsePayload, err := doRequest(&request{\n\t\tmethod: \"POST\",\n\t\tresource: \"v1/organisation/accounts/\",\n\t\tpayload: payload,\n\t})\n\n\tif err != nil {\n\t\treturn Data{}, err\n\t}\n\n\treturn handleResponseData(responsePayload, http.StatusCreated, responseStatus)\n}", "func NewAccount(email, password string) (*Account, error) {\n\thashedPassword, err := hash(password)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Account{\n\t\tEmail: email,\n\t\tPassword: hashedPassword,\n\t\tVerified: false,\n\t\tVerificationID: uuid.New().String(),\n\t}, nil\n}", "func NewAccount(val string) AccountField {\n\treturn AccountField{quickfix.FIXString(val)}\n}", "func NewAccount(email string, password string) (*Account, error) {\n\taccount := &Account{Email: email}\n\tencryptedPassword, err := utils.Encrypt(password)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\taccount.ID = primitive.NewObjectID()\n\taccount.Password = string(encryptedPassword)\n\ttNow := time.Now()\n\taccount.CreatedAt = &tNow\n\taccount.repo = repoimpl.GetAccountRepo()\n\treturn account, nil\n}", "func createAccount(w http.ResponseWriter, req *http.Request) {\n\n\t// Redirect to home page if already signed in. We run getUserAndSession()\n\t// at the beginning of each route to capture user and session state\n\t// for processing.\n\tuserData, _ := getUserAndSession(w, req)\n\tif _, ok := dbUsers[userData.UserID]; ok {\n\t\thttp.Redirect(w, req, \"/\", http.StatusSeeOther)\n\t\treturn\n\t}\n\n\t// Process form submission\n\tif req.Method == http.MethodPost {\n\n\t\tuserID := req.FormValue(\"userid\")\n\t\tpassword := req.FormValue(\"password\")\n\t\tfirst := req.FormValue(\"first\")\n\t\tlast := req.FormValue(\"last\")\n\n\t\t// Check that the userID isn't already taken. If it is, reply to\n\t\t// the request with the specified error message and HTTP code.\n\t\tif _, ok := dbUsers[userID]; ok {\n\t\t\thttp.Error(w, \"Username already taken\", http.StatusForbidden)\n\t\t\treturn\n\t\t}\n\n\t\t// Ecrypt password before storing it. If the encryption errs, respond with\n\t\t// an Internal Server Error.\n\t\tencryptedPassword, err := getEncryptedPassword(password)\n\t\tif err != nil {\n\t\t\thttp.Error(w, \"Internal server error\", http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\n\t\t// Register user in dbUsers map\n\t\tnewuser := user{userID, first, last, encryptedPassword, false}\n\t\tdbUsers[userID] = newuser\n\n\t\t// Redirect to home page after signup\n\t\thttp.Redirect(w, req, \"/\", http.StatusSeeOther)\n\t\treturn\n\t}\n\n\t// Render page if not a POST\n\ttpl.ExecuteTemplate(w, \"createAccount.gohtml\", nil)\n}", "func CreateUser(id int, email string, pass string, bName string,\n\tbAccNum int, bRoutNum int) {\n\tvar count int = 0\n\n\tfor count < len(userList) {\n\t\tif userList[count].uID == id {\n\t\t\tfmt.Println(\"That user id is taken. Please choose a new ID.\")\n\t\t\treturn\n\t\t} else {\n\t\t\tcount++\n\t\t}\n\t}\n\n\tpANew := payAccount{\n\t\tbankName: bName,\n\t\taccountNumber: bAccNum,\n\t\troutingNumber: bRoutNum,\n\t}\n\tuNew := user{\n\t\tuID: id,\n\t\tuEmail: email,\n\t\tuPassword: pass,\n\t\tuBankAccount: pANew,\n\t}\n\tAddUserToDatabase(uNew)\n}", "func (s *Service) CreateAccount(budgetID string, accountPayload PayloadAccount) (*Account, error) {\n\tresModel := struct {\n\t\tData struct {\n\t\t\tAccount *Account `json:\"account\"`\n\t\t} `json:\"data\"`\n\t}{}\n\n\tpayload := struct {\n\t\tAccount PayloadAccount `json:\"account\"`\n\t}{\n\t\taccountPayload,\n\t}\n\n\tbuf, err := json.Marshal(&payload)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\turl := fmt.Sprintf(\"/budgets/%s/accounts/\", budgetID)\n\tif err := s.c.POST(url, &resModel, buf); err != nil {\n\t\treturn nil, err\n\t}\n\treturn resModel.Data.Account, nil\n}", "func (a *APIClient) Create(account AccountData) (created AccountData, err error) {\n\n\trel := &url.URL{Path: \"/v1/organisation/accounts\"}\n\turl := a.BaseURL.ResolveReference(rel)\n\n\tjsonPayload, err := json.Marshal(account)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn AccountData{}, err\n\t}\n\n\treq, err := http.NewRequest(\"POST\", url.String(), bytes.NewBuffer(jsonPayload))\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn AccountData{}, err\n\t}\n\n\tresp, err := a.HTTPClient.Do(req)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn AccountData{}, err\n\t}\n\n\tdefer resp.Body.Close()\n\n\tbody, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn AccountData{}, err\n\t}\n\n\terr = json.Unmarshal(body, &created)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn AccountData{}, err\n\t}\n\n\treturn created, nil\n}", "func (a *Account) CreateNameNew(name string) (*Transaction, []byte, error) {\n\tsb := txscript.NewScriptBuilder()\n\n\t// read a 64-bit random number to use as the salt\n\tbuffer := make([]byte, 8)\n\t_, err := io.ReadFull(rand.Reader, buffer)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\t// Hash the name with the salt.\n\tnameBy := []byte(name)\n\thash := btcutil.Hash160(append(buffer, nameBy...))\n\n\t// Build the name script\n\tsb.AddOp(txscript.OP_1).AddData(hash).AddOp(txscript.OP_2DROP)\n\n\ttx, err := a.finishNameTransaction(name, sb, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treturn tx, buffer, err\n}", "func NewAccount(address string) *Account {\n\treturn &Account{\n\t\tAddress: address,\n\t\tHeight: \"0\",\n\t\tGoldTokenBalance: \"0\",\n\t\tTotalLockedGoldBalance: \"0\",\n\t\tNonVotingLockedGoldBalance: \"0\",\n\t\tVotingLockedGoldBalance: \"0\",\n\t\tPendingWithdrawalBalance: \"0\",\n\t\tCeloUSDValue: \"0\",\n\t\tDelegations: []*Delegation{},\n\t}\n\n}", "func createAccount(gm *gomatrix.Client) (accessToken, userID string, err error) {\n username := \"testing-\" + randString(5)\n // Get the session token\n req := &gomatrix.ReqRegister{\n Username: username,\n Password: testPass,\n }\n _, respInt, err := gm.Register(req)\n if err != nil {\n return\n }\n\n // Make a dummy register request\n req = &gomatrix.ReqRegister{\n Username: username,\n Password: testPass,\n Auth: struct {\n Session string\n }{\n Session: respInt.Session,\n },\n }\n resp, err := gm.RegisterDummy(req)\n if err != nil {\n return\n }\n\n // Save the access token and UserID\n accessToken = resp.AccessToken\n userID = resp.UserID\n return\n}", "func (repo *Repository) Create(ctx context.Context, claims auth.Claims, req UserAccountCreateRequest, now time.Time) (*UserAccount, error) {\n\tspan, ctx := tracer.StartSpanFromContext(ctx, \"internal.user_account.Create\")\n\tdefer span.Finish()\n\n\t// Validate the request.\n\tv := webcontext.Validator()\n\terr := v.Struct(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Ensure the claims can modify the account specified in the request.\n\terr = repo.CanModifyAccount(ctx, claims, req.AccountID)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// If now empty set it to the current time.\n\tif now.IsZero() {\n\t\tnow = time.Now()\n\t}\n\n\t// Always store the time as UTC.\n\tnow = now.UTC()\n\n\t// Postgres truncates times to milliseconds when storing. We and do the same\n\t// here so the value we return is consistent with what we store.\n\tnow = now.Truncate(time.Millisecond)\n\n\t// Check to see if there is an existing user account, including archived.\n\texistQuery := selectQuery()\n\texistQuery.Where(existQuery.And(\n\t\texistQuery.Equal(\"account_id\", req.AccountID),\n\t\texistQuery.Equal(\"user_id\", req.UserID),\n\t))\n\texisting, err := find(ctx, claims, repo.DbConn, existQuery, []interface{}{}, true)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// If there is an existing entry, then update instead of insert.\n\tvar ua UserAccount\n\tif len(existing) > 0 {\n\t\tupReq := UserAccountUpdateRequest{\n\t\t\tUserID: req.UserID,\n\t\t\tAccountID: req.AccountID,\n\t\t\tRoles: &req.Roles,\n\t\t\tunArchive: true,\n\t\t}\n\t\terr = repo.Update(ctx, claims, upReq, now)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tua = *existing[0]\n\t\tua.Roles = req.Roles\n\t\tua.UpdatedAt = now\n\t\tua.ArchivedAt = nil\n\t} else {\n\t\tuaID := uuid.NewRandom().String()\n\n\t\tua = UserAccount{\n\t\t\t//ID: uaID,\n\t\t\tUserID: req.UserID,\n\t\t\tAccountID: req.AccountID,\n\t\t\tRoles: req.Roles,\n\t\t\tStatus: UserAccountStatus_Active,\n\t\t\tCreatedAt: now,\n\t\t\tUpdatedAt: now,\n\t\t}\n\n\t\tif req.Status != nil {\n\t\t\tua.Status = *req.Status\n\t\t}\n\n\t\t// Build the insert SQL statement.\n\t\tquery := sqlbuilder.NewInsertBuilder()\n\t\tquery.InsertInto(userAccountTableName)\n\t\tquery.Cols(\"id\", \"user_id\", \"account_id\", \"roles\", \"status\", \"created_at\", \"updated_at\")\n\t\tquery.Values(uaID, ua.UserID, ua.AccountID, ua.Roles, ua.Status.String(), ua.CreatedAt, ua.UpdatedAt)\n\n\t\t// Execute the query with the provided context.\n\t\tsql, args := query.Build()\n\t\tsql = repo.DbConn.Rebind(sql)\n\t\t_, err = repo.DbConn.ExecContext(ctx, sql, args...)\n\t\tif err != nil {\n\t\t\terr = errors.Wrapf(err, \"query - %s\", query.String())\n\t\t\terr = errors.WithMessagef(err, \"add account %s to user %s failed\", req.AccountID, req.UserID)\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn &ua, nil\n}" ]
[ "0.7277593", "0.70907176", "0.7073786", "0.69714516", "0.6833372", "0.681707", "0.6791376", "0.6759559", "0.6725389", "0.668418", "0.66642153", "0.66445816", "0.66366863", "0.6588414", "0.6569843", "0.6567917", "0.65641284", "0.6529776", "0.6523798", "0.6477096", "0.64761114", "0.64688593", "0.64646465", "0.64593244", "0.6421299", "0.6396168", "0.6382722", "0.6373423", "0.6357339", "0.6332455", "0.6313364", "0.63130635", "0.63130635", "0.63046235", "0.6303551", "0.6302528", "0.6295307", "0.6264435", "0.62402254", "0.62328124", "0.6227253", "0.6222935", "0.6210288", "0.62052506", "0.6194088", "0.6185462", "0.6165076", "0.6163827", "0.6162172", "0.61603403", "0.61509573", "0.61506206", "0.6142831", "0.613685", "0.61277837", "0.6126556", "0.612428", "0.6104761", "0.6101935", "0.6088537", "0.60848844", "0.6082885", "0.6078512", "0.6076024", "0.60619015", "0.60320807", "0.6029069", "0.6022338", "0.6003041", "0.5998748", "0.59897304", "0.59712666", "0.5969249", "0.59103036", "0.5906122", "0.58899575", "0.5889636", "0.5855763", "0.58405024", "0.583969", "0.58385384", "0.5832726", "0.583188", "0.58172387", "0.5816115", "0.5809304", "0.58076775", "0.580547", "0.58039844", "0.57995754", "0.5798836", "0.5798439", "0.57905877", "0.57709974", "0.5769224", "0.5754645", "0.57510173", "0.57474065", "0.5731949", "0.5729429" ]
0.6330037
30
ShowAccount shows a, A, AT.
func (account *Account) ShowAccount() { fmt.Println("a:", account.a) fmt.Println("A:", account.A) fmt.Println("AT:", account.AT) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (r Runner) ShowAccount(ctx context.Context, name string) (Account, error) {\n\tb := &bytes.Buffer{}\n\n\topt := []step.Option{\n\t\tr.cc.ShowKeyAddressCommand(name),\n\t}\n\n\tif r.cc.KeyringPassword != \"\" {\n\t\tinput := &bytes.Buffer{}\n\t\tfmt.Fprintln(input, r.cc.KeyringPassword)\n\t\topt = append(opt, step.Write(input.Bytes()))\n\t}\n\n\tif err := r.run(ctx, runOptions{stdout: b}, opt...); err != nil {\n\t\tif strings.Contains(err.Error(), \"item could not be found\") ||\n\t\t\tstrings.Contains(err.Error(), \"not a valid name or address\") {\n\t\t\treturn Account{}, ErrAccountDoesNotExist\n\t\t}\n\t\treturn Account{}, err\n\t}\n\n\treturn Account{\n\t\tName: name,\n\t\tAddress: strings.TrimSpace(b.String()),\n\t}, nil\n}", "func (c *AccountController) Show(ctx *app.ShowAccountContext) error {\n\tu, err := repositories.GetUserByID(ctx.AccountID)\n\tif err != nil {\n\t\treturn ctx.NotFound()\n\t}\n\tres := &app.FtAccount{\n\t\tID: u.ID,\n\t\tFirstName: u.FirstName,\n\t\tLastName: u.LastName,\n\t\tEmail: u.Email,\n\t}\n\treturn ctx.OK(res)\n}", "func (c *AccountController) Show(ctx *app.ShowAccountContext) error {\n\tres := &app.Account{}\n\treturn ctx.OK(res, \"default\")\n}", "func displayAccount(w io.Writer) {\n\taccount := doAccount()\n\tif account.Status != \"active\" {\n\t\tlog.Errorf(\"DO Account issue expected status (active) got (%s) \", account.Status)\n\t\treturn\n\t}\n\tfmt.Fprintf(w, account.String())\n}", "func (a *Account) Show() (*Account, *http.Response, []error) {\n\tk := kumoru.New()\n\n\tk.Get(fmt.Sprintf(\"%v/v1/accounts/%v\", k.EndPoint.Authorization, a.Email))\n\tk.SignRequest(true)\n\n\tresp, body, errs := k.End()\n\n\tif len(errs) > 0 {\n\t\treturn a, resp, errs\n\t}\n\n\tif resp.StatusCode >= 400 {\n\t\terrs = append(errs, fmt.Errorf(\"%s\", resp.Status))\n\t\treturn a, resp, errs\n\t}\n\n\terr := json.Unmarshal([]byte(body), &a)\n\n\tif err != nil {\n\t\terrs = append(errs, err)\n\t\treturn a, resp, errs\n\t}\n\n\treturn a, resp, nil\n}", "func (a Account) ShowOwner() string {\n\treturn a.owner\n}", "func displayAccountAlgoBalance(account string, client *algod.Client) {\n\taccountInfo, err := client.AccountInformation(account).Do(context.Background())\n\tif err != nil {\n\t\tfmt.Printf(\"failed to get account info: %v\\n\", err)\n\t\treturn\n\t}\n\tfmt.Printf(\"%s: %v microAlgos\\n\", accountInfo.Address, accountInfo.Amount)\n}", "func (c *Jrpc) ShowPrivacyAccountInfo(in *pty.ReqPrivacyAccount, result *json.RawMessage) error {\n\treply, err := c.cli.ExecWalletFunc(pty.PrivacyX, \"ShowPrivacyAccountInfo\", in)\n\tif err != nil {\n\t\treturn err\n\t}\n\t*result, err = types.PBToJSON(reply)\n\treturn err\n}", "func (a Account) ShowBalance() int {\n\treturn a.balance\n}", "func (a Account) String() string {\n\treturn fmt.Sprint(\"owner : \", a.owner, \", balance : \", a.balance)\n}", "func (c *Jrpc) ShowPrivacyAccountSpend(in *pty.ReqPrivBal4AddrToken, result *json.RawMessage) error {\n\tif 0 == len(in.Addr) {\n\t\treturn types.ErrInvalidParam\n\t}\n\treply, err := c.cli.ExecWalletFunc(pty.PrivacyX, \"ShowPrivacyAccountSpend\", in)\n\tif err != nil {\n\t\tlog.Info(\"ShowPrivacyAccountSpend\", \"return err info\", err)\n\t\treturn err\n\t}\n\t*result, err = types.PBToJSON(reply)\n\treturn err\n}", "func (a *Account) ToString() string {\n\treturn a.ID + \" \" + a.Name\n}", "func (t *TezTracker) GetAccount(id string) (acc models.AccountListView, err error) {\n\tr := t.repoProvider.GetAccount()\n\n\tfilter := models.Account{AccountID: null.StringFrom(id)}\n\n\tfound, acc, err := r.Find(filter)\n\tif err != nil {\n\t\treturn acc, err\n\t}\n\tif !found {\n\t\treturn acc, ErrNotFound\n\t}\n\n\tcounts, err := t.repoProvider.GetOperation().AccountOperationCount(acc.AccountID.String)\n\tif err != nil {\n\t\treturn acc, err\n\t}\n\n\tvar total int64\n\tfor i := range counts {\n\t\tif counts[i].Kind == \"transaction\" {\n\t\t\tacc.Transactions = counts[i].Count\n\t\t}\n\t\tif counts[i].Kind == \"reveal\" {\n\t\t\tacc.IsRevealed = true\n\t\t}\n\n\t\ttotal += counts[i].Count\n\t}\n\n\tacc.Operations = total\n\n\tbi, err := t.GetBakerInfo(id)\n\tif err != nil {\n\t\treturn acc, err\n\t}\n\n\tacc.BakerInfo = bi\n\n\t//Account identified as baker\n\tif bi != nil {\n\t\t//Set real value for front\n\t\tacc.IsBaker = true\n\t}\n\n\treturn acc, nil\n}", "func (t *TezTracker) GetAccount(id string) (acc models.AccountListView, err error) {\n\tr := t.repoProvider.GetAccount()\n\n\tfilter := models.Account{AccountID: null.StringFrom(id)}\n\n\tfound, acc, err := r.Find(filter)\n\tif err != nil {\n\t\treturn acc, err\n\t}\n\tif !found {\n\t\treturn acc, ErrNotFound\n\t}\n\n\tcounts, err := t.repoProvider.GetOperation().AccountOperationCount(acc.AccountID.String)\n\tif err != nil {\n\t\treturn acc, err\n\t}\n\n\tvar total int64\n\tfor i := range counts {\n\t\tif counts[i].Kind == \"transaction\" {\n\t\t\tacc.Transactions = counts[i].Count\n\t\t}\n\t\tif counts[i].Kind == \"reveal\" {\n\t\t\tacc.IsRevealed = true\n\t\t}\n\n\t\ttotal += counts[i].Count\n\t}\n\n\tacc.Operations = total\n\n\tbi, err := t.GetBakerInfo(id)\n\tif err != nil {\n\t\treturn acc, err\n\t}\n\n\tacc.BakerInfo = bi\n\n\t//Account identified as baker\n\tif bi != nil {\n\t\t//Set real value for front\n\t\tacc.IsBaker = true\n\t}\n\n\treturn acc, nil\n}", "func (h *UserRepos) Account(ctx context.Context, w http.ResponseWriter, r *http.Request, params map[string]string) error {\n\n\tdata := make(map[string]interface{})\n\tf := func() error {\n\n\t\tclaims, err := auth.ClaimsFromContext(ctx)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tacc, err := h.AccountRepo.ReadByID(ctx, claims, claims.Audience)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdata[\"account\"] = acc.Response(ctx)\n\n\t\treturn nil\n\t}\n\n\tif err := f(); err != nil {\n\t\treturn web.RenderError(ctx, w, r, err, h.Renderer, TmplLayoutBase, TmplContentErrorGeneric, web.MIMETextHTMLCharsetUTF8)\n\t}\n\n\treturn h.Renderer.Render(ctx, w, r, TmplLayoutBase, \"user-account.gohtml\", web.MIMETextHTMLCharsetUTF8, http.StatusOK, data)\n}", "func Account(client *ticketmatic.Client) (*ticketmatic.AccountInfo, error) {\n\tr := client.NewRequest(\"GET\", \"/{accountname}/tools/account\", \"json\")\n\n\tvar obj *ticketmatic.AccountInfo\n\terr := r.Run(&obj)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn obj, nil\n}", "func AchAccount() string {\n\treturn Numerify(\"############\")\n}", "func (a *Funcs) Account() (string, error) {\n\ta.stsInit.Do(a.initSTS)\n\treturn a.sts.Account()\n}", "func (account *Account) ShowWalletIndex() {\n\tfmt.Println(\"\\033[1;31;40m\", account.WalletIndex, \"\\033[0m\")\n\n}", "func (account *Account) ShowTx(P string) bool {\n\tfor i := 0; i < account.WalletIndex; i++ {\n\t\tif P == account.Wallet[i].P {\n\t\t\tJSON, _ := json.MarshalIndent(account.Wallet[i], \"\", \" \")\n\n\t\t\tfmt.Println(string(JSON))\n\t\t\treturn true\n\n\t\t}\n\t}\n\treturn false\n}", "func ShowAccountPrivacyInfo() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"keys\",\n\t\tShort: \"get account privacy keys for mix note\",\n\t\tRun: accountPrivacy,\n\t}\n\taccountPrivacyCmdFlags(cmd)\n\treturn cmd\n}", "func (gb *CurrentGrantBuilder) Show() string {\n\treturn fmt.Sprintf(`SHOW GRANTS ON %v %v`, gb.grantType, gb.qualifiedName)\n}", "func (ge *CurrentGrantExecutable) Show() string {\n\treturn fmt.Sprintf(`SHOW GRANTS OF %v \"%v\"`, ge.granteeType, ge.granteeName)\n}", "func (dstv Dstv) AccountInfo(accountNumber string,\n\tcountry ...string) (*AccountInfoResponse, error) {\n\n\tcustomerNumber, smartCardNumber := \"\", \"\"\n\taccountNumberType(accountNumber, &customerNumber, &smartCardNumber)\n\n\tif len(country) > 0 {\n\t\tdstv.AddQueryData(paymenex.PCountry, country[0])\n\t} else {\n\t\tdstv.AddQueryData(paymenex.PCountry, \"GH\")\n\t}\n\tdstv.AddQueryData(paymenex.PActId, \"LKDCT\")\n\tdstv.AddQueryData(paymenex.PDstvCustomerNo, customerNumber)\n\tdstv.AddQueryData(paymenex.PDstvSmartCardNo, smartCardNumber)\n\txml, err := dstv.MakeRequest()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// writeFile(\"account-info.xml\", xml) // DEBUG\n\tresponse := new(AccountInfoResponse)\n\tok := dstv.ParseAndVerifyResponse(xml, response)\n\tif !ok {\n\t\treturn response, errors.New(errVerifyMsg)\n\t}\n\treturn response, nil\n}", "func (a *Account) Name() string { return a.name }", "func printBalance(account horizon.Account) {\n\taddress := account.AccountID\n\tbalance := getBalance(account)\n\n\tfmt.Println(\"Balances for account:\", address)\n\tfmt.Println(\"XLM Balance:\", balance)\n}", "func (s Account) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s Account) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (t *TezTracker) AccountList(before string, limits Limiter, favorites []string) (accs []models.AccountListView, count int64, err error) {\n\tr := t.repoProvider.GetAccount()\n\tfilter := models.AccountFilter{\n\t\tType: models.AccountTypeAccount,\n\t\tOrderBy: models.AccountOrderFieldCreatedAt,\n\t\tAfter: before,\n\t\tFavorites: favorites,\n\t}\n\tcount, accs, err = r.List(limits.Limit(), limits.Offset(), filter)\n\treturn accs, count, err\n}", "func (t *TezTracker) AccountList(before string, limits Limiter) (accs []models.AccountListView, count int64, err error) {\n\tr := t.repoProvider.GetAccount()\n\tfilter := models.AccountFilter{\n\t\tType: models.AccountTypeAccount,\n\t\tOrderBy: models.AccountOrderFieldCreatedAt,\n\t\tAfter: before,\n\t}\n\tcount, accs, err = r.List(limits.Limit(), limits.Offset(), filter)\n\treturn accs, count, err\n}", "func Account(name, mnemonic string) Option {\n\treturn func(f *Faucet) {\n\t\tf.accountName = name\n\t\tf.accountMnemonic = mnemonic\n\t}\n}", "func (c *TenantController) Show(ctx *app.ShowTenantContext) error {\n\ttoken := goajwt.ContextJWT(ctx)\n\tif token == nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewUnauthorizedError(\"Missing JWT token\"))\n\t}\n\n\tttoken := &TenantToken{token: token}\n\ttenantID := ttoken.Subject()\n\ttenant, err := c.tenantService.GetTenant(tenantID)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewNotFoundError(\"tenants\", tenantID.String()))\n\t}\n\n\tnamespaces, err := c.tenantService.GetNamespaces(tenantID)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, err)\n\t}\n\n\tresult := &app.TenantSingle{Data: convertTenant(ctx, tenant, namespaces, c.resolveCluster)}\n\treturn ctx.OK(result)\n}", "func (c *TenantController) Show(ctx *app.ShowTenantContext) error {\n\tuserToken := goajwt.ContextJWT(ctx)\n\tif userToken == nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewUnauthorizedError(\"Missing JWT token\"))\n\t}\n\n\tttoken := &auth.TenantToken{Token: userToken}\n\ttenantID := ttoken.Subject()\n\ttenant, err := c.tenantService.GetTenant(tenantID)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewNotFoundError(\"tenants\", tenantID.String()))\n\t}\n\n\tnamespaces, err := c.tenantService.GetNamespaces(tenantID)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, err)\n\t}\n\n\tresult := &app.TenantSingle{Data: convertTenant(ctx, tenant, namespaces, c.clusterService.GetCluster)}\n\treturn ctx.OK(result)\n}", "func (c *Client) Account(account string) (*response.Account, error) {\n\tvar data *d3.Account\n\n\tep := endpointAccount(c.region, account)\n\n\tq, err := c.get(ep, &data)\n\n\tif nil != err {\n\t\treturn nil, err\n\t}\n\n\treturn &response.Account{\n\t\tData: data,\n\t\tEndpoint: ep,\n\t\tQuota: q,\n\t\tRegion: c.region,\n\t}, nil\n}", "func (o *NumbersACH) GetAccount() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.Account\n}", "func Show(ctx context.Context) (string, bool) {\n\tacc, ok := ctx.Value(oauthToken).(*access)\n\tif !ok {\n\t\treturn \"\", ok\n\t}\n\treturn acc.Login, true\n}", "func (s AccountInfo) String() string {\n\treturn awsutil.Prettify(s)\n}", "func GetAccount(w http.ResponseWriter, r *http.Request) {\n\tlogin := mux.Vars(r)[\"login\"]\n\n\taccount, ok := data.GetAccountByLogin(login)\n\tif !ok {\n\t\tPrintErrorJSON(w, r, \"The requested account does not exist\", http.StatusNotFound)\n\t\treturn\n\t}\n\n\tisAdmin := false\n\tisOwner := false\n\tif oauth, ok := OAuthToken(r); ok {\n\t\tisAdmin = oauth.Match.Contains(\"account-admin\")\n\t\tisOwner = oauth.Token.AccountUUID.String == account.UUID && oauth.Match.Contains(\"account-write\")\n\t}\n\n\tmarshal := &data.AccountMarshaler{\n\t\tWithMail: account.IsEmailPublic || isOwner || isAdmin,\n\t\tWithAffiliation: account.IsAffiliationPublic || isOwner || isAdmin,\n\t\tAccount: account,\n\t}\n\n\tw.Header().Add(\"Cache-Control\", \"no-cache\")\n\tw.Header().Add(\"Content-Type\", \"application/json\")\n\tenc := json.NewEncoder(w)\n\terr := enc.Encode(marshal)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func (s *Service) asoAccountInfo(c context.Context, userid string) (info *model.AsoAccount, err error) {\n\tvar acs []*model.AsoAccount\n\tif acs, err = s.d.AsoAccount(c, userid, model.DefaultHash(userid)); err != nil {\n\t\treturn\n\t}\n\tif len(acs) == 0 {\n\t\terr = ecode.UserNotExist\n\t\treturn\n\t}\n\tif len(acs) > 1 {\n\t\terr = ecode.UserDuplicate\n\t\treturn\n\t}\n\tinfo = acs[0]\n\treturn\n}", "func (o AdminAccountOutput) AccountId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *AdminAccount) pulumi.StringOutput { return v.AccountId }).(pulumi.StringOutput)\n}", "func (c *Client) AccountInfo() (AccInfo, error) {\n\tdata, err := c.apiCall(\"account.info\", nil)\n\tif err != nil {\n\t\treturn AccInfo{}, err\n\t}\n\n\tvar info AccInfo\n\terr = json.Unmarshal(data, &info)\n\tif err != nil {\n\t\treturn AccInfo{}, err\n\t}\n\n\treturn info, nil\n}", "func (digitalocean *DigitalOcean) Account() (*godo.Account, error) {\n\tclient, err := DigitalOceanClient()\n\n\taccount, _, err := client.client.Account.Get(client.context)\n\n\treturn account, err\n}", "func (cmd *ShowAccountCommand) Run(c *client.Client, args []string) error {\n\tvar path string\n\tif len(args) > 0 {\n\t\tpath = args[0]\n\t} else {\n\t\tpath = fmt.Sprintf(\"/api/accounts/%v\", url.QueryEscape(cmd.User))\n\t}\n\tlogger := goa.NewLogger(log.New(os.Stderr, \"\", log.LstdFlags))\n\tctx := goa.WithLogger(context.Background(), logger)\n\tresp, err := c.ShowAccount(ctx, path)\n\tif err != nil {\n\t\tgoa.LogError(ctx, \"failed\", \"err\", err)\n\t\treturn err\n\t}\n\n\tgoaclient.HandleResponse(c.Client, resp, cmd.PrettyPrint)\n\treturn nil\n}", "func (gw2 *GW2Api) Account() (acc Account, err error) {\n\tver := \"v2\"\n\ttag := \"account\"\n\terr = gw2.fetchAuthenticatedEndpoint(ver, tag, PermAccount, nil, &acc)\n\treturn\n}", "func (o DelegatedAdminAccountOutput) AccountId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *DelegatedAdminAccount) pulumi.StringOutput { return v.AccountId }).(pulumi.StringOutput)\n}", "func (tc *TransactionsController) Show(c *gin.Context) {\n\thash := common.HexToHash(c.Param(\"TxHash\"))\n\tif tx, err := tc.App.GetStore().FindTxByAttempt(hash); err == orm.ErrorNotFound {\n\t\tpublicError(c, http.StatusNotFound, errors.New(\"Transaction not found\"))\n\t} else if err != nil {\n\t\tc.AbortWithError(http.StatusInternalServerError, err)\n\t} else if doc, err := jsonapi.Marshal(presenters.NewTx(tx)); err != nil {\n\t\tc.AbortWithError(http.StatusInternalServerError, err)\n\t} else {\n\t\tc.Data(http.StatusOK, MediaType, doc)\n\t}\n}", "func (c *Client) ShowAccountbbb(ctx context.Context, path string) (*http.Response, error) {\n\treq, err := c.NewShowAccountbbbRequest(ctx, path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn c.Client.Do(ctx, req)\n}", "func (o EntitlementOutput) BillingAccount() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Entitlement) pulumi.StringOutput { return v.BillingAccount }).(pulumi.StringOutput)\n}", "func (sc Funcs) Accounts(ctx wasmlib.ScViewClientContext) *AccountsCall {\n\tf := &AccountsCall{Func: wasmlib.NewScView(ctx, HScName, HViewAccounts)}\n\twasmlib.NewCallResultsProxy(f.Func, &f.Results.Proxy)\n\treturn f\n}", "func ShowAccountNoteInfo() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"notes\",\n\t\tShort: \"show account notes\",\n\t\tRun: accountNote,\n\t}\n\taccountNoteCmdFlags(cmd)\n\treturn cmd\n}", "func (o AggregatorAggregatorAccountOutput) AccountType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v AggregatorAggregatorAccount) string { return v.AccountType }).(pulumi.StringOutput)\n}", "func ListAccounts(w http.ResponseWriter, r *http.Request) {\n\tisAdmin := false\n\tif oauth, ok := OAuthToken(r); ok {\n\t\tisAdmin = oauth.Match.Contains(\"account-admin\")\n\t}\n\n\tvar accounts []data.Account\n\tsearch := r.URL.Query().Get(\"q\")\n\tif search != \"\" {\n\t\taccounts = data.SearchAccounts(search)\n\t} else {\n\t\taccounts = data.ListAccounts()\n\t}\n\n\tmarshal := make([]data.AccountMarshaler, 0, len(accounts))\n\tfor i := 0; i < len(accounts); i++ {\n\t\tacc := &accounts[i]\n\t\tmarshal = append(marshal, data.AccountMarshaler{\n\t\t\tWithMail: isAdmin || acc.IsEmailPublic,\n\t\t\tWithAffiliation: isAdmin || acc.IsAffiliationPublic,\n\t\t\tAccount: acc,\n\t\t})\n\t}\n\n\tw.Header().Add(\"Cache-Control\", \"no-cache\")\n\tw.Header().Add(\"Content-Type\", \"application/json\")\n\tenc := json.NewEncoder(w)\n\terr := enc.Encode(marshal)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func (e Account) String() string { return ent.EntString(&e) }", "func (e Account) String() string { return ent.EntString(&e) }", "func (c *GethClient) Accounts(ctx context.Context) ([]string, error) {\n\tvar result []string\n\terr := c.rpcCli.CallContext(ctx, &result, \"personal_listAccounts\")\n\treturn result, err\n}", "func (t *ATrains) Show() {\n\tfmt.Println(\"wszystkie pociagi\")\n\n\tfor i := 0; i < len(t.trains); i++ {\n\t\tt.trains[i].Show()\n\t}\n\n\tfmt.Println(\"\")\n}", "func (o PartnerOutput) AccountId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Partner) pulumi.StringOutput { return v.AccountId }).(pulumi.StringOutput)\n}", "func (this *FamilyAccount) showDetails() {\n\tfmt.Println(\"------My Income and Expense Detail-------\")\n\tif this.flag {\n\t\t//因为我们用的是FamilyAccount结构体里传过来的字段,所以不能直接yongflag, 要用this. , 表示调用这个方法的结构体变量里面的字段\n\t\tfmt.Println(this.details)\n\t} else {\n\t\tfmt.Println(\"No current income and expenditure details!\")\n\t}\n}", "func ShowPerson(aPerson *Person) {\n\tfmt.Printf(\"Name: %s\\n\", aPerson.Name)\n\tfmt.Printf(\"Address: %s\\n\", aPerson.Address)\n\tfmt.Printf(\"Phone: %d\\n\", aPerson.Phone)\n}", "func (o MemberOutput) AccountId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Member) pulumi.StringOutput { return v.AccountId }).(pulumi.StringOutput)\n}", "func GetAccount(db gorm.DB, account_id int)(*AccountView, error) {\n\n\tfmt.Println(\"account_id=\", account_id)\n\tvar row *AccountView = new(AccountView)\n\tdb.Table(ACCOUNT_VIEW).Select(ACCOUNT_VIEW_COLS).Where(\"account_id = ?\", account_id).Scan(row)\n\n\treturn row, nil\n}", "func (p *Player) AccountName() string {\n\treturn p.Account\n}", "func (o AccessPointOutput) AccountId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *AccessPoint) pulumi.StringOutput { return v.AccountId }).(pulumi.StringOutput)\n}", "func index(c echo.Context) error {\n\treturn c.String(http.StatusOK, \"Account service\")\n}", "func (c *PartyController) Show(ctx *app.ShowPartyContext) error {\n\t// PartyController_Show: start_implement\n\n\tp, err := c.P.PinService().Party(pinbase.Hash(ctx.PartyHash))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tres := &app.PinbaseParty{\n\t\tHash: string(p.ID),\n\t\tDescription: p.Description,\n\t}\n\n\t// PartyController_Show: end_implement\n\treturn ctx.OK(res)\n}", "func (o LookupSnapshotPolicyResultOutput) AccountName() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupSnapshotPolicyResult) string { return v.AccountName }).(pulumi.StringOutput)\n}", "func ShowTenant(ctx context.Context, config tenantConfig) (*tenant.TenantSingle, error) {\n\n\tc, err := createClient(ctx, config)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tres, err := c.ShowTenant(goasupport.ForwardContextRequestID(ctx), tenant.ShowTenantPath())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif res.StatusCode == http.StatusOK {\n\t\ttenant, err := c.DecodeTenantSingle(res)\n\t\tif err != nil {\n\t\t\treturn nil, errors.NewInternalError(ctx, err)\n\t\t}\n\t\treturn tenant, nil\n\t} else if res.StatusCode > 400 {\n\t\tjsonErr, err := c.DecodeJSONAPIErrors(res)\n\t\tif err == nil {\n\t\t\tif len(jsonErr.Errors) > 0 {\n\t\t\t\treturn nil, errors.NewInternalError(ctx, fmt.Errorf(jsonErr.Errors[0].Detail))\n\t\t\t}\n\t\t}\n\t}\n\treturn nil, errors.NewInternalError(ctx, fmt.Errorf(\"Unknown response \"+res.Status))\n}", "func doAccount() (a *godo.Account) {\n\tctx := context.TODO()\n\ta, _, err := client.Account.Get(ctx)\n\tpanicIfError(err)\n\n\t// Check if something may be wrong with the account\n\tif a.Status != \"active\" {\n\t\tlog.Errorf(\"URGENT! expected account status (active) got (%s)\", a.Status)\n\t}\n\treturn a\n}", "func (o GetUsersUserOutput) AccountName() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetUsersUser) string { return v.AccountName }).(pulumi.StringOutput)\n}", "func GetAccount(w http.ResponseWriter, r *http.Request) {\n\temail := mux.Vars(r)[\"email\"]\n\n\tacc, err := models.GetAccount(email)\n\n\tif err != nil {\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\t_, _ = fmt.Fprintf(w, err.Error())\n\t\treturn\n\t}\n\n\tif reflect.DeepEqual(models.Account{}, acc) {\n\t\tw.WriteHeader(http.StatusNotFound)\n\t\treturn\n\t}\n\t_ = json.NewEncoder(w).Encode(acc)\n}", "func (r Virtual_Guest) GetAccount() (resp datatypes.Account, err error) {\n\terr = r.Session.DoRequest(\"SoftLayer_Virtual_Guest\", \"getAccount\", nil, &r.Options, &resp)\n\treturn\n}", "func viewAccounts(ctx iscp.SandboxView) (dict.Dict, error) {\n\treturn getAccountsIntern(ctx.State()), nil\n}", "func main() {\n\taccount := accounts.NewAccount(\"hskimim\")\n\taccount.Deposit(10)\n\terr := account.Withdraw(100)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\tfmt.Println(account)\n}", "func GetAccount(resp http.ResponseWriter, req *http.Request) {\n\tvars := mux.Vars(req)\n\n\tresp.Header().Add(\"Content-Type\", \"application/json\")\n\n\t//test if param is a number\n\tid, err := strconv.Atoi(vars[\"id\"])\n\tif err != nil {\n\t\tapiErr := &utils.ApplicationError{\n\t\t\tMessage: \"account_id must be a number/cannot be empty\",\n\t\t\tStatusCode: http.StatusBadRequest,\n\t\t\tCode: \"bad_request\",\n\t\t}\n\n\t\tjsonValue, _ := json.Marshal(apiErr)\n\t\tresp.WriteHeader(apiErr.StatusCode)\n\t\tresp.Write(jsonValue)\n\n\t\treturn\n\t}\n\n\taccount, apiErr := service.GetAccount(int64(id))\n\n\tif apiErr != nil {\n\n\t\tjsonValue, _ := json.Marshal(apiErr)\n\t\tresp.WriteHeader(apiErr.StatusCode)\n\t\tresp.Write(jsonValue)\n\n\t\treturn\n\t}\n\n\tresp.WriteHeader(http.StatusOK)\n\n\tencoder := json.NewEncoder(resp)\n\tencoder.Encode(account)\n\n}", "func (o *ContentProviderReadDetailed) GetAccount() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.Account\n}", "func getAccount(c echo.Context) error {\n\tvar errResp ErrorResponseData\n\tvar resp UserResponseData\n\n\tid := strings.TrimSpace(c.Param(\"id\"))\n\tif len(id) == 0 {\n\t\terrResp.Data.Code = \"invalid_param_error\"\n\t\terrResp.Data.Description = \"Value for account id not set in request\"\n\t\terrResp.Data.Status = strconv.Itoa(http.StatusBadRequest)\n\t\treturn c.JSON(http.StatusBadRequest, errResp)\n\t}\n\n\taccount, err := storage.GetAccount(id)\n\n\tif err != nil {\n\t\terrResp.Data.Code = \"get_account_error\"\n\t\terrResp.Data.Description = \"Unable to fetch account details\"\n\t\terrResp.Data.Status = strconv.Itoa(http.StatusInternalServerError)\n\t\treturn c.JSON(http.StatusInternalServerError, errResp)\n\t}\n\n\tif account == nil {\n\t\terrResp.Data.Code = \"no_account_found\"\n\t\terrResp.Data.Description = \"No account with id \" + id + \" exists\"\n\t\terrResp.Data.Status = strconv.Itoa(http.StatusNotFound)\n\t\treturn c.JSON(http.StatusNotFound, errResp)\n\t}\n\tresp.mapFromModel(account)\n\n\treturn c.JSON(http.StatusOK, resp)\n}", "func (o AccessCustomPageOutput) AccountId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *AccessCustomPage) pulumi.StringPtrOutput { return v.AccountId }).(pulumi.StringPtrOutput)\n}", "func (o StorageAccountOutput) AccountName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v StorageAccount) *string { return v.AccountName }).(pulumi.StringPtrOutput)\n}", "func (s AwsAccount) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (mt *Account) Dump(view AccountViewEnum) (res map[string]interface{}, err error) {\n\tif view == AccountDefaultView {\n\t\tres, err = MarshalAccount(mt, err)\n\t}\n\tif view == AccountLinkView {\n\t\tres, err = MarshalAccountLink(mt, err)\n\t}\n\tif view == AccountTinyView {\n\t\tres, err = MarshalAccountTiny(mt, err)\n\t}\n\treturn\n}", "func (o *NumbersACH) GetAccountOk() (*string, bool) {\n\tif o == nil {\n\t\treturn nil, false\n\t}\n\treturn &o.Account, true\n}", "func (c *TradeController) Show(ctx *app.ShowTradeContext) error {\n\t// TradeController_Show: start_implement\n\n\t// Put your logic here\n\tt, ok := tradeRegistry[ctx.TradeID]\n\tif !ok {\n\t\treturn ctx.NotFound()\n\t}\n\tres := &app.GoaTrade{\n\t\tTradeID: t.TradeID,\n\t\tContractID: t.ContractID,\n\t\tCounterpartyID: t.CounterpartyID,\n\t}\n\treturn ctx.OK(res)\n\t// TradeController_Show: end_implement\n}", "func (o AzureTableOutputDataSourceOutput) AccountName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v AzureTableOutputDataSource) *string { return v.AccountName }).(pulumi.StringPtrOutput)\n}", "func (s *Single) Account(account_id interface{}) (account *Account) {\n\taccount = &Account{}\n\terr := DB.BelongsToThrough(s, \"users\").Find(account, account_id)\n\tif err != nil {\n\t\treturn nil\n\t}\n\treturn\n}", "func GetAccount(w http.ResponseWriter, r *http.Request) {\n\tvar accountID = mux.Vars(r)[\"accountID\"]\n\n\taccount, err := DBClient.QueryAccount(accountID)\n\n\tif err != nil {\n\t\tw.WriteHeader(http.StatusNotFound)\n\t\treturn\n\t}\n\n\tdata, _ := json.Marshal(account)\n\twriteJSONResponse(w, http.StatusOK, data)\n}", "func (_Storage *StorageCallerSession) AccountAt(index *big.Int) (common.Address, uint8, bool, common.Address, error) {\n\treturn _Storage.Contract.AccountAt(&_Storage.CallOpts, index)\n}", "func GetAccount(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\n\tparams := mux.Vars(r)\n\tvar account entity.Account\n\tdb.DBCon.First(&account, params[\"id\"])\n\n\tjson.NewEncoder(w).Encode(account)\n}", "func (o LookupAccountTeamResultOutput) AccountId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupAccountTeamResult) string { return v.AccountId }).(pulumi.StringOutput)\n}", "func (o GetAggregatorsAggregatorAggregatorAccountOutput) AccountType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetAggregatorsAggregatorAggregatorAccount) string { return v.AccountType }).(pulumi.StringOutput)\n}", "func (k KuCLIContext) GetAccountInfo() (exported.Account, error) {\n\treturn NewAccountRetriever(k).GetAccount(k.GetAccountID())\n}", "func (o *NumbersACH) SetAccount(v string) {\n\to.Account = v\n}", "func (a *DefaultApiService) ShowTenant(ctx _context.Context, id string) ApiShowTenantRequest {\n\treturn ApiShowTenantRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\tid: id,\n\t}\n}", "func DisplayUAAInfo(w http.ResponseWriter, r *http.Request) {\n\tapiQuery := \"/info\"\n\tuaaRespBytes, flash, userNameVal := ClientRequest(w, r, apiQuery)\n\tuaaResp := ServerInfo{}\n\tif uaaServErr := json.Unmarshal([]byte(uaaRespBytes), &uaaResp); uaaServErr != nil {\n\t\tfmt.Println(uaaServErr)\n\t}\n\tdata := CredentialPageData{\n\t\tPageTitle: \"UAA Info\",\n\t\tServerInfo: uaaResp,\n\t\tUserName: userNameVal,\n\t\tFlash: flash,\n\t}\n\ttmpl := template.Must(template.ParseFiles(\"templates/uaainfo.html\", \"templates/base.html\"))\n\ttmpl.ExecuteTemplate(w, \"base\", data)\n}", "func (am *AccountManager) Account(name string) (*Account, error) {\n\trespChan := make(chan *Account)\n\tam.cmdChan <- &accessAccountRequest{\n\t\tname: name,\n\t\tresp: respChan,\n\t}\n\tresp := <-respChan\n\tif resp == nil {\n\t\treturn nil, ErrNotFound\n\t}\n\treturn resp, nil\n}", "func (o SnapshotOutput) OwnerAccount() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Snapshot) pulumi.StringOutput { return v.OwnerAccount }).(pulumi.StringOutput)\n}", "func (a *AccountClient) List(paging PagingParams) (*Resources, error) {\n\n\tr := a.client.R().SetResult(&Resources{})\n\n\tif paging.number != \"\" {\n\t\tr.SetQueryParam(\"page[number]\", paging.number)\n\t}\n\n\tif paging.size != \"\" {\n\t\tr.SetQueryParam(\"page[size]\", paging.size)\n\t}\n\tresp, err := r.Get(\"/v1/organisation/accounts\")\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"list accounts failed: %s\", err)\n\t}\n\n\tif resp.Error() != nil {\n\t\treturn nil, getAPIError(resp)\n\t}\n\n\treturn resp.Result().(*Resources), nil\n}", "func (a *managedAddress) Account() uint32 {\n\treturn a.account\n}", "func (s ManagedAccount) String() string {\n\treturn awsutil.Prettify(s)\n}", "func FindAcc(w http.ResponseWriter, req *http.Request) {\n\t//io.WriteString(w, \"Alle Accounts Anzeigen!\"+\"\\n\")\n\tallAcc(w)\n}", "func (a *AmoCrm) GetAccount(with []string) (*Account, error) {\n\tvar account *Account\n\treturn account, a.getItem([]string{accountEntity}, nil, &entitiesQuery{With: with}, &account)\n}" ]
[ "0.8005922", "0.7624846", "0.76141137", "0.7244993", "0.7196593", "0.67903006", "0.67458165", "0.66855305", "0.66137165", "0.64803433", "0.6379123", "0.6314034", "0.6280776", "0.6280776", "0.62725013", "0.6227875", "0.61869156", "0.6158084", "0.6129593", "0.6111697", "0.61007607", "0.5993273", "0.5949654", "0.59281677", "0.5867765", "0.5865872", "0.5838524", "0.5838524", "0.5832362", "0.58087665", "0.580274", "0.5772795", "0.57532746", "0.57515895", "0.5669796", "0.56495625", "0.5622842", "0.5618926", "0.55898255", "0.5563874", "0.5533741", "0.5525161", "0.55099154", "0.55038834", "0.55017984", "0.5476875", "0.5474639", "0.54609466", "0.5453077", "0.54487014", "0.544619", "0.5440565", "0.5414209", "0.5414209", "0.53983814", "0.539623", "0.5393991", "0.53894794", "0.53840846", "0.53804785", "0.5379108", "0.53432316", "0.53254896", "0.53223234", "0.5321424", "0.52940434", "0.52936715", "0.52886915", "0.5288468", "0.52823335", "0.5272504", "0.52571225", "0.52531946", "0.525099", "0.5240815", "0.524038", "0.523691", "0.52362746", "0.5228993", "0.522807", "0.5222328", "0.5219256", "0.52180773", "0.5208148", "0.5198049", "0.5197814", "0.51959705", "0.5193926", "0.5189523", "0.51882964", "0.5177747", "0.5168336", "0.5168124", "0.51679415", "0.5167756", "0.516275", "0.5162161", "0.516039", "0.51603484", "0.51545274" ]
0.89442307
0
Mine is used to mine a coinbase transaction. Notice: this is not real mining work.
func (account *Account) Mine() transaction.Transaction { params, _ := pbc.NewParamsFromString(crypto.Params) pairing := params.NewPairing() var tx transaction.Transaction RA := pairing.NewG1() RT := pairing.NewG1() V := pairing.NewG1() P := pairing.NewG1() r := pairing.NewZr() r.Rand() T := pairing.NewG1() T.SetString(crypto.T, 10) A := pairing.NewG1() A.SetString(account.A, 10) tx.Base = true RA.PowZn(A, r) RT.PowZn(T, r) P.SetString(crypto.GetP(r.String(), account.AT, account.A), 10) V.SetString(crypto.GetV(r.String(), account.AT), 10) tx.RA = RA.String() tx.RT = RT.String() tx.P = P.String() tx.V = V.String() tx.Setp(crypto.Getp(account.a, RT.String())) account.Wallet[account.WalletIndex] = tx account.WalletIndex++ fmt.Println("\033[1;31;40m", "We mined a new tx:", tx.P, "\033[0m") return tx }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (self *BlockChain) Mine() {\n\t//Get the last transactions proof\n\tprintln(\"Starting to mine...\")\n\tlast := self.LastBlock()\n\tlastProof := last.Proof\n\n\t//Work out the proof\n\tfmt.Printf(\"Last proof = %d \\n\", lastProof)\n\tnewProof := self.ProofOfWork(lastProof)\n\tself.NewTransaction(\"0\", \"dest\", 1)\n\n\t//Add to blockchain with Proof + HASH\n\t//TODO get hash of previous\n\tfmt.Printf(\"new proof = %d \\n\", newProof)\n\tself.NewBlock(newProof, \"xxx\")\n\n}", "func (d *POW) Mine(tracer *tracing.Tracer, nonce []uint8, numTrailingZeros uint) error {\n\ttracer.RecordAction(PowlibMiningBegin{Nonce: nonce, NumTrailingZeros: numTrailingZeros})\n\n\tgo func(nonce []uint8, numTrailingZeros uint) {\n\t\tvar secret []uint8\n\t\targs := PowlibMine{Nonce: nonce, NumTrailingZeros: numTrailingZeros}\n\t\ttracer.RecordAction(args)\n\t\td.client.Call(\"Coordinator.Mine\", args, &secret)\n\t\t//if err != nil {\n\t\t//\treturn err\n\t\t//}\n\t\ttracer.RecordAction(PowlibSuccess{\n\t\t\tNonce: nonce,\n\t\t\tNumTrailingZeros: numTrailingZeros,\n\t\t\tSecret: secret,\n\t\t})\n\t\ttracer.RecordAction(PowlibMiningComplete{\n\t\t\tNonce: nonce,\n\t\t\tNumTrailingZeros: numTrailingZeros,\n\t\t\tSecret: secret,\n\t\t})\n\t\td.nc <- MineResult{\n\t\t\tNonce: nonce,\n\t\t\tNumTrailingZeros: numTrailingZeros,\n\t\t\tSecret: secret,\n\t\t}\n\t}(nonce, numTrailingZeros)\n\treturn nil\n}", "func (back *backend) Mine(ctx context.Context,\n\ttx *types.Transaction) (*types.Receipt, error) {\n\tswitch conn := back.connect.(type) {\n\tcase *ethclient.Client:\n\t\treturn bind.WaitMined(ctx, conn, tx)\n\tcase *backends.SimulatedBackend:\n\t\tconn.Commit()\n\t\treturn bind.WaitMined(ctx, conn, tx)\n\t}\n\treturn nil, ErrInvalidBackend\n}", "func Mine(chanLB chan Block, chanT chan Transaction, chanB chan Block, log util.Logger) {\n\tvar t Transaction // incoming transaction\n\tvar lb Block // current last block on the chain\n\tvar block Block // block to mine\n\n\tvar counterInt32 uint32\n\tcounter := []byte{0, 0, 0, 0}\n\tsuccess := false\n\tfor {\n\t\tselect {\n\t\tcase t = <-chanT:\n\t\t\tblock.Transactions.AddTransaction(t)\n\t\t\tlog.Infof(\"MINER RECEIVED TRANSACTION: %v\", t)\n\t\tcase lb = <-chanLB:\n\t\t\tblock = lb.NewBlock()\n\t\t\tlog.Infof(\"MINER RECEIVED BLOCK: %v\", lb)\n\t\tdefault:\n\t\t\tif block.Transactions == nil || &lb == nil {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\t// increment counter\n\t\t\tcounterInt32 = binary.LittleEndian.Uint32(counter)\n\t\t\tcounterInt32++\n\t\t\tbinary.LittleEndian.PutUint32(counter, counterInt32)\n\n\t\t\tsuccess = ProofOfWork(lb.Hash(), counter)\n\t\t\tif success == true {\n\t\t\t\tblock.Proof = append([]byte(nil), counter...)\n\t\t\t\tchanB <- block\n\t\t\t\tlog.Infof(\"MINED NEW BLOCK: %v\", block)\n\t\t\t\tlb = block\n\t\t\t\tblock = lb.NewBlock()\n\t\t\t}\n\t\t}\n\t}\n}", "func (bi *Blockchainidentifier) Mine(w http.ResponseWriter, r *http.Request) {\n\tlog.Println(r.RemoteAddr + \" GET /mine\")\n\n\tlastblock := bi.lastBlock()\n\tlastblockhash := blockHasher(lastblock)\n\tlastproof := lastblock.Proof\n\tnewproof := bi.proofOfWork(lastproof)\n\tnewblockindex := bi.newTransaction(transaction{\n\t\tAmount: 1,\n\t\tRecipient: bi.Nodeidentifier,\n\t\tSender: \"0\",\n\t\tTimestamp: time.Now().UTC().Format(\"2006-01-02 15:04:05\"),\n\t})\n\n\tblockforged := bi.newBlock(newblockindex, newproof, lastblockhash)\n\n\tresponseMessage := map[string]interface{}{\n\t\t\"message\": \"New Block Forged\",\n\t\t\"index\": blockforged.Index,\n\t\t\"previous_block_hash\": blockforged.Previousblockhash,\n\t\t\"proof\": blockforged.Proof,\n\t\t\"transactions\": blockforged.Transactions,\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=utf-8\")\n\tw.WriteHeader(http.StatusCreated)\n\tjson.NewEncoder(w).Encode(responseMessage)\n}", "func Test_MineBlock_MineBlock(t *testing.T) {\n\tvar block = &Block{\n\t\tIndex: 0,\n\t\tHash: \"\",\n\t\tTransactions: []*transactions.Transaction{\n\t\t\t{\n\t\t\t\tID: \"transactionID\",\n\t\t\t\tInputs: []*transactions.TransactionInput{\n\t\t\t\t\t{\n\t\t\t\t\t\tOutputID: \"output-ID\",\n\t\t\t\t\t\tOutputIndex: 1,\n\t\t\t\t\t\tSignature: \"signature\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tOutputs: []*transactions.TransactionOutput{\n\t\t\t\t\t{\n\t\t\t\t\t\tAddress: \"address\",\n\t\t\t\t\t\tAmount: 25,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\tMessage: \"my genesis block!!\",\n\t\tPreviousHash: \"\",\n\t\tTimestamp: time.Unix(1465154705, 0),\n\t\tDifficulty: 5,\n\t\tNonce: 0,\n\t}\n\n\t// hash block\n\tblock.MineBlock()\n\n\t// result should be correct hash\n\texpectedHash := \"01cae462faef5b5132df4a29cba801c620813bc7033ad8ae7d4ad8a8806bb7ca\"\n\tif block.Hash != expectedHash {\n\t\tt.Errorf(\"mining result is incorrect, Actual: %s Expected: %s\", block.Hash, expectedHash)\n\t}\n\texpectedNonce := 4\n\tif block.Nonce != expectedNonce {\n\t\tt.Errorf(\"mining result is incorrect, Actual: %d Expected: %d\", block.Nonce, expectedNonce)\n\t}\n}", "func (b *Block) Mine(diff int) string {\n\tfor {\n\t\th := b.Hash()\n\t\tif countStartZeroBits(h) == diff {\n\t\t\treturn mustEncodeBase64URL(h)\n\t\t}\n\t\tb.Nonce++\n\t\tif b.Nonce%10000 == 0 {\n\t\t\tfmt.Println(b.Nonce)\n\t\t}\n\t}\n}", "func mineNewBlock (block *Block) {\n proofPrefix := strings.Repeat(\"0\", difficulty)\n for calculateHash(*block)[:difficulty] != proofPrefix {\n block.Nonce ++\n }\n\n block.Hash = calculateHash(*block)\n}", "func (miner *Miner) Mine(team *Team, index int) {\n\tticker := time.NewTicker(time.Second * time.Duration(miner.interval))\n\tdefer ticker.Stop()\n\tfor {\n\t\tselect {\n\t\tcase <-ticker.C:\n\t\t\tif !team.DiscoveredNodes[index].Active || !miner.Process.Alive {\n\t\t\t\treturn\n\t\t\t}\n\t\t\t//team.mutex.Unlock()\n\t\t\tteam.mutex.Lock()\n\t\t\tdefer team.mutex.Unlock()\n\t\t\tswitch miner.minerType {\n\t\t\tcase \"Bandwidth\":\n\t\t\t\tteam.Bandwidth = team.Bandwidth + miner.amount\n\t\t\t\tteam.mutex.Unlock()\n\t\t\t\tcontinue\n\t\t\tcase \"IO\":\n\t\t\t\tteam.Io = team.Io + miner.amount\n\t\t\t\tteam.mutex.Unlock()\n\t\t\t\tcontinue\n\t\t\tcase \"Entropy\":\n\t\t\t\tteam.Entropy = team.Entropy + miner.amount\n\t\t\t\tteam.mutex.Unlock()\n\t\t\t\tcontinue\n\t\t\tcase \"CPU\":\n\t\t\t\tteam.Cpu = team.Cpu + miner.amount\n\t\t\t\tteam.mutex.Unlock()\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\t//team.mutex.Unlock()\n\t\t}\n\t}\n}", "func (blk *Block) Mine(workers uint64) bool {\n\treasonableRangeEnd := uint64(4 * 1 << blk.Difficulty) // 4 * 2^(bits that must be zero)\n\tmr := blk.MineRange(0, reasonableRangeEnd, workers, 4567)\n\tif mr.Found {\n\t\tblk.SetProof(mr.Proof)\n\t}\n\treturn mr.Found\n}", "func (blk *Block) Mine(workers uint64) bool {\n\treasonableRangeEnd := uint64(4 * 1 << blk.Difficulty) // 4 * 2^(bits that must be zero)\n\tmr := blk.MineRange(0, reasonableRangeEnd, workers, 4567)\n\tif mr.Found {\n\t\tblk.SetProof(mr.Proof)\n\t}\n\treturn mr.Found\n}", "func (blk *Block) Mine(workers uint64) bool {\n\treasonableRangeEnd := uint64(4 * 1 << blk.Difficulty) // 4 * 2^(bits that must be zero)\n\tmr := blk.MineRange(0, reasonableRangeEnd, workers, 4321)\n\tif mr.Found {\n\t\tblk.SetProof(mr.Proof)\n\t}\n\treturn mr.Found\n}", "func Mine(backend *backends.SimulatedBackend, blockTime time.Duration) (stopMining func()) {\n\ttimer := time.NewTicker(blockTime)\n\tchStop := make(chan struct{})\n\twg := sync.WaitGroup{}\n\twg.Add(1)\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-timer.C:\n\t\t\t\tbackend.Commit()\n\t\t\tcase <-chStop:\n\t\t\t\twg.Done()\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\treturn func() { close(chStop); timer.Stop(); wg.Wait() }\n}", "func TestMine(t *testing.T) {\n\ttest := tests[4] // only one we have complete data for\n\n\t// note that this has no MerkleRoot, as above ..\n\tserverData := coin.InComing{\n\t\tVersion: test.v,\n\t\tPrevBlock: test.pb,\n\t\tTimeStamp: test.tme,\n\t\tBits: test.bits,\n\t\tCoinBase: txHashes[0], // but refers to txn hash list\n\t\tTxHashes: txHashes[1:],\n\t\tShare: 3,\n\t}\n\ttask := server.SetTask(serverData) // we get this data from the server\n\n\t// for success, you need to make Cls == 856192328 (Mod N)\n\tif (test.nce)%N != Cls {\n\t\tfmt.Printf(\"\\n** Note you will not find a solution! needs Cls == %d (Mod N)\\n\", test.nce)\n\t}\n\tSearch(task)\n}", "func Mine(block *model.Block, difficulty int, ctl chan commands.Command) (commands.Command, error) {\n\tfor i := 0; i < int(^uint(0)>>1); i++ {\n\t\t// time.Sleep(time.Second)\n\t\tselect {\n\t\tcase c := <-ctl:\n\t\t\treturn c, errors.New(\"mining terminated, new block received or explicit termination\")\n\t\tdefault:\n\t\t\tblock.Nounce = int64(i)\n\t\t\tisMatched, digest := MatchDifficulty(block, difficulty)\n\t\t\tif isMatched {\n\t\t\t\tblock.Hash = digest\n\t\t\t\treturn commands.NewDefaultCommand(), nil\n\t\t\t}\n\t\t}\n\t}\n\treturn commands.NewDefaultCommand(), errors.New(\"failed to find any nounce\")\n}", "func (b *BlockChain) MineBlock(txns []*Transaction) {\n\t// construct new block and prev hash will be current tip of db\n\tblock := NewBlock(txns, b.tip)\n\n\terr := b.db.Update(func(tx *bolt.Tx) error {\n\t\tbckt := tx.Bucket([]byte(blocksBucket))\n\t\tif err := bckt.Put(block.Hash, block.Serialize()); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif err := bckt.Put([]byte(\"l\"), block.Hash); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tb.tip = block.Hash\n\t\treturn nil\n\t})\n\n\tif err != nil {\n\t\tlog.Fatal(\"AddBlock :\", err)\n\t}\n}", "func (blockchain *Blockchain) MineNewBlock(originalTxs []*Transaction) *Block {\n\t// Reward of mining a block\n\tcoinBaseTransaction := NewRewardTransacion()\n\ttxs := []*Transaction{coinBaseTransaction}\n\ttxs = append(txs, originalTxs...)\n\t// Verify transactions\n\tfor _, tx := range txs {\n\t\tif !tx.IsCoinBaseTransaction() {\n\t\t\tif blockchain.VerifityTransaction(tx, txs) == false {\n\t\t\t\tlog.Panic(\"Verify transaction failed...\")\n\t\t\t}\n\t\t}\n\t}\n\n\tDBName := fmt.Sprintf(DBName, os.Getenv(\"NODE_ID\"))\n\tdb, err := bolt.Open(DBName, 0600, nil)\n\tif err != nil {\n\t\tlog.Panic(err)\n\t}\n\tdefer db.Close()\n\t// Get the latest block\n\tvar block Block\n\terr = db.View(func(tx *bolt.Tx) error {\n\t\tb := tx.Bucket([]byte(BlockBucketName))\n\t\tif b != nil {\n\t\t\thash := b.Get([]byte(\"l\"))\n\t\t\tblockBytes := b.Get(hash)\n\t\t\tgobDecode(blockBytes, &block)\n\t\t}\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\tlog.Panic(err)\n\t}\n\n\t// Mine a new block\n\tnewBlock := NewBlock(txs, block.Height+1, block.BlockHash)\n\n\treturn newBlock\n}", "func mine(w *wallet.Wallet, bl Block) (validNonce uint32) {\n\tfor i := 0; i < math.MaxInt64; i += 1 {\n\t\tif bl.CheckNonce(uint32(i)) {\n\t\t\tlog.Println(i, \" Success!\")\n\t\t\treturn uint32(i)\n\t\t}\n\t}\n\treturn 0\n}", "func (b *Block) Mine() bool {\n\tnonce := uint32(b.Nonce)\n\tfor nonce = nonce; nonce < MaxUint; nonce++ {\n\t\tb.Nonce = nonce\n\n\t\tif b.IsValid() {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func MineBlock(w http.ResponseWriter, r *http.Request) {\n\t// Checks for the block in data field\n\tvar data BlockData\n\terr := json.NewDecoder(r.Body).Decode(&data)\n\tif err != nil {\n\t\tlog.Println(\"MineBlock: Received block failed to prase(JSON)\")\n\t}\n\n\tsuccess := b.GenerateNextBlock(data.Data)\n\tif success {\n\t\thub.broadcastMsg(RespLatestMsg())\n\t}\n}", "func (unitImpl *UnitImpl) Mine(body stardash.Body) bool {\n\treturn unitImpl.RunOnServer(\"mine\", map[string]interface{}{\n\t\t\"body\": body,\n\t}).(bool)\n}", "func (ethash *Ethash) mine(block *types.Block, id int, seed uint64, abort chan struct{}, found chan *types.Block) {\n\t// Extract some data from the header\n\tvar (\n\t\theader = block.Header()\n\t\thash = ethash.SealHash(header).Bytes()\n\t\ttarget = new(big.Int).Div(two256, header.Difficulty)\n\t\tnumber = header.Number.Uint64()\n\t\tdataset = ethash.dataset(number, false)\n\t)\n\t// Start generating random nonces until we abort or find a good one\n\tvar (\n\t\tattempts = int64(0)\n\t\tnonce = seed\n\t)\n\tlogger := ethash.config.Log.New(\"miner\", id)\n\tlogger.Trace(\"Started ethash search for new nonces\", \"seed\", seed)\nsearch:\n\tfor {\n\t\tselect {\n\t\tcase <-abort:\n\t\t\t// Mining terminated, update stats and abort\n\t\t\tlogger.Trace(\"Ethash nonce search aborted\", \"attempts\", nonce-seed)\n\t\t\tethash.hashrate.Mark(attempts)\n\t\t\tbreak search\n\n\t\tdefault:\n\t\t\t// We don't have to update hash rate on every nonce, so update after after 2^X nonces\n\t\t\tattempts++\n\t\t\tif (attempts % (1 << 15)) == 0 {\n\t\t\t\tethash.hashrate.Mark(attempts)\n\t\t\t\tattempts = 0\n\t\t\t}\n\t\t\t// Compute the PoW value of this nonce\n\t\t\tdigest, result := hashimotoFull(dataset.dataset, hash, nonce)\n\t\t\tif new(big.Int).SetBytes(result).Cmp(target) <= 0 {\n\t\t\t\t// Correct nonce found, create a new header with it\n\t\t\t\theader = types.CopyHeader(header)\n\t\t\t\theader.Nonce = types.EncodeNonce(nonce)\n\t\t\t\theader.MixDigest = common.BytesToHash(digest)\n\n\t\t\t\t// Seal and return a block (if still needed)\n\t\t\t\tselect {\n\t\t\t\tcase found <- block.WithSeal(header):\n\t\t\t\t\tlogger.Trace(\"Ethash nonce found and reported\", \"attempts\", nonce-seed, \"nonce\", nonce)\n\t\t\t\tcase <-abort:\n\t\t\t\t\tlogger.Trace(\"Ethash nonce found but discarded\", \"attempts\", nonce-seed, \"nonce\", nonce)\n\t\t\t\t}\n\t\t\t\tbreak search\n\t\t\t}\n\t\t\tnonce++\n\t\t}\n\t}\n\t// Datasets are unmapped in a finalizer. Ensure that the dataset stays live\n\t// during sealing so it's not unmapped while being read.\n\truntime.KeepAlive(dataset)\n}", "func (m *Miner) Mine() {\n\tm.miningWorkChannel = make(chan *miningWork, len(m.ClDevices))\n\tgo m.createWork()\n\tfor minerID, device := range m.ClDevices {\n\t\tsdm := &singleDeviceMiner{\n\t\t\tClDevice: device,\n\t\t\tMinerID: minerID,\n\t\t\tHashRateReports: m.HashRateReports,\n\t\t\tminingWorkChannel: m.miningWorkChannel,\n\t\t\tGlobalItemSize: m.GlobalItemSize,\n\t\t\tClient: m.Client,\n\t\t}\n\n\t\tgo sdm.mine()\n\t}\n}", "func MineBlock(bc *blockchain.Blockchain, transactions []*transaction.Transaction) *blockchain.Block {\n\tvar lastHash []byte\n\tvar lastHeight int\n\n\tfor _, tx := range transactions {\n\t\t// TODO: ignore transaction if it's not valid\n\t\tif bc.VerifyTransactionSig(tx) != true {\n\t\t\tlog.Panic(\"ERROR: Invalid transaction\")\n\t\t}\n\t}\n\n\tlastBlock := bc.LastBlockInfo()\n\tlastHash = lastBlock.Hash\n\tlastHeight = lastBlock.Height\n\n\tnewBlock := blockchain.NewBlock(transactions, lastHash, lastHeight+1, blockchain.GetBlockchain().GetCurrentDifficult())\n\n\tpow := NewProofOfWork(newBlock)\n\tnonce, hash := pow.Run()\n\n\tnewBlock.Hash = hash[:]\n\tnewBlock.Nonce = nonce\n\n\tbc.AddBlock(newBlock)\n\n\treturn newBlock\n}", "func (c BaseController) CoinMint(store weave.KVStore,\n\tdest weave.Address, amount coin.Coin) error {\n\n\trecipient, err := c.bucket.GetOrCreate(store, dest)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = Add(AsCoinage(recipient), amount)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn c.bucket.Save(store, recipient)\n}", "func (s *service) MineNewBlock(lastBlock *Block, data []Transaction) (*Block, error) {\n\t// validations\n\tif lastBlock == nil {\n\t\treturn nil, ErrMissingLastBlock\n\t}\n\n\tdifficulty := lastBlock.Difficulty\n\tvar nonce uint32\n\tvar timestamp int64\n\tvar hash string\n\tfor {\n\t\tnonce++\n\t\ttimestamp = time.Now().UnixNano()\n\t\tdifficulty = adjustBlockDifficulty(*lastBlock, timestamp, s.MineRate)\n\t\thash = hashing.SHA256Hash(timestamp, *lastBlock.Hash, data, nonce, difficulty)\n\t\tif hexStringToBinary(hash)[:difficulty] == strings.Repeat(\"0\", int(difficulty)) {\n\t\t\tbreak\n\t\t}\n\t}\n\n\treturn yieldBlock(timestamp, lastBlock.Hash, &hash, data, nonce, difficulty), nil\n}", "func (_Lelecoin *LelecoinTransactor) Mint(opts *bind.TransactOpts, to common.Address, value *big.Int) (*types.Transaction, error) {\n\treturn _Lelecoin.contract.Transact(opts, \"mint\", to, value)\n}", "func (t *trusteeImpl) NewMiningRewardTx(block consensus.Block) *consensus.Transaction {\n\tvar tx *consensus.Transaction\n\t// build list of miner nodes for uncle blocks\n\tuncleMiners := make([][]byte, len(block.UncleMiners()))\n\tfor i, uncleMiner := range block.UncleMiners() {\n\t\tuncleMiners[i] = uncleMiner\n\t}\n\t\n\tops := make([]Op, 1 + len(uncleMiners))\n\t// first add self's mining reward\n\tops[0] = *t.myReward\n\t\n\t// now add award for each uncle\n\tfor i, uncleMiner := range uncleMiners {\n\t\top := NewOp(OpReward)\n\t\top.Params[ParamUncle] = bytesToHexString(uncleMiner)\n\t\top.Params[ParamAward] = UncleAward\n\t\tops[i+1] = *op \n\t}\n\t// serialize ops into payload\n\tif payload,err := common.Serialize(ops); err != nil {\n\t\tt.log.Error(\"Failed to serialize ops into payload: %s\", err)\n\t\treturn nil\n\t} else {\n\t\t// make a signed transaction out of payload\n\t\tif signature := t.sign(payload); len(signature) > 0 {\n\t\t\t// return the signed transaction\n\t\t\ttx = consensus.NewTransaction(payload, signature, t.myAddress)\n\t\t\tblock.AddTransaction(tx)\n\t\t\tt.process(block, tx)\n\t\t}\n\t}\n\treturn tx\n}", "func (g *testGenerator) createPremineBlock(blockName string, additionalAmount dcrutil.Amount) *wire.MsgBlock {\n\tcoinbaseTx := wire.NewMsgTx()\n\tcoinbaseTx.AddTxIn(&wire.TxIn{\n\t\tPreviousOutPoint: *wire.NewOutPoint(&chainhash.Hash{},\n\t\t\twire.MaxPrevOutIndex, wire.TxTreeRegular),\n\t\tSequence: wire.MaxTxInSequenceNum,\n\t\tValueIn: 0, // Updated below.\n\t\tBlockHeight: wire.NullBlockHeight,\n\t\tBlockIndex: wire.NullBlockIndex,\n\t\tSignatureScript: coinbaseSigScript,\n\t})\n\n\t// Add each required output and tally the total payouts for the coinbase\n\t// in order to set the input value appropriately.\n\tvar totalSubsidy dcrutil.Amount\n\tfor _, payout := range g.params.BlockOneLedger {\n\t\tpayoutAddr, err := dcrutil.DecodeAddress(payout.Address, g.params)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\tpkScript, err := txscript.PayToAddrScript(payoutAddr)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\tcoinbaseTx.AddTxOut(&wire.TxOut{\n\t\t\tValue: payout.Amount + int64(additionalAmount),\n\t\t\tVersion: 0,\n\t\t\tPkScript: pkScript,\n\t\t})\n\n\t\ttotalSubsidy += dcrutil.Amount(payout.Amount)\n\t}\n\tcoinbaseTx.TxIn[0].ValueIn = int64(totalSubsidy)\n\n\t// Generate the block with the specially created regular transactions.\n\treturn g.nextBlock(blockName, nil, nil, func(b *wire.MsgBlock) {\n\t\tb.Transactions = []*wire.MsgTx{coinbaseTx}\n\t})\n}", "func (w *Worker) Mine(ctx context.Context, data []byte, targetScore float64) (uint64, error) {\n\tvar (\n\t\tdone uint32\n\t\tcounter uint64\n\t\twg sync.WaitGroup\n\t\tresults = make(chan uint64, w.numWorkers)\n\t\tclosing = make(chan struct{})\n\t)\n\n\t// compute the digest\n\th := Hash.New()\n\th.Write(data)\n\tpowDigest := h.Sum(nil)\n\n\t// stop when the context has been canceled\n\tgo func() {\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\tatomic.StoreUint32(&done, 1)\n\t\tcase <-closing:\n\t\t\treturn\n\t\t}\n\t}()\n\n\t// compute the minimum numbers of trailing zeros required to get a PoW score ≥ targetScore\n\ttargetZeros := uint(math.Ceil(math.Log(float64(len(data)+nonceBytes)*targetScore) / ln3))\n\n\tworkerWidth := math.MaxUint64 / uint64(w.numWorkers)\n\tfor i := 0; i < w.numWorkers; i++ {\n\t\tstartNonce := uint64(i) * workerWidth\n\t\twg.Add(1)\n\t\tgo func() {\n\t\t\tdefer wg.Done()\n\n\t\t\tnonce, workerErr := w.worker(powDigest, startNonce, targetZeros, &done, &counter)\n\t\t\tif workerErr != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tatomic.StoreUint32(&done, 1)\n\t\t\tresults <- nonce\n\t\t}()\n\t}\n\twg.Wait()\n\tclose(results)\n\tclose(closing)\n\n\tnonce, ok := <-results\n\tif !ok {\n\t\treturn 0, ErrCancelled\n\t}\n\treturn nonce, nil\n}", "func (o *ViewReactionsForObject) SetMine(v string) {\n\to.Mine = &v\n}", "func TestIntegrationMiner(t *testing.T) {\n\tif testing.Short() {\n\t\tt.SkipNow()\n\t}\n\tmt, err := createMinerTester(t.Name())\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\t// Check that the wallet has money.\n\tsiacoins, err := mt.wallet.ConfirmedBalance()\n\tif err != nil {\n\t\tt.Error(err)\n\t}\n\tif siacoins.IsZero() {\n\t\tt.Error(\"expecting mining full balance to not be zero\")\n\t}\n\n\t// Mine a bunch of blocks.\n\tfor i := 0; i < 50; i++ {\n\t\tb, _ := mt.miner.FindBlock()\n\t\terr = mt.cs.AcceptBlock(b)\n\t\tif err != nil {\n\t\t\tt.Fatal(err)\n\t\t}\n\t}\n\tmorecoins, err := mt.wallet.ConfirmedBalance()\n\tif err != nil {\n\t\tt.Error(err)\n\t}\n\tif siacoins.Cmp(morecoins) >= 0 {\n\t\tt.Error(\"wallet is not gaining balance while mining\")\n\t}\n}", "func MineBlockWithSpecialsAndAttestations(b *beacon.Blockchain, specials []transaction.Transaction, attestations []transaction.AttestationRecord) (*primitives.Block, error) {\n\tlastBlock, err := b.LastBlock()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tblock1 := primitives.Block{\n\t\tSlotNumber: lastBlock.SlotNumber + 1,\n\t\tRandaoReveal: chainhash.HashH([]byte(fmt.Sprintf(\"test test %d\", lastBlock.SlotNumber))),\n\t\tAncestorHashes: beacon.UpdateAncestorHashes(lastBlock.AncestorHashes, lastBlock.SlotNumber, lastBlock.Hash()),\n\t\tActiveStateRoot: zeroHash,\n\t\tCrystallizedStateRoot: zeroHash,\n\t\tSpecials: specials,\n\t\tAttestations: attestations,\n\t}\n\n\terr = b.ProcessBlock(&block1)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &block1, nil\n}", "func (s *Store) insertMinedTx(ns walletdb.ReadWriteBucket, rec *TxRecord,\n\tblock *BlockMeta) error {\n\n\t// If a transaction record for this hash and block already exists, we\n\t// can exit early.\n\tif _, v := existsTxRecord(ns, &rec.Hash, &block.Block); v != nil {\n\t\treturn ErrDuplicateTx\n\t}\n\n\t// If a block record does not yet exist for any transactions from this\n\t// block, insert a block record first. Otherwise, update it by adding\n\t// the transaction hash to the set of transactions from this block.\n\tvar err error\n\tblockKey, blockValue := existsBlockRecord(ns, block.Height)\n\tif blockValue == nil {\n\t\terr = putBlockRecord(ns, block, &rec.Hash)\n\t} else {\n\t\tblockValue, err = appendRawBlockRecord(blockValue, &rec.Hash)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\terr = putRawBlockRecord(ns, blockKey, blockValue)\n\t}\n\tif err != nil {\n\t\treturn err\n\t}\n\tif err := putTxRecord(ns, rec, &block.Block); err != nil {\n\t\treturn err\n\t}\n\n\t// Determine if this transaction has affected our balance, and if so,\n\t// update it.\n\tif err := s.updateMinedBalance(ns, rec, block); err != nil {\n\t\treturn err\n\t}\n\n\t// If this transaction previously existed within the store as unmined,\n\t// we'll need to remove it from the unmined bucket.\n\tif v := existsRawUnmined(ns, rec.Hash[:]); v != nil {\n\t\tlog.Infof(\"Marking unconfirmed transaction %v mined in block %d\",\n\t\t\t&rec.Hash, block.Height)\n\n\t\tif err := s.deleteUnminedTx(ns, rec); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// As there may be unconfirmed transactions that are invalidated by this\n\t// transaction (either being duplicates, or double spends), remove them\n\t// from the unconfirmed set. This also handles removing unconfirmed\n\t// transaction spend chains if any other unconfirmed transactions spend\n\t// outputs of the removed double spend.\n\tif err := s.removeDoubleSpends(ns, rec); err != nil {\n\t\treturn err\n\t}\n\n\t// Clear any locked outputs since we now have a confirmed spend for\n\t// them, making them not eligible for coin selection anyway.\n\tfor _, txIn := range rec.MsgTx.TxIn {\n\t\tif err := unlockOutput(ns, txIn.PreviousOutPoint); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (cm *coinsMempool) Add(item *primitives.Tx, state *primitives.CoinsState) error {\n\tcm.lock.Lock()\n\tdefer cm.lock.Unlock()\n\tfpkh, err := item.FromPubkeyHash()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif item.Nonce != state.Nonces[fpkh]+1 {\n\t\treturn errors.New(\"invalid nonce\")\n\t}\n\n\tif item.Fee < 5000 {\n\t\treturn errors.New(\"transaction doesn't include enough fee\")\n\t}\n\n\tmpi, ok := cm.mempool[fpkh]\n\tif !ok {\n\t\tcm.mempool[fpkh] = newCoinMempoolItem()\n\t\tmpi = cm.mempool[fpkh]\n\t}\n\tif err := mpi.add(item, state.Balances[fpkh]); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (bc *Blockchain) Mining() bool {\n\tbc.AddTransaction(MiningSender, bc.blockchainAddress, MiningReward, nil, nil)\n\tbc.createBlock()\n\tutils.Logger.Tracef(\"action=mining, status=success\")\n\treturn true\n}", "func (_Bep20 *Bep20Transactor) Mint(opts *bind.TransactOpts, _to common.Address, _amount *big.Int) (*types.Transaction, error) {\n\treturn _Bep20.contract.Transact(opts, \"mint\", _to, _amount)\n}", "func (m *Miner) threadedMine() {\n\t// Increment the number of threads running, because this thread is spinning\n\t// up. Also grab a number that will tell us when to shut down.\n\tm.mu.Lock()\n\tm.runningThreads++\n\tmyThread := m.runningThreads\n\tm.mu.Unlock()\n\n\t// Try to solve a block repeatedly.\n\tfor {\n\t\t// Grab the number of threads that are supposed to be running.\n\t\tm.mu.Lock()\n\t\tdesiredThreads := m.desiredThreads\n\t\tm.mu.Unlock()\n\n\t\t// If we are allowed to be running, mine a block, otherwise shut down.\n\t\tif desiredThreads >= myThread {\n\t\t\t// Grab the necessary variables for mining, and then attempt to\n\t\t\t// mine a block.\n\t\t\tm.mu.Lock()\n\t\t\tbfw, blockMerkleRoot, target := m.blockForWork()\n\t\t\tm.increaseAttempts()\n\t\t\tm.mu.Unlock()\n\t\t\tm.solveBlock(bfw, blockMerkleRoot, target)\n\t\t} else {\n\t\t\tm.mu.Lock()\n\t\t\t// Need to check the mining status again, something might have\n\t\t\t// changed while waiting for the lock.\n\t\t\tif desiredThreads < myThread {\n\t\t\t\tm.runningThreads--\n\t\t\t\tm.mu.Unlock()\n\t\t\t\treturn\n\t\t\t}\n\t\t\tm.mu.Unlock()\n\t\t}\n\t}\n}", "func NewMining(minter sdk.AccAddress, tally int64) Mining {\n\treturn Mining{\n\t\tMinter: minter,\n\t\tLastTime: 0,\n\t\tTally: tally,\n\t}\n}", "func TestIntegrationBlocksMined(t *testing.T) {\n\tif testing.Short() {\n\t\tt.SkipNow()\n\t}\n\tmt, err := createMinerTester(t.Name())\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\t// Get an unsolved header.\n\tunsolvedHeader, target, err := mt.miner.HeaderForWork()\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\t// Solve the header - necessary because the target is very low when\n\t// mining.\n\tfor {\n\t\tunsolvedHeader.Nonce[0]++\n\t\tid := crypto.HashObject(unsolvedHeader)\n\t\tif bytes.Compare(target[:], id[:]) < 0 {\n\t\t\tbreak\n\t\t}\n\t}\n\n\t// Get two solved headers.\n\theader1, target, err := mt.miner.HeaderForWork()\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\theader1 = solveHeader(header1, target)\n\theader2, target, err := mt.miner.HeaderForWork()\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\theader2 = solveHeader(header2, target)\n\n\t// Submit the unsolved header followed by the two solved headers, this\n\t// should result in 1 real block mined and 1 stale block mined.\n\terr = mt.miner.SubmitHeader(unsolvedHeader)\n\tif err != modules.ErrBlockUnsolved {\n\t\tt.Fatal(err)\n\t}\n\terr = mt.miner.SubmitHeader(header1)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\terr = mt.miner.SubmitHeader(header2)\n\tif err != modules.ErrNonExtendingBlock {\n\t\tt.Fatal(err)\n\t}\n\tgoodBlocks, staleBlocks := mt.miner.BlocksMined()\n\tif goodBlocks != 1 {\n\t\tt.Error(\"expecting 1 good block\")\n\t}\n\tif staleBlocks != 1 {\n\t\tt.Error(\"expecting 1 stale block, got\", staleBlocks)\n\t}\n\n\t// Reboot the miner and verify that the block record has persisted.\n\terr = mt.miner.Close()\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\trebootMiner, err := New(mt.cs, mt.tpool, mt.wallet, filepath.Join(mt.persistDir, modules.MinerDir))\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\tgoodBlocks, staleBlocks = rebootMiner.BlocksMined()\n\tif goodBlocks != 1 {\n\t\tt.Error(\"expecting 1 good block\")\n\t}\n\tif staleBlocks != 1 {\n\t\tt.Error(\"expecting 1 stale block, got\", staleBlocks)\n\t}\n}", "func (o *QuotePaymentMethodManagementV1SetPutMineParams) WithContext(ctx context.Context) *QuotePaymentMethodManagementV1SetPutMineParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func sendCoins(ctx sdk.Context, am sdk.AccountMapper, fromAddr sdk.Address, toAddr sdk.Address, amt sdk.Coins) sdk.Error {\n\t_, err := subtractCoins(ctx, am, fromAddr, amt)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = addCoins(ctx, am, toAddr, amt)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (r *wsContext) AddMinedTxRequest(wallet walletChan, txID *btcwire.ShaHash) {\n\tr.Lock()\n\tdefer r.Unlock()\n\n\tclist, ok := r.minedTxNotifications[*txID]\n\tif !ok {\n\t\tclist = list.New()\n\t\tr.minedTxNotifications[*txID] = clist\n\t}\n\tclist.PushBack(wallet)\n\n\trc := r.connections[wallet]\n\trc.minedTxRequests[*txID] = struct{}{}\n}", "func (p *Provence) MineMetal() int {\n\tlevel := int(p.Level)\n\tcountModifier := float64(p.MetalCount) / MetalMineRate\n\tadjustedModifier := float64((4-level)*10 + p.MineMod)\n\tpopModifier := float64(p.Population) / PopWorkerModifer\n\tMinedMetalAmount := int(countModifier * adjustedModifier * popModifier)\n\tif MinedMetalAmount > p.MetalCount {\n\t\tMinedMetalAmount = p.MetalCount\n\t}\n\tp.MetalCount -= MinedMetalAmount\n\treturn MinedMetalAmount\n}", "func (api *PublicEthereumAPI) Mining() bool {\n\tapi.logger.Debug(\"eth_mining\")\n\treturn false\n}", "func (wlt *WhitelistableToken) egMint(address, amount string) (uint64, error) {\n\t// method\n\ttransferFnSignature := []byte(\"mint(address,uint256)\")\n\thash := sha3.NewLegacyKeccak256()\n\thash.Write(transferFnSignature)\n\tmethodID := hash.Sum(nil)[:4]\n\n\t// address\n\taddr := common.HexToAddress(address)\n\tpaddedAddress := common.LeftPadBytes(addr.Bytes(), 32)\n\n\t// amount\n\tamountBN := new(big.Int)\n\tamountBN.SetString(amount, 10)\n\tpaddedAmount := common.LeftPadBytes(amountBN.Bytes(), 32)\n\n\tvar data []byte\n\tdata = append(data, methodID...)\n\tdata = append(data, paddedAddress...)\n\tdata = append(data, paddedAmount...)\n\n\tgasLimit, err := wlt.EthClient.EstimateGas(context.Background(), ethereum.CallMsg{\n\t\tFrom: *wlt.CallerAddres,\n\t\tTo: wlt.ContractAddress,\n\t\tData: data,\n\t})\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn gasLimit, nil\n}", "func (f *HubTransactor) Mint(opts *bind.TransactOpts, iERC20TokenContract common.Address, fromCosmosAddress string, recipient common.Address, amount *big.Int) (*types.Transaction, error) {\n\treturn f.contract.Transact(opts, \"mint\", iERC20TokenContract, fromCosmosAddress, recipient, amount)\n}", "func (_BREMToken *BREMTokenTransactor) Mint(opts *bind.TransactOpts, _to common.Address, _amount *big.Int) (*types.Transaction, error) {\n\treturn _BREMToken.contract.Transact(opts, \"mint\", _to, _amount)\n}", "func (o *QuotePaymentMethodManagementV1SetPutMineParams) WithHTTPClient(client *http.Client) *QuotePaymentMethodManagementV1SetPutMineParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func (pow *ProofOfWork)Run() ([]byte, uint64) {\n\t/*\n\tThe steps of algorithm\n\t1.assemble data\n\t2.calculate hash\n\t3.compare to the exit criteria(that is target)\n\t*/\n\n\tvar nonce uint64\n\tvar hash [32]byte\n\tblock := pow.block\n\tfmt.Println(\"start mine!\")\n\tfor {\n\t\t//1.assemble data\n\t\titem := [][]byte{\n\t\t\tUint64ToBtye(block.Version),\n\t\t\tblock.PrevHash,\n\t\t\tblock.Merkel,\n\t\t\tUint64ToBtye(block.TimeStamp),\n\t\t\tUint64ToBtye(block.Difficulty),\n\t\t\tUint64ToBtye(nonce),\n\t\t\t//block.Data,\n\t\t}\n\t\tblockInfo := bytes.Join(item,[]byte{})\n\t\t//2.calculate hash\n\t\thash = sha256.Sum256(blockInfo)\n\t\t//3.compare to the exit criteria(that is target)\n\t\tbigTmp := big.Int{}\n\t\tbigTmp.SetBytes(hash[:])\n\t\tif (bigTmp.Cmp(pow.target)) == -1{\n\t\t\tfmt.Printf(\"succeed:%x,\",hash)\n\t\t\tfmt.Printf(\"nonce:%d\\n\",nonce)\n\t\t\tbreak\n\t\t} else {\n\t\t\tnonce++\n\t\t}\n\n\n\t}\n\treturn hash[:], nonce\n}", "func (_BurnableToken *BurnableTokenTransactor) Mint(opts *bind.TransactOpts, _to common.Address, _amount *big.Int) (*types.Transaction, error) {\n\treturn _BurnableToken.contract.Transact(opts, \"mint\", _to, _amount)\n}", "func (_IERC20Minter *IERC20MinterTransactor) Mint(opts *bind.TransactOpts, to common.Address, amount *big.Int) (*types.Transaction, error) {\n\treturn _IERC20Minter.contract.Transact(opts, \"mint\", to, amount)\n}", "func (t *trusteeImpl) VerifyMiningRewardTx(block consensus.Block) bool {\n\t// pop first transaction from the block\n\ttx := block.Transactions()[0]\n\n\t// validate transaction signature\n\tif !t.verify(tx.Payload, tx.Signature, tx.Submitter) {\n\t\treturn false\n\t}\n\n\t// validate that mining award is for block's miner\n\tif string(tx.Submitter) != string(block.Miner()) {\n\t\tt.log.Error(\"mining award owner is not the block miner\")\n\t\treturn false\n\t}\n\n\t// process transaction and update block\n\treturn t.process(block, &tx)\n}", "func (_Harberger *HarbergerTransactor) Mint(opts *bind.TransactOpts, _metaURI string, _tax *big.Int) (*types.Transaction, error) {\n\treturn _Harberger.contract.Transact(opts, \"mint\", _metaURI, _tax)\n}", "func Mint(ctx contractapi.TransactionContextInterface, amount int) error {\n\n\t// Get ID of submitting client identity\n\tminter, err := ctx.GetClientIdentity().GetID()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to get client id: %v\", err)\n\t}\n\n\tif amount <= 0 {\n\t\treturn fmt.Errorf(\"mint amount must be a positive integer\")\n\t}\n\n\tcurrentBalanceBytes, err := ctx.GetStub().GetState(minter)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to read minter account %s from world state: %v\", minter, err)\n\t}\n\n\tvar currentBalance int\n\n\t// If minter current balance doesn't yet exist, we'll create it with a current balance of 0\n\tif currentBalanceBytes == nil {\n\t\tcurrentBalance = 0\n\t} else {\n\t\tcurrentBalance, _ = strconv.Atoi(string(currentBalanceBytes)) // Error handling not needed since Itoa() was used when setting the account balance, guaranteeing it was an integer.\n\t}\n\n\tupdatedBalance := currentBalance + amount\n\n\terr = ctx.GetStub().PutState(minter, []byte(strconv.Itoa(updatedBalance)))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Update the totalSupply\n\ttotalSupplyBytes, err := ctx.GetStub().GetState(totalSupplyKey)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to retrieve total token supply: %v\", err)\n\t}\n\n\tvar totalSupply int\n\n\t// If no tokens have been minted, initialize the totalSupply\n\tif totalSupplyBytes == nil {\n\t\ttotalSupply = 0\n\t} else {\n\t\ttotalSupply, _ = strconv.Atoi(string(totalSupplyBytes)) // Error handling not needed since Itoa() was used when setting the totalSupply, guaranteeing it was an integer.\n\t}\n\n\t// Add the mint amount to the total supply and update the state\n\ttotalSupply += amount\n\terr = ctx.GetStub().PutState(totalSupplyKey, []byte(strconv.Itoa(totalSupply)))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Emit the Transfer event\n\ttransferEvent := event{\"0x0\", minter, amount}\n\ttransferEventJSON, err := json.Marshal(transferEvent)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to obtain JSON encoding: %v\", err)\n\t}\n\terr = ctx.GetStub().SetEvent(\"Transfer\", transferEventJSON)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to set event: %v\", err)\n\t}\n\n\tlog.Printf(\"minter account %s balance updated from %d to %d\", minter, currentBalance, updatedBalance)\n\n\treturn nil\n}", "func (blk Block) MineRange(start uint64, end uint64, workers uint64, chunks uint64) MiningResult {\n\t// TODO\n\tchunkSize := ((end + 1) - start)/ chunks\n\tif chunkSize == 0{\n\t\tchunkSize = 1\n\t}\n\tqueue := work_queue.Create(uint(workers),uint(chunks))\n\tfor i := uint64(0); i < chunks; i ++ {\n\t\tmx := (i+1) * chunkSize\n\t\tif mx > end{\n\t\t\tmx = end\n\t\t}\n\n\t\tworker:= miningWorker{blk, i*chunkSize, mx}\n\t\tqueue.Enqueue(worker)\n\t}\n\tfor i:=uint64(0); i < chunks; i ++{\n\t\tmr := <-queue.Results\n\t\tres := mr.(MiningResult)\n\t\tif res.Found == true{\n\t\t\tqueue.Shutdown()\n\t\t\treturn res\n\t\t}\n\t}\n\tqueue.Shutdown()\n\treturn MiningResult{Found:false}\n}", "func (_ERC721Contract *ERC721ContractTransactor) Mint(opts *bind.TransactOpts, _owner common.Address, _tokenId *big.Int, _approvedAddress common.Address, _metadata string) (*types.Transaction, error) {\n\treturn _ERC721Contract.contract.Transact(opts, \"mint\", _owner, _tokenId, _approvedAddress, _metadata)\n}", "func (_Registry *RegistryTransactor) Mint(opts *bind.TransactOpts, _id *big.Int, _to common.Address, _quantity *big.Int) (*types.Transaction, error) {\n\treturn _Registry.contract.Transact(opts, \"mint\", _id, _to, _quantity)\n}", "func handleWalletSendRawTransaction(s *rpcServer, cmd btcjson.Cmd, wallet walletChan) error {\n\tresult, err := handleSendRawTransaction(s, cmd)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// The result is already guaranteed to be a valid hash string if no\n\t// error was returned above, so it's safe to ignore the error here.\n\ttxSha, _ := btcwire.NewShaHashFromStr(result.(string))\n\n\t// Request to be notified when the transaction is mined.\n\ts.ws.AddMinedTxRequest(wallet, txSha)\n\treturn nil\n}", "func (powb *ProofOfWorkBlock) GenesisBlock(minter *Node) *ProofOfWorkBlock {\n\tvar totalMiningPower int64 = 0\n\t//计算所有区块的算力总和\n\t//for (Node node : getSimulatedNodes()) {\n\t//\ttotalMiningPower += node.getMiningPower();\n\t//}\n\tSimulator.GetSimulatedNodes().Each(func(index int, v interface{}) {\n\t\ttotalMiningPower += v.(*Node).GetMiningPower()\n\t})\n\n\t//难度 = 总算力*区块间隔时间\n\t// 相当于难度没有调整:\n\t// 调整策略,取前几个区块的出块时间,与预计的总出块时间\n\t// 难度(非难度值) = (预计的总出块时间/前几个区块的出块时间)*当前难度\n\t// 比特币总每次调整不超过4倍\n\tGenesisNextDifficulty = GenesisNextDifficulty.Mul(big.NewInt(totalMiningPower), big.NewInt(Simulator.GetTargetInterval()))\n\n\treturn NewProofOfWorkBlock(nil, minter, 0, big.NewInt(0))\n}", "func (_MintableToken *MintableTokenTransactor) Mint(opts *bind.TransactOpts, _to common.Address, _amount *big.Int) (*types.Transaction, error) {\n\treturn _MintableToken.contract.Transact(opts, \"mint\", _to, _amount)\n}", "func (_Lelecoin *LelecoinSession) Mint(to common.Address, value *big.Int) (*types.Transaction, error) {\n\treturn _Lelecoin.Contract.Mint(&_Lelecoin.TransactOpts, to, value)\n}", "func (coin Coin) Minus(coinB Coin) Coin {\n\tr := coin.Amount.Sub(coinB.Amount)\n\treturn Coin{r}\n}", "func (o *QuotePaymentMethodManagementV1SetPutMineParams) WithTimeout(timeout time.Duration) *QuotePaymentMethodManagementV1SetPutMineParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func (blk Block) MineRange(start uint64, end uint64, workers uint64, chunks uint64) MiningResult {\n\tvar noResult, myResult MiningResult\n\tqueue := work_queue.Create(uint(workers), uint(chunks))\n\tlength := uint64(math.Ceil(float64((end - start) / chunks) + 0.0001))\n\n\tfor i := start; i < end; i+= length {\n\t\tmineWorker := miningWorker{i, i + length, blk}\n\t\tqueue.Enqueue(mineWorker)\n\t\t//fmt.Println(i, i+length)\n\t}\n\n\t//fmt.Printf(\"type of queue %T, value of queue %v\", queue, queue)\n\t//fmt.Println(len(queue.Jobs))\n\t//fmt.Println(len(queue.Results))\n\n\n\tfor i := range queue.Results {\n\t\ttemp := i // channel of interface{} type\n\t\t//MiningResult(temp) would not work, but type assert would\n\t\tmyResult = temp.(MiningResult)\n\t\tif (myResult.Found == true) {\n\t\t\tqueue.Shutdown()\n\t\t\t//fmt.Println(\"got something\")\n\t\t\treturn myResult\n\t\t}\n\t}\n\tqueue.Shutdown()\n\tfmt.Println(\"freed\")\n\treturn noResult\n}", "func (blk Block) MineRange(start uint64, end uint64, workers uint64, chunks uint64) MiningResult {\n\tresult := new(MiningResult)\n\tchuckSize := (end-start) / chunks\n\tqueue := work_queue.Create(uint(workers), uint(chunks))\n\n\tstartIndex := start\n\tfor i:=uint64(0); i<chunks; i++{\n\t\tendIndex := startIndex+chuckSize\n\t\tif endIndex > end {\n\t\t\tendIndex = end\n\t\t}\n\t\tminer := miningWorker{startIndex, endIndex, blk}\n\t\tqueue.Enqueue(miner)\n\t\tstartIndex = endIndex+1\n\t}\n\tfor res := range queue.Results{\n\t\tresult = res.(*MiningResult)\n\t\tif result.Found {\n\t\t\tqueue.Shutdown()\n\t\t\tbreak\n\t\t}\n\t}\n\treturn *result\n}", "func (w Wallet) Send(to string, amount uint, utxoSetPath string) error {\n\treturn nil\n}", "func (k Keeper) MintCoins(ctx sdk.Context, moduleName string, amt sdk.Coins) error {\n\tacc := k.GetModuleAccount(ctx, moduleName)\n\tif acc == nil {\n\t\tpanic(sdkerrors.Wrapf(sdkerrors.ErrUnknownAddress, \"module account %s does not exist\", moduleName))\n\t}\n\n\tif !acc.HasPermission(types.Minter) {\n\t\tpanic(sdkerrors.Wrapf(sdkerrors.ErrUnauthorized, \"module account %s does not have permissions to mint tokens\", moduleName))\n\t}\n\n\t_, err := k.bk.AddCoins(ctx, acc.GetAddress(), amt)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// update total supply\n\tsupply := k.GetSupply(ctx)\n\tsupply = supply.Inflate(amt)\n\n\tk.SetSupply(ctx, supply)\n\n\tlogger := k.Logger(ctx)\n\tlogger.Info(fmt.Sprintf(\"minted %s from %s module account\", amt.String(), moduleName))\n\n\treturn nil\n}", "func (_IERC721Minter *IERC721MinterTransactor) Mint(opts *bind.TransactOpts, to common.Address, tokenURI string) (*types.Transaction, error) {\n\treturn _IERC721Minter.contract.Transact(opts, \"mint\", to, tokenURI)\n}", "func (cm *chainManager) MintNewBlock(timestamp time.Time) (*block.Block, error) {\n\treturn cm.bc.MintNewBlock(timestamp)\n}", "func (_Token *TokenTransactor) Mint(opts *bind.TransactOpts, account common.Address, amount *big.Int) (*types.Transaction, error) {\n\treturn _Token.contract.Transact(opts, \"mint\", account, amount)\n}", "func NewCoinBaseTx(to, data string) *Tx {\n\tif data == \"\" {\n\t\tdata = fmt.Sprintf(\"Reward to '%s'\", to)\n\t}\n\tin := TxInput{\n\t\tTxID: []byte{},\n\t\tVout: -1,\n\t\tPubKey: []byte(data),\n\t}\n\tout := NewTXOutput(10, to)\n\n\ttx := Tx{\n\t\tVin: []TxInput{in},\n\t\tVout: []TxOutput{*out},\n\t}\n\ttx.ID = tx.Hash()\n\n\treturn &tx\n}", "func additionalCoinbasePoW(amount dcrutil.Amount) func(*wire.MsgBlock) {\n\treturn func(b *wire.MsgBlock) {\n\t\t// Increase the first proof-of-work coinbase subsidy by the\n\t\t// provided amount.\n\t\tb.Transactions[0].TxOut[2].Value += int64(amount)\n\t}\n}", "func additionalCoinbasePoW(amount dcrutil.Amount) func(*wire.MsgBlock) {\n\treturn func(b *wire.MsgBlock) {\n\t\t// Increase the first proof-of-work coinbase subsidy by the\n\t\t// provided amount.\n\t\tb.Transactions[0].TxOut[2].Value += int64(amount)\n\t}\n}", "func (mgr *MiningMgr) newWork() {\n\tgo func() {\n\t\t// instantSubmit means 15 mins have passed so\n\t\t// the difficulty now is zero and any solution/nonce will work so\n\t\t// can just submit without sending to the miner.\n\t\twork, instantSubmit := mgr.tasker.GetWork()\n\t\tif instantSubmit {\n\t\t\tmgr.solutionOutput <- &pow.Result{Work: work, Nonce: \"anything will work\"}\n\t\t} else {\n\t\t\t// It sends even nil work to indicate that no new challenge is available.\n\t\t\tif work == nil {\n\t\t\t\tmgr.solutionOutput <- nil\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tvar ids []int64\n\t\t\tfor _, id := range work.Challenge.RequestIDs {\n\t\t\t\tids = append(ids, id.Int64())\n\t\t\t}\n\t\t\tlevel.Debug(mgr.logger).Log(\"msg\", \"sending new chalenge for mining\", \"reqIDs\", fmt.Sprintf(\"%+v\", ids))\n\t\t\tmgr.toMineInput <- work\n\t\t}\n\t}()\n}", "func GiveMyMine(path string) Mine {\n\tpanic(\"non implementato\")\n}", "func (_ERC721 *ERC721Transactor) Mint(opts *bind.TransactOpts, _to common.Address, _tokenId *big.Int) (*types.Transaction, error) {\n\treturn _ERC721.contract.Transact(opts, \"mint\", _to, _tokenId)\n}", "func (k Keeper) MintCoins(ctx sdk.Context, moduleName string, amt *Coins) error {\n\tacc := k.GetModuleAccount(ctx, moduleName)\n\tif acc == nil {\n\t\tpanic(sdkerrors.Wrapf(sdkerrors.ErrUnknownAddress, \"module account %s does not exist\", moduleName))\n\t}\n\n\tif !acc.HasPermission(types.Minter) {\n\t\tpanic(sdkerrors.Wrapf(sdkerrors.ErrUnauthorized, \"module account %s does not have permissions to mint tokens\", moduleName))\n\t}\n\n\t_, err := k.bk.IssueCoinPower(ctx, acc.GetID(), *amt)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tk.Logger(ctx).Info(fmt.Sprintf(\"minted %s from %s module account\", amt.String(), moduleName))\n\n\treturn nil\n}", "func (k Keeper) MintCoins(ctx sdk.Context, newCoins sdk.Coins) error {\n\tif newCoins.Empty() {\n\t\t// skip as no coins need to be minted\n\t\treturn nil\n\t}\n\treturn k.supplyKeeper.MintCoins(ctx, types.ModuleName, newCoins)\n}", "func (_Lelecoin *LelecoinTransactorSession) Mint(to common.Address, value *big.Int) (*types.Transaction, error) {\n\treturn _Lelecoin.Contract.Mint(&_Lelecoin.TransactOpts, to, value)\n}", "func (c *Client) Mining() (*BooleanResponse, error) {\n\trequest := c.newRequest(EthMining)\n\n\tresponse := &BooleanResponse{}\n\n\treturn response, c.send(request, response)\n}", "func (c BitcoinCoreChain) RawTx(cxt context.Context, from, to, amount, memo, asset string) (string, error) {\n if configure.ChainAssets[asset] != Bitcoin {\n return \"\", fmt.Errorf(\"Unsupport %s in bitcoincore\", asset)\n }\n amountF, err := strconv.ParseFloat(amount, 64)\n if err != nil {\n return \"\", err\n }\n txAmountSatoshi, err := btcutil.NewAmount(amountF)\n if err != nil {\n return \"\", err\n }\n\n fromPkScript, err := BitcoincoreAddressP2AS(from, c.Mode)\n if err != nil {\n return \"\", err\n }\n toPkScript, err := BitcoincoreAddressP2AS(to, c.Mode)\n if err != nil {\n return \"\", err\n }\n\n // query bitcoin chain info\n chaininfo, err := c.Client.GetBlockChainInfo()\n if err != nil {\n return \"\", err\n }\n // feeKB, err := c.Client.EstimateFee(int64(6))\n feeKB, err := c.Client.EstimateSmartFee(int64(6))\n if err != nil {\n return \"\", err\n }\n feeRate := mempool.SatoshiPerByte(feeKB.FeeRate)\n\n if feeKB.FeeRate <= 0 {\n feeRate = mempool.SatoshiPerByte(100)\n }\n\n var (\n selectedutxos, unselectedutxos []db.UTXO\n selectedCoins coinset.Coins\n )\n\n // Coin Select\n if strings.ToLower(configure.ChainsInfo[Bitcoin].Coin) == strings.ToLower(asset) {\n // select coins for BTC transfer\n if selectedutxos, unselectedutxos, selectedCoins, err = CoinSelect(int64(chaininfo.Headers), txAmountSatoshi, c.Wallet.Address.UTXOs); err != nil {\n return \"\", fmt.Errorf(\"Select UTXO for tx %s\", err)\n }\n }else {\n // select coins for Token transfer\n // 300: https://bitcoin.stackexchange.com/questions/1195/how-to-calculate-transaction-size-before-sending-legacy-non-segwit-p2pkh-p2sh\n inputAmount := feeRate.Fee(uint32(300))\n if selectedutxos, unselectedutxos, selectedCoins, err = CoinSelect(int64(chaininfo.Headers), inputAmount, c.Wallet.Address.UTXOs); err != nil {\n return \"\", fmt.Errorf(\"Select UTXO for tx %s\", err)\n }\n }\n\n var vinAmount int64\n for _, coin := range selectedCoins.Coins() {\n vinAmount += int64(coin.Value())\n }\n msgTx := coinset.NewMsgTxWithInputCoins(wire.TxVersion, selectedCoins)\n\n token := configure.ChainsInfo[Bitcoin].Tokens[strings.ToLower(asset)]\n if token != \"\" && strings.ToLower(asset) != strings.ToLower(configure.ChainsInfo[Bitcoin].Coin) {\n // OmniToken transfer\n b := txscript.NewScriptBuilder()\n b.AddOp(txscript.OP_RETURN)\n\n omniVersion := util.Int2byte(uint64(0), 2)\t// omnicore version\n txType := util.Int2byte(uint64(0), 2)\t// omnicore tx type: simple send\n propertyID := configure.ChainsInfo[Bitcoin].Tokens[asset]\n tokenPropertyid, err := strconv.Atoi(propertyID)\n if err != nil {\n return \"\", fmt.Errorf(\"tokenPropertyid to int %s\", err)\n }\n // tokenPropertyid := configure.Config.OmniToken[\"omni_first_token\"].(int)\n tokenIdentifier := util.Int2byte(uint64(tokenPropertyid), 4)\t// omni token identifier\n tokenAmount := util.Int2byte(uint64(txAmountSatoshi), 8)\t// omni token transfer amount\n\n b.AddData([]byte(\"omni\"))\t// transaction maker\n b.AddData(omniVersion)\n b.AddData(txType)\n b.AddData(tokenIdentifier)\n b.AddData(tokenAmount)\n pkScript, err := b.Script()\n if err != nil {\n return \"\", fmt.Errorf(\"Bitcoin Token pkScript %s\", err)\n }\n msgTx.AddTxOut(wire.NewTxOut(0, pkScript))\n txOutReference := wire.NewTxOut(0, toPkScript)\n msgTx.AddTxOut(txOutReference)\n }else {\n // BTC transfer\n txOutTo := wire.NewTxOut(int64(txAmountSatoshi), toPkScript)\n msgTx.AddTxOut(txOutTo)\n\n // recharge\n // 181, 34: https://bitcoin.stackexchange.com/questions/1195/how-to-calculate-transaction-size-before-sending-legacy-non-segwit-p2pkh-p2sh\n fee := feeRate.Fee(uint32(msgTx.SerializeSize() + 181 + 34))\n if (vinAmount - int64(txAmountSatoshi) - int64(fee)) > 0 {\n txOutReCharge := wire.NewTxOut((vinAmount-int64(txAmountSatoshi) - int64(fee)), fromPkScript)\n msgTx.AddTxOut(txOutReCharge)\n }else {\n selectedutxoForFee, _, selectedCoinsForFee, err := CoinSelect(int64(chaininfo.Headers), fee, unselectedutxos)\n if err != nil {\n return \"\", fmt.Errorf(\"Select UTXO for fee %s\", err)\n }\n for _, coin := range selectedCoinsForFee.Coins() {\n vinAmount += int64(coin.Value())\n }\n txOutReCharge := wire.NewTxOut((vinAmount - int64(txAmountSatoshi) - int64(fee)), fromPkScript)\n msgTx.AddTxOut(txOutReCharge)\n selectedutxos = append(selectedutxos, selectedutxoForFee...)\n }\n }\n\n buf := bytes.NewBuffer(make([]byte, 0, msgTx.SerializeSize()))\n msgTx.Serialize(buf)\n rawTxHex := hex.EncodeToString(buf.Bytes())\n c.Wallet.SelectedUTXO = selectedutxos\n return rawTxHex, nil\n}", "func newCoins() sdk.Coins {\n\treturn sdk.Coins{\n\t\tsdk.NewInt64Coin(\"atom\", 10000000),\n\t}\n}", "func (_Univ2 *Univ2Transactor) Mint(opts *bind.TransactOpts, to common.Address) (*types.Transaction, error) {\n\treturn _Univ2.contract.Transact(opts, \"mint\", to)\n}", "func (f *FakeUtxoVM) GenerateTx(from string, to string, t *testing.T, amount string, preHash []byte, desc []byte, frozenHeight int64) ([]*pb.Transaction, *pb.Transaction, *ecdsa.PrivateKey, error) {\n\tutxoVM := f.U\n\tt.Logf(\"preHash of this block: %x\", preHash)\n\ttxReq := &pb.TxData{}\n\ttxReq.Bcname = \"xuper-chain\"\n\ttxReq.FromAddr = f.Users[from].Address\n\ttxReq.FromPubkey = f.Users[from].Pubkey\n\ttxReq.FromScrkey = f.Users[from].PrivateKey\n\ttxReq.Nonce = \"nonce\"\n\ttxReq.Timestamp = time.Now().UnixNano()\n\ttxReq.Desc = desc\n\ttxReq.Account = []*pb.TxDataAccount{\n\t\t{Address: f.Users[to].Address, Amount: amount, FrozenHeight: frozenHeight},\n\t\t{Address: FeePlaceholder, Amount: \"300000000\"},\n\t}\n\ttx, err := utxoVM.GenerateTx(txReq)\n\tif err != nil {\n\t\treturn nil, nil, nil, err\n\t}\n\n\tverifyOK, vErr := utxoVM.ImmediateVerifyTx(tx, false)\n\tif !verifyOK || vErr != nil {\n\t\tt.Log(\"verify tx fail, ignore in unit test here\", vErr)\n\t\tt.Error(\"verify tx failed\")\n\t}\n\terrDo := utxoVM.DoTx(tx)\n\tif errDo != nil {\n\t\treturn nil, nil, nil, errDo\n\t}\n\ttxlist, packErr := utxoVM.GetUnconfirmedTx(true)\n\tif packErr != nil {\n\t\treturn nil, nil, nil, packErr\n\t}\n\n\tawardTx, minerErr := utxoVM.GenerateAwardTx([]byte(\"miner-1\"), \"1\", []byte(\"award,onyeah!\"))\n\tif minerErr != nil {\n\t\treturn nil, nil, nil, minerErr\n\t}\n\ttxlist = append(txlist, awardTx)\n\tecdsdPk, _ := ecdsa.GenerateKey(elliptic.P256(), rand.Reader)\n\treturn txlist, tx, ecdsdPk, nil\n}", "func coinbaseGet(L *lua.LState) int {\n\tp := checkCoinBase(L, 1)\n\tfmt.Println(p)\n\n\treturn 0\n}", "func (cli *CommandLine) send(from, to string, amount int) {\n\tchain := blockchain.ContinueBlockChain(from)\n\tdefer chain.Database.Close()\n\n\ttx := blockchain.NewTransaction(from, to, amount, chain)\n\tchain.AddBlock([]*blockchain.Transaction{tx})\n\tfmt.Println(\"Success send token\")\n}", "func (s *Store) updateMinedBalance(ns walletdb.ReadWriteBucket, rec *TxRecord,\n\tblock *BlockMeta) er.R {\n\n\t// Add a debit record for each unspent credit spent by this transaction.\n\t// The index is set in each iteration below.\n\tspender := indexedIncidence{\n\t\tincidence: incidence{\n\t\t\ttxHash: rec.Hash,\n\t\t\tblock: block.Block,\n\t\t},\n\t}\n\n\tspentByAddress := map[string]btcutil.Amount{}\n\n\tfor i, input := range rec.MsgTx.TxIn {\n\t\tunspentKey, credKey := existsUnspent(ns, &input.PreviousOutPoint)\n\t\tif credKey == nil {\n\t\t\t// Debits for unmined transactions are not explicitly\n\t\t\t// tracked. Instead, all previous outputs spent by any\n\t\t\t// unmined transaction are added to a map for quick\n\t\t\t// lookups when it must be checked whether a mined\n\t\t\t// output is unspent or not.\n\t\t\t//\n\t\t\t// Tracking individual debits for unmined transactions\n\t\t\t// could be added later to simplify (and increase\n\t\t\t// performance of) determining some details that need\n\t\t\t// the previous outputs (e.g. determining a fee), but at\n\t\t\t// the moment that is not done (and a db lookup is used\n\t\t\t// for those cases instead). There is also a good\n\t\t\t// chance that all unmined transaction handling will\n\t\t\t// move entirely to the db rather than being handled in\n\t\t\t// memory for atomicity reasons, so the simplist\n\t\t\t// implementation is currently used.\n\t\t\tcontinue\n\t\t}\n\n\t\tprevAddr := \"unknown\"\n\t\tif prevPk, err := AddressForOutPoint(ns, &input.PreviousOutPoint); err != nil {\n\t\t\tlog.Warnf(\"Error decoding address spent from because [%s]\", err.String())\n\t\t} else if prevPk != nil {\n\t\t\tprevAddr = txscript.PkScriptToAddress(prevPk, s.chainParams).String()\n\t\t}\n\n\t\t// If this output is relevant to us, we'll mark the it as spent\n\t\t// and remove its amount from the store.\n\t\tspender.index = uint32(i)\n\t\tamt, err := spendCredit(ns, credKey, &spender)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\terr = putDebit(\n\t\t\tns, &rec.Hash, uint32(i), amt, &block.Block, credKey,\n\t\t)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif err := DeleteRawUnspent(ns, unspentKey); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tspentByAddress[prevAddr] += amt\n\t}\n\n\tfor addr, amt := range spentByAddress {\n\t\tlog.Infof(\"📩 %s [%s] from [%s] tx [%s] @ [%s]\",\n\t\t\tlog.GreenBg(\"Confirmed spend\"),\n\t\t\tlog.Coins(amt.ToBTC()),\n\t\t\tlog.Address(addr),\n\t\t\tlog.Txid(rec.Hash.String()),\n\t\t\tlog.Height(block.Height))\n\t}\n\n\treturn nil\n}", "func (c *Chain) GiveTxBackFromBlock() {\n\n}", "func (s SmesherService) Coinbase(context.Context, *empty.Empty) (*pb.CoinbaseResponse, error) {\n\tlog.Info(\"GRPC SmesherService.Coinbase\")\n\n\t_, _, coinbase, _ := s.Mining.MiningStats()\n\taddr, err := types.StringToAddress(coinbase)\n\tif err != nil {\n\t\tlog.Error(\"error converting coinbase: %s\", err)\n\t\treturn nil, status.Errorf(codes.Internal, \"error reading coinbase data\")\n\t}\n\treturn &pb.CoinbaseResponse{AccountId: &pb.AccountId{Address: addr.Bytes()}}, nil\n}", "func (s *ethereumPaymentObligation) sendMintTransaction(contract ethereumPaymentObligationContract, opts *bind.TransactOpts, requestData *MintRequest) error {\n\ttx, err := s.ethClient.SubmitTransactionWithRetries(contract.Mint, opts, requestData.To, requestData.TokenID, requestData.TokenURI, requestData.AnchorID,\n\t\trequestData.MerkleRoot, requestData.Values, requestData.Salts, requestData.Proofs)\n\tif err != nil {\n\t\treturn err\n\t}\n\tlog.Infof(\"Sent off tx to mint [tokenID: %x, anchor: %x, registry: %x] to payment obligation contract. Ethereum transaction hash [%x] and Nonce [%v] and Check [%v]\",\n\t\trequestData.TokenID, requestData.AnchorID, requestData.To, tx.Hash(), tx.Nonce(), tx.CheckNonce())\n\tlog.Infof(\"Transfer pending: 0x%x\\n\", tx.Hash())\n\treturn nil\n}", "func (s *Store) insertMinedTx(ns walletdb.ReadWriteBucket, rec *TxRecord,\n\tblock *BlockMeta) er.R {\n\t// this might get called on a tx which is already in the db\n\n\t// If a block record does not yet exist for any transactions from this\n\t// block, insert a block record first. Otherwise, update it by adding\n\t// the transaction hash to the set of transactions from this block.\n\tbr, err := fetchBlockRecord(ns, block.Height)\n\tif err != nil && !ErrNoExists.Is(err) {\n\t\treturn err\n\t}\n\tvar transactions []chainhash.Hash\n\tif br != nil {\n\t\ttransactions = make([]chainhash.Hash, 0, len(br.transactions)+1)\n\t\tfor _, txid := range br.transactions {\n\t\t\tif br.Block.Hash.IsEqual(&block.Block.Hash) {\n\t\t\t\tif !txid.IsEqual(&rec.Hash) {\n\t\t\t\t\ttransactions = append(transactions, txid)\n\t\t\t\t}\n\t\t\t}\n\t\t\t// Ideally we would do some sort of transaction rollback operation\n\t\t\t// but we're dealing with a corrupt db and rollbackTransaction()\n\t\t\t// is not going to rollback everything it can and ignore what it\n\t\t\t// can't, instead it will fail with an error when any of the things\n\t\t\t// it expects to be present are not. Also deleting the tx object\n\t\t\t// without all of the associated credits and debits is risky because\n\t\t\t// it can cause errors in other code. So we're just going to detach\n\t\t\t// it from the block and let it be with the caviats:\n\t\t\t//\n\t\t\t// 1. there may be dangling credits and no associated block record\n\t\t\t// ForEachUnspentOutput will filter these out\n\t\t\t// 2. there may be debits which spent credits that should not have\n\t\t\t// been spent.\n\t\t}\n\t}\n\ttransactions = append(transactions, rec.Hash)\n\tif err := putBlockRecord(ns, &blockRecord{\n\t\tBlock: block.Block,\n\t\tTime: block.Time,\n\t\ttransactions: transactions,\n\t}); err != nil {\n\t\treturn err\n\t}\n\n\t// no harm in putting the tx record again\n\tif err := putTxRecord(ns, rec, &block.Block); err != nil {\n\t\treturn err\n\t}\n\n\t// Determine if this transaction has affected our balance, and if so,\n\t// update it.\n\tif err := s.updateMinedBalance(ns, rec, block); err != nil {\n\t\treturn err\n\t}\n\n\t// If this transaction previously existed within the store as unmined,\n\t// we'll need to remove it from the unmined bucket.\n\tif v := existsRawUnmined(ns, rec.Hash[:]); v != nil {\n\t\tlog.Debugf(\"Marking unconfirmed transaction [%s] mined in block [%d]\",\n\t\t\trec.Hash.String(),\n\t\t\tblock.Height)\n\n\t\tif err := s.deleteUnminedTx(ns, rec); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// As there may be unconfirmed transactions that are invalidated by this\n\t// transaction (either being duplicates, or double spends), remove them\n\t// from the unconfirmed set. This also handles removing unconfirmed\n\t// transaction spend chains if any other unconfirmed transactions spend\n\t// outputs of the removed double spend.\n\tif err := s.removeDoubleSpends(ns, rec); err != nil {\n\t\treturn err\n\t}\n\n\t// Clear any locked outputs since we now have a confirmed spend for\n\t// them, making them not eligible for coin selection anyway.\n\tfor _, txIn := range rec.MsgTx.TxIn {\n\t\tif err := unlockOutput(ns, txIn.PreviousOutPoint); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (_IUniswapV2Pair *IUniswapV2PairTransactor) Mint(opts *bind.TransactOpts, to common.Address) (*types.Transaction, error) {\r\n\treturn _IUniswapV2Pair.contract.Transact(opts, \"mint\", to)\r\n}", "func (t *Transaction) IsCoinBase() bool {\n\t// A coin base must only have one transaction input.\n\tif len(t.Inputs) != 1 {\n\t\treturn false\n\t}\n\n\t// The previous output of a coin base must have a max value index and\n\t// a zero hash.\n\tprevOut := &t.Inputs[0].PreviousOutPoint\n\tif prevOut.Index != math.MaxUint32 || prevOut.Hash != zeroHash {\n\t\treturn false\n\t}\n\n\treturn true\n}", "func (_LifToken *LifTokenTransactor) Mint(opts *bind.TransactOpts, _to common.Address, _amount *big.Int) (*types.Transaction, error) {\n\treturn _LifToken.contract.Transact(opts, \"mint\", _to, _amount)\n}", "func (entry *UtxoEntry) IsCoinBase() bool {\n\treturn entry.packedFlags&utxoFlagCoinBase == utxoFlagCoinBase\n}", "func (_DetailedTestToken *DetailedTestTokenTransactor) Mint(opts *bind.TransactOpts, _to common.Address, _amount *big.Int) (*types.Transaction, error) {\n\treturn _DetailedTestToken.contract.Transact(opts, \"mint\", _to, _amount)\n}", "func main() {\n\tvar chainHead *Block\n\tgenesis := BlockData{Transactions: []string{\"S2E\", \"S2Z\"}}\n\t//fmt.Printf(\"helllooo\")\n\t//fmt.Println(genesis)\n\tchainHead = InsertBlock(genesis, chainHead)\n\t//var x string=CalculateHash(chainHead)\n\t//fmt.Printf(\"%x\\n\", x)\n\n\tsecondBlock := BlockData{Transactions: []string{\"E2Alice\", \"E2Bob\", \"S2John\"}}\n\tchainHead = InsertBlock(secondBlock, chainHead)\n\n\tListBlocks(chainHead)\n\n\t//ChangeBlock(\"S2E\", \"S2Trudy\", chainHead)\n\n\tListBlocks(chainHead)\n\n\tVerifyChain(chainHead)\n\n}", "func TestInsertMempoolTxAfterSpentOutput(t *testing.T) {\n\tt.Parallel()\n\n\tstore, db, teardown, err := testStore()\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\tdefer teardown()\n\n\t// First we add a confirmed transaction to the wallet.\n\tb100 := BlockMeta{\n\t\tBlock: Block{Height: 100},\n\t\tTime: time.Now(),\n\t}\n\tcb := newCoinBase(1e8)\n\tcbRec, err := NewTxRecordFromMsgTx(cb, b100.Time)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\tcommitDBTx(t, store, db, func(ns walletdb.ReadWriteBucket) {\n\t\tif err := store.InsertTx(ns, cbRec, &b100); err != nil {\n\t\t\tt.Fatal(err)\n\t\t}\n\t\terr := store.AddCredit(ns, cbRec, &b100, 0, false)\n\t\tif err != nil {\n\t\t\tt.Fatal(err)\n\t\t}\n\t})\n\n\t// Then create a transaction that spends the previous tx output.\n\tb101 := BlockMeta{\n\t\tBlock: Block{Height: 101},\n\t\tTime: time.Now(),\n\t}\n\tamt := int64(1e7)\n\tspend := spendOutput(&cbRec.Hash, 0, amt)\n\tspendRec, err := NewTxRecordFromMsgTx(spend, time.Now())\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tcommitDBTx(t, store, db, func(ns walletdb.ReadWriteBucket) {\n\t\t// We add the spending tx to the wallet as confirmed.\n\t\terr := store.InsertTx(ns, spendRec, &b101)\n\t\tif err != nil {\n\t\t\tt.Fatal(err)\n\t\t}\n\t\terr = store.AddCredit(ns, spendRec, &b101, 0, false)\n\t\tif err != nil {\n\t\t\tt.Fatal(err)\n\t\t}\n\n\t\t// We now adding the original transaction as mempool to simulate\n\t\t// a real case where trying to broadcast a tx after it has been\n\t\t// confirmed and spent.\n\t\tif err := store.InsertTx(ns, cbRec, nil); err != nil {\n\t\t\tt.Fatal(err)\n\t\t}\n\t\terr = store.AddCredit(ns, cbRec, nil, 0, false)\n\t\tif err != nil {\n\t\t\tt.Fatal(err)\n\t\t}\n\t})\n\n\t// now we check that there no unminedCredit exists for the original tx.\n\tcommitDBTx(t, store, db, func(ns walletdb.ReadWriteBucket) {\n\t\tk := canonicalOutPoint(&cbRec.Hash, 0)\n\t\tif existsRawUnminedCredit(ns, k) != nil {\n\t\t\tt.Fatalf(\"expected output to not exist \" +\n\t\t\t\t\"in unmined credit bucket\")\n\t\t}\n\t})\n}", "func (_Erc20Mock *Erc20MockTransactor) Mint(opts *bind.TransactOpts, to common.Address, value *big.Int) (*types.Transaction, error) {\n\treturn _Erc20Mock.contract.Transact(opts, \"mint\", to, value)\n}" ]
[ "0.7742158", "0.7670547", "0.7275985", "0.710107", "0.6993326", "0.6694833", "0.6483202", "0.6464114", "0.6412588", "0.6320174", "0.6320174", "0.62685657", "0.6245498", "0.6235786", "0.61825", "0.6181117", "0.6157046", "0.6126204", "0.6116468", "0.60582566", "0.6024877", "0.6010123", "0.5934115", "0.58344275", "0.58166015", "0.5714872", "0.5632067", "0.559181", "0.55896366", "0.55724066", "0.55154556", "0.54921854", "0.54693294", "0.54482275", "0.5435304", "0.54260033", "0.5407048", "0.54008657", "0.539322", "0.5379277", "0.53734034", "0.534367", "0.53284836", "0.5306337", "0.5301946", "0.52854675", "0.5283567", "0.52614206", "0.5223442", "0.52225107", "0.5220883", "0.5208523", "0.5206915", "0.52051747", "0.51973206", "0.5194759", "0.51901484", "0.5186883", "0.51831317", "0.51781213", "0.51746184", "0.51726884", "0.51726264", "0.51694113", "0.5151417", "0.51428145", "0.51397693", "0.51002836", "0.5100103", "0.5094566", "0.50848264", "0.507745", "0.5074212", "0.5074212", "0.5074017", "0.506902", "0.5067675", "0.50614727", "0.50597167", "0.5059496", "0.5055317", "0.5053916", "0.50525564", "0.50311625", "0.5026968", "0.5016972", "0.5009308", "0.49933028", "0.4991544", "0.49891257", "0.49853", "0.49848688", "0.4980553", "0.49794313", "0.4964069", "0.49479815", "0.49409753", "0.49405426", "0.49158064", "0.4901427" ]
0.70467657
4
Check is used to check if the tx belongs to me.
func (account *Account) Check(tx transaction.Transaction) bool { P := crypto.GetP(account.a, tx.RT, account.A) if P == tx.P { fmt.Println("\033[1;31;40m", "It belongs to us!", "\033[0m") tx.Setp(crypto.Getp(account.a, tx.RT)) account.Wallet[account.WalletIndex] = tx account.WalletIndex++ return true } return false }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (ok OKHandler) CheckTx(ctx sdk.Context, store state.SimpleDB, tx sdk.Tx) (res sdk.CheckResult, err error) {\n\treturn sdk.CheckResult{Log: ok.Log}, nil\n}", "func (EchoHandler) CheckTx(ctx sdk.Context, store state.SimpleDB, tx sdk.Tx) (res sdk.CheckResult, err error) {\n\tdata, err := data.ToWire(tx)\n\treturn sdk.CheckResult{Data: data}, err\n}", "func (s *TXPoolServer) checkTx(hash common.Uint256) bool {\n\t// Check if the tx is in pending list\n\ts.mu.RLock()\n\tif ok := s.allPendingTxs[hash]; ok != nil {\n\t\ts.mu.RUnlock()\n\t\treturn true\n\t}\n\ts.mu.RUnlock()\n\n\t// Check if the tx is in txn pool\n\tif res := s.txPool.GetTransaction(hash); res != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func CheckBuyerTx(w http.ResponseWriter, r *http.Request) {\n\tdefer r.Body.Close()\n\n\tvar query = r.URL.Query()\n\tsendedAmount := query.Get(\"received_amount\")\n\n\ttx := &model.Tx{Timestamp: query.Get(\"timestamp\"), Sender: query.Get(\"sender\"), Receiver: query.Get(\"receiver\")}\n\tamount, err := strconv.ParseFloat(sendedAmount, 64)\n\tif err != nil {\n\t\tjson.NewEncoder(w).Encode(&model.JsonResponse{\n\t\t\tSuccess: false,\n\t\t\tComment: err.Error(),\n\t\t})\n\t\treturn\n\t}\n\ttx.Amount = amount\n\n\tif query.Get(\"ticker\") == \"\" || tx.Sender == \"\" || tx.Receiver == \"\" || sendedAmount == \"\" {\n\t\tjson.NewEncoder(w).Encode(&model.JsonResponse{\n\t\t\tSuccess: false,\n\t\t\tComment: \"please provide required parameters (ticker, senderAddress, receiverAddress, sendedAmount)\",\n\t\t})\n\t\treturn\n\t}\n\n\tcoin, err := storage.GetSingleCoinExplorer(query.Get(\"ticker\"))\n\tif err != nil {\n\t\tjson.NewEncoder(w).Encode(&model.JsonResponse{\n\t\t\tSuccess: false,\n\t\t\tComment: err.Error(),\n\t\t})\n\t\treturn\n\t}\n\ttx.Explorer = coin.Explorer\n\ttx.Ticker = coin.Ticker\n\n\tif err = explorers.NewAggregator().CheckIfTxExist(tx);err == nil {\n\t\tjson.NewEncoder(w).Encode(&model.JsonResponse{\n\t\t\tSuccess: true,\n\t\t})\n\t\treturn\n\t} else {\n\t\tfmt.Println(err.Error())\n\n\n\t\tjson.NewEncoder(w).Encode(&model.JsonResponse{\n\t\t\tSuccess: false,\n\t\t})\n\t\treturn\n\t}\n\n}", "func (tbl DbCompoundTable) IsTx() bool {\n\t_, ok := tbl.db.(*sql.Tx)\n\treturn ok\n}", "func (Logger) CheckTx(ctx sdk.Context, store sdk.SimpleDB, tx interface{}, next sdk.Checker) (res sdk.CheckResult, err error) {\n\tstart := time.Now()\n\tres, err = next.CheckTx(ctx, store, tx)\n\tdelta := time.Now().Sub(start)\n\t// TODO: log some info on the tx itself?\n\tl := ctx.With(\"duration\", micros(delta))\n\tif err == nil {\n\t\tl.Debug(\"CheckTx\", \"log\", res.Log)\n\t} else {\n\t\tl.Info(\"CheckTx\", \"err\", err)\n\t}\n\treturn\n}", "func (c CheckHandler) CheckTx(ctx sdk.Context, store state.SimpleDB, tx sdk.Tx) (res sdk.CheckResult, err error) {\n\tcheck, ok := tx.Unwrap().(CheckTx)\n\tif !ok {\n\t\treturn res, errors.ErrUnknownTxType(tx)\n\t}\n\n\tfor _, perm := range check.Required {\n\t\tif !ctx.HasPermission(perm) {\n\t\t\treturn res, errors.ErrUnauthorized()\n\t\t}\n\t}\n\treturn res, nil\n}", "func (tbl RecordTable) IsTx() bool {\n\treturn tbl.db.IsTx()\n}", "func NeedThisTxExt(id *btc.Uint256, cb func()) (why_not int) {\n\tTxMutex.Lock()\n\tif _, present := TransactionsToSend[id.BIdx()]; present {\n\t\twhy_not = 1\n\t} else if _, present := TransactionsRejected[id.BIdx()]; present {\n\t\twhy_not = 2\n\t} else if _, present := TransactionsPending[id.BIdx()]; present {\n\t\twhy_not = 3\n\t} else if common.BlockChain.Unspent.TxPresent(id) {\n\t\twhy_not = 4\n\t\t// This assumes that tx's out #0 has not been spent yet, which may not always be the case, but well...\n\t\tcommon.CountSafe(\"TxAlreadyMined\")\n\t} else {\n\t\t// why_not = 0\n\t\tif cb != nil {\n\t\t\tcb()\n\t\t}\n\t}\n\tTxMutex.Unlock()\n\treturn\n}", "func (_TreasureHunt *TreasureHuntTransactor) IsTreasureHere(opts *bind.TransactOpts, latitude *big.Int, longitude *big.Int) (*types.Transaction, error) {\n\treturn _TreasureHunt.contract.Transact(opts, \"IsTreasureHere\", latitude, longitude)\n}", "func (app *JSONStoreApplication) CheckTx(tx types.RequestCheckTx) types.ResponseCheckTx {\n\n\tvar temp interface{}\n\terr := json.Unmarshal(tx.Tx, &temp)\n\n\tif err != nil {\n\t\treturn types.ResponseCheckTx{Code: code.CodeTypeEncodingError,Log: fmt.Sprint(err)}\n\t}\n\n\tmessage := temp.(map[string]interface{})\n\tPublicKey:= message[\"publicKey\"].(string)\n\tByteString:= message[\"msg\"].(string)\n\tSignaTure:= message[\"sig\"].(string) \n\tcount := checkUserPublic(db,PublicKey)\n\n if signatureValidate(PublicKey,SignaTure,ByteString) && count !=0 {\n return types.ResponseCheckTx{Code: code.CodeTypeOK}\n}\n return types.ResponseCheckTx{Code: code.CodeTypeBadData}\n}", "func (mp *TxPool) HaveTransaction(txId Uint256) bool {\n\tmp.RLock()\n\t_, ok := mp.txnList[txId]\n\tmp.RUnlock()\n\treturn ok\n}", "func (f FailHandler) CheckTx(ctx sdk.Context, store state.SimpleDB, tx sdk.Tx) (res sdk.CheckResult, err error) {\n\treturn res, errors.Wrap(f.Err)\n}", "func (g *Generator) CheckTx(statements []Statement) bool {\n\tvar hasTxStatement bool\n\tfor _, v := range statements {\n\t\tif v.Definition.IsTx {\n\t\t\thasTxStatement = true\n\t\t\tbreak\n\t\t}\n\t}\n\treturn hasTxStatement\n}", "func (mp *TxPool) haveTransaction(hash *chainhash.Hash) bool {\n\treturn mp.isTransactionInPool(hash) || mp.isOrphanInPool(hash)\n}", "func (app *App) txChecker() txChecker {\n\n\treturn func(msg RequestCheckTx) ResponseCheckTx {\n\t\tdefer app.handlePanic()\n\n\t\tif app.VerifyCache(msg.Tx) {\n\t\t\tloginfo := fmt.Sprintf(\"checkTx duplicated tx: %s\", hex.EncodeToString(utils.GetTransactionHash(msg.Tx)))\n\t\t\tapp.logger.Detail(loginfo)\n\t\t\treturn ResponseCheckTx{\n\t\t\t\tCode: CodeNotOK.uint32(),\n\t\t\t\tLog: loginfo,\n\t\t\t}\n\t\t}\n\n\t\tapp.Context.check.BeginTxSession()\n\n\t\ttx := &action.SignedTx{}\n\n\t\terr := serialize.GetSerializer(serialize.NETWORK).Deserialize(msg.Tx, tx)\n\t\tif err != nil {\n\t\t\tapp.logger.Errorf(\"checkTx failed to deserialize msg: %v, error: %s \", msg, err)\n\t\t}\n\t\ttxCtx := app.Context.Action(&app.header, app.Context.check)\n\t\thandler := txCtx.Router.Handler(tx.Type)\n\n\t\tgas := txCtx.State.ConsumedGas()\n\n\t\tok, err := handler.Validate(txCtx, *tx)\n\t\tif err != nil {\n\t\t\tapp.logger.Debug(\"Check Tx invalid: \", err.Error())\n\t\t\treturn ResponseCheckTx{\n\t\t\t\tCode: getCode(ok).uint32(),\n\t\t\t\tLog: err.Error(),\n\t\t\t}\n\t\t}\n\n\t\tok, response := handler.ProcessCheck(txCtx, tx.RawTx)\n\n\t\tfeeOk, feeResponse := handler.ProcessFee(txCtx, *tx, gas, storage.Gas(len(msg.Tx)), storage.Gas(response.GasUsed))\n\n\t\tlogString := marshalLog(ok, response, feeResponse)\n\n\t\tresult := ResponseCheckTx{\n\t\t\tCode: getCode(ok && feeOk).uint32(),\n\t\t\tData: response.Data,\n\t\t\tLog: logString,\n\t\t\tInfo: response.Info,\n\t\t\tGasWanted: feeResponse.GasWanted,\n\t\t\tGasUsed: feeResponse.GasUsed,\n\t\t\tEvents: response.Events,\n\t\t\tCodespace: \"\",\n\t\t}\n\n\t\tif !(ok && feeOk) {\n\t\t\tapp.Context.check.DiscardTxSession()\n\t\t} else {\n\t\t\tapp.Context.check.CommitTxSession()\n\t\t}\n\n\t\tapp.logger.Detail(\"Check Tx: \", result, \"log\", response.Log)\n\t\treturn result\n\n\t}\n}", "func (tbl AssociationTable) IsTx() bool {\n\treturn tbl.db.IsTx()\n}", "func (g GoroutineLock) Check() {\n\tif getGoroutineID() != goroutineID(g) {\n\t\tpanic(\"running on the wrong goroutine\")\n\t}\n}", "func (this *Node) IsMe(nodeName string) bool {\n\treturn this.NodeInfo.Name == nodeName\n}", "func checkValidTx(t *testing.T, anteHandler sdk.AnteHandler, ctx sdk.Context, tx sdk.Tx, simulate bool) {\n\t_, result, abort := anteHandler(ctx, tx, simulate)\n\trequire.False(t, abort)\n\trequire.Equal(t, sdk.CodeOK, result.Code)\n\trequire.True(t, result.IsOK())\n}", "func checkValidTx(t *testing.T, anteHandler sdk.AnteHandler, ctx sdk.Context, tx sdk.Tx, simulate bool) {\n\t_, result, abort := anteHandler(ctx, tx, simulate)\n\trequire.False(t, abort)\n\trequire.Equal(t, sdk.CodeOK, result.Code)\n\trequire.True(t, result.IsOK())\n}", "func (p *PromptStore) Check(_ string) bool {\n\treturn false\n}", "func (as *accountSet) containsTx(tx *types.Transaction) bool {\n\tif addr, err := types.Sender(as.signer, tx); err == nil {\n\t\treturn as.contains(addr)\n\t}\n\treturn false\n}", "func (_TreasureHunt *TreasureHuntSession) IsTreasureHere(latitude *big.Int, longitude *big.Int) (*types.Transaction, error) {\n\treturn _TreasureHunt.Contract.IsTreasureHere(&_TreasureHunt.TransactOpts, latitude, longitude)\n}", "func (_TreasureHunt *TreasureHuntTransactorSession) IsTreasureHere(latitude *big.Int, longitude *big.Int) (*types.Transaction, error) {\n\treturn _TreasureHunt.Contract.IsTreasureHere(&_TreasureHunt.TransactOpts, latitude, longitude)\n}", "func (blockchain *Blockchain) CheckTx(req abciTypes.RequestCheckTx) abciTypes.ResponseCheckTx {\n\tresponse := blockchain.executor.RunTx(blockchain.CurrentState(), req.Tx, nil, blockchain.Height()+1, blockchain.currentMempool, blockchain.MinGasPrice(), true)\n\n\treturn abciTypes.ResponseCheckTx{\n\t\tCode: response.Code,\n\t\tData: response.Data,\n\t\tLog: response.Log,\n\t\tInfo: response.Info,\n\t\tGasWanted: response.GasWanted,\n\t\tGasUsed: response.GasUsed,\n\t\tEvents: []abciTypes.Event{\n\t\t\t{\n\t\t\t\tType: \"tags\",\n\t\t\t\tAttributes: response.Tags,\n\t\t\t},\n\t\t},\n\t}\n}", "func (database *Database) CheckVerification(userID, code string) (bool, error) {\n\ttx, err := database.db.Begin(false)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\tidToVerify, err := tx.Get(fmt.Sprintf(VerificationCodeToUser, code))\n\tif err == buntdb.ErrNotFound {\n\t\treturn false, tx.Rollback()\n\t} else if err != nil {\n\t\t_ = tx.Rollback()\n\t\treturn false, err\n\t}\n\n\treturn idToVerify == userID, tx.Rollback()\n}", "func (app *BaseApp) CheckTx(req abci.RequestCheckTx) (res abci.ResponseCheckTx) {\n\t// Decode the Tx.\n\tvar result types.Result\n\tvar tx, err = types.DecoderTx(app.cdc, req.Tx)\n\n\tif err != nil {\n\t\treturn toResponseCheckTx(err.Result())\n\t}\n\n\n\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\tswitch r.(type) {\n\t\t\tcase types.ErrorOutOfGas:\n\t\t\t\tlog := \"checkTxStd out of gas\"\n\t\t\t\tresult = types.ErrOutOfGas(log).Result()\n\t\t\tdefault:\n\t\t\t\tlog := fmt.Sprintf(\"checkTxStd recovered: %v\\nstack:\\n%v\", r, string(debug.Stack()))\n\t\t\t\tresult = types.ErrInternal(log).Result()\n\t\t\t}\n\n\t\t\tres = toResponseCheckTx(result)\n\t\t}\n\t}()\n\n\t// 初始化context相关数据\n\tctx := app.checkState.ctx.WithTxBytes(req.Tx)\n\n\tswitch implTx := tx.(type) {\n\tcase *txs.TxStd:\n\t\tctx = setGasMeter(ctx, implTx)\n\t\tresult, _ = app.checkTxStd(ctx, implTx, \"\")\n\tcase *txs.TxQcp:\n\t\tctx = setGasMeter(ctx, implTx.TxStd)\n\t\tresult = app.checkTxQcp(ctx, implTx)\n\tdefault:\n\t\tresult = types.ErrInternal(\"not support itx type\").Result()\n\t}\n\n\treturn toResponseCheckTx(result)\n}", "func (uv *UtxoVM) verifyUTXOPermission(tx *pb.Transaction, verifiedID map[string]bool) (bool, error) {\n\t// verify tx input\n\tconUtxoInputs, err := xmodel.ParseContractUtxoInputs(tx)\n\tif err != nil {\n\t\tuv.xlog.Warn(\"verifyUTXOPermission error, parseContractUtxo \")\n\t\treturn false, ErrParseContractUtxos\n\t}\n\tconUtxoInputsMap := map[string]bool{}\n\tfor _, conUtxoInput := range conUtxoInputs {\n\t\taddr := conUtxoInput.GetFromAddr()\n\t\ttxid := conUtxoInput.GetRefTxid()\n\t\toffset := conUtxoInput.GetRefOffset()\n\t\tutxoKey := genUtxoKey(addr, txid, offset)\n\t\tconUtxoInputsMap[utxoKey] = true\n\t}\n\n\tfor _, txInput := range tx.TxInputs {\n\t\t// if transfer from contract\n\t\taddr := txInput.GetFromAddr()\n\t\ttxid := txInput.GetRefTxid()\n\t\toffset := txInput.GetRefOffset()\n\t\tutxoKey := genUtxoKey(addr, txid, offset)\n\t\tif conUtxoInputsMap[utxoKey] {\n\t\t\t// this utxo transfer from contract, will verify in rwset verify\n\t\t\tcontinue\n\t\t}\n\n\t\tname := string(txInput.FromAddr)\n\t\tif verifiedID[name] {\n\t\t\t// this ID(either AK or Account) is verified before\n\t\t\tcontinue\n\t\t}\n\t\takType := acl.IsAccount(name)\n\t\tif akType == 1 {\n\t\t\t// Identify account\n\t\t\tacl, err := uv.queryAccountACL(name)\n\t\t\tif err != nil || acl == nil {\n\t\t\t\t// valid account should have ACL info, so this account might not exsit\n\t\t\t\tuv.xlog.Warn(\"verifyUTXOPermission error, account might not exist\", \"account\", name, \"error\", err)\n\t\t\t\treturn false, ErrInvalidAccount\n\t\t\t}\n\t\t\tif ok, err := pm.IdentifyAccount(string(name), tx.AuthRequire, uv.aclMgr); !ok {\n\t\t\t\tuv.xlog.Warn(\"verifyUTXOPermission error, failed to IdentifyAccount\", \"error\", err)\n\t\t\t\treturn false, ErrACLNotEnough\n\t\t\t}\n\t\t} else if akType == 0 {\n\t\t\t// Identify address failed, if address not in verifiedID then it must have no signature\n\t\t\tuv.xlog.Warn(\"verifyUTXOPermission error, address has no signature\", \"address\", name)\n\t\t\treturn false, ErrInvalidSignature\n\t\t} else {\n\t\t\tuv.xlog.Warn(\"verifyUTXOPermission error, Invalid account/address name\", \"name\", name)\n\t\t\treturn false, ErrInvalidAccount\n\t\t}\n\t\tverifiedID[name] = true\n\t}\n\treturn true, nil\n}", "func (txh runMsgsTxHandler) CheckTx(ctx context.Context, req tx.Request, checkReq tx.RequestCheckTx) (tx.Response, tx.ResponseCheckTx, error) {\n\t// Don't run Msgs during CheckTx.\n\treturn tx.Response{}, tx.ResponseCheckTx{}, nil\n}", "func (m *TrxMgr) checkTrx(e *TrxEntry, blockTime uint32, checkTapos bool) (err error) {\n\tif err = e.CheckExpiration(blockTime); err != nil {\n\t\treturn err\n\t}\n\tif checkTapos {\n\t\tif err = e.CheckTapos(m.tapos); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tif err = e.CheckSignerKey(m.auth); err != nil {\n\t\treturn err\n\t}\n\tif err = e.CheckInBlockTrxs(m.history); err != nil {\n\t\treturn err\n\t}\n\treturn\n}", "func (server *Server) checkIfTxnPossible(txn *common.TransferTxn) bool {\n\tblockchainPrint := utils.GetBlockchainPrint(server.Blockchain)\n\tbalance := server.getLocalBalance()\n\tlog.WithFields(log.Fields{\n\t\t\"currentLocalLog\": utils.GetLocalLogPrint(server.Log),\n\t\t\"blockchain\": blockchainPrint,\n\t\t\"newTxn\": txn.Amount,\n\t\t\"balance\": balance,\n\t}).Info(\"checking transaction possibility.....\")\n\tif balance < txn.Amount {\n\t\tlog.Error(\"transaction not possible\")\n\t\treturn false\n\t} else {\n\t\treturn true\n\t}\n}", "func (a *Account) Check() bool {\n\treturn len(a.KeyID) > 0 && len(a.Key) > 0\n}", "func (uv *UtxoVM) checkRelyOnMarkedTxid(reftxid []byte, blockid []byte) (bool, bool, error) {\n\tisRely := false\n\treftx, err := uv.ledger.QueryTransaction(reftxid)\n\tif err != nil {\n\t\treturn true, isRely, nil\n\t}\n\tif reftx.GetModifyBlock() != nil && reftx.ModifyBlock.Marked {\n\t\tisRely = true\n\t\tif string(blockid) != \"\" {\n\t\t\tib, err := uv.ledger.QueryBlock(blockid)\n\t\t\tif err != nil {\n\t\t\t\treturn false, isRely, err\n\t\t\t}\n\t\t\tif ib.Height <= reftx.ModifyBlock.EffectiveHeight {\n\t\t\t\treturn true, isRely, nil\n\t\t\t}\n\t\t}\n\t\treturn false, isRely, nil\n\t}\n\treturn true, isRely, nil\n}", "func (self ID) Check() error {\n if !self.Valid() {\n return fmt.Errorf(\"Invalid ID: %#v\", self)\n }\n\n return nil\n}", "func (l *LedgerState) CheckTransaction(transaction *ledgerstate.Transaction) (err error) {\n\treturn l.UTXODAG.CheckTransaction(transaction)\n}", "func (mp *TxPool) HaveTransaction(hash *chainhash.Hash) bool {\n\t// Protect concurrent access.\n\tmp.mtx.RLock()\n\thaveTx := mp.haveTransaction(hash)\n\tmp.mtx.RUnlock()\n\n\treturn haveTx\n}", "func isOnMempool(uTxOut UTxOut) bool {\n\texists := false\nOuter:\n\tfor _, tx := range Mempool().Txs {\n\t\tfor _, txIn := range tx.TxIns {\n\t\t\tif txIn.TxId == uTxOut.TxId && txIn.Index == uTxOut.Index {\n\t\t\t\texists = true // uTxOut is already being used on the mempool\n\t\t\t\tbreak Outer // using labels, break from outer for loop\n\t\t\t}\n\t\t}\n\t}\n\treturn exists\n}", "func (n *Node) HandleTx(ctx context.Context, tx *wire.MsgTx) (bool, error) {\n\tctx = logger.ContextWithOutLogSubSystem(ctx)\n\n\tt, err := n.wallet.GetTx(ctx, *tx.TxHash())\n\tif err == nil && t != nil {\n\t\treturn true, nil // already have tx. this happens when reprocessing a block\n\t}\n\n\tfor index, input := range tx.TxIn {\n\t\t// Check for owned public keys in unlock scripts.\n\t\tpubkeys, err := bitcoin.PubKeysFromSigScript(input.SignatureScript)\n\t\tif err != nil {\n\t\t\treturn false, err\n\t\t}\n\t\tpkhs := make([]bitcoin.Hash20, 0, len(pubkeys))\n\t\tfor _, pubkey := range pubkeys {\n\t\t\tpkh, err := bitcoin.NewHash20FromData(pubkey)\n\t\t\tif err != nil {\n\t\t\t\treturn false, err\n\t\t\t}\n\n\t\t\tpkhs = append(pkhs, *pkh)\n\t\t}\n\n\t\tm, ra := n.wallet.AreHashesMonitored(pkhs)\n\t\tif m {\n\t\t\taddress := bitcoin.NewAddressFromRawAddress(ra, n.cfg.Net)\n\t\t\tlogger.Info(ctx, \"Found tx input %d for %s : %s\", index, address.String(),\n\t\t\t\ttx.TxHash().String())\n\t\t\tif err := n.PreprocessTx(ctx, tx); err != nil {\n\t\t\t\tlogger.Error(ctx, \"Failed preprocess : %s : %s\", err, tx.TxHash().String())\n\t\t\t} else {\n\t\t\t\tn.wallet.AddWireTx(ctx, tx)\n\t\t\t}\n\t\t\treturn true, nil\n\t\t}\n\n\t\t// Check for owned utxos.\n\t\tutxo, err := n.wallet.FindUTXO(ctx, input.PreviousOutPoint.Hash, input.PreviousOutPoint.Index)\n\t\tif err != nil {\n\t\t\treturn false, err\n\t\t}\n\t\tif utxo != nil {\n\t\t\tlogger.Info(ctx, \"Found tx input %d for utxo %s %d : %s\", index,\n\t\t\t\tutxo.UTXO.Hash.String(), utxo.UTXO.Index, tx.TxHash().String())\n\t\t\tif err := n.PreprocessTx(ctx, tx); err != nil {\n\t\t\t\tlogger.Error(ctx, \"Failed preprocess : %s : %s\", err, tx.TxHash().String())\n\t\t\t} else {\n\t\t\t\tn.wallet.AddWireTx(ctx, tx)\n\t\t\t}\n\t\t\treturn true, nil\n\t\t}\n\t}\n\n\tfor index, output := range tx.TxOut {\n\t\t// Check for owned public keys or public key hashes in locking scripts.\n\t\tpkhs, err := bitcoin.PKHsFromLockingScript(output.PkScript)\n\t\tif err != nil {\n\t\t\treturn false, err\n\t\t}\n\n\t\tm, ra := n.wallet.AreHashesMonitored(pkhs)\n\t\tif m {\n\t\t\taddress := bitcoin.NewAddressFromRawAddress(ra, n.cfg.Net)\n\t\t\tlogger.Info(ctx, \"Found tx output %d for %s : %s\", index, address.String(),\n\t\t\t\ttx.TxHash().String())\n\t\t\tif err := n.PreprocessTx(ctx, tx); err != nil {\n\t\t\t\tlogger.Error(ctx, \"Failed preprocess : %s : %s\", err, tx.TxHash().String())\n\t\t\t} else {\n\t\t\t\tn.wallet.AddWireTx(ctx, tx)\n\t\t\t}\n\t\t\treturn true, nil\n\t\t}\n\n\t\t// Check for flags for known relationships.\n\t\tflag, err := protocol.DeserializeFlagOutputScript(output.PkScript)\n\t\tif err == nil {\n\t\t\tr := n.rs.FindRelationshipForFlag(ctx, flag)\n\t\t\tif r != nil {\n\t\t\t\tlogger.Info(ctx, \"Found tx output %d for flag %x : %s\", index, flag,\n\t\t\t\t\ttx.TxHash().String())\n\t\t\t\tif err := n.PreprocessTx(ctx, tx); err != nil {\n\t\t\t\t\tlogger.Error(ctx, \"Failed preprocess : %s : %s\", err, tx.TxHash().String())\n\t\t\t\t} else {\n\t\t\t\t\tn.wallet.AddWireTx(ctx, tx)\n\t\t\t\t}\n\t\t\t\treturn true, nil\n\t\t\t}\n\t\t}\n\t}\n\n\treturn false, nil\n}", "func (a Account) CheckOwner () string {\n\treturn a.owner\n}", "func (v *Verifier) isSpent(spentKey string, simulator ledger.LedgerReader) (bool, error) {\n\tverifierLogger.Debugf(\"checking if input with ID '%s' has been spent\", spentKey)\n\tresult, err := simulator.GetState(tokenNameSpace, spentKey)\n\treturn result != nil, err\n}", "func (app *JSONStateApplication) CheckTx(txBytes []byte) types.ResponseCheckTx {\n\ttx := &transaction.Transaction{}\n\tif err := tx.FromBytes(txBytes); err != nil{\n\t\tfmt.Println(err)\n\t\treturn types.ResponseCheckTx{Code: code.CodeTypeBadData}\n\t}\n\t\n\tswitch tx.Type{\n\t\tcase transaction.AccountAdd:\n\t\t\t{\n\t\t\t\tif err := checkAccountAddTransaction(tx, app.store); err != nil {\n\t\t\t\t\treturn types.ResponseCheckTx{Code: code.CodeTypeBadData, Log:err.Error()}\n\t\t\t\t}\n\t\t\t}\n\n\t\tcase transaction.AccountDel:\n\t\t\t{\n\t\t\t\tif err := checkAccountDelTransaction(tx, app.store); err != nil {\n\t\t\t\t\treturn types.ResponseCheckTx{Code: code.CodeTypeBadData, Log:err.Error()}\n\t\t\t\t}\n\t\t\t}\n\t\tcase transaction.AttendPatient:\n\t\t\t{\n\t\t\t\tif err := checkAttendPatientTransaction(tx, app.db); err != nil {\n\t\t\t\t\treturn types.ResponseCheckTx{Code: code.CodeTypeBadData, Log:err.Error()}\n\t\t\t\t}\n\t\t\t}\n\t\tcase transaction.SecretAdd:\n\t\t\t{\n\t\t\t\tif err := checkSecretAddTransaction(tx, app.store); err != nil {\n\t\t\t\t\treturn types.ResponseCheckTx{Code: code.CodeTypeBadData, Log:err.Error()}\n\t\t\t\t}\n\t\t\t}\n\t\tcase transaction.SecretUpdate:\n\t\t\t{\n\t\t\tif err := checkSecretUpdateTransaction(tx, app.store); err != nil {\n\t\t\t\t\treturn types.ResponseCheckTx{Code: code.CodeTypeBadData, Log:err.Error()}\n\t\t\t\t}\n\t\t\t}\n\t\tcase transaction.SecretDel:\n\t\t\t{\n\t\t\t\tif err := checkSecretDelTransaction(tx, app.store); err != nil {\n\t\t\t\t\treturn types.ResponseCheckTx{Code: code.CodeTypeBadData, Log:err.Error()}\n\t\t\t\t}\n\t\t\t}\n\t\tcase transaction.SecretShare:\n\t\t\t{\n\t\t\t\tif err := checkSecretShareTransaction(tx, app.store); err != nil {\n\t\t\t\t\treturn types.ResponseCheckTx{Code: code.CodeTypeBadData, Log:err.Error()}\n\t\t\t\t}\n\t\t\t}\n\t\tcase transaction.PatientAdd:\n\t\t\t{\n\t\t\t\tif err := checkPatientAddTransaction(tx, app.store); err != nil {\n\t\t\t\t\treturn types.ResponseCheckTx{Code: code.CodeTypeBadData, Log:err.Error()}\n\t\t\t\t}\n\t\t\t}\n\t\tcase transaction.DoctorAdd:\n\t\t\t{\n\t\t\t\tif err := checkDoctorAddTransaction(tx, app.store); err != nil {\n\t\t\t\t\treturn types.ResponseCheckTx{Code: code.CodeTypeBadData, Log:err.Error()}\n\t\t\t\t}\n\t\t\t}\n\t\tcase transaction.MedicalAppointmentAdd:\n\t\t\t{\n\t\t\t\tif err := checkMedicalAppointmentAddTransaction(tx, app.db); err != nil {\n\t\t\t\t\treturn types.ResponseCheckTx{Code: code.CodeTypeBadData, Log:err.Error()}\n\t\t\t\t}\n\t\t\t}\n\t}\n\tcodeType := code.CodeTypeOK\n\n\treturn types.ResponseCheckTx{Code: codeType}\n}", "func (s *SendError) IsTransactionAlreadyInMempool() bool {\n\treturn s.is(TransactionAlreadyInMempool)\n}", "func (tu *TransactionUpdate) check() error {\n\tif _, ok := tu.mutation.DetailID(); tu.mutation.DetailCleared() && !ok {\n\t\treturn errors.New(`ent: clearing a required unique edge \"Transaction.detail\"`)\n\t}\n\treturn nil\n}", "func checkTransferTx(txId string) (string, error) {\n\tif \"\" == txId {\n\t\treturn \"\", fault.TransactionHexDataIsRequired\n\t}\n\n\treturn txId, nil\n}", "func (q *MaxLevel) Check(task *task.Task) bool {\n\tif task == nil {\n\t\treturn false\n\t}\n\n\treturn q.CheckSubQuery(task, task.Level < q.level)\n}", "func (h UpdateHandler) Check(ctx weave.Context, store weave.KVStore,\n\ttx weave.Tx) (weave.CheckResult, error) {\n\tvar res weave.CheckResult\n\t_, err := h.validate(ctx, store, tx)\n\treturn res, err\n}", "func (tuo *TransactionUpdateOne) check() error {\n\tif _, ok := tuo.mutation.DetailID(); tuo.mutation.DetailCleared() && !ok {\n\t\treturn errors.New(`ent: clearing a required unique edge \"Transaction.detail\"`)\n\t}\n\treturn nil\n}", "func (p *Privmsg) IsFromSelf() bool {\n\t// if the from user is empty, then this is a self privmsg\n\treturn p.From == User{}\n}", "func (mr *MockMempoolMockRecorder) CheckTx(tx, callback, txInfo interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"CheckTx\", reflect.TypeOf((*MockMempool)(nil).CheckTx), tx, callback, txInfo)\n}", "func (g GoroutineLock) CheckNotOn() {\n\tif getGoroutineID() == goroutineID(g) {\n\t\tpanic(\"running on the wrong goroutine\")\n\t}\n}", "func (app *BftApplication) CheckTx(tx []byte) types.ResponseCheckTx {\n\treturn types.ResponseCheckTx{Code: code.CodeTypeOK}\n\t//if cpcash.validate(\"BFTXafe2242d45cc5e54041b2b52913ef9a1aede4998a32e3fee128cf7d1e7575a41\") {\n\t//\treturn types.ResponseCheckTx{Code: code.CodeTypeOK}\n\t//}\n\t//return types.ResponseCheckTx{Code: code.NotPaid}\n\n}", "func (wc *WalletCreate) check() error {\n\tif _, ok := wc.mutation.Seed(); !ok {\n\t\treturn &ValidationError{Name: \"seed\", err: errors.New(`ent: missing required field \"Wallet.seed\"`)}\n\t}\n\tif v, ok := wc.mutation.Seed(); ok {\n\t\tif err := wallet.SeedValidator(v); err != nil {\n\t\t\treturn &ValidationError{Name: \"seed\", err: fmt.Errorf(`ent: validator failed for field \"Wallet.seed\": %w`, err)}\n\t\t}\n\t}\n\tif v, ok := wc.mutation.Representative(); ok {\n\t\tif err := wallet.RepresentativeValidator(v); err != nil {\n\t\t\treturn &ValidationError{Name: \"representative\", err: fmt.Errorf(`ent: validator failed for field \"Wallet.representative\": %w`, err)}\n\t\t}\n\t}\n\tif _, ok := wc.mutation.Encrypted(); !ok {\n\t\treturn &ValidationError{Name: \"encrypted\", err: errors.New(`ent: missing required field \"Wallet.encrypted\"`)}\n\t}\n\tif _, ok := wc.mutation.Work(); !ok {\n\t\treturn &ValidationError{Name: \"work\", err: errors.New(`ent: missing required field \"Wallet.work\"`)}\n\t}\n\tif _, ok := wc.mutation.CreatedAt(); !ok {\n\t\treturn &ValidationError{Name: \"created_at\", err: errors.New(`ent: missing required field \"Wallet.created_at\"`)}\n\t}\n\treturn nil\n}", "func (m Middleware) CheckTx(ctx sdk.Context, store state.SimpleDB, tx sdk.Tx, next sdk.Checker) (res sdk.CheckResult, err error) {\n\t// if it is not a PostPacket, just let it go through\n\tpost, ok := tx.Unwrap().(PostPacketTx)\n\tif !ok {\n\t\treturn next.CheckTx(ctx, store, tx)\n\t}\n\n\t// parse this packet and get the ibc-enhanced tx and context\n\tictx, itx, err := m.verifyPost(ctx, store, post)\n\tif err != nil {\n\t\treturn res, err\n\t}\n\treturn next.CheckTx(ictx, store, itx)\n}", "func isVoteTx(tx *wire.MsgTx) bool {\n\tif len(tx.TxOut) < 3 {\n\t\treturn false\n\t}\n\ttxOut := tx.TxOut[2]\n\tscriptClass := txscript.GetScriptClass(txOut.Version, txOut.PkScript)\n\treturn scriptClass == txscript.StakeGenTy\n}", "func (d *Discord) IsMe(message Message) bool {\n\tif d.Session.State.User == nil {\n\t\treturn false\n\t}\n\treturn message.UserID() == d.Session.State.User.ID\n}", "func (d *Discord) IsMe(message Message) bool {\n\tif d.Session.State.User == nil {\n\t\treturn false\n\t}\n\treturn message.UserID() == d.Session.State.User.ID\n}", "func (t *Commit) DBExistsTx(ctx context.Context, tx Tx) (bool, error) {\n\tq := \"SELECT `id` FROM `commit` WHERE `id` = ? LIMIT 1\"\n\tvar _ID sql.NullString\n\terr := tx.QueryRowContext(ctx, q, orm.ToSQLString(t.ID)).Scan(&_ID)\n\tif err != nil && err != sql.ErrNoRows {\n\t\treturn false, err\n\t}\n\treturn _ID.Valid, nil\n}", "func (s *orm) Check(ctx context.Context) {\n\ts.client.DB().PingContext(ctx)\n}", "func (this *DefaultNode) IsMe(nodeName string) bool {\n\treturn this.NodeInfo.Name == nodeName\n}", "func (a API) GetTxOutChk() (isNew bool) {\n\tselect {\n\tcase o := <-a.Ch.(chan GetTxOutRes):\n\t\tif o.Err != nil {\n\t\t\ta.Result = o.Err\n\t\t} else {\n\t\t\ta.Result = o.Res\n\t\t}\n\t\tisNew = true\n\tdefault:\n\t}\n\treturn\n}", "func (uv *UtxoVM) verifyRelyOnMarkedTxs(tx *pb.Transaction) (bool, bool, error) {\n\tisRelyOnMarkedTx := false\n\tfor _, txInput := range tx.GetTxInputs() {\n\t\treftxid := txInput.RefTxid\n\t\tif string(reftxid) == \"\" {\n\t\t\tcontinue\n\t\t}\n\t\tok, isRelyOnMarkedTx, err := uv.checkRelyOnMarkedTxid(reftxid, tx.Blockid)\n\t\tif err != nil || !ok {\n\t\t\treturn ok, isRelyOnMarkedTx, err\n\t\t}\n\t}\n\tfor _, txIn := range tx.GetTxInputsExt() {\n\t\treftxid := txIn.RefTxid\n\t\tif string(reftxid) == \"\" {\n\t\t\tcontinue\n\t\t}\n\t\tok, isRelyOnMarkedTx, err := uv.checkRelyOnMarkedTxid(reftxid, tx.Blockid)\n\t\tif !ok || err != nil {\n\t\t\treturn ok, isRelyOnMarkedTx, err\n\t\t}\n\t}\n\n\treturn true, isRelyOnMarkedTx, nil\n}", "func (t *Transaction) IsCoinbase() bool {\n\t// Check to see there is just 1 input and that it is not linked to any other transactions\n\treturn len(t.Inputs) == 1 && len(t.Inputs[0].ID) == 0 && t.Inputs[0].Out == -1\n}", "func UtxoExists(ctx context.Context, exec boil.ContextExecutor, rowid int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"utxo\\\" where \\\"rowid\\\"=$1 limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, rowid)\n\t}\n\n\trow := exec.QueryRowContext(ctx, sql, rowid)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if utxo exists\")\n\t}\n\n\treturn exists, nil\n}", "func (a API) GetRawTransactionChk() (isNew bool) {\n\tselect {\n\tcase o := <-a.Ch.(chan GetRawTransactionRes):\n\t\tif o.Err != nil {\n\t\t\ta.Result = o.Err\n\t\t} else {\n\t\t\ta.Result = o.Res\n\t\t}\n\t\tisNew = true\n\tdefault:\n\t}\n\treturn\n}", "func (uv *UtxoVM) verifyTxRWSets(tx *pb.Transaction) (bool, error) {\n\tif uv.verifyReservedWhitelist(tx) {\n\t\tuv.xlog.Info(\"verifyReservedWhitelist true\", \"txid\", fmt.Sprintf(\"%x\", tx.GetTxid()))\n\t\treturn true, nil\n\t}\n\n\treq := tx.GetContractRequests()\n\treservedRequests, err := uv.getReservedContractRequests(tx.GetContractRequests(), false)\n\tif err != nil {\n\t\tuv.xlog.Error(\"getReservedContractRequests error\", \"error\", err.Error())\n\t\treturn false, err\n\t}\n\n\tif !uv.verifyReservedContractRequests(reservedRequests, req) {\n\t\tuv.xlog.Error(\"verifyReservedContractRequests error\", \"reservedRequests\", reservedRequests, \"req\", req)\n\t\treturn false, fmt.Errorf(\"verify reservedContracts error\")\n\t}\n\n\tif req == nil {\n\t\tif tx.GetTxInputsExt() != nil || tx.GetTxOutputsExt() != nil {\n\t\t\tuv.xlog.Error(\"verifyTxRWSets error\", \"error\", ErrInvalidTxExt.Error())\n\t\t\treturn false, ErrInvalidTxExt\n\t\t}\n\t\treturn true, nil\n\t}\n\t// transfer in contract\n\ttransContractName, transAmount, err := txn.ParseContractTransferRequest(req)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\tenv, err := uv.model3.PrepareEnv(tx)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tcontextConfig := &contract.ContextConfig{\n\t\tXMCache: env.GetModelCache(),\n\t\tInitiator: tx.GetInitiator(),\n\t\tAuthRequire: tx.GetAuthRequire(),\n\t\tContractName: \"\",\n\t\tCore: contractChainCore{\n\t\t\tManager: uv.aclMgr,\n\t\t\tUtxoVM: uv,\n\t\t\tLedger: uv.ledger,\n\t\t},\n\t\tBCName: uv.bcname,\n\t}\n\tgasLimit, err := getGasLimitFromTx(tx)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tuv.xlog.Trace(\"get gas limit from tx\", \"gasLimit\", gasLimit, \"txid\", hex.EncodeToString(tx.Txid))\n\n\t// get gas rate to utxo\n\tgasPrice := uv.GetGasPrice()\n\n\tfor i, tmpReq := range tx.GetContractRequests() {\n\t\tmoduleName := tmpReq.GetModuleName()\n\t\tvm, err := uv.vmMgr3.GetVM(moduleName)\n\t\tif err != nil {\n\t\t\treturn false, err\n\t\t}\n\n\t\tlimits := contract.FromPbLimits(tmpReq.GetResourceLimits())\n\t\tif i >= len(reservedRequests) {\n\t\t\tgasLimit -= limits.TotalGas(gasPrice)\n\t\t}\n\t\tif gasLimit < 0 {\n\t\t\tuv.xlog.Error(\"virifyTxRWSets error:out of gas\", \"contractName\", tmpReq.GetContractName(),\n\t\t\t\t\"txid\", hex.EncodeToString(tx.Txid))\n\t\t\treturn false, errors.New(\"out of gas\")\n\t\t}\n\t\tcontextConfig.ResourceLimits = limits\n\t\tcontextConfig.ContractName = tmpReq.GetContractName()\n\t\tif transContractName == tmpReq.GetContractName() {\n\t\t\tcontextConfig.TransferAmount = transAmount.String()\n\t\t} else {\n\t\t\tcontextConfig.TransferAmount = \"\"\n\t\t}\n\n\t\tctx, err := vm.NewContext(contextConfig)\n\t\tif err != nil {\n\t\t\t// FIXME zq @icexin: need to return contract not found\n\t\t\tuv.xlog.Error(\"verifyTxRWSets NewContext error\", \"err\", err, \"contractName\", tmpReq.GetContractName())\n\t\t\tif i < len(reservedRequests) && (err.Error() == \"leveldb: not found\" || strings.HasSuffix(err.Error(), \"not found\")) {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\treturn false, err\n\t\t}\n\n\t\tctxResponse, ctxErr := ctx.Invoke(tmpReq.MethodName, tmpReq.Args)\n\t\tif ctxErr != nil {\n\t\t\tctx.Release()\n\t\t\tuv.xlog.Error(\"verifyTxRWSets Invoke error\", \"error\", ctxErr, \"contractName\", tmpReq.GetContractName())\n\t\t\treturn false, ctxErr\n\t\t}\n\t\t// 判断合约调用的返回码\n\t\tif ctxResponse.Status >= 400 && i < len(reservedRequests) {\n\t\t\tctx.Release()\n\t\t\tuv.xlog.Error(\"verifyTxRWSets Invoke error\", \"status\", ctxResponse.Status, \"contractName\", tmpReq.GetContractName())\n\t\t\treturn false, errors.New(ctxResponse.Message)\n\t\t}\n\n\t\tctx.Release()\n\t}\n\tutxoInputs, utxoOutputs := env.GetModelCache().GetUtxoRWSets()\n\terr = env.GetModelCache().PutUtxos(utxoInputs, utxoOutputs)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tcrossQuery := env.GetModelCache().GetCrossQueryRWSets()\n\terr = env.GetModelCache().PutCrossQueries(crossQuery)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\t_, writeSet, err := env.GetModelCache().GetRWSets()\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tuv.xlog.Trace(\"verifyTxRWSets\", \"env.output\", env.GetOutputs(), \"writeSet\", writeSet)\n\tok := xmodel.Equal(env.GetOutputs(), writeSet)\n\tif !ok {\n\t\treturn false, fmt.Errorf(\"write set not equal\")\n\t}\n\treturn true, nil\n}", "func (back *backend) GoodTransaction(tx *types.Transaction) bool {\n\ttr, err := back.Mine(context.Background(), tx)\n\tif err != nil {\n\t\treturn false\n\t}\n\n\t//fmt.Printf(\"gas: %d\\n\", tr.GasUsed)\n\n\tif tr.Status != 1 {\n\t\treturn false\n\t}\n\treturn true\n}", "func (p *PgDb) EmpExistsTx(id int, tx *sqlx.Tx) (bool, error) {\n\tcnt := \"postgres (p *PgDb) EmpExistsTx\" // Имя текущего метода для логирования\n\t//mylog.PrintfDebug(\"[DEBUG] %v - START, param: '%+v'\", cnt, id)\n\n\t// =====================================================================\n\t// Переменная часть кода\n\t// =====================================================================\n\tvar stm *sqlx.Stmt\n\tif tx == nil {\n\t\t// Работаем без транзакции\n\t\tstm = p.sqlSelectEmpExists\n\t} else {\n\t\t// Помещаем запрос в транзакцию\n\t\tstm = tx.Stmtx(p.sqlSelectEmpExists)\n\t}\n\t// возможен случай TWO_MANY_ROWS, поэтому определяем срез\n\tpk := model.GetEmpPK()\n\tdefer model.PutEmpPK(pk) // возвращаем в pool\n\t// =====================================================================\n\n\t//Выполняем запрос\n\terr := stm.Select(&pk, id)\n\tif err != nil {\n\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - stm.Select(&pk, id), args = '%v'\", cnt, id)\n\t\tlog.Printf(errM)\n\t\treturn false, errors.Wrap(err, errM)\n\t}\n\n\t//mylog.PrintfDebug(\"[DEBUG] %v - get %v rows\", cnt, len(pk))\n\n\t// Маловероятная ситуация - только если проблема с БД\n\tif len(pk) > 1 {\n\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - TWO_MANY_ROWS\", cnt)\n\t\tlog.Printf(errM)\n\t\treturn false, errors.Wrap(err, errM)\n\t}\n\n\t//mylog.PrintfDebug(\"[DEBUG] %v - SUCCESS\", cnt)\n\n\t// проверим количество полученных строк\n\tif len(pk) == 0 {\n\t\treturn false, nil\n\t}\n\treturn true, nil\n}", "func (otuo *OutboundTransactionUpdateOne) check() error {\n\tif v, ok := otuo.mutation.Channel(); ok {\n\t\tif err := outboundtransaction.ChannelValidator(v); err != nil {\n\t\t\treturn &ValidationError{Name: \"channel\", err: fmt.Errorf(\"ent: validator failed for field \\\"channel\\\": %w\", err)}\n\t\t}\n\t}\n\treturn nil\n}", "func (gWal *GenericWAL) IsLostOwnership() bool {\n // FIXME (bvk) This is not thread-safe.\n return gWal.lostOwnership\n}", "func (q queryManager) checkQueryNeedsTransaction(qp dbquery.QueryParsed) (bool, error) {\n\n\tif qp.IsSelect() {\n\t\treturn false, nil\n\t}\n\t// transaction for any update\n\treturn true, nil\n}", "func (t *Transaction) IsValid() bool {\n\treturn t.Id != uuid.Nil && t.SenderAddress.IsValid() && t.RecipientAddress.IsValid() && t.Value > 0\n}", "func (self *TokenService) Check(userToken *UserToken) bool {\n\tif common.IsEmpty(userToken.Id, userToken.UserId) == true {\n\t\treturn false\n\t}\n\tsql := `SELECT * FROM \"TOKEN\"\n\t\tWHERE \"UUID\" = ?\n\t\tAND \"USER_ID\" = ?\n\t \tAND \"DEAD_TIME\" > ?`\n\ttokenTable := new(table.TokenTable)\n\tgetFlag, getErr := self.Session.Sql(sql, userToken.Id, userToken.UserId, time.Now()).Get(tokenTable)\n\tif getErr != nil {\n\t\tself.Log.Println(getErr)\n\t}\n\t// update the dead time\n\tif getFlag == true {\n\t\tu := new(table.TokenTable)\n\t\tu.DeadTime = time.Now().Add(24 * time.Hour)\n\t\t_, updateErr := self.Session.Id(tokenTable.UUID).Update(u)\n\t\tif updateErr != nil {\n\t\t\tself.Log.Println(updateErr)\n\t\t}\n\t}\n\treturn getFlag\n}", "func (uv *UtxoVM) ImmediateVerifyTx(tx *pb.Transaction, isRootTx bool) (bool, error) {\n\t// Pre processing of tx data\n\tif !isRootTx && tx.Version == RootTxVersion {\n\t\treturn false, ErrVersionInvalid\n\t}\n\tif tx.Version > BetaTxVersion || tx.Version < RootTxVersion {\n\t\treturn false, ErrVersionInvalid\n\t}\n\t// autogen tx should not run ImmediateVerifyTx, this could be a fake tx\n\tif tx.Autogen {\n\t\treturn false, ErrInvalidAutogenTx\n\t}\n\tMaxTxSizePerBlock, MaxTxSizePerBlockErr := uv.MaxTxSizePerBlock()\n\tif MaxTxSizePerBlockErr != nil {\n\t\treturn false, MaxTxSizePerBlockErr\n\t}\n\tif proto.Size(tx) > MaxTxSizePerBlock {\n\t\tuv.xlog.Warn(\"tx too large, should not be greater than half of max blocksize\", \"size\", proto.Size(tx))\n\t\treturn false, ErrTxTooLarge\n\t}\n\n\t// Start transaction verification workflow\n\tif tx.Version > RootTxVersion {\n\t\t// verify txid\n\t\ttxid, err := txhash.MakeTransactionID(tx)\n\t\tif err != nil {\n\t\t\tuv.xlog.Warn(\"ImmediateVerifyTx: call MakeTransactionID failed\", \"error\", err)\n\t\t\treturn false, err\n\t\t}\n\t\tif bytes.Compare(tx.Txid, txid) != 0 {\n\t\t\tuv.xlog.Warn(\"ImmediateVerifyTx: txid not match\", \"tx.Txid\", tx.Txid, \"txid\", txid)\n\t\t\treturn false, fmt.Errorf(\"Txid verify failed\")\n\t\t}\n\n\t\t// get digestHash\n\t\tdigestHash, err := txhash.MakeTxDigestHash(tx)\n\t\tif err != nil {\n\t\t\tuv.xlog.Warn(\"ImmediateVerifyTx: call MakeTxDigestHash failed\", \"error\", err)\n\t\t\treturn false, err\n\t\t}\n\n\t\t// verify signatures\n\t\tok, verifiedID, err := uv.verifySignatures(tx, digestHash)\n\t\tif !ok {\n\t\t\tuv.xlog.Warn(\"ImmediateVerifyTx: verifySignatures failed\", \"error\", err)\n\t\t\treturn ok, ErrInvalidSignature\n\t\t}\n\n\t\t// get all authenticated users\n\t\tauthUsers := uv.removeDuplicateUser(tx.GetInitiator(), tx.GetAuthRequire())\n\n\t\t// veify tx UTXO input permission (Account ACL)\n\t\tok, err = uv.verifyUTXOPermission(tx, verifiedID)\n\t\tif !ok {\n\t\t\tuv.xlog.Warn(\"ImmediateVerifyTx: verifyUTXOPermission failed\", \"error\", err)\n\t\t\treturn ok, ErrACLNotEnough\n\t\t}\n\n\t\t// verify contract requests' permission using ACL\n\t\tok, err = uv.verifyContractPermission(tx, authUsers)\n\t\tif !ok {\n\t\t\tuv.xlog.Warn(\"ImmediateVerifyTx: verifyContractPermission failed\", \"error\", err)\n\t\t\treturn ok, ErrACLNotEnough\n\t\t}\n\n\t\t// verify amount of transfer within contract\n\t\tok, err = uv.verifyContractTxAmount(tx)\n\t\tif !ok {\n\t\t\tuv.xlog.Warn(\"ImmediateVerifyTx: verifyContractTxAmount failed\", \"error\", err)\n\t\t\treturn ok, ErrContractTxAmout\n\t\t}\n\n\t\t// verify the permission of RWSet using ACL\n\t\tok, err = uv.verifyRWSetPermission(tx, verifiedID)\n\t\tif !ok {\n\t\t\tuv.xlog.Warn(\"ImmediateVerifyTx: verifyRWSetPermission failed\", \"error\", err)\n\t\t\treturn ok, ErrACLNotEnough\n\t\t}\n\t\t// verify RWSet(run contracts and compare RWSet)\n\t\tok, err = uv.verifyTxRWSets(tx)\n\t\tif err != nil {\n\t\t\tuv.xlog.Warn(\"ImmediateVerifyTx: verifyTxRWSets failed\", \"error\", err)\n\t\t\t// reset error message\n\t\t\tif strings.HasPrefix(err.Error(), \"Gas not enough\") {\n\t\t\t\terr = ErrGasNotEnough\n\t\t\t} else {\n\t\t\t\terr = ErrRWSetInvalid\n\t\t\t}\n\t\t\treturn ok, err\n\t\t}\n\t\tif !ok {\n\t\t\t// always return RWSet Invalid Error if verification not passed\n\t\t\treturn ok, ErrRWSetInvalid\n\t\t}\n\t}\n\treturn true, nil\n}", "func (tx Transaction) IsCoinbase() bool {\n\treturn len(tx.Vin) == 1 && len(tx.Vin[0].Txid) == 0 && tx.Vin[0].Vout == -1\n}", "func (p *TestPager) Check() (bool, error) {\n return false, nil\n}", "func (a API) SendRawTransactionChk() (isNew bool) {\n\tselect {\n\tcase o := <-a.Ch.(chan SendRawTransactionRes):\n\t\tif o.Err != nil {\n\t\t\ta.Result = o.Err\n\t\t} else {\n\t\t\ta.Result = o.Res\n\t\t}\n\t\tisNew = true\n\tdefault:\n\t}\n\treturn\n}", "func CurrentlyOwns(\n\ttrx storage.Transaction,\n\towner *account.Account,\n\ttxId merkle.Digest,\n\tpool storage.Handle,\n) bool {\n\tdKey := append(owner.Bytes(), txId[:]...)\n\n\tif nil == trx {\n\t\treturn pool.Has(dKey)\n\t}\n\treturn trx.Has(pool, dKey)\n}", "func (a *ActStatus) Check(msg types.IMessage, strict bool) error {\n\tif msg.Time() > uint64(utils.NowUnix()) {\n\t\treturn errors.New(\"incorrect message time\")\n\t}\n\n\taccount := a.Account(msg.From())\n\treturn account.Check(msg, strict)\n}", "func hasCoins(ctx sdk.Context, am sdk.AccountMapper, addr sdk.Address, amt sdk.Coins) bool {\n\treturn getCoins(ctx, am, addr).IsGTE(amt)\n}", "func (otu *OutboundTransactionUpdate) check() error {\n\tif v, ok := otu.mutation.Channel(); ok {\n\t\tif err := outboundtransaction.ChannelValidator(v); err != nil {\n\t\t\treturn &ValidationError{Name: \"channel\", err: fmt.Errorf(\"ent: validator failed for field \\\"channel\\\": %w\", err)}\n\t\t}\n\t}\n\treturn nil\n}", "func (check *Check) Tx(tx string) *Check {\n\tcheck.tx = &tx\n\treturn check\n}", "func (tx *Tx) Check(record []byte) (err error) {\n\tif tx.extract==nil { err = EInternal ; return }\n\tdef := &checker{tx,nil}\n\ttx.extract.Extract(record,def)\n\terr = def.damage\n\treturn\n}", "func (t *trusteeImpl) VerifyMiningRewardTx(block consensus.Block) bool {\n\t// pop first transaction from the block\n\ttx := block.Transactions()[0]\n\n\t// validate transaction signature\n\tif !t.verify(tx.Payload, tx.Signature, tx.Submitter) {\n\t\treturn false\n\t}\n\n\t// validate that mining award is for block's miner\n\tif string(tx.Submitter) != string(block.Miner()) {\n\t\tt.log.Error(\"mining award owner is not the block miner\")\n\t\treturn false\n\t}\n\n\t// process transaction and update block\n\treturn t.process(block, &tx)\n}", "func CheckTransactionContext(txn *tx.Transaction, ledger *Ledger) ErrCode {\n\t// check if duplicated with transaction in ledger\n\tif exist := ledger.Store.IsTxHashDuplicate(txn.Hash()); exist {\n\t\tlog.Info(\"[CheckTransactionContext] duplicate transaction check faild.\")\n\t\treturn ErrTxHashDuplicate\n\t}\n\n\tif txn.IsCoinBaseTx() {\n\t\treturn Success\n\t}\n\n\t// check double spent transaction\n\tif IsDoubleSpend(txn, ledger) {\n\t\tlog.Info(\"[CheckTransactionContext] IsDoubleSpend check faild.\")\n\t\treturn ErrDoubleSpend\n\t}\n\n\tif err := CheckTransactionUTXOLock(txn); err != nil {\n\t\tlog.Warn(\"[CheckTransactionUTXOLock],\", err)\n\t\treturn ErrUTXOLocked\n\t}\n\n\tif err := CheckTransactionBalance(txn); err != nil {\n\t\tlog.Warn(\"[CheckTransactionBalance],\", err)\n\t\treturn ErrTransactionBalance\n\t}\n\n\tif err := CheckTransactionSignature(txn); err != nil {\n\t\tlog.Warn(\"[CheckTransactionSignature],\", err)\n\t\treturn ErrTransactionSignature\n\t}\n\t// check referenced Output value\n\tfor _, input := range txn.UTXOInputs {\n\t\treferHash := input.ReferTxID\n\t\treferTxnOutIndex := input.ReferTxOutputIndex\n\t\treferTxn, _, err := ledger.Store.GetTransaction(referHash)\n\t\tif err != nil {\n\t\t\tlog.Warn(\"Referenced transaction can not be found\", common.BytesToHexString(referHash.ToArray()))\n\t\t\treturn ErrUnknownReferedTxn\n\t\t}\n\t\treferTxnOut := referTxn.Outputs[referTxnOutIndex]\n\t\tif referTxnOut.Value <= 0 {\n\t\t\tlog.Warn(\"Value of referenced transaction output is invalid\")\n\t\t\treturn ErrInvalidReferedTxn\n\t\t}\n\t\t// coinbase transaction only can be spent after got SpendCoinbaseSpan times confirmations\n\t\tif referTxn.IsCoinBaseTx() {\n\t\t\tlockHeight := referTxn.LockTime\n\t\t\tcurrentHeight := ledger.Store.GetHeight()\n\t\t\tif currentHeight-lockHeight < config.Parameters.ChainParam.SpendCoinbaseSpan {\n\t\t\t\treturn ErrIneffectiveCoinbase\n\t\t\t}\n\t\t}\n\t}\n\n\treturn Success\n}", "func CurrentlyOwns(owner *account.Account, txId merkle.Digest) bool {\n\tdKey := append(owner.Bytes(), txId[:]...)\n\treturn nil != storage.Pool.OwnerDigest.Get(dKey)\n}", "func (s *State) SpecialTxExists(tx *types.Transaction) bool {\n\tillegalData, ok := tx.Payload.(payload.DPOSIllegalData)\n\tif !ok {\n\t\tlog.Error(\"special tx payload cast failed, tx:\", tx.Hash())\n\t\treturn false\n\t}\n\n\thash := illegalData.Hash()\n\ts.mtx.RLock()\n\t_, ok = s.SpecialTxHashes[hash]\n\ts.mtx.RUnlock()\n\treturn ok\n}", "func (keeper BaseViewKeeper) HasCoins(ctx sdk.Context, addr sdk.AccAddress, amt sdk.Coins) bool {\n\treturn hasCoins(ctx, keeper.am, addr, amt)\n}", "func (a API) CreateRawTransactionChk() (isNew bool) {\n\tselect {\n\tcase o := <-a.Ch.(chan CreateRawTransactionRes):\n\t\tif o.Err != nil {\n\t\t\ta.Result = o.Err\n\t\t} else {\n\t\t\ta.Result = o.Res\n\t\t}\n\t\tisNew = true\n\tdefault:\n\t}\n\treturn\n}", "func (t *Token) IsRedeem() bool {\n\treturn len(t.Owner) == 0\n}", "func (_PBridge *PBridgeCallerSession) IsCompletedTx(txKey string) (bool, error) {\n\treturn _PBridge.Contract.IsCompletedTx(&_PBridge.CallOpts, txKey)\n}", "func (n *NodeBlockMaker) checkGoodTimeToMakeBlock() bool {\n\treturn true\n}", "func (a *account) callNeedsToSync() bool {\n\ta.mu.Lock()\n\tdefer a.mu.Unlock()\n\treturn a.syncAt.Before(time.Now())\n}", "func CheckGenTx(\n\tt *testing.T, app *baseapp.BaseApp, msgs []sdk.Msg, accNums []uint64,\n\tseq []uint64, expPass bool, priv ...crypto.PrivKey,\n) sdk.Result {\n\ttx := GenTx(msgs, accNums, seq, priv...)\n\tres := app.Check(tx)\n\n\tif expPass {\n\t\trequire.Equal(t, sdk.CodeOK, res.Code, res.Log)\n\t} else {\n\t\trequire.NotEqual(t, sdk.CodeOK, res.Code, res.Log)\n\t}\n\n\treturn res\n}", "func (s *SimpleBlockFactory) IsTransactionValid(tx *types.Tx) bool {\n\t// SimpleBlockFactory has no tx valid check.\n\treturn true\n}", "func (tx *Transaction) IsCoinbase() bool {\n\treturn len(tx.Vin) == 1 && len(tx.Vin[0].Txid) == 0 && tx.Vin[0].Vout == -1\n}", "func (_Token *TokenCaller) IsOwner(opts *bind.CallOpts) (bool, error) {\n\tvar (\n\t\tret0 = new(bool)\n\t)\n\tout := ret0\n\terr := _Token.contract.Call(opts, out, \"isOwner\")\n\treturn *ret0, err\n}", "func validate(tx *Tx) bool {\n\tfor _, txIn := range tx.TxIns {\n\t\t// Find the transaction that created the transaction input\n\t\tprevTx := FindTx(Blockchain(), txIn.TxId)\n\t\tif prevTx == nil {\n\t\t\t// Fake input: not on the blockchain\n\t\t\treturn false\n\t\t}\n\t\tprevTxOut := prevTx.TxOuts[txIn.Index]\n\t\taddress := prevTxOut.Address\n\t\t// If the public key (address) cannot verify the signature that I just\n\t\t// created w/ my wallet, that means the TxOuts/funds are not actually mine\n\t\tif !wallet.Verify(tx.Id, txIn.Signature, address) {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true // All transaction inputs verfied\n}", "func (t *ACLRole) DBExistsTx(ctx context.Context, tx Tx) (bool, error) {\n\tq := \"SELECT `id` FROM `acl_role` WHERE `id` = ? LIMIT 1\"\n\tvar _ID sql.NullString\n\terr := tx.QueryRowContext(ctx, q, orm.ToSQLString(t.ID)).Scan(&_ID)\n\tif err != nil && err != sql.ErrNoRows {\n\t\treturn false, err\n\t}\n\treturn _ID.Valid, nil\n}", "func (n *NodeManager) CheckByUUIDEnvID(uuid string, envID int) bool {\n\tvar results int64\n\tn.DB.Model(&OsqueryNode{}).Where(\"uuid = ? AND environment_id = ?\", strings.ToUpper(uuid), envID).Count(&results)\n\treturn (results > 0)\n}" ]
[ "0.6219588", "0.6043592", "0.5973341", "0.57881826", "0.5786513", "0.5681203", "0.56810457", "0.5648593", "0.563845", "0.5616632", "0.5613117", "0.55977416", "0.5552989", "0.55476046", "0.55441684", "0.5538579", "0.55384165", "0.5532735", "0.552402", "0.5520972", "0.5520972", "0.54907477", "0.54651034", "0.5450362", "0.54374385", "0.5418135", "0.5409281", "0.5408374", "0.5394832", "0.5394424", "0.5384628", "0.53837806", "0.5375425", "0.53750193", "0.53332067", "0.533312", "0.5308207", "0.530671", "0.5268558", "0.5264312", "0.5254385", "0.5254319", "0.52533895", "0.52251107", "0.52246726", "0.5215446", "0.5205147", "0.5187886", "0.5177429", "0.517716", "0.5168154", "0.51656884", "0.51544046", "0.5152159", "0.51508874", "0.5135981", "0.5135981", "0.5133472", "0.51205033", "0.51203126", "0.51169765", "0.51119524", "0.5091511", "0.508997", "0.5087836", "0.50849754", "0.5084127", "0.5083664", "0.5078645", "0.5073383", "0.50713396", "0.5060232", "0.5056126", "0.5054952", "0.5047792", "0.5044771", "0.5037405", "0.5027353", "0.50270903", "0.50270844", "0.50264984", "0.502349", "0.5012572", "0.5011563", "0.501067", "0.50086224", "0.5006434", "0.50064045", "0.49903965", "0.49867904", "0.49803817", "0.49798828", "0.49784136", "0.49751094", "0.49735534", "0.49733463", "0.49723896", "0.49691346", "0.4964139", "0.49580264" ]
0.6790171
0
ShowWalletIndex is uesd to show wallet index.
func (account *Account) ShowWalletIndex() { fmt.Println("\033[1;31;40m", account.WalletIndex, "\033[0m") }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func ShowIndex() {\n\tfmt.Printf(\"%v\\n\", indexText)\n}", "func (p *planner) ShowIndex(ctx context.Context, n *tree.ShowIndex) (planNode, error) {\n\tconst getIndexes = `\n\t\t\t\tSELECT\n\t\t\t\t\ttable_name AS \"Table\",\n\t\t\t\t\tindex_name AS \"Name\",\n\t\t\t\t\tNOT non_unique::BOOL AS \"Unique\",\n\t\t\t\t\tseq_in_index AS \"Seq\",\n\t\t\t\t\tcolumn_name AS \"Column\",\n\t\t\t\t\tdirection AS \"Direction\",\n\t\t\t\t\tstoring::BOOL AS \"Storing\",\n\t\t\t\t\timplicit::BOOL AS \"Implicit\"\n\t\t\t\tFROM %[4]s.information_schema.statistics\n\t\t\t\tWHERE table_catalog=%[1]s AND table_schema=%[5]s AND table_name=%[2]s`\n\treturn p.showTableDetails(ctx, \"SHOW INDEX\", n.Table, getIndexes)\n}", "func ShowIndex(ctx context.Context, db QueryExecutor, schemaName string, table string) ([]*IndexInfo, error) {\n\t/*\n\t\tshow index example result:\n\t\tmysql> show index from test;\n\t\t+-------+------------+----------+--------------+-------------+-----------+-------------+----------+--------+------+------------+---------+---------------+\n\t\t| Table | Non_unique | Key_name | Seq_in_index | Column_name | Collation | Cardinality | Sub_part | Packed | Null | Index_type | Comment | Index_comment |\n\t\t+-------+------------+----------+--------------+-------------+-----------+-------------+----------+--------+------+------------+---------+---------------+\n\t\t| test | 0 | PRIMARY | 1 | id | A | 0 | NULL | NULL | | BTREE | | |\n\t\t| test | 0 | aid | 1 | aid | A | 0 | NULL | NULL | YES | BTREE | | |\n\t\t+-------+------------+----------+--------------+-------------+-----------+-------------+----------+--------+------+------------+---------+---------------+\n\t*/\n\tindices := make([]*IndexInfo, 0, 3)\n\tquery := fmt.Sprintf(\"SHOW INDEX FROM %s\", TableName(schemaName, table))\n\trows, err := db.QueryContext(ctx, query)\n\tif err != nil {\n\t\treturn nil, errors.Trace(err)\n\t}\n\tdefer rows.Close()\n\n\tfor rows.Next() {\n\t\tfields, err1 := ScanRow(rows)\n\t\tif err1 != nil {\n\t\t\treturn nil, errors.Trace(err1)\n\t\t}\n\t\tseqInIndex, err1 := strconv.Atoi(string(fields[\"Seq_in_index\"].Data))\n\t\tif err1 != nil {\n\t\t\treturn nil, errors.Trace(err1)\n\t\t}\n\t\tcardinality, err1 := strconv.Atoi(string(fields[\"Cardinality\"].Data))\n\t\tif err1 != nil {\n\t\t\treturn nil, errors.Trace(err1)\n\t\t}\n\t\tindex := &IndexInfo{\n\t\t\tTable: string(fields[\"Table\"].Data),\n\t\t\tNoneUnique: string(fields[\"Non_unique\"].Data) == \"1\",\n\t\t\tKeyName: string(fields[\"Key_name\"].Data),\n\t\t\tColumnName: string(fields[\"Column_name\"].Data),\n\t\t\tSeqInIndex: seqInIndex,\n\t\t\tCardinality: cardinality,\n\t\t}\n\t\tindices = append(indices, index)\n\t}\n\n\treturn indices, nil\n}", "func (account *Account) ShowTx(P string) bool {\n\tfor i := 0; i < account.WalletIndex; i++ {\n\t\tif P == account.Wallet[i].P {\n\t\t\tJSON, _ := json.MarshalIndent(account.Wallet[i], \"\", \" \")\n\n\t\t\tfmt.Println(string(JSON))\n\t\t\treturn true\n\n\t\t}\n\t}\n\treturn false\n}", "func tableInfoIndexShow(meta *halproto.TableMetadata) {\n\tiMeta := meta.GetIndexMeta()\n\tinsStr := fmt.Sprintf(\"%d[%d]\", iMeta.GetNumInserts(), iMeta.GetNumInsertFailures())\n\tupdStr := fmt.Sprintf(\"%d[%d]\", iMeta.GetNumUpdates(), iMeta.GetNumUpdateFailures())\n\tdelStr := fmt.Sprintf(\"%d[%d]\", iMeta.GetNumDeletes(), iMeta.GetNumDeleteFailures())\n\tfmt.Printf(\"%-30s%-10d%-10s%-10d%-10s%-10d%-10s%-10s%-10s%-10s\\n\",\n\t\tmeta.GetTableName(),\n\t\tmeta.GetTableId(),\n\t\tutils.TableKindToStr(meta.GetKind()),\n\t\tiMeta.GetCapacity(),\n\t\t\"-\",\n\t\tiMeta.GetUsage(),\n\t\t\"-\",\n\t\tinsStr,\n\t\tupdStr,\n\t\tdelStr)\n}", "func ShowWalletsPath() string {\n\n\treturn fmt.Sprintf(\"/v1/wallets\")\n}", "func showIndex(res http.ResponseWriter, req *http.Request, params httprouter.Params) {\n\tServeTemplateWithParams(res, req, \"index.html\", nil)\n}", "func (tc *TransactionsController) Index(c *gin.Context, size, page, offset int) {\n\ttxs, count, err := tc.App.GetStore().Transactions(offset, size)\n\tptxs := make([]presenters.Tx, len(txs))\n\tfor i, tx := range txs {\n\t\tptxs[i] = presenters.NewTx(&tx)\n\t}\n\tpaginatedResponse(c, \"Transactions\", size, page, ptxs, count, err)\n}", "func Index(txn *cheshire.Txn) {\n\t//create a context map to be passed to the template\n\tcontext := make(map[string]interface{})\n\tcontext[\"services\"] = Servs.RouterTables()\n\tcheshire.RenderInLayout(txn, \"/index.html\", \"/template.html\", context)\n}", "func (ekc *ETHKeysController) Index(c *gin.Context) {\n\tstore := ekc.App.GetStore()\n\tkeys, err := store.AllKeys()\n\tif err != nil {\n\t\terr = errors.Errorf(\"error fetching ETH keys from database: %v\", err)\n\t\tjsonAPIError(c, http.StatusInternalServerError, err)\n\t\treturn\n\t}\n\tvar pkeys []presenters.ETHKey\n\tfor _, key := range keys {\n\t\tethBalance, err := store.EthClient.BalanceAt(c.Request.Context(), key.Address.Address(), nil)\n\t\tif err != nil {\n\t\t\terr = errors.Errorf(\"error calling getEthBalance on Ethereum node: %v\", err)\n\t\t\tjsonAPIError(c, http.StatusInternalServerError, err)\n\t\t\treturn\n\t\t}\n\n\t\tlinkAddress := common.HexToAddress(store.Config.LinkContractAddress())\n\t\tlinkBalance, err := store.EthClient.GetLINKBalance(linkAddress, key.Address.Address())\n\t\tif err != nil {\n\t\t\terr = errors.Errorf(\"error calling getLINKBalance on Ethereum node: %v\", err)\n\t\t\tjsonAPIError(c, http.StatusInternalServerError, err)\n\t\t\treturn\n\t\t}\n\n\t\tk, err := store.ORM.KeyByAddress(key.Address.Address())\n\t\tif err != nil {\n\t\t\terr = errors.Errorf(\"error fetching ETH key from DB: %v\", err)\n\t\t\tjsonAPIError(c, http.StatusInternalServerError, err)\n\t\t\treturn\n\t\t}\n\t\tpkeys = append(pkeys, presenters.ETHKey{\n\t\t\tAddress: k.Address.Hex(),\n\t\t\tEthBalance: (*assets.Eth)(ethBalance),\n\t\t\tLinkBalance: linkBalance,\n\t\t\tNextNonce: k.NextNonce,\n\t\t\tLastUsed: k.LastUsed,\n\t\t\tIsFunding: k.IsFunding,\n\t\t\tCreatedAt: k.CreatedAt,\n\t\t\tUpdatedAt: k.UpdatedAt,\n\t\t\tDeletedAt: k.DeletedAt,\n\t\t})\n\t}\n\tjsonAPIResponse(c, pkeys, \"keys\")\n}", "func (t *Table) SetShowIndex(index bool) {\n\tt.showIndex = index\n}", "func showIndex(c *gin.Context) {\n\trender(\n\t\tc,\n\t\tgin.H{\n\t\t\t\"title\": \"Home Page\",\n\t\t\t\"payload\": films,\n\t\t},\n\t\ttemplates.Index,\n\t)\n}", "func index(c echo.Context) error {\n\treturn c.String(http.StatusOK, \"Account service\")\n}", "func index(w http.ResponseWriter, r *http.Request) {\n\tfmt.Fprint(w, \"index de uma função\")\n}", "func (s *Store) RetrieveAccountsIndex(walletID uuid.UUID) ([]byte, error) {\n\tpath := s.walletIndexPath(walletID)\n\tdata, err := os.ReadFile(path)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to read wallet index\")\n\t}\n\t// Do not decrypt empty index.\n\tif len(data) == 2 {\n\t\treturn data, nil\n\t}\n\n\treturn s.decryptIfRequired(data)\n}", "func (tc *TransactionsController) Show(c *gin.Context) {\n\thash := common.HexToHash(c.Param(\"TxHash\"))\n\tif tx, err := tc.App.GetStore().FindTxByAttempt(hash); err == orm.ErrorNotFound {\n\t\tpublicError(c, http.StatusNotFound, errors.New(\"Transaction not found\"))\n\t} else if err != nil {\n\t\tc.AbortWithError(http.StatusInternalServerError, err)\n\t} else if doc, err := jsonapi.Marshal(presenters.NewTx(tx)); err != nil {\n\t\tc.AbortWithError(http.StatusInternalServerError, err)\n\t} else {\n\t\tc.Data(http.StatusOK, MediaType, doc)\n\t}\n}", "func (t *ATrains) Show() {\n\tfmt.Println(\"wszystkie pociagi\")\n\n\tfor i := 0; i < len(t.trains); i++ {\n\t\tt.trains[i].Show()\n\t}\n\n\tfmt.Println(\"\")\n}", "func (w *Wallet) Info() modules.WalletInfo {\n\twi := modules.WalletInfo{\n\t\tBalance: w.Balance(false),\n\t\tFullBalance: w.Balance(true),\n\t}\n\n\tcounter := w.mu.RLock()\n\twi.NumAddresses = len(w.keys)\n\tw.mu.RUnlock(counter)\n\n\tvar sortingSpace crypto.HashSlice\n\tfor va := range w.visibleAddresses {\n\t\tsortingSpace = append(sortingSpace, crypto.Hash(va))\n\t}\n\tsort.Sort(sortingSpace)\n\tfor _, va := range sortingSpace {\n\t\twi.VisibleAddresses = append(wi.VisibleAddresses, types.UnlockHash(va))\n\t}\n\treturn wi\n}", "func Index(writer http.ResponseWriter, _ *http.Request) {\n\trespond(writer, \"{Welcome to the Bitcoin Arbitrage Detector}\", nil)\n}", "func (a *App) Index(w http.ResponseWriter, r *http.Request, _ httprouter.Params) {\n\trawBankStatements, err := models.AllRawBankStatements(a.db)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\thttp.Error(w, http.StatusText(500), 500)\n\t\treturn\n\t}\n\n\tdata, err := json.Marshal(rawBankStatements)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\thttp.Error(w, http.StatusText(500), 500)\n\t\treturn\n\t}\n\tw.Write(data)\n\tfmt.Println(\"Index\")\n}", "func Index(w http.ResponseWriter, data *IndexData) {\n\trender(tpIndex, w, data)\n}", "func Index(w http.ResponseWriter, data *IndexData) {\n\trender(tpIndex, w, data)\n}", "func (s *Store) StoreAccountsIndex(walletID uuid.UUID, data []byte) error {\n\t// Ensure wallet path exists.\n\tvar err error\n\tif err = s.ensureWalletPathExists(walletID); err != nil {\n\t\treturn errors.Wrap(err, \"wallet path does not exist\")\n\t}\n\n\t// Do not encrypt empty index.\n\tif len(data) != 2 {\n\t\tdata, err = s.encryptIfRequired(data)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"failed to encrypt index\")\n\t\t}\n\t}\n\n\tpath := s.walletIndexPath(walletID)\n\n\treturn os.WriteFile(path, data, 0o600)\n}", "func Index(w http.ResponseWriter, data interface{}) {\n\trender(tpIndex, w, data)\n}", "func (c *Client) ShowMapping() *Client {\n\tif checkIndexName(c) {\n\t\treturn c\n\t}\n\tc.hostDB.Path = path.Join(c.hostDB.Path, \"_mapping\")\n\treturn c.exec(c.hostDB.String())\n}", "func (ctrl SettingsController) Index(c *gin.Context) {\n\tmenu := gui.GetMenu(c)\n\ttemplateParams := gin.H{\"menu\": menu}\n\ttemplateParams[\"title\"] = \"Settings\"\n\ttemplateParams[\"appconfig\"], _ = conf.GetAppConfigData()\n\tc.HTML(http.StatusOK, \"settings/index\", templateParams)\n}", "func displayTransactions(scid string, option int, ID string) {\n\t\t\n\tscKeys:= []string{\"numberOfOwners\", \"txCount\"}\n\tresult:= getKeysFromDaemon(scid, scKeys)\n\tif result == \"\" {return}\n\n\n\t//Response ok, extract keys from JSON\n\t\n\n\ttxCount := gjson.Get(result, \"txs.#.sc_keys.txCount\")\n\ttxCountArray:= txCount.Array()[0]\n\ttxCountInt:= txCountArray.Int()\n\t//fmt.Printf(\"Tx Count: %d\\n\", txCountInt)\n\n\t//Make a slice of keys so we can request in RPC call\n\tx:= int(txCountInt) //txCount in wallet smart contract is always 1 ahead of actual number of transactions\t\n\tx4:= x * 4\t\n\tkeySlice:= make([]string, x4) \n\t\n\tfor i:=0; i<x; i++ {\n\t\tz:= strconv.Itoa(i)\n\t\tkeySlice[i] = \"txIndex_\" + z\n\t\tkeySlice[i+x] = \"recipient_\" + z\n\t\tkeySlice[i+(x*2)] = \"amount_\" + z\n\t\tkeySlice[i+(x*3)] = \"sent_\" + z\n\t}\n\t\t\n\t//fmt.Println(keySlice)\n\tdisplayData(scid, keySlice, x, option, ID)\n\n\n}", "func (a Account) ShowBalance() int {\n\treturn a.balance\n}", "func WalletInfo(chains []kmdgo.AppType) []WInfo {\n\tvar wallets []WInfo\n\n\t// fmt.Println(chains)\n\n\tstats, err := kmdgo.NewAppType(\"DEX\").DEXStats()\n\tif err != nil {\n\t\tfmt.Printf(\"Code: %v\\n\", stats.Error.Code)\n\t\tfmt.Printf(\"Message: %v\\n\\n\", stats.Error.Message)\n\t\t// log.Fatalln(\"Err happened\", err)\n\t}\n\n\t// fmt.Println(\"stats value\", stats)\n\t// fmt.Println(\"Recvaddr\", stats.Result.Recvaddr)\n\t// fmt.Println(\"RecvZaddr\", stats.Result.RecvZaddr)\n\n\tfor _, v := range chains {\n\t\t// fmt.Println(i)\n\t\t// fmt.Println(v)\n\t\tif v == \"KMD\" {\n\t\t\tv = \"komodo\"\n\t\t}\n\n\t\tappName := kmdgo.NewAppType(v)\n\n\t\tvar info kmdgo.GetInfo\n\n\t\tinfo, err := appName.GetInfo()\n\t\t// fmt.Println(info.Error.Message)\n\t\tif err != nil {\n\t\t\t// fmt.Printf(\"Code: %v\\n\", info.Error.Code)\n\t\t\t// fmt.Printf(\"Message: %v\\n\\n\", info.Error.Message)\n\t\t\tif info.Error.Message == \"Loading block index...\" {\n\t\t\t\tfmt.Println(v, \"- Err happened:\", info.Error.Message)\n\t\t\t\twallets = append(wallets, WInfo{string(v), \"Loading...\", 0.0, 0, false})\n\t\t\t} else if info.Error.Message == \"Rescanning...\" {\n\t\t\t\tfmt.Println(v, \"- Err happened:\", info.Error.Message)\n\t\t\t\twallets = append(wallets, WInfo{string(v), \"Rescanning...\", 0.0, 0, false})\n\t\t\t} else {\n\t\t\t\tfmt.Println(v, \"- Err happened:\", err)\n\t\t\t\twallets = append(wallets, WInfo{string(v), \"Offline\", 0.0, 0, false})\n\t\t\t}\n\t\t} else {\n\t\t\tif info.Error.Message == \"connection refused\" {\n\t\t\t\tfmt.Println(v, \"- Err happened:\", info.Error.Message)\n\t\t\t\twallets = append(wallets, WInfo{string(v), \"Offline\", 0.0, 0, false})\n\t\t\t} else {\n\t\t\t\t// Check status of the blockchain sync\n\t\t\t\tvar tempSyncStatus bool\n\t\t\t\tif info.Result.Longestchain != info.Result.Blocks {\n\t\t\t\t\ttempSyncStatus = false\n\t\t\t\t} else {\n\t\t\t\t\ttempSyncStatus = true\n\t\t\t\t}\n\n\t\t\t\tif v == \"PIRATE\" {\n\t\t\t\t\t// fmt.Println(\"it is PIRATE\")\n\t\t\t\t\tvar zblc kmdgo.ZGetBalance\n\n\t\t\t\t\targs := make(kmdgo.APIParams, 2)\n\t\t\t\t\targs[0] = stats.Result.RecvZaddr\n\t\t\t\t\t//args[1] = 1\n\t\t\t\t\t// fmt.Println(args)\n\n\t\t\t\t\tzblc, err := appName.ZGetBalance(args)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tfmt.Printf(\"Code: %v\\n\", zblc.Error.Code)\n\t\t\t\t\t\tfmt.Printf(\"Message: %v\\n\\n\", zblc.Error.Message)\n\t\t\t\t\t\t// log.Fatalln(\"Err happened\", err)\n\t\t\t\t\t}\n\n\t\t\t\t\t// fmt.Println(\"zblc value\", zblc)\n\t\t\t\t\t// fmt.Println(\"-------\")\n\t\t\t\t\t// fmt.Printf(\"\\n%0.8f\\n\", zblc.Result)\n\n\t\t\t\t\twallets = append(wallets, WInfo{\n\t\t\t\t\t\tTicker: info.Result.Name,\n\t\t\t\t\t\tStatus: \"Online\",\n\t\t\t\t\t\tBalance: zblc.Result,\n\t\t\t\t\t\tBlocks: info.Result.Longestchain,\n\t\t\t\t\t\tSynced: tempSyncStatus,\n\t\t\t\t\t})\n\n\t\t\t\t} else {\n\t\t\t\t\twallets = append(wallets, WInfo{\n\t\t\t\t\t\tTicker: info.Result.Name,\n\t\t\t\t\t\tStatus: \"Online\",\n\t\t\t\t\t\tBalance: info.Result.Balance,\n\t\t\t\t\t\tBlocks: info.Result.Longestchain,\n\t\t\t\t\t\tSynced: tempSyncStatus,\n\t\t\t\t\t})\n\t\t\t\t}\n\n\t\t\t}\n\t\t}\n\t}\n\n\t// fmt.Println(wallets)\n\treturn wallets\n}", "func(peers *PeerList) Show() string {\n\tpeers.mux.Lock()\n\tdefer peers.mux.Unlock()\n\tvar result string\n\tfor key, entry := range peers.peerMap {\n\t\tresult += \"Address = \" + string(key) + \" Id = \" + string(entry) + \"\\n\"\n\t}\n\treturn result\n}", "func blockchainViewHandler(w http.ResponseWriter, r *http.Request) {\n\n\t// Take the URL beyond /blockchain/ and split into request and value strings\n\trequestAction := strings.Split(r.URL.String(), \"/\")\n\trequestItem, err := strconv.Atoi(requestAction[3])\n\trequestItem = requestItem - 1\n\n\tif err != nil {\n\t\tlog.Println(\"INFO: Unable to convert argument to integer, assume this is a request for entire chain\")\n\t}\n\n\tif requestItem == -1 { //Request item is invalid so display that blockID only\n\t\tblockString, err := json.MarshalIndent(Blockchain, \"\", \"\\t\")\n\t\tif err != nil {\n\t\t\tlog.Println(\"ERROR: blockchainViewHandler(): Cannot print Blockchain\")\n\t\t}\n\t\tfmt.Fprintf(w, \"\\n %s\", blockString)\n\t} else {\n\t\tblockItemString, _ := json.MarshalIndent(Blockchain[requestItem], \"\", \"\\t\") // Do nothing if index too high\n\t\tfmt.Fprintf(w, \"\\n %s.\", blockItemString)\n\t}\n}", "func index(w http.ResponseWriter, req *http.Request, _ httprouter.Params) {\n\tvar returnResponse = map[string]interface{}{\"message\": \"Welcome to the TonicPow API!\"}\n\tapirouter.ReturnResponse(w, req, http.StatusOK, returnResponse)\n}", "func (p *Proxy) handleShowTxnz(session *driver.Session, query string, node sqlparser.Statement) (*sqltypes.Result, error) {\n\treturn nil, sqldb.NewSQLErrorf(sqldb.ER_SPECIFIC_ACCESS_DENIED_ERROR, \"Access denied; lacking super privilege for the operation\")\n\t// privilegePlug := spanner.plugins.PlugPrivilege()\n\t// if !privilegePlug.IsSuperPriv(session.User()) {\n\t// \treturn nil, sqldb.NewSQLErrorf(sqldb.ER_SPECIFIC_ACCESS_DENIED_ERROR, \"Access denied; lacking super privilege for the operation\")\n\t// }\n\n\t// qr := &sqltypes.Result{}\n\t// qr.Fields = []*querypb.Field{\n\t// \t{Name: \"TxnID\", Type: querypb.Type_INT64},\n\t// \t{Name: \"Start\", Type: querypb.Type_VARCHAR},\n\t// \t{Name: \"Duration\", Type: querypb.Type_INT32},\n\t// \t{Name: \"XaState\", Type: querypb.Type_VARCHAR},\n\t// \t{Name: \"TxnState\", Type: querypb.Type_VARCHAR},\n\t// }\n\n\t// rows := spanner.scatter.Txnz().GetTxnzRows()\n\t// for _, row := range rows {\n\t// \trow := []sqltypes.Value{\n\t// \t\tsqltypes.MakeTrusted(querypb.Type_INT64, []byte(fmt.Sprintf(\"%v\", uint64(row.TxnID)))),\n\t// \t\tsqltypes.MakeTrusted(querypb.Type_VARCHAR, []byte(row.Start.Format(\"20060102150405.000\"))),\n\t// \t\tsqltypes.MakeTrusted(querypb.Type_INT32, []byte(fmt.Sprintf(\"%v\", row.Duration))),\n\t// \t\tsqltypes.MakeTrusted(querypb.Type_VARCHAR, []byte(row.XaState)),\n\t// \t\tsqltypes.MakeTrusted(querypb.Type_VARCHAR, []byte(row.State)),\n\t// \t}\n\t// \tqr.Rows = append(qr.Rows, row)\n\t// }\n\t// return qr, nil\n}", "func (api *Api) TxByIndex(blockNumber *int, txIndex int) (*models.Tx, error) {\n\tgetter := &proto.TxGetter{\n\t\tGetter: &proto.TxGetter_Index{\n\t\t\tIndex: &proto.TxGetterIndex{\n\t\t\t\tIndex: int64(txIndex),\n\t\t\t\tBlock: parseBlockGetter(blockNumber),\n\t\t\t}},\n\t}\n\n\treturn api.fetchTx(getter)\n}", "func (account *Account) ShowAccount() {\n\tfmt.Println(\"a:\", account.a)\n\tfmt.Println(\"A:\", account.A)\n\tfmt.Println(\"AT:\", account.AT)\n}", "func (node *ShowIndex) Format(buf *bytes.Buffer, f FmtFlags) {\n\tbuf.WriteString(\"SHOW INDEXES FROM \")\n\tFormatNode(buf, f, node.Table)\n}", "func (b *Builder) ChainShow(ctx context.Context, chainID string) (spn.Chain, error) {\n\taccount, err := b.AccountInUse()\n\tif err != nil {\n\t\treturn spn.Chain{}, err\n\t}\n\treturn b.spnclient.ChainGet(ctx, account.Name, chainID)\n}", "func (showTxCommand ShowTransactionCommand) Run(ctx context.Context, wallet walletcore.Wallet) error {\n\ttransaction, err := wallet.GetTransaction(showTxCommand.Args.TxHash)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tbasicOutput := \"Hash\\t%s\\n\" +\n\t\t\"Confirmations\\t%d\\n\" +\n\t\t\"Included in block\\t%d\\n\" +\n\t\t\"Type\\t%s\\n\" +\n\t\t\"Amount %s\\t%s\\n\" +\n\t\t\"Date\\t%s\\n\" +\n\t\t\"Size\\t%s\\n\" +\n\t\t\"Fee\\t%s\\n\" +\n\t\t\"Rate\\t%s/kB\\n\"\n\n\ttxDirection := strings.ToLower(transaction.Direction.String())\n\ttxSize := fmt.Sprintf(\"%.1f kB\", float64(transaction.Size)/1000)\n\tbasicOutput = fmt.Sprintf(basicOutput,\n\t\ttransaction.Hash,\n\t\ttransaction.Confirmations,\n\t\ttransaction.BlockHeight,\n\t\ttransaction.Type,\n\t\ttxDirection, transaction.Amount,\n\t\ttransaction.FormattedTime,\n\t\ttxSize,\n\t\ttransaction.Fee,\n\t\ttransaction.FeeRate)\n\n\tif showTxCommand.Detailed {\n\t\tdetailedOutput := strings.Builder{}\n\t\tdetailedOutput.WriteString(\"General Info\\n\")\n\t\tdetailedOutput.WriteString(basicOutput)\n\t\tdetailedOutput.WriteString(\"\\nInputs\\n\")\n\t\tfor _, input := range transaction.Inputs {\n\t\t\tdetailedOutput.WriteString(fmt.Sprintf(\"%s\\t%s\\n\", dcrutil.Amount(input.AmountIn).String(), input.PreviousOutpoint))\n\t\t}\n\t\tdetailedOutput.WriteString(\"\\nOutputs\\n\")\n\t\tfor _, out := range transaction.Outputs {\n\t\t\tif len(out.Addresses) == 0 {\n\t\t\t\tdetailedOutput.WriteString(fmt.Sprintf(\"%s\\t (no address)\\n\", dcrutil.Amount(out.Value).String()))\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tdetailedOutput.WriteString(fmt.Sprintf(\"%s\", dcrutil.Amount(out.Value).String()))\n\t\t\tfor _, address := range out.Addresses {\n\t\t\t\taccountName := address.AccountName\n\t\t\t\tif !address.IsMine {\n\t\t\t\t\taccountName = \"external\"\n\t\t\t\t}\n\t\t\t\tdetailedOutput.WriteString(fmt.Sprintf(\"\\t%s (%s)\\n\", address.Address, accountName))\n\t\t\t}\n\t\t}\n\t\ttermio.PrintStringResult(strings.TrimRight(detailedOutput.String(), \" \\n\\r\"))\n\t} else {\n\t\ttermio.PrintStringResult(basicOutput)\n\t}\n\treturn nil\n}", "func ValidateTransactionIndex(transactionIndex uint32) (bool) {\n if transactionIndex < MaxReasonableTransactionIndex || transactionIndex == SatoshiConst {\n return true\n }\n return false\n}", "func ShowSchema(c *mgin.Context) {\n\tindex := c.Param(\"index\")\n\tif schema, err := conf.LoadSchema(index); err != nil {\n\t\tc.Error(http.StatusInternalServerError, err.Error())\n\t} else {\n\t\tc.JSON(http.StatusOK, schema.SchemaConf)\n\t}\n}", "func (h *HandlersApp01sqVendor) ListShow(w http.ResponseWriter, offset int, msg string) {\n\tvar err error\n\tvar rcds []App01sqVendor.App01sqVendor\n\tvar name = \"App01sq.Vendor.list.gohtml\"\n\tvar str strings.Builder\n\n\tlog.Printf(\"hndlrVendor.ListShow(%d)\\n\", offset)\n\tlog.Printf(\"\\tname: %s\\n\", name)\n\tw2 := io.MultiWriter(w, &str)\n\n\t// Get the records to display\n\trcds, err = h.db.RowPage(offset, h.rowsPerPage)\n\tif err != nil {\n\n\t\tlog.Printf(\"...end hndlrVendor.ListShow(Error:400) - No Key\\n\")\n\n\t\thttp.Error(w, http.StatusText(400), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tdata := struct {\n\t\tRcds []App01sqVendor.App01sqVendor\n\t\tOffset int\n\t\tMsg string\n\t}{rcds, offset, msg}\n\n\tlog.Printf(\"\\tData: %+v\\n\", data)\n\n\tlog.Printf(\"\\tExecuting template: %s\\n\", name)\n\terr = h.Tmpls.Tmpls.ExecuteTemplate(w2, name, data)\n\tif err != nil {\n\t\tfmt.Fprintf(w, err.Error())\n\t}\n\n\tlog.Printf(\"\\t output: %s\\n\", str.String())\n\tlog.Printf(\"...end hndlrVendor.ListShow(%s)\\n\", util.ErrorString(err))\n}", "func (w *Wallet) List() ([]string, error) {\n\treturn w.store.List()\n}", "func Index(w http.ResponseWriter, r *http.Request) {\n\tc := flight.Context(w, r)\n\n\titems, sum, cnt, _, err := code.All(c.DB)\n\tif err != nil {\n\t\tc.FlashErrorGeneric(err)\n\t\titems = []code.Item{}\n\t}\n\n\tdefaultFormat := \"Mon, 02-Jan-2006\"\n\n\tfor i := 0; i < len(items); i++ {\n\t\titems[i].Trans_Datetime_Formatted = items[i].Trans_Datetime.Time.Format(defaultFormat)\n\t\titems[i].Details_Split = strings.Split(items[i].Details, \"|\")\n\t\titems[i].Amount_String = u.DisplayPrettyNullFloat64(items[i].Amount)\n\t}\n\n\tprettysum := u.DisplayPrettyFloat(sum)\n\tprettycnt := u.DisplayPrettyFloat(cnt)\n\n\tv := c.View.New(\"code/index\")\n\tv.Vars[\"items\"] = items\n\tv.Vars[\"sum\"] = prettysum\n\tv.Vars[\"cnt\"] = prettycnt\n\tv.Render(w, r)\n}", "func (d *Diagnosis) indexHandler(w http.ResponseWriter, r *http.Request) {\n\tvar profiles []profile\n\tfor _, p := range pprof.Profiles() {\n\t\tprofiles = append(profiles, profile{\n\t\t\tName: p.Name(),\n\t\t\tHref: p.Name() + \"?debug=1\",\n\t\t\tDesc: profileDescriptions[p.Name()],\n\t\t\tCount: p.Count(),\n\t\t})\n\t}\n\n\t// Adding other profiles exposed from within this package\n\tfor _, p := range []string{\"cmdline\", \"profile\", \"trace\"} {\n\t\tprofiles = append(profiles, profile{\n\t\t\tName: p,\n\t\t\tHref: p,\n\t\t\tDesc: profileDescriptions[p],\n\t\t})\n\t}\n\n\tsort.Slice(profiles, func(i, j int) bool {\n\t\treturn profiles[i].Name < profiles[j].Name\n\t})\n\n\tif err := indexTmpl.Execute(w, map[string]interface{}{\n\t\t\"AppName\": d.appName,\n\t\t\"PathPrefix\": d.pathPrefix,\n\t\t\"Profiles\": profiles,\n\t}); err != nil {\n\t\td.log.Error(err)\n\t}\n}", "func (c *TradeController) Show(ctx *app.ShowTradeContext) error {\n\t// TradeController_Show: start_implement\n\n\t// Put your logic here\n\tt, ok := tradeRegistry[ctx.TradeID]\n\tif !ok {\n\t\treturn ctx.NotFound()\n\t}\n\tres := &app.GoaTrade{\n\t\tTradeID: t.TradeID,\n\t\tContractID: t.ContractID,\n\t\tCounterpartyID: t.CounterpartyID,\n\t}\n\treturn ctx.OK(res)\n\t// TradeController_Show: end_implement\n}", "func Index(c echo.Context) error {\n\treturn c.Render(http.StatusOK, \"index\", echo.Map{})\n}", "func (b *Blueprint) Index(columns []string, name string, algorithm string) *Blueprint {\n\treturn b.indexCommand(\"index\", columns, name, algorithm)\n}", "func index(w http.ResponseWriter, r *http.Request) {\n\tfmt.Fprintln(w, \"RadioChecker API v0.4\\n\\nCopyright (C) 2017 Paul Haunschmied.\\nwww.radiochecker.com\")\n}", "func (c *Jrpc) ShowPrivacyAccountSpend(in *pty.ReqPrivBal4AddrToken, result *json.RawMessage) error {\n\tif 0 == len(in.Addr) {\n\t\treturn types.ErrInvalidParam\n\t}\n\treply, err := c.cli.ExecWalletFunc(pty.PrivacyX, \"ShowPrivacyAccountSpend\", in)\n\tif err != nil {\n\t\tlog.Info(\"ShowPrivacyAccountSpend\", \"return err info\", err)\n\t\treturn err\n\t}\n\t*result, err = types.PBToJSON(reply)\n\treturn err\n}", "func ShowIndexPage(c *gin.Context) {\n\tisLoggedIn, _ := c.Get(\"is_logged_in\")\n\t//fresh index page according to the values\n\trender(c, gin.H{\n\t\t\"is_logged_in\": isLoggedIn,\n\t\t\"title\": \"Welcome to Micro Message!\"}, \"index.html\")\n}", "func index(w http.ResponseWriter, r *http.Request) {\n\tdata := page{Title: \"School Database\", Header: \"Welcome, please select an option\"}\n\ttemplateInit(w, \"index.html\", data)\n\n}", "func (h *Stocks) Index(ctx context.Context, w http.ResponseWriter, r *http.Request, _ map[string]string) error {\n\n\tclaims, err := auth.ClaimsFromContext(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfields := []datatable.DisplayField{\n\t\t{Field: \"id\", Title: \"ID\", Visible: false, Searchable: true, Orderable: true, Filterable: false},\n\t\t{Field: \"product_id\", Title: \"Product ID\", Visible: false, Searchable: false, Orderable: false, Filterable: false},\n\t\t{Field: \"product_name\", Title: \"Product\", Visible: true, Searchable: true, Orderable: true, Filterable: true, FilterPlaceholder: \"filter Name\"},\n\t\t{Field: \"branch_id\", Title: \"Branch ID\", Visible: false, Searchable: false, Orderable: false, Filterable: false},\n\t\t{Field: \"branch_name\", Title: \"Branch\", Visible: true, Searchable: true, Orderable: true, Filterable: true, FilterPlaceholder: \"filter Branch\"},\n\t\t{Field: \"quantity\", Title: \"Quantity\", Visible: true, Searchable: false, Orderable: true, Filterable: false, },\n\t\t{Field: \"type\", Title: \"Type\", Visible: true, Searchable: false, Orderable: true, Filterable: true, FilterPlaceholder: \"filter Transaction Type\"},\n\t\t{Field: \"narration\", Title: \"Narration\", Visible: true, Searchable: false, Orderable: true, Filterable: true, FilterPlaceholder: \"filter Transaction Type\"},\n\t\t{Field: \"created_at\", Title: \"Date\", Visible: true, Searchable: false, Orderable: true, Filterable: true, },\n\t\t{Field: \"sales_rep\", Title: \"Added By\", Visible: true, Searchable: false, Orderable: true, Filterable: true, },\n\t}\n\n\tmapFunc := func(q *inventory.Response, cols []datatable.DisplayField) (resp []datatable.ColumnValue, err error) {\n\t\tfor i := 0; i < len(cols); i++ {\n\t\t\tcol := cols[i]\n\t\t\tvar v datatable.ColumnValue\n\t\t\tswitch col.Field {\n\t\t\tcase \"id\":\n\t\t\t\tv.Value = fmt.Sprintf(\"%s\", q.ID)\n\t\t\tcase \"product_id\":\n\t\t\t\tv.Value = fmt.Sprintf(\"%s\", q.ProductID)\n\t\t\tcase \"product_name\":\n\t\t\t\tv.Value = q.Product\n\t\t\t\tv.Formatted = fmt.Sprintf(\"<a href='%s'>%s</a>\", urlProductsView(q.ProductID), v.Value)\n\t\t\tcase \"branch_id\":\n\t\t\t\tv.Value = fmt.Sprintf(\"%s\", q.BranchID)\n\t\t\tcase \"branch_name\":\n\t\t\t\tv.Value = q.Branch\n\t\t\t\tv.Formatted = fmt.Sprintf(\"<a href='%s'>%s</a>\", urlBranchesView(q.BranchID), v.Value)\n\t\t\tcase \"quantity\":\n\t\t\t\tv.Value = fmt.Sprintf(\"%d\", q.Quantity)\n\t\t\t\tv.Formatted = fmt.Sprintf(\"<a href='%s'>%s</a>\", urlStocksView(q.ID), v.Value)\n\t\t\tcase \"type\":\n\t\t\t\tv.Value = q.TXType\n\t\t\t\tv.Formatted = v.Value\n\t\t\tcase \"narration\":\n\t\t\t\tv.Value = q.Narration\n\t\t\t\tv.Formatted = formatInventoryNarration(q.Narration)\n\t\t\tcase \"created_at\":\n\t\t\t\tv.Value = q.CreatedAt.Local\n\t\t\t\tv.Formatted = fmt.Sprintf(\"<a href='%s'>%s</a>\", urlStocksView(q.ID), v.Value)\n\t\t\tcase \"sales_rep\":\n\t\t\t\tv.Value = q.SalesRep\n\t\t\t\tv.Formatted = fmt.Sprintf(\"<a href='%s'>%s</a>\", urlUsersView(q.SalesRepID), v.Value)\n\n\t\t\tdefault:\n\t\t\t\treturn resp, errors.Errorf(\"Failed to map value for %s.\", col.Field)\n\t\t\t}\n\t\t\tresp = append(resp, v)\n\t\t}\n\n\t\treturn resp, nil\n\t}\n\n\tloadFunc := func(ctx context.Context, sorting string, fields []datatable.DisplayField) (resp [][]datatable.ColumnValue, err error) {\n\t\tif len(sorting) == 0 {\n\t\t\tsorting = \"created_at desc\"\n\t\t}\n\t\tres, err := h.Repo.Find(ctx, claims, inventory.FindRequest{\n\t\t\tOrder: strings.Split(sorting, \",\"), IncludeProduct: true, IncludeBranch: true, IncludeSalesRep: true,\n\t\t})\n\n\t\tif err != nil {\n\t\t\treturn resp, err\n\t\t}\n\n\t\tfor _, a := range res.Transactions {\n\t\t\tl, err := mapFunc(a, fields)\n\t\t\tif err != nil {\n\t\t\t\treturn resp, errors.Wrapf(err, \"Failed to map category for display.\")\n\t\t\t}\n\n\t\t\tresp = append(resp, l)\n\t\t}\n\n\t\treturn resp, nil\n\t}\n\n\tdt, err := datatable.New(ctx, w, r, h.Redis, fields, loadFunc)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif dt.HasCache() {\n\t\treturn nil\n\t}\n\n\tif ok, err := dt.Render(); ok {\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t}\n\n\tdata := map[string]interface{}{\n\t\t\"datatable\": dt.Response(),\n\t\t\"urlStocksCreate\": urlStocksCreate(),\n\t\t\"urlStocksRemove\": urlStocksRemove(),\n\t\t\"urlStocksIndex\": urlStocksIndex(),\n\t}\n\n\treturn h.Renderer.Render(ctx, w, r, TmplLayoutBase, \"stocks-index.gohtml\", web.MIMETextHTMLCharsetUTF8, http.StatusOK, data)\n}", "func (si ServeIndex) Index(w http.ResponseWriter, r *http.Request) {\n\tpara := params.NewParams()\n\tdata, _, err := Updates(r, para)\n\tif err != nil {\n\t\tif _, ok := err.(params.RenamedConstError); ok {\n\t\t\thttp.Redirect(w, r, err.Error(), http.StatusFound)\n\t\t\treturn\n\t\t}\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tdata[\"Distrib\"] = si.StaticData.Distrib\n\tdata[\"Exporting\"] = exportingCopy() // from ./ws.go\n\tdata[\"OstentUpgrade\"] = OstentUpgrade.Get()\n\tdata[\"OstentVersion\"] = si.StaticData.OstentVersion\n\tdata[\"TAGGEDbin\"] = si.StaticData.TAGGEDbin\n\n\tsi.IndexTemplate.Apply(w, struct{ Data IndexData }{Data: data})\n}", "func (api *API) GetIndex(w http.ResponseWriter, r *http.Request) {\n\n\tinfo := Info{Port: api.Session.Config.API.Port, Versions: Version}\n\td := Metadata{Info: info}\n\n\tres := CodeToResult[CodeOK]\n\tres.Data = d\n\tres.Message = \"Documentation available at https://github.com/netm4ul/netm4ul\"\n\tw.WriteHeader(res.HTTPCode)\n\tjson.NewEncoder(w).Encode(res)\n}", "func TestOpenWallet(t *testing.T) {\n\tassert := assert.New(t)\n\tindysdk.CreateWallet(config, credential)\n\twh, e := indysdk.OpenWallet(config, credential)\n\tassert.Equal(nil, e)\n\tassert.Equal(1, wh)\n\tcleanUp(wh)\n}", "func Show (w http.ResponseWriter, r *http.Request) {\n\tfmt.Fprintln(w, \"This is the Peer List: \")\n\tfmt.Fprintf(w, \"%s\\n%s\", Peers.Show(), SBC.Show())\n}", "func ShowNodeList(w http.ResponseWriter, r *http.Request) {\n\tpeerJSON := nodeList.GetNodeListJSON()\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(http.StatusOK)\n\tw.Write([]byte(peerJSON))\n}", "func (t *Trie) Show() {\n\tfmt.Println(\"root\")\n\tfmt.Println(\"|\")\n\tfor _, n := range t.root.getChilds() {\n\t\tn.show()\n\t}\n}", "func Index(w http.ResponseWriter, req *http.Request) {\n\tfmt.Fprint(w, \"El servidor esta funcionando\")\n\n\tmatriz := pedidosAnuales.BuscarNodo(2017).meses.searchByContent(04).data\n\tfmt.Println(matriz)\n\tmatriz.lst_h.print()\n\tmatriz.lst_v.print()\n\tfmt.Println(matriz.getGraphviz())\n}", "func (v *RealVerifier) Index(vrfProof []byte, directoryID, userID string) ([]byte, error) {\n\tindex, err := v.vrf.ProofToHash([]byte(userID), vrfProof)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"vrf.ProofToHash(): %v\", err)\n\t}\n\treturn index[:], nil\n}", "func (p2pkc *P2PKeysController) Index(c *gin.Context) {\n\tkeys, err := p2pkc.App.GetKeyStore().P2P().GetAll()\n\tif err != nil {\n\t\tjsonAPIError(c, http.StatusInternalServerError, err)\n\t\treturn\n\t}\n\tjsonAPIResponse(c, presenters.NewP2PKeyResources(keys), \"p2pKey\")\n}", "func index(w http.ResponseWriter, r *http.Request){\n\terr := templ.ExecuteTemplate(w, \"index\", nil)\n\tif err != nil {\n\t\tfmt.Print(err.Error())\n\t}\n}", "func (client *WoWClient) MythicKeystoneDungeonIndex(options *RequestOptions) *ApiResponse {\n\treturn client.ApiRequest(DATA, DYNAMIC_NS, \"/mythic-keystone/dungeon/index\", options)\n}", "func (a *TarArchiver) Index(fn func(k string) error) error {\n\treturn fn(slashpath.Join(a.keyPrefix, TarArchiverKey))\n}", "func (k *Index) String() string {\n\treturn fmt.Sprintf(\"%d\", uint32(*k))\n}", "func (_Token *TokenCaller) ShowExchange(opts *bind.CallOpts, amount *big.Int) (struct {\n\tTokenList [][32]byte\n\tAmounts []*big.Int\n}, error) {\n\tret := new(struct {\n\t\tTokenList [][32]byte\n\t\tAmounts []*big.Int\n\t})\n\tout := ret\n\terr := _Token.contract.Call(opts, out, \"showExchange\", amount)\n\treturn *ret, err\n}", "func IndexHandler(w http.ResponseWriter, r *http.Request) {\n\ttitle := \"Visualizing Wizard\"\n buttonContent := \"Sätt igång\"\n\tbt := &ButtonText{title, buttonContent}\n template.Must(template.ParseFiles(\"static/index.html\", \"static/templates/start.tmp\")).Execute(w, bt)\n}", "func Index(w http.ResponseWriter, r *http.Request) {\n\tif err := json.NewEncoder(w).Encode(ResultIndex{Connect: true}); err != nil {\n\t\tpanic(err)\n\t}\n}", "func (u UserConfig) IndexType() string {\n\treturn \"hnsw\"\n}", "func GetAccountsIndex(db gorm.DB, search_vars fp.SearchVars) ([]AccountView, error) {\n\n\tvar rows []AccountView\n\tfmt.Println(\"getttttts=\", search_vars)\n\n\twhere := search_vars.GetSQL(\"company\", \"acc_active\")\n\tfmt.Println(\"where=\", where)\n\tdb.Table(ACCOUNT_VIEW).Select(ACCOUNT_VIEW_COLS).Where(where).Scan(&rows)\n\n\treturn rows, nil\n\n}", "func (b *Blueprint) indexCommand(typ string, columns []string, index string, algorithm string) *Blueprint {\n\t// if no name was specified for this index, we will create one using a bsaic\n\t// convention of the table name, followed by the columns, followd by an\n\t// index type, such as primary or index, which makes the index unique.\n\tif index == \"\" {\n\t\tindex = b.createIndexName(typ, columns)\n\t}\n\n\treturn b.addCommand(typ, &CommandOptions{\n\t\tIndex: index,\n\t\tColumns: columns,\n\t\tAlgorithm: algorithm,\n\t})\n}", "func (ctrl *CSAKeysController) Index(c *gin.Context) {\n\tkeys, err := ctrl.App.GetKeyStore().CSA().ListCSAKeys()\n\tif err != nil {\n\t\tjsonAPIError(c, http.StatusInternalServerError, err)\n\t\treturn\n\t}\n\tjsonAPIResponse(c, presenters.NewCSAKeyResources(keys), \"csaKeys\")\n}", "func (c Project) Index(Id int64) revel.Result {\n\tc.connected()\n\n\t// get the project\n\tobj, err := c.Txn.Get(models.Project{}, Id)\n\tif err != nil || obj == nil {\n\t\t// redirect the user if the project is not found\n\t\treturn c.NotFound(\"Project not found.\")\n\t}\n\tproject := obj.(*models.Project)\n\n\t// count the number of contributors\n\tresults, err := c.Txn.SelectInt(\"select count(DISTINCT user_id) from Transaction WHERE project_id=?\", Id)\n\tvar nbPledge int64 = 0\n\tif err == nil {\n\t\tnbPledge = results\n\t}\n\n\t// count the total amount pledged\n\tsumResults, err := c.Txn.SelectInt(\"select sum(amount) from Transaction WHERE project_id=?\", Id)\n\tvar pledged int64 = 0\n\tif err == nil {\n\t\tpledged = sumResults\n\t}\n\n\t// get the owner of the project\n\townerResult, err := c.Txn.Get(models.User{}, project.OwnerId)\n\towner := ownerResult.(*models.User)\n\n\t// display the page\n\treturn c.Render(project, nbPledge, pledged, owner)\n}", "func (c *AccountController) Show(ctx *app.ShowAccountContext) error {\n\tres := &app.Account{}\n\treturn ctx.OK(res, \"default\")\n}", "func (s *Store) kvsListTxn(tx *memdb.Txn, ws memdb.WatchSet, prefix string) (uint64, structs.DirEntries, error) {\n\t// Get the table indexes.\n\tidx := maxIndexTxn(tx, \"kvs\", \"tombstones\")\n\n\t// Query the prefix and list the available keys\n\tentries, err := tx.Get(\"kvs\", \"id_prefix\", prefix)\n\tif err != nil {\n\t\treturn 0, nil, fmt.Errorf(\"failed kvs lookup: %s\", err)\n\t}\n\tws.Add(entries.WatchCh())\n\n\t// Gather all of the keys found in the store\n\tvar ents structs.DirEntries\n\tvar lindex uint64\n\tfor entry := entries.Next(); entry != nil; entry = entries.Next() {\n\t\te := entry.(*structs.DirEntry)\n\t\tents = append(ents, e)\n\t\tif e.ModifyIndex > lindex {\n\t\t\tlindex = e.ModifyIndex\n\t\t}\n\t}\n\n\t// Check for the highest index in the graveyard. If the prefix is empty\n\t// then just use the full table indexes since we are listing everything.\n\tif prefix != \"\" {\n\t\tgindex, err := s.kvsGraveyard.GetMaxIndexTxn(tx, prefix)\n\t\tif err != nil {\n\t\t\treturn 0, nil, fmt.Errorf(\"failed graveyard lookup: %s\", err)\n\t\t}\n\t\tif gindex > lindex {\n\t\t\tlindex = gindex\n\t\t}\n\t} else {\n\t\tlindex = idx\n\t}\n\n\t// Use the sub index if it was set and there are entries, otherwise use\n\t// the full table index from above.\n\tif lindex != 0 {\n\t\tidx = lindex\n\t}\n\treturn idx, ents, nil\n}", "func NodeIndex(index uint64) (res IMerkleIndex) {\n\tv := nodeIndex(index)\n\treturn &v\n}", "func (cli *CLI) listAddresses() {\n\twallets, _ := wallet.CreateWallets()\n\tfor address := range wallets {\n\t\tfmt.Println(address)\n\t}\n}", "func (u MainController) Index(c *gin.Context) {\n\tuserState := c.MustGet(\"userstate\").(*permissionbolt.UserState)\n\tusername := userState.Username(c.Request)\n\n\tc.HTML(http.StatusOK, \"index\", gin.H{\n\t\t\"username\": username,\n\t\t\"isLogged\": (len(username) > 0),\n\t\t\"isAdmin\": userState.IsAdmin(username),\n\t})\n}", "func (_Token *TokenSession) ShowExchange(amount *big.Int) (struct {\n\tTokenList [][32]byte\n\tAmounts []*big.Int\n}, error) {\n\treturn _Token.Contract.ShowExchange(&_Token.CallOpts, amount)\n}", "func (ge *CurrentGrantExecutable) Show() string {\n\treturn fmt.Sprintf(`SHOW GRANTS OF %v \"%v\"`, ge.granteeType, ge.granteeName)\n}", "func (h *Handler) GetWallet(w http.ResponseWriter, r *http.Request) {\n\twalletID := chi.URLParam(r, \"walletID\")\n\n\twallet := model.Wallet{}\n\terr := wallet.GetWalletByID(walletID)\n\n\tif err != nil {\n\t\tlogger.LogError(fmt.Sprintf(\"%s [%s]\", err.Error(), reqID))\n\t\tw.WriteHeader(http.StatusNotFound)\n\t\treturn\n\t}\n\n\treqID = r.Context().Value(middleware.RequestIDKey)\n\tuser := r.Context().Value(h.cfg.UserCtxKey)\n\tretrieveUser(user)\n\n\tif wallet.UserID != userToken.UserID {\n\t\t//user cannot retrieve other user's transactions\n\t\tlogger.LogDebug(fmt.Sprintf(\"user %s cannot retrieve wallet %s [%s]\", userToken.UserID, wallet.ID, reqID))\n\t\tw.WriteHeader(http.StatusNotFound)\n\t\treturn\n\t}\n\n\trender.Status(r, http.StatusOK)\n\trender.DefaultResponder(w, r, wallet)\n\treturn\n}", "func (client *WoWClient) MythicKeystoneLeaderboardIndex(connectedRealmId int, options *RequestOptions) *ApiResponse {\n\treturn client.ApiRequest(DATA, DYNAMIC_NS, \"/connected-realm/%d/mythic-leaderboard/index\", options, connectedRealmId)\n}", "func (c *Controller) Index(w http.ResponseWriter, r *http.Request) {\n\tw.Write([]byte(\"listing users\"))\n}", "func Index(w http.ResponseWriter, r *http.Request) {\n\tmessage := \"Welcome to Recipe Book!\"\n\tindexT.Execute(w, message)\n}", "func Index(w http.ResponseWriter, r *http.Request) {\r\n\tdb := dbConn()\r\n\tselDB, err := db.Query(\"SELECT * FROM Pegawai ORDER BY id ASC\")\r\n\tif err != nil {\r\n\t\tpanic(err.Error())\r\n\t}\r\n\r\n\t/* Penggunaan Struct ,dipanggil*/\r\n\temp := Pegawai{}\r\n\tres := []Pegawai{}\r\n\r\n\t/* Looping untuk mengambil data */\r\n\tfor selDB.Next() {\r\n\r\n\t\t//buat variabel untuk menampung data\r\n\t\t//sesuaikan sama nama kolom database (huruf kecil)\r\n\t\tvar id int\r\n\t\tvar nama, alamat, jabatan string\r\n\r\n\t\terr = selDB.Scan(&id, &nama, &alamat, &jabatan)\r\n\t\tif err != nil {\r\n\t\t\tpanic(err.Error())\r\n\t\t}\r\n\r\n\t\t//kanan nama var struct - kiri nama kolom database yang diinisialisasikan diatas\r\n\t\temp.Id = id\r\n\t\temp.Nama = nama\r\n\t\temp.Alamat = alamat\r\n\t\temp.Jabatan = jabatan\r\n\t\tres = append(res, emp)\r\n\t}\r\n\r\n\t//kirim data ke view Index\r\n\ttmpl.ExecuteTemplate(w, \"Index\", res)\r\n\tdefer db.Close()\r\n}", "func index() string {\n\tvar buffer bytes.Buffer\n\tvar id = 0\n\tvar class = 0\n\tbuffer.WriteString(indexTemplate)\n\tlock.Lock()\n\tfor folderName, folder := range folders {\n\t\tbuffer.WriteString(fmt.Sprintf(\"<h2>%s</h2>\", folderName))\n\t\tfor _, source := range folder {\n\t\t\tif !anyNonRead(source) {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tsort.Sort(source)\n\t\t\tbuffer.WriteString(fmt.Sprintf(\"<h3>%s</h3>\", source.Title))\n\t\t\tbuffer.WriteString(fmt.Sprintf(`<button onClick=\"hideAll('source_%d'); return false\">Mark all as read</button>`, class))\n\t\t\tbuffer.WriteString(\"<ul>\")\n\n\t\t\tfor _, entry := range source.Entries {\n\t\t\t\tif entry.Read {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tbuffer.WriteString(fmt.Sprintf(`<li id=\"entry_%d\">`, id))\n\t\t\t\tbuffer.WriteString(fmt.Sprintf(`<button class=\"source_%d\" onClick=\"hide('entry_%d', '%s'); return false\">Mark Read</button> `, class, id, entry.Url))\n\t\t\t\tbuffer.WriteString(fmt.Sprintf(`<a href=\"%s\">%s</a>`, entry.Url, entry.Title))\n\t\t\t\tbuffer.WriteString(\"</li>\")\n\t\t\t\tid += 1\n\t\t\t}\n\t\t\tbuffer.WriteString(\"</ul>\")\n\t\t\tclass += 1\n\t\t}\n\t}\n\tlock.Unlock()\n\tbuffer.WriteString(\"</body></html>\")\n\treturn buffer.String()\n}", "func (srv *Server) walletHandler(w http.ResponseWriter, req *http.Request, _ httprouter.Params) {\n\tsiacoinBal, siafundBal, siaclaimBal := srv.wallet.ConfirmedBalance()\n\tsiacoinsOut, siacoinsIn := srv.wallet.UnconfirmedBalance()\n\twriteJSON(w, WalletGET{\n\t\tEncrypted: srv.wallet.Encrypted(),\n\t\tUnlocked: srv.wallet.Unlocked(),\n\n\t\tConfirmedSiacoinBalance: siacoinBal,\n\t\tUnconfirmedOutgoingSiacoins: siacoinsOut,\n\t\tUnconfirmedIncomingSiacoins: siacoinsIn,\n\n\t\tSiafundBalance: siafundBal,\n\t\tSiacoinClaimBalance: siaclaimBal,\n\t})\n}", "func (o SavedAttachedDiskResponseOutput) Index() pulumi.IntOutput {\n\treturn o.ApplyT(func(v SavedAttachedDiskResponse) int { return v.Index }).(pulumi.IntOutput)\n}", "func (_Token *TokenCallerSession) ShowExchange(amount *big.Int) (struct {\n\tTokenList [][32]byte\n\tAmounts []*big.Int\n}, error) {\n\treturn _Token.Contract.ShowExchange(&_Token.CallOpts, amount)\n}", "func (lbq *LatestBlockQuery) OnlyIDX(ctx context.Context) int {\n\tid, err := lbq.OnlyID(ctx)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn id\n}", "func BankCommand(store *wallet.Store) cli.Command {\n\tbankExecutor := func(args []string) error {\n\t\twalletFile, err := wallet.ReadWallet()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(walletFile) == 0 {\n\t\t\tfmt.Println(\"No credentials in wallet.\")\n\t\t\treturn nil\n\t\t}\n\t\tfor _, key := range walletFile {\n\t\t\tpubKey := key.PublicKey\n\t\t\tif len(args) == 1 || args[1] != \"full\" {\n\t\t\t\tpubKey = pubKey[:6]\n\t\t\t}\n\t\t\tfmt.Printf(\"Public key %x:\", pubKey)\n\t\t\tif currencies, ok := store.WalletStore[util.ToBytes(key.PublicKey)]; ok {\n\t\t\t\tfmt.Print(\"\\n\")\n\t\t\t\tfor currency, amount := range currencies {\n\t\t\t\t\tfmt.Printf(\" - %s: %s\\n\", currency, amount)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tfmt.Print(\" no currency\\n\")\n\t\t\t}\n\t\t}\n\t\treturn nil\n\t}\n\n\treturn cli.CreateCommand(\"bank\", \"Display the current balance in each of the local wallets\", bankExecutor, nil)\n}", "func (c *TenantController) Show(ctx *app.ShowTenantContext) error {\n\ttoken := goajwt.ContextJWT(ctx)\n\tif token == nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewUnauthorizedError(\"Missing JWT token\"))\n\t}\n\n\tttoken := &TenantToken{token: token}\n\ttenantID := ttoken.Subject()\n\ttenant, err := c.tenantService.GetTenant(tenantID)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewNotFoundError(\"tenants\", tenantID.String()))\n\t}\n\n\tnamespaces, err := c.tenantService.GetNamespaces(tenantID)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, err)\n\t}\n\n\tresult := &app.TenantSingle{Data: convertTenant(ctx, tenant, namespaces, c.resolveCluster)}\n\treturn ctx.OK(result)\n}", "func (wg *WalletGUI) RecentTransactions(n int, listName string) l.Widget {\n\twg.txMx.Lock()\n\tdefer wg.txMx.Unlock()\n\t// wg.ready.Store(false)\n\tvar out []l.Widget\n\tfirst := true\n\t// out = append(out)\n\tvar wga []btcjson.ListTransactionsResult\n\tswitch listName {\n\tcase \"history\":\n\t\twga = wg.txHistoryList\n\tcase \"recent\":\n\t\twga = wg.txRecentList\n\t}\n\tif len(wga) == 0 {\n\t\treturn func(gtx l.Context) l.Dimensions {\n\t\t\treturn l.Dimensions{Size: gtx.Constraints.Max}\n\t\t}\n\t}\n\tDebug(\">>>>>>>>>>>>>>>> iterating transactions\", n, listName)\n\tfor x := range wga {\n\t\tif x > n && n > 0 {\n\t\t\tbreak\n\t\t}\n\t\t\n\t\ti := x\n\t\ttxs := wga[i]\n\t\t// spacer\n\t\tif !first {\n\t\t\tout = append(out,\n\t\t\t\twg.Inset(0.25, gui.EmptyMaxWidth()).Fn,\n\t\t\t)\n\t\t} else {\n\t\t\tfirst = false\n\t\t}\n\t\tout = append(out,\n\t\t\twg.Fill(\"DocBg\", l.W, 0, 0,\n\t\t\t\twg.Inset(0.25,\n\t\t\t\t\twg.Flex().\n\t\t\t\t\t\tRigid(\n\t\t\t\t\t\t\twg.Body1(fmt.Sprintf(\"%-6.8f DUO\", txs.Amount)).Color(\"PanelText\").Fn,\n\t\t\t\t\t\t).\n\t\t\t\t\t\tFlexed(1,\n\t\t\t\t\t\t\twg.Inset(0.25,\n\t\t\t\t\t\t\t\twg.Caption(txs.Address).\n\t\t\t\t\t\t\t\t\tFont(\"go regular\").\n\t\t\t\t\t\t\t\t\tColor(\"PanelText\").\n\t\t\t\t\t\t\t\t\tTextScale(0.66).\n\t\t\t\t\t\t\t\t\tAlignment(text.End).\n\t\t\t\t\t\t\t\t\tFn,\n\t\t\t\t\t\t\t).Fn,\n\t\t\t\t\t\t).Fn,\n\t\t\t\t).Fn,\n\t\t\t).Fn,\n\t\t)\n\t\t// out = append(out,\n\t\t// \twg.Caption(txs.TxID).\n\t\t// \t\tFont(\"go regular\").\n\t\t// \t\tColor(\"PanelText\").\n\t\t// \t\tTextScale(0.5).Fn,\n\t\t// )\n\t\tout = append(out,\n\t\t\twg.Fill(\"DocBg\", l.W, 0, 0,\n\t\t\t\twg.Inset(0.25,\n\t\t\t\t\twg.Flex().Flexed(1,\n\t\t\t\t\t\twg.Flex().\n\t\t\t\t\t\t\tRigid(\n\t\t\t\t\t\t\t\twg.Flex().\n\t\t\t\t\t\t\t\t\tRigid(\n\t\t\t\t\t\t\t\t\t\twg.Icon().Color(\"PanelText\").Scale(1).Src(&icons2.DeviceWidgets).Fn,\n\t\t\t\t\t\t\t\t\t).\n\t\t\t\t\t\t\t\t\t// Rigid(\n\t\t\t\t\t\t\t\t\t// \twg.Caption(fmt.Sprint(*txs.BlockIndex)).Fn,\n\t\t\t\t\t\t\t\t\t// \t// wg.buttonIconText(txs.clickBlock,\n\t\t\t\t\t\t\t\t\t// \t// \tfmt.Sprint(*txs.BlockIndex),\n\t\t\t\t\t\t\t\t\t// \t// \t&icons2.DeviceWidgets,\n\t\t\t\t\t\t\t\t\t// \t// \twg.blockPage(*txs.BlockIndex)),\n\t\t\t\t\t\t\t\t\t// ).\n\t\t\t\t\t\t\t\t\tRigid(\n\t\t\t\t\t\t\t\t\t\twg.Caption(fmt.Sprintf(\"%d \", txs.BlockIndex)).Fn,\n\t\t\t\t\t\t\t\t\t).\n\t\t\t\t\t\t\t\t\tFn,\n\t\t\t\t\t\t\t).\n\t\t\t\t\t\t\tRigid(\n\t\t\t\t\t\t\t\twg.Flex().\n\t\t\t\t\t\t\t\t\tRigid(\n\t\t\t\t\t\t\t\t\t\twg.Icon().Color(\"PanelText\").Scale(1).Src(&icons2.ActionCheckCircle).Fn,\n\t\t\t\t\t\t\t\t\t).\n\t\t\t\t\t\t\t\t\tRigid(\n\t\t\t\t\t\t\t\t\t\twg.Caption(fmt.Sprintf(\"%d \", txs.Confirmations)).Fn,\n\t\t\t\t\t\t\t\t\t).\n\t\t\t\t\t\t\t\t\tFn,\n\t\t\t\t\t\t\t).\n\t\t\t\t\t\t\tRigid(\n\t\t\t\t\t\t\t\twg.Flex().\n\t\t\t\t\t\t\t\t\tRigid(\n\t\t\t\t\t\t\t\t\t\tfunc(gtx l.Context) l.Dimensions {\n\t\t\t\t\t\t\t\t\t\t\tswitch txs.Category {\n\t\t\t\t\t\t\t\t\t\t\tcase \"generate\":\n\t\t\t\t\t\t\t\t\t\t\t\treturn wg.Icon().Color(\"PanelText\").Scale(1).Src(&icons2.ActionStars).Fn(gtx)\n\t\t\t\t\t\t\t\t\t\t\tcase \"immature\":\n\t\t\t\t\t\t\t\t\t\t\t\treturn wg.Icon().Color(\"PanelText\").Scale(1).Src(&icons2.ImageTimeLapse).Fn(gtx)\n\t\t\t\t\t\t\t\t\t\t\tcase \"receive\":\n\t\t\t\t\t\t\t\t\t\t\t\treturn wg.Icon().Color(\"PanelText\").Scale(1).Src(&icons2.ActionPlayForWork).Fn(gtx)\n\t\t\t\t\t\t\t\t\t\t\tcase \"unknown\":\n\t\t\t\t\t\t\t\t\t\t\t\treturn wg.Icon().Color(\"PanelText\").Scale(1).Src(&icons2.AVNewReleases).Fn(gtx)\n\t\t\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t\t\t\treturn l.Dimensions{}\n\t\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t\t).\n\t\t\t\t\t\t\t\t\tRigid(\n\t\t\t\t\t\t\t\t\t\twg.Caption(txs.Category+\" \").Fn,\n\t\t\t\t\t\t\t\t\t).\n\t\t\t\t\t\t\t\t\tFn,\n\t\t\t\t\t\t\t).\n\t\t\t\t\t\t\tRigid(\n\t\t\t\t\t\t\t\twg.Flex().\n\t\t\t\t\t\t\t\t\tRigid(\n\t\t\t\t\t\t\t\t\t\twg.Icon().Color(\"PanelText\").Scale(1).Src(&icons2.DeviceAccessTime).Fn,\n\t\t\t\t\t\t\t\t\t).\n\t\t\t\t\t\t\t\t\tRigid(\n\t\t\t\t\t\t\t\t\t\twg.Caption(\n\t\t\t\t\t\t\t\t\t\t\ttime.Unix(txs.Time,\n\t\t\t\t\t\t\t\t\t\t\t\t0).Format(\"02 Jan 06 15:04:05 MST\"),\n\t\t\t\t\t\t\t\t\t\t).Color(\"PanelText\").Fn,\n\t\t\t\t\t\t\t\t\t).\n\t\t\t\t\t\t\t\t\tFn,\n\t\t\t\t\t\t\t).Fn,\n\t\t\t\t\t).Fn,\n\t\t\t\t).Fn,\n\t\t\t).Fn,\n\t\t)\n\t}\n\tle := func(gtx l.Context, index int) l.Dimensions {\n\t\treturn out[index](gtx)\n\t}\n\two := func(gtx l.Context) l.Dimensions {\n\t\treturn wg.lists[listName].\n\t\t\tVertical().\n\t\t\tLength(len(out)).\n\t\t\tListElement(le).\n\t\t\tFn(gtx)\n\t}\n\tDebug(\">>>>>>>>>>>>>>>> history widget completed\", n, listName)\n\tswitch listName {\n\tcase \"history\":\n\t\twg.HistoryWidget = wo\n\t\tif !wg.ready.Load() {\n\t\t\twg.ready.Store(true)\n\t\t}\n\tcase \"recent\":\n\t\twg.RecentTransactionsWidget = wo\n\t}\n\treturn func(gtx l.Context) l.Dimensions {\n\t\treturn wo(gtx)\n\t}\n}", "func (e *Card) Index(index bleve.Index) error {\n\terr := index.Index(string(e.Id), e)\n\treturn err\n}", "func Index(w http.ResponseWriter, r *http.Request) {\n\tfmt.Fprint(w, \"s-senpai, please don't hurt me ;_;\\n\")\n}", "func index(w http.ResponseWriter, req *http.Request, _ httprouter.Params) {\n\terr := tpl.ExecuteTemplate(w, \"index.html\", nil)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\tlog.Fatalln(err)\n\t}\n\tfmt.Println(\"HERE INDEX\")\n}", "func (c *AccountController) Show(ctx *app.ShowAccountContext) error {\n\tu, err := repositories.GetUserByID(ctx.AccountID)\n\tif err != nil {\n\t\treturn ctx.NotFound()\n\t}\n\tres := &app.FtAccount{\n\t\tID: u.ID,\n\t\tFirstName: u.FirstName,\n\t\tLastName: u.LastName,\n\t\tEmail: u.Email,\n\t}\n\treturn ctx.OK(res)\n}", "func (gw2 *GW2Api) AccountWallet() (currency []WalletCurrency, err error) {\n\tver := \"v2\"\n\ttag := \"account/materials\"\n\terr = gw2.fetchAuthenticatedEndpoint(ver, tag, PermWallet, nil, &currency)\n\treturn\n}", "func (bc *Blockchain) PrintBlockInfo(index int) {\n block := bc.Chain[index]\n fmt.Println(\"Index of the block is \" + strconv.Itoa(block.Index))\n fmt.Println(\"Timestamp of the block is \" + time.Unix(block.Timestamp, 0).Format(time.UnixDate))\n fmt.Println(\"Proof of the block is \" + strconv.Itoa(block.Proof))\n fmt.Println(\"Hash of the previous block is \" + block.PreviousHash)\n fmt.Println(\"Hash of the current block is \" + bc.HashBlock(block))\n fmt.Println(\"Difficulty of the block is \" + block.Difficulty)\n fmt.Println(\"\\n\\n\")\n}", "func (s MutableGetRequestReceiptResults) BlockIndex() wasmtypes.ScMutableUint32 {\n\treturn wasmtypes.NewScMutableUint32(s.Proxy.Root(ResultBlockIndex))\n}" ]
[ "0.63500696", "0.5926195", "0.5621318", "0.55879503", "0.558767", "0.5469524", "0.5401371", "0.5312564", "0.5297335", "0.5275967", "0.5268884", "0.52070117", "0.5178347", "0.5161064", "0.51266795", "0.5098225", "0.5020761", "0.500083", "0.49379826", "0.49024478", "0.4888832", "0.4888832", "0.4881241", "0.48564103", "0.48448625", "0.48184088", "0.48038423", "0.47926414", "0.4756553", "0.4710004", "0.47051758", "0.46973905", "0.46914223", "0.46901146", "0.46794942", "0.46792668", "0.46509543", "0.46477604", "0.4644129", "0.46433142", "0.46054837", "0.46009773", "0.45896596", "0.45596185", "0.45515382", "0.4548419", "0.45483738", "0.45325315", "0.45205247", "0.45166403", "0.45158684", "0.45115936", "0.44783035", "0.4475332", "0.44712564", "0.44634706", "0.4462393", "0.4450767", "0.4450516", "0.44460097", "0.44350994", "0.4433389", "0.44312778", "0.44255224", "0.44254673", "0.44233507", "0.44207066", "0.44166276", "0.44112366", "0.44061217", "0.43969375", "0.4375884", "0.4373718", "0.43687788", "0.43660405", "0.43637532", "0.43627068", "0.43582797", "0.43564597", "0.43464202", "0.43353924", "0.43327773", "0.4328596", "0.4327562", "0.4327201", "0.43227914", "0.43226367", "0.43207863", "0.43063414", "0.4305409", "0.43021196", "0.42963517", "0.42934772", "0.42933774", "0.42930925", "0.42863733", "0.4283911", "0.42790955", "0.42751083", "0.4266733" ]
0.833931
0
ShowTx is used to show a transaction belongs to us.
func (account *Account) ShowTx(P string) bool { for i := 0; i < account.WalletIndex; i++ { if P == account.Wallet[i].P { JSON, _ := json.MarshalIndent(account.Wallet[i], "", " ") fmt.Println(string(JSON)) return true } } return false }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (tc *TransactionsController) Show(c *gin.Context) {\n\thash := common.HexToHash(c.Param(\"TxHash\"))\n\tif tx, err := tc.App.GetStore().FindTxByAttempt(hash); err == orm.ErrorNotFound {\n\t\tpublicError(c, http.StatusNotFound, errors.New(\"Transaction not found\"))\n\t} else if err != nil {\n\t\tc.AbortWithError(http.StatusInternalServerError, err)\n\t} else if doc, err := jsonapi.Marshal(presenters.NewTx(tx)); err != nil {\n\t\tc.AbortWithError(http.StatusInternalServerError, err)\n\t} else {\n\t\tc.Data(http.StatusOK, MediaType, doc)\n\t}\n}", "func (account *Account) ShowAccount() {\n\tfmt.Println(\"a:\", account.a)\n\tfmt.Println(\"A:\", account.A)\n\tfmt.Println(\"AT:\", account.AT)\n}", "func (c *PurchaseController) Show(ctx *app.ShowPurchaseContext) error {\n\n\tsession := Database.Session.Copy()\n\tdefer session.Close()\n\n\tresult := app.Purchase{}\n\n\terr := session.DB(\"services-pos\").C(\"Purchase\").FindId(bson.ObjectIdHex(ctx.TransactionID)).One(&result)\n\n\tif err != nil {\n\t\treturn ctx.NotFound()\n\t}\n\n\tresult.TransactionID = ctx.TransactionID\n\tresult.Href = app.PurchaseHref(ctx.TransactionID)\n\n\treturn ctx.OK(&result)\n}", "func (showTxCommand ShowTransactionCommand) Run(ctx context.Context, wallet walletcore.Wallet) error {\n\ttransaction, err := wallet.GetTransaction(showTxCommand.Args.TxHash)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tbasicOutput := \"Hash\\t%s\\n\" +\n\t\t\"Confirmations\\t%d\\n\" +\n\t\t\"Included in block\\t%d\\n\" +\n\t\t\"Type\\t%s\\n\" +\n\t\t\"Amount %s\\t%s\\n\" +\n\t\t\"Date\\t%s\\n\" +\n\t\t\"Size\\t%s\\n\" +\n\t\t\"Fee\\t%s\\n\" +\n\t\t\"Rate\\t%s/kB\\n\"\n\n\ttxDirection := strings.ToLower(transaction.Direction.String())\n\ttxSize := fmt.Sprintf(\"%.1f kB\", float64(transaction.Size)/1000)\n\tbasicOutput = fmt.Sprintf(basicOutput,\n\t\ttransaction.Hash,\n\t\ttransaction.Confirmations,\n\t\ttransaction.BlockHeight,\n\t\ttransaction.Type,\n\t\ttxDirection, transaction.Amount,\n\t\ttransaction.FormattedTime,\n\t\ttxSize,\n\t\ttransaction.Fee,\n\t\ttransaction.FeeRate)\n\n\tif showTxCommand.Detailed {\n\t\tdetailedOutput := strings.Builder{}\n\t\tdetailedOutput.WriteString(\"General Info\\n\")\n\t\tdetailedOutput.WriteString(basicOutput)\n\t\tdetailedOutput.WriteString(\"\\nInputs\\n\")\n\t\tfor _, input := range transaction.Inputs {\n\t\t\tdetailedOutput.WriteString(fmt.Sprintf(\"%s\\t%s\\n\", dcrutil.Amount(input.AmountIn).String(), input.PreviousOutpoint))\n\t\t}\n\t\tdetailedOutput.WriteString(\"\\nOutputs\\n\")\n\t\tfor _, out := range transaction.Outputs {\n\t\t\tif len(out.Addresses) == 0 {\n\t\t\t\tdetailedOutput.WriteString(fmt.Sprintf(\"%s\\t (no address)\\n\", dcrutil.Amount(out.Value).String()))\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tdetailedOutput.WriteString(fmt.Sprintf(\"%s\", dcrutil.Amount(out.Value).String()))\n\t\t\tfor _, address := range out.Addresses {\n\t\t\t\taccountName := address.AccountName\n\t\t\t\tif !address.IsMine {\n\t\t\t\t\taccountName = \"external\"\n\t\t\t\t}\n\t\t\t\tdetailedOutput.WriteString(fmt.Sprintf(\"\\t%s (%s)\\n\", address.Address, accountName))\n\t\t\t}\n\t\t}\n\t\ttermio.PrintStringResult(strings.TrimRight(detailedOutput.String(), \" \\n\\r\"))\n\t} else {\n\t\ttermio.PrintStringResult(basicOutput)\n\t}\n\treturn nil\n}", "func displayTransactions(scid string, option int, ID string) {\n\t\t\n\tscKeys:= []string{\"numberOfOwners\", \"txCount\"}\n\tresult:= getKeysFromDaemon(scid, scKeys)\n\tif result == \"\" {return}\n\n\n\t//Response ok, extract keys from JSON\n\t\n\n\ttxCount := gjson.Get(result, \"txs.#.sc_keys.txCount\")\n\ttxCountArray:= txCount.Array()[0]\n\ttxCountInt:= txCountArray.Int()\n\t//fmt.Printf(\"Tx Count: %d\\n\", txCountInt)\n\n\t//Make a slice of keys so we can request in RPC call\n\tx:= int(txCountInt) //txCount in wallet smart contract is always 1 ahead of actual number of transactions\t\n\tx4:= x * 4\t\n\tkeySlice:= make([]string, x4) \n\t\n\tfor i:=0; i<x; i++ {\n\t\tz:= strconv.Itoa(i)\n\t\tkeySlice[i] = \"txIndex_\" + z\n\t\tkeySlice[i+x] = \"recipient_\" + z\n\t\tkeySlice[i+(x*2)] = \"amount_\" + z\n\t\tkeySlice[i+(x*3)] = \"sent_\" + z\n\t}\n\t\t\n\t//fmt.Println(keySlice)\n\tdisplayData(scid, keySlice, x, option, ID)\n\n\n}", "func (p *Proxy) handleShowTxnz(session *driver.Session, query string, node sqlparser.Statement) (*sqltypes.Result, error) {\n\treturn nil, sqldb.NewSQLErrorf(sqldb.ER_SPECIFIC_ACCESS_DENIED_ERROR, \"Access denied; lacking super privilege for the operation\")\n\t// privilegePlug := spanner.plugins.PlugPrivilege()\n\t// if !privilegePlug.IsSuperPriv(session.User()) {\n\t// \treturn nil, sqldb.NewSQLErrorf(sqldb.ER_SPECIFIC_ACCESS_DENIED_ERROR, \"Access denied; lacking super privilege for the operation\")\n\t// }\n\n\t// qr := &sqltypes.Result{}\n\t// qr.Fields = []*querypb.Field{\n\t// \t{Name: \"TxnID\", Type: querypb.Type_INT64},\n\t// \t{Name: \"Start\", Type: querypb.Type_VARCHAR},\n\t// \t{Name: \"Duration\", Type: querypb.Type_INT32},\n\t// \t{Name: \"XaState\", Type: querypb.Type_VARCHAR},\n\t// \t{Name: \"TxnState\", Type: querypb.Type_VARCHAR},\n\t// }\n\n\t// rows := spanner.scatter.Txnz().GetTxnzRows()\n\t// for _, row := range rows {\n\t// \trow := []sqltypes.Value{\n\t// \t\tsqltypes.MakeTrusted(querypb.Type_INT64, []byte(fmt.Sprintf(\"%v\", uint64(row.TxnID)))),\n\t// \t\tsqltypes.MakeTrusted(querypb.Type_VARCHAR, []byte(row.Start.Format(\"20060102150405.000\"))),\n\t// \t\tsqltypes.MakeTrusted(querypb.Type_INT32, []byte(fmt.Sprintf(\"%v\", row.Duration))),\n\t// \t\tsqltypes.MakeTrusted(querypb.Type_VARCHAR, []byte(row.XaState)),\n\t// \t\tsqltypes.MakeTrusted(querypb.Type_VARCHAR, []byte(row.State)),\n\t// \t}\n\t// \tqr.Rows = append(qr.Rows, row)\n\t// }\n\t// return qr, nil\n}", "func ShowTxStatusTracker(stdout io.Writer, hash string, rpcClient types2.Client) error {\n\ts := spinner.New(spinner.CharSets[14], 100*time.Millisecond)\n\ts.Writer = stdout\n\ts.Prefix = \" \"\n\ts.Start()\n\tlastStatus := \"\"\n\n\tvar err error\n\tvar resp *api2.ResultTx\n\tattempts := 0\n\tfor {\n\t\tattempts += 1\n\t\tif attempts == 3 {\n\t\t\treturn err\n\t\t}\n\n\t\ttime.Sleep(1 * time.Second)\n\t\tresp, err = api.GetTransaction(hash, rpcClient)\n\t\tif err != nil {\n\t\t\ts.Stop()\n\t\t\tcontinue\n\t\t}\n\n\t\tif lastStatus == resp.Status {\n\t\t\tcontinue\n\t\t}\n\n\t\tlastStatus = resp.Status\n\t\tif resp.Status == types3.TxStatusInMempool {\n\t\t\ts.Suffix = colorfmt.YellowStringf(\" In mempool\")\n\t\t} else if resp.Status == types3.TxStatusInPushpool {\n\t\t\ts.Suffix = colorfmt.YellowStringf(\" In pushpool\")\n\t\t} else {\n\t\t\ts.FinalMSG = colorfmt.GreenString(\" Confirmed!\\n\")\n\t\t\ts.Stop()\n\t\t\tbreak\n\t\t}\n\t}\n\treturn nil\n}", "func (t *Trans) QueryTx(txid string) (*pb.TxStatus, error) {\n\treturn t.Xchain.QueryTx(txid)\n}", "func (c *TradeController) Show(ctx *app.ShowTradeContext) error {\n\t// TradeController_Show: start_implement\n\n\t// Put your logic here\n\tt, ok := tradeRegistry[ctx.TradeID]\n\tif !ok {\n\t\treturn ctx.NotFound()\n\t}\n\tres := &app.GoaTrade{\n\t\tTradeID: t.TradeID,\n\t\tContractID: t.ContractID,\n\t\tCounterpartyID: t.CounterpartyID,\n\t}\n\treturn ctx.OK(res)\n\t// TradeController_Show: end_implement\n}", "func (f Factory) PrintUnsignedTx(clientCtx client.Context, msgs ...sdk.Msg) error {\n\tif f.SimulateAndExecute() {\n\t\tif clientCtx.Offline {\n\t\t\treturn errors.New(\"cannot estimate gas in offline mode\")\n\t\t}\n\n\t\t// Prepare TxFactory with acc & seq numbers as CalculateGas requires\n\t\t// account and sequence numbers to be set\n\t\tpreparedTxf, err := f.Prepare(clientCtx)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t_, adjusted, err := CalculateGas(clientCtx, preparedTxf, msgs...)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tf = f.WithGas(adjusted)\n\t\t_, _ = fmt.Fprintf(os.Stderr, \"%s\\n\", GasEstimateResponse{GasEstimate: f.Gas()})\n\t}\n\n\tunsignedTx, err := f.BuildUnsignedTx(msgs...)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tjson, err := clientCtx.TxConfig.TxJSONEncoder()(unsignedTx.GetTx())\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn clientCtx.PrintString(fmt.Sprintf(\"%s\\n\", json))\n}", "func (o *APIServiceSendTransactionParams) SetTx(tx string) {\n\to.Tx = tx\n}", "func (st *SignedTx) Tx() *btcutil.Tx {\n\treturn st.tx\n}", "func (tx Transaction) String() string {\n\tvar lines []string\n\n\tlines = append(lines, fmt.Sprintf(\"--- Transaction %x: \", tx.ID))\n\tfor i, input := range tx.Inputs {\n\t\tlines = append(lines, fmt.Sprintf(\" Input %d: \", i))\n\t\tlines = append(lines, fmt.Sprintf(\" TxID: %x\", input.ID))\n\t\tlines = append(lines, fmt.Sprintf(\" Out : %d\", input.Out))\n\t\tlines = append(lines, fmt.Sprintf(\" Signature: %x\", input.Signature))\n\t\tlines = append(lines, fmt.Sprintf(\" PubKey: %x\", input.PubKey))\n\t}\n\n\tfor i, output := range tx.Outputs {\n\t\tlines = append(lines, fmt.Sprintf(\" Output %d:\", i))\n\t\tlines = append(lines, fmt.Sprintf(\" Value: %d\", output.Value))\n\t\tlines = append(lines, fmt.Sprintf(\" PubKeyHash: %x\", output.PubKeyHash))\n\t}\n\n\treturn strings.Join(lines, \"\\n\")\n}", "func (c *Jrpc) ShowPrivacyAccountSpend(in *pty.ReqPrivBal4AddrToken, result *json.RawMessage) error {\n\tif 0 == len(in.Addr) {\n\t\treturn types.ErrInvalidParam\n\t}\n\treply, err := c.cli.ExecWalletFunc(pty.PrivacyX, \"ShowPrivacyAccountSpend\", in)\n\tif err != nil {\n\t\tlog.Info(\"ShowPrivacyAccountSpend\", \"return err info\", err)\n\t\treturn err\n\t}\n\t*result, err = types.PBToJSON(reply)\n\treturn err\n}", "func (db database) Tx(txHash id.Hash) (tx.Tx, error) {\n\tscript := \"SELECT hash, selector, txid, txindex, amount, payload, phash, to_address, nonce, nhash, gpubkey, ghash, version FROM txs WHERE hash = $1\"\n\trow := db.db.QueryRow(script, txHash.String())\n\terr := row.Err()\n\tif err != nil {\n\t\treturn tx.Tx{}, err\n\t}\n\treturn rowToTx(row)\n}", "func ViewTransaction(rw http.ResponseWriter, r *http.Request) {\n\t// get the token\n\treqToken := r.Header.Get(\"Authorization\")\n\n\t// get the claims\n\tclaims, isNotValid := GetClaims(reqToken, rw)\n\tif isNotValid {\n\t\treturn\n\t}\n\n\tdt, err := db.GetUserTransaction(claims.Roll)\n\tif err != nil {\n\t\trw.WriteHeader(http.StatusInternalServerError)\n\t\trw.Write(Rsp(err.Error(), \"Server Error\"))\n\t\treturn\n\t}\n\trw.WriteHeader(http.StatusOK)\n\tres := c.RespData{\n\t\tMessage: \"All data\",\n\t\tData: dt,\n\t}\n\tjson.NewEncoder(rw).Encode(res)\n}", "func (f *userFactory) Tx(tx *ent.Tx) *userFactory {\n\treturn f.Client(tx.Client())\n}", "func (check *Check) Tx(tx string) *Check {\n\tcheck.tx = &tx\n\treturn check\n}", "func (tx Transaction) String() string {\n\tvar lines []string\n\n\tlines = append(lines, fmt.Sprintf(\"--- Transaction %x:\", tx.ID))\n\n\tfor i, input := range tx.Vin {\n\n\t\tlines = append(lines, fmt.Sprintf(\" Input %d:\", i))\n\t\tlines = append(lines, fmt.Sprintf(\" TXID: %x\", input.Txid))\n\t\tlines = append(lines, fmt.Sprintf(\" Out: %d\", input.Vout))\n\t\tlines = append(lines, fmt.Sprintf(\" Signature: %x\", input.Signature))\n\t\tlines = append(lines, fmt.Sprintf(\" PubKey: %x\", input.PubKey))\n\t}\n\n\tfor i, output := range tx.Vout {\n\t\tlines = append(lines, fmt.Sprintf(\" Output %d:\", i))\n\t\tlines = append(lines, fmt.Sprintf(\" Value: %d\", output.Value))\n\t\tlines = append(lines, fmt.Sprintf(\" Script: %x\", output.PubKeyHash))\n\t}\n\n\treturn strings.Join(lines, \"\\n\")\n}", "func (tx Transaction) String() string {\n\tvar lines []string\n\n\tlines = append(lines, fmt.Sprintf(\"--- Transaction %x:\", tx.ID))\n\tfor i, input := range tx.Inputs {\n\t\tlines = append(lines, fmt.Sprintf(\" Input %d:\", i))\n\t\tlines = append(lines, fmt.Sprintf(\" TXID: %x\", input.ID))\n\t\tlines = append(lines, fmt.Sprintf(\" Out: %d\", input.Out))\n\t\tlines = append(lines, fmt.Sprintf(\" Signature: %x\", input.Signature))\n\t\tlines = append(lines, fmt.Sprintf(\" PubKey: %x\", input.PubKey))\n\t}\n\n\tfor i, output := range tx.Outputs {\n\t\tlines = append(lines, fmt.Sprintf(\" Output %d:\", i))\n\t\tlines = append(lines, fmt.Sprintf(\" Value: %d\", output.Value))\n\t\tlines = append(lines, fmt.Sprintf(\" Script: %x\", output.PubKeyHash))\n\t}\n\n\treturn strings.Join(lines, \"\\n\")\n}", "func (c *AccountController) Show(ctx *app.ShowAccountContext) error {\n\tu, err := repositories.GetUserByID(ctx.AccountID)\n\tif err != nil {\n\t\treturn ctx.NotFound()\n\t}\n\tres := &app.FtAccount{\n\t\tID: u.ID,\n\t\tFirstName: u.FirstName,\n\t\tLastName: u.LastName,\n\t\tEmail: u.Email,\n\t}\n\treturn ctx.OK(res)\n}", "func (me TxsdShow) String() string { return xsdt.String(me).String() }", "func (me TxsdShow) String() string { return xsdt.String(me).String() }", "func (Logger) CheckTx(ctx sdk.Context, store sdk.SimpleDB, tx interface{}, next sdk.Checker) (res sdk.CheckResult, err error) {\n\tstart := time.Now()\n\tres, err = next.CheckTx(ctx, store, tx)\n\tdelta := time.Now().Sub(start)\n\t// TODO: log some info on the tx itself?\n\tl := ctx.With(\"duration\", micros(delta))\n\tif err == nil {\n\t\tl.Debug(\"CheckTx\", \"log\", res.Log)\n\t} else {\n\t\tl.Info(\"CheckTx\", \"err\", err)\n\t}\n\treturn\n}", "func (dao *Dao) Txn(opts *sql.TxOptions) (*DaoTxnContext, error) {\n\treturn dao.TxnWithContext(context.Background(), opts)\n}", "func (xc *AChainCore) GenerateTx(in *pb.TxData, hd *global.XContext) *pb.TxStatus {\n\tout := &pb.TxStatus{Header: in.Header}\n\tout.Header.Error = pb.AChainErrorEnum_SUCCESS\n\tif xc.Status() != global.Normal {\n\t\tout.Header.Error = pb.AChainErrorEnum_CONNECT_REFUSE // 拒绝\n\t\treturn out\n\t}\n\tt, err := xc.Utxovm.GenerateTx(in)\n\txc.Speed.Add(\"GenerateTx\")\n\tif err != nil {\n\t\tout.Header.Error = HandlerUtxoError(err)\n\t} else {\n\t\tout.Tx = t\n\t\tout.Bcname = in.Bcname\n\t\tout.Txid = out.Tx.Txid\n\t}\n\treturn out\n}", "func txHexString(tx *wire.MsgTx) string {\n\tbuf := bytes.NewBuffer(make([]byte, 0, tx.SerializeSize()))\n\t// Ignore Serialize's error, as writing to a bytes.buffer cannot fail.\n\ttx.Serialize(buf)\n\treturn hex.EncodeToString(buf.Bytes())\n}", "func (st *SignedTx) TxInfo(account string, chainHeight int32, net btcwire.BitcoinNet) []map[string]interface{} {\n\treply := make([]map[string]interface{}, len(st.tx.MsgTx().TxOut))\n\n\tvar confirmations int32\n\tif st.block != nil {\n\t\tconfirmations = chainHeight - st.block.Height + 1\n\t}\n\n\tfor i, txout := range st.tx.MsgTx().TxOut {\n\t\taddress := \"Unknown\"\n\t\t_, addrs, _, _ := btcscript.ExtractPkScriptAddrs(txout.PkScript, net)\n\t\tif len(addrs) == 1 {\n\t\t\taddress = addrs[0].EncodeAddress()\n\t\t}\n\t\tinfo := map[string]interface{}{\n\t\t\t\"account\": account,\n\t\t\t\"address\": address,\n\t\t\t\"category\": \"send\",\n\t\t\t\"amount\": float64(-txout.Value) / float64(btcutil.SatoshiPerBitcoin),\n\t\t\t\"fee\": float64(st.Fee()) / float64(btcutil.SatoshiPerBitcoin),\n\t\t\t\"confirmations\": float64(confirmations),\n\t\t\t\"txid\": st.txSha.String(),\n\t\t\t\"time\": float64(st.created.Unix()),\n\t\t\t\"timereceived\": float64(st.created.Unix()),\n\t\t}\n\t\tif st.block != nil {\n\t\t\tinfo[\"blockhash\"] = st.block.Hash.String()\n\t\t\tinfo[\"blockindex\"] = float64(st.block.Index)\n\t\t\tinfo[\"blocktime\"] = float64(st.block.Time.Unix())\n\t\t}\n\t\treply[i] = info\n\t}\n\n\treturn reply\n}", "func (t *Transaction) String() string {\n\treturn t.From + \" -> \" + t.To + \" : \" + strconv.Itoa(t.Amount)\n}", "func view(ctx context.Context, db transactor, fn func(*bolt.Tx) error) error {\n\ttx, ok := ctx.Value(transactionKey{}).(*bolt.Tx)\n\tif !ok {\n\t\treturn db.View(fn)\n\t}\n\treturn fn(tx)\n}", "func (b *Builder) ChainShow(ctx context.Context, chainID string) (spn.Chain, error) {\n\taccount, err := b.AccountInUse()\n\tif err != nil {\n\t\treturn spn.Chain{}, err\n\t}\n\treturn b.spnclient.ChainGet(ctx, account.Name, chainID)\n}", "func (t *Transaction) String() string {\n\tswitch t.Type {\n\tcase TIN, TEP:\n\t\treturn fmt.Sprintf(\"%22s {%d} %s %.3f %s [%s]\", t.Datetime, t.Index, t.Category.FullName, t.Amount, t.AmountCurrency, t.Type.Name)\n\tcase TOB:\n\t\treturn fmt.Sprintf(\"%22s {%d} Opening new Account\", t.Datetime, t.Index)\n\tcase TBJ:\n\t\treturn fmt.Sprintf(\"%22s {%d} Update balance by %.3f %s\", t.Datetime, t.Index, t.Amount, t.AmountCurrency)\n\tcase TMT:\n\t\treturn fmt.Sprintf(\"%22s {%d} Move %.3f %s to '%s' (%.3f %s)\", t.Datetime, t.Index, t.Amount, t.AmountCurrency, t.AccountTo, t.AmountTo, t.AmountToCurrency)\n\tdefault:\n\t\treturn fmt.Sprintf(\"%22s {%d} [%s] TODO: DIDN'T IMPLEMENT THIS TYPE YET\", t.Datetime, t.Index, t.Type.Name)\n\t}\n}", "func Tx(ctx context.Context, db *sqlx.DB, opts *sqltx.Options, fn TXFn) (err error) {\n\treturn sqltx.TxHandler(ctx, &sqlxDB{db}, opts, func(tx sqltx.TXer) error {\n\t\treturn fn(tx.(*sqlx.Tx))\n\t})\n}", "func (t *ATrains) Show() {\n\tfmt.Println(\"wszystkie pociagi\")\n\n\tfor i := 0; i < len(t.trains); i++ {\n\t\tt.trains[i].Show()\n\t}\n\n\tfmt.Println(\"\")\n}", "func (a Account) ShowOwner() string {\n\treturn a.owner\n}", "func (t *Transacter) WithTx(ctx context.Context) context.Context {\r\n\treturn WithTxContext(ctx)\r\n}", "func (EchoHandler) CheckTx(ctx sdk.Context, store state.SimpleDB, tx sdk.Tx) (res sdk.CheckResult, err error) {\n\tdata, err := data.ToWire(tx)\n\treturn sdk.CheckResult{Data: data}, err\n}", "func (s Transaction) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (ok OKHandler) CheckTx(ctx sdk.Context, store state.SimpleDB, tx sdk.Tx) (res sdk.CheckResult, err error) {\n\treturn sdk.CheckResult{Log: ok.Log}, nil\n}", "func NodeTx(nodeTx Tx) Option {\n\treturn func(options *options) {\n\t\toptions.nodeTx = nodeTx\n\t}\n}", "func (w *xcWallet) SendTransaction(tx []byte) ([]byte, error) {\n\tbonder, ok := w.Wallet.(asset.Broadcaster)\n\tif !ok {\n\t\treturn nil, errors.New(\"wallet is not a Broadcaster\")\n\t}\n\treturn bonder.SendTransaction(tx)\n}", "func (c *TenantController) Show(ctx *app.ShowTenantContext) error {\n\tuserToken := goajwt.ContextJWT(ctx)\n\tif userToken == nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewUnauthorizedError(\"Missing JWT token\"))\n\t}\n\n\tttoken := &auth.TenantToken{Token: userToken}\n\ttenantID := ttoken.Subject()\n\ttenant, err := c.tenantService.GetTenant(tenantID)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewNotFoundError(\"tenants\", tenantID.String()))\n\t}\n\n\tnamespaces, err := c.tenantService.GetNamespaces(tenantID)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, err)\n\t}\n\n\tresult := &app.TenantSingle{Data: convertTenant(ctx, tenant, namespaces, c.clusterService.GetCluster)}\n\treturn ctx.OK(result)\n}", "func (T transaction) String() string {\n\treturn fmt.Sprintf(\"{\\n\\t\\t\\tsender:%v,\\n\\t\\t\\treceiver:%v,\\n\\t\\t\\tamount:%v\\n\\t\\t}\", T.sender, T.receiver, T.amount)\n}", "func (c *TenantController) Show(ctx *app.ShowTenantContext) error {\n\ttoken := goajwt.ContextJWT(ctx)\n\tif token == nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewUnauthorizedError(\"Missing JWT token\"))\n\t}\n\n\tttoken := &TenantToken{token: token}\n\ttenantID := ttoken.Subject()\n\ttenant, err := c.tenantService.GetTenant(tenantID)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewNotFoundError(\"tenants\", tenantID.String()))\n\t}\n\n\tnamespaces, err := c.tenantService.GetNamespaces(tenantID)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, err)\n\t}\n\n\tresult := &app.TenantSingle{Data: convertTenant(ctx, tenant, namespaces, c.resolveCluster)}\n\treturn ctx.OK(result)\n}", "func (tbl RecordTable) IsTx() bool {\n\treturn tbl.db.IsTx()\n}", "func (me TSAFPTTransactionID) String() string { return xsdt.String(me).String() }", "func (c *TrackerController) Show(ctx *app.ShowTrackerContext) error {\n\treturn application.Transactional(c.db, func(appl application.Application) error {\n\t\tt, err := appl.Trackers().Load(ctx.Context, ctx.ID)\n\t\tif err != nil {\n\t\t\tswitch err.(type) {\n\t\t\tcase remoteworkitem.NotFoundError:\n\t\t\t\tlog.Printf(\"not found, id=%s\", ctx.ID)\n\t\t\t\treturn goa.ErrNotFound(err.Error())\n\t\t\tdefault:\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t\treturn ctx.OK(t)\n\t})\n}", "func (as *accountSet) addTx(tx *types.Transaction) {\n\tif addr, err := types.Sender(as.signer, tx); err == nil {\n\t\tas.add(addr)\n\t}\n}", "func (res BackTestResult) Show() TradestatPort {\n\t//\tp := NewPortfolio()\n\tmds := bean.NewRPCMDSConnC(\"tcp\", res.dbhost+\":\"+res.dbport)\n\tratesbook := make(ReferenceRateBook)\n\n\t// FIXME: think about how to show multi pair result\n\tvar stat TradestatPort\n\tif len(res.pairs) > 0 {\n\t\tp := res.pairs[0]\n\t\ttxn, _ := mds.GetTransactions(p, res.start, res.end)\n\t\tratesbook[p] = RefRatesFromTxn(txn)\n\t\t//\t\tsnapts.Print()\n\t\t//\t\tperfts.Print()\n\n\t\tstat = *Tradestat(p.Base, res.Txn, NewPortfolio(), ratesbook)\n\t\tstat.Print()\n\t}\n\treturn stat\n}", "func (tbl AssociationTable) IsTx() bool {\n\treturn tbl.db.IsTx()\n}", "func (as *accountSet) containsTx(tx *types.Transaction) bool {\n\tif addr, err := types.Sender(as.signer, tx); err == nil {\n\t\treturn as.contains(addr)\n\t}\n\treturn false\n}", "func (gb *CurrentGrantBuilder) Show() string {\n\treturn fmt.Sprintf(`SHOW GRANTS ON %v %v`, gb.grantType, gb.qualifiedName)\n}", "func (c Client) SendTransaction(t txn.Transaction) (string, error) {\n\tvar result string\n\terr := c.Call(&result, \"eth_sendRawTransaction\", \"0x\"+hex.EncodeToString(t.Encode()))\n\treturn result, err\n}", "func (ge *CurrentGrantExecutable) Show() string {\n\treturn fmt.Sprintf(`SHOW GRANTS OF %v \"%v\"`, ge.granteeType, ge.granteeName)\n}", "func (cc *ChainConfig) Show() {\n\tlog.Println(\"Name is \", cc.Name)\n\tlog.Println(\"Prefix is \", cc.Prefix)\n\tlog.Println(\"Suffix is \", cc.Suffix)\n\tlog.Println(\"Interval is \", cc.Interval)\n\tlog.Printf(\"Inputer is %v\\n\", cc.Chain)\n}", "func (b *DatabaseTestSuiteBase) Tx() sqlx.Ext {\n\treturn b.tx\n}", "func (a *Ethereum) Tx(hashstr string) (Tx, error) {\n\tvar (\n\t\thash = common.HexToHash(hashstr)\n\t\tctx, _ = context.WithDeadline(context.Background(), time.Now().Add(time.Second*30))\n\t\ttx, isPending, err = ethclient.NewClient(a.rpcclient).TransactionByHash(ctx, hash)\n\t)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif isPending {\n\n\t}\n\treturn EtherTx(*tx), errors.New(\"is pending\")\n}", "func (r *Repository) TxGet(tx *dbr.Tx, userID int64) (*pb.User, error) {\n\treturn r.get(tx, userID)\n}", "func (d *Driver) Tx(ctx context.Context) (dialect.Tx, error) {\n\treturn d.BeginTx(ctx, nil)\n}", "func (blockchain *Blockchain) CheckTx(req abciTypes.RequestCheckTx) abciTypes.ResponseCheckTx {\n\tresponse := blockchain.executor.RunTx(blockchain.CurrentState(), req.Tx, nil, blockchain.Height()+1, blockchain.currentMempool, blockchain.MinGasPrice(), true)\n\n\treturn abciTypes.ResponseCheckTx{\n\t\tCode: response.Code,\n\t\tData: response.Data,\n\t\tLog: response.Log,\n\t\tInfo: response.Info,\n\t\tGasWanted: response.GasWanted,\n\t\tGasUsed: response.GasUsed,\n\t\tEvents: []abciTypes.Event{\n\t\t\t{\n\t\t\t\tType: \"tags\",\n\t\t\t\tAttributes: response.Tags,\n\t\t\t},\n\t\t},\n\t}\n}", "func EncodeTx(tx *SignedTransaction) ([]byte, error) {\n\treturn proto.Marshal(tx)\n}", "func profileTx(TBTx byte, address []byte, name string) []byte {\n\ttx := types.TxProfile{\n\t\tAddress: address,\n\t\tName: name,\n\t\tAcceptedCur: viper.GetString(trcmn.FlagCur),\n\t\tDepositInfo: viper.GetString(trcmn.FlagDepositInfo),\n\t\tDueDurationDays: viper.GetInt(trcmn.FlagDueDurationDays),\n\t}\n\treturn invoicer.MarshalWithTB(tx, TBTx)\n}", "func (tx *Tx) Hex() string {\n\treturn hex.EncodeToString(tx.Bytes())\n}", "func (c *AccountController) Show(ctx *app.ShowAccountContext) error {\n\tres := &app.Account{}\n\treturn ctx.OK(res, \"default\")\n}", "func (s *Store) TxDetails(ns walletdb.ReadBucket, txHash *chainhash.Hash) (*TxDetails, error) {\n\t// First, check whether there exists an unmined transaction with this\n\t// hash. Use it if found.\n\tv := existsRawUnmined(ns, txHash[:])\n\tif v != nil {\n\t\treturn s.unminedTxDetails(ns, txHash, v)\n\t}\n\n\t// Otherwise, if there exists a mined transaction with this matching\n\t// hash, skip over to the newest and begin fetching all details.\n\tk, v := latestTxRecord(ns, txHash)\n\tif v == nil {\n\t\t// not found\n\t\treturn nil, nil\n\t}\n\treturn s.minedTxDetails(ns, txHash, k, v)\n}", "func (c *dummyWavesMDLrpcclient) GetTransaction(txid string) (*model.Transactions, error) {\n\ttransaction, _, err := client.NewTransactionsService(c.MainNET).GetTransactionsInfoID(txid)\n\treturn transaction, err\n}", "func (store *Store) TransferTx(ctx context.Context, arg TransferTxParams) (TransferTxResult, error) {\n\n\tvar result TransferTxResult\n\n\terr := store.execTx(ctx, func(q *Queries) error {\n\t\tvar err error\n\n\t\tresult.Transfer, err = q.CreateTransfer(ctx, CreateTransferParams{\n\t\t\tFromAccountID: arg.FromAccountID,\n\t\t\tToAccountID: arg.ToAccountID,\n\t\t\tAmount: arg.Amount,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tresult.FromEntry, err = q.CreateEntry(ctx, CreateEntryParams{\n\t\t\tAccountID: arg.FromAccountID,\n\t\t\tAmount: -arg.Amount,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tresult.ToEntry, err = q.CreateEntry(ctx, CreateEntryParams{\n\t\t\tAccountID: arg.ToAccountID,\n\t\t\tAmount: arg.Amount,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// TODO: update account balances\n\n\t\treturn err\n\t})\n\n\treturn result, err\n\n}", "func (tx *Transaction) TxID() string {\n\treturn tx.jsonTransaction.Hash.Hex()\n}", "func (rt *RecvTxOut) Tx() *btcutil.Tx {\n\treturn rt.tx\n}", "func (account *Account) ShowWalletIndex() {\n\tfmt.Println(\"\\033[1;31;40m\", account.WalletIndex, \"\\033[0m\")\n\n}", "func (f FailHandler) CheckTx(ctx sdk.Context, store state.SimpleDB, tx sdk.Tx) (res sdk.CheckResult, err error) {\n\treturn res, errors.Wrap(f.Err)\n}", "func (r Runner) ShowAccount(ctx context.Context, name string) (Account, error) {\n\tb := &bytes.Buffer{}\n\n\topt := []step.Option{\n\t\tr.cc.ShowKeyAddressCommand(name),\n\t}\n\n\tif r.cc.KeyringPassword != \"\" {\n\t\tinput := &bytes.Buffer{}\n\t\tfmt.Fprintln(input, r.cc.KeyringPassword)\n\t\topt = append(opt, step.Write(input.Bytes()))\n\t}\n\n\tif err := r.run(ctx, runOptions{stdout: b}, opt...); err != nil {\n\t\tif strings.Contains(err.Error(), \"item could not be found\") ||\n\t\t\tstrings.Contains(err.Error(), \"not a valid name or address\") {\n\t\t\treturn Account{}, ErrAccountDoesNotExist\n\t\t}\n\t\treturn Account{}, err\n\t}\n\n\treturn Account{\n\t\tName: name,\n\t\tAddress: strings.TrimSpace(b.String()),\n\t}, nil\n}", "func (s *Store) getDepositInfoTx(tx *bolt.Tx, Txid string) (DepositInfo, error) {\n\tvar dpi DepositInfo\n\n\tif err := dbutil.GetBucketObject(tx, DepositInfoBkt, Txid, &dpi); err != nil {\n\t\treturn DepositInfo{}, err\n\t}\n\n\treturn dpi, nil\n}", "func (c BitcoinCoreChain) RawTx(cxt context.Context, from, to, amount, memo, asset string) (string, error) {\n if configure.ChainAssets[asset] != Bitcoin {\n return \"\", fmt.Errorf(\"Unsupport %s in bitcoincore\", asset)\n }\n amountF, err := strconv.ParseFloat(amount, 64)\n if err != nil {\n return \"\", err\n }\n txAmountSatoshi, err := btcutil.NewAmount(amountF)\n if err != nil {\n return \"\", err\n }\n\n fromPkScript, err := BitcoincoreAddressP2AS(from, c.Mode)\n if err != nil {\n return \"\", err\n }\n toPkScript, err := BitcoincoreAddressP2AS(to, c.Mode)\n if err != nil {\n return \"\", err\n }\n\n // query bitcoin chain info\n chaininfo, err := c.Client.GetBlockChainInfo()\n if err != nil {\n return \"\", err\n }\n // feeKB, err := c.Client.EstimateFee(int64(6))\n feeKB, err := c.Client.EstimateSmartFee(int64(6))\n if err != nil {\n return \"\", err\n }\n feeRate := mempool.SatoshiPerByte(feeKB.FeeRate)\n\n if feeKB.FeeRate <= 0 {\n feeRate = mempool.SatoshiPerByte(100)\n }\n\n var (\n selectedutxos, unselectedutxos []db.UTXO\n selectedCoins coinset.Coins\n )\n\n // Coin Select\n if strings.ToLower(configure.ChainsInfo[Bitcoin].Coin) == strings.ToLower(asset) {\n // select coins for BTC transfer\n if selectedutxos, unselectedutxos, selectedCoins, err = CoinSelect(int64(chaininfo.Headers), txAmountSatoshi, c.Wallet.Address.UTXOs); err != nil {\n return \"\", fmt.Errorf(\"Select UTXO for tx %s\", err)\n }\n }else {\n // select coins for Token transfer\n // 300: https://bitcoin.stackexchange.com/questions/1195/how-to-calculate-transaction-size-before-sending-legacy-non-segwit-p2pkh-p2sh\n inputAmount := feeRate.Fee(uint32(300))\n if selectedutxos, unselectedutxos, selectedCoins, err = CoinSelect(int64(chaininfo.Headers), inputAmount, c.Wallet.Address.UTXOs); err != nil {\n return \"\", fmt.Errorf(\"Select UTXO for tx %s\", err)\n }\n }\n\n var vinAmount int64\n for _, coin := range selectedCoins.Coins() {\n vinAmount += int64(coin.Value())\n }\n msgTx := coinset.NewMsgTxWithInputCoins(wire.TxVersion, selectedCoins)\n\n token := configure.ChainsInfo[Bitcoin].Tokens[strings.ToLower(asset)]\n if token != \"\" && strings.ToLower(asset) != strings.ToLower(configure.ChainsInfo[Bitcoin].Coin) {\n // OmniToken transfer\n b := txscript.NewScriptBuilder()\n b.AddOp(txscript.OP_RETURN)\n\n omniVersion := util.Int2byte(uint64(0), 2)\t// omnicore version\n txType := util.Int2byte(uint64(0), 2)\t// omnicore tx type: simple send\n propertyID := configure.ChainsInfo[Bitcoin].Tokens[asset]\n tokenPropertyid, err := strconv.Atoi(propertyID)\n if err != nil {\n return \"\", fmt.Errorf(\"tokenPropertyid to int %s\", err)\n }\n // tokenPropertyid := configure.Config.OmniToken[\"omni_first_token\"].(int)\n tokenIdentifier := util.Int2byte(uint64(tokenPropertyid), 4)\t// omni token identifier\n tokenAmount := util.Int2byte(uint64(txAmountSatoshi), 8)\t// omni token transfer amount\n\n b.AddData([]byte(\"omni\"))\t// transaction maker\n b.AddData(omniVersion)\n b.AddData(txType)\n b.AddData(tokenIdentifier)\n b.AddData(tokenAmount)\n pkScript, err := b.Script()\n if err != nil {\n return \"\", fmt.Errorf(\"Bitcoin Token pkScript %s\", err)\n }\n msgTx.AddTxOut(wire.NewTxOut(0, pkScript))\n txOutReference := wire.NewTxOut(0, toPkScript)\n msgTx.AddTxOut(txOutReference)\n }else {\n // BTC transfer\n txOutTo := wire.NewTxOut(int64(txAmountSatoshi), toPkScript)\n msgTx.AddTxOut(txOutTo)\n\n // recharge\n // 181, 34: https://bitcoin.stackexchange.com/questions/1195/how-to-calculate-transaction-size-before-sending-legacy-non-segwit-p2pkh-p2sh\n fee := feeRate.Fee(uint32(msgTx.SerializeSize() + 181 + 34))\n if (vinAmount - int64(txAmountSatoshi) - int64(fee)) > 0 {\n txOutReCharge := wire.NewTxOut((vinAmount-int64(txAmountSatoshi) - int64(fee)), fromPkScript)\n msgTx.AddTxOut(txOutReCharge)\n }else {\n selectedutxoForFee, _, selectedCoinsForFee, err := CoinSelect(int64(chaininfo.Headers), fee, unselectedutxos)\n if err != nil {\n return \"\", fmt.Errorf(\"Select UTXO for fee %s\", err)\n }\n for _, coin := range selectedCoinsForFee.Coins() {\n vinAmount += int64(coin.Value())\n }\n txOutReCharge := wire.NewTxOut((vinAmount - int64(txAmountSatoshi) - int64(fee)), fromPkScript)\n msgTx.AddTxOut(txOutReCharge)\n selectedutxos = append(selectedutxos, selectedutxoForFee...)\n }\n }\n\n buf := bytes.NewBuffer(make([]byte, 0, msgTx.SerializeSize()))\n msgTx.Serialize(buf)\n rawTxHex := hex.EncodeToString(buf.Bytes())\n c.Wallet.SelectedUTXO = selectedutxos\n return rawTxHex, nil\n}", "func (c *swapCoin) TxID() string {\n\treturn c.txid\n}", "func (ec *Client) SendTransaction(ctx context.Context, tx *types.Transaction) error {\n\tdata, err := rlp.EncodeToBytes(tx)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn ec.c.CallContext(ctx, nil, \"siot_sendRawTransaction\", helper.ToHex(data))\n}", "func (tbl AssociationTable) Tx() sqlapi.SqlTx {\n\treturn tbl.db.(sqlapi.SqlTx)\n}", "func (c *Client) SendTransaction(ctx context.Context, tx types.Transaction) (string, error) {\n\trawTx, err := tx.Serialize()\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"failed to serialize tx, err: %v\", err)\n\t}\n\tres, err := c.RpcClient.SendTransactionWithConfig(\n\t\tctx,\n\t\tbase64.StdEncoding.EncodeToString(rawTx),\n\t\trpc.SendTransactionConfig{\n\t\t\tEncoding: rpc.SendTransactionConfigEncodingBase64,\n\t\t},\n\t)\n\terr = checkRpcResult(res.GeneralResponse, err)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn res.Result, nil\n}", "func (res BackTestResult) Show() TradestatPort {\n\t//\tp := NewPortfolio()\n\t// mds := bean.NewRPCMDSConnC(\"tcp\", res.dbhost+\":\"+res.dbport)\n\tratesbook := make(ReferenceRateBook)\n\n\t// FIXME: think about how to show multi pair result\n\tvar stat TradestatPort\n\tif len(res.pairs) > 0 {\n\t\tp := res.pairs[0]\n\t\t//\t\ttxn, _ := mds.GetTransactions2(NameFcoin, p, res.start, res.end)\n\t\tratesbook[p] = RefRatesFromTxn(res.Txn)\n\t\t//\t\tsnapts.Print()\n\t\t//\t\tperfts.Print()\n\n\t\tstat = *Tradestat(p.Base, res.Txn, NewPortfolio(), ratesbook)\n\t\tssTS := GenerateSnapshotTS(res.Txn, NewPortfolio())\n\t\tmaxPos := 0.0\n\t\tmaxNeg := 0.0\n\t\tfor _, ss := range ssTS {\n\t\t\tmaxPos = math.Max(maxPos, ss.Port.Balance(p.Coin))\n\t\t\tmaxNeg = math.Min(maxNeg, ss.Port.Balance(p.Coin))\n\t\t}\n\t\tfmt.Println(\"max position\", maxPos, maxNeg)\n\t\tstat.Print()\n\t}\n\treturn stat\n}", "func (s *Store) GetTx(txid common.Hash) *types.Transaction {\n\ttx, _ := s.rlp.Get(s.table.Txs, txid.Bytes(), &types.Transaction{}).(*types.Transaction)\n\n\treturn tx\n}", "func (app *BftApplication) CheckTx(tx []byte) types.ResponseCheckTx {\n\treturn types.ResponseCheckTx{Code: code.CodeTypeOK}\n\t//if cpcash.validate(\"BFTXafe2242d45cc5e54041b2b52913ef9a1aede4998a32e3fee128cf7d1e7575a41\") {\n\t//\treturn types.ResponseCheckTx{Code: code.CodeTypeOK}\n\t//}\n\t//return types.ResponseCheckTx{Code: code.NotPaid}\n\n}", "func (tx *CreateLegalEntityTx) TxType() byte {\n\treturn TxTypeCreateLegalEntity\n}", "func (c *Client) Tx(ctx context.Context) (*Tx, error) {\n\tif _, ok := c.driver.(*txDriver); ok {\n\t\treturn nil, fmt.Errorf(\"ent: cannot start a transaction within a transaction\")\n\t}\n\ttx, err := newTx(ctx, c.driver)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"ent: starting a transaction: %v\", err)\n\t}\n\tcfg := config{driver: tx, log: c.log, debug: c.debug, hooks: c.hooks}\n\treturn &Tx{\n\t\tctx: ctx,\n\t\tconfig: cfg,\n\t\tDNSBLQuery: NewDNSBLQueryClient(cfg),\n\t\tDNSBLResponse: NewDNSBLResponseClient(cfg),\n\t\tIP: NewIPClient(cfg),\n\t\tOperation: NewOperationClient(cfg),\n\t\tUser: NewUserClient(cfg),\n\t}, nil\n}", "func (c *SyscallService) QueryTx(ctx context.Context, in *pb.QueryTxRequest) (*pb.QueryTxResponse, error) {\n\tnctx, ok := c.ctxmgr.Context(in.GetHeader().Ctxid)\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"bad ctx id:%d\", in.Header.Ctxid)\n\t}\n\n\trawTxid, err := hex.DecodeString(in.Txid)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttx, confirmed, err := nctx.Cache.QueryTx(rawTxid)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif !confirmed {\n\t\treturn nil, fmt.Errorf(\"Unconfirm tx:%s\", in.Txid)\n\t}\n\n\ttxsdk := ConvertTxToSDKTx(tx)\n\n\treturn &pb.QueryTxResponse{\n\t\tTx: txsdk,\n\t}, nil\n}", "func (s *Session) TransactionTx(f func(*Session) (interface{}, error), opts *sql.TxOptions) (interface{}, error) {\n\terr := s.BeginTx(opts)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\td, err := f(s)\n\tif err != nil {\n\t\ts.RollBack()\n\t} else {\n\t\ts.Commit()\n\t}\n\treturn d, err\n}", "func (c *Client) Tx(ctx context.Context) (*Tx, error) {\n\tif _, ok := c.driver.(*txDriver); ok {\n\t\treturn nil, fmt.Errorf(\"ent: cannot start a transaction within a transaction\")\n\t}\n\ttx, err := newTx(ctx, c.driver)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"ent: starting a transaction: %v\", err)\n\t}\n\tcfg := config{driver: tx, log: c.log, debug: c.debug, hooks: c.hooks}\n\treturn &Tx{\n\t\tctx: ctx,\n\t\tconfig: cfg,\n\t\tPatient: NewPatientClient(cfg),\n\t\tPatientofphysician: NewPatientofphysicianClient(cfg),\n\t\tPatientroom: NewPatientroomClient(cfg),\n\t\tPhysician: NewPhysicianClient(cfg),\n\t}, nil\n}", "func (s *signer) SignTx(magic netmode.Magic, tx *transaction.Transaction) error {\n\treturn (*wallet.Account)(s).SignTx(magic, tx)\n}", "func (app *JSONStoreApplication) CheckTx(tx types.RequestCheckTx) types.ResponseCheckTx {\n\n\tvar temp interface{}\n\terr := json.Unmarshal(tx.Tx, &temp)\n\n\tif err != nil {\n\t\treturn types.ResponseCheckTx{Code: code.CodeTypeEncodingError,Log: fmt.Sprint(err)}\n\t}\n\n\tmessage := temp.(map[string]interface{})\n\tPublicKey:= message[\"publicKey\"].(string)\n\tByteString:= message[\"msg\"].(string)\n\tSignaTure:= message[\"sig\"].(string) \n\tcount := checkUserPublic(db,PublicKey)\n\n if signatureValidate(PublicKey,SignaTure,ByteString) && count !=0 {\n return types.ResponseCheckTx{Code: code.CodeTypeOK}\n}\n return types.ResponseCheckTx{Code: code.CodeTypeBadData}\n}", "func SprintTransaction(tx *types.Transaction) (string, error) {\n\ttxbody, err := tx.MarshalJSON()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tvar txjson bytes.Buffer\n\tif err := json.Indent(&txjson, txbody, \"\", \"\\t\"); err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn string(txjson.Bytes()), nil\n}", "func (store *SQLStore) TransferTx(ctx context.Context, arg TransferTxParams) (TransferTxResult, error) {\n\t//create empty result \n\tvar result TransferTxResult\n\n\terr := store.execTx(ctx, func(q *Queries) error {\n\t\tvar err error \n\t\t\n\n\t\tresult.Transfer, err = q.CreateTransfer(ctx, CreateTransferParams{\n\t\t\tFromAccountID: arg.FromAccountID,\n\t\t\tToAccountID: arg.ToAccountID,\n\t\t\tAmount: arg.Amount,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t\n\n\t\t// moving money out of account\n\t\tresult.FromEntry, err = q.CreateEntry(ctx, CreateEntryParams{\n\t\t\tAccountID: arg.FromAccountID,\n\t\t\tAmount: -arg.Amount,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\n\t\t// moving money into an account\n\t\tresult.ToEntry, err = q.CreateEntry(ctx, CreateEntryParams{\n\t\t\tAccountID: arg.ToAccountID,\n\t\t\tAmount: arg.Amount,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t\n\t\t// make sure we always execute transactions in a standard order \n\t\tif arg.FromAccountID < arg.ToAccountID {\n\t\t\tresult.FromAccount, result.ToAccount, err = addMoney(ctx, q, arg.FromAccountID, -arg.Amount, arg.ToAccountID, arg.Amount)\n\t\t} else {\n\t\t\tresult.ToAccount, result.FromAccount, err = addMoney(ctx, q, arg.ToAccountID, arg.Amount, arg.FromAccountID, -arg.Amount)\n\n\t}\n\n\t\treturn nil\n\t})\n\n\treturn result, err\n}", "func (c *Client) Tx(ctx context.Context) (*Tx, error) {\n\tif _, ok := c.driver.(*txDriver); ok {\n\t\treturn nil, fmt.Errorf(\"ent: cannot start a transaction within a transaction\")\n\t}\n\ttx, err := newTx(ctx, c.driver)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"ent: starting a transaction: %w\", err)\n\t}\n\tcfg := c.config\n\tcfg.driver = tx\n\treturn &Tx{\n\t\tctx: ctx,\n\t\tconfig: cfg,\n\t\tUser: NewUserClient(cfg),\n\t\tUserCard: NewUserCardClient(cfg),\n\t\tUserWallet: NewUserWalletClient(cfg),\n\t}, nil\n}", "func (tx *Tx) Tx(pkt *mbuf.Mbuf) int {\n\treturn int(C.go_wr_tx((*C.struct_tx_port)(tx),\n\t\t(*C.struct_rte_mbuf)(unsafe.Pointer(pkt))))\n}", "func (db *DB) GetTx() *GetTx {\n\treturn &GetTx{\n\t\tdb: db,\n\t}\n}", "func (o *APIServiceSendTransactionParams) WithTx(tx string) *APIServiceSendTransactionParams {\n\to.SetTx(tx)\n\treturn o\n}", "func (me TxsdShow) ToXsdtString() xsdt.String { return xsdt.String(me) }", "func (me TxsdShow) ToXsdtString() xsdt.String { return xsdt.String(me) }", "func displayAccount(w io.Writer) {\n\taccount := doAccount()\n\tif account.Status != \"active\" {\n\t\tlog.Errorf(\"DO Account issue expected status (active) got (%s) \", account.Status)\n\t\treturn\n\t}\n\tfmt.Fprintf(w, account.String())\n}", "func GetTx(txhash string) (*model.Tx, error) {\n\turl := fmt.Sprintf(bchapi.TxUrl, txhash)\n\tresult, err := bchapi.HttpGet(url, bchapi.ConnTimeoutMS, bchapi.ServeTimeoutMS)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ttx, err := model.StringToTx(result)\n\treturn tx, err\n}", "func (e Endpoints) SendETHTx(ctx context.Context, signedTxHex string) (txID string, err error) {\n\trequest := SendETHTxRequest{SignedTxHex: signedTxHex}\n\tresponse, err := e.SendETHTxEndpoint(ctx, request)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn response.(SendETHTxResponse).TxID, nil\n}", "func (c *Client) Tx(ctx context.Context) (*Tx, error) {\n\tif _, ok := c.driver.(*txDriver); ok {\n\t\treturn nil, fmt.Errorf(\"ent: cannot start a transaction within a transaction\")\n\t}\n\ttx, err := newTx(ctx, c.driver)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"ent: starting a transaction: %v\", err)\n\t}\n\tcfg := config{driver: tx, log: c.log, debug: c.debug}\n\treturn &Tx{\n\t\tconfig: cfg,\n\t\tModule: NewModuleClient(cfg),\n\t\tModuleVersion: NewModuleVersionClient(cfg),\n\t}, nil\n}" ]
[ "0.7389181", "0.6147807", "0.60307246", "0.60012263", "0.59897405", "0.5976033", "0.5919237", "0.57918423", "0.5711443", "0.5654345", "0.5603889", "0.55986404", "0.5598449", "0.5588051", "0.5583727", "0.5568362", "0.55514675", "0.55491924", "0.5546502", "0.5537545", "0.5520107", "0.5514006", "0.5514006", "0.548742", "0.5464789", "0.5462737", "0.5462653", "0.5461838", "0.5439106", "0.54335946", "0.5390721", "0.53899944", "0.53854996", "0.53821176", "0.5373848", "0.5362792", "0.5333694", "0.53256166", "0.5297932", "0.52731276", "0.52490896", "0.52484417", "0.52428913", "0.5229313", "0.5228682", "0.5224055", "0.5205882", "0.51992834", "0.518638", "0.5183744", "0.5160607", "0.51552457", "0.5152202", "0.5142116", "0.5141962", "0.51397145", "0.5135687", "0.5126441", "0.5100012", "0.50992405", "0.50973433", "0.50842834", "0.50807345", "0.5078648", "0.5076424", "0.5073661", "0.5070471", "0.50683516", "0.50346357", "0.50332886", "0.50314707", "0.5029291", "0.5028215", "0.5020823", "0.5018584", "0.50185615", "0.501738", "0.5013903", "0.50106317", "0.5009323", "0.5008196", "0.50075287", "0.4995199", "0.49946406", "0.4993971", "0.4988983", "0.49872118", "0.49841848", "0.49812612", "0.49807772", "0.49807325", "0.49779215", "0.49750525", "0.49736673", "0.49723735", "0.49723735", "0.49655643", "0.4965486", "0.49612835", "0.4960393" ]
0.73721284
1
ReadResponse reads a server response into the received o.
func (o *GetComputeRackUnitsMoidReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) { switch response.Code() { case 200: result := NewGetComputeRackUnitsMoidOK() if err := result.readResponse(response, consumer, o.formats); err != nil { return nil, err } return result, nil case 404: result := NewGetComputeRackUnitsMoidNotFound() if err := result.readResponse(response, consumer, o.formats); err != nil { return nil, err } return nil, result default: result := NewGetComputeRackUnitsMoidDefault(response.Code()) if err := result.readResponse(response, consumer, o.formats); err != nil { return nil, err } if response.Code()/100 == 2 { return result, nil } return nil, result } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (d *ResourceHandler) ReadResponse(dataOut unsafe.Pointer, bytesToRead int32, bytesRead *int32, callback *Callback) int32 {\n\treturn lookupResourceHandlerProxy(d.Base()).ReadResponse(d, dataOut, bytesToRead, bytesRead, callback)\n}", "func (o *GetServerReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetServerOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *InteractionBindReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewInteractionBindOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 404:\n\t\tresult := NewInteractionBindNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 500:\n\t\tresult := NewInteractionBindInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *InteractionUnbindReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewInteractionUnbindOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 404:\n\t\tresult := NewInteractionUnbindNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 500:\n\t\tresult := NewInteractionUnbindInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (r *ResponseReader) ReadResponse(req *Request) (res *Response, err error) {\n\tres = CreateEmptyResponse(req)\n\t_, err = readFirstLine(r, res)\n\tif err != nil {\n\t\treturn\n\t}\n\n\terr = readHeaders(r, res)\n\tif err != nil {\n\t\treturn\n\t}\n\n\t_, err = readBodyContent(r, res)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn res, nil\n}", "func (c *Conn) ReadResponse(rmsg *Response) error {\n\tdata, err := c.ReadDataUnit()\n\tif err != nil {\n\t\treturn err\n\t}\n\tcolor.Printf(\"@{c}<!-- RESPONSE -->\\n%s\\n\\n\", string(data))\n\terr = xml.Unmarshal(data, rmsg)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// color.Fprintf(os.Stderr, \"@{y}%s\\n\", spew.Sprintf(\"%+v\", msg))\n\tif len(rmsg.Results) != 0 {\n\t\tr := rmsg.Results[0]\n\t\tif r.IsError() {\n\t\t\treturn r\n\t\t}\n\t}\n\treturn nil\n}", "func (o *VerifyConnectionReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewVerifyConnectionOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *GetAvailableReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewGetAvailableOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *ClosePositionReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewClosePositionOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 400:\n\t\tresult := NewClosePositionBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 401:\n\t\tresult := NewClosePositionUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 404:\n\t\tresult := NewClosePositionNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 405:\n\t\tresult := NewClosePositionMethodNotAllowed()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *DescribeServerReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewDescribeServerOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewDescribeServerBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 404:\n\t\tresult := NewDescribeServerNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 409:\n\t\tresult := NewDescribeServerConflict()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 500:\n\t\tresult := NewDescribeServerInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *GetServerSessionReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetServerSessionOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewGetServerSessionBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 401:\n\t\tresult := NewGetServerSessionUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewGetServerSessionNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 500:\n\t\tresult := NewGetServerSessionInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\tdata, err := ioutil.ReadAll(response.Body())\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\treturn nil, fmt.Errorf(\"Requested GET /dsmcontroller/namespaces/{namespace}/servers/{podName}/session returns an error %d: %s\", response.Code(), string(data))\n\t}\n}", "func (o *StartReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewStartOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (resp *PharosResponse) readResponse() {\n\tif !resp.hasBeenRead && resp.Response != nil && resp.Response.Body != nil {\n\t\tresp.data, resp.Error = ioutil.ReadAll(resp.Response.Body)\n\t\tresp.Response.Body.Close()\n\t\tresp.hasBeenRead = true\n\t}\n}", "func (o *HelloWorldReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewHelloWorldOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewHelloWorldBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 500:\n\t\tresult := NewHelloWorldInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (reader *BasicRpcReader) ReadResponse(r io.Reader, method string, requestID int32, resp proto.Message) error {\n\trrh := &hadoop.RpcResponseHeaderProto{}\n\terr := readRPCPacket(r, rrh, resp)\n\tif err != nil {\n\t\treturn err\n\t} else if int32(rrh.GetCallId()) != requestID {\n\t\treturn errors.New(\"unexpected sequence number\")\n\t} else if rrh.GetStatus() != hadoop.RpcResponseHeaderProto_SUCCESS {\n\t\treturn &NamenodeError{\n\t\t\tmethod: method,\n\t\t\tmessage: rrh.GetErrorMsg(),\n\t\t\tcode: int(rrh.GetErrorDetail()),\n\t\t\texception: rrh.GetExceptionClassName(),\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o *UpdateAntivirusServerReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 204:\n\t\tresult := NewUpdateAntivirusServerNoContent()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\tresult := NewUpdateAntivirusServerDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *HasEventsReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewHasEventsOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 401:\n\t\tresult := NewHasEventsUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 403:\n\t\tresult := NewHasEventsForbidden()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *GetV2Reader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetV2OK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 500:\n\t\tresult := NewGetV2InternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *SaveReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 204:\n\t\tresult := NewSaveNoContent()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 500:\n\t\tresult := NewSaveInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *TestWriteReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewTestWriteOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 401:\n\t\tresult := NewTestWriteUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *AllConnectionsReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewAllConnectionsOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewAllConnectionsBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 404:\n\t\tresult := NewAllConnectionsNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *SendDataToDeviceReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewSendDataToDeviceOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewSendDataToDeviceBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 500:\n\t\tresult := NewSendDataToDeviceInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *HealthNoopReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewHealthNoopOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *PutOutOfRotationReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 204:\n\t\tresult := NewPutOutOfRotationNoContent()\n\t\tresult.HttpResponse = response\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\terrorResult := kbcommon.NewKillbillError(response.Code())\n\t\tif err := consumer.Consume(response.Body(), &errorResult); err != nil && err != io.EOF {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, errorResult\n\t}\n}", "func (o *GetReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *StatusReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewStatusOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 401:\n\t\tresult := NewStatusUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 403:\n\t\tresult := NewStatusForbidden()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *ReplaceServerReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewReplaceServerOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 202:\n\t\tresult := NewReplaceServerAccepted()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewReplaceServerBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 404:\n\t\tresult := NewReplaceServerNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\tresult := NewReplaceServerDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func ReadResponse(r *bfe_bufio.Reader, req *Request) (*Response, error) {\n\ttp := textproto.NewReader(r)\n\tresp := &Response{\n\t\tRequest: req,\n\t}\n\n\t// Parse the first line of the response.\n\tline, err := tp.ReadLine()\n\tif err != nil {\n\t\tif err == io.EOF {\n\t\t\terr = io.ErrUnexpectedEOF\n\t\t}\n\t\treturn nil, err\n\t}\n\tf := strings.SplitN(line, \" \", 3)\n\tif len(f) < 2 {\n\t\treturn nil, &badStringError{\"malformed HTTP response\", line}\n\t}\n\treasonPhrase := \"\"\n\tif len(f) > 2 {\n\t\treasonPhrase = f[2]\n\t}\n\tresp.Status = f[1] + \" \" + reasonPhrase\n\tresp.StatusCode, err = strconv.Atoi(f[1])\n\tif err != nil {\n\t\treturn nil, &badStringError{\"malformed HTTP status code\", f[1]}\n\t}\n\n\tresp.Proto = f[0]\n\tvar ok bool\n\tif resp.ProtoMajor, resp.ProtoMinor, ok = ParseHTTPVersion(resp.Proto); !ok {\n\t\treturn nil, &badStringError{\"malformed HTTP version\", resp.Proto}\n\t}\n\n\t// Parse the response headers.\n\tmimeHeader, err := tp.ReadMIMEHeader()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresp.Header = Header(mimeHeader)\n\n\tfixPragmaCacheControl(resp.Header)\n\n\terr = readTransfer(resp, r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn resp, nil\n}", "func (o *PostChatroomsChannelHashReadReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewPostChatroomsChannelHashReadOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 403:\n\t\tresult := NewPostChatroomsChannelHashReadForbidden()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *TogglePacketGeneratorsReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 201:\n\t\tresult := NewTogglePacketGeneratorsCreated()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *FrontPutBinaryReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewFrontPutBinaryOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *SystemPingReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewSystemPingOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 500:\n\t\tresult := NewSystemPingInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *SendDummyAlertReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewSendDummyAlertOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 400:\n\t\tresult := NewSendDummyAlertBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 404:\n\t\tresult := NewSendDummyAlertNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *GetViewsConnectionReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetViewsConnectionOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewGetViewsConnectionBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *SyncCopyReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewSyncCopyOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewSyncCopyDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *PostPatientsReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewPostPatientsOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 400:\n\t\tresult := NewPostPatientsBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 405:\n\t\tresult := NewPostPatientsMethodNotAllowed()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (c *Conn) readResponse(res *response_) error {\n\terr := c.readDataUnit()\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = IgnoreEOF(scanResponse.Scan(c.decoder, res))\n\tif err != nil {\n\t\treturn err\n\t}\n\tif res.Result.IsError() {\n\t\treturn res.Result\n\t}\n\treturn nil\n}", "func (o *AllConnectionsReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n switch response.Code() {\n \n case 200:\n result := NewAllConnectionsOK()\n if err := result.readResponse(response, consumer, o.formats); err != nil {\n return nil, err\n }\n return result, nil\n \n case 400:\n result := NewAllConnectionsBadRequest()\n if err := result.readResponse(response, consumer, o.formats); err != nil {\n return nil, err\n }\n return nil, result\n \n case 404:\n result := NewAllConnectionsNotFound()\n if err := result.readResponse(response, consumer, o.formats); err != nil {\n return nil, err\n }\n return nil, result\n \n default:\n return nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n }\n}", "func (o *GetMsgVpnReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewGetMsgVpnOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\tresult := NewGetMsgVpnDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (r *Response) Read(p []byte) (n int, err error) {\n\n\tif r.Error != nil {\n\t\treturn -1, r.Error\n\t}\n\n\treturn r.RawResponse.Body.Read(p)\n}", "func (o *PostPciLinksMoidReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 201:\n\t\tresult := NewPostPciLinksMoidCreated()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewPostPciLinksMoidDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *THSRAPIODFare2121Reader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewTHSRAPIODFare2121OK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 299:\n\t\tresult := NewTHSRAPIODFare2121Status299()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 304:\n\t\tresult := NewTHSRAPIODFare2121NotModified()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *PostGatewayConnectNetaddressReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 204:\n\t\tresult := NewPostGatewayConnectNetaddressNoContent()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewPostGatewayConnectNetaddressDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *DNSGetReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewDNSGetOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewDNSGetDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *GetGreetStatusReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetGreetStatusOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *PostAPIV2EventsReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 201:\n\t\tresult := NewPostAPIV2EventsNoContent()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewPostAPIV2EventsBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 403:\n\t\tresult := NewPostAPIV2EventsForbidden()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *CreateAntivirusServerReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewCreateAntivirusServerOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\tresult := NewCreateAntivirusServerDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *PostCarsReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewPostCarsOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 405:\n\t\tresult := NewPostCarsMethodNotAllowed()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *LogReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewLogOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewLogNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *ChatGetConnectedReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewChatGetConnectedOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 400:\n\t\tresult := NewChatGetConnectedBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 401:\n\t\tresult := NewChatGetConnectedUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 404:\n\t\tresult := NewChatGetConnectedNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *WebModifyReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewWebModifyOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 202:\n\t\tresult := NewWebModifyAccepted()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewWebModifyDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *GetHyperflexServerModelsMoidReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetHyperflexServerModelsMoidOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewGetHyperflexServerModelsMoidNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\tresult := NewGetHyperflexServerModelsMoidDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *KillQueryReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 204:\n\t\tresult := NewKillQueryNoContent()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewKillQueryBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 404:\n\t\tresult := NewKillQueryNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 422:\n\t\tresult := NewKillQueryUnprocessableEntity()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *GetProgressionViewReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetProgressionViewOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewGetProgressionViewBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *UpdateRackTopoReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewUpdateRackTopoOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewUpdateRackTopoBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *UtilTestReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewUtilTestOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *GetByUIDReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetByUIDOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewGetByUIDNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *GetMeReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetMeOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewGetMeDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *Delete1Reader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 204:\n\t\tresult := NewDelete1NoContent()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewDelete1NotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *RevokeReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewRevokeOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 401:\n\t\tresult := NewRevokeUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 404:\n\t\tresult := NewRevokeNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *PostGatewayDisconnectNetaddressReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 204:\n\t\tresult := NewPostGatewayDisconnectNetaddressNoContent()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewPostGatewayDisconnectNetaddressDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *GetProtocolsUsingGETReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetProtocolsUsingGETOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *DestroySessionUsingPOSTReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewDestroySessionUsingPOSTOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *CompleteTransactionReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 204:\n\t\tresult := NewCompleteTransactionNoContent()\n\t\tresult.HttpResponse = response\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\terrorResult := kbcommon.NewKillbillError(response.Code())\n\t\tif err := consumer.Consume(response.Body(), &errorResult); err != nil && err != io.EOF {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, errorResult\n\t}\n}", "func (o *GetMapNameEventsReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetMapNameEventsOK(o.writer)\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewGetMapNameEventsNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *RecoveryReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewRecoveryOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 500:\n\t\tresult := NewRecoveryInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *GetPeersReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewGetPeersOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 403:\n\t\tresult := NewGetPeersForbidden()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *InstallEventsReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewInstallEventsOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *SetMemoRequiredReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewSetMemoRequiredOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewSetMemoRequiredBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 500:\n\t\tresult := NewSetMemoRequiredInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *UpdateRackTopoReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewUpdateRackTopoOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewUpdateRackTopoBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 404:\n\t\tresult := NewUpdateRackTopoNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 500:\n\t\tresult := NewUpdateRackTopoInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *GetVoicesReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetVoicesOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *PatchHyperflexServerModelsMoidReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 201:\n\t\tresult := NewPatchHyperflexServerModelsMoidCreated()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewPatchHyperflexServerModelsMoidDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *BounceReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tresult := NewBounceDefault(response.Code())\n\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\treturn nil, err\n\t}\n\tif response.Code()/100 == 2 {\n\t\treturn result, nil\n\t}\n\treturn nil, result\n}", "func (o *PostHyperflexHxdpVersionsMoidReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 201:\n\t\tresult := NewPostHyperflexHxdpVersionsMoidCreated()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewPostHyperflexHxdpVersionsMoidDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *GetObmsLibraryIdentifierReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewGetObmsLibraryIdentifierOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 404:\n\t\tresult := NewGetObmsLibraryIdentifierNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\tresult := NewGetObmsLibraryIdentifierDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *DeleteApplianceRestoresMoidReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewDeleteApplianceRestoresMoidOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewDeleteApplianceRestoresMoidNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\tresult := NewDeleteApplianceRestoresMoidDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *UserQuerySessionReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewUserQuerySessionOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewUserQuerySessionBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 500:\n\t\tresult := NewUserQuerySessionInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\tdata, err := ioutil.ReadAll(response.Body())\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\treturn nil, fmt.Errorf(\"Requested GET /sessionbrowser/namespaces/{namespace}/gamesession returns an error %d: %s\", response.Code(), string(data))\n\t}\n}", "func (o *GetDiscoverReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewGetDiscoverOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (r *overwriteConsumerReader) ReadResponse(resp runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tif r.forStatusCode == ForAllStatusCodes || resp.Code() == r.forStatusCode {\n\t\treturn r.requestReader.ReadResponse(resp, r.consumer)\n\t}\n\n\treturn r.requestReader.ReadResponse(resp, consumer)\n}", "func (o *UnclaimTrafficFilterLinkIDReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewUnclaimTrafficFilterLinkIDOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewUnclaimTrafficFilterLinkIDBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 500:\n\t\tresult := NewUnclaimTrafficFilterLinkIDInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *GetDebugRequestReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewGetDebugRequestOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 404:\n\t\tresult := NewGetDebugRequestNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *ChangeaspecificSpeedDialReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 204:\n\t\tresult := NewChangeaspecificSpeedDialNoContent()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *PostMemoryArraysMoidReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 201:\n\t\tresult := NewPostMemoryArraysMoidCreated()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewPostMemoryArraysMoidDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (c *Client) readResponse(conn net.Conn) ([]byte, error) {\n\tif c.Timeout > 0 {\n\t\t_ = conn.SetReadDeadline(time.Now().Add(c.Timeout))\n\t}\n\n\tproto := \"udp\"\n\tif _, ok := conn.(*net.TCPConn); ok {\n\t\tproto = \"tcp\"\n\t}\n\n\tif proto == \"udp\" {\n\t\tbufSize := c.UDPSize\n\t\tif bufSize == 0 {\n\t\t\tbufSize = dns.MinMsgSize\n\t\t}\n\t\tresponse := make([]byte, bufSize)\n\t\tn, err := conn.Read(response)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn response[:n], nil\n\t}\n\n\t// If we got here, this is a TCP connection\n\t// so we should read a 2-byte prefix first\n\treturn readPrefixed(conn)\n}", "func (o *PayReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewPayOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewPayBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewPayNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 409:\n\t\tresult := NewPayConflict()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\tdata, err := ioutil.ReadAll(response.Body())\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\treturn nil, fmt.Errorf(\"Requested POST /platform/public/namespaces/{namespace}/payment/orders/{paymentOrderNo}/pay returns an error %d: %s\", response.Code(), string(data))\n\t}\n}", "func (o *CountReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewCountOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 400:\n\t\tresult := NewCountBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *PostNodesIdentifierObmIdentifyReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 201:\n\t\tresult := NewPostNodesIdentifierObmIdentifyCreated()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 404:\n\t\tresult := NewPostNodesIdentifierObmIdentifyNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\tresult := NewPostNodesIdentifierObmIdentifyDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *GetInterpreterReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetInterpreterOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewGetInterpreterNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *DeleteEventsEventIDReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 204:\n\t\tresult := NewDeleteEventsEventIDNoContent()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 401:\n\t\tresult := NewDeleteEventsEventIDUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 404:\n\t\tresult := NewDeleteEventsEventIDNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *UtilityServiceReadyReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewUtilityServiceReadyOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewUtilityServiceReadyDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *HTTPGetPersistenceItemDataReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewHTTPGetPersistenceItemDataOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewHTTPGetPersistenceItemDataNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *SubscriptionReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewSubscriptionOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *PostEquipmentIoExpandersMoidReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 201:\n\t\tresult := NewPostEquipmentIoExpandersMoidCreated()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewPostEquipmentIoExpandersMoidDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *FrontSessionReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewFrontSessionOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (pr *PushedRequest) ReadResponse(ctx context.Context) (*http.Response, error) {\n\tselect {\n\tcase <-ctx.Done():\n\t\tpr.Cancel()\n\t\tpr.pushedStream.bufPipe.CloseWithError(ctx.Err())\n\t\treturn nil, ctx.Err()\n\tcase <-pr.pushedStream.peerReset:\n\t\treturn nil, pr.pushedStream.resetErr\n\tcase resErr := <-pr.pushedStream.resc:\n\t\tif resErr.err != nil {\n\t\t\tfmt.Println(resErr.err.Error())\n\t\t\tpr.Cancel()\n\t\t\tpr.pushedStream.bufPipe.CloseWithError(resErr.err)\n\t\t\treturn nil, resErr.err\n\t\t}\n\t\tresErr.res.Request = pr.Promise\n\t\tresErr.res.TLS = pr.pushedStream.cc.tlsState\n\t\treturn resErr.res, resErr.err\n\t}\n}", "func (o *GetZippedReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tresult := NewGetZippedDefault(response.Code())\n\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\treturn nil, err\n\t}\n\tif response.Code()/100 == 2 {\n\t\treturn result, nil\n\t}\n\treturn nil, result\n}", "func (o *DeleteFirmwareUpgradesMoidReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewDeleteFirmwareUpgradesMoidOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewDeleteFirmwareUpgradesMoidNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\tresult := NewDeleteFirmwareUpgradesMoidDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *GetEtherPhysicalPortsReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetEtherPhysicalPortsOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewGetEtherPhysicalPortsDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *ZoneStreamReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewZoneStreamOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *ByNamespaceReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewByNamespaceOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 404:\n\t\tresult := NewByNamespaceNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *GetRequestTrackerReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetRequestTrackerOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 403:\n\t\tresult := NewGetRequestTrackerForbidden()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 404:\n\t\tresult := NewGetRequestTrackerNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}" ]
[ "0.7640225", "0.7607834", "0.75210214", "0.7509121", "0.74803215", "0.74724656", "0.7433606", "0.74244606", "0.7375357", "0.7367311", "0.73589337", "0.73551553", "0.7350114", "0.7347274", "0.7346054", "0.733966", "0.7336042", "0.73239547", "0.7315819", "0.73155594", "0.7310195", "0.730769", "0.72904205", "0.7287086", "0.72826135", "0.72742283", "0.7274111", "0.72655845", "0.726384", "0.7262403", "0.7255057", "0.72496617", "0.72492456", "0.72479755", "0.72409135", "0.7224629", "0.722366", "0.7219326", "0.7216009", "0.72122216", "0.72110355", "0.72099286", "0.7209348", "0.72004783", "0.71978456", "0.719778", "0.71926844", "0.7177653", "0.71745974", "0.71737057", "0.716626", "0.7155474", "0.71500206", "0.7149536", "0.7148374", "0.7143972", "0.7143686", "0.7141745", "0.71397567", "0.713703", "0.7136677", "0.7136661", "0.7135863", "0.7135147", "0.71337897", "0.71312535", "0.7124361", "0.7123878", "0.71200526", "0.7120036", "0.7119569", "0.71148854", "0.7104891", "0.7100936", "0.70989054", "0.70989", "0.70984536", "0.70977753", "0.709657", "0.70961034", "0.70941985", "0.70932794", "0.70886916", "0.70850074", "0.7083912", "0.7080819", "0.7078785", "0.70775825", "0.70765215", "0.7076268", "0.7070042", "0.70699906", "0.7068155", "0.7068122", "0.7066828", "0.70625323", "0.70621973", "0.70599294", "0.70577264", "0.7054454", "0.70509636" ]
0.0
-1
NewGetComputeRackUnitsMoidOK creates a GetComputeRackUnitsMoidOK with default headers values
func NewGetComputeRackUnitsMoidOK() *GetComputeRackUnitsMoidOK { return &GetComputeRackUnitsMoidOK{} }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewGetComputeRackUnitsMoidDefault(code int) *GetComputeRackUnitsMoidDefault {\n\treturn &GetComputeRackUnitsMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetComputeRackUnitsMoidNotFound() *GetComputeRackUnitsMoidNotFound {\n\treturn &GetComputeRackUnitsMoidNotFound{}\n}", "func NewGetRacksOK() *GetRacksOK {\n\treturn &GetRacksOK{}\n}", "func (o *DcimRacksListParams) WithOuterUnitn(outerUnitn *string) *DcimRacksListParams {\n\to.SetOuterUnitn(outerUnitn)\n\treturn o\n}", "func NewGetRacksUnauthorized() *GetRacksUnauthorized {\n\treturn &GetRacksUnauthorized{}\n}", "func (o *DcimRacksListParams) WithDescUnits(descUnits *string) *DcimRacksListParams {\n\to.SetDescUnits(descUnits)\n\treturn o\n}", "func (o *DcimRacksListParams) WithOuterUnit(outerUnit *string) *DcimRacksListParams {\n\to.SetOuterUnit(outerUnit)\n\treturn o\n}", "func (o *MemoryArrayAllOf) GetComputeRackUnitOk() (*ComputeRackUnitRelationship, bool) {\n\tif o == nil || o.ComputeRackUnit == nil {\n\t\treturn nil, false\n\t}\n\treturn o.ComputeRackUnit, true\n}", "func NewGetLargeScreenDataRequestWithoutParam() *GetLargeScreenDataRequest {\n\n return &GetLargeScreenDataRequest{\n JDCloudRequest: core.JDCloudRequest{\n URL: \"/regions/{regionId}/getData\",\n Method: \"GET\",\n Header: nil,\n Version: \"v1\",\n },\n }\n}", "func NewGetRacksIDOK() *GetRacksIDOK {\n\treturn &GetRacksIDOK{}\n}", "func (o *DcimRacksListParams) SetOuterUnitn(outerUnitn *string) {\n\to.OuterUnitn = outerUnitn\n}", "func NewGetHyperflexServerModelsMoidOK() *GetHyperflexServerModelsMoidOK {\n\treturn &GetHyperflexServerModelsMoidOK{}\n}", "func (m ComputeRackUnit) MarshalJSON() ([]byte, error) {\n\t_parts := make([][]byte, 0, 2)\n\n\taO0, err := swag.WriteJSON(m.ComputePhysical)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t_parts = append(_parts, aO0)\n\n\tvar dataAO1 struct {\n\t\tAdapters []*AdapterUnitRef `json:\"Adapters\"`\n\n\t\tBiosBootmode *BiosBootModeRef `json:\"BiosBootmode,omitempty\"`\n\n\t\tBiosunits []*BiosUnitRef `json:\"Biosunits\"`\n\n\t\tBmc *ManagementControllerRef `json:\"Bmc,omitempty\"`\n\n\t\tBoard *ComputeBoardRef `json:\"Board,omitempty\"`\n\n\t\tBootDeviceBootmode *BootDeviceBootModeRef `json:\"BootDeviceBootmode,omitempty\"`\n\n\t\tFanmodules []*EquipmentFanModuleRef `json:\"Fanmodules\"`\n\n\t\tGenericInventoryHolders []*InventoryGenericInventoryHolderRef `json:\"GenericInventoryHolders\"`\n\n\t\tLocatorLed *EquipmentLocatorLedRef `json:\"LocatorLed,omitempty\"`\n\n\t\tPciDevices []*PciDeviceRef `json:\"PciDevices\"`\n\n\t\tPsus []*EquipmentPsuRef `json:\"Psus\"`\n\n\t\tRackEnclosureSlot *EquipmentRackEnclosureSlotRef `json:\"RackEnclosureSlot,omitempty\"`\n\n\t\tRegisteredDevice *AssetDeviceRegistrationRef `json:\"RegisteredDevice,omitempty\"`\n\n\t\tSasExpanders []*StorageSasExpanderRef `json:\"SasExpanders\"`\n\n\t\tServerID int64 `json:\"ServerId,omitempty\"`\n\n\t\tStorageEnclosures []*StorageEnclosureRef `json:\"StorageEnclosures\"`\n\n\t\tTopSystem *TopSystemRef `json:\"TopSystem,omitempty\"`\n\t}\n\n\tdataAO1.Adapters = m.Adapters\n\n\tdataAO1.BiosBootmode = m.BiosBootmode\n\n\tdataAO1.Biosunits = m.Biosunits\n\n\tdataAO1.Bmc = m.Bmc\n\n\tdataAO1.Board = m.Board\n\n\tdataAO1.BootDeviceBootmode = m.BootDeviceBootmode\n\n\tdataAO1.Fanmodules = m.Fanmodules\n\n\tdataAO1.GenericInventoryHolders = m.GenericInventoryHolders\n\n\tdataAO1.LocatorLed = m.LocatorLed\n\n\tdataAO1.PciDevices = m.PciDevices\n\n\tdataAO1.Psus = m.Psus\n\n\tdataAO1.RackEnclosureSlot = m.RackEnclosureSlot\n\n\tdataAO1.RegisteredDevice = m.RegisteredDevice\n\n\tdataAO1.SasExpanders = m.SasExpanders\n\n\tdataAO1.ServerID = m.ServerID\n\n\tdataAO1.StorageEnclosures = m.StorageEnclosures\n\n\tdataAO1.TopSystem = m.TopSystem\n\n\tjsonDataAO1, errAO1 := swag.WriteJSON(dataAO1)\n\tif errAO1 != nil {\n\t\treturn nil, errAO1\n\t}\n\t_parts = append(_parts, jsonDataAO1)\n\n\treturn swag.ConcatJSON(_parts...), nil\n}", "func NewGetEquipmentRackEnclosureSlotsOK() *GetEquipmentRackEnclosureSlotsOK {\n\treturn &GetEquipmentRackEnclosureSlotsOK{}\n}", "func (o *GetComputeRackUnitsMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *DcimRacksListParams) SetOuterUnit(outerUnit *string) {\n\to.OuterUnit = outerUnit\n}", "func (o *StorageEnclosure) GetComputeRackUnitOk() (*ComputeRackUnitRelationship, bool) {\n\tif o == nil || o.ComputeRackUnit == nil {\n\t\treturn nil, false\n\t}\n\treturn o.ComputeRackUnit, true\n}", "func (m *ComputeRackUnit) Validate(formats strfmt.Registry) error {\n\tvar res []error\n\n\t// validation for a type composition with ComputePhysical\n\tif err := m.ComputePhysical.Validate(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateAdapters(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateBiosBootmode(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateBiosunits(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateBmc(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateBoard(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateBootDeviceBootmode(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateFanmodules(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateGenericInventoryHolders(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateLocatorLed(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validatePciDevices(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validatePsus(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateRackEnclosureSlot(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateRegisteredDevice(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateSasExpanders(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateStorageEnclosures(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateTopSystem(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func NewCreateRackTopoOK() *CreateRackTopoOK {\n\treturn &CreateRackTopoOK{}\n}", "func (m *ComputeRackUnit) UnmarshalJSON(raw []byte) error {\n\t// AO0\n\tvar aO0 ComputePhysical\n\tif err := swag.ReadJSON(raw, &aO0); err != nil {\n\t\treturn err\n\t}\n\tm.ComputePhysical = aO0\n\n\t// AO1\n\tvar dataAO1 struct {\n\t\tAdapters []*AdapterUnitRef `json:\"Adapters\"`\n\n\t\tBiosBootmode *BiosBootModeRef `json:\"BiosBootmode,omitempty\"`\n\n\t\tBiosunits []*BiosUnitRef `json:\"Biosunits\"`\n\n\t\tBmc *ManagementControllerRef `json:\"Bmc,omitempty\"`\n\n\t\tBoard *ComputeBoardRef `json:\"Board,omitempty\"`\n\n\t\tBootDeviceBootmode *BootDeviceBootModeRef `json:\"BootDeviceBootmode,omitempty\"`\n\n\t\tFanmodules []*EquipmentFanModuleRef `json:\"Fanmodules\"`\n\n\t\tGenericInventoryHolders []*InventoryGenericInventoryHolderRef `json:\"GenericInventoryHolders\"`\n\n\t\tLocatorLed *EquipmentLocatorLedRef `json:\"LocatorLed,omitempty\"`\n\n\t\tPciDevices []*PciDeviceRef `json:\"PciDevices\"`\n\n\t\tPsus []*EquipmentPsuRef `json:\"Psus\"`\n\n\t\tRackEnclosureSlot *EquipmentRackEnclosureSlotRef `json:\"RackEnclosureSlot,omitempty\"`\n\n\t\tRegisteredDevice *AssetDeviceRegistrationRef `json:\"RegisteredDevice,omitempty\"`\n\n\t\tSasExpanders []*StorageSasExpanderRef `json:\"SasExpanders\"`\n\n\t\tServerID int64 `json:\"ServerId,omitempty\"`\n\n\t\tStorageEnclosures []*StorageEnclosureRef `json:\"StorageEnclosures\"`\n\n\t\tTopSystem *TopSystemRef `json:\"TopSystem,omitempty\"`\n\t}\n\tif err := swag.ReadJSON(raw, &dataAO1); err != nil {\n\t\treturn err\n\t}\n\n\tm.Adapters = dataAO1.Adapters\n\n\tm.BiosBootmode = dataAO1.BiosBootmode\n\n\tm.Biosunits = dataAO1.Biosunits\n\n\tm.Bmc = dataAO1.Bmc\n\n\tm.Board = dataAO1.Board\n\n\tm.BootDeviceBootmode = dataAO1.BootDeviceBootmode\n\n\tm.Fanmodules = dataAO1.Fanmodules\n\n\tm.GenericInventoryHolders = dataAO1.GenericInventoryHolders\n\n\tm.LocatorLed = dataAO1.LocatorLed\n\n\tm.PciDevices = dataAO1.PciDevices\n\n\tm.Psus = dataAO1.Psus\n\n\tm.RackEnclosureSlot = dataAO1.RackEnclosureSlot\n\n\tm.RegisteredDevice = dataAO1.RegisteredDevice\n\n\tm.SasExpanders = dataAO1.SasExpanders\n\n\tm.ServerID = dataAO1.ServerID\n\n\tm.StorageEnclosures = dataAO1.StorageEnclosures\n\n\tm.TopSystem = dataAO1.TopSystem\n\n\treturn nil\n}", "func NewUpdateRackTopoOK() *UpdateRackTopoOK {\n\treturn &UpdateRackTopoOK{}\n}", "func NewUpdateRackTopoOK() *UpdateRackTopoOK {\n\treturn &UpdateRackTopoOK{}\n}", "func NewDescribeBillSummarysRequestWithoutParam() *DescribeBillSummarysRequest {\n\n return &DescribeBillSummarysRequest{\n JDCloudRequest: core.JDCloudRequest{\n URL: \"/regions/{regionId}/describeBillSummarys\",\n Method: \"POST\",\n Header: nil,\n Version: \"v1\",\n },\n }\n}", "func NewGetRackTopoesParams() *GetRackTopoesParams {\n\treturn &GetRackTopoesParams{\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func NewGetStoragePureDisksMoidOK() *GetStoragePureDisksMoidOK {\n\treturn &GetStoragePureDisksMoidOK{}\n}", "func NewGetRacksParams() *GetRacksParams {\n\tvar ()\n\treturn &GetRacksParams{\n\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func NewGetStoragePureControllersMoidOK() *GetStoragePureControllersMoidOK {\n\treturn &GetStoragePureControllersMoidOK{}\n}", "func NewGetHyperflexServerModelsMoidDefault(code int) *GetHyperflexServerModelsMoidDefault {\n\treturn &GetHyperflexServerModelsMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func newAwsMediaStoreContainers(c *TrussleV1Client, namespace string) *awsMediaStoreContainers {\n\treturn &awsMediaStoreContainers{\n\t\tclient: c.RESTClient(),\n\t\tns: namespace,\n\t}\n}", "func NewGetHardwareOK() *GetHardwareOK {\n\treturn &GetHardwareOK{}\n}", "func NewGetMempoolOperationsOK() *GetMempoolOperationsOK {\n\n\treturn &GetMempoolOperationsOK{}\n}", "func (o *MemoryArrayAllOf) GetComputeRackUnit() ComputeRackUnitRelationship {\n\tif o == nil || o.ComputeRackUnit == nil {\n\t\tvar ret ComputeRackUnitRelationship\n\t\treturn ret\n\t}\n\treturn *o.ComputeRackUnit\n}", "func NewCmdShowRack() *cobra.Command {\n\tvar dc string\n\tvar floor string\n\tvar hall string\n\tvar row string\n\tvar pdu string\n\tvar rowCmd = &cobra.Command{\n\t\tUse: \"rack [RACK_NAME]\",\n\t\tAliases: []string{\"\"},\n\t\tShort: \"show rack\",\n\t\tArgs: func(cmd *cobra.Command, args []string) error {\n\t\t\tif len(args) < 1 {\n\t\t\t\tif cmd.Flag(\"dc\").Value.String() == \"\" {\n\t\t\t\t\tcmd.Help()\n\t\t\t\t\treturn fmt.Errorf(\"'dc' flag is required\")\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn nil\n\t\t},\n\t\tRun: showRack,\n\t\tPreRunE: checkServerVersion,\n\t}\n\trowCmd.Flags().StringVarP(&dc, \"dc\", \"\", \"\", \"specify datacenter [REQUIRED]\")\n\trowCmd.Flags().StringVarP(&floor, \"floor\", \"\", \"\", \"specify datacenter floor\")\n\trowCmd.Flags().StringVarP(&hall, \"hall\", \"\", \"\", \"specify datacenter hall\")\n\trowCmd.Flags().StringVarP(&row, \"row\", \"\", \"\", \"specify rack row\")\n\trowCmd.Flags().StringVarP(&pdu, \"row-pdu\", \"\", \"\", \"specify source row-pdu\")\n\t//rowCmd.MarkFlagRequired(\"dc\")\n\n\treturn rowCmd\n}", "func (o *GetComputeRackUnitsMoidReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetComputeRackUnitsMoidOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewGetComputeRackUnitsMoidNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\tresult := NewGetComputeRackUnitsMoidDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func NewGetVMVolumeMetricsOK() *GetVMVolumeMetricsOK {\n\treturn &GetVMVolumeMetricsOK{}\n}", "func (o *EquipmentLocatorLed) GetComputeRackUnitOk() (*ComputeRackUnitRelationship, bool) {\n\tif o == nil || o.ComputeRackUnit == nil {\n\t\treturn nil, false\n\t}\n\treturn o.ComputeRackUnit, true\n}", "func NewGetRacksServiceUnavailable() *GetRacksServiceUnavailable {\n\treturn &GetRacksServiceUnavailable{}\n}", "func NewGetEquipmentRackEnclosureSlotsDefault(code int) *GetEquipmentRackEnclosureSlotsDefault {\n\treturn &GetEquipmentRackEnclosureSlotsDefault{\n\t\t_statusCode: code,\n\t}\n}", "func (o *StorageSasExpander) GetComputeRackUnitOk() (*ComputeRackUnitRelationship, bool) {\n\tif o == nil || o.ComputeRackUnit == nil {\n\t\treturn nil, false\n\t}\n\treturn o.ComputeRackUnit, true\n}", "func newKubeNodeStatusCapacityMemoryBytesMetric(fqname string, node string, bytes float64) KubeNodeStatusCapacityMemoryBytesMetric {\n\treturn KubeNodeStatusCapacityMemoryBytesMetric{\n\t\tfqName: fqname,\n\t\thelp: \"kube_node_status_capacity_memory_bytes Node Capacity Memory Bytes\",\n\t\tnode: node,\n\t\tbytes: bytes,\n\t}\n}", "func NewRackUnitPersonality(classId string, objectType string) *RackUnitPersonality {\n\tthis := RackUnitPersonality{}\n\tthis.ClassId = classId\n\tthis.ObjectType = objectType\n\treturn &this\n}", "func (client *AvailabilitySetsClient) listAvailableSizesCreateRequest(ctx context.Context, resourceGroupName string, availabilitySetName string, options *AvailabilitySetsListAvailableSizesOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Compute/availabilitySets/{availabilitySetName}/vmSizes\"\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif availabilitySetName == \"\" {\n\t\treturn nil, errors.New(\"parameter availabilitySetName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{availabilitySetName}\", url.PathEscape(availabilitySetName))\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-07-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func NewGetRacksIDUnauthorized() *GetRacksIDUnauthorized {\n\treturn &GetRacksIDUnauthorized{}\n}", "func NewGetMemoryArraysMoidOK() *GetMemoryArraysMoidOK {\n\treturn &GetMemoryArraysMoidOK{}\n}", "func newKubeListRequest(values url.Values, site, resourceKind string) (*kubeproto.ListKubernetesResourcesRequest, error) {\n\tlimit, err := queryLimitAsInt32(values, \"limit\", defaults.MaxIterationLimit)\n\tif err != nil {\n\t\treturn nil, trace.Wrap(err)\n\t}\n\n\tsortBy := types.GetSortByFromString(values.Get(\"sort\"))\n\n\tstartKey := values.Get(\"startKey\")\n\treq := &kubeproto.ListKubernetesResourcesRequest{\n\t\tResourceType: resourceKind,\n\t\tLimit: limit,\n\t\tStartKey: startKey,\n\t\tSortBy: &sortBy,\n\t\tPredicateExpression: values.Get(\"query\"),\n\t\tSearchKeywords: client.ParseSearchKeywords(values.Get(\"search\"), ' '),\n\t\tUseSearchAsRoles: values.Get(\"searchAsRoles\") == \"yes\",\n\t\tTeleportCluster: site,\n\t\tKubernetesCluster: values.Get(\"kubeCluster\"),\n\t\tKubernetesNamespace: values.Get(\"kubeNamespace\"),\n\t}\n\treturn req, nil\n}", "func (o *MemoryArrayAllOf) HasComputeRackUnit() bool {\n\tif o != nil && o.ComputeRackUnit != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o StorageNodeStatusGeographyOutput) Rack() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v StorageNodeStatusGeography) *string { return v.Rack }).(pulumi.StringPtrOutput)\n}", "func NewDeviceOperatingSystemSummary()(*DeviceOperatingSystemSummary) {\n m := &DeviceOperatingSystemSummary{\n }\n m.backingStore = ie8677ce2c7e1b4c22e9c3827ecd078d41185424dd9eeb92b7d971ed2d49a392e.BackingStoreFactoryInstance();\n m.SetAdditionalData(make(map[string]any))\n return m\n}", "func NewGetAdapterHostEthInterfacesMoidOK() *GetAdapterHostEthInterfacesMoidOK {\n\treturn &GetAdapterHostEthInterfacesMoidOK{}\n}", "func NewCmdShowRackPDU() *cobra.Command {\n\tvar dc string\n\tvar ups string\n\tvar pdu string\n\tvar location string\n\tvar pduCmd = &cobra.Command{\n\t\tUse: \"rack-pdu [RACK_PDU_NAME]\",\n\t\tShort: \"show rack-pdu\",\n\t\tArgs: cobra.MaximumNArgs(1),\n\t\tRun: showRackPDU,\n\t\tPreRunE: checkServerVersion,\n\t}\n\tpduCmd.Flags().StringVarP(&dc, \"dc\", \"\", \"\", \"specify datacenter name\")\n\tpduCmd.Flags().StringVarP(&ups, \"ups\", \"\", \"\", \"specify ups name\")\n\tpduCmd.Flags().StringVarP(&pdu, \"pdu\", \"\", \"\", \"specify datacenter pdu name\")\n\tpduCmd.Flags().StringVarP(&location, \"location\", \"l\", \"\", \"specify rack-pdu location. use format '{DC}/{FLOOR}/{HALL}/{ROW}/{RACK}'\")\n\n\treturn pduCmd\n}", "func NewDcimRackRolesBulkPartialUpdateOK() *DcimRackRolesBulkPartialUpdateOK {\n\treturn &DcimRackRolesBulkPartialUpdateOK{}\n}", "func NewGetMachineSnapshotsOK() *GetMachineSnapshotsOK {\n\treturn &GetMachineSnapshotsOK{}\n}", "func (r Virtual_Guest) GetVirtualRack() (resp datatypes.Network_Bandwidth_Version1_Allotment, err error) {\n\terr = r.Session.DoRequest(\"SoftLayer_Virtual_Guest\", \"getVirtualRack\", nil, &r.Options, &resp)\n\treturn\n}", "func NewGetInventoryOK(body *GetResponseBody) *inventoryviews.InventoryView {\n\tv := &inventoryviews.InventoryView{\n\t\tID: body.ID,\n\t\tNumber: body.Number,\n\t\tCode: body.Code,\n\t\tType: body.Type,\n\t\tInventoryDate: body.InventoryDate,\n\t\tInAndOut: body.InAndOut,\n\t\tNote: body.Note,\n\t}\n\tv.Product = unmarshalProductResponseBodyToInventoryviewsProductView(body.Product)\n\tv.Warehouse = unmarshalWarehouseResponseBodyToInventoryviewsWarehouseView(body.Warehouse)\n\tv.Head = unmarshalHeadResponseBodyToInventoryviewsHeadView(body.Head)\n\tv.Founder = unmarshalFounderResponseBodyToInventoryviewsFounderView(body.Founder)\n\n\treturn v\n}", "func NewRackUnitPersonalityWithDefaults() *RackUnitPersonality {\n\tthis := RackUnitPersonality{}\n\tvar classId string = \"rack.UnitPersonality\"\n\tthis.ClassId = classId\n\tvar objectType string = \"rack.UnitPersonality\"\n\tthis.ObjectType = objectType\n\treturn &this\n}", "func (o *MemoryArrayAllOf) SetComputeRackUnit(v ComputeRackUnitRelationship) {\n\to.ComputeRackUnit = &v\n}", "func NewCiMacOsVersionsGetCollectionRequest(server string, params *CiMacOsVersionsGetCollectionParams) (*http.Request, error) {\n\tvar err error\n\n\tserverURL, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\toperationPath := fmt.Sprintf(\"/v1/ciMacOsVersions\")\n\tif operationPath[0] == '/' {\n\t\toperationPath = \".\" + operationPath\n\t}\n\n\tqueryURL, err := serverURL.Parse(operationPath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tqueryValues := queryURL.Query()\n\n\tif params.FieldsCiMacOsVersions != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"fields[ciMacOsVersions]\", runtime.ParamLocationQuery, *params.FieldsCiMacOsVersions); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.Limit != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", true, \"limit\", runtime.ParamLocationQuery, *params.Limit); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.Include != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"include\", runtime.ParamLocationQuery, *params.Include); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.FieldsCiXcodeVersions != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", false, \"fields[ciXcodeVersions]\", runtime.ParamLocationQuery, *params.FieldsCiXcodeVersions); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.LimitXcodeVersions != nil {\n\n\t\tif queryFrag, err := runtime.StyleParamWithLocation(\"form\", true, \"limit[xcodeVersions]\", runtime.ParamLocationQuery, *params.LimitXcodeVersions); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tqueryURL.RawQuery = queryValues.Encode()\n\n\treq, err := http.NewRequest(\"GET\", queryURL.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn req, nil\n}", "func (o *RackUnitPersonality) GetComputeRackUnitOk() (*ComputeRackUnitRelationship, bool) {\n\tif o == nil || o.ComputeRackUnit == nil {\n\t\treturn nil, false\n\t}\n\treturn o.ComputeRackUnit, true\n}", "func (me *masterExtension) Units() ([]k8scloudconfig.UnitAsset, error) {\n\tunitsMeta := []k8scloudconfig.UnitMetadata{\n\t\t{\n\t\t\tAssetContent: ignition.AzureCNINatRules,\n\t\t\tName: \"azure-cni-nat-rules.service\",\n\t\t\tEnabled: true,\n\t\t},\n\t\t{\n\t\t\tAssetContent: ignition.CertificateDecrypterUnit,\n\t\t\tName: \"certificate-decrypter.service\",\n\t\t\tEnabled: true,\n\t\t},\n\t\t{\n\t\t\tAssetContent: ignition.EtcdMountUnit,\n\t\t\tName: \"var-lib-etcd.mount\",\n\t\t\tEnabled: true,\n\t\t},\n\t\t{\n\t\t\tAssetContent: ignition.DockerMountUnit,\n\t\t\tName: \"var-lib-docker.mount\",\n\t\t\tEnabled: true,\n\t\t},\n\t\t{\n\t\t\tAssetContent: ignition.KubeletMountUnit,\n\t\t\tName: \"var-lib-kubelet.mount\",\n\t\t\tEnabled: true,\n\t\t},\n\t\t{\n\t\t\tAssetContent: ignition.VNICConfigurationUnit,\n\t\t\tName: \"vnic-configuration.service\",\n\t\t\tEnabled: true,\n\t\t},\n\t}\n\n\tdata := me.templateData(me.certFiles)\n\n\t// To use the certificate decrypter unit for the etcd data encryption config file.\n\tdata.certificateDecrypterUnitParams.CertsPaths = append(data.certificateDecrypterUnitParams.CertsPaths, encryptionConfigFilePath)\n\n\tvar newUnits []k8scloudconfig.UnitAsset\n\n\tfor _, fm := range unitsMeta {\n\t\tc, err := k8scloudconfig.RenderAssetContent(fm.AssetContent, data)\n\t\tif err != nil {\n\t\t\treturn nil, microerror.Mask(err)\n\t\t}\n\n\t\tunitAsset := k8scloudconfig.UnitAsset{\n\t\t\tMetadata: fm,\n\t\t\tContent: c,\n\t\t}\n\n\t\tnewUnits = append(newUnits, unitAsset)\n\t}\n\n\treturn newUnits, nil\n}", "func CreateGetTotalQueueReportRequest() (request *GetTotalQueueReportRequest) {\n\trequest = &GetTotalQueueReportRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"EHPC\", \"2018-04-12\", \"GetTotalQueueReport\", \"ehs\", \"openAPI\")\n\treturn\n}", "func NewGetRacksNotFound() *GetRacksNotFound {\n\treturn &GetRacksNotFound{}\n}", "func mockedVirtualMachines(reqUrl string) (virtualMachines []armcompute.VirtualMachine, err error) {\n\tvar mockedVirtualMachinesResponse mockedVirtualMachinesResponse\n\n\tm := newMockComputeSender()\n\treq, err := http.NewRequest(\"GET\", reqUrl, nil)\n\tif err != nil {\n\t\treturn virtualMachines, fmt.Errorf(\"error creating new request: %w\", err)\n\t}\n\tresp, err := m.Do(req)\n\tif err != nil || resp.StatusCode == 404 {\n\t\treturn virtualMachines, fmt.Errorf(\"error getting mock http response: %w\", err)\n\t}\n\n\tdefer func(Body io.ReadCloser) {\n\t\terr := Body.Close()\n\t\tif err != nil {\n\t\t\tfmt.Println(\"error io.ReadCloser: %w\", err)\n\t\t}\n\t}(resp.Body)\n\tresponseBody, err := io.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn virtualMachines, fmt.Errorf(\"error read all: %w\", err)\n\t}\n\terr = json.Unmarshal(responseBody, &mockedVirtualMachinesResponse)\n\tif err != nil {\n\t\treturn virtualMachines, fmt.Errorf(\"error unmarshalling: %w\", err)\n\t}\n\n\tvirtualMachines = mockedVirtualMachinesResponse.Value\n\n\treturn virtualMachines, nil\n}", "func newMachineSet(name string, cluster *clusteroperator.Cluster, properlyOwned bool) *clusteroperator.MachineSet {\n\tvar controllerReference metav1.OwnerReference\n\tif properlyOwned {\n\t\ttrueVar := true\n\t\tcontrollerReference = metav1.OwnerReference{\n\t\t\tUID: testClusterUUID,\n\t\t\tAPIVersion: clusteroperator.SchemeGroupVersion.String(),\n\t\t\tKind: \"Cluster\",\n\t\t\tName: cluster.Name,\n\t\t\tController: &trueVar,\n\t\t}\n\t}\n\treturn &clusteroperator.MachineSet{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tName: name,\n\t\t\tNamespace: cluster.Namespace,\n\t\t\tOwnerReferences: []metav1.OwnerReference{controllerReference},\n\t\t},\n\t}\n}", "func (o *GetRacksParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.AssetNo != nil {\n\n\t\t// query param asset_no\n\t\tvar qrAssetNo string\n\t\tif o.AssetNo != nil {\n\t\t\tqrAssetNo = *o.AssetNo\n\t\t}\n\t\tqAssetNo := qrAssetNo\n\t\tif qAssetNo != \"\" {\n\t\t\tif err := r.SetQueryParam(\"asset_no\", qAssetNo); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Building != nil {\n\n\t\t// query param building\n\t\tvar qrBuilding string\n\t\tif o.Building != nil {\n\t\t\tqrBuilding = *o.Building\n\t\t}\n\t\tqBuilding := qrBuilding\n\t\tif qBuilding != \"\" {\n\t\t\tif err := r.SetQueryParam(\"building\", qBuilding); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.BuildingID != nil {\n\n\t\t// query param building_id\n\t\tvar qrBuildingID string\n\t\tif o.BuildingID != nil {\n\t\t\tqrBuildingID = *o.BuildingID\n\t\t}\n\t\tqBuildingID := qrBuildingID\n\t\tif qBuildingID != \"\" {\n\t\t\tif err := r.SetQueryParam(\"building_id\", qBuildingID); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Manufacturer != nil {\n\n\t\t// query param manufacturer\n\t\tvar qrManufacturer string\n\t\tif o.Manufacturer != nil {\n\t\t\tqrManufacturer = *o.Manufacturer\n\t\t}\n\t\tqManufacturer := qrManufacturer\n\t\tif qManufacturer != \"\" {\n\t\t\tif err := r.SetQueryParam(\"manufacturer\", qManufacturer); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Name != nil {\n\n\t\t// query param name\n\t\tvar qrName string\n\t\tif o.Name != nil {\n\t\t\tqrName = *o.Name\n\t\t}\n\t\tqName := qrName\n\t\tif qName != \"\" {\n\t\t\tif err := r.SetQueryParam(\"name\", qName); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Room != nil {\n\n\t\t// query param room\n\t\tvar qrRoom string\n\t\tif o.Room != nil {\n\t\t\tqrRoom = *o.Room\n\t\t}\n\t\tqRoom := qrRoom\n\t\tif qRoom != \"\" {\n\t\t\tif err := r.SetQueryParam(\"room\", qRoom); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.RoomID != nil {\n\n\t\t// query param room_id\n\t\tvar qrRoomID string\n\t\tif o.RoomID != nil {\n\t\t\tqrRoomID = *o.RoomID\n\t\t}\n\t\tqRoomID := qrRoomID\n\t\tif qRoomID != \"\" {\n\t\t\tif err := r.SetQueryParam(\"room_id\", qRoomID); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Row != nil {\n\n\t\t// query param row\n\t\tvar qrRow string\n\t\tif o.Row != nil {\n\t\t\tqrRow = *o.Row\n\t\t}\n\t\tqRow := qrRow\n\t\tif qRow != \"\" {\n\t\t\tif err := r.SetQueryParam(\"row\", qRow); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.Size != nil {\n\n\t\t// query param size\n\t\tvar qrSize int64\n\t\tif o.Size != nil {\n\t\t\tqrSize = *o.Size\n\t\t}\n\t\tqSize := swag.FormatInt64(qrSize)\n\t\tif qSize != \"\" {\n\t\t\tif err := r.SetQueryParam(\"size\", qSize); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func CreateGetHealthMonitorLogsRequest() (request *GetHealthMonitorLogsRequest) {\n\trequest = &GetHealthMonitorLogsRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"EHPC\", \"2018-04-12\", \"GetHealthMonitorLogs\", \"\", \"\")\n\trequest.Method = requests.GET\n\treturn\n}", "func NewIpspacesGetOK() *IpspacesGetOK {\n\treturn &IpspacesGetOK{}\n}", "func CreateDescribeCloudMonitorAgentStatusesRequest() (request *DescribeCloudMonitorAgentStatusesRequest) {\n\trequest = &DescribeCloudMonitorAgentStatusesRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"SWAS-OPEN\", \"2020-06-01\", \"DescribeCloudMonitorAgentStatuses\", \"SWAS-OPEN\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func (r ApiGetMetricsForBitlinkByDevicesRequest) Units(units int32) ApiGetMetricsForBitlinkByDevicesRequest {\n\tr.units = &units\n\treturn r\n}", "func NewGetCloudSystemMetricsOK() *GetCloudSystemMetricsOK {\n\treturn &GetCloudSystemMetricsOK{}\n}", "func NewGetVMMetricsOK() *GetVMMetricsOK {\n\treturn &GetVMMetricsOK{}\n}", "func (o *GetRacksParams) WithRoom(room *string) *GetRacksParams {\n\to.SetRoom(room)\n\treturn o\n}", "func NewGetVMFoldersOK() *GetVMFoldersOK {\n\treturn &GetVMFoldersOK{}\n}", "func (*stockKeepingUnitContentR) NewStruct() *stockKeepingUnitContentR {\n\treturn &stockKeepingUnitContentR{}\n}", "func NewGetHyperflexClusterStoragePoliciesMoidOK() *GetHyperflexClusterStoragePoliciesMoidOK {\n\treturn &GetHyperflexClusterStoragePoliciesMoidOK{}\n}", "func NewGetRacksBadRequest() *GetRacksBadRequest {\n\treturn &GetRacksBadRequest{}\n}", "func NewReadFromMicrostorageOK() *ReadFromMicrostorageOK {\n\treturn &ReadFromMicrostorageOK{}\n}", "func (o *EquipmentFanModule) GetComputeRackUnitOk() (*ComputeRackUnitRelationship, bool) {\n\tif o == nil || o.ComputeRackUnit == nil {\n\t\treturn nil, false\n\t}\n\treturn o.ComputeRackUnit, true\n}", "func newKubeNodeStatusCapacityCPUCoresMetric(fqname string, node string, cores float64) KubeNodeStatusCapacityCPUCoresMetric {\n\treturn KubeNodeStatusCapacityCPUCoresMetric{\n\t\tfqName: fqname,\n\t\thelp: \"kube_node_status_capacity_cpu_cores Node Capacity CPU Cores\",\n\t\tcores: cores,\n\t\tnode: node,\n\t}\n}", "func NewListSmsSendOverviewUsingGETRequestWithoutParam() *ListSmsSendOverviewUsingGETRequest {\n\n return &ListSmsSendOverviewUsingGETRequest{\n JDCloudRequest: core.JDCloudRequest{\n URL: \"/smsApps/{appId}:overview\",\n Method: \"GET\",\n Header: nil,\n Version: \"v1\",\n },\n }\n}", "func (o *EquipmentLocatorLed) GetComputeRackUnit() ComputeRackUnitRelationship {\n\tif o == nil || o.ComputeRackUnit == nil {\n\t\tvar ret ComputeRackUnitRelationship\n\t\treturn ret\n\t}\n\treturn *o.ComputeRackUnit\n}", "func (c *UPHostClient) NewDescribeBaremetalMachineTypeRequest() *DescribeBaremetalMachineTypeRequest {\n\treq := &DescribeBaremetalMachineTypeRequest{}\n\n\t// setup request with client config\n\tc.Client.SetupRequest(req)\n\n\t// setup retryable with default retry policy (retry for non-create action and common error)\n\treq.SetRetryable(true)\n\treturn req\n}", "func NewGetHardwareDefault(code int) *GetHardwareDefault {\n\treturn &GetHardwareDefault{\n\t\t_statusCode: code,\n\t}\n}", "func (b *CompanyRequestBuilder) UnitsOfMeasure() *CompanyUnitsOfMeasureCollectionRequestBuilder {\n\tbb := &CompanyUnitsOfMeasureCollectionRequestBuilder{BaseRequestBuilder: b.BaseRequestBuilder}\n\tbb.baseURL += \"/unitsOfMeasure\"\n\treturn bb\n}", "func NewDeleteRackTopoOK() *DeleteRackTopoOK {\n\treturn &DeleteRackTopoOK{}\n}", "func NewGetAdapterHostFcInterfacesMoidOK() *GetAdapterHostFcInterfacesMoidOK {\n\treturn &GetAdapterHostFcInterfacesMoidOK{}\n}", "func NewStatusReportRequestWithoutParam() *StatusReportRequest {\n\n return &StatusReportRequest{\n JDCloudRequest: core.JDCloudRequest{\n URL: \"/regions/{regionId}/statusReport\",\n Method: \"POST\",\n Header: nil,\n Version: \"v1\",\n },\n }\n}", "func (r ApiGetSortedBitlinksRequest) Units(units int32) ApiGetSortedBitlinksRequest {\n\tr.units = &units\n\treturn r\n}", "func (o *StorageSasExpander) GetComputeRackUnit() ComputeRackUnitRelationship {\n\tif o == nil || o.ComputeRackUnit == nil {\n\t\tvar ret ComputeRackUnitRelationship\n\t\treturn ret\n\t}\n\treturn *o.ComputeRackUnit\n}", "func NewMosn() *Mosn {\n\tlog.StartLogger.Infof(\"[mosn start] create an empty mosn structure\")\n\tm := &Mosn{\n\t\tUpgrade: UpgradeData{},\n\t}\n\treturn m\n}", "func (o *StorageEnclosure) GetComputeRackUnit() ComputeRackUnitRelationship {\n\tif o == nil || o.ComputeRackUnit == nil {\n\t\tvar ret ComputeRackUnitRelationship\n\t\treturn ret\n\t}\n\treturn *o.ComputeRackUnit\n}", "func NewGetOSListRequest(page int, size int) *GetOSListRequest {\n\treturn &GetOSListRequest{\n\t\tPage: page,\n\t\tSize: size,\n\t}\n}", "func NewGetNiaapiDcnmHweolsMoidOK() *GetNiaapiDcnmHweolsMoidOK {\n\treturn &GetNiaapiDcnmHweolsMoidOK{}\n}", "func (o *EquipmentLocatorLed) SetComputeRackUnit(v ComputeRackUnitRelationship) {\n\to.ComputeRackUnit = &v\n}", "func newKubeNodeStatusAllocatableMemoryBytesMetric(fqname, node string, value float64) KubeNodeStatusAllocatableMemoryBytesMetric {\n\treturn KubeNodeStatusAllocatableMemoryBytesMetric{\n\t\tfqName: fqname,\n\t\thelp: \"kube_node_status_allocatable_memory_bytes node allocatable memory in bytes\",\n\t\tnode: node,\n\t\tvalue: value,\n\t}\n}", "func (o OceanLaunchSpecSchedulingTaskTaskHeadroomOutput) NumOfUnits() pulumi.IntOutput {\n\treturn o.ApplyT(func(v OceanLaunchSpecSchedulingTaskTaskHeadroom) int { return v.NumOfUnits }).(pulumi.IntOutput)\n}", "func (a *HyperflexApiService) GetHyperflexStorageContainerByMoidExecute(r ApiGetHyperflexStorageContainerByMoidRequest) (*HyperflexStorageContainer, *http.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = http.MethodGet\n\t\tlocalVarPostBody interface{}\n\t\tformFiles []formFile\n\t\tlocalVarReturnValue *HyperflexStorageContainer\n\t)\n\n\tlocalBasePath, err := a.client.cfg.ServerURLWithContext(r.ctx, \"HyperflexApiService.GetHyperflexStorageContainerByMoid\")\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, &GenericOpenAPIError{error: err.Error()}\n\t}\n\n\tlocalVarPath := localBasePath + \"/api/v1/hyperflex/StorageContainers/{Moid}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"Moid\"+\"}\", url.PathEscape(parameterToString(r.moid, \"\")), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\", \"text/csv\", \"application/vnd.openxmlformats-officedocument.spreadsheetml.sheet\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\treq, err := a.client.prepareRequest(r.ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, formFiles)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(req)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tlocalVarHTTPResponse.Body = ioutil.NopCloser(bytes.NewBuffer(localVarBody))\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := &GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 400 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 401 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 403 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tvar v Error\n\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\tif err != nil {\n\t\t\tnewErr.error = err.Error()\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tnewErr.model = v\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := &GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func (o ElastigroupIntegrationDockerSwarmAutoscaleHeadroomOutput) NumOfUnits() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v ElastigroupIntegrationDockerSwarmAutoscaleHeadroom) *int { return v.NumOfUnits }).(pulumi.IntPtrOutput)\n}", "func newClusterWithSizes(masterSize int, computes ...clusteroperator.ClusterMachineSet) *clusteroperator.Cluster {\n\tcluster := &clusteroperator.Cluster{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tUID: testClusterUUID,\n\t\t\tName: testClusterName,\n\t\t\tNamespace: testNamespace,\n\t\t},\n\t\tSpec: clusteroperator.ClusterSpec{\n\t\t\tMachineSets: append(computes, clusteroperator.ClusterMachineSet{\n\t\t\t\tName: \"master\",\n\t\t\t\tMachineSetConfig: clusteroperator.MachineSetConfig{\n\t\t\t\t\tInfra: true,\n\t\t\t\t\tSize: masterSize,\n\t\t\t\t\tNodeType: clusteroperator.NodeTypeMaster,\n\t\t\t\t},\n\t\t\t}),\n\t\t},\n\t\tStatus: clusteroperator.ClusterStatus{\n\t\t\tMasterMachineSetName: testClusterName + \"-master-random\",\n\t\t\tInfraMachineSetName: testClusterName + \"-master-random\",\n\t\t},\n\t}\n\treturn cluster\n}", "func NewGetVMVolumeOK() *GetVMVolumeOK {\n\n\treturn &GetVMVolumeOK{}\n}", "func NewReportsMicrosoftGraphNetworkaccessGetDeviceUsageSummaryWithStartDateTimeWithEndDateTimeWithActivityPivotDateTimeRequestBuilderInternal(pathParameters map[string]string, requestAdapter i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.RequestAdapter, activityPivotDateTime *i336074805fc853987abe6f7fe3ad97a6a6f3077a16391fec744f671a015fbd7e.Time, endDateTime *i336074805fc853987abe6f7fe3ad97a6a6f3077a16391fec744f671a015fbd7e.Time, startDateTime *i336074805fc853987abe6f7fe3ad97a6a6f3077a16391fec744f671a015fbd7e.Time)(*ReportsMicrosoftGraphNetworkaccessGetDeviceUsageSummaryWithStartDateTimeWithEndDateTimeWithActivityPivotDateTimeRequestBuilder) {\n m := &ReportsMicrosoftGraphNetworkaccessGetDeviceUsageSummaryWithStartDateTimeWithEndDateTimeWithActivityPivotDateTimeRequestBuilder{\n BaseRequestBuilder: *i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.NewBaseRequestBuilder(requestAdapter, \"{+baseurl}/networkAccess/reports/microsoft.graph.networkaccess.getDeviceUsageSummary(startDateTime={startDateTime},endDateTime={endDateTime},activityPivotDateTime={activityPivotDateTime})\", pathParameters),\n }\n if activityPivotDateTime != nil {\n m.BaseRequestBuilder.PathParameters[\"activityPivotDateTime\"] = (*activityPivotDateTime).Format(i336074805fc853987abe6f7fe3ad97a6a6f3077a16391fec744f671a015fbd7e.RFC3339)\n }\n if endDateTime != nil {\n m.BaseRequestBuilder.PathParameters[\"endDateTime\"] = (*endDateTime).Format(i336074805fc853987abe6f7fe3ad97a6a6f3077a16391fec744f671a015fbd7e.RFC3339)\n }\n if startDateTime != nil {\n m.BaseRequestBuilder.PathParameters[\"startDateTime\"] = (*startDateTime).Format(i336074805fc853987abe6f7fe3ad97a6a6f3077a16391fec744f671a015fbd7e.RFC3339)\n }\n return m\n}" ]
[ "0.7532864", "0.60939354", "0.52469087", "0.5219634", "0.51336026", "0.5010661", "0.5009418", "0.4823311", "0.4805804", "0.4771977", "0.47522828", "0.4608592", "0.4553403", "0.45409843", "0.4539188", "0.44669628", "0.44094688", "0.44045812", "0.43989533", "0.43974108", "0.43724528", "0.43724528", "0.43713176", "0.4337442", "0.43272978", "0.43250078", "0.43013284", "0.42998865", "0.4286935", "0.42820773", "0.42518625", "0.42382768", "0.42336962", "0.42322826", "0.4231724", "0.42247137", "0.42191416", "0.4204338", "0.42032462", "0.41970697", "0.41699716", "0.41610128", "0.41504157", "0.41452017", "0.4126155", "0.41141942", "0.41119608", "0.4096311", "0.40945664", "0.40904713", "0.40865505", "0.40835574", "0.40705958", "0.40664744", "0.40628275", "0.40592512", "0.40567255", "0.405091", "0.40470827", "0.4042026", "0.4025793", "0.40155023", "0.40142798", "0.4008058", "0.40069512", "0.40041998", "0.39939982", "0.39882344", "0.39880288", "0.39804325", "0.39737737", "0.39721236", "0.39698198", "0.39667585", "0.39636886", "0.39566556", "0.39536524", "0.39481196", "0.39461908", "0.39427614", "0.39412415", "0.39400908", "0.3939755", "0.3938014", "0.39260036", "0.39248103", "0.39235902", "0.39190918", "0.39180696", "0.39144936", "0.3914386", "0.39138758", "0.390682", "0.3904627", "0.39044064", "0.39018643", "0.3899901", "0.38982877", "0.38974082", "0.38971588" ]
0.75314784
1
NewGetComputeRackUnitsMoidNotFound creates a GetComputeRackUnitsMoidNotFound with default headers values
func NewGetComputeRackUnitsMoidNotFound() *GetComputeRackUnitsMoidNotFound { return &GetComputeRackUnitsMoidNotFound{} }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewGetComputeRackUnitsMoidDefault(code int) *GetComputeRackUnitsMoidDefault {\n\treturn &GetComputeRackUnitsMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetComputeRackUnitsMoidOK() *GetComputeRackUnitsMoidOK {\n\treturn &GetComputeRackUnitsMoidOK{}\n}", "func NewGetRacksNotFound() *GetRacksNotFound {\n\treturn &GetRacksNotFound{}\n}", "func NewGetHyperflexServerModelsMoidNotFound() *GetHyperflexServerModelsMoidNotFound {\n\treturn &GetHyperflexServerModelsMoidNotFound{}\n}", "func NewGetRacksIDNotFound() *GetRacksIDNotFound {\n\treturn &GetRacksIDNotFound{}\n}", "func NewGetStoragePureDisksMoidNotFound() *GetStoragePureDisksMoidNotFound {\n\treturn &GetStoragePureDisksMoidNotFound{}\n}", "func NewGetStoragePureControllersMoidNotFound() *GetStoragePureControllersMoidNotFound {\n\treturn &GetStoragePureControllersMoidNotFound{}\n}", "func NewGetNiaapiDcnmHweolsMoidNotFound() *GetNiaapiDcnmHweolsMoidNotFound {\n\treturn &GetNiaapiDcnmHweolsMoidNotFound{}\n}", "func NewGetAdapterHostEthInterfacesMoidNotFound() *GetAdapterHostEthInterfacesMoidNotFound {\n\treturn &GetAdapterHostEthInterfacesMoidNotFound{}\n}", "func (o *GetComputeRackUnitsMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func NewGetRacksUnauthorized() *GetRacksUnauthorized {\n\treturn &GetRacksUnauthorized{}\n}", "func NewNotFound() error {\n\treturn requestError{\n\t\tClientError: ClientError{\n\t\t\tErrors: []clientErrorSubError{{Message: \"status code 404\"}},\n\t\t},\n\t}\n}", "func NewGetAdapterHostFcInterfacesMoidNotFound() *GetAdapterHostFcInterfacesMoidNotFound {\n\treturn &GetAdapterHostFcInterfacesMoidNotFound{}\n}", "func NewGetCloudSystemMetricsNotFound() *GetCloudSystemMetricsNotFound {\n\treturn &GetCloudSystemMetricsNotFound{}\n}", "func NewGetNiaapiDcnmSweolsMoidNotFound() *GetNiaapiDcnmSweolsMoidNotFound {\n\treturn &GetNiaapiDcnmSweolsMoidNotFound{}\n}", "func NewGetMempoolOperationsNotFound() *GetMempoolOperationsNotFound {\n\n\treturn &GetMempoolOperationsNotFound{}\n}", "func NewGetNiaapiApicSweolsMoidNotFound() *GetNiaapiApicSweolsMoidNotFound {\n\treturn &GetNiaapiApicSweolsMoidNotFound{}\n}", "func NewGetManagementControllersMoidNotFound() *GetManagementControllersMoidNotFound {\n\treturn &GetManagementControllersMoidNotFound{}\n}", "func NewGetMachineSnapshotsNotFound() *GetMachineSnapshotsNotFound {\n\treturn &GetMachineSnapshotsNotFound{}\n}", "func NewGetNicsNotFound() *GetNicsNotFound {\n\treturn &GetNicsNotFound{}\n}", "func NewGetHyperflexClusterStoragePoliciesMoidNotFound() *GetHyperflexClusterStoragePoliciesMoidNotFound {\n\treturn &GetHyperflexClusterStoragePoliciesMoidNotFound{}\n}", "func NewUpdateRackTopoNotFound() *UpdateRackTopoNotFound {\n\treturn &UpdateRackTopoNotFound{}\n}", "func NewGetMemoryArraysMoidNotFound() *GetMemoryArraysMoidNotFound {\n\treturn &GetMemoryArraysMoidNotFound{}\n}", "func NewGetVMMetricsNotFound() *GetVMMetricsNotFound {\n\treturn &GetVMMetricsNotFound{}\n}", "func NewGetLargeScreenDataRequestWithoutParam() *GetLargeScreenDataRequest {\n\n return &GetLargeScreenDataRequest{\n JDCloudRequest: core.JDCloudRequest{\n URL: \"/regions/{regionId}/getData\",\n Method: \"GET\",\n Header: nil,\n Version: \"v1\",\n },\n }\n}", "func NewReadFromMicrostorageNotFound() *ReadFromMicrostorageNotFound {\n\treturn &ReadFromMicrostorageNotFound{}\n}", "func NewSystemGetNotFound() *SystemGetNotFound {\n\treturn &SystemGetNotFound{}\n}", "func NewGetObmsLibraryIdentifierNotFound() *GetObmsLibraryIdentifierNotFound {\n\treturn &GetObmsLibraryIdentifierNotFound{}\n}", "func NewGetHostGroupsNotFound() *GetHostGroupsNotFound {\n\treturn &GetHostGroupsNotFound{}\n}", "func NewGetHyperflexServerModelsMoidDefault(code int) *GetHyperflexServerModelsMoidDefault {\n\treturn &GetHyperflexServerModelsMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetNodeNotFound() *GetNodeNotFound {\n\treturn &GetNodeNotFound{}\n}", "func NewGetVMVolumeMetricsNotFound() *GetVMVolumeMetricsNotFound {\n\treturn &GetVMVolumeMetricsNotFound{}\n}", "func NewGetRacksIDUnauthorized() *GetRacksIDUnauthorized {\n\treturn &GetRacksIDUnauthorized{}\n}", "func NewCreateRackTopoNotFound() *CreateRackTopoNotFound {\n\treturn &CreateRackTopoNotFound{}\n}", "func NewUploadMediaNotFound(body *UploadMediaNotFoundResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func (o *MemoryArrayAllOf) GetComputeRackUnitOk() (*ComputeRackUnitRelationship, bool) {\n\tif o == nil || o.ComputeRackUnit == nil {\n\t\treturn nil, false\n\t}\n\treturn o.ComputeRackUnit, true\n}", "func NewNotFound(s string, v ...interface{}) error {\n\treturn asNotFound(fmt.Errorf(s, v...))\n}", "func NewGetNiaapiDcnmHweolsMoidDefault(code int) *GetNiaapiDcnmHweolsMoidDefault {\n\treturn &GetNiaapiDcnmHweolsMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetTableSizeNotFound() *GetTableSizeNotFound {\n\treturn &GetTableSizeNotFound{}\n}", "func NewGetRacksBadRequest() *GetRacksBadRequest {\n\treturn &GetRacksBadRequest{}\n}", "func NewGetRacksServiceUnavailable() *GetRacksServiceUnavailable {\n\treturn &GetRacksServiceUnavailable{}\n}", "func NewGetServerConfigImportsMoidNotFound() *GetServerConfigImportsMoidNotFound {\n\treturn &GetServerConfigImportsMoidNotFound{}\n}", "func NewGetVMVolumeDefault(code int) *GetVMVolumeDefault {\n\tif code <= 0 {\n\t\tcode = 500\n\t}\n\n\treturn &GetVMVolumeDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewNotFound(parameters ...wparams.ParamStorer) Error {\n\treturn newGenericError(nil, DefaultNotFound, wparams.NewParamStorer(parameters...))\n}", "func NewShowNotFound(body *ShowNotFoundResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func NewShowNotFound(body *ShowNotFoundResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func NewNotFoundError(message string)*RestErr{\n\treturn &RestErr{\n\t\tMessage: message,\n\t\tStatus: http.StatusNotFound,\n\t\tError: \"Not Found\",\n\t}\n}", "func NewNotFoundErr(err error) Error {\n\treturn StatusError{\n\t\tCode: http.StatusNotFound,\n\t\tErr: err,\n\t}\n}", "func NewGetNiaapiDcnmCcoPostsMoidNotFound() *GetNiaapiDcnmCcoPostsMoidNotFound {\n\treturn &GetNiaapiDcnmCcoPostsMoidNotFound{}\n}", "func NewDeviceLayoutNotFound(body DeviceLayoutNotFoundResponseBody) information.NotFound {\n\tv := information.NotFound(body)\n\treturn v\n}", "func NewGetModelRegistryNotFound() *GetModelRegistryNotFound {\n\treturn &GetModelRegistryNotFound{}\n}", "func NewNotFound(err error, msg ...string) *Errs {\n\tif err == nil {\n\t\terr = ErrNotFound\n\t}\n\treturn &Errs{\n\t\tcodeHTTP: http.StatusNotFound,\n\t\terr: err,\n\t\tkind: trace(2),\n\t\tmessage: msg,\n\t}\n}", "func NewGetTerminalAuditLogsMoidNotFound() *GetTerminalAuditLogsMoidNotFound {\n\treturn &GetTerminalAuditLogsMoidNotFound{}\n}", "func NewNotFound(msg string) error {\n\treturn &ELBError{\n\t\tmsg: msg,\n\t\tCode: http.StatusNotFound,\n\t}\n}", "func (a *HyperflexApiService) GetHyperflexClusterBackupPolicyInventoryByMoid(ctx context.Context, moid string) ApiGetHyperflexClusterBackupPolicyInventoryByMoidRequest {\n\treturn ApiGetHyperflexClusterBackupPolicyInventoryByMoidRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\tmoid: moid,\n\t}\n}", "func NewDownloadMediaNotFound(body *DownloadMediaNotFoundResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func NewDescribeCacheInstancesRequestWithoutParam() *DescribeCacheInstancesRequest {\n\n return &DescribeCacheInstancesRequest{\n JDCloudRequest: core.JDCloudRequest{\n URL: \"/regions/{regionId}/cacheInstance\",\n Method: \"GET\",\n Header: nil,\n Version: \"v1\",\n },\n }\n}", "func NewGetStoragePureDisksMoidDefault(code int) *GetStoragePureDisksMoidDefault {\n\treturn &GetStoragePureDisksMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetLimitsNotFound() *GetLimitsNotFound {\n\treturn &GetLimitsNotFound{}\n}", "func (a *HyperflexApiService) GetHyperflexIscsiNetworkByMoid(ctx context.Context, moid string) ApiGetHyperflexIscsiNetworkByMoidRequest {\n\treturn ApiGetHyperflexIscsiNetworkByMoidRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\tmoid: moid,\n\t}\n}", "func NewGetSepainstantNotFound() *GetSepainstantNotFound {\n\treturn &GetSepainstantNotFound{}\n}", "func NewGetHyperflexSysConfigPoliciesMoidNotFound() *GetHyperflexSysConfigPoliciesMoidNotFound {\n\treturn &GetHyperflexSysConfigPoliciesMoidNotFound{}\n}", "func (a *HyperflexApiService) GetHyperflexStorageContainerByMoid(ctx context.Context, moid string) ApiGetHyperflexStorageContainerByMoidRequest {\n\treturn ApiGetHyperflexStorageContainerByMoidRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\tmoid: moid,\n\t}\n}", "func NewGetDevicesMacMacNotFound() *GetDevicesMacMacNotFound {\n\treturn &GetDevicesMacMacNotFound{}\n}", "func notFound(resp *ApiResponse, msg string) error {\n resp.StatusCode = http.StatusNotFound\n resp.Message = []byte(msg)\n resp.ErrorMessage = http.StatusText(http.StatusNotFound)\n\n return nil\n}", "func (nsn ErrNoSuchNetwork) NotFound() {}", "func NewGetSearchSearchItemsMoidNotFound() *GetSearchSearchItemsMoidNotFound {\n\treturn &GetSearchSearchItemsMoidNotFound{}\n}", "func NewGetMemoryArraysMoidDefault(code int) *GetMemoryArraysMoidDefault {\n\treturn &GetMemoryArraysMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetNotFound(body *GetNotFoundResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func NewGetNotFound(body *GetNotFoundResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func (une *UnknownNetworkError) NotFound() {}", "func NewGetNiaapiDcnmHweolsMoidOK() *GetNiaapiDcnmHweolsMoidOK {\n\treturn &GetNiaapiDcnmHweolsMoidOK{}\n}", "func (Http) getNotFound(name string) *Http {\n\treturn &Http{\n\t\tCode: http.StatusNotFound,\n\t\tStatus: http.StatusText(http.StatusNotFound),\n\t\tMessage: fmt.Sprintf(\"%s not found\", toUpperFirstChar(name)),\n\t}\n}", "func NewObjectsGetNotFound() *ObjectsGetNotFound {\n\treturn &ObjectsGetNotFound{}\n}", "func noFound(msg string) error {\n\treturn status.Error(codes.NotFound, msg)\n}", "func NewGetNvmfNamespaceSnapshotsNotFound() *GetNvmfNamespaceSnapshotsNotFound {\n\treturn &GetNvmfNamespaceSnapshotsNotFound{}\n}", "func NewFindMaterialByIDDefault(code int) *FindMaterialByIDDefault {\n\treturn &FindMaterialByIDDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewFirmwareStatisticsNotFound(body FirmwareStatisticsNotFoundResponseBody) information.NotFound {\n\tv := information.NotFound(body)\n\treturn v\n}", "func (m *mockAPI) notFound() *expectation {\n\treturn &expectation{code: http.StatusNotFound, body: nil}\n}", "func NewGetSearchTagItemsMoidNotFound() *GetSearchTagItemsMoidNotFound {\n\treturn &GetSearchTagItemsMoidNotFound{}\n}", "func NewFindMaterialByIDNotFound() *FindMaterialByIDNotFound {\n\treturn &FindMaterialByIDNotFound{}\n}", "func NewGetRacksOK() *GetRacksOK {\n\treturn &GetRacksOK{}\n}", "func NotFound(msg string) Error {\n\te := err{msg: msg, code: notFoundCode, group: generic, kind: notFound}\n\treturn &e\n}", "func NewGetV1PullmetricsNotFound() *GetV1PullmetricsNotFound {\n\n\treturn &GetV1PullmetricsNotFound{}\n}", "func NewGetDebugRequestNotFound() *GetDebugRequestNotFound {\n\treturn &GetDebugRequestNotFound{}\n}", "func (o *DcimRacksListParams) WithOuterUnitn(outerUnitn *string) *DcimRacksListParams {\n\to.SetOuterUnitn(outerUnitn)\n\treturn o\n}", "func NewGetRacksIDOK() *GetRacksIDOK {\n\treturn &GetRacksIDOK{}\n}", "func NewGetNiaapiApicSweolsMoidDefault(code int) *GetNiaapiApicSweolsMoidDefault {\n\treturn &GetNiaapiApicSweolsMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewNotFound(a Attributes) error {\n\tname, resource, err := extractResourceName(a)\n\tif err != nil {\n\t\treturn apierrors.NewInternalError(err)\n\t}\n\treturn apierrors.NewNotFound(resource, name)\n}", "func NewGetStoragePureDisksMoidOK() *GetStoragePureDisksMoidOK {\n\treturn &GetStoragePureDisksMoidOK{}\n}", "func NewAllDashboardsNotFound() *AllDashboardsNotFound {\n return &AllDashboardsNotFound{\n }\n}", "func NewGetNiaapiDcnmSweolsMoidDefault(code int) *GetNiaapiDcnmSweolsMoidDefault {\n\treturn &GetNiaapiDcnmSweolsMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewByNamespaceNotFound() *ByNamespaceNotFound {\n\treturn &ByNamespaceNotFound{}\n}", "func NFailedUnits(w http.ResponseWriter) error {\n\tv, err := getProperty(\"NFailedUnits\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tprop := Property{\n\t\tProperty: \"NFailedUnits\",\n\t\tValue: fmt.Sprint(v.Value().(uint32)),\n\t}\n\n\treturn share.JSONResponse(prop, w)\n}", "func NewNotFound(name, group, resource string) error {\n\treturn errors.NewNotFound(schema.GroupResource{Group: group, Resource: resource}, name)\n}", "func NewGetCustomNameByProjectIDNotFound() *GetCustomNameByProjectIDNotFound {\n\treturn &GetCustomNameByProjectIDNotFound{}\n}", "func NewBatchGetWaitEstimatesNotFound(body *BatchGetWaitEstimatesNotFoundResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func NewSystemDetailFindByIDNotFound() *SystemDetailFindByIDNotFound {\n\treturn &SystemDetailFindByIDNotFound{}\n}", "func NewGetNicsBadRequest() *GetNicsBadRequest {\n\treturn &GetNicsBadRequest{}\n}", "func NewDescribeBillSummarysRequestWithoutParam() *DescribeBillSummarysRequest {\n\n return &DescribeBillSummarysRequest{\n JDCloudRequest: core.JDCloudRequest{\n URL: \"/regions/{regionId}/describeBillSummarys\",\n Method: \"POST\",\n Header: nil,\n Version: \"v1\",\n },\n }\n}" ]
[ "0.6873871", "0.6207839", "0.5926978", "0.5767671", "0.5551007", "0.5232645", "0.5211905", "0.5198357", "0.5130382", "0.5066114", "0.5060463", "0.5014112", "0.5006994", "0.5006573", "0.49573246", "0.49514273", "0.49505547", "0.49388602", "0.48527807", "0.4849585", "0.48404494", "0.4819281", "0.4808902", "0.4800759", "0.47348446", "0.47098088", "0.46693045", "0.46672443", "0.46370387", "0.463702", "0.46049893", "0.45914638", "0.45478573", "0.4538731", "0.45281217", "0.451556", "0.4513306", "0.45017886", "0.4476967", "0.44742385", "0.44686878", "0.44595993", "0.44594872", "0.4445581", "0.44438767", "0.44438767", "0.44339597", "0.44263554", "0.4424368", "0.44182166", "0.44133666", "0.44063643", "0.4403042", "0.44010803", "0.43990603", "0.4398166", "0.43968177", "0.43941227", "0.43923226", "0.43787053", "0.4364602", "0.4357499", "0.435674", "0.43556437", "0.43470284", "0.43434757", "0.43423545", "0.43411636", "0.43393165", "0.43393165", "0.43335107", "0.43287677", "0.43281895", "0.43172356", "0.43110678", "0.43088657", "0.43072698", "0.42986184", "0.42803234", "0.42800257", "0.42775407", "0.426522", "0.4258041", "0.42529", "0.4252572", "0.42495707", "0.4245371", "0.42395464", "0.42394388", "0.4238428", "0.42276102", "0.42246088", "0.42220724", "0.4215653", "0.41961426", "0.41927212", "0.41921994", "0.41814777", "0.4177409", "0.41771737" ]
0.7762601
0
NewGetComputeRackUnitsMoidDefault creates a GetComputeRackUnitsMoidDefault with default headers values
func NewGetComputeRackUnitsMoidDefault(code int) *GetComputeRackUnitsMoidDefault { return &GetComputeRackUnitsMoidDefault{ _statusCode: code, } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewGetComputeRackUnitsMoidOK() *GetComputeRackUnitsMoidOK {\n\treturn &GetComputeRackUnitsMoidOK{}\n}", "func NewGetHyperflexServerModelsMoidDefault(code int) *GetHyperflexServerModelsMoidDefault {\n\treturn &GetHyperflexServerModelsMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetComputeRackUnitsMoidNotFound() *GetComputeRackUnitsMoidNotFound {\n\treturn &GetComputeRackUnitsMoidNotFound{}\n}", "func NewGetStoragePureDisksMoidDefault(code int) *GetStoragePureDisksMoidDefault {\n\treturn &GetStoragePureDisksMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetMemoryArraysMoidDefault(code int) *GetMemoryArraysMoidDefault {\n\treturn &GetMemoryArraysMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetHardwareDefault(code int) *GetHardwareDefault {\n\treturn &GetHardwareDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetAdapterHostEthInterfacesMoidDefault(code int) *GetAdapterHostEthInterfacesMoidDefault {\n\treturn &GetAdapterHostEthInterfacesMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetHyperflexClusterStoragePoliciesMoidDefault(code int) *GetHyperflexClusterStoragePoliciesMoidDefault {\n\treturn &GetHyperflexClusterStoragePoliciesMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetEquipmentRackEnclosureSlotsDefault(code int) *GetEquipmentRackEnclosureSlotsDefault {\n\treturn &GetEquipmentRackEnclosureSlotsDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetAdapterHostFcInterfacesMoidDefault(code int) *GetAdapterHostFcInterfacesMoidDefault {\n\treturn &GetAdapterHostFcInterfacesMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewPatchHyperflexServerModelsMoidDefault(code int) *PatchHyperflexServerModelsMoidDefault {\n\treturn &PatchHyperflexServerModelsMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func getDefaults(resourceType voyager.ResourceType, location voyager.ClusterLocation) map[string]interface{} {\n\tswitch resourceType {\n\tcase voyager.ResourceType(\"DynamoDB\"):\n\t\treturn map[string]interface{}{\n\t\t\t\"BackupPeriod\": \"1 hours\",\n\t\t}\n\tcase voyager.ResourceType(\"KubeCompute\"):\n\t\tvar minReplicas int32\n\t\tswitch location.EnvType {\n\t\tcase voyager.EnvTypeProduction, voyager.EnvTypeStaging:\n\t\t\tminReplicas = defaultKubeComputeMinReplicasProd\n\t\tdefault:\n\t\t\tminReplicas = defaultKubeComputeMinReplicas\n\t\t}\n\t\treturn map[string]interface{}{\n\t\t\t\"Scaling\": map[string]interface{}{\n\t\t\t\t\"MinReplicas\": minReplicas,\n\t\t\t\t\"MaxReplicas\": defaultKubeComputeMaxReplicas,\n\t\t\t\t\"Metrics\": []map[string]interface{}{\n\t\t\t\t\t{\n\t\t\t\t\t\t\"Type\": \"Resource\",\n\t\t\t\t\t\t\"Resource\": map[string]interface{}{\n\t\t\t\t\t\t\t\"Name\": \"cpu\",\n\t\t\t\t\t\t\t\"TargetAverageUtilization\": defaultKubeComputeResourceMetricTargetUtilization,\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t\t\"Container\": map[string]interface{}{\n\t\t\t\t\"ImagePullPolicy\": defaultKubeComputeImagePullPolicy,\n\t\t\t\t\"LivenessProbe\": map[string]interface{}{\n\t\t\t\t\t\"TimeoutSeconds\": defaultKubeComputeProbeTimeoutSeconds,\n\t\t\t\t\t\"PeriodSeconds\": defaultKubeComputeProbePeriodSeconds,\n\t\t\t\t\t\"SuccessThreshold\": defaultKubeComputeProbeSuccessThreshold,\n\t\t\t\t\t\"FailureThreshold\": defaultKubeComputeProbeFailureThreshold,\n\t\t\t\t\t\"HTTPGet\": map[string]interface{}{\n\t\t\t\t\t\t\"Path\": defaultKubeComputeHTTPGetPath,\n\t\t\t\t\t\t\"Scheme\": defaultKubeComputeHTTPGetScheme,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\t\"ReadinessProbe\": map[string]interface{}{\n\t\t\t\t\t\"TimeoutSeconds\": defaultKubeComputeProbeTimeoutSeconds,\n\t\t\t\t\t\"PeriodSeconds\": defaultKubeComputeProbePeriodSeconds,\n\t\t\t\t\t\"SuccessThreshold\": defaultKubeComputeProbeSuccessThreshold,\n\t\t\t\t\t\"FailureThreshold\": defaultKubeComputeProbeFailureThreshold,\n\t\t\t\t\t\"HTTPGet\": map[string]interface{}{\n\t\t\t\t\t\t\"Path\": defaultKubeComputeHTTPGetPath,\n\t\t\t\t\t\t\"Scheme\": defaultKubeComputeHTTPGetScheme,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\t\"Resources\": map[string]interface{}{\n\t\t\t\t\t\"Requests\": map[string]interface{}{\n\t\t\t\t\t\t\"cpu\": defaultKubeComputeResourceRequestCPU,\n\t\t\t\t\t\t\"memory\": defaultKubeComputeResourceRequestMemory,\n\t\t\t\t\t},\n\t\t\t\t\t\"Limits\": map[string]interface{}{\n\t\t\t\t\t\t\"cpu\": defaultKubeComputeResourceLimitCPU,\n\t\t\t\t\t\t\"memory\": defaultKubeComputeResourceLimitMemory,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t\t\"Port\": map[string]interface{}{\n\t\t\t\t\"Protocol\": defaultKubeComputeProtocol,\n\t\t\t},\n\t\t}\n\tcase voyager.ResourceType(\"KubeIngress\"):\n\t\treturn map[string]interface{}{\n\t\t\t\"timeoutSeconds\": 60,\n\t\t}\n\tdefault:\n\t\treturn map[string]interface{}{}\n\t}\n\n}", "func NewGetEquipmentIoCardsDefault(code int) *GetEquipmentIoCardsDefault {\n\treturn &GetEquipmentIoCardsDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetStoragePureControllersMoidDefault(code int) *GetStoragePureControllersMoidDefault {\n\treturn &GetStoragePureControllersMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetVMVolumeDefault(code int) *GetVMVolumeDefault {\n\tif code <= 0 {\n\t\tcode = 500\n\t}\n\n\treturn &GetVMVolumeDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewPostEquipmentIoExpandersMoidDefault(code int) *PostEquipmentIoExpandersMoidDefault {\n\treturn &PostEquipmentIoExpandersMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewPostMemoryArraysMoidDefault(code int) *PostMemoryArraysMoidDefault {\n\treturn &PostMemoryArraysMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetHardwareFcportsDefault(code int) *GetHardwareFcportsDefault {\n\treturn &GetHardwareFcportsDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetEtherPhysicalPortsDefault(code int) *GetEtherPhysicalPortsDefault {\n\treturn &GetEtherPhysicalPortsDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewPostInventoryGenericInventoryHoldersMoidDefault(code int) *PostInventoryGenericInventoryHoldersMoidDefault {\n\treturn &PostInventoryGenericInventoryHoldersMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetManagementControllersMoidDefault(code int) *GetManagementControllersMoidDefault {\n\treturn &GetManagementControllersMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetHyperflexSysConfigPoliciesMoidDefault(code int) *GetHyperflexSysConfigPoliciesMoidDefault {\n\treturn &GetHyperflexSysConfigPoliciesMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func (in *KubeadmControlPlane) Default() {\n\tif in.Spec.Replicas == nil {\n\t\treplicas := int32(1)\n\t\tin.Spec.Replicas = &replicas\n\t}\n\n\tif in.Spec.InfrastructureTemplate.Namespace == \"\" {\n\t\tin.Spec.InfrastructureTemplate.Namespace = in.Namespace\n\t}\n}", "func NewPatchEquipmentIoCardsMoidDefault(code int) *PatchEquipmentIoCardsMoidDefault {\n\treturn &PatchEquipmentIoCardsMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func newDefaultContainerConfig() ContainerConfig {\n\treturn ContainerConfig{\n\t\tCPU: newMinMaxAllocation(),\n\t\tMemory: newMinMaxAllocation(),\n\t\tBlockRead: newMinMaxAllocation(),\n\t\tBlockWrite: newMinMaxAllocation(),\n\t\tNetworkRx: newMinMaxAllocation(),\n\t\tNetworkTx: newMinMaxAllocation(),\n\t}\n}", "func NewGetAvailabilityZonesUUIDDefault(code int) *GetAvailabilityZonesUUIDDefault {\n\treturn &GetAvailabilityZonesUUIDDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetNodesDefault(code int) *GetNodesDefault {\n\treturn &GetNodesDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewPostHyperflexClusterProfilesMoidDefault(code int) *PostHyperflexClusterProfilesMoidDefault {\n\treturn &PostHyperflexClusterProfilesMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewPatchEquipmentIoExpandersMoidDefault(code int) *PatchEquipmentIoExpandersMoidDefault {\n\treturn &PatchEquipmentIoExpandersMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func (r *Unit) Default() {\n\tunitlog.Info(\"default\", \"name\", r.Name)\n\n\t// TODO(user): fill in your defaulting logic.\n\t// 这里可以加入一些Unit 结构体对象初始化之前的一些默认的逻辑,比如给一些字段填充默认值\n\n\t// default replicas set to 1\n\tunitlog.Info(\"default\", \"name\", r.Name)\n\n\tif r.Spec.Replicas == nil {\n\t\tdefaultReplicas := int32(1)\n\t\tr.Spec.Replicas = &defaultReplicas\n\t}\n\n\t// add default selector label\n\tlabelMap := make(map[string]string, 1)\n\tlabelMap[\"app\"] = r.Name\n\tr.Spec.Selector = &metav1.LabelSelector{\n\t\tMatchLabels: labelMap,\n\t}\n\n\t// add default template label\n\tr.Spec.Template.Labels = labelMap\n\n\tr.Status.LastUpdateTime = metav1.Now()\n\n\t// 当然,还可以根据需求加一些适合在初始化时做的逻辑,例如为pod注入sidecar\n\t// ...\n\n}", "func NewGetHyperflexNodeProfilesDefault(code int) *GetHyperflexNodeProfilesDefault {\n\treturn &GetHyperflexNodeProfilesDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetLargeScreenDataRequestWithoutParam() *GetLargeScreenDataRequest {\n\n return &GetLargeScreenDataRequest{\n JDCloudRequest: core.JDCloudRequest{\n URL: \"/regions/{regionId}/getData\",\n Method: \"GET\",\n Header: nil,\n Version: \"v1\",\n },\n }\n}", "func (o *ColumnFamilyMetricsTotalDiskSpaceUsedByNameGetParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func NewGetNiaapiDcnmHweolsMoidDefault(code int) *GetNiaapiDcnmHweolsMoidDefault {\n\treturn &GetNiaapiDcnmHweolsMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewDcimRackRolesBulkPartialUpdateDefault(code int) *DcimRackRolesBulkPartialUpdateDefault {\n\treturn &DcimRackRolesBulkPartialUpdateDefault{\n\t\t_statusCode: code,\n\t}\n}", "func (o *GetRackTopoesParams) SetDefaults() {\n\tvar (\n\t\tcontentLanguageDefault = string(\"en-US\")\n\t)\n\n\tval := GetRackTopoesParams{\n\t\tContentLanguage: &contentLanguageDefault,\n\t}\n\n\tval.timeout = o.timeout\n\tval.Context = o.Context\n\tval.HTTPClient = o.HTTPClient\n\t*o = val\n}", "func NewPostHyperflexHxdpVersionsMoidDefault(code int) *PostHyperflexHxdpVersionsMoidDefault {\n\treturn &PostHyperflexHxdpVersionsMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetObmsLibraryIdentifierDefault(code int) *GetObmsLibraryIdentifierDefault {\n\treturn &GetObmsLibraryIdentifierDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetTerminalAuditLogsMoidDefault(code int) *GetTerminalAuditLogsMoidDefault {\n\treturn &GetTerminalAuditLogsMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewNvmeSubsystemMapCreateDefault(code int) *NvmeSubsystemMapCreateDefault {\n\treturn &NvmeSubsystemMapCreateDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetSearchTagItemsMoidDefault(code int) *GetSearchTagItemsMoidDefault {\n\treturn &GetSearchTagItemsMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetTimezoneRegionDefault(code int) *GetTimezoneRegionDefault {\n\treturn &GetTimezoneRegionDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetHostGroupsDefault(code int) *GetHostGroupsDefault {\n\treturn &GetHostGroupsDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewPatchApplianceDiagSettingsMoidDefault(code int) *PatchApplianceDiagSettingsMoidDefault {\n\treturn &PatchApplianceDiagSettingsMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewPatchVnicEthNetworkPoliciesMoidDefault(code int) *PatchVnicEthNetworkPoliciesMoidDefault {\n\treturn &PatchVnicEthNetworkPoliciesMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetSummarySystemDefault(code int) *GetSummarySystemDefault {\n\treturn &GetSummarySystemDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewDefault(m map[string]interface{}) (share.Manager, error) {\n\tc := &config{}\n\tif err := mapstructure.Decode(m, c); err != nil {\n\t\terr = errors.Wrap(err, \"error creating a new manager\")\n\t\treturn nil, err\n\t}\n\n\ts, err := metadata.NewCS3Storage(c.GatewayAddr, c.ProviderAddr, c.ServiceUserID, c.ServiceUserIdp, c.MachineAuthAPIKey)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tindexer := indexer.CreateIndexer(s)\n\n\tclient, err := pool.GetGatewayServiceClient(c.GatewayAddr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn New(client, s, indexer)\n}", "func NewGetVnicEthAdapterPoliciesMoidDefault(code int) *GetVnicEthAdapterPoliciesMoidDefault {\n\treturn &GetVnicEthAdapterPoliciesMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetNiaapiApicSweolsMoidDefault(code int) *GetNiaapiApicSweolsMoidDefault {\n\treturn &GetNiaapiApicSweolsMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func (machine *VirtualMachine) defaultImpl() { machine.defaultAzureName() }", "func NewGetNetworkPoolsDefault(code int) *GetNetworkPoolsDefault {\n\treturn &GetNetworkPoolsDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetNiaapiDcnmSweolsMoidDefault(code int) *GetNiaapiDcnmSweolsMoidDefault {\n\treturn &GetNiaapiDcnmSweolsMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetRacksUnauthorized() *GetRacksUnauthorized {\n\treturn &GetRacksUnauthorized{}\n}", "func (c *TestClusterGKE) Default() {\n\tif c.Name != \"\" { // avoid loging internal annonymous objects\n\t\tlog.Info(\"applying defaults\", \"name\", c.Name, \"namespace\", c.Namespace)\n\t}\n\n\tif c.Spec.Project == nil {\n\t\tc.Spec.Project = new(string)\n\t\t*c.Spec.Project = \"cilium-ci\"\n\t}\n\n\tif c.Spec.ConfigTemplate == nil {\n\t\tc.Spec.ConfigTemplate = new(string)\n\t\t*c.Spec.ConfigTemplate = \"basic\"\n\t}\n\n\tif c.Spec.Location == nil {\n\t\tc.Spec.Location = new(string)\n\t\t*c.Spec.Location = \"europe-west2-b\"\n\t}\n\n\tif c.Spec.Region == nil {\n\t\tc.Spec.Region = new(string)\n\t\t*c.Spec.Region = \"europe-west2\"\n\t}\n\n\tif c.Spec.JobSpec != nil {\n\t\tif c.Spec.JobSpec.Runner == nil {\n\t\t\tc.Spec.JobSpec.Runner = &TestClusterGKEJobRunnerSpec{}\n\t\t}\n\n\t\tif c.Spec.JobSpec.Runner.Image == nil {\n\t\t\tc.Spec.JobSpec.Runner.Image = new(string)\n\t\t\t*c.Spec.JobSpec.Runner.Image = \"quay.io/isovalent/gke-test-cluster-gcloud:803ff83d3786eb38ef05c95768060b0c7ae0fc4d\"\n\t\t}\n\n\t\tif c.Spec.JobSpec.Runner.InitImage == nil {\n\t\t\tc.Spec.JobSpec.Runner.InitImage = new(string)\n\t\t\t*c.Spec.JobSpec.Runner.InitImage = \"quay.io/isovalent/gke-test-cluster-initutil:854733411778d633350adfa1ae66bf11ba658a3f\"\n\t\t}\n\t}\n\n\tif c.Spec.MachineType == nil {\n\t\tc.Spec.MachineType = new(string)\n\t\t*c.Spec.MachineType = \"n1-standard-4\"\n\t}\n\n\tif c.Spec.Nodes == nil {\n\t\tc.Spec.Nodes = new(int)\n\t\t*c.Spec.Nodes = 2\n\t}\n}", "func GetDefaultOsOptions(providerType string) tkgconfigbom.OSInfo {\n\tswitch providerType {\n\tcase constants.InfrastructureProviderVSphere:\n\t\treturn tkgconfigbom.OSInfo{Name: \"ubuntu\", Version: \"20.04\", Arch: \"amd64\"}\n\n\tcase constants.InfrastructureProviderAWS:\n\t\treturn tkgconfigbom.OSInfo{Name: \"ubuntu\", Version: \"20.04\", Arch: \"amd64\"}\n\n\tcase constants.InfrastructureProviderAzure:\n\t\treturn tkgconfigbom.OSInfo{Name: \"ubuntu\", Version: \"20.04\", Arch: \"amd64\"}\n\t}\n\treturn tkgconfigbom.OSInfo{}\n}", "func getDefaultMachineVolumes() []string {\n\treturn []string{}\n}", "func DefaultRequest() Request {\n\treturn Request{\n\t\tName: \"\",\n\n\t\tOwner: \"\",\n\n\t\tKubernetesVersion: \"\",\n\n\t\tMasters: []request.Master{},\n\t\tVault: request.DefaultVault(),\n\t\tWorkers: []request.Worker{},\n\t}\n}", "func NewRackUnitPersonalityWithDefaults() *RackUnitPersonality {\n\tthis := RackUnitPersonality{}\n\tvar classId string = \"rack.UnitPersonality\"\n\tthis.ClassId = classId\n\tvar objectType string = \"rack.UnitPersonality\"\n\tthis.ObjectType = objectType\n\treturn &this\n}", "func NewDescribeDefault(code int) *DescribeDefault {\n\tif code <= 0 {\n\t\tcode = 500\n\t}\n\n\treturn &DescribeDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetServerConfigImportsMoidDefault(code int) *GetServerConfigImportsMoidDefault {\n\treturn &GetServerConfigImportsMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetNiaapiDcnmCcoPostsMoidDefault(code int) *GetNiaapiDcnmCcoPostsMoidDefault {\n\treturn &GetNiaapiDcnmCcoPostsMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewPatchHyperflexSoftwareVersionPoliciesMoidDefault(code int) *PatchHyperflexSoftwareVersionPoliciesMoidDefault {\n\treturn &PatchHyperflexSoftwareVersionPoliciesMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetSearchSearchItemsMoidDefault(code int) *GetSearchSearchItemsMoidDefault {\n\treturn &GetSearchSearchItemsMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetNetworkInterfacesDefault(code int) *GetNetworkInterfacesDefault {\n\treturn &GetNetworkInterfacesDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewPatchGraphicsControllersMoidDefault(code int) *PatchGraphicsControllersMoidDefault {\n\treturn &PatchGraphicsControllersMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func Default() *oceanv1.OceanToolkit {\n\treturn &oceanv1.OceanToolkit{\n\t\tTypeMeta: metav1.TypeMeta{\n\t\t\tKind: \"ocean.spot.io.Toolkit\",\n\t\t\tAPIVersion: \"v1\",\n\t\t},\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tName: \"ocean-toolkit\",\n\t\t\tNamespace: NamespaceOceanSystem,\n\t\t},\n\t\tSpec: oceanv1.OceanToolkitSpec{\n\t\t\tComponents: map[string]*oceanv1.OceanToolkitComponent{\n\t\t\t\t\"controller\": {\n\t\t\t\t\tEnabled: true,\n\t\t\t\t},\n\t\t\t\t\"operator\": {\n\t\t\t\t\tEnabled: true,\n\t\t\t\t},\n\t\t\t\t\"metrics-server\": {\n\t\t\t\t\tEnabled: true,\n\t\t\t\t},\n\t\t\t},\n\t\t\tValues: map[string]interface{}{\n\t\t\t\t\"spot\": map[string]interface{}{\n\t\t\t\t\t\"token\": os.Getenv(credentials.EnvCredentialsVarToken),\n\t\t\t\t\t\"account\": os.Getenv(credentials.EnvCredentialsVarAccount),\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n}", "func NewDefaultClient() QueueClient {\n\treturn &inMemoryQueue{queues: make(map[string][]string)}\n}", "func (client IdentityClient) createTagDefault(ctx context.Context, request common.OCIRequest, binaryReqBody *common.OCIReadSeekCloser, extraHeaders map[string]string) (common.OCIResponse, error) {\n\n\thttpRequest, err := request.HTTPRequest(http.MethodPost, \"/tagDefaults\", binaryReqBody, extraHeaders)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar response CreateTagDefaultResponse\n\tvar httpResponse *http.Response\n\thttpResponse, err = client.Call(ctx, &httpRequest)\n\tdefer common.CloseBodyIfValid(httpResponse)\n\tresponse.RawResponse = httpResponse\n\tif err != nil {\n\t\treturn response, err\n\t}\n\n\terr = common.UnmarshalResponse(httpResponse, &response)\n\treturn response, err\n}", "func (record *PrivateDnsZonesSRVRecord) defaultImpl() { record.defaultAzureName() }", "func (f *FaasController) DefaultSpec() interface{} {\n\treturn &spec.Admin{\n\t\tSyncInterval: \"10s\",\n\t\tProvider: spec.ProviderKnative,\n\t\tKnative: &spec.Knative{\n\t\t\tNamespace: \"default\",\n\t\t\tTimeout: \"2s\",\n\t\t},\n\t}\n}", "func (o KubernetesClusterNodePoolLinuxOsConfigSysctlConfigOutput) NetCoreWmemDefault() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v KubernetesClusterNodePoolLinuxOsConfigSysctlConfig) *int { return v.NetCoreWmemDefault }).(pulumi.IntPtrOutput)\n}", "func TestApplyDefaultMachineSetHardwareSpec(t *testing.T) {\n\n\tawsSpec := func(amiName, instanceType string) *clusteroperator.MachineSetHardwareSpec {\n\t\treturn &clusteroperator.MachineSetHardwareSpec{\n\t\t\tAWS: &clusteroperator.MachineSetAWSHardwareSpec{\n\t\t\t\tAMIName: amiName,\n\t\t\t\tInstanceType: instanceType,\n\t\t\t},\n\t\t}\n\t}\n\tcases := []struct {\n\t\tname string\n\t\tdefaultSpec *clusteroperator.MachineSetHardwareSpec\n\t\tspecific *clusteroperator.MachineSetHardwareSpec\n\t\texpected *clusteroperator.MachineSetHardwareSpec\n\t}{\n\t\t{\n\t\t\tname: \"no default\",\n\t\t\tdefaultSpec: nil,\n\t\t\tspecific: awsSpec(\"base-ami\", \"large-instance\"),\n\t\t\texpected: awsSpec(\"base-ami\", \"large-instance\"),\n\t\t},\n\t\t{\n\t\t\tname: \"only default\",\n\t\t\tdefaultSpec: awsSpec(\"base-ami\", \"small-instance\"),\n\t\t\tspecific: &clusteroperator.MachineSetHardwareSpec{},\n\t\t\texpected: awsSpec(\"base-ami\", \"small-instance\"),\n\t\t},\n\t\t{\n\t\t\tname: \"override default\",\n\t\t\tdefaultSpec: awsSpec(\"base-ami\", \"large-instance\"),\n\t\t\tspecific: awsSpec(\"\", \"specific-instance\"),\n\t\t\texpected: awsSpec(\"base-ami\", \"specific-instance\"),\n\t\t},\n\t\t{\n\t\t\tname: \"partial default\",\n\t\t\tdefaultSpec: awsSpec(\"base-ami\", \"\"),\n\t\t\tspecific: awsSpec(\"\", \"large-instance\"),\n\t\t\texpected: awsSpec(\"base-ami\", \"large-instance\"),\n\t\t},\n\t}\n\n\tfor _, tc := range cases {\n\t\tresult, err := applyDefaultMachineSetHardwareSpec(tc.specific, tc.defaultSpec)\n\t\tif err != nil {\n\t\t\tt.Errorf(\"%s: unexpected error: %v\", tc.name, err)\n\t\t\tcontinue\n\t\t}\n\t\tif !reflect.DeepEqual(result, tc.expected) {\n\t\t\tt.Errorf(\"%s: unexpected result. Expected: %v, Got: %v\", tc.name, tc.expected, result)\n\t\t}\n\t}\n}", "func (o *GetComputeRackUnitsMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (client IdentityClient) getTagDefault(ctx context.Context, request common.OCIRequest, binaryReqBody *common.OCIReadSeekCloser, extraHeaders map[string]string) (common.OCIResponse, error) {\n\n\thttpRequest, err := request.HTTPRequest(http.MethodGet, \"/tagDefaults/{tagDefaultId}\", binaryReqBody, extraHeaders)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar response GetTagDefaultResponse\n\tvar httpResponse *http.Response\n\thttpResponse, err = client.Call(ctx, &httpRequest)\n\tdefer common.CloseBodyIfValid(httpResponse)\n\tresponse.RawResponse = httpResponse\n\tif err != nil {\n\t\treturn response, err\n\t}\n\n\terr = common.UnmarshalResponse(httpResponse, &response)\n\treturn response, err\n}", "func NewGetNetworkSubnetsDefault(code int) *GetNetworkSubnetsDefault {\n\treturn &GetNetworkSubnetsDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetIOAUsersDefault(code int) *GetIOAUsersDefault {\n\treturn &GetIOAUsersDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewIpspacesGetDefault(code int) *IpspacesGetDefault {\n\treturn &IpspacesGetDefault{\n\t\t_statusCode: code,\n\t}\n}", "func (fp *ListRegionsResponse_FieldTerminalPath) GetDefault() interface{} {\n\tswitch fp.selector {\n\tcase ListRegionsResponse_FieldPathSelectorRegions:\n\t\treturn ([]*region.Region)(nil)\n\tcase ListRegionsResponse_FieldPathSelectorPrevPageToken:\n\t\treturn (*region.PagerCursor)(nil)\n\tcase ListRegionsResponse_FieldPathSelectorNextPageToken:\n\t\treturn (*region.PagerCursor)(nil)\n\tcase ListRegionsResponse_FieldPathSelectorCurrentOffset:\n\t\treturn int32(0)\n\tcase ListRegionsResponse_FieldPathSelectorTotalResultsCount:\n\t\treturn int32(0)\n\tdefault:\n\t\tpanic(fmt.Sprintf(\"Invalid selector for ListRegionsResponse: %d\", fp.selector))\n\t}\n}", "func NewGetStorageFlexUtilControllersDefault(code int) *GetStorageFlexUtilControllersDefault {\n\treturn &GetStorageFlexUtilControllersDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetStoragePureVolumesDefault(code int) *GetStoragePureVolumesDefault {\n\treturn &GetStoragePureVolumesDefault{\n\t\t_statusCode: code,\n\t}\n}", "func (o KubernetesClusterDefaultNodePoolLinuxOsConfigSysctlConfigOutput) NetCoreWmemDefault() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v KubernetesClusterDefaultNodePoolLinuxOsConfigSysctlConfig) *int { return v.NetCoreWmemDefault }).(pulumi.IntPtrOutput)\n}", "func NewDescribeClustersDefault(code int) *DescribeClustersDefault {\n\tif code <= 0 {\n\t\tcode = 500\n\t}\n\n\treturn &DescribeClustersDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetSdcardPoliciesMoidDefault(code int) *GetSdcardPoliciesMoidDefault {\n\treturn &GetSdcardPoliciesMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewFindConfigMemtableOffheapSpaceInMbDefault(code int) *FindConfigMemtableOffheapSpaceInMbDefault {\n\treturn &FindConfigMemtableOffheapSpaceInMbDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewGetHyperflexServerFirmwareVersionsDefault(code int) *GetHyperflexServerFirmwareVersionsDefault {\n\treturn &GetHyperflexServerFirmwareVersionsDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewDefaultRawFileSystem() RawFileSystem {\n\treturn (*defaultRawFileSystem)(nil)\n}", "func (o KafkaMirrorMakerOutput) DiskSpaceDefault() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *KafkaMirrorMaker) pulumi.StringOutput { return v.DiskSpaceDefault }).(pulumi.StringOutput)\n}", "func NewGetNodeUpgradesDefault(code int) *GetNodeUpgradesDefault {\n\treturn &GetNodeUpgradesDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewPatchManagementEntitiesMoidDefault(code int) *PatchManagementEntitiesMoidDefault {\n\treturn &PatchManagementEntitiesMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func GetDefaultHeaders() map[string]string {\n\treturn map[string]string{\n\t\t\"content-type\": \"application/json\",\n\t}\n}", "func NewListOfDevicesDefault(code int) *ListOfDevicesDefault {\n\tif code <= 0 {\n\t\tcode = 500\n\t}\n\n\treturn &ListOfDevicesDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewPostNodesIdentifierObmIdentifyDefault(code int) *PostNodesIdentifierObmIdentifyDefault {\n\treturn &PostNodesIdentifierObmIdentifyDefault{\n\t\t_statusCode: code,\n\t}\n}", "func newKubeNodeStatusAllocatableMemoryBytesMetric(fqname, node string, value float64) KubeNodeStatusAllocatableMemoryBytesMetric {\n\treturn KubeNodeStatusAllocatableMemoryBytesMetric{\n\t\tfqName: fqname,\n\t\thelp: \"kube_node_status_allocatable_memory_bytes node allocatable memory in bytes\",\n\t\tnode: node,\n\t\tvalue: value,\n\t}\n}", "func NewDeleteFirmwareUpgradesMoidDefault(code int) *DeleteFirmwareUpgradesMoidDefault {\n\treturn &DeleteFirmwareUpgradesMoidDefault{\n\t\t_statusCode: code,\n\t}\n}", "func newKubeNodeStatusCapacityMemoryBytesMetric(fqname string, node string, bytes float64) KubeNodeStatusCapacityMemoryBytesMetric {\n\treturn KubeNodeStatusCapacityMemoryBytesMetric{\n\t\tfqName: fqname,\n\t\thelp: \"kube_node_status_capacity_memory_bytes Node Capacity Memory Bytes\",\n\t\tnode: node,\n\t\tbytes: bytes,\n\t}\n}", "func (o M3DbOutput) DiskSpaceDefault() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *M3Db) pulumi.StringOutput { return v.DiskSpaceDefault }).(pulumi.StringOutput)\n}", "func (in *ManagedCluster) Default() {\n}", "func NewRegisterBareMetalHostDefault(code int) *RegisterBareMetalHostDefault {\n\treturn &RegisterBareMetalHostDefault{\n\t\t_statusCode: code,\n\t}\n}", "func (o *DcimRacksListParams) SetDefaults() {\n\t// no default values defined for this parameter\n}", "func newDefaultHTTPClient() (cev2.Client, error) {\n\tp, err := cev2.NewHTTP()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn cev2.NewClientObserved(p,\n\t\tcev2.WithUUIDs(),\n\t\tcev2.WithTimeNow(),\n\t\tcev2.WithTracePropagation,\n\t)\n}" ]
[ "0.6030046", "0.57353795", "0.5700413", "0.53284496", "0.52793485", "0.5262371", "0.52255535", "0.51287174", "0.50712514", "0.5070867", "0.5001611", "0.49926227", "0.4939943", "0.4937771", "0.48900813", "0.48466387", "0.48183927", "0.4815518", "0.48105782", "0.4809492", "0.48088977", "0.4799291", "0.47958612", "0.47734654", "0.4758105", "0.4738616", "0.4731484", "0.47108716", "0.4702862", "0.46987152", "0.46854308", "0.46730894", "0.46659487", "0.46630263", "0.46553463", "0.4654183", "0.46413788", "0.46350512", "0.4617842", "0.45921454", "0.45837086", "0.45694134", "0.45622316", "0.45571795", "0.4555527", "0.45507368", "0.45223704", "0.4514345", "0.45092297", "0.45079818", "0.45058143", "0.44795102", "0.44785285", "0.4477025", "0.44681945", "0.44331104", "0.44207212", "0.44148093", "0.44102693", "0.4408998", "0.44084078", "0.43992662", "0.43919513", "0.4390891", "0.43908158", "0.43903482", "0.43845302", "0.4379572", "0.43730694", "0.43581808", "0.4354143", "0.4342764", "0.43359217", "0.4335664", "0.4333414", "0.43312746", "0.4330135", "0.4328438", "0.43164334", "0.43134058", "0.43088108", "0.43083587", "0.42945457", "0.42930996", "0.42879018", "0.42873794", "0.4285788", "0.42841077", "0.42835265", "0.42796797", "0.42769137", "0.42725518", "0.42647678", "0.42634073", "0.4255637", "0.42530933", "0.42477602", "0.4243992", "0.42389166", "0.42374992" ]
0.81235754
0
Code gets the status code for the get compute rack units moid default response
func (o *GetComputeRackUnitsMoidDefault) Code() int { return o._statusCode }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *GetHyperflexServerModelsMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetHardwareDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetMemoryArraysMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *NvmeSubsystemMapCreateDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetStorageFlexUtilControllersDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *RegisterBareMetalHostDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetNiaapiDcnmHweolsMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetStoragePureDisksMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetManagementControllersMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetNiaapiApicSweolsMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetStoragePureControllersMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (bstr BlobsSetTierResponse) StatusCode() int {\n\treturn bstr.rawResponse.StatusCode\n}", "func (o *NvmeSubsystemMapCollectionGetDefault) Code() int {\n\treturn o._statusCode\n}", "func (ccr ContainersCreateResponse) StatusCode() int {\n\treturn ccr.rawResponse.StatusCode\n}", "func (o *GetEquipmentIoCardsDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetEquipmentRackEnclosureSlotsDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PostEquipmentIoExpandersMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *MetroclusterSvmGetDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *DetermineDatalakeDataSizesDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchGraphicsControllersMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetNiaapiDcnmSweolsMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetHyperflexClusterStoragePoliciesMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchEquipmentIoExpandersMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchEquipmentIoCardsMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PostHyperflexClusterProfilesMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *DescribeServerOK) Code() int {\n\treturn 200\n}", "func (o *WebSvmGetDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetMachineDeploymentDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *IpspacesGetDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetEtherPhysicalPortsDefault) Code() int {\n\treturn o._statusCode\n}", "func (cgpr ContainersGetPropertiesResponse) StatusCode() int {\n\treturn cgpr.rawResponse.StatusCode\n}", "func (o *ObjectsGetOK) Code() int {\n\treturn 200\n}", "func (o *PatchHyperflexServerModelsMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (csmr ContainersSetMetadataResponse) StatusCode() int {\n\treturn csmr.rawResponse.StatusCode\n}", "func (o *GetStoragePureVolumesDefault) Code() int {\n\treturn o._statusCode\n}", "func (lcr ListContainersResponse) StatusCode() int {\n\treturn lcr.rawResponse.StatusCode\n}", "func (o *FindConfigMemtableOffheapSpaceInMbDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *DescribeDNSServerOK) Code() int {\n\treturn 200\n}", "func (o *PostThermalSimulationDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetAvailabilityZonesUUIDDefault) Code() int {\n\treturn o._statusCode\n}", "func (pbrr PageBlobsResizeResponse) StatusCode() int {\n\treturn pbrr.rawResponse.StatusCode\n}", "func (o *GetAdapterHostEthInterfacesMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (sss StorageServiceStats) StatusCode() int {\n\treturn sss.rawResponse.StatusCode\n}", "func (o *StorageServicePendingRangeByKeyspaceGetDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PostInventoryGenericInventoryHoldersMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PostClustersMulticlusterConfigDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetQuotaQuotaDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *NfsCollectionPerformanceMetricsGetDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *CreateAntivirusServerDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetSummarySystemDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetNodesDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetDeviceDetailsV2OK) Code() int {\n\treturn 200\n}", "func (o *GetMalQueryMetadataV1Default) Code() int {\n\treturn o._statusCode\n}", "func (o *PostMemoryArraysMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (bsmr BlobsSetMetadataResponse) StatusCode() int {\n\treturn bsmr.rawResponse.StatusCode\n}", "func (o *AddClusterV5Default) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchManagementEntitiesMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetSearchTagItemsMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (s *UnsupportedAvailabilityZone) StatusCode() int {\n\treturn s.RespMetadata.StatusCode\n}", "func (o *NvmeServiceCollectionGetDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetTimezoneRegionDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetHyperflexNodeProfilesDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *QosWorkloadCollectionGetDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PostHyperflexHxdpVersionsMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetIaasUcsdInfosDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetPartGeometryURLDefault) Code() int {\n\treturn o._statusCode\n}", "func (s *FileSystemInUse) StatusCode() int {\n\treturn s.RespMetadata.StatusCode\n}", "func (o *NodesGetDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *DeleteApplianceRestoresMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetObmsLibraryIdentifierDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetHardwareFcportsDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetHyperflexSysConfigPoliciesMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *UtilityServiceReadyDefault) Code() int {\n\treturn o._statusCode\n}", "func (cfr CreateFilesystemResponse) StatusCode() int {\n\treturn cfr.rawResponse.StatusCode\n}", "func (o *GetSdcardPoliciesMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PatchMachineConfigurationDefault) Code() int {\n\treturn o._statusCode\n}", "func (lbhr ListBlobsHierarchyResponse) StatusCode() int {\n\treturn lbhr.rawResponse.StatusCode\n}", "func (o *GetClusterOidcDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetAPIFeaturesNameDefault) Code() int {\n\treturn o._statusCode\n}", "func (cblr ContainersBreakLeaseResponse) StatusCode() int {\n\treturn cblr.rawResponse.StatusCode\n}", "func (s *InsufficientThroughputCapacity) StatusCode() int {\n\treturn s.RespMetadata.StatusCode\n}", "func (abcr AppendBlobsCreateResponse) StatusCode() int {\n\treturn abcr.rawResponse.StatusCode\n}", "func (o *ObjectsGetInternalServerError) Code() int {\n\treturn 500\n}", "func (o *GetHyperflexServerFirmwareVersionsDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetAdapterHostFcInterfacesMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *UpdateMTOPostCounselingInformationOK) Code() int {\n\treturn 200\n}", "func (o *UpdateDeviceTagsDefault) Code() int {\n\treturn o._statusCode\n}", "func (ssp StorageServiceProperties) StatusCode() int {\n\treturn ssp.rawResponse.StatusCode\n}", "func (o *GetSearchSearchItemsMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *DeleteFirmwareUpgradesMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetModelRegistryDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetServicesServiceIDOperationsDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *UpdateAntivirusServerDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *FindMaterialByIDDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *PostApplianceUpgradesMoidDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *DeletePorositySimulationDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GetVersionOK) Code() int {\n\treturn 200\n}", "func (o *QueryAWSAccountsForIDsInternalServerError) Code() int {\n\treturn 500\n}", "func (cclr ContainersChangeLeaseResponse) StatusCode() int {\n\treturn cclr.rawResponse.StatusCode\n}", "func (o *AddAPIDefault) Code() int {\n\treturn o._statusCode\n}" ]
[ "0.71472275", "0.71210796", "0.69848347", "0.6958648", "0.6957823", "0.69504213", "0.69414854", "0.6933363", "0.69294435", "0.6923509", "0.6920206", "0.689982", "0.68875915", "0.6860156", "0.6856944", "0.68548477", "0.6851181", "0.68449444", "0.6826638", "0.68255234", "0.6815843", "0.68142694", "0.6803688", "0.6801363", "0.67817724", "0.67610955", "0.67524415", "0.674775", "0.6741059", "0.6737007", "0.6715849", "0.6715098", "0.67140335", "0.6712856", "0.6710929", "0.6706132", "0.67015475", "0.66962653", "0.6694413", "0.66920316", "0.668364", "0.6665041", "0.66606957", "0.6656194", "0.66515523", "0.6642541", "0.6637443", "0.6636077", "0.66321176", "0.66306406", "0.6629452", "0.6624831", "0.6616204", "0.6614542", "0.66011375", "0.6600824", "0.66002864", "0.6593393", "0.65882087", "0.65845764", "0.6584185", "0.6578444", "0.65775716", "0.6573827", "0.6569982", "0.65685755", "0.6566118", "0.6561628", "0.65558946", "0.6552881", "0.6550544", "0.6548341", "0.6539505", "0.65392876", "0.6539279", "0.65382355", "0.65379655", "0.6537655", "0.653452", "0.65326333", "0.6531451", "0.65311414", "0.65236384", "0.65229934", "0.65222603", "0.65222484", "0.6517955", "0.6517434", "0.6513995", "0.6507803", "0.6504504", "0.649559", "0.6495278", "0.64913607", "0.6490442", "0.6487966", "0.6485856", "0.6482254", "0.64738715", "0.6469595" ]
0.8536312
0
SetRegistry will store the registry by name
func SetRegistry(name string, newRegFunc func(model.Registry, common2.RegistryEventListener) (Registry, error)) { registryMap[name] = newRegFunc }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func RegistrySet(mm MX4JMetric, mb MX4JData) {\n\treglock.Lock()\n\tdefer reglock.Unlock()\n\tlog.Debugf(\"RegistrySet: %s\", mm.HumanName)\n\n\tmm.Data = mb\n\tregistry[mm.HumanName] = mm\n}", "func SetRegistry(spec *v1.ServiceBrokerConfigSpec, name string, expression interface{}) {\n\tvar str string\n\n\tswitch t := expression.(type) {\n\tcase Function:\n\t\tstr = string(t)\n\tcase Pipeline:\n\t\tstr = string(t)\n\tcase string, int, bool, nil:\n\t\tstr = argument(t)\n\tdefault:\n\t\tfmt.Println(\"fail\")\n\t}\n\n\tspec.Bindings[0].ServiceInstance.Registry = []v1.RegistryValue{\n\t\t{\n\t\t\tName: name,\n\t\t\tValue: `{{` + str + `}}`,\n\t\t},\n\t}\n}", "func (r *DefaultInstallerRegistry) Registry(name string, installer core.Installer) {\n\t_, ok := r.installerMap[name]\n\tif ok {\n\t\tr.installerMap[name] = append(r.installerMap[name], installer)\n\t} else {\n\t\tr.installerMap[name] = []core.Installer{installer}\n\t}\n}", "func (i *Config) SetRegistry(registry string) {\n\ti.registry = registry\n}", "func (sc *ShamClient) setLocalRegistry(endpoints []string) {\n\tsc.lrMutex.Lock()\n\tdefer sc.lrMutex.Unlock()\n\n\tsc.localRegistry = endpoints\n}", "func (m *Win32LobAppRegistryDetection) SetValueName(value *string)() {\n err := m.GetBackingStore().Set(\"valueName\", value)\n if err != nil {\n panic(err)\n }\n}", "func RegistrySet() {\n\tif *Registry == \"\" {\n\t\tApp.Errorf(\"required flag --%s not provided, try --help\", RegistryName)\n\t\tos.Exit(1)\n\t}\n}", "func (me *TxsdRegistryHandleSimpleContentExtensionRegistry) Set(s string) { (*xsdt.Nmtoken)(me).Set(s) }", "func SetDefaultRegistry(uri string) {\n\tregistryURI = uri\n}", "func (_Registry *RegistryTransactor) SetRegistryKeeper(opts *bind.TransactOpts, _registryKeeper common.Address) (*types.Transaction, error) {\n\treturn _Registry.contract.Transact(opts, \"setRegistryKeeper\", _registryKeeper)\n}", "func (c *ClientOptions) SetRegistry(registry *bsoncodec.Registry) *ClientOptions {\n\tc.Registry = registry\n\treturn c\n}", "func (_WyvernExchange *WyvernExchangeCallerSession) Registry() (common.Address, error) {\n\treturn _WyvernExchange.Contract.Registry(&_WyvernExchange.CallOpts)\n}", "func (_WyvernExchange *WyvernExchangeSession) Registry() (common.Address, error) {\n\treturn _WyvernExchange.Contract.Registry(&_WyvernExchange.CallOpts)\n}", "func Set(name string, l Logger) error {\n\tregMutex.Lock()\n\tdefer regMutex.Unlock()\n\tlOld, ok := reg[name]\n\tif !ok {\n\t\treg[name] = l\n\t\treturn nil\n\t}\n\tll, ok := l.(*logger)\n\tif !ok {\n\t\treturn errors.New(\"unsupported logger type to overwrite already esisting logger\")\n\t}\n\tllOld, ok := lOld.(*logger)\n\tif !ok {\n\t\treturn errors.New(\"unsupported logger type to overwrite already esisting logger\")\n\t}\n\t*llOld = *ll\n\treturn nil\n}", "func Registry(r registry.Registry) Option {\n\treturn func(o *Options) {\n\t\to.Registry = r\n\t}\n}", "func Registry(r registry.RegistryI) Option {\n\treturn func(o *Options) {\n\t\to.Registry = r\n\t\t// Update Client and Server\n\t\to.Client.Init(client.Registry(r))\n\t\to.Server.Init(server.Registry(r))\n\t\t// Update Broker\n\t\to.Broker.Init(broker.Registry(r))\n\t}\n}", "func RegistrySetAndExists(settings *config.Settings) {\n\tRegistrySet()\n\tif _, ok := settings.Auths[*Registry]; !ok {\n\t\tApp.Errorf(\"Registry doesn't exist. maybe add it?\")\n\t\tos.Exit(1)\n\t}\n}", "func (sr *Registry) Register(name string, suite Suite) {\n\tif _, found := sr.registry[name]; found {\n\t\tlog.Fatal(fmt.Sprintf(\"Trying to register the suite %s multiple times\", name))\n\t}\n\n\tsr.registry[name] = suite\n}", "func Register(name string, builder Builder) {\n\tregistry[name] = builder\n}", "func (_KeepRegistry *KeepRegistryTransactor) SetRegistryKeeper(opts *bind.TransactOpts, _registryKeeper common.Address) (*types.Transaction, error) {\n\treturn _KeepRegistry.contract.Transact(opts, \"setRegistryKeeper\", _registryKeeper)\n}", "func (m *Win32LobAppRegistryRule) SetValueName(value *string)() {\n m.valueName = value\n}", "func (c *contractInsecureDarc) SetRegistry(r ismartc.ReadOnlyContractRegistry) {\n\tc.contracts = r\n}", "func NewRegistry(scheme *runtime.Scheme, codec serializer.CodecFactory, serializer *json.Serializer) *Registry {\n\tvar groupVersions []schema.GroupVersion\n\tfor k := range scheme.AllKnownTypes() {\n\t\tgroupVersions = append(groupVersions, k.GroupVersion())\n\t}\n\n\treturn &Registry{\n\t\tscheme: scheme,\n\t\tcodec: codec.CodecForVersions(serializer, serializer, schema.GroupVersions(groupVersions), schema.GroupVersions(groupVersions)),\n\t\tnameToObject: make(map[string]*object),\n\t}\n}", "func (s *Server) Register(name string, h Handler) error {\n if _, ok := s.registry[name]; ok {\n return fmt.Errorf(\"cannot register name %q twice\", name)\n }\n s.registry[name] = &register{handler: h}\n return nil\n}", "func (r *Registrar) writeRegistry() error {\n\tr.gcStates()\n\n\tlogp.Debug(\"registrar\", \"Write registry file: %s\", r.registryFile)\n\n\ttempfile := r.registryFile + \".new\"\n\tf, err := os.OpenFile(tempfile, os.O_RDWR|os.O_CREATE|os.O_TRUNC|os.O_SYNC, 0600)\n\tif err != nil {\n\t\tlogp.Err(\"Failed to create tempfile (%s) for writing: %s\", tempfile, err)\n\t\treturn err\n\t}\n\n\t// First clean up states\n\tstates := r.states.GetStates()\n\n\tencoder := json.NewEncoder(f)\n\terr = encoder.Encode(states)\n\tif err != nil {\n\t\tf.Close()\n\t\tlogp.Err(\"Error when encoding the states: %s\", err)\n\t\treturn err\n\t}\n\n\t// Directly close file because of windows\n\tf.Close()\n\n\terr = helper.SafeFileRotate(r.registryFile, tempfile)\n\n\tlogp.Debug(\"registrar\", \"Registry file updated. %d states written.\", len(states))\n\t \n\n\treturn err\n}", "func Register(name string, f driver.UpdaterSetFactory) {\n\tpkg.Lock()\n\tdefer pkg.Unlock()\n\tif _, ok := pkg.fs[name]; ok {\n\t\tpanic(\"\")\n\t}\n\tpkg.fs[name] = f\n}", "func (a *AssetRegistry) SetRegistryURL(url string) {\n\ta.client = registryclient.NewHTTPClient(resty.New().SetHostURL(url))\n}", "func Register(name string, port int) (err error) {\n\tr := RegistryRequest{name, port}\n\n\tbyt, err := json.Marshal(r)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tresp, err := http.Post(\"http://127.0.0.1\" + _LOCAL_PORT, \"text/json\", bytes.NewBuffer(byt))\n\tif err != nil {\n\t\treturn\n\t}\n\tif resp.StatusCode != http.StatusOK {\n\t\terr = fmt.Errorf(\"non 200 response %d: %s\", resp.StatusCode, resp.Status)\n\t\treturn\n\t}\n\n\tbyt, err = ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func (p *Pack) Registry() (reg *registry.Registry) {\n\treturn p.reg\n}", "func (m *RegistryKeyState) SetValueName(value *string)() {\n m.valueName = value\n}", "func (_WyvernExchange *WyvernExchangeCaller) Registry(opts *bind.CallOpts) (common.Address, error) {\n\tvar (\n\t\tret0 = new(common.Address)\n\t)\n\tout := ret0\n\terr := _WyvernExchange.contract.Call(opts, out, \"registry\")\n\treturn *ret0, err\n}", "func register(name string, p Plugin) {\n\tdirectory[name] = p\n}", "func (r *Registry) Store(value interface{}) {\n\t_, filename, _, _ := runtime.Caller(1)\n\tr.db.Store(filename, value)\n}", "func (_Erc1820Registry *Erc1820RegistryTransactor) SetManager(opts *bind.TransactOpts, _addr common.Address, _newManager common.Address) (*types.Transaction, error) {\n\treturn _Erc1820Registry.contract.Transact(opts, \"setManager\", _addr, _newManager)\n}", "func (r *Registry) Register(school data.School) {\n\tr.schoolList = append(r.schoolList, school)\n}", "func (_Registry *RegistryTransactorSession) SetRegistryKeeper(_registryKeeper common.Address) (*types.Transaction, error) {\n\treturn _Registry.Contract.SetRegistryKeeper(&_Registry.TransactOpts, _registryKeeper)\n}", "func RegisterStore(name string, f func(interface{}) Storage) {\n\tif Registry[name] != nil {\n\t\tpanic(\"storage registry name conflict: \" + name)\n\t}\n\tRegistry[name] = f\n}", "func (reg *registrar) Make(name string) (interface{}, error) {\n\treg.lock.Lock()\n\tdefer reg.lock.Unlock()\n\treturn reg.Registry.Make(name)\n}", "func NewRegistry(s Storage) Registry {\n\treturn &storage{s}\n}", "func NewRegistry(s Storage) Registry {\n\treturn &storage{s}\n}", "func NewRegistry() *Registry {\n\treturn &Registry{\n\t\toutlets: make([]*Outlet, 0),\n\t\toutletMap: make(map[string]*Outlet),\n\t\tgroups: make([]*Group, 0),\n\t\tgroupMap: make(map[string]*Group),\n\t}\n}", "func (_Registry *RegistrySession) SetRegistryKeeper(_registryKeeper common.Address) (*types.Transaction, error) {\n\treturn _Registry.Contract.SetRegistryKeeper(&_Registry.TransactOpts, _registryKeeper)\n}", "func Assign(key, file, name string) error {\n\tcl, err := sockfile.NewRpcClient(\"redis-reg\")\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer cl.Close()\n\treturn cl.Call(\"Reg.Assign\", args.Assign{key, file, name}, &empty)\n}", "func Register(name string, driver Driver) {\n\tdriversMu.Lock()\n\tdefer driversMu.Unlock()\n\n\tdrivers[name] = driver\n}", "func (_Coordinatorproxy *CoordinatorproxySession) NameRegistry() (common.Address, error) {\n\treturn _Coordinatorproxy.Contract.NameRegistry(&_Coordinatorproxy.CallOpts)\n}", "func (_ElvTradableLocal *ElvTradableLocalTransactor) SetProxyRegistryAddress(opts *bind.TransactOpts, _newProxy common.Address) (*types.Transaction, error) {\n\treturn _ElvTradableLocal.contract.Transact(opts, \"setProxyRegistryAddress\", _newProxy)\n}", "func (_Coordinatorproxy *CoordinatorproxyCallerSession) NameRegistry() (common.Address, error) {\n\treturn _Coordinatorproxy.Contract.NameRegistry(&_Coordinatorproxy.CallOpts)\n}", "func NewRegistry(opts ...registry.Option) registry.Registry {\r\n\treturn registry.NewRegistry(opts...)\r\n}", "func (_ElvTradable *ElvTradableTransactor) SetProxyRegistryAddress(opts *bind.TransactOpts, _newProxy common.Address) (*types.Transaction, error) {\n\treturn _ElvTradable.contract.Transact(opts, \"setProxyRegistryAddress\", _newProxy)\n}", "func InitializeRegistry(\n\tcommandMap stringmap.StringMap,\n\tkarmaMap stringmap.StringMap,\n\tvoteMap stringmap.StringMap,\n\tgist api.Gist,\n\tconfig *config.Config,\n\tclock model.UTCClock,\n\ttimer model.UTCTimer,\n\tcommandChannel chan<- *model.Command) *feature.Registry {\n\n\t// Initializing builtin features.\n\t// TODO(jvoytko): investigate the circularity that emerged to see if there's\n\t// a better pattern here.\n\tfeatureRegistry := feature.NewRegistry()\n\tallFeatures := []feature.Feature{\n\t\tfactsphere.NewFeature(featureRegistry),\n\t\thelp.NewFeature(featureRegistry),\n\t\tkarma.NewFeature(featureRegistry, karmaMap),\n\t\tkarmalist.NewFeature(featureRegistry, karmaMap, gist),\n\t\tlearn.NewFeature(featureRegistry, commandMap),\n\t\tlist.NewFeature(featureRegistry, commandMap, gist),\n\t\tmoderation.NewFeature(featureRegistry, config),\n\t\tvote.NewFeature(featureRegistry, voteMap, clock, timer, commandChannel),\n\t}\n\n\tfor _, f := range allFeatures {\n\t\terr := featureRegistry.Register(f)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t}\n\treturn featureRegistry\n}", "func (d *Devices) Registry() *RegistryServiceClient {\n\treturn &RegistryServiceClient{getConn: d.getConn}\n}", "func NewRegistry(ctx *pulumi.Context,\n\tname string, args *RegistryArgs, opts ...pulumi.ResourceOption) (*Registry, error) {\n\tif args == nil {\n\t\targs = &RegistryArgs{}\n\t}\n\tvar resource Registry\n\terr := ctx.RegisterResource(\"gcp:kms/registry:Registry\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func NewRegistry(s rest.StandardStorage) Registry {\n\treturn &storage{s}\n}", "func NewRegistry(opts ...registry.Option) registry.Registry {\n\treturn mdns.NewRegistry(opts...)\n}", "func (r *Registry) Register(ds ...*Object) {\n\tr.Lock()\n\tdefer r.Unlock()\n\tr.initDataStores()\n\n\tfor _, obj := range ds {\n\t\tname := obj.Name\n\t\tr.dataStores[name] = obj\n\t\tr.dataStores[name].Enabled = true\n\t}\n}", "func AddRegistry(spec *v1.ServiceBrokerConfigSpec, name string, expression interface{}) {\n\tvar str string\n\n\tswitch t := expression.(type) {\n\tcase Function:\n\t\tstr = string(t)\n\tcase Pipeline:\n\t\tstr = string(t)\n\tcase string, int, bool, nil:\n\t\tstr = argument(t)\n\tdefault:\n\t\tfmt.Println(\"fail\")\n\t}\n\n\tspec.Bindings[0].ServiceInstance.Registry = append(spec.Bindings[0].ServiceInstance.Registry, v1.RegistryValue{\n\t\tName: name,\n\t\tValue: `{{` + str + `}}`,\n\t})\n}", "func (reg *reflectRegistry) set(rt reflect.Type, info *reflectInfo) {\n\treg.Lock()\n\tif exist := reg.infoMap[rt]; exist == nil {\n\t\treg.infoMap[rt] = info\n\t}\n\treg.Unlock()\n}", "func newRegistry(logger logr.Logger, config globalregistry.RegistryConfig) (globalregistry.Registry, error) {\n\tvar err error\n\tc := &registry{\n\t\tlogger: logger,\n\t\tRegistryConfig: config,\n\t\tClient: http.DefaultClient,\n\t}\n\tc.projects, err = newProjectAPI(c)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tc.remoteRegistries = newRemoteRegistries(c)\n\tc.replications = newReplicationAPI(c)\n\tc.parsedUrl, err = url.Parse(config.GetAPIEndpoint())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tc.scanners = newScannerAPI(c)\n\treturn c, nil\n}", "func Register(s string, new NewHash) {\n\t// TODO: check registry\n\n\t// check the Hash interface is matched\n\tvar _ Hash = new()\n\n\tsupportedHashes[s] = new\n}", "func RegistryFlush() {\n\treglock.Lock()\n\tdefer reglock.Unlock()\n\n\t// Replace the registry with a new map\n\tfor k, mm := range registry {\n\t\tlog.Debugf(\"Blanking gowrapmx4j.registry data of %s\", k)\n\t\tmm.Data = nil\n\t\tregistry[mm.HumanName] = mm\n\t}\n}", "func (asr *sessionRegistry) register(clt *Client) {\n\tasr.lock.Lock()\n\tasr.registry[clt.Session.Key] = clt\n\tasr.lock.Unlock()\n}", "func NewRegistry(settings ...Settings) *Registry {\n\treturn NewSwarmRegistry(nil, nil, settings...)\n}", "func (_Posminer *PosminerTransactor) MinerRegistry(opts *bind.TransactOpts, MinerPool common.Address, RegisterFingure string) (*types.Transaction, error) {\n\treturn _Posminer.contract.Transact(opts, \"MinerRegistry\", MinerPool, RegisterFingure)\n}", "func (_Depositmanager *DepositmanagerCallerSession) NameRegistry() (common.Address, error) {\n\treturn _Depositmanager.Contract.NameRegistry(&_Depositmanager.CallOpts)\n}", "func (r *LocalRegistry) save() {\n\tregBytes := core.ToJsonBytes(r)\n\tcore.CheckErr(ioutil.WriteFile(r.file(), regBytes, os.ModePerm), \"fail to update local registry metadata\")\n}", "func bindRegistry(address common.Address, caller bind.ContractCaller, transactor bind.ContractTransactor, filterer bind.ContractFilterer) (*bind.BoundContract, error) {\n\tparsed, err := abi.JSON(strings.NewReader(RegistryABI))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn bind.NewBoundContract(address, parsed, caller, transactor, filterer), nil\n}", "func bindRegistry(address common.Address, caller bind.ContractCaller, transactor bind.ContractTransactor, filterer bind.ContractFilterer) (*bind.BoundContract, error) {\n\tparsed, err := abi.JSON(strings.NewReader(RegistryABI))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn bind.NewBoundContract(address, parsed, caller, transactor, filterer), nil\n}", "func Register(name string, drv Driver) error {\n\tif _, f := drivers[name]; f {\n\t\treturn fmt.Errorf(\"morsel: driver '%s' already registered\", name)\n\t}\n\tdrivers[name] = drv\n\treturn nil\n}", "func (r *Registry) Register(name string, element interface{}) {\n\tr.mu.Lock()\n\tdefer r.mu.Unlock()\n\tr.types[name] = reflect.TypeOf(element)\n}", "func NewRegistry(name string) (*Registry, error) {\n\tref := registry.Reference{\n\t\tRegistry: name,\n\t}\n\tif err := ref.ValidateRegistry(); err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Registry{\n\t\tRepositoryOptions: RepositoryOptions{\n\t\t\tReference: ref,\n\t\t},\n\t}, nil\n}", "func NewRegistry() *Registry {\n\treturn &Registry{\n\t\ttypeEncoders: new(typeEncoderCache),\n\t\ttypeDecoders: new(typeDecoderCache),\n\t\tkindEncoders: new(kindEncoderCache),\n\t\tkindDecoders: new(kindDecoderCache),\n\t}\n}", "func (cs *configSet) Register(name string, data IConfigurable, doc string) error {\n\tif len(name) < 1 {\n\t\treturn fmt.Errorf(\"config.Register(%s, %T) without a name\", name, data)\n\t}\n\n\tif _, ok := cs.all[name]; ok {\n\t\treturn fmt.Errorf(\"config.Register(%s, %T) with duplicate name\", name, data)\n\t}\n\n\tcs.all[name] = &config{\n\t\tdata: data,\n\t\tdoc: doc,\n\t\trt: nil,\n\t\tviperConfig: nil,\n\t\tloaded: false,\n\t}\n\tlog.Debugf(\"Registered config(%s)=%T\", name, data)\n\treturn nil\n}", "func (mgr *Manager) Register(register string) error {\n\tif register == \"unchanged\" {\n\t\treturn nil\n\t}\n\tif register == \"unwanted\" || register == \"unknown\" {\n\t\terr := mgr.RemoveNodeDef(NODE_DEFS_WANTED)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif register == \"unknown\" {\n\t\t\terr := mgr.RemoveNodeDef(NODE_DEFS_KNOWN)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\tif register == \"known\" || register == \"knownForce\" ||\n\t\tregister == \"wanted\" || register == \"wantedForce\" {\n\t\t// Save our nodeDef (with our UUID) into the Cfg as a known node.\n\t\terr := mgr.SaveNodeDef(NODE_DEFS_KNOWN, register == \"knownForce\")\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif register == \"wanted\" || register == \"wantedForce\" {\n\t\t\t// Save our nodeDef (with our UUID) into the Cfg as a wanted node.\n\t\t\terr := mgr.SaveNodeDef(NODE_DEFS_WANTED, register == \"wantedForce\")\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func InitRegistry() {\n\tsubscriberRegistry = make(map[Topic][]chan ExecutionEvent)\n\tsubscriberRegistry[SuiteStart] = make([]chan ExecutionEvent, 0)\n\tsubscriberRegistry[ScenarioStart] = make([]chan ExecutionEvent, 0)\n\tsubscriberRegistry[ConceptStart] = make([]chan ExecutionEvent, 0)\n\tsubscriberRegistry[StepStart] = make([]chan ExecutionEvent, 0)\n\tsubscriberRegistry[SuiteEnd] = make([]chan ExecutionEvent, 0)\n\tsubscriberRegistry[ConceptEnd] = make([]chan ExecutionEvent, 0)\n\tsubscriberRegistry[ScenarioEnd] = make([]chan ExecutionEvent, 0)\n\tsubscriberRegistry[SpecEnd] = make([]chan ExecutionEvent, 0)\n\tsubscriberRegistry[SuiteEnd] = make([]chan ExecutionEvent, 0)\n}", "func NewRegistry(userNS *auth.UserNamespace) *Registry {\n\treturn &Registry{\n\t\treg: ipc.NewRegistry(userNS),\n\t}\n}", "func (rr *Registry) SetWanted(ctx context.Context, graphType, graphName string, state State) error {\n\trr.mu.Lock()\n\tdefer rr.mu.Unlock()\n\n\tkey := rr.keyFromGraphTypeAndName(graphType, graphName)\n\n\tgetRes, err := rr.kv.Get(ctx, key, etcdv3.WithLimit(2))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif getRes.Count == 0 {\n\t\treturn nil\n\t}\n\tif getRes.Count > 1 {\n\t\treturn ErrMultipleValues\n\t}\n\n\tkv := getRes.Kvs[0]\n\trec := &Registration{}\n\terr = json.Unmarshal(kv.Value, rec)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif rec.Wanted == string(state) {\n\t\treturn nil\n\t}\n\trec.Wanted = string(state)\n\n\treturn rr.insert(ctx, getRes.Kvs[0].Version, key, rec)\n}", "func RegisterStore(name string, newFunc func(*Robot) (Store, error)) {\n\tStores[name] = store{\n\t\tname: name,\n\t\tnewFunc: newFunc,\n\t}\n}", "func (c *LiveCollector) Registry() fslib.FS {\n\treturn c.registryfs\n}", "func Set(name string, value interface{}) {\n\tinitialize()\n\tlock.Lock()\n\tnotGlobals[name] = value\n\tlock.Unlock()\n}", "func NewRegistry() *Registry {\n\treturn &Registry{\n\t\tnodesView: make(map[identity.ID]*View),\n\t}\n}", "func (_Depositmanager *DepositmanagerSession) NameRegistry() (common.Address, error) {\n\treturn _Depositmanager.Contract.NameRegistry(&_Depositmanager.CallOpts)\n}", "func NewRegistry() Registry {\n\treturn make(Registry)\n}", "func (r registry) Register(year int, day int, f Solver) {\n\tr[getKey(year, day)] = f\n}", "func RegisterTypes(registry interface {\n\t RegisterType(name string, obj any)\n}) {\n\n}", "func (_Erc1820Registry *Erc1820RegistrySession) SetManager(_addr common.Address, _newManager common.Address) (*types.Transaction, error) {\n\treturn _Erc1820Registry.Contract.SetManager(&_Erc1820Registry.TransactOpts, _addr, _newManager)\n}", "func (in *Registry) DeepCopy() *Registry {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(Registry)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (in *Registry) DeepCopy() *Registry {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(Registry)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (m *RegistryKeyState) SetKey(value *string)() {\n m.key = value\n}", "func AddRegistryFor[T any](cmpType components.PluggableType, regFunc func(componentFactory func(logger.Logger) T, names ...string), factory func(logger.Logger, components.Pluggable) T) {\n\tregistries[cmpType] = func(pc components.Pluggable) {\n\t\tregFunc(func(l logger.Logger) T {\n\t\t\treturn factory(l, pc)\n\t\t}, pc.Name)\n\t}\n}", "func (_Posminer *PosminerSession) MinerRegistry(MinerPool common.Address, RegisterFingure string) (*types.Transaction, error) {\n\treturn _Posminer.Contract.MinerRegistry(&_Posminer.TransactOpts, MinerPool, RegisterFingure)\n}", "func NewRegistry() *Registry {\n\tr := &Registry{}\n\tr.functions = make(map[string]*function)\n\treturn r\n}", "func (cmd *createRegistryCommand) CreateRegistry(cfg *config.Config, arguments []string) error {\n\tdatabase, err := cfg.OpenDatabase()\n\tif database != nil {\n\t\tdefer cfg.CloseDatabase()\n\t}\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// detect type\n\tfactory, found := registry.Types[cmd.Type]\n\tif !found {\n\t\treturn ErrorRegistryTypeNotSupported\n\t}\n\n\treg, err := factory.Create(registry.Data{\n\t\tName: cmd.Name,\n\t}, arguments)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// create path for registry\n\tpath, err := filepath.Abs(filepath.Join(cfg.ConfigDir, \"registries\", reg.Data().Name))\n\tif err != nil {\n\t\treturn err\n\t}\n\treg.Data().Path = path\n\n\tregistries := registry.New(config.Database{DB: database})\n\n\texists, err := registries.Exists(reg.Data().Name)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif exists && cmd.Force {\n\t\tregistries.Remove(reg.Data().Name)\n\t}\n\n\terr = reg.Initialize()\n\tif err != nil {\n\t\t// remove the content there\n\t\tos.RemoveAll(path)\n\t\tlogrus.Warningln(\"Could not initialize registry\")\n\t\treturn err\n\t}\n\n\terr = registries.Add(reg)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn GetRegistries(cfg.WithDatabase(database))\n}", "func (registry *Registry) Register(t *task.Task) string {\n\t// For the purposes of this exercise we're using a stringified UUID as\n\t// the task ID. This should be statistically unique - certainly we\n\t// should not expect to see a collision in the lifetime of the server.\n\t// It's reasonably safe to treat the generated handles as unique without\n\t// further checking. In fact, if we see duplicate handles we probably\n\t// have bigger problems on the system than this toy service misbehaving.\n\n\thandle := handleFromUUID()\n\n\tregistry.lock.Lock()\n\tdefer registry.lock.Unlock()\n\n\tregistry.db[handle] = t\n\treturn handle\n}", "func NewRegistry() *Registry {\n\treturn &Registry{\n\t\trateLimiters: make(map[string]*rate.Limiter),\n\t}\n}", "func RegistryGet(humanName string) MX4JMetric {\n\treglock.RLock()\n\tdefer reglock.RUnlock()\n\n\treturn registry[humanName]\n}", "func RegisterRCS(id RCSBackend, name string, loader RCSLoader) {\n\trcsRegistry[id] = loader\n\trcsNameToBackendMap[name] = id\n\trcsBackendToNameMap[id] = name\n}", "func (d *Debugger) WriteRegByName(name string, value uint64) error {\n\tvar reg Register\n\tif attr, err := d.arch.register(name); err == nil {\n\t\treg.attr = attr\n\t\treg.Value = value\n\t\treturn d.WriteReg(reg)\n\t} else {\n\t\treturn err\n\t}\n}", "func (z *ZkRegistry) register() error {\n\terr := z.connect()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// 1. Check root node existed or not. If not, create the root node\n\tisExist, _, err := z.Conn.Exists(z.NodeParams.Root)\n\tif err != nil {\n\t\treturn errors.New(\"Check root path error: \\n\" + err.Error())\n\t}\n\n\tif !isExist {\n\t\terr = z.createNode(z.NodeParams.Root, nil, 0)\n\t}\n\tif err != nil {\n\t\treturn errors.New(\"Create root path error: \\n\" + err.Error())\n\t}\n\n\t// 2. Use CreateProtectedEphemeralSequential method to create the node as a ephemeral sequential node.\n\t// Then the rule for the name like this: _c_ + guid + - + srvName + index\n\t// All about this, please check the code at here: https://github.com/samuel/go-zookeeper/blob/master/zk/conn.go#L1061\n\tsrvFullPath := z.NodeParams.Root + \"/\" + z.NodeParams.SrvName\n\trealSrvFullPath, err := z.Conn.CreateProtectedEphemeralSequential(srvFullPath, z.NodeParams.Data, z.acl())\n\n\tif err != nil {\n\t\treturn errors.New(\"Create zk node for srv failed: \\n\" + err.Error())\n\t}\n\n\trealSrvParts := strings.Split(realSrvFullPath, \"/\")\n\tz.node.RealSrvPath = realSrvParts[len(realSrvParts)-1]\n\treturn nil\n}", "func (r OpenstackAdapter) RegistryName() string {\n\tif r.Config.URL.Host == \"\" {\n\t\treturn r.Config.URL.Path\n\t}\n\treturn r.Config.URL.Host\n}", "func (s *servicecenter) Registry(clusterName string, data *pb.SyncData) {\n\tmapping := s.storage.GetMapByCluster(clusterName)\n\tfor _, inst := range data.Instances {\n\t\tsvc := searchService(inst, data.Services)\n\t\tif svc == nil {\n\t\t\terr := errors.New(\"service does not exist\")\n\t\t\tlog.Error(fmt.Sprintf(\"servicecenter.Registry, serviceID = %s, instanceId = %s\", inst.ServiceId, inst.InstanceId), err)\n\t\t\tcontinue\n\t\t}\n\n\t\t// If the svc is in the mapping, just do nothing, if not, created it in servicecenter and get the new serviceID\n\t\tsvcID, err := s.createService(svc)\n\t\tif err != nil {\n\t\t\tlog.Error(\"create service failed\", err)\n\t\t\tcontinue\n\t\t}\n\n\t\t// If inst is in the mapping, just heart beat it in servicecenter\n\t\tlog.Debug(fmt.Sprintf(\"trying to do registration of instance, instanceID = %s\", inst.InstanceId))\n\t\tif s.heartbeatInstances(mapping, inst) {\n\t\t\tcontinue\n\t\t}\n\n\t\t// If inst is not in the mapping, that is because this the first time syncer get the instance data\n\t\t// in this case, we should registry it to the servicecenter and get the new instanceID\n\t\titem := &pb.MappingEntry{\n\t\t\tClusterName: clusterName,\n\t\t\tDomainProject: svc.DomainProject,\n\t\t\tOrgServiceID: svc.ServiceId,\n\t\t\tOrgInstanceID: inst.InstanceId,\n\t\t\tCurServiceID: svcID,\n\t\t\tCurInstanceID: s.registryInstances(svc.DomainProject, svcID, inst),\n\t\t}\n\n\t\t// Use new serviceID and instanceID to update mapping data in this servicecenter\n\t\tif item.CurInstanceID != \"\" {\n\t\t\tmapping = append(mapping, item)\n\t\t}\n\t}\n\t// UnRegistry instances that is not in the data which means the instance in the mapping is no longer actived\n\tmapping = s.unRegistryInstances(data, mapping)\n\t// Update mapping data of the cluster to the storage of the servicecenter\n\ts.storage.UpdateMapByCluster(clusterName, mapping)\n}" ]
[ "0.7304189", "0.7187912", "0.6942264", "0.6334491", "0.62176913", "0.62176603", "0.59973913", "0.59799457", "0.5962236", "0.58829594", "0.5862283", "0.5826678", "0.58098185", "0.58006924", "0.5781043", "0.57669806", "0.57097864", "0.5626036", "0.5607278", "0.55465734", "0.5542331", "0.54581046", "0.5439968", "0.5432317", "0.5431688", "0.54257506", "0.539838", "0.5391556", "0.53692055", "0.53484386", "0.5347309", "0.5332048", "0.5309293", "0.5305312", "0.5304532", "0.52915055", "0.5291274", "0.52883697", "0.528723", "0.528723", "0.5273813", "0.52647734", "0.5250946", "0.5233693", "0.5219987", "0.5213758", "0.5210409", "0.52007604", "0.51978", "0.51870203", "0.5176823", "0.51507777", "0.51470983", "0.5146603", "0.512349", "0.5118321", "0.51125574", "0.51120424", "0.51107407", "0.50976896", "0.5094408", "0.5091656", "0.50890166", "0.5079567", "0.50742006", "0.5073585", "0.5073585", "0.5068642", "0.50596946", "0.5052978", "0.5051225", "0.5048859", "0.50485766", "0.5047108", "0.50463355", "0.50408703", "0.50382453", "0.50379425", "0.50377655", "0.50325024", "0.50235265", "0.50231856", "0.50174", "0.5015265", "0.50127125", "0.50096047", "0.50096047", "0.5006749", "0.50031835", "0.49981433", "0.49928617", "0.49906117", "0.498946", "0.49886805", "0.49861056", "0.49859467", "0.4976173", "0.49761102", "0.49747193", "0.49725056" ]
0.77865237
0
GetRegistry will return the registry if not found, it will panic
func GetRegistry(name string, regConfig model.Registry, listener common2.RegistryEventListener) (Registry, error) { if registry, ok := registryMap[name]; ok { reg, err := registry(regConfig, listener) if err != nil { panic("Initialize Registry" + name + "failed due to: " + err.Error()) } return reg, nil } return nil, errors.New("Registry " + name + " does not support yet") }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func GetRegistry(registryURL string, username string, password string) (*registry.Registry, error) {\n\thub, err := registry.New(registryURL, username, password)\n\tif err != nil {\n\t\treturn hub, errors.Wrap(err, \"error initializing registry\")\n\t}\n\treturn hub, nil\n}", "func GetRegistry(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *RegistryState, opts ...pulumi.ResourceOption) (*Registry, error) {\n\tvar resource Registry\n\terr := ctx.ReadResource(\"aws-native:eventschemas:Registry\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func GetRegistry(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *RegistryState, opts ...pulumi.ResourceOption) (*Registry, error) {\n\tvar resource Registry\n\terr := ctx.ReadResource(\"gcp:kms/registry:Registry\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func GetRegistry(ctx context.Context) (*Registry, bool) {\n\treg, ok := ctx.Value(registryContextKey{}).(*Registry)\n\treturn reg, ok\n}", "func (v *VersionClient) GetRegistry(ctx context.Context) (*apitypes.Version, error) {\n\treturn v.fetchVersion(ctx, true)\n}", "func (o *WlDisplay) GetRegistry(registry *WlRegistry) error {\n\tmsg, err := wire.NewMessage(o.ID(), 1)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err = msg.Write(registry.ID()); err != nil {\n\t\treturn err\n\t}\n\n\tif err = o.Base.Conn.Write(msg); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func RegistryGet(humanName string) MX4JMetric {\n\treglock.RLock()\n\tdefer reglock.RUnlock()\n\n\treturn registry[humanName]\n}", "func (r *Registry) GetRegistry(rw http.ResponseWriter, req *http.Request) {\n\tvar items = []*W{}\n\tfor i := 0; i < r.m.list.Len(); i++ {\n\t\titems = append(items, r.m.list.Get(i))\n\t}\n\n\tif msg, err := json.Marshal(items); err != nil {\n\t\tjsonResponse(rw, http.StatusInternalServerError, err.Error())\n\t} else {\n\t\trw.Header().Set(\"Content-Type\", \"application/json\")\n\t\trw.Write(msg)\n\t}\n}", "func (ctx *serverRequestContextImpl) GetRegistry() user.Registry {\n\treturn ctx.ca.registry\n}", "func (c *client) getRegistryByName(registryName string) (*Registry, error) {\n\tregistryJSON, err := c.runCmd(\"registry\", \"list\", \"-o\", \"json\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tregistryList := &RegistryList{}\n\tif err := registryList.Unmarshal([]byte(registryJSON)); err != nil {\n\t\treturn nil, err\n\t}\n\n\tfor _, registry := range registryList.Registries {\n\t\tif registry.Name == registryName {\n\t\t\tif c.verbose {\n\t\t\t\tfmt.Printf(\"k3d registry '%s' exists\", registryName)\n\t\t\t}\n\t\t\treturn &registry, nil\n\t\t}\n\t}\n\n\tif c.verbose {\n\t\tfmt.Printf(\"k3d registry '%s' does not exist\", registryName)\n\t}\n\treturn nil, nil\n}", "func (m RegistryMap) Get(name string) (component.Registration, bool) {\n\treg, ok := m[name]\n\treturn reg, ok\n}", "func GetPrivateRegistry(cluster *v3.Cluster) *rketypes.PrivateRegistry {\n\tprivateClusterLevelRegistry := GetPrivateClusterLevelRegistry(cluster)\n\tif privateClusterLevelRegistry != nil {\n\t\treturn privateClusterLevelRegistry\n\t}\n\tif settings.SystemDefaultRegistry.Get() != \"\" {\n\t\treturn &rketypes.PrivateRegistry{\n\t\t\tURL: settings.SystemDefaultRegistry.Get(),\n\t\t}\n\t}\n\treturn nil\n}", "func GetRegistryAddress() (*net.TCPAddr, error) {\n\tif registryAddress != nil {\n\t\treturn registryAddress, nil\n\t}\n\n\tch := make(chan *net.TCPAddr, 1)\n\tintf, err := net.Interfaces()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t\n\tgo GetRegistryAddressFromInterface(net.Interface{}, true, ch)\n\tfor _, i := range intf {\n\t\tgo GetRegistryAddressFromInterface(i, false, ch)\n\t}\n\t\n\tselect {\n case address := <-ch:\n\t\tregistryAddress = address\n return address, nil\n case <-time.After(6 * time.Second):\n return nil, errors.New(\"error: no registry found!\")\n }\n\t\n\treturn nil, nil\n}", "func (r *Registry) Get(name string) interface{} {\n\treturn r.registrants[name]\n}", "func GetRegistries(ctx *types.SystemContext) ([]Registry, error) {\n\tconfig, err := getConfig(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn config.Registries, nil\n}", "func MustGetRegistry(ctx context.Context) *Registry {\n\treg, ok := ctx.Value(registryContextKey{}).(*Registry)\n\tif !ok {\n\t\tlog.S().Panic(\"Miss registry context\")\n\t}\n\treturn reg\n}", "func (_RandomBeacon *RandomBeaconCaller) GetGroupsRegistry(opts *bind.CallOpts) ([][32]byte, error) {\n\tvar out []interface{}\n\terr := _RandomBeacon.contract.Call(opts, &out, \"getGroupsRegistry\")\n\n\tif err != nil {\n\t\treturn *new([][32]byte), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new([][32]byte)).(*[][32]byte)\n\n\treturn out0, err\n\n}", "func (p *Pack) Registry() (reg *registry.Registry) {\n\treturn p.reg\n}", "func (reg DefaultComponentRegistry) Get(name string) (component.Registration, bool) {\n\treturn component.Get(name)\n}", "func (e *NodeStatusWSHandler) Registry() *ice9.ChannelRegistry {\n\treturn e.registry\n}", "func (a *Client) GetRegistryFile(params *GetRegistryFileParams) (*GetRegistryFileOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetRegistryFileParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"getRegistryFile\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/Registries/{identifier}\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &GetRegistryFileReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*GetRegistryFileOK), nil\n\n}", "func NewRegistry() Registry {\n\treturn make(Registry)\n}", "func (_RandomBeacon *RandomBeaconCallerSession) GetGroupsRegistry() ([][32]byte, error) {\n\treturn _RandomBeacon.Contract.GetGroupsRegistry(&_RandomBeacon.CallOpts)\n}", "func (_WyvernExchange *WyvernExchangeSession) Registry() (common.Address, error) {\n\treturn _WyvernExchange.Contract.Registry(&_WyvernExchange.CallOpts)\n}", "func GetConnectedRegistry(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *ConnectedRegistryState, opts ...pulumi.ResourceOption) (*ConnectedRegistry, error) {\n\tvar resource ConnectedRegistry\n\terr := ctx.ReadResource(\"azure:containerservice/connectedRegistry:ConnectedRegistry\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (sr *Registry) Get(name string) Suite {\n\ts, found := sr.registry[name]\n\tif !found {\n\t\tlog.Fatal(fmt.Sprintf(\"The suite %s does not exist\", name))\n\t}\n\n\treturn s\n}", "func GetRegistryURL() (URL string, hubAPI bool, BasicAuth string, XRegistryAuth string) {\n\tbasicAuth, fullRegistry, XRegistryAuth := RegAuth(RegistrySpec)\n\tif *AuthRegistry == true {\n\t\tif basicAuth == \"\" {\n\t\t\texcept.Fail(\"Registry auth could not be determined from docker config.\")\n\t\t}\n\t\tBasicAuth = basicAuth\n\t}\n\tif *HTTPSRegistry == false {\n\t\tURL = \"http://\" + RegistrySpec\n\t} else {\n\t\t// HTTPS is required\n\t\tif strings.HasPrefix(fullRegistry, \"https://\") {\n\t\t\tURL = fullRegistry\n\t\t} else {\n\t\t\tURL = \"https://\" + RegistrySpec\n\t\t}\n\t\tif *RegistryTokenAuthV1 == true {\n\t\t\t// only for original Docker Hub registry, due to be phased out soon.\n\t\t\thubAPI = true\n\t\t}\n\t}\n\treturn\n}", "func (_WyvernExchange *WyvernExchangeCallerSession) Registry() (common.Address, error) {\n\treturn _WyvernExchange.Contract.Registry(&_WyvernExchange.CallOpts)\n}", "func (_RandomBeacon *RandomBeaconSession) GetGroupsRegistry() ([][32]byte, error) {\n\treturn _RandomBeacon.Contract.GetGroupsRegistry(&_RandomBeacon.CallOpts)\n}", "func (_Harberger *HarbergerCaller) GetEVMScriptRegistry(opts *bind.CallOpts) (common.Address, error) {\n\tvar (\n\t\tret0 = new(common.Address)\n\t)\n\tout := ret0\n\terr := _Harberger.contract.Call(opts, out, \"getEVMScriptRegistry\")\n\treturn *ret0, err\n}", "func GetDefaultRegistry() *RegistryConfig {\n\tnewReg := DefaultRegistry\n\treturn &newReg\n}", "func (c *LiveCollector) Registry() fslib.FS {\n\treturn c.registryfs\n}", "func (d *Devices) Registry() *RegistryServiceClient {\n\treturn &RegistryServiceClient{getConn: d.getConn}\n}", "func NewRegistry() *Registry {\n\treturn new(Registry)\n}", "func (_WyvernExchange *WyvernExchangeCaller) Registry(opts *bind.CallOpts) (common.Address, error) {\n\tvar (\n\t\tret0 = new(common.Address)\n\t)\n\tout := ret0\n\terr := _WyvernExchange.contract.Call(opts, out, \"registry\")\n\treturn *ret0, err\n}", "func NewRegistry() *Registry {\n\tt := Registry{}\n\tt.Reset()\n\n\treturn &t\n}", "func GetGlobalLoggerRegistry() LoggerRegistry { return globalLoggerRegistry }", "func (e *PollerRegistry) Get(key string) (*poller.Poller, error) {\n\tif val, ok := e.Registry[key]; ok {\n\t\treturn val, nil\n\t}\n\treturn nil, ErrElementNotFound\n}", "func (moa *MultiOrgAlertmanager) GetOrCreateOrgRegistry(id int64) prometheus.Registerer {\n\treturn moa.registries.GetOrCreateOrgRegistry(id)\n}", "func NewRegistry() *Registry {\n\tregistry := Registry{}\n\tregistry.lock = &sync.Mutex{}\n\treturn &registry\n}", "func (o *VersionedFlowCoordinates) GetRegistryUrl() string {\n\tif o == nil || o.RegistryUrl == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.RegistryUrl\n}", "func (r *DefaultInstallerRegistry) Registry(name string, installer core.Installer) {\n\t_, ok := r.installerMap[name]\n\tif ok {\n\t\tr.installerMap[name] = append(r.installerMap[name], installer)\n\t} else {\n\t\tr.installerMap[name] = []core.Installer{installer}\n\t}\n}", "func GetRegistryClient(locator string) RegistryClient {\n\tif mockClient != nil {\n\t\treturn mockClient\n\t}\n\tclient := &Tars.Tarsregistry{}\n\tif err := StringToProxy(locator, \"tars.tarsregistry.Registry\", client); err != nil {\n\t\treturn nil\n\t}\n\tclient.TarsSetTimeout(rpcTimeout)\n\timpClient = &registryClientImp{\n\t\tclient: client,\n\t\tretry: retry.New(retry.MaxTimeoutOpt(time.Second*100, time.Second*3)),\n\t}\n\treturn impClient\n}", "func (s *Store) Registry() *metric.Registry {\n\treturn s.metrics.registry\n}", "func DefaultRegistry() *Registry {\n\treturn registry\n}", "func GetRegistrySamples(registry string) ([]byte, error) {\n\tif registry == DEVFILE_REGISTRY_PLACEHOLDER_URL {\n\t\treturn []byte(SamplePlaceholderJSON), nil\n\t} else {\n\t\treturn nil, fmt.Errorf(\"registry %s is invalid\", registry)\n\t}\n}", "func (c *Cache) Get(key string) *Registry {\n\treturn c.Map[key]\n}", "func RegistryGetHRMap() map[string]MX4JMetric {\n\treglock.RLock()\n\tdefer reglock.RUnlock()\n\n\tmetrics := make(map[string]MX4JMetric)\n\tfor _, mm := range registry {\n\t\tmetrics[mm.HumanName] = mm\n\t}\n\treturn metrics\n}", "func NewGetModelRegistryNotFound() *GetModelRegistryNotFound {\n\treturn &GetModelRegistryNotFound{}\n}", "func (_Harberger *HarbergerSession) GetEVMScriptRegistry() (common.Address, error) {\n\treturn _Harberger.Contract.GetEVMScriptRegistry(&_Harberger.CallOpts)\n}", "func ExampleRegistriesClient_Get() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclientFactory, err := armcontainerregistry.NewClientFactory(\"<subscription-id>\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tres, err := clientFactory.NewRegistriesClient().Get(ctx, \"myResourceGroup\", \"myRegistry\", nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\t// You could use response here. We use blank identifier for just demo purposes.\n\t_ = res\n\t// If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes.\n\t// res.Registry = armcontainerregistry.Registry{\n\t// \tName: to.Ptr(\"myRegistry\"),\n\t// \tType: to.Ptr(\"Microsoft.ContainerRegistry/registries\"),\n\t// \tID: to.Ptr(\"/subscriptions/00000000-0000-0000-0000-000000000000/resourceGroups/myResourceGroup/providers/Microsoft.ContainerRegistry/registries/myRegistry\"),\n\t// \tLocation: to.Ptr(\"westus\"),\n\t// \tTags: map[string]*string{\n\t// \t\t\"key\": to.Ptr(\"value\"),\n\t// \t},\n\t// \tProperties: &armcontainerregistry.RegistryProperties{\n\t// \t\tAdminUserEnabled: to.Ptr(false),\n\t// \t\tAnonymousPullEnabled: to.Ptr(false),\n\t// \t\tCreationDate: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2021-06-15T21:38:26.1537861Z\"); return t}()),\n\t// \t\tDataEndpointEnabled: to.Ptr(false),\n\t// \t\tDataEndpointHostNames: []*string{\n\t// \t\t},\n\t// \t\tEncryption: &armcontainerregistry.EncryptionProperty{\n\t// \t\t\tStatus: to.Ptr(armcontainerregistry.EncryptionStatusDisabled),\n\t// \t\t},\n\t// \t\tLoginServer: to.Ptr(\"myRegistry.azurecr-test.io\"),\n\t// \t\tNetworkRuleBypassOptions: to.Ptr(armcontainerregistry.NetworkRuleBypassOptionsAzureServices),\n\t// \t\tNetworkRuleSet: &armcontainerregistry.NetworkRuleSet{\n\t// \t\t\tDefaultAction: to.Ptr(armcontainerregistry.DefaultActionAllow),\n\t// \t\t\tIPRules: []*armcontainerregistry.IPRule{\n\t// \t\t\t},\n\t// \t\t},\n\t// \t\tPolicies: &armcontainerregistry.Policies{\n\t// \t\t\tExportPolicy: &armcontainerregistry.ExportPolicy{\n\t// \t\t\t\tStatus: to.Ptr(armcontainerregistry.ExportPolicyStatusEnabled),\n\t// \t\t\t},\n\t// \t\t\tQuarantinePolicy: &armcontainerregistry.QuarantinePolicy{\n\t// \t\t\t\tStatus: to.Ptr(armcontainerregistry.PolicyStatusDisabled),\n\t// \t\t\t},\n\t// \t\t\tRetentionPolicy: &armcontainerregistry.RetentionPolicy{\n\t// \t\t\t\tDays: to.Ptr[int32](7),\n\t// \t\t\t\tLastUpdatedTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2021-06-15T21:40:12.8506835+00:00\"); return t}()),\n\t// \t\t\t\tStatus: to.Ptr(armcontainerregistry.PolicyStatusDisabled),\n\t// \t\t\t},\n\t// \t\t\tTrustPolicy: &armcontainerregistry.TrustPolicy{\n\t// \t\t\t\tType: to.Ptr(armcontainerregistry.TrustPolicyTypeNotary),\n\t// \t\t\t\tStatus: to.Ptr(armcontainerregistry.PolicyStatusDisabled),\n\t// \t\t\t},\n\t// \t\t},\n\t// \t\tPrivateEndpointConnections: []*armcontainerregistry.PrivateEndpointConnection{\n\t// \t\t},\n\t// \t\tProvisioningState: to.Ptr(armcontainerregistry.ProvisioningStateSucceeded),\n\t// \t\tPublicNetworkAccess: to.Ptr(armcontainerregistry.PublicNetworkAccessEnabled),\n\t// \t\tZoneRedundancy: to.Ptr(armcontainerregistry.ZoneRedundancyDisabled),\n\t// \t},\n\t// \tSKU: &armcontainerregistry.SKU{\n\t// \t\tName: to.Ptr(armcontainerregistry.SKUNameStandard),\n\t// \t\tTier: to.Ptr(armcontainerregistry.SKUTierStandard),\n\t// \t},\n\t// }\n}", "func NewRegistry(s rest.StandardStorage) Registry {\n\treturn &storage{s}\n}", "func (s *appConnectorE2EState) GetRegistryClient() *testkit.RegistryClient {\n\treturn s.registryClient\n}", "func (daemon *Daemon) RegistryService() *registry.Service {\n\treturn daemon.registryService\n}", "func GetContainerRegistry(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *ContainerRegistryState, opts ...pulumi.ResourceOption) (*ContainerRegistry, error) {\n\tvar resource ContainerRegistry\n\terr := ctx.ReadResource(\"sakuracloud:index/containerRegistry:ContainerRegistry\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (a *Client) ListRegistry(params *ListRegistryParams) (*ListRegistryOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewListRegistryParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"listRegistry\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/Registries\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &ListRegistryReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*ListRegistryOK), nil\n\n}", "func (m *OrgRegistries) GetOrCreateOrgRegistry(orgID int64) prometheus.Registerer {\n\tm.regsMu.Lock()\n\tdefer m.regsMu.Unlock()\n\n\torgRegistry, ok := m.regs[orgID]\n\tif !ok {\n\t\treg := prometheus.NewRegistry()\n\t\tm.regs[orgID] = reg\n\t\treturn reg\n\t}\n\treturn orgRegistry\n}", "func newRegistry(logger logr.Logger, config globalregistry.RegistryConfig) (globalregistry.Registry, error) {\n\tvar err error\n\tc := &registry{\n\t\tlogger: logger,\n\t\tRegistryConfig: config,\n\t\tClient: http.DefaultClient,\n\t}\n\tc.projects, err = newProjectAPI(c)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tc.remoteRegistries = newRemoteRegistries(c)\n\tc.replications = newReplicationAPI(c)\n\tc.parsedUrl, err = url.Parse(config.GetAPIEndpoint())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tc.scanners = newScannerAPI(c)\n\treturn c, nil\n}", "func (r *GoMetricsRegistry) GetOrRegister(name string, metric interface{}) interface{} {\n\tr.mutex.Lock()\n\tdefer r.mutex.Unlock()\n\n\tv := r.get(name)\n\tif v != nil {\n\t\treturn v\n\t}\n\n\treturn r.doRegister(name, metric)\n}", "func NewRegistry() Registry {\n\tfr := make(Registry)\n\tfr.mustRegister(BuiltIns...)\n\treturn fr\n}", "func NewRegistry() *Registry {\n\treturn &Registry{\n\t\tServicesMap: make(map[string]ServiceList),\n\t}\n}", "func validateAndFetchRegistry(registryName string) (catalog.DevfileComponentTypeList, error) {\n\t// TODO(feloy) Get from DI\n\tprefClient, err := preference.NewClient()\n\tif err != nil {\n\t\todoutil.LogErrorAndExit(err, \"unable to set preference, something is wrong with odo, kindly raise an issue at https://github.com/redhat-developer/odo/issues/new?template=Bug.md\")\n\t}\n\tcatalogClient := catalog.NewCatalogClient(filesystem.DefaultFs{}, prefClient)\n\n\t// Validate if the component type is available\n\tif registryName != \"\" {\n\t\tregistryExistSpinner := log.Spinnerf(\"Checking if the registry %q exists\", registryName)\n\t\tdefer registryExistSpinner.End(false)\n\t\tregistryList, e := catalogClient.GetDevfileRegistries(registryName)\n\t\tif e != nil {\n\t\t\treturn catalog.DevfileComponentTypeList{}, errors.Wrap(e, \"failed to get registry\")\n\t\t}\n\t\tif len(registryList) == 0 {\n\t\t\treturn catalog.DevfileComponentTypeList{}, errors.Errorf(\"registry %s doesn't exist, please specify a valid registry via --registry\", registryName)\n\t\t}\n\t\tregistryExistSpinner.End(true)\n\t}\n\n\tklog.V(4).Infof(\"Fetching the available devfile components\")\n\t// Get available devfile components for checking devfile compatibility\n\tcatalogDevfileList, err := catalogClient.ListDevfileComponents(registryName)\n\tif err != nil {\n\t\treturn catalog.DevfileComponentTypeList{}, err\n\t}\n\n\tif registryName != \"\" && catalogDevfileList.Items == nil {\n\t\treturn catalog.DevfileComponentTypeList{}, errors.Errorf(\"can't create devfile component from registry %s\", registryName)\n\t}\n\n\tif len(catalogDevfileList.DevfileRegistries) == 0 {\n\t\treturn catalog.DevfileComponentTypeList{}, errors.New(\"Registry is empty, please run `odo registry add <registry name> <registry URL>` to add a registry\\n\")\n\t}\n\treturn catalogDevfileList, nil\n}", "func (_Harberger *HarbergerCallerSession) GetEVMScriptRegistry() (common.Address, error) {\n\treturn _Harberger.Contract.GetEVMScriptRegistry(&_Harberger.CallOpts)\n}", "func regGetString(hKey HKEY, subKey string, value string) string {\r\n\tvar bufLen uint32\r\n\tprocRegGetValue.Call(\r\n\t\tuintptr(hKey),\r\n\t\tuintptr(unsafe.Pointer(syscall.StringToUTF16Ptr(subKey))),\r\n\t\tuintptr(unsafe.Pointer(syscall.StringToUTF16Ptr(value))),\r\n\t\tuintptr(RRF_RT_REG_SZ),\r\n\t\t0,\r\n\t\t0,\r\n\t\tuintptr(unsafe.Pointer(&bufLen)))\r\n\r\n\tif bufLen == 0 {\r\n\t\treturn \"\"\r\n\t}\r\n\r\n\tbuf := make([]uint16, bufLen)\r\n\tret, _, _ := procRegGetValue.Call(\r\n\t\tuintptr(hKey),\r\n\t\tuintptr(unsafe.Pointer(syscall.StringToUTF16Ptr(subKey))),\r\n\t\tuintptr(unsafe.Pointer(syscall.StringToUTF16Ptr(value))),\r\n\t\tuintptr(RRF_RT_REG_SZ),\r\n\t\t0,\r\n\t\tuintptr(unsafe.Pointer(&buf[0])),\r\n\t\tuintptr(unsafe.Pointer(&bufLen)))\r\n\r\n\tif ret != 0 {\r\n\t\treturn \"\"\r\n\t}\r\n\r\n\treturn syscall.UTF16ToString(buf)\r\n}", "func NewRegistry() *bsoncodec.Registry {\n\treturn NewRegistryBuilder().Build()\n}", "func Get(name string) Logger {\n\tregMutex.Lock()\n\tdefer regMutex.Unlock()\n\tl, ok := reg[name]\n\tif ok {\n\t\treturn l\n\t}\n\tl = New()\n\treg[name] = l\n\treturn l\n}", "func (m Message) GetInstrRegistry(f *field.InstrRegistryField) quickfix.MessageRejectError {\n\treturn m.Body.Get(f)\n}", "func (m Message) GetInstrRegistry(f *field.InstrRegistryField) quickfix.MessageRejectError {\n\treturn m.Body.Get(f)\n}", "func (m Message) GetInstrRegistry(f *field.InstrRegistryField) quickfix.MessageRejectError {\n\treturn m.Body.Get(f)\n}", "func (m Message) GetInstrRegistry(f *field.InstrRegistryField) quickfix.MessageRejectError {\n\treturn m.Body.Get(f)\n}", "func NewRegistry(ctx *pulumi.Context,\n\tname string, args *RegistryArgs, opts ...pulumi.ResourceOption) (*Registry, error) {\n\tif args == nil {\n\t\targs = &RegistryArgs{}\n\t}\n\tvar resource Registry\n\terr := ctx.RegisterResource(\"gcp:kms/registry:Registry\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func DefaultRegistry() *Registry {\n\treturn &defaultRegistry\n}", "func (fn DefaultRegistryFunc) DefaultRegistry() (string, bool) {\n\treturn fn()\n}", "func NewRegistry() *Registry {\n\treturn &Registry{\n\t\tall: make(map[string]*Job),\n\t}\n}", "func NewRegistry(s Storage) Registry {\n\treturn &storage{s}\n}", "func NewRegistry(s Storage) Registry {\n\treturn &storage{s}\n}", "func (c *ImageRegistryCollection) Get(name common.ID) (*ImageRegistryResource, error) {\n\tr := c.New()\n\tif err := c.core.db.get(c, name, r); err != nil {\n\t\treturn nil, err\n\t}\n\treturn r, nil\n}", "func (o IopingSpecVolumeVolumeSourceQuobytePtrOutput) Registry() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *IopingSpecVolumeVolumeSourceQuobyte) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Registry\n\t}).(pulumi.StringPtrOutput)\n}", "func NewRegistry() *Registry {\n\treturn &Registry{\n\t\toutlets: make([]*Outlet, 0),\n\t\toutletMap: make(map[string]*Outlet),\n\t\tgroups: make([]*Group, 0),\n\t\tgroupMap: make(map[string]*Group),\n\t}\n}", "func GetGatherer() prometheus.Gatherer {\n\treturn registry\n}", "func registerCRGet(L *lua.LState) int {\n\tp := checkRegisterCR(L, 1)\n\tfmt.Println(p)\n\n\treturn 0\n}", "func (sr *ServicedStatsReporter) getOrCreateContainerRegistry(serviceID string, instanceID int) metrics.Registry {\n\tkey := registryKey{serviceID, instanceID}\n\tif registry, ok := sr.containerRegistries[key]; ok {\n\t\treturn registry\n\t}\n\tsr.Lock()\n\tdefer sr.Unlock()\n\tsr.containerRegistries[key] = metrics.NewRegistry()\n\treturn sr.containerRegistries[key]\n}", "func GetPrometheusRegistry() *prometheus.Registry {\n\treturn prometheusRegistry\n}", "func NewRegistry() *Registry {\n\treturn &Registry{\n\t\ttypes: make(map[string]reflect.Type),\n\t\tmu: sync.Mutex{},\n\t}\n}", "func NewRegistry(ctx context.Context, config *configuration.Configuration) (*Registry, error) {\n\tvar err error\n\tctx, err = configureLogging(ctx, config)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error configuring logger: %v\", err)\n\t}\n\n\tconfigureBugsnag(config)\n\n\t// inject a logger into the uuid library. warns us if there is a problem\n\t// with uuid generation under low entropy.\n\tuuid.Loggerf = dcontext.GetLogger(ctx).Warnf\n\n\tapp := handlers.NewApp(ctx, config)\n\t// TODO(aaronl): The global scope of the health checks means NewRegistry\n\t// can only be called once per process.\n\tapp.RegisterHealthChecks()\n\thandler := configureReporting(app)\n\thandler = alive(\"/\", handler)\n\thandler = health.Handler(handler)\n\thandler = panicHandler(handler)\n\tif !config.Log.AccessLog.Disabled {\n\t\thandler = gorhandlers.CombinedLoggingHandler(os.Stdout, handler)\n\t}\n\n\tfor _, applyHandlerMiddleware := range handlerMiddlewares {\n\t\thandler = applyHandlerMiddleware(config, handler)\n\t}\n\n\tserver := &http.Server{\n\t\tHandler: handler,\n\t}\n\n\treturn &Registry{\n\t\tapp: app,\n\t\tconfig: config,\n\t\tserver: server,\n\t}, nil\n}", "func (r *Registry) Get(baseURL string) *rate.Limiter {\n\treturn r.GetOrSet(baseURL, nil)\n}", "func NewRegistry() *Registry {\n\tr := &Registry{}\n\tr.functions = make(map[string]*function)\n\treturn r\n}", "func NewRegistry() *Registry {\n\treturn &Registry{\n\t\ttypeEncoders: new(typeEncoderCache),\n\t\ttypeDecoders: new(typeDecoderCache),\n\t\tkindEncoders: new(kindEncoderCache),\n\t\tkindDecoders: new(kindDecoderCache),\n\t}\n}", "func (opts *RegistryOption) getRegistryDigest(imageName string) *digest.Digest {\n\n\t_, exists := opts.DigestChache[imageName]\n\tif exists {\n\t\treturn opts.DigestChache[imageName]\n\t}\n\n\treference, err := docker.ParseReference(\"//\" + imageName)\n\tif err != nil {\n\t\tlog.Fatalf(\"Can't parse reference on %v, because %v\", imageName, err)\n\t}\n\n\tctx := context.Background()\n\n\timage, err := reference.NewImage(ctx, opts.SysCtx)\n\tif err != nil {\n\t\tlog.Fatalf(\"Can't create new Image, because %v\", err)\n\t}\n\n\t// Close Image on exit\n\tdefer func(image types.ImageCloser) {\n\t\terr := image.Close()\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Can't close Image %v, because %v\", image, err)\n\t\t}\n\t}(image)\n\n\tmanifestBytes, _, err := image.Manifest(ctx)\n\tif err != nil {\n\t\tlog.Errorf(\"Can't get manifest of image %v, because %v\", imageName, err)\n\t}\n\n\timageDigest, err := manifest.Digest(manifestBytes)\n\tif err != nil {\n\t\tlog.Errorf(\"Can't get Digest of image %v, because %v\", imageName, err)\n\t}\n\n\topts.DigestChache[imageName] = &imageDigest\n\n\treturn &imageDigest\n}", "func (o FioSpecVolumeVolumeSourceQuobytePtrOutput) Registry() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *FioSpecVolumeVolumeSourceQuobyte) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Registry\n\t}).(pulumi.StringPtrOutput)\n}", "func (r DockerhubRegistry) GetCatalog() (*RegistryCatalog, error) {\n\trepos := r.RepoConfig.Repositories\n\tbaseURL := r.RepoConfig.AddressOrDefault(dockerhubApiBase)\n\ttoken := \"\"\n\tif r.RepoConfig.Password != \"\" && r.RepoConfig.Username != \"\" {\n\t\ttok, err := r.getAuthToken()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\ttoken = tok\n\t}\n\tfor _, org := range r.RepoConfig.Orgs {\n\t\tcatalog, err := getRepositoryPage(fmt.Sprintf(\"%s/repositories/%s?page=1&page_size=2\", baseURL, org), token)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\trepos = append(repos, catalog...)\n\t}\n\n\treturn &RegistryCatalog{Repositories: repos}, nil\n}", "func (o IopingSpecVolumeVolumeSourceQuobyteOutput) Registry() pulumi.StringOutput {\n\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceQuobyte) string { return v.Registry }).(pulumi.StringOutput)\n}", "func DetectEnvRegistry(reg *RegistryObj) (error) {\n if reg.Env == \"\" || registry_struct[reg.Env].Environment == \"\" {\n return errors.New(\"Environment registry not found\")\n }\n\n return nil\n}", "func getConfig(ctx *types.SystemContext) (*V2RegistriesConf, error) {\n\tconfigPath := ConfigPath(ctx)\n\n\tconfigMutex.Lock()\n\t// if the config has already been loaded, return the cached registries\n\tif config, inCache := configCache[configPath]; inCache {\n\t\tconfigMutex.Unlock()\n\t\treturn config, nil\n\t}\n\tconfigMutex.Unlock()\n\n\treturn TryUpdatingCache(ctx)\n}", "func GetRegistryCtlURL() string {\n\turl := os.Getenv(\"REGISTRY_CONTROLLER_URL\")\n\tif len(url) == 0 {\n\t\treturn common.DefaultRegistryCtlURL\n\t}\n\treturn url\n}", "func Get(name string) *Subcommand {\n\treturn defaultRegistry.get(name)\n}", "func NewRegistry(url *url.URL) *Registry {\n\tclient := Client{BaseURL: url}\n\treturn &Registry{client: &client}\n}", "func NewRegistry() *Registry {\n\treturn &Registry{\n\t\trateLimiters: make(map[string]*rate.Limiter),\n\t}\n}", "func getRegBin(root registry.Key, path string, name string) ([]byte, error) {\n\tkey, err := registry.OpenKey(root, path, registry.READ)\n\tdefer key.Close()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbin, _, err := key.GetBinaryValue(name)\n\treturn bin, err\n}", "func (cc *CredsConfig) GetRegistryCreds(id string) map[string]string {\n\treturn cc.Registry[id]\n}" ]
[ "0.76218796", "0.7515332", "0.7475501", "0.72384256", "0.68892515", "0.68240374", "0.6817327", "0.6678539", "0.653758", "0.64635944", "0.6416455", "0.6396177", "0.63797766", "0.635688", "0.63549465", "0.6328571", "0.6263573", "0.61694133", "0.61538637", "0.6127399", "0.61179096", "0.6113246", "0.6050782", "0.6046544", "0.60286635", "0.5998846", "0.597463", "0.59625405", "0.5939577", "0.5930443", "0.5921979", "0.58521986", "0.58453774", "0.58281094", "0.58093166", "0.57685596", "0.5741447", "0.57294", "0.5700417", "0.5659892", "0.55986893", "0.5594007", "0.5584448", "0.5559802", "0.5552793", "0.55477417", "0.5543554", "0.55132014", "0.5509721", "0.55011505", "0.54935396", "0.5489555", "0.5485015", "0.5477401", "0.5462802", "0.544941", "0.54449785", "0.5444726", "0.5434401", "0.54275113", "0.5425343", "0.5424576", "0.54214907", "0.54203403", "0.5418669", "0.5418409", "0.5414096", "0.5414096", "0.5414096", "0.5414096", "0.5412314", "0.5409069", "0.5407007", "0.54001063", "0.53816444", "0.53816444", "0.5378534", "0.5374073", "0.5367559", "0.5359406", "0.5356114", "0.53503937", "0.53423053", "0.53398246", "0.53372455", "0.5308703", "0.5297366", "0.5291887", "0.528959", "0.5274985", "0.5268717", "0.52666366", "0.52565444", "0.52561396", "0.5255249", "0.52518374", "0.5247768", "0.5228363", "0.5221987", "0.5220473" ]
0.7815176
0
CreateAPIConfig returns router.API struct base on the input
func CreateAPIConfig(urlPattern string, dboBackendConfig config.DubboBackendConfig, methodString string, mappingParams []config.MappingParam) router.API { dboBackendConfig.Method = methodString url := strings.Join([]string{urlPattern, methodString}, constant.PathSlash) method := config.Method{ Enable: true, Timeout: 3 * time.Second, Mock: false, HTTPVerb: config.MethodPost, InboundRequest: config.InboundRequest{ RequestType: config.HTTPRequest, }, IntegrationRequest: config.IntegrationRequest{ RequestType: config.DubboRequest, DubboBackendConfig: dboBackendConfig, MappingParams: mappingParams, }, } return router.API{ URLPattern: url, Method: method, } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func CreateAPI(config JSONConfigurationAPI, insecure bool) *OsctrlAPI {\n\tvar a *OsctrlAPI\n\t// Prepare URL\n\tu, err := url.Parse(config.URL)\n\tif err != nil {\n\t\tlog.Fatalf(\"invalid url: %v\", err)\n\t}\n\t// Define client with correct TLS settings\n\tclient := &http.Client{}\n\tif u.Scheme == \"https\" {\n\t\tcertPool, err := x509.SystemCertPool()\n\t\tif err != nil {\n\t\t\tlog.Fatalf(\"error loading x509 certificate pool: %v\", err)\n\t\t}\n\t\ttlsCfg := &tls.Config{RootCAs: certPool}\n\t\tif insecure {\n\t\t\ttlsCfg.InsecureSkipVerify = true\n\t\t}\n\t\tclient.Transport = &http.Transport{TLSClientConfig: tlsCfg}\n\t}\n\t// Prepare authentication\n\theaders := make(map[string]string)\n\theaders[Authorization] = fmt.Sprintf(\"Bearer %s\", config.Token)\n\theaders[ContentType] = JSONApplicationUTF8\n\ta = &OsctrlAPI{\n\t\tConfiguration: config,\n\t\tClient: client,\n\t\tHeaders: headers,\n\t}\n\treturn a\n}", "func NewAPI(\n\tpromURL string,\n\tproxy *Proxy,\n\treadConfig func() ([]byte, error),\n\tlg logrus.FieldLogger) *API {\n\tw := &API{\n\t\tConfigReload: make(chan *config.Config, 2),\n\t\tTargetReload: make(chan map[string][]*target.Target, 2),\n\t\tEngine: gin.Default(),\n\t\tlg: lg,\n\t\tpromURL: promURL,\n\t\tpromCli: prom.NewClient(promURL),\n\t\tproxy: proxy,\n\t}\n\n\tw.GET(w.path(\"/api/v1/shard/runtimeinfo/\"), api.Wrap(w.lg, w.runtimeInfo))\n\tw.GET(w.path(\"/api/v1/shard/targets/\"), api.Wrap(w.lg, w.getTargets))\n\tw.POST(w.path(\"/api/v1/shard/targets/\"), api.Wrap(w.lg, w.updateTargets))\n\tw.GET(w.path(\"/api/v1/status/config/\"), api.Wrap(lg, func(ctx *gin.Context) *api.Result {\n\t\treturn prom.APIReadConfig(readConfig)\n\t}))\n\tw.POST(w.path(\"/-/reload/\"), api.Wrap(lg, func(ctx *gin.Context) *api.Result {\n\t\treturn prom.APIReloadConfig(readConfig, w.ConfigReload)\n\t}))\n\treturn w\n}", "func createOpenAPIBuilderConfig() *common.Config {\n\treturn &common.Config{\n\t\tProtocolList: []string{\"https\"},\n\t\tIgnorePrefixes: []string{\"/swaggerapi\"},\n\t\tInfo: &spec.Info{\n\t\t\tInfoProps: spec.InfoProps{\n\t\t\t\tTitle: \"Argo-Events\",\n\t\t\t\tVersion: \"v0.6\",\n\t\t\t},\n\t\t},\n\t}\n}", "func NewAPI(config config.Config, logger log.Logger) *API {\n\tapi := &API{\n\t\tmux.NewRouter().StrictSlash(false),\n\t\tnewAuthMiddleware(config, logger),\n\t\tlog.NewLoggerWith(logger, \"component\", \"api\"),\n\t\tconfig,\n\t}\n\n\tapi.addRoute(\"\", http.MethodGet, \"/\", func(w http.ResponseWriter, r *http.Request) {\n\t\tif err := json.NewEncoder(w).Encode(NewAPIInfo(config.ExternalURL)); err != nil {\n\t\t\tjson.NewEncoder(w).Encode(Error{Code: 500, Message: err.Error()})\n\t\t}\n\t})\n\n\treturn api\n}", "func NewAPIFromConfig(c Config) API {\n\treturn NewNameAPI(c.Username, c.Token, c.Dev)\n}", "func CreateAPI(s server.AtlasServer) pb.AtlasServer {\n\treturn api{s: s}\n}", "func NewAPI(config *config.APIConfig) (*API, error) {\n\tapi := &API{}\n\tapi.log = log.NewLogger(&config.Log, \"api\").WithFields(logrus.Fields{\n\t\tlog.BaseURLField: config.BaseURL,\n\t\tlog.PortField: config.HTTP.Port,\n\t\tlog.UseTLSField: config.HTTP.UseTLS,\n\t\tlog.CertFileField: config.HTTP.CertFile,\n\t\tlog.KeyFileField: config.HTTP.KeyFile,\n\t})\n\tcontextLogger := api.log.WithField(log.FuncField, ref.GetFuncName())\n\n\tserver, err := createAPIServer(&config.HTTP, api)\n\tif err != nil {\n\t\tcontextLogger.WithError(err).Error(\"error creating mock API\")\n\t\treturn nil, err\n\t}\n\n\tapi.server = wrapper.NewServerOps(server)\n\tapi.baseURL = config.BaseURL\n\tapi.endpoints = config.Endpoints\n\tapi.handlers = make(map[string]map[string]func(http.ResponseWriter, *http.Request))\n\tapi.routeTree = route.NewRouteTree()\n\tapi.httpConfig = config.HTTP\n\tapi.file = &wrapper.FileOps{}\n\tapi.creator = newCreator(api.log)\n\n\tcontextLogger.Info(\"successfully created mock API\")\n\treturn api, nil\n}", "func newApiDefinitionFromAPI(a utils.API) *api {\n\treturn &api{a.ID, a.Name, a.Context, a.Version, a.Provider,\n\t\ta.LifeCycleStatus}\n}", "func createConfig(apiServerURL, kubeCfgPath string, qps float32, burst int) (*rest.Config, error) {\n\tvar (\n\t\tconfig *rest.Config\n\t\terr error\n\t)\n\tcmdName := \"cilium\"\n\tif len(os.Args[0]) != 0 {\n\t\tcmdName = filepath.Base(os.Args[0])\n\t}\n\tuserAgent := fmt.Sprintf(\"%s/%s\", cmdName, version.Version)\n\n\tswitch {\n\t// If the apiServerURL and the kubeCfgPath are empty then we can try getting\n\t// the rest.Config from the InClusterConfig\n\tcase apiServerURL == \"\" && kubeCfgPath == \"\":\n\t\tif config, err = rest.InClusterConfig(); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\tcase kubeCfgPath != \"\":\n\t\tif config, err = clientcmd.BuildConfigFromFlags(\"\", kubeCfgPath); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\tcase strings.HasPrefix(apiServerURL, \"https://\"):\n\t\tif config, err = rest.InClusterConfig(); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tconfig.Host = apiServerURL\n\tdefault:\n\t\tconfig = &rest.Config{Host: apiServerURL, UserAgent: userAgent}\n\t}\n\n\tsetConfig(config, userAgent, qps, burst)\n\treturn config, nil\n}", "func NewApiConfig() *ApiConfig {\n\tmodel := new(ApiConfig)\n\n\treturn model\n}", "func (k *Kong) CreateAPI(api Data) error {\n\t// Create the api object\n\tcreateAPI := &API{Name: api.Name, Upstream: api.UpstreamURL, Hosts: strings.Join(api.Hosts, \",\")}\n\n\t// Create JSON\n\tbuf := new(bytes.Buffer)\n\tjson.NewEncoder(buf).Encode(createAPI)\n\n\t// Setup URL\n\turl := fmt.Sprintf(\"%s/apis/\", k.KongAdminURL)\n\n\tresp, err := k.client.Post(url, \"application/json\", buf)\n\n\tif err != nil {\n\t\tlogrus.Error(\"Could not create kong api: \", err)\n\t\treturn err\n\t}\n\n\tdefer resp.Body.Close()\n\n\tif resp.StatusCode != 201 {\n\t\tbodyBytes, _ := ioutil.ReadAll(resp.Body)\n\t\tlogrus.Errorf(\"Create api returned: %d Response: %s\", resp.StatusCode, string(bodyBytes))\n\t\treturn fmt.Errorf(\"Create api returned: %d\", resp.StatusCode)\n\t}\n\n\tlogrus.Infof(\"Created api: %s\", api.Name)\n\n\treturn nil\n}", "func CreateConfig(apiServerURL, kubeCfgPath string) (*rest.Config, error) {\n\tvar (\n\t\tconfig *rest.Config\n\t\terr error\n\t)\n\n\tswitch {\n\t// If the apiServerURL and the kubeCfgPath are empty then we can try getting\n\t// the rest.Config from the InClusterConfig\n\tcase apiServerURL == \"\" && kubeCfgPath == \"\":\n\t\tif config, err = inClusterConfig(); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\tcase kubeCfgPath != \"\":\n\t\tif config, err = clientcmd.BuildConfigFromFlags(\"\", kubeCfgPath); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\tcase strings.HasPrefix(apiServerURL, \"https://\"):\n\t\tif config, err = rest.InClusterConfig(); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tconfig.Host = apiServerURL\n\tdefault:\n\t\tconfig = &rest.Config{Host: apiServerURL}\n\t}\n\n\treturn config, nil\n}", "func New(c *core.Core, r *mux.Router, version string) *API {\n\ta := &API{}\n\ta.AppVersion = version\n\ta.c = c\n\ta.reqDecoder = schema.NewDecoder()\n\twireupRoutes(r, a)\n\treturn a\n}", "func NewAPI(mySQLConfig mysqlDriver.Config, logWriter io.Writer) (*API, error) {\n\tdb, err := sql.Open(\"mysql\", mySQLConfig.FormatDSN())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tapiObj := API{\n\t\trouter: mux.NewRouter(),\n\t\tdb: db,\n\t\tlogWriter: logWriter,\n\t}\n\n\t// Room\n\troomDBFacade := mysqlEntities.NewRoomMySQL(apiObj.db)\n\troomRoutes := api.CreateRoomRoutes(roomDBFacade)\n\tapiObj.CreateRoutes(roomRoutes...)\n\n\t// Session\n\tsessionDBFacade := mysqlEntities.NewSessionMySQL(apiObj.db)\n\tsessionRoutes := api.CreateSessionRoutes(sessionDBFacade)\n\tapiObj.CreateRoutes(sessionRoutes...)\n\n\t// Speaker\n\tspeakerDBFacade := mysqlEntities.NewSpeakerMySQL(apiObj.db)\n\tspeakerRoutes := api.CreateSpeakerRoutes(speakerDBFacade)\n\tapiObj.CreateRoutes(speakerRoutes...)\n\n\t// Timeslot\n\ttimeslotDBFacade := mysqlEntities.NewTimeslotMySQL(apiObj.db)\n\ttimeslotRoutes := api.CreateTimeslotRoutes(timeslotDBFacade)\n\tapiObj.CreateRoutes(timeslotRoutes...)\n\n\t// Count\n\tcountDBFacade := mysqlEntities.NewCountMySQL(apiObj.db)\n\tcountRoutes := api.CreateCountRoutes(countDBFacade)\n\tapiObj.CreateRoutes(countRoutes...)\n\n\treturn &apiObj, nil\n}", "func NewAPIConfig(ids, keys []string) APIConfig {\n\tif len(ids) != len(keys) || len(ids) == 0 {\n\t\tpanic(errors.New(\"length of ids and keys should be equal and not 0\"))\n\t}\n\n\tlq, err := queue.NewLinkedQueue()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tfor i, id := range ids {\n\t\tlq.Add(Node{\n\t\t\tid: id,\n\t\t\tkey: keys[i],\n\t\t})\n\t}\n\n\treturn APIConfig{\n\t\tlqueue: lq,\n\t}\n}", "func createRouterConfig(routeMode int) *conf.RouterConfig {\n\tdomainStrategy := \"IPIfNonMatch\"\n\tbypassLAN, _ := json.Marshal(v2ray.Rules{\n\t\tType: \"field\",\n\t\tOutboundTag: \"direct\",\n\t\tIP: []string{\"geoip:private\"},\n\t})\n\tbypassChinaIP, _ := json.Marshal(v2ray.Rules{\n\t\tType: \"field\",\n\t\tOutboundTag: \"direct\",\n\t\tIP: []string{\"geoip:cn\"},\n\t})\n\tbypassChinaSite, _ := json.Marshal(v2ray.Rules{\n\t\tType: \"field\",\n\t\tOutboundTag: \"direct\",\n\t\tDomain: []string{\"geosite:cn\"},\n\t})\n\tblockDomain, _ := json.Marshal(v2ray.Rules{\n\t\tType: \"field\",\n\t\tOutboundTag: \"blocked\",\n\t\tDomain: v2ray.BlockDomains,\n\t})\n\t// blockAd, _ := json.Marshal(v2ray.Rules{\n\t// \tType: \"field\",\n\t// \tOutboundTag: \"blocked\",\n\t// \tDomain: []string{\"geosite:category-ads-all\"},\n\t// })\n\tgfwList, _ := json.Marshal(v2ray.Rules{\n\t\tType: \"field\",\n\t\tOutboundTag: \"proxy\",\n\t\tDomain: []string{\"geosite:geolocation-!cn\"},\n\t})\n\tchinaListSite, _ := json.Marshal(v2ray.Rules{\n\t\tType: \"field\",\n\t\tOutboundTag: \"proxy\",\n\t\tDomain: []string{\"geosite:cn\"},\n\t})\n\tchinaListIP, _ := json.Marshal(v2ray.Rules{\n\t\tType: \"field\",\n\t\tOutboundTag: \"proxy\",\n\t\tIP: []string{\"geoip:cn\"},\n\t})\n\tgoogleAPI, _ := json.Marshal(v2ray.Rules{\n\t\tType: \"field\",\n\t\tOutboundTag: \"proxy\",\n\t\tDomain: []string{\"domain:googleapis.cn\"},\n\t})\n\t// all\n\trules := []json.RawMessage{}\n\tif routeMode == 1 {\n\t\trules = []json.RawMessage{\n\t\t\tjson.RawMessage(googleAPI),\n\t\t\tjson.RawMessage(bypassLAN),\n\t\t\tjson.RawMessage(blockDomain),\n\t\t}\n\t}\n\tif routeMode == 2 {\n\t\trules = []json.RawMessage{\n\t\t\tjson.RawMessage(googleAPI),\n\t\t\tjson.RawMessage(blockDomain),\n\t\t\tjson.RawMessage(bypassChinaSite),\n\t\t\tjson.RawMessage(bypassChinaIP),\n\t\t}\n\t}\n\tif routeMode == 3 {\n\t\trules = []json.RawMessage{\n\t\t\tjson.RawMessage(googleAPI),\n\t\t\tjson.RawMessage(blockDomain),\n\t\t\tjson.RawMessage(bypassLAN),\n\t\t\tjson.RawMessage(bypassChinaSite),\n\t\t\tjson.RawMessage(bypassChinaIP),\n\t\t}\n\t}\n\tif routeMode == 4 {\n\t\trules = []json.RawMessage{\n\t\t\tjson.RawMessage(googleAPI),\n\t\t\tjson.RawMessage(blockDomain),\n\t\t\tjson.RawMessage(gfwList),\n\t\t}\n\t}\n\tif routeMode == 5 {\n\t\trules = []json.RawMessage{\n\t\t\tjson.RawMessage(googleAPI),\n\t\t\tjson.RawMessage(blockDomain),\n\t\t\tjson.RawMessage(chinaListSite),\n\t\t\tjson.RawMessage(chinaListIP),\n\t\t}\n\t}\n\tif routeMode >= 5 {\n\t\trules = []json.RawMessage{\n\t\t\tjson.RawMessage(googleAPI),\n\t\t\tjson.RawMessage(bypassLAN),\n\t\t\tjson.RawMessage(bypassChinaSite),\n\t\t\tjson.RawMessage(bypassChinaIP),\n\t\t\tjson.RawMessage(blockDomain),\n\t\t\t// json.RawMessage(blockAd),\n\t\t}\n\t}\n\treturn &conf.RouterConfig{\n\t\tDomainStrategy: &domainStrategy,\n\t\tRuleList: rules,\n\t}\n}", "func initApiConfig() (*model.Bootstrap, bool, error) {\n\tbootstrap := config.Load(configPath)\n\treturn bootstrap, false, nil\n}", "func createAPIs() (apis APIs, err error) {\n\n\tparam := os.Getenv(\"OPSMAN\")\n\n\tvar config helpers.OpsManConfig\n\tif err = json.Unmarshal([]byte(param), &config); err == nil {\n\n\t\t// First find the opsman API\n\t\tvar opsMan *helpers.OpsManAPI\n\t\tif opsMan, err = helpers.NewOpsManAPI(config); err == nil {\n\t\t\tapis = APIs{\n\t\t\t\thelpers.None: nil,\n\t\t\t\thelpers.OpsMan: opsMan,\n\t\t\t}\n\n\t\t\tvar api helpers.API\n\t\t\tif api, err = helpers.NewAppsManAPI(opsMan); err == nil {\n\t\t\t\tapis[helpers.AppsMan] = api\n\t\t\t}\n\t\t}\n\t}\n\n\tif err != nil {\n\t\tlog.Printf(\"API Error: %s\", err)\n\t}\n\n\treturn apis, err\n}", "func NewAPI(svc Service, config *Config) *API {\n\tif !config.Debug {\n\t\tgin.SetMode(gin.ReleaseMode)\n\t}\n\n\tapi := &API{\n\t\tRoutes: configureRoutes(svc, database.DB, !config.AuthDisabled, config.TokenURL),\n\t\tconfig: config,\n\t\t// database: svc.DataSource,\n\t\tTitle: \"DataHive RESTful APIs\",\n\t\tVersion: \"1.0\",\n\t}\n\n\t// enable pprof http endpoints in debug mode\n\tif config.Debug {\n\t\tpprof.Register(api.Routes.Engine, nil)\n\t}\n\n\t// set logrus logger to TextFormatter with no colors\n\tlog.SetFormatter(&log.TextFormatter{DisableColors: true})\n\n\tapi.server = &http.Server{\n\t\tAddr: config.Address,\n\t\tHandler: api.Routes.Engine,\n\t\tReadTimeout: 10 * time.Second,\n\t\tWriteTimeout: 10 * time.Second,\n\t}\n\n\tif !config.WellKnownDisabled {\n\t\tapi.Routes.configureWellKnown(svc.Healthy)\n\t}\n\n\t// configure healthz endpoint\n\tapi.Routes.GET(\"/healthz\", healthHandler(svc.Healthy))\n\n\t// configure info endpoint\n\t// api.Routes.GET(\"/info\", infoHandler(svc.Info))\n\n\treturn api\n}", "func ToAPI(api *Api) *v1.Api {\n\tcrd := &v1.Api{}\n\tcrd.TypeMeta.Kind = \"Api\"\n\tcrd.TypeMeta.APIVersion = v1.GroupVersion.Group + \"/\" + v1.GroupVersion.Version\n\n\tcrd.ObjectMeta.Name = api.ID\n\tcrd.ObjectMeta.Namespace = api.Namespace\n\tcrd.ObjectMeta.Labels = make(map[string]string)\n\tcrd.ObjectMeta.Labels[v1.ServiceunitLabel] = api.Serviceunit.ID\n\tcrd.Spec = v1.ApiSpec{\n\t\tName: api.Name,\n\t\tDescription: api.Description,\n\t\tServiceunit: api.Serviceunit,\n\t\tApplications: api.Applications,\n\t\tFrequency: api.Frequency,\n\t\tApiType: api.ApiType,\n\t\tAuthType: api.AuthType,\n\t\tTags: api.Tags,\n\t\tApiBackendType: api.Serviceunit.Type,\n\t\tMethod: api.Method,\n\t\tProtocol: api.Protocol,\n\t\tReturnType: api.ReturnType,\n\t\tRDBQuery: api.RDBQuery,\n\t\tDataWarehouseQuery: api.DataWarehouseQuery,\n\t\tApiDefineInfo: api.ApiDefineInfo,\n\t\tKongApi: api.KongApi,\n\t\tApiQueryInfo: api.ApiQueryInfo,\n\t\tApiReturnInfo: api.ApiReturnInfo,\n\t\tTraffic: api.Traffic,\n\t\tRestriction: api.Restriction,\n\t\tPublishInfo: api.PublishInfo,\n\t\tResponseTransformer: api.ResTransformer,\n\t}\n\n\tcrd.Status = v1.ApiStatus{\n\t\tStatus: v1.Init,\n\t\tAction: v1.Create,\n\t\t//create api update status to unreleased\n\t\tPublishStatus: v1.UnRelease,\n\t\tAccessLink: api.AccessLink,\n\t\tUpdatedAt: metav1.Now(),\n\t\tReleasedAt: metav1.Now(),\n\t\tApplicationCount: api.ApplicationCount,\n\t\tCalledCount: api.CalledCount,\n\t\tFailedCount: api.FailedCount,\n\t\tLatencyCount: api.LatencyCount,\n\t\tCallFrequency: api.CallFrequency,\n\t}\n\t// add user labels\n\tcrd.ObjectMeta.Labels = user.AddUsersLabels(api.Users, crd.ObjectMeta.Labels)\n\treturn crd\n}", "func createFromResources(dp *v1alpha1.EdgeDataplane, da *v1alpha1.EdgeTraceabilityAgent) (*APIGatewayConfiguration, error) {\n\n\tcfg := &APIGatewayConfiguration{\n\t\tHost: dp.Spec.ApiGatewayManager.Host,\n\t\tPort: int(dp.Spec.ApiGatewayManager.Port),\n\t\tEnableAPICalls: da.Spec.Config.ProcessHeaders,\n\t\tPollInterval: 1 * time.Minute,\n\t}\n\n\tif dp.Spec.ApiGatewayManager.PollInterval != \"\" {\n\t\tresCfgPollInterval, err := time.ParseDuration(dp.Spec.ApiGatewayManager.PollInterval)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tcfg.PollInterval = resCfgPollInterval\n\t}\n\treturn cfg, nil\n}", "func NewAPI(volplugin Volplugin, hostname string, client *config.Client, global **config.Global) *API {\n\treturn &API{\n\t\tVolplugin: volplugin,\n\t\tHostname: hostname,\n\t\tClient: client,\n\t\tGlobal: global,\n\t\tLock: lock.NewDriver(client),\n\t\tMountCollection: mount.NewCollection(),\n\t\tMountCounter: mount.NewCounter(),\n\t\tlockStopChans: map[string]chan struct{}{},\n\t}\n}", "func New(config Config) (*API, error) {\n\tif config.VerboseLogging {\n\t\tlog.SetLevel(log.DebugLevel)\n\t}\n\tlog.Debugf(\"%s API config: %+v\", fn(), config)\n\n\tif config.ConcurrentRequests == 0 {\n\t\tconfig.ConcurrentRequests = 1\n\t}\n\tremote.ConcurrentRequests = config.ConcurrentRequests\n\tremote.TraceRequests = config.TraceRequests\n\tremote.RetryRequests = config.RetryRequests\n\tremote.RetryDelay = config.RetryDelay\n\n\tdockerclient.RetryPulls = config.RetryRequests\n\tdockerclient.RetryDelay = config.RetryDelay\n\n\tif config.InsecureRegistryEx != \"\" {\n\t\trepository.InsecureRegistryEx = config.InsecureRegistryEx\n\t}\n\n\tif config.DockerJSONConfigFile == \"\" {\n\t\tconfig.DockerJSONConfigFile = dockerconfig.DefaultDockerJSON\n\t}\n\tdockerConfig, err := dockerconfig.Load(config.DockerJSONConfigFile)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdockerClient, err := dockerclient.New(dockerConfig)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &API{\n\t\tconfig: config,\n\t\tdockerClient: dockerClient,\n\t}, nil\n}", "func NewApi(c *ApiConfig) (*Api, error) {\n\tctx := &Context{\n\t\tstore: c.Store,\n\t}\n\tapi := newRouter(ctx)\n\tserver := &http.Server{\n\t\tHandler: api,\n\t}\n\n\tprotoAddrParts := strings.SplitN(c.Address, \"://\", 2)\n\tif len(protoAddrParts) == 1 {\n\t\tprotoAddrParts = []string{\"tcp\", protoAddrParts[0]}\n\t}\n\n\tlistener, err := newListener(protoAddrParts[0], protoAddrParts[1], c.TlsConfig)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tserver.Addr = protoAddrParts[1]\n\n\treturn &Api{\n\t\thttpserver: server,\n\t\tlistener: listener,\n\t}, nil\n}", "func CreateOpenAPIIndexConfig() *SpecIndexConfig {\n\tcw, _ := os.Getwd()\n\treturn &SpecIndexConfig{\n\t\tBasePath: cw,\n\t\tAllowRemoteLookup: true,\n\t\tAllowFileLookup: true,\n\t\tseenRemoteSources: &syncmap.Map{},\n\t}\n}", "func createConfig(ctx context.Context, req *events.Request) *config {\n\treturn &config{\n\t\tlog: l.Create(ctx, l.Input{\n\t\t\t\"service\": service,\n\t\t\t\"function\": function,\n\t\t\t\"env\": os.Getenv(\"ENVIRONMENT\"),\n\t\t\t\"stackId\": req.StackID,\n\t\t\t\"requestType\": req.RequestType,\n\t\t\t\"requestId\": req.RequestID,\n\t\t\t\"resourceType\": req.ResourceType,\n\t\t\t\"logicalResourceId\": req.LogicalResourceID,\n\t\t\t\"resourceProperties\": req.ResourceProperties,\n\t\t\t\"oldResourceProperties\": req.OldResourceProperties,\n\t\t}),\n\t\tphysicalID: \"NotAviable\",\n\t\tresourceProperties: &Client{},\n\t\toldResourceProperties: &Client{},\n\t}\n}", "func CreateConfig() *Config {\n return &Config{\n HeaderName: defaultHeader,\n }\n}", "func NewAPI(env *cfenv.App, b Broker, user, pass string, catalog *cf.Catalog) http.Handler {\n\t// create new router\n\tg := gin.Default()\n\n\tauthorised, err := withAuth(g, user, pass)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tapi := API{\n\t\tEnv: env,\n\t\tcatalog: catalog,\n\t\tBroker: &validatingBroker{\n\t\t\tBroker: b,\n\t\t},\n\t}\n\n\tauthorised.GET(\"/v2/catalog\", api.Catalog)\n\tauthorised.PUT(\"/v2/service_instances/:instance_id\", api.Provision)\n\tauthorised.DELETE(\"/v2/service_instances/:instance_id\", api.Deprovision)\n\tauthorised.PUT(\"/v2/service_instances/:instance_id/service_bindings/:binding_id\", api.Bind)\n\tauthorised.DELETE(\"/v2/service_instances/:instance_id/service_bindings/:binding_id\", api.Unbind)\n\n\treturn g\n}", "func New(kubeconfig *rest.Config, opa opa_client.Data, ns types.ResourceType, name string, owner metav1.OwnerReference) *Initializer {\n\tcpy := *kubeconfig\n\tif ns.Group == \"\" {\n\t\tcpy.APIPath = \"/api\"\n\t} else {\n\t\tcpy.APIPath = \"/apis\"\n\t}\n\tcpy.GroupVersion = &schema.GroupVersion{\n\t\tGroup: ns.Group,\n\t\tVersion: ns.Version,\n\t}\n\tcpy.NegotiatedSerializer = dynamic.ContentConfig().NegotiatedSerializer\n\treturn &Initializer{\n\t\tkubeconfig: &cpy,\n\t\tname: name,\n\t\tns: ns,\n\t\topa: opa,\n\t\towner: owner,\n\t}\n}", "func New(config *Config, db mongodb.DB, router *mux.Router) (*API, error) {\n\tapi := &API{\n\t\tconfig: config,\n\t\tdb: db,\n\t\tRouter: router,\n\t}\n\n\t// Endpoint for browser preflight requests\n\tapi.Router.Methods(\"OPTIONS\").HandlerFunc(api.corsMiddleware(api.preflightHandler))\n\n\t// Endpoint for healtcheck\n\tapi.Router.HandleFunc(\"/api/v1/health\", api.corsMiddleware(api.logMiddleware(api.healthHandler))).Methods(\"GET\")\n\n\t// User\n\tapi.Router.HandleFunc(\"/api/v1/auth\", api.corsMiddleware(api.logMiddleware(api.jwtMiddleware(api.UserValidate)))).Methods(\"GET\")\n\tapi.Router.HandleFunc(\"/api/v1/auth/login\", api.corsMiddleware(api.logMiddleware(api.userLoginPostHandler))).Methods(\"POST\")\n\tapi.Router.HandleFunc(\"/api/v1/auth/signup\", api.corsMiddleware(api.logMiddleware(api.userSignupPostHandler))).Methods(\"POST\")\n\tapi.Router.HandleFunc(\"/api/v1/profile\", api.corsMiddleware(api.logMiddleware(api.jwtMiddleware(api.userProfileGetHandler)))).Methods(\"GET\")\n\tapi.Router.HandleFunc(\"/api/v1/profile/{userid}\", api.corsMiddleware(api.logMiddleware(api.jwtMiddleware(api.userProfileUpdateHandler)))).Methods(\"PUT\")\n\n\t// Home\n\tapi.Router.HandleFunc(\"/api/v1/f/{userid}\", api.corsMiddleware(api.logMiddleware(api.jwtMiddleware(api.GetAllFormsHandler)))).Methods(\"GET\")\n\tapi.Router.HandleFunc(\"/api/v1/f/{userid}/less\", api.corsMiddleware(api.logMiddleware(api.jwtMiddleware(api.GetAllFormsWithLessValueHandler)))).Methods(\"GET\")\n\tapi.Router.HandleFunc(\"/api/v1/f/{userid}\", api.corsMiddleware(api.logMiddleware(api.jwtMiddleware(api.CreateFormHandler)))).Methods(\"POST\")\n\n\t// Form\n\tapi.Router.HandleFunc(\"/api/v1/f/{formid}/view\", api.corsMiddleware(api.logMiddleware(api.formViewGetHandler))).Methods(\"GET\")\n\tapi.Router.HandleFunc(\"/api/v1/f/{formid}/view\", api.corsMiddleware(api.logMiddleware(api.formViewPostHandler))).Methods(\"POST\")\n\tapi.Router.HandleFunc(\"/api/v1/f/{userid}/{formid}\", api.corsMiddleware(api.logMiddleware(api.jwtMiddleware(api.formGetHandler)))).Methods(\"GET\")\n\tapi.Router.HandleFunc(\"/api/v1/f/{userid}/{formid}\", api.corsMiddleware(api.logMiddleware(api.jwtMiddleware(api.formPutHandler)))).Methods(\"PUT\")\n\tapi.Router.HandleFunc(\"/api/v1/f/{userid}/{formid}\", api.corsMiddleware(api.logMiddleware(api.jwtMiddleware(api.formDeleteHandler)))).Methods(\"DELETE\")\n\n\tapi.Router.HandleFunc(\"/api/v1/f/{userid}/{formid}/response\", api.corsMiddleware(api.logMiddleware(api.jwtMiddleware(api.formResponseGetHandler)))).Methods(\"GET\")\n\n\tapi.Router.HandleFunc(\"/api/v1/q/{userid}/{formid}\", api.corsMiddleware(api.logMiddleware(api.jwtMiddleware(api.questionPostHandler)))).Methods(\"POST\")\n\tapi.Router.HandleFunc(\"/api/v1/q/{userid}/{formid}/copy\", api.corsMiddleware(api.logMiddleware(api.jwtMiddleware(api.questionPostCopyHandler)))).Methods(\"POST\")\n\tapi.Router.HandleFunc(\"/api/v1/q/{userid}/{formid}/{questionid}\", api.corsMiddleware(api.logMiddleware(api.jwtMiddleware(api.questionPutHandler)))).Methods(\"PUT\")\n\tapi.Router.HandleFunc(\"/api/v1/q/{userid}/{formid}/{questionid}\", api.corsMiddleware(api.logMiddleware(api.jwtMiddleware(api.questionDeleteHandler)))).Methods(\"DELETE\")\n\n\tapi.Router.HandleFunc(\"/api/v1/o/{userid}/{formid}/{questionid}\", api.corsMiddleware(api.logMiddleware(api.jwtMiddleware(api.optionPostHandler)))).Methods(\"POST\")\n\tapi.Router.HandleFunc(\"/api/v1/o/{userid}/{formid}/{questionid}/{optionid}\", api.corsMiddleware(api.logMiddleware(api.jwtMiddleware(api.optionPutHandler)))).Methods(\"PUT\")\n\tapi.Router.HandleFunc(\"/api/v1/o/{userid}/{formid}/{questionid}/{optionid}\", api.corsMiddleware(api.logMiddleware(api.jwtMiddleware(api.optionDeleteHandler)))).Methods(\"DELETE\")\n\n\treturn api, nil\n}", "func InitApi(\n\tcfg Config,\n\ttemplates models.Templates,\n\tlocaleManager LocaleManager,\n) *Api {\n\treturn &Api{\n\t\tConfig: cfg,\n\t\ttemplates: templates,\n\t\tlocaleManager: localeManager,\n\t}\n}", "func NewAPI(root string, logger logger.Logger) *API {\n\t// init\n\tapi := &API{\n\t\troot: root,\n\t\tcorsHosts: \"*\",\n\t\tcorsHeaders: \"*\",\n\t\tendpoints: make([]APIEndpoint, 0), // explicitly define an empty array\n\t\turlBuilder: URLDefaultBuilder,\n\t\tlogger: logger,\n\t}\n\n\treturn api\n}", "func (r *K8sRESTConfigFactory) Create(token string) (*rest.Config, error) {\n\tshallowCopy := *r.cfg\n\tshallowCopy.BearerToken = token\n\tif r.insecure {\n\t\tshallowCopy.TLSClientConfig = rest.TLSClientConfig{\n\t\t\tInsecure: r.insecure,\n\t\t}\n\t}\n\treturn &shallowCopy, nil\n}", "func NewAPI(log *logrus.Entry, config *conf.Config, db IApiDb, fs IApiFileSystem) *API {\n\techoServer := EchoServer{e: echo.New()}\n\n\t// create the api\n\tapi := &API{\n\t\tconfig: config,\n\t\tlog: log.WithField(\"component\", \"api\"),\n\t\tServer: echoServer,\n\t\tDb: db,\n\t\tFs: fs,\n\t}\n\n\treturn api\n}", "func SetupAPI(db *sql.DB, APIConfig *config.APIConfig) *API {\n\t// Setup Websocket hub\n\thub := ws.Hub{\n\t\tBroadcast: make(chan ws.Message),\n\t\tRegister: make(chan ws.Subscription),\n\t\tUnregister: make(chan ws.Subscription),\n\t\tRooms: make(map[string]map[*ws.Connection]bool),\n\t}\n\tgo hub.Run()\n\n\treturn &API{\n\t\tAPIConfig: APIConfig,\n\t\tDB: db,\n\t\tHub: &hub,\n\t}\n}", "func NewServiceConfigByAPI(opts ...ServiceConfigOpt) *ServiceConfig {\n\tdefaultServiceConfig := NewDefaultServiceConfig()\n\tfor _, v := range opts {\n\t\tv(defaultServiceConfig)\n\t}\n\treturn defaultServiceConfig\n}", "func Init(appName string, router *echo.Group) (*API, error) {\n\tvar err error\n\tvar c Config\n\tvar a API\n\n\tif err = envconfig.Process(appName, &c); err != nil {\n\t\treturn nil, err\n\t}\n\tconnSTR := c.connSTR()\n\tif a.driver, err = sql.Open(\"postgres\", connSTR); err != nil {\n\t\treturn nil, err\n\t}\n\n\trouter.POST(routeFizzBuzz, a.FizzBuzz)\n\trouter.GET(routeStats, a.Stats)\n\treturn &a, nil\n}", "func newPostBuildOpenAPIObjectFunc(config ServerConfig, container *restful.Container) restfulSpec.PostBuildSwaggerObjectFunc {\n\treturn func(swo *spec.Swagger) {\n\t\tswo.Host = config.OpenAPI.Host\n\t\tswo.BasePath = config.OpenAPI.BasePath\n\t\tswo.Schemes = config.OpenAPI.Schemas\n\n\t\tvar title, description string\n\t\tif config.Name != \"\" {\n\t\t\ttitle = config.Name\n\t\t} else {\n\t\t\ttitle = config.OpenAPI.Spec.Title\n\t\t}\n\t\tif config.Description != \"\" {\n\t\t\tdescription = config.Description\n\t\t} else {\n\t\t\tdescription = config.OpenAPI.Spec.Description\n\t\t}\n\t\tswo.Info = &spec.Info{\n\t\t\tInfoProps: spec.InfoProps{\n\t\t\t\tTitle: title,\n\t\t\t\tDescription: description,\n\t\t\t\tContact: &spec.ContactInfo{\n\t\t\t\t\tContactInfoProps: spec.ContactInfoProps{\n\t\t\t\t\t\tName: config.OpenAPI.Spec.Contact.Name,\n\t\t\t\t\t\tEmail: config.OpenAPI.Spec.Contact.Email,\n\t\t\t\t\t\tURL: config.OpenAPI.Spec.Contact.URL,\n\t\t\t\t\t},\n\t\t\t\t},\n\n\t\t\t\tLicense: &spec.License{\n\t\t\t\t\tLicenseProps: spec.LicenseProps{\n\t\t\t\t\t\tName: config.OpenAPI.Spec.License.Name,\n\t\t\t\t\t\tURL: config.OpenAPI.Spec.License.URL,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tVersion: config.OpenAPI.Spec.Version,\n\t\t\t},\n\t\t}\n\n\t\tvar nTags []spec.Tag\n\t\tvar tags []OpenapiTagConfig\n\t\tif len(config.OpenAPI.Tags) > 0 {\n\t\t\ttags = config.OpenAPI.Tags\n\t\t} else {\n\t\t\ttags = config.OpenAPI.Tags\n\t\t}\n\t\tfor _, tag := range tags {\n\t\t\tnTag := spec.Tag{TagProps: spec.TagProps{Name: tag.Name, Description: tag.Description}}\n\n\t\t\tnTags = append(nTags, nTag)\n\t\t}\n\t\tswo.Tags = nTags\n\t\t// setup security definitions\n\t\tif config.OpenAPI.Auth == \"basic\" {\n\t\t\tswo.SecurityDefinitions = map[string]*spec.SecurityScheme{\n\t\t\t\t\"basicAuth\": spec.BasicAuth(),\n\t\t\t}\n\t\t\tauth := make(map[string][]string)\n\t\t\tauth[\"basicAuth\"] = []string{}\n\t\t\tswo.Security = append(swo.Security, auth)\n\t\t} else if config.OpenAPI.Auth == \"jwt\" {\n\t\t\tswo.SecurityDefinitions = map[string]*spec.SecurityScheme{\n\t\t\t\t\"jwt\": spec.APIKeyAuth(\"Authorization\", \"header\"),\n\t\t\t}\n\t\t\tenrichSwaggerObjectSecurity(swo, container)\n\t\t}\n\n\t}\n}", "func NewAPI(handler APIHandler, address string, port uint) (*API, error) {\n\tif handler == nil {\n\t\treturn nil, errors.New(\"An API handler is required\")\n\t}\n\tapi := new(API)\n\tapi.address = address\n\tapi.port = port\n\tapi.router = mux.NewRouter().StrictSlash(true)\n\tapi.router.HandleFunc(\"/\", handler.SlashRoot)\n\tapi.router.HandleFunc(\"/clients\", handler.ClientsHandler)\n\tapi.router.HandleFunc(\"/clients/{id}\", handler.ClientHandler)\n\tapi.router.HandleFunc(\"/clients/{id}/history\", handler.ClientHistoryHandler)\n\tapi.router.HandleFunc(\"/jobs\", handler.PendingJobsHandler)\n\tapi.router.HandleFunc(\"/jobs/{id}\", handler.JobHandler)\n\tapi.router.HandleFunc(\"/clients/kick/{id}\", handler.KickClientHandler).Methods(\"POST\")\n\tapi.router.HandleFunc(\"/terminate\", handler.TerminateHandler).Methods(\"POST\")\n\tapi.router.HandleFunc(\"/report\", handler.ReportHandler)\n\tapi.router.HandleFunc(\"/history\", handler.HistoryHandler)\n\tapi.router.HandleFunc(\"/configure\", handler.ConfigureHandler).Methods(\"POST\")\n\tapi.router.HandleFunc(\"/config\", handler.ConfigHandler)\n\tapi.router.HandleFunc(\"/target\", handler.TargetHandler)\n\n\treturn api, nil\n}", "func New(config Config) API {\n\treturn API{\n\t\tClient: &http.Client{},\n\t\tConfig: config,\n\t}\n}", "func buildApiplex(config ApiplexConfig) (*apiplex, error) {\n\tif config.Serve.API == \"\" {\n\t\tconfig.Serve.API = \"/\"\n\t}\n\n\tif config.Serve.SigningKey == \"\" {\n\t\tconfig.Serve.SigningKey = uniuri.NewLen(64)\n\t}\n\n\t// TODO make everything configurable\n\tap := apiplex{\n\t\tapipath: ensureFinalSlash(config.Serve.API),\n\t\tauthCacheMins: 10,\n\t\tsigningKey: config.Serve.SigningKey,\n\t}\n\n\tif _, ok := config.Quotas[\"default\"]; !ok {\n\t\treturn nil, fmt.Errorf(\"Your configuration must specify at least a 'default' quota.\")\n\t}\n\tif kl, ok := config.Quotas[\"keyless\"]; ok {\n\t\tif kl.MaxKey != 0 {\n\t\t\treturn nil, fmt.Errorf(\"You cannot set a per-key maximum for the 'keyless' quota.\")\n\t\t}\n\t\tap.allowKeyless = true\n\t} else {\n\t\tap.allowKeyless = false\n\t}\n\tap.quotas = config.Quotas\n\n\t// auth plugins\n\tauth, err := buildPlugins(config.Plugins.Auth, reflect.TypeOf((*AuthPlugin)(nil)).Elem())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tap.auth = make([]AuthPlugin, len(auth))\n\tfor i, p := range auth {\n\t\tcp := p.(AuthPlugin)\n\t\tap.auth[i] = cp\n\t}\n\n\t// backend plugins\n\tbackend, err := buildPlugins(config.Plugins.Backend, reflect.TypeOf((*BackendPlugin)(nil)).Elem())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tap.backends = make([]BackendPlugin, len(backend))\n\tfor i, p := range backend {\n\t\tcp := p.(BackendPlugin)\n\t\tap.backends[i] = cp\n\t}\n\n\t// The first ManagementBackendPlugin (i.e. one with additional user/key management) gets special\n\t// treatment: if the portal API is enabled, it will connect directly to this plugin and use that\n\t// to perform portal actions.\n\tfor _, plugin := range ap.backends {\n\t\t// must use reflection here since type switch will see ap.backends as implementing\n\t\t// BackendPlugin only\n\t\tif reflect.TypeOf(plugin).Implements(reflect.TypeOf((*ManagementBackendPlugin)(nil)).Elem()) {\n\t\t\tmgmt := plugin.(ManagementBackendPlugin)\n\t\t\tap.usermgmt = mgmt\n\t\t\tbreak\n\t\t}\n\t}\n\n\t// postauth plugins\n\tpostauth, err := buildPlugins(config.Plugins.PostAuth, reflect.TypeOf((*PostAuthPlugin)(nil)).Elem())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tap.postauth = make([]PostAuthPlugin, len(postauth))\n\tfor i, p := range postauth {\n\t\tcp := p.(PostAuthPlugin)\n\t\tap.postauth[i] = cp\n\t}\n\n\t// preupstream plugins\n\tpreupstream, err := buildPlugins(config.Plugins.PreUpstream, reflect.TypeOf((*PreUpstreamPlugin)(nil)).Elem())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tap.preupstream = make([]PreUpstreamPlugin, len(preupstream))\n\tfor i, p := range preupstream {\n\t\tcp := p.(PreUpstreamPlugin)\n\t\tap.preupstream[i] = cp\n\t}\n\n\t// postupstream plugins\n\tpostupstream, err := buildPlugins(config.Plugins.PostUpstream, reflect.TypeOf((*PostUpstreamPlugin)(nil)).Elem())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tap.postupstream = make([]PostUpstreamPlugin, len(postupstream))\n\tfor i, p := range postupstream {\n\t\tcp := p.(PostUpstreamPlugin)\n\t\tap.postupstream[i] = cp\n\t}\n\n\t// logging plugins\n\tlogging, err := buildPlugins(config.Plugins.Logging, reflect.TypeOf((*LoggingPlugin)(nil)).Elem())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tap.logging = make([]LoggingPlugin, len(logging))\n\tfor i, p := range logging {\n\t\tcp := p.(LoggingPlugin)\n\t\tap.logging[i] = cp\n\t}\n\n\t// upstreams\n\tap.upstreams = make([]APIUpstream, len(config.Serve.Upstreams))\n\tfor i, us := range config.Serve.Upstreams {\n\t\tu, err := url.Parse(us)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"Invalid upstream address: %s\", us)\n\t\t}\n\t\tap.upstreams[i] = APIUpstream{\n\t\t\tClient: &http.Client{},\n\t\t\tAddress: u,\n\t\t}\n\t}\n\n\tap.redis = &redis.Pool{\n\t\tMaxIdle: 3,\n\t\tIdleTimeout: 240 * time.Second,\n\t\tDial: func() (redis.Conn, error) {\n\t\t\tc, err := redis.Dial(\"tcp\", config.Redis.Host+\":\"+strconv.Itoa(config.Redis.Port))\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tc.Do(\"SELECT\", config.Redis.DB)\n\t\t\treturn c, err\n\t\t},\n\t\tTestOnBorrow: func(c redis.Conn, t time.Time) error {\n\t\t\t_, err := c.Do(\"PING\")\n\t\t\treturn err\n\t\t},\n\t}\n\t// test connection\n\trd := ap.redis.Get()\n\t_, err = rd.Do(\"PING\")\n\tif err != nil {\n\t\tlog.Fatalf(\"Couldn't connect to Redis. %s\", err.Error())\n\t}\n\n\treturn &ap, nil\n}", "func New(log *util.Logger, user, password string, cache time.Duration, config Config) (*API, error) {\n\tif err := mergo.Merge(&config, defaults); err != nil {\n\t\treturn nil, err\n\t}\n\n\tv := &API{\n\t\tlog: log,\n\t\tHelper: request.NewHelper(log),\n\t\tconfig: config,\n\t\tuser: user,\n\t\tpassword: password,\n\t}\n\n\t// api is unbelievably slow when retrieving status\n\tv.Helper.Client.Timeout = 120 * time.Second\n\n\tv.apiG = provider.NewCached(v.statusAPI, cache).InterfaceGetter()\n\n\treturn v, nil\n}", "func NewAPI(t authorization.TokenGetter, client *http.Client, clientID string) (*API, error) {\n\tapi, err := i.NewAPI(t, client, clientID, i.DefaultHTTPRequestGenerator(\"configuration\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &API{api}, nil\n}", "func CreateConfig(host string, verifyTLS bool, apiKey string, project string, name string, environment string) (models.ConfigInfo, Error) {\n\tpostBody := map[string]interface{}{\"name\": name, \"environment\": environment}\n\tbody, err := json.Marshal(postBody)\n\tif err != nil {\n\t\treturn models.ConfigInfo{}, Error{Err: err, Message: \"Invalid config info\"}\n\t}\n\n\tvar params []queryParam\n\tparams = append(params, queryParam{Key: \"project\", Value: project})\n\n\turl, err := generateURL(host, \"/v3/configs\", params)\n\tif err != nil {\n\t\treturn models.ConfigInfo{}, Error{Err: err, Message: \"Unable to generate url\"}\n\t}\n\n\tstatusCode, _, response, err := PostRequest(url, verifyTLS, apiKeyHeader(apiKey), body)\n\tif err != nil {\n\t\treturn models.ConfigInfo{}, Error{Err: err, Message: \"Unable to create config\", Code: statusCode}\n\t}\n\n\tvar result map[string]interface{}\n\terr = json.Unmarshal(response, &result)\n\tif err != nil {\n\t\treturn models.ConfigInfo{}, Error{Err: err, Message: \"Unable to parse API response\", Code: statusCode}\n\t}\n\n\tconfig, ok := result[\"config\"].(map[string]interface{})\n\tif !ok {\n\t\treturn models.ConfigInfo{}, Error{Err: fmt.Errorf(\"Unexpected type parsing config, expected map[string]interface{}, got %T\", result[\"config\"]), Message: \"Unable to parse API response\", Code: statusCode}\n\t}\n\tinfo := models.ParseConfigInfo(config)\n\treturn info, Error{}\n}", "func (c *Config) generateConfigForHTTPAPI() (string, error) {\n\tvar err error\n\tcfg := configHttpAPI{}\n\tcfg.Databases = map[string]*ConfigDatabase{}\n\n\tfor dbName, db := range c.Databases {\n\t\tif dbName == indexDatabaseName || dbName == c.getIndexDbAlias() {\n\t\t\tcontinue\n\t\t}\n\t\tdbCopy := &ConfigDatabase{}\n\t\tdbCopy.Tables = db.Tables\n\t\tcfg.Databases[dbName] = dbCopy\n\t}\n\n\tjson, err := json.MarshalIndent(&cfg, \"\", \"\\t\")\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn string(json), nil\n}", "func New(token, api string) (*API, error) {\n\tu, err := url.Parse(api)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tswitch u.Scheme {\n\tcase \"http\":\n\t\tfallthrough\n\tcase \"https\":\n\t\treturn NewHTTP(token, api)\n\tcase \"ws\":\n\t\tfallthrough\n\tcase \"wss\":\n\t\treturn NewWebsocket(token, api)\n\tdefault:\n\t\treturn nil, errors.New(\"invalid api's scheme\")\n\t}\n}", "func NewConfig(conf dynamic.Configuration) *Configuration {\n\tif conf.HTTP == nil && conf.TCP == nil && conf.UDP == nil {\n\t\treturn &Configuration{}\n\t}\n\n\truntimeConfig := &Configuration{}\n\n\tif conf.HTTP != nil {\n\t\trouters := conf.HTTP.Routers\n\t\tif len(routers) > 0 {\n\t\t\truntimeConfig.Routers = make(map[string]*RouterInfo, len(routers))\n\t\t\tfor k, v := range routers {\n\t\t\t\truntimeConfig.Routers[k] = &RouterInfo{Router: v, Status: StatusEnabled}\n\t\t\t}\n\t\t}\n\n\t\tservices := conf.HTTP.Services\n\t\tif len(services) > 0 {\n\t\t\truntimeConfig.Services = make(map[string]*ServiceInfo, len(services))\n\t\t\tfor k, v := range services {\n\t\t\t\truntimeConfig.Services[k] = &ServiceInfo{Service: v, Status: StatusEnabled}\n\t\t\t}\n\t\t}\n\n\t\tmiddlewares := conf.HTTP.Middlewares\n\t\tif len(middlewares) > 0 {\n\t\t\truntimeConfig.Middlewares = make(map[string]*MiddlewareInfo, len(middlewares))\n\t\t\tfor k, v := range middlewares {\n\t\t\t\truntimeConfig.Middlewares[k] = &MiddlewareInfo{Middleware: v, Status: StatusEnabled}\n\t\t\t}\n\t\t}\n\t}\n\n\tif conf.TCP != nil {\n\t\tif len(conf.TCP.Routers) > 0 {\n\t\t\truntimeConfig.TCPRouters = make(map[string]*TCPRouterInfo, len(conf.TCP.Routers))\n\t\t\tfor k, v := range conf.TCP.Routers {\n\t\t\t\truntimeConfig.TCPRouters[k] = &TCPRouterInfo{TCPRouter: v, Status: StatusEnabled}\n\t\t\t}\n\t\t}\n\n\t\tif len(conf.TCP.Services) > 0 {\n\t\t\truntimeConfig.TCPServices = make(map[string]*TCPServiceInfo, len(conf.TCP.Services))\n\t\t\tfor k, v := range conf.TCP.Services {\n\t\t\t\truntimeConfig.TCPServices[k] = &TCPServiceInfo{TCPService: v, Status: StatusEnabled}\n\t\t\t}\n\t\t}\n\n\t\tif len(conf.TCP.Middlewares) > 0 {\n\t\t\truntimeConfig.TCPMiddlewares = make(map[string]*TCPMiddlewareInfo, len(conf.TCP.Middlewares))\n\t\t\tfor k, v := range conf.TCP.Middlewares {\n\t\t\t\truntimeConfig.TCPMiddlewares[k] = &TCPMiddlewareInfo{TCPMiddleware: v, Status: StatusEnabled}\n\t\t\t}\n\t\t}\n\t}\n\n\tif conf.UDP != nil {\n\t\tif len(conf.UDP.Routers) > 0 {\n\t\t\truntimeConfig.UDPRouters = make(map[string]*UDPRouterInfo, len(conf.UDP.Routers))\n\t\t\tfor k, v := range conf.UDP.Routers {\n\t\t\t\truntimeConfig.UDPRouters[k] = &UDPRouterInfo{UDPRouter: v, Status: StatusEnabled}\n\t\t\t}\n\t\t}\n\n\t\tif len(conf.UDP.Services) > 0 {\n\t\t\truntimeConfig.UDPServices = make(map[string]*UDPServiceInfo, len(conf.UDP.Services))\n\t\t\tfor k, v := range conf.UDP.Services {\n\t\t\t\truntimeConfig.UDPServices[k] = &UDPServiceInfo{UDPService: v, Status: StatusEnabled}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn runtimeConfig\n}", "func initConfig(centralConfig corecfg.CentralConfig) (interface{}, error) {\n\trootProps := RootCmd.GetProperties()\n\t// Parse the config from bound properties and setup gateway config\n\tgatewayConfig = &config.GatewayConfig{\n\t\tSpecPath: rootProps.StringPropertyValue(\"3scale-api-gateway.specPath\"),\n\t\tConfigKey1: rootProps.StringPropertyValue(\"3scale-api-gateway.config_key_1\"),\n\t\tConfigKey2: rootProps.StringPropertyValue(\"3scale-api-gateway.config_key_2\"),\n\t\tConfigKey3: rootProps.StringPropertyValue(\"3scale-api-gateway.config_key_3\"),\n\t}\n\n\tagentConfig := config.AgentConfig{\n\t\tCentralCfg: centralConfig,\n\t\tGatewayCfg: gatewayConfig,\n\t}\n\treturn agentConfig, nil\n}", "func NewRouting(cfg *config.APICfg, brk brokers.Broker, str stores.Store, mgr *oldPush.Manager, c push.Client, routes []APIRoute) *API {\n\t// Create the api Object\n\tar := API{}\n\t// Create a new router and reference him in API object\n\tar.Router = mux.NewRouter().StrictSlash(false)\n\t// reference routes input in API object too keep info centralized\n\tar.Routes = routes\n\n\t// For each route\n\tfor _, route := range ar.Routes {\n\n\t\t// prepare handle wrappers\n\t\tvar handler http.HandlerFunc\n\t\thandler = route.Handler\n\n\t\thandler = WrapLog(handler, route.Name)\n\n\t\t// skip authentication/authorization for the health status and profile api calls\n\t\tif route.Name != \"ams:healthStatus\" && \"users:profile\" != route.Name && route.Name != \"version:list\" {\n\t\t\thandler = WrapAuthorize(handler, route.Name)\n\t\t\thandler = WrapAuthenticate(handler)\n\t\t}\n\n\t\thandler = WrapValidate(handler)\n\t\thandler = WrapConfig(handler, cfg, brk, str, mgr, c)\n\n\t\tar.Router.\n\t\t\tPathPrefix(\"/v1\").\n\t\t\tMethods(route.Method).\n\t\t\tPath(route.Path).\n\t\t\tName(route.Name).\n\t\t\tHandler(gorillaContext.ClearHandler(handler))\n\t}\n\n\tlog.Info(\"API\", \"\\t\", \"API Router initialized! Ready to start listening...\")\n\t// Return reference to API object\n\treturn &ar\n}", "func LoadAPIConfig() error {\n\treturn loadConfig([]func() error{\n\t\tloadAppConfig,\n\t\tloadDBConfig,\n\t})\n}", "func NewAPIRouter(db *database.Service, game *game.Service, config *config.Config) http.Handler {\n\tapi := &APIHandler{\n\t\tDB: db,\n\t\tGame: game,\n\t\tConfig: config,\n\t}\n\n\tr := chi.NewRouter()\n\n\t// CORS Header\n\tcorsConfig := cors.New(cors.Options{\n\t\tAllowedOrigins: []string{\"https://fantasymarket.netlify.app\", \"https://develop--fantasymarket.netlify.app\", \"http://localhost:3000\"},\n\t\tAllowedMethods: []string{\"GET\", \"POST\", \"PUT\", \"DELETE\", \"HEAD\", \"PATCH\"},\n\t\tAllowCredentials: true,\n\t\tDebug: true,\n\t})\n\n\t// Middleware\n\tr.Use(corsConfig.Handler)\n\tr.Use(middleware.Logger)\n\n\tr.Get(\"/events\", api.getEvents)\n\n\tr.Get(\"/time\", api.getTime)\n\n\tr.Route(\"/stocks\", func(r chi.Router) {\n\n\t\tr.Get(\"/\", api.getAllStocks)\n\n\t\tr.Get(\"/{symbol}\", api.getStockDetails)\n\n\t})\n\n\tr.Route(\"/orders\", func(r chi.Router) {\n\n\t\tr.Get(\"/\", api.ordersForUser)\n\n\t\tr.Post(\"/\", api.addOrder)\n\n\t\tr.Get(\"/{orderID}\", api.getOrdersID)\n\n\t\tr.Delete(\"/{orderID}\", api.deleteOrder)\n\t})\n\n\tr.Route(\"/user\", func(r chi.Router) {\n\t\tr.Group(func(r chi.Router) {\n\t\t\tr.Use(jwt.Middleware(api.Config.TokenSecret, true))\n\t\t\tr.Get(\"/{username}\", api.getUser)\n\t\t\tr.Put(\"/\", api.createUser)\n\t\t\tr.Post(\"/login\", api.loginUser)\n\t\t})\n\n\t\tr.Group(func(r chi.Router) {\n\t\t\tr.Use(jwt.Middleware(api.Config.TokenSecret, false))\n\t\t\tr.Get(\"/\", api.getSelf)\n\t\t\tr.Post(\"/\", api.updateSelf)\n\t\t})\n\t})\n\n\treturn r\n}", "func (config *Config) Api() (*fc.Api, error) {\n\tapi := fc.NewApi(nil)\n\tapi.SetAccessCredentials(config.AccessKey, config.SecretKey)\n\tif config.FindGatewayTimeout != \"\" {\n\t\tapi.SetFindGatewayTimeout(config.FindGatewayTimeout)\n\t}\n\tlog.Printf(\"[INFO] Forty Cloud Client configured.\")\n\treturn api, nil\n}", "func createAPIEndpointService(namespace, serviceName string, api *API) error {\n\n\tclientset, err := getClientSet()\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar stype v1.ServiceType\n\n\tswitch api.ServiceType {\n\n\tcase \"clusterIP\":\n\t\tstype = v1.ServiceTypeClusterIP\n\n\tcase \"nodePort\":\n\t\tstype = v1.ServiceTypeNodePort\n\n\tcase \"loadBalancer\":\n\t\tstype = v1.ServiceTypeLoadBalancer\n\n\tdefault:\n\t\tstype = v1.ServiceTypeClusterIP\n\n\t}\n\n\t_, err = clientset.CoreV1().Services(namespace).Create(&v1.Service{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tName: serviceName,\n\t\t\tLabels: map[string]string{\n\t\t\t\t\"kubez-type\": \"api-endpoint\",\n\t\t\t\t\"kubez-name\": api.Name,\n\t\t\t},\n\t\t},\n\t\tSpec: v1.ServiceSpec{\n\t\t\tSelector: map[string]string{\n\t\t\t\t\"kubez-type\": \"api-endpoint\",\n\t\t\t\t\"kubez-name\": api.Name,\n\t\t\t},\n\t\t\tType: stype,\n\t\t\tPorts: []v1.ServicePort{\n\t\t\t\t{\n\t\t\t\t\tName: \"kubez-api\",\n\t\t\t\t\tPort: api.Port,\n\t\t\t\t\tTargetPort: intstr.IntOrString{IntVal: api.Port},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func GenerateCreateRestApiInput(cr *svcapitypes.RestAPI) *svcsdk.CreateRestApiInput {\n\tres := &svcsdk.CreateRestApiInput{}\n\n\tif cr.Spec.ForProvider.APIKeySource != nil {\n\t\tres.SetApiKeySource(*cr.Spec.ForProvider.APIKeySource)\n\t}\n\tif cr.Spec.ForProvider.BinaryMediaTypes != nil {\n\t\tf1 := []*string{}\n\t\tfor _, f1iter := range cr.Spec.ForProvider.BinaryMediaTypes {\n\t\t\tvar f1elem string\n\t\t\tf1elem = *f1iter\n\t\t\tf1 = append(f1, &f1elem)\n\t\t}\n\t\tres.SetBinaryMediaTypes(f1)\n\t}\n\tif cr.Spec.ForProvider.CloneFrom != nil {\n\t\tres.SetCloneFrom(*cr.Spec.ForProvider.CloneFrom)\n\t}\n\tif cr.Spec.ForProvider.Description != nil {\n\t\tres.SetDescription(*cr.Spec.ForProvider.Description)\n\t}\n\tif cr.Spec.ForProvider.DisableExecuteAPIEndpoint != nil {\n\t\tres.SetDisableExecuteApiEndpoint(*cr.Spec.ForProvider.DisableExecuteAPIEndpoint)\n\t}\n\tif cr.Spec.ForProvider.EndpointConfiguration != nil {\n\t\tf5 := &svcsdk.EndpointConfiguration{}\n\t\tif cr.Spec.ForProvider.EndpointConfiguration.Types != nil {\n\t\t\tf5f0 := []*string{}\n\t\t\tfor _, f5f0iter := range cr.Spec.ForProvider.EndpointConfiguration.Types {\n\t\t\t\tvar f5f0elem string\n\t\t\t\tf5f0elem = *f5f0iter\n\t\t\t\tf5f0 = append(f5f0, &f5f0elem)\n\t\t\t}\n\t\t\tf5.SetTypes(f5f0)\n\t\t}\n\t\tif cr.Spec.ForProvider.EndpointConfiguration.VPCEndpointIDs != nil {\n\t\t\tf5f1 := []*string{}\n\t\t\tfor _, f5f1iter := range cr.Spec.ForProvider.EndpointConfiguration.VPCEndpointIDs {\n\t\t\t\tvar f5f1elem string\n\t\t\t\tf5f1elem = *f5f1iter\n\t\t\t\tf5f1 = append(f5f1, &f5f1elem)\n\t\t\t}\n\t\t\tf5.SetVpcEndpointIds(f5f1)\n\t\t}\n\t\tres.SetEndpointConfiguration(f5)\n\t}\n\tif cr.Spec.ForProvider.MinimumCompressionSize != nil {\n\t\tres.SetMinimumCompressionSize(*cr.Spec.ForProvider.MinimumCompressionSize)\n\t}\n\tif cr.Spec.ForProvider.Name != nil {\n\t\tres.SetName(*cr.Spec.ForProvider.Name)\n\t}\n\tif cr.Spec.ForProvider.Policy != nil {\n\t\tres.SetPolicy(*cr.Spec.ForProvider.Policy)\n\t}\n\tif cr.Spec.ForProvider.Tags != nil {\n\t\tf9 := map[string]*string{}\n\t\tfor f9key, f9valiter := range cr.Spec.ForProvider.Tags {\n\t\t\tvar f9val string\n\t\t\tf9val = *f9valiter\n\t\t\tf9[f9key] = &f9val\n\t\t}\n\t\tres.SetTags(f9)\n\t}\n\tif cr.Spec.ForProvider.Version != nil {\n\t\tres.SetVersion(*cr.Spec.ForProvider.Version)\n\t}\n\n\treturn res\n}", "func New() *API {\n\tapi := &API{\n\t\trouter: router.New(),\n\t\tchain: []alice.Constructor{},\n\t\tEncoders: encoding.New(),\n\t}\n\tapi.RouteGroup = &RouteGroup{[]HandleFunc{}, \"/\", nil, api}\n\treturn api\n}", "func initAPI() {\n\t// 1. init router\n\tr := chi.NewRouter()\n\tr.Post(\"/sub\", subscribe)\n\tr.Post(\"/unsub\", unsubscribe)\n\tr.Post(\"/pub\", publish)\n\tr.Post(\"/unpub\", unpublish)\n\n\t_ = dubbologger.InitLog(path.Join(logPath, \"dubbo.log\"))\n\n\t// FIXME make port configurable\n\tutils.GoWithRecover(func() {\n\t\tif err := http.ListenAndServe(\":\"+apiPort, r); err != nil {\n\t\t\tlog.DefaultLogger.Infof(\"auto write config when updated\")\n\t\t}\n\t}, nil)\n\n\t// 2. init dubbo router\n\tinitRouterManager()\n}", "func (cfg *Config) ToUAPI(w io.Writer, prev *Config) error {\n\tvar stickyErr error\n\tset := func(key, value string) {\n\t\tif stickyErr != nil {\n\t\t\treturn\n\t\t}\n\t\t_, err := fmt.Fprintf(w, \"%s=%s\\n\", key, value)\n\t\tif err != nil {\n\t\t\tstickyErr = err\n\t\t}\n\t}\n\tsetUint16 := func(key string, value uint16) {\n\t\tset(key, strconv.FormatUint(uint64(value), 10))\n\t}\n\tsetPeer := func(peer Peer) {\n\t\tset(\"public_key\", peer.PublicKey.UntypedHexString())\n\t}\n\n\t// Device config.\n\tif !prev.PrivateKey.Equal(cfg.PrivateKey) {\n\t\tset(\"private_key\", cfg.PrivateKey.UntypedHexString())\n\t}\n\n\told := make(map[key.NodePublic]Peer)\n\tfor _, p := range prev.Peers {\n\t\told[p.PublicKey] = p\n\t}\n\n\t// Add/configure all new peers.\n\tfor _, p := range cfg.Peers {\n\t\toldPeer, wasPresent := old[p.PublicKey]\n\t\tsetPeer(p)\n\t\tset(\"protocol_version\", \"1\")\n\n\t\t// Avoid setting endpoints if the correct one is already known\n\t\t// to WireGuard, because doing so generates a bit more work in\n\t\t// calling magicsock's ParseEndpoint for effectively a no-op.\n\t\tif !wasPresent {\n\t\t\tset(\"endpoint\", p.PublicKey.UntypedHexString())\n\t\t}\n\n\t\t// TODO: replace_allowed_ips is expensive.\n\t\t// If p.AllowedIPs is a strict superset of oldPeer.AllowedIPs,\n\t\t// then skip replace_allowed_ips and instead add only\n\t\t// the new ipps with allowed_ip.\n\t\tif !cidrsEqual(oldPeer.AllowedIPs, p.AllowedIPs) {\n\t\t\tset(\"replace_allowed_ips\", \"true\")\n\t\t\tfor _, ipp := range p.AllowedIPs {\n\t\t\t\tset(\"allowed_ip\", ipp.String())\n\t\t\t}\n\t\t}\n\n\t\t// Set PersistentKeepalive after the peer is otherwise configured,\n\t\t// because it can trigger handshake packets.\n\t\tif oldPeer.PersistentKeepalive != p.PersistentKeepalive {\n\t\t\tsetUint16(\"persistent_keepalive_interval\", p.PersistentKeepalive)\n\t\t}\n\t}\n\n\t// Remove peers that were present but should no longer be.\n\tfor _, p := range cfg.Peers {\n\t\tdelete(old, p.PublicKey)\n\t}\n\tfor _, p := range old {\n\t\tsetPeer(p)\n\t\tset(\"remove\", \"true\")\n\t}\n\n\tif stickyErr != nil {\n\t\tstickyErr = fmt.Errorf(\"ToUAPI: %w\", stickyErr)\n\t}\n\treturn stickyErr\n}", "func (builder *RoomBuilder) BuildAPI(a *RoomRequestsI) {\n\t*a = builder.api\n}", "func (work *Worker) CreateAPIServer(port int) {\n\trouter := mux.NewRouter()\n\n\t// Setup all of the endpoints\n\trouter.HandleFunc(\"/startPolling\", work.PollerExample)\n\trouter.HandleFunc(\"/startStream\", work.StartStreamRequest)\n\trouter.HandleFunc(\"/stopStream\", work.StopStreamRequest)\n\thttp.Handle(\"/\", router)\n\n\twork.apiServer = &http.Server{\n\t\tAddr: fmt.Sprintf(\":%d\", port),\n\t\tHandler: handlers.CORS()(router),\n\t}\n}", "func (swagger *MgwSwagger) PopulateFromAPIYaml(apiYaml APIYaml) error {\n\n\tdata := apiYaml.Data\n\t// UUID in the generated api.yaml file is considered as swagger.id\n\tswagger.id = data.ID\n\tswagger.apiType = data.APIType\n\t// name and version in api.yaml corresponds to title and version respectively.\n\tswagger.title = data.Name\n\tswagger.version = data.Version\n\t// context value in api.yaml is assigned as xWso2Basepath\n\tswagger.xWso2Basepath = data.Context + \"/\" + swagger.version\n\tswagger.LifecycleStatus = data.LifeCycleStatus\n\tswagger.IsDefaultVersion = data.IsDefaultVersion\n\n\t// Added with both HTTP and WS APIs. x-throttling-tier is not used with WS.\n\tswagger.xWso2ThrottlingTier = data.APIThrottlingPolicy\n\n\t// productionURL & sandBoxURL values are extracted from endpointConfig in api.yaml\n\tendpointConfig := data.EndpointConfig\n\n\tif endpointConfig.ImplementationStatus == constants.Prototyped {\n\t\tswagger.IsPrototyped = true\n\t}\n\n\tswagger.EndpointType = endpointConfig.EndpointType\n\tswagger.EndpointImplementationType = data.EndpointImplementationType\n\n\t// from here onwards it will process endpoint info\n\t// So discontinue if the implementation type is mocked_oas\n\tif data.EndpointImplementationType == constants.MockedOASEndpointType {\n\t\treturn nil\n\t}\n\n\tif len(endpointConfig.ProductionEndpoints) > 0 {\n\t\tvar endpoints []Endpoint\n\t\tendpointType := constants.LoadBalance\n\t\tvar unProcessedURLs []interface{}\n\t\tfor _, endpointConfig := range endpointConfig.ProductionEndpoints {\n\t\t\tif swagger.apiType == constants.WS {\n\t\t\t\tprodEndpoint, err := getWebSocketEndpoint(endpointConfig.Endpoint)\n\t\t\t\tif err == nil {\n\t\t\t\t\tendpoints = append(endpoints, *prodEndpoint)\n\t\t\t\t} else {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tunProcessedURLs = append(unProcessedURLs, endpointConfig.Endpoint)\n\t\t\t}\n\t\t}\n\t\tif len(endpointConfig.ProductionFailoverEndpoints) > 0 {\n\t\t\tendpointType = constants.FailOver\n\t\t\tfor _, endpointConfig := range endpointConfig.ProductionFailoverEndpoints {\n\t\t\t\tif swagger.apiType == constants.WS {\n\t\t\t\t\tfailoverEndpoint, err := getWebSocketEndpoint(endpointConfig.Endpoint)\n\t\t\t\t\tif err == nil {\n\t\t\t\t\t\tendpoints = append(endpoints, *failoverEndpoint)\n\t\t\t\t\t} else {\n\t\t\t\t\t\treturn err\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\tunProcessedURLs = append(unProcessedURLs, endpointConfig.Endpoint)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tif swagger.apiType != constants.WS {\n\t\t\tproductionEndpoints, err := processEndpointUrls(unProcessedURLs)\n\t\t\tif err == nil {\n\t\t\t\tendpoints = append(endpoints, productionEndpoints...)\n\t\t\t} else {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t\tswagger.productionEndpoints = generateEndpointCluster(constants.ProdClustersConfigNamePrefix, endpoints, endpointType)\n\t}\n\n\tif len(endpointConfig.SandBoxEndpoints) > 0 {\n\t\tvar endpoints []Endpoint\n\t\tendpointType := constants.LoadBalance\n\t\tvar unProcessedURLs []interface{}\n\t\tfor _, endpointConfig := range endpointConfig.SandBoxEndpoints {\n\t\t\tif swagger.apiType == constants.WS {\n\t\t\t\tsandBoxEndpoint, err := getWebSocketEndpoint(endpointConfig.Endpoint)\n\t\t\t\tif err == nil {\n\t\t\t\t\tendpoints = append(endpoints, *sandBoxEndpoint)\n\t\t\t\t} else {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tunProcessedURLs = append(unProcessedURLs, endpointConfig.Endpoint)\n\t\t\t}\n\t\t}\n\t\tif len(endpointConfig.SandboxFailoverEndpoints) > 0 {\n\t\t\tendpointType = constants.FailOver\n\t\t\tfor _, endpointConfig := range endpointConfig.SandboxFailoverEndpoints {\n\t\t\t\tif swagger.apiType == constants.WS {\n\t\t\t\t\tfailoverEndpoint, err := getWebSocketEndpoint(endpointConfig.Endpoint)\n\t\t\t\t\tif err == nil {\n\t\t\t\t\t\tendpoints = append(endpoints, *failoverEndpoint)\n\t\t\t\t\t} else {\n\t\t\t\t\t\treturn err\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\tunProcessedURLs = append(unProcessedURLs, endpointConfig.Endpoint)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tif swagger.apiType != constants.WS {\n\t\t\tsandboxEndpoints, err := processEndpointUrls(unProcessedURLs)\n\t\t\tif err == nil {\n\t\t\t\tendpoints = append(endpoints, sandboxEndpoints...)\n\t\t\t} else {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t\tswagger.sandboxEndpoints = generateEndpointCluster(constants.SandClustersConfigNamePrefix, endpoints, endpointType)\n\t}\n\n\t// if yaml has production security, setting it\n\tif swagger.productionEndpoints != nil && endpointConfig.APIEndpointSecurity.Production.Enabled {\n\t\tif endpointConfig.APIEndpointSecurity.Production.Type != \"\" && strings.EqualFold(\"BASIC\", endpointConfig.APIEndpointSecurity.Production.Type) {\n\t\t\tswagger.productionEndpoints.SecurityConfig = endpointConfig.APIEndpointSecurity.Production\n\t\t} else {\n\t\t\tendpointConfig.APIEndpointSecurity.Production.Enabled = false\n\t\t\tlogger.LoggerXds.Errorf(\"endpoint security type given in api.yaml : %v is not currently supported with WSO2 Choreo Connect\",\n\t\t\t\tendpointConfig.APIEndpointSecurity.Production.Type)\n\t\t}\n\t}\n\t// if yaml has sandbox security, setting it\n\tif swagger.sandboxEndpoints != nil && endpointConfig.APIEndpointSecurity.Sandbox.Enabled {\n\t\tif endpointConfig.APIEndpointSecurity.Sandbox.Type != \"\" && strings.EqualFold(\"BASIC\", endpointConfig.APIEndpointSecurity.Sandbox.Type) {\n\t\t\tswagger.sandboxEndpoints.SecurityConfig = endpointConfig.APIEndpointSecurity.Sandbox\n\t\t} else {\n\t\t\tendpointConfig.APIEndpointSecurity.Sandbox.Enabled = false\n\t\t\tlogger.LoggerXds.Errorf(\"endpoint security type given in api.yaml : %v is not currently supported with WSO2 Choreo Connect\",\n\t\t\t\tendpointConfig.APIEndpointSecurity.Sandbox.Type)\n\t\t}\n\t}\n\n\tif apiYaml.Data.APIType == constants.GRAPHQL {\n\t\terr := swagger.SetInfoGraphQLAPI(apiYaml)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func NewAPIWithVersion(config *conf.Configuration, db *gorm.DB, paypal *paypalsdk.Client, mailer *mailer.Mailer, assets assetstores.Store, version string) *API {\n\tapi := &API{\n\t\tlog: logrus.WithField(\"component\", \"api\"),\n\t\tconfig: config,\n\t\tdb: db,\n\t\tpaypal: paypal,\n\t\tmailer: mailer,\n\t\thttpClient: &http.Client{},\n\t\tassets: assets,\n\t\tversion: version,\n\t}\n\n\tmux := kami.New()\n\tmux.Use(\"/\", api.populateContext)\n\tmux.Use(\"/\", api.withToken)\n\tmux.LogHandler = api.logCompleted\n\n\t// endpoints\n\tmux.Get(\"/\", api.Index)\n\n\tmux.Get(\"/orders\", api.OrderList)\n\tmux.Post(\"/orders\", api.OrderCreate)\n\tmux.Get(\"/orders/:id\", api.OrderView)\n\tmux.Put(\"/orders/:id\", api.OrderUpdate)\n\tmux.Get(\"/orders/:order_id/payments\", api.PaymentListForOrder)\n\tmux.Post(\"/orders/:order_id/payments\", api.PaymentCreate)\n\tmux.Post(\"/orders/:order_id/receipt\", api.ResendOrderReceipt)\n\n\tmux.Get(\"/users\", api.UserList)\n\tmux.Get(\"/users/:user_id\", api.UserView)\n\tmux.Get(\"/users/:user_id/payments\", api.PaymentListForUser)\n\tmux.Delete(\"/users/:user_id\", api.UserDelete)\n\tmux.Get(\"/users/:user_id/addresses\", api.AddressList)\n\tmux.Get(\"/users/:user_id/addresses/:addr_id\", api.AddressView)\n\tmux.Delete(\"/users/:user_id/addresses/:addr_id\", api.AddressDelete)\n\tmux.Get(\"/users/:user_id/orders\", api.OrderList)\n\n\tmux.Get(\"/downloads/:id\", api.DownloadURL)\n\tmux.Get(\"/downloads\", api.DownloadList)\n\tmux.Get(\"/orders/:order_id/downloads\", api.DownloadList)\n\n\tmux.Get(\"/vatnumbers/:number\", api.VatnumberLookup)\n\n\tmux.Get(\"/payments\", api.PaymentList)\n\tmux.Get(\"/payments/:pay_id\", api.PaymentView)\n\tmux.Post(\"/payments/:pay_id/refund\", api.PaymentRefund)\n\n\tmux.Post(\"/paypal\", api.PaypalCreatePayment)\n\tmux.Get(\"/paypal/:payment_id\", api.PaypalGetPayment)\n\n\tmux.Get(\"/reports/sales\", api.SalesReport)\n\tmux.Get(\"/reports/products\", api.ProductsReport)\n\n\tmux.Get(\"/coupons/:code\", api.CouponView)\n\n\tmux.Post(\"/claim\", api.ClaimOrders)\n\n\tcorsHandler := cors.New(cors.Options{\n\t\tAllowedMethods: []string{\"GET\", \"POST\", \"PATCH\", \"PUT\", \"DELETE\"},\n\t\tAllowedHeaders: []string{\"Accept\", \"Authorization\", \"Content-Type\"},\n\t\tExposedHeaders: []string{\"Link\", \"X-Total-Count\"},\n\t\tAllowCredentials: true,\n\t})\n\n\tapi.handler = corsHandler.Handler(mux)\n\n\treturn api\n}", "func NewAPI(endpoint, from, accessKeyID, secretAccessKey string) *Option {\n\top := new(Option)\n\n\top.endpoint = endpoint\n\top.source = from\n\top.accessKeyID = accessKeyID\n\top.secretAccessKey = []byte(secretAccessKey)\n\n\treturn op\n}", "func NewAPI(root string) *API {\n\tapi := &API{\n\t\troot: root,\n\t}\n\treturn api\n}", "func NewAPI(c AlertmanagerClient) *API {\n\ta := &API{client: c}\n\tr := mux.NewRouter()\n\ta.RegisterRoutes(r)\n\ta.Handler = r\n\treturn a\n}", "func importAPI(lambdaArn string) string {\n\tsvc := apigateway.New(getAWSSession())\n\n\t// First check to see if there's already an API by the same name\n\t// (only pulls up to 100 APIs, so this isn't a great long term solution)\n\tapisResp, err := svc.GetRestApis(&apigateway.GetRestApisInput{\n\t\tLimit: aws.Int64(100),\n\t})\n\tif err != nil {\n\t\tfmt.Println(\"There was a problem creating the API.\")\n\t\tfmt.Println(err.Error())\n\t\tos.Exit(-1)\n\t}\n\tfor key := range apisResp.Items {\n\t\tif *apisResp.Items[key].Name == cfg.API.Name {\n\t\t\t// TODO: Prompt user to continue and add a new API anyway. Or remove/overwrite/ignore?\n\t\t\t// Inspect the same named APIs and see if there's a {proxy+} path?\n\t\t\t// It's possible to have multiple APIs with the same name. I hate to break this into\n\t\t\t// multiple commands/steps, it's nice just running `up` and nothing else...But it's not\n\t\t\t// perfect because the user doesn't set the unique identifier for the API.\n\t\t\tfmt.Println(\"API already exists.\")\n\t\t\treturn *apisResp.Items[key].Id\n\t\t}\n\t}\n\n\t// Build Swagger\n\tswaggerDefinition, swaggerErr := swagger.NewSwagger(&swagger.SwaggerConfig{\n\t\tTitle: cfg.API.Name,\n\t\tLambdaURI: swagger.GetLambdaURI(lambdaArn),\n\t\tResourceTimeoutMs: cfg.API.ResourceTimeoutMs,\n\t\t// BinaryMediaTypes: cfg.API.BinaryMediaTypes,\n\t})\n\tif swaggerErr != nil {\n\t\tfmt.Println(swaggerErr.Error())\n\t\tos.Exit(-1)\n\t}\n\n\tswaggerBytes, err := json.Marshal(swaggerDefinition)\n\tif err != nil {\n\t\tfmt.Println(\"There was a problem creating the API.\")\n\t\tfmt.Println(err.Error())\n\t\tos.Exit(-1)\n\t}\n\n\t// Import from Swagger\n\tresp, err := svc.ImportRestApi(&apigateway.ImportRestApiInput{\n\t\tBody: swaggerBytes, // Required\n\t\tFailOnWarnings: aws.Bool(true),\n\t})\n\tif err != nil {\n\t\tfmt.Println(\"There was a problem creating the API.\")\n\t\tfmt.Println(err.Error())\n\t\tos.Exit(-1)\n\t}\n\n\treturn *resp.Id\n}", "func New(\n\tctx context.Context,\n\tconf *apisettings.Service,\n\thandlers []Handler) *API {\n\n\tbaseHandler := newBaseHandler()\n\thandlers = append(handlers, baseHandler)\n\n\tapi := API{\n\t\tconf: conf,\n\t\thandlers: handlers}\n\n\tapi.initRouter(ctx)\n\treturn &api\n}", "func NewAPI(daprID string, appChannel channel.AppChannel, stateStores map[string]state.Store, pubSub pubsub.PubSub, directMessaging messaging.DirectMessaging, actor actors.Actors, sendToOutputBindingFn func(name string, req *bindings.WriteRequest) error, componentHandler components.ComponentHandler) API {\n\treturn &api{\n\t\tdirectMessaging: directMessaging,\n\t\tcomponentsHandler: componentHandler,\n\t\tactor: actor,\n\t\tid: daprID,\n\t\tappChannel: appChannel,\n\t\tpubSub: pubSub,\n\t\tstateStores: stateStores,\n\t\tsendToOutputBindingFn: sendToOutputBindingFn,\n\t}\n}", "func NewApp(config models.Config) (MainWebAPI, error) {\n\tvar err error\n\tvar wapp MainWebAPI\n\n\tmux := mux.NewRouter().StrictSlash(true)\n\n\tlog := log.New(os.Stdout, \"API\", log.LstdFlags)\n\twapp.Mux = mux\n\twapp.Config = config\n\twapp.Log = log\n\n\tif err != nil {\n\t\tlog.Println(err)\n\t}\n\treturn wapp, err\n}", "func CreateApi(g *gorm.DB) http.Handler {\n\th := ManagerHandler{g: g}\n\tr := httprouter.New()\n\tr.GET(\"/task\", h.TasksAll)\n\tr.PUT(\"/task\", h.TaskCreate)\n\tr.POST(\"/task/:id\", h.TaskUpdate)\n\tr.GET(\"/task/:id/detail\", h.TaskDetail)\n\tr.GET(\"/task/:id\", h.TaskSingle)\n\treturn r\n}", "func NewAPI(heartbeat time.Duration) (conn *API,\n\terr error) {\n\n\t// Create the API and initialize the connection listener\n\tconn = &API{\n\t\tlistener: make(chan net.Conn),\n\t}\n\n\tstompCfg := stompServer.Server{\n\t\tHeartBeat: heartbeat,\n\t}\n\t// Start a new stomp server and wait for new connections\n\tgo stompCfg.Serve(conn.listener)\n\n\t// Connect to the server locally\n\terr = conn.connectLocal()\n\tif err != nil {\n\t\tfmt.Println(\"Errrr\")\n\t\treturn nil, err\n\t}\n\treturn\n}", "func CreateAPI() GetSamplesAllAPI {\n\tif api == nil {\n\t\treturn GetSamplesAllAPIImpl{\n\t\t\tInteractor: core.CreateDefaultGetSamples(\"Kaka\", \"Ronaldo\"),\n\t\t}\n\t}\n\treturn api\n}", "func (api *API) setupRoutes() error {\n\tvar (\n\t\tconnLimit int\n\t\terr error\n\t)\n\tif api.cfg.API.Connection.Limit == \"\" {\n\t\tconnLimit = 50\n\t} else {\n\t\t// setup the connection limit\n\t\tconnLimit, err = strconv.Atoi(api.cfg.API.Connection.Limit)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\t// ensure we have valid cors configuration, otherwise default to allow all\n\tvar allowedOrigins []string\n\tif len(api.cfg.API.Connection.CORS.AllowedOrigins) > 0 {\n\t\tallowedOrigins = api.cfg.API.Connection.CORS.AllowedOrigins\n\t}\n\t// set up defaults\n\tapi.r.Use(\n\t\t// allows for automatic xss removal\n\t\t// greater than what can be configured with HTTP Headers\n\t\txssMdlwr.RemoveXss(),\n\t\t// rate limiting\n\t\tlimit.MaxAllowed(connLimit),\n\t\t// security middleware\n\t\tmiddleware.NewSecWare(dev),\n\t\t// cors middleware\n\t\tmiddleware.CORSMiddleware(dev, allowedOrigins),\n\t\t// request id middleware\n\t\tmiddleware.RequestID(),\n\t\t// stats middleware\n\t\tstats.RequestStats())\n\n\t// set up middleware\n\tginjwt := middleware.JwtConfigGenerate(api.cfg.JWT.Key, api.cfg.JWT.Realm, api.dbm.DB, api.l)\n\tauthware := []gin.HandlerFunc{ginjwt.MiddlewareFunc()}\n\n\t// V2 API\n\tv2 := api.r.Group(\"/v2\")\n\n\t// system checks used to verify the integrity of our services\n\tsystemChecks := v2.Group(\"/systems\")\n\t{\n\t\tsystemChecks.GET(\"/check\", api.SystemsCheck)\n\t}\n\n\t// authless account recovery routes\n\tforgot := v2.Group(\"/forgot\")\n\t{\n\t\tforgot.POST(\"/username\", api.forgotUserName)\n\t\tforgot.POST(\"/password\", api.resetPassword)\n\t}\n\n\t// authentication\n\tauth := v2.Group(\"/auth\")\n\t{\n\t\tauth.POST(\"/register\", api.registerUserAccount)\n\t\tauth.POST(\"/login\", ginjwt.LoginHandler)\n\t\tauth.GET(\"/refresh\", ginjwt.RefreshHandler)\n\t}\n\n\t// statistics\n\tstatistics := v2.Group(\"/statistics\").Use(authware...)\n\t{\n\t\tstatistics.GET(\"/stats\", api.getStats)\n\t}\n\n\t// lens search engine\n\tlens := v2.Group(\"/lens\")\n\t{\n\t\tlens.POST(\"/index\", api.submitIndexRequest)\n\t\t// only allow registered users to search\n\t\tlens.POST(\"/search\", api.submitSearchRequest)\n\t}\n\n\t// payments\n\tpayments := v2.Group(\"/payments\", authware...)\n\t{\n\t\tdash := payments.Group(\"/dash\")\n\t\t{\n\t\t\tdash.POST(\"/create\", api.CreateDashPayment)\n\t\t}\n\t\teth := payments.Group(\"/eth\")\n\t\t{\n\t\t\teth.POST(\"/request\", api.RequestSignedPaymentMessage)\n\t\t\teth.POST(\"/confirm\", api.ConfirmETHPayment)\n\t\t}\n\t\tbch := payments.Group(\"/bch\")\n\t\t{\n\t\t\tbch.POST(\"/create\", api.createBchPayment)\n\t\t\tbch.POST(\"/confirm\", api.confirmBchPayment)\n\t\t}\n\t\tstripe := payments.Group(\"/stripe\")\n\t\t{\n\t\t\tstripe.POST(\"/charge\", api.stripeCharge)\n\t\t}\n\t\tpayments.GET(\"/status/:number\", api.getPaymentStatus)\n\t}\n\n\t// accounts\n\taccount := v2.Group(\"/account\")\n\t{\n\t\ttoken := account.Group(\"/token\", authware...)\n\t\t{\n\t\t\ttoken.GET(\"/username\", api.getUserFromToken)\n\t\t}\n\t\tpassword := account.Group(\"/password\", authware...)\n\t\t{\n\t\t\tpassword.POST(\"/change\", api.changeAccountPassword)\n\t\t}\n\t\tkey := account.Group(\"/key\", authware...)\n\t\t{\n\t\t\tkey.GET(\"/export/:name\", api.exportKey)\n\t\t\tipfs := key.Group(\"/ipfs\")\n\t\t\t{\n\t\t\t\tipfs.GET(\"/get\", api.getIPFSKeyNamesForAuthUser)\n\t\t\t\tipfs.POST(\"/new\", api.createIPFSKey)\n\t\t\t}\n\t\t}\n\t\tcredits := account.Group(\"/credits\", authware...)\n\t\t{\n\t\t\tcredits.GET(\"/available\", api.getCredits)\n\t\t}\n\t\temail := account.Group(\"/email\")\n\t\t{\n\t\t\t// auth-less account email routes\n\t\t\ttoken := email.Group(\"/verify\")\n\t\t\t{\n\t\t\t\ttoken.GET(\"/:user/:token\", api.verifyEmailAddress)\n\t\t\t}\n\t\t\t// authenticatoin email routes\n\t\t\tauth := email.Use(authware...)\n\t\t\t{\n\t\t\t\tauth.POST(\"/forgot\", api.forgotEmail)\n\t\t\t}\n\t\t}\n\t\tauth := account.Use(authware...)\n\t\t{\n\t\t\t// used to upgrade account to light tier\n\t\t\tauth.POST(\"/upgrade\", api.upgradeAccount)\n\t\t\tauth.GET(\"/usage\", api.usageData)\n\t\t}\n\t}\n\n\t// ipfs routes\n\tipfs := v2.Group(\"/ipfs\", authware...)\n\t{\n\t\t// public ipfs routes\n\t\tpublic := ipfs.Group(\"/public\")\n\t\t{\n\t\t\t// pinning routes\n\t\t\tpin := public.Group(\"/pin\")\n\t\t\t{\n\t\t\t\tpin.POST(\"/:hash\", api.pinHashLocally)\n\t\t\t\tpin.POST(\"/:hash/extend\", api.extendPin)\n\t\t\t}\n\t\t\t// file upload routes\n\t\t\tfile := public.Group(\"/file\")\n\t\t\t{\n\t\t\t\tfile.POST(\"/add\", api.addFile)\n\t\t\t\tfile.POST(\"/add/directory\", api.uploadDirectory)\n\t\t\t}\n\t\t\t// pubsub routes\n\t\t\tpubsub := public.Group(\"/pubsub\")\n\t\t\t{\n\t\t\t\tpubsub.POST(\"/publish/:topic\", api.ipfsPubSubPublish)\n\t\t\t}\n\t\t\t// general routes\n\t\t\tpublic.GET(\"/stat/:hash\", api.getObjectStatForIpfs)\n\t\t\tpublic.GET(\"/dag/:hash\", api.getDagObject)\n\t\t}\n\n\t\t// private ipfs routes\n\t\tprivate := ipfs.Group(\"/private\")\n\t\t{\n\t\t\t// network management routes\n\t\t\tprivate.GET(\"/networks\", api.getAuthorizedPrivateNetworks)\n\t\t\tnetwork := private.Group(\"/network\")\n\t\t\t{\n\t\t\t\tusers := network.Group(\"/users\")\n\t\t\t\t{\n\t\t\t\t\tusers.DELETE(\"/remove\", api.removeUsersFromNetwork)\n\t\t\t\t\tusers.POST(\"/add\", api.addUsersToNetwork)\n\t\t\t\t}\n\t\t\t\towners := network.Group(\"/owners\")\n\t\t\t\t{\n\t\t\t\t\towners.POST(\"/add\", api.addOwnersToNetwork)\n\t\t\t\t}\n\t\t\t\tnetwork.GET(\"/:name\", api.getIPFSPrivateNetworkByName)\n\t\t\t\tnetwork.POST(\"/new\", api.createIPFSNetwork)\n\t\t\t\tnetwork.POST(\"/stop\", api.stopIPFSPrivateNetwork)\n\t\t\t\tnetwork.POST(\"/start\", api.startIPFSPrivateNetwork)\n\t\t\t\tnetwork.DELETE(\"/remove\", api.removeIPFSPrivateNetwork)\n\t\t\t}\n\t\t\t// pinning routes\n\t\t\tpin := private.Group(\"/pin\")\n\t\t\t{\n\t\t\t\tpin.POST(\"/:hash\", api.pinToHostedIPFSNetwork)\n\t\t\t\tpin.GET(\"/check/:hash/:networkName\", api.checkLocalNodeForPinForHostedIPFSNetwork)\n\t\t\t}\n\t\t\t// file upload routes\n\t\t\tfile := private.Group(\"/file\")\n\t\t\t{\n\t\t\t\tfile.POST(\"/add\", api.addFileToHostedIPFSNetwork)\n\t\t\t}\n\t\t\t// pubsub routes\n\t\t\tpubsub := private.Group(\"/pubsub\")\n\t\t\t{\n\t\t\t\tpubsub.POST(\"/publish/:topic\", api.ipfsPubSubPublishToHostedIPFSNetwork)\n\t\t\t}\n\t\t\t// object stat route\n\t\t\tprivate.GET(\"/stat/:hash/:networkName\", api.getObjectStatForIpfsForHostedIPFSNetwork)\n\t\t\t// general routes\n\t\t\tprivate.GET(\"/dag/:hash/:networkName\", api.getDagObjectForHostedIPFSNetwork)\n\t\t\tprivate.GET(\"/uploads/:networkName\", api.getUploadsByNetworkName)\n\t\t}\n\t\t// utility routes\n\t\tutils := ipfs.Group(\"/utils\")\n\t\t{\n\t\t\t// generic download\n\t\t\tutils.POST(\"/download/:hash\", api.downloadContentHash)\n\t\t\tlaser := utils.Group(\"/laser\")\n\t\t\t{\n\t\t\t\tlaser.POST(\"/beam\", api.beamContent)\n\t\t\t}\n\t\t}\n\t}\n\n\t// ipns\n\tipns := v2.Group(\"/ipns\", authware...)\n\t{\n\t\t// public ipns routes\n\t\tpublic := ipns.Group(\"/public\")\n\t\t{\n\t\t\tpublic.POST(\"/publish/details\", api.publishToIPNSDetails)\n\t\t\t// used to handle pinning of IPNS records on public ipfs\n\t\t\t// this involves first resolving the record, parsing it\n\t\t\t// and extracting the hash to pin\n\t\t\tpublic.POST(\"/pin\", api.pinIPNSHash)\n\t\t}\n\t\t// general routes\n\t\tipns.GET(\"/records\", api.getIPNSRecordsPublishedByUser)\n\t}\n\n\t// database\n\tdatabase := v2.Group(\"/database\", authware...)\n\t{\n\t\tdatabase.GET(\"/uploads\", api.getUploadsForUser)\n\t\tdatabase.GET(\"/uploads/encrypted\", api.getEncryptedUploadsForUser)\n\t}\n\n\t// frontend\n\tfrontend := v2.Group(\"/frontend\", authware...)\n\t{\n\t\tcost := frontend.Group(\"/cost\")\n\t\t{\n\t\t\tcalculate := cost.Group(\"/calculate\")\n\t\t\t{\n\t\t\t\tcalculate.GET(\"/:hash/:hold_time\", api.calculatePinCost)\n\t\t\t\tcalculate.POST(\"/file\", api.calculateFileCost)\n\t\t\t}\n\t\t}\n\t}\n\n\tapi.l.Info(\"Routes initialized\")\n\treturn nil\n}", "func NewApp(ctx context.Context, appCfg AppConfig) (*API, error) {\n\tappCfg.checkConfig()\n\n\tlog.Debug().Interface(\"api app config\", appCfg).Msg(\"starting initialize api application\")\n\n\te := echo.New()\n\n\ta := &API{\n\t\te: e,\n\t\taddr: appCfg.NetInterface,\n\t}\n\n\te.Use(func(next echo.HandlerFunc) echo.HandlerFunc {\n\t\treturn func(c echo.Context) error {\n\t\t\tcc := &Context{\n\t\t\t\tContext: c,\n\t\t\t\tCtx: ctx,\n\t\t\t}\n\t\t\treturn next(cc)\n\t\t}\n\t})\n\te.Validator = &Validator{validator: validator.New()}\n\te.Use(logMiddleware)\n\n\te.GET(\"/healthcheck\", a.handleHealthcheck)\n\n\tg := e.Group(\"/api\")\n\tg.GET(\"/:collection/documents\", a.handleSearch)\n\tg.POST(\"/:collection/documents\", a.handleAddDocuments)\n\n\tlog.Debug().Msg(\"endpoints registered\")\n\n\treturn a, nil\n}", "func CreateClosedAPIIndexConfig() *SpecIndexConfig {\n\tcw, _ := os.Getwd()\n\treturn &SpecIndexConfig{\n\t\tBasePath: cw,\n\t\tAllowRemoteLookup: false,\n\t\tAllowFileLookup: false,\n\t\tseenRemoteSources: &syncmap.Map{},\n\t}\n}", "func New(l *logrus.Logger, c *configurations.Collector, d *debugger.Debugger, path string) *APIServer {\n\n\t// create new instance of configurations interface\n\tcfg, err := configurations.New(configurations.CONF_TYPE_API_SERVER, path)\n\tif err != nil {\n\t\td.Verbose(fmt.Sprintf(\"[%d]-%s: (%v)\",\n\t\t\tconfigurations.ERROR_READ_CONFIG.Int(),\n\t\t\tconfigurations.ERROR_READ_CONFIG, err),\n\t\t\tlogrus.ErrorLevel,\n\t\t)\n\t\tos.Exit(configurations.ERROR_READ_CONFIG.Int())\n\t}\n\n\t// Read config & return the requested struct type\n\tcf, err := cfg.Read()\n\tif err != nil {\n\t\td.Verbose(fmt.Sprintf(\"[%d]-%s: (%v)\",\n\t\t\tconfigurations.ERROR_READ_CONFIG.Int(),\n\t\t\tconfigurations.ERROR_READ_CONFIG, err),\n\t\t\tlogrus.ErrorLevel,\n\t\t)\n\t\tos.Exit(configurations.ERROR_READ_CONFIG.Int())\n\t}\n\n\tconfigs := cf.(*configurations.APIServer)\n\n\t// open http access log file\n\talfHasError := false\n\taccessLogfile, err := os.OpenFile(configs.AccessLogFile, os.O_RDWR|os.O_CREATE|os.O_APPEND, 0666)\n\tif err != nil {\n\t\talfHasError = true\n\t\td.Verbose(fmt.Sprintf(\"[%d]-%s: (%v)\",\n\t\t\tconfigurations.ERROR_CAN_T_OPEN_OR_CREATE_HTTP_ACCESS_LOG_FILE.Int(),\n\t\t\tconfigurations.ERROR_CAN_T_OPEN_OR_CREATE_HTTP_ACCESS_LOG_FILE, err),\n\t\t\tlogrus.ErrorLevel,\n\t\t)\n\t}\n\n\t// open http error log file\n\telfHasError := false\n\terrorLogfile, err := os.OpenFile(configs.ErrorLogFile, os.O_RDWR|os.O_CREATE|os.O_APPEND, 0666)\n\tif err != nil {\n\t\telfHasError = true\n\t\td.Verbose(fmt.Sprintf(\"[%d]-%s: (%v)\",\n\t\t\tconfigurations.ERROR_CAN_T_OPEN_OR_CREATE_HTTP_ERROR_LOG_FILE.Int(),\n\t\t\tconfigurations.ERROR_CAN_T_OPEN_OR_CREATE_HTTP_ERROR_LOG_FILE, err),\n\t\t\tlogrus.ErrorLevel,\n\t\t)\n\t}\n\n\t// getIP2location conf\n\t// create new instance of configurations interface\n\tcfgLoc, err := configurations.New(configurations.CONF_TYPE_IP2LOCATION, path)\n\tif err != nil {\n\t\td.Verbose(fmt.Sprintf(\"[%d]-%s: (%v)\",\n\t\t\tconfigurations.ERROR_READ_CONFIG.Int(),\n\t\t\tconfigurations.ERROR_READ_CONFIG, err),\n\t\t\tlogrus.ErrorLevel,\n\t\t)\n\t\tos.Exit(configurations.ERROR_READ_CONFIG.Int())\n\t}\n\n\t// Read config & return the requested strucut type\n\tcfl, err := cfgLoc.Read()\n\tif err != nil {\n\t\td.Verbose(fmt.Sprintf(\"[%d]-%s: (%v)\",\n\t\t\tconfigurations.ERROR_READ_CONFIG.Int(),\n\t\t\tconfigurations.ERROR_READ_CONFIG, err),\n\t\t\tlogrus.ErrorLevel,\n\t\t)\n\t\tos.Exit(configurations.ERROR_READ_CONFIG.Int())\n\t}\n\n\t// make new instance of ip2location\n\ti2l := location.New(cfl.(*configurations.IP2Location), d)\n\n\t// collector will connect to pg-bouncer\n\t// and API server will connect to postgres db directly\n\t// so Exporter configs will be ommited and commented and\n\t// after that we will initialize our pgsql\n\n\t// var pgDB database.Postgres\n\t// for _, ex := range c.Exporter.Postgres {\n\t// \t// create new Postgres\n\t// \tpgDB = database.New(ex.Host, ex.User, ex.Password, ex.DB, c.IPReputation, ex.Port, d, i2l, 20, 50, 1*time.Hour)\n\n\t// \t// just get the very first one inside api server\n\t// \tbreak\n\t// }\n\n\tvar pgDB database.Postgres\n\tfor _, pg := range configs.Postgres {\n\t\tpgDB = database.New(pg.Host, pg.User, pg.Password, pg.DB, c.IPReputation, pg.Port, d, i2l, 20, 50, 1*time.Hour)\n\n\t\t// just get the very first one inside api server\n\t\tbreak\n\t}\n\n\tapi := &APIServer{\n\t\thost: configs.Listen.Address,\n\t\tport: configs.Listen.Port,\n\t\tl: l,\n\t\tc: c,\n\t\tapiConf: configs,\n\t\td: d,\n\n\t\tpgdb: pgDB,\n\t\tdb: pgDB.GetDB(),\n\n\t\tip2l: i2l,\n\n\t\thttpAccessHasError: alfHasError,\n\t\thttpErrorHasError: elfHasError,\n\t\thttpAccessLog: accessLogfile,\n\t\thttpErrorLog: errorLogfile,\n\n\t\tch: make(chan os.Signal, 1),\n\t\twaitGroup: &sync.WaitGroup{},\n\t}\n\n\tapi.ch = make(chan os.Signal, 1)\n\tsignal.Notify(api.ch,\n\t\t// https://www.gnu.org/software/libc/manual/html_node/Termination-Signals.html\n\t\tsyscall.SIGTERM, // \"the normal way to politely ask a program to terminate\"\n\t\tsyscall.SIGINT, // Ctrl+C\n\t\tsyscall.SIGQUIT, // Ctrl-\\\n\t\tsyscall.SIGKILL, // \"always fatal\", \"SIGKILL and SIGSTOP may not be caught by a program\"\n\t\tsyscall.SIGHUP, // \"terminal is disconnected\"\n\t)\n\n\t// catch signals\n\tgo func() {\n\t\t<-api.ch\n\n\t\tapi.d.Verbose(\"Stopping API HTTP server...!\", logrus.InfoLevel)\n\n\t\t// close http log files\n\t\tdefer api.closeLogFiles()\n\n\t\tif err := api.httpSrv.Close(); err != nil {\n\t\t\tapi.d.Verbose(fmt.Sprintf(\"[%d]-%s: (%v)\",\n\t\t\t\tconfigurations.ERROR_CAN_T_STOP_API_HTTP_SERVER.Int(),\n\t\t\t\tconfigurations.ERROR_CAN_T_STOP_API_HTTP_SERVER, err),\n\t\t\t\tlogrus.ErrorLevel,\n\t\t\t)\n\t\t\tos.Exit(configurations.ERROR_CAN_T_STOP_API_HTTP_SERVER.Int())\n\t\t} else {\n\t\t\tapi.d.Verbose(\"API Server has stopped!\", logrus.InfoLevel)\n\t\t\tos.Exit(0)\n\t\t}\n\t}()\n\n\treturn api\n}", "func New(config Config) IAPI {\n\tif config.BaseURL == \"\" {\n\t\tconfig.BaseURL = defaultAPIURL\n\t}\n\tif config.Timeout == nil {\n\t\tconfig.Timeout = newDuration(defaultTimeout)\n\t}\n\n\tapi := &API{config: config}\n\tapi.client = &http.Client{\n\t\tTransport: api,\n\t\tTimeout: *config.Timeout,\n\t}\n\treturn api\n}", "func NewAPI(port int, authKey string, zone string, storage Storage) *API {\n\treturn &API{port: port, authKey: authKey, zone: zone, storage: storage}\n}", "func NewAPI() API {\n\treturn API{}\n}", "func RunCreateAPI(cfg *config.Config, gvk config.GVK) error {\n\tif !HasPluginConfig(cfg) {\n\t\treturn nil\n\t}\n\treturn manifests.RunCreateAPI(cfg, gvk)\n}", "func CreateConfig(confPath string) (JSONConfig, error) {\n\tc := JSONConfig{\n\t\tHostname: \"streamer.xplex.online\",\n\t\tRigURL: \"https://rig-dev.xplex.online\",\n\t\tRTMPPort: 1935,\n\t\tHTTPPort: 8086,\n\t\tAgentKey: \"alongsharedkeyhereusedforagents\",\n\t}\n\tj, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\treturn c, err\n\t}\n\terr = ioutil.WriteFile(confPath, j, 0644)\n\tif err != nil {\n\t\treturn c, err\n\t}\n\treturn c, nil\n}", "func ExampleConfiguration(pluginNames []string) (*ApiplexConfig, error) {\n\tc := ApiplexConfig{\n\t\tRedis: apiplexConfigRedis{\n\t\t\tHost: \"127.0.0.1\",\n\t\t\tPort: 6379,\n\t\t\tDB: 0,\n\t\t},\n\t\tQuotas: map[string]apiplexQuota{\n\t\t\t\"default\": apiplexQuota{\n\t\t\t\tMinutes: 5,\n\t\t\t\tMaxIP: 50,\n\t\t\t\tMaxKey: 5000,\n\t\t\t},\n\t\t\t\"keyless\": apiplexQuota{\n\t\t\t\tMinutes: 5,\n\t\t\t\tMaxIP: 20,\n\t\t\t},\n\t\t},\n\t\tServe: apiplexConfigServe{\n\t\t\tPort: 5000,\n\t\t\tAPI: \"/\",\n\t\t\tUpstreams: []string{\"http://your-actual-api:8000/\"},\n\t\t\tPortalAPI: \"/portal/api/\",\n\t\t\tPortal: \"/portal/\",\n\t\t\tSigningKey: uniuri.NewLen(64),\n\t\t},\n\t}\n\tplugins := apiplexConfigPlugins{}\n\tfor _, pname := range pluginNames {\n\t\tpInfo, ok := registeredPlugins[pname]\n\t\tif !ok {\n\t\t\treturn nil, fmt.Errorf(\"No plugin '%s' available.\", pname)\n\t\t}\n\n\t\tpluginPtr := reflect.New(pInfo.pluginType)\n\t\tdefConfig := pluginPtr.MethodByName(\"DefaultConfig\").Call([]reflect.Value{})[0].Interface().(map[string]interface{})\n\t\tpconfig := apiplexPluginConfig{Plugin: pname, Config: defConfig}\n\n\t\tswitch pluginPtr.Interface().(type) {\n\t\tcase AuthPlugin:\n\t\t\tplugins.Auth = append(plugins.Auth, pconfig)\n\t\tcase ManagementBackendPlugin:\n\t\t\tplugins.Backend = append(plugins.Backend, pconfig)\n\t\tcase BackendPlugin:\n\t\t\tplugins.Backend = append(plugins.Backend, pconfig)\n\t\tcase PreUpstreamPlugin:\n\t\t\tplugins.PreUpstream = append(plugins.PreUpstream, pconfig)\n\t\tcase PostUpstreamPlugin:\n\t\t\tplugins.PostUpstream = append(plugins.PostUpstream, pconfig)\n\t\tcase PostAuthPlugin:\n\t\t\tplugins.PostAuth = append(plugins.PostAuth, pconfig)\n\t\tcase LoggingPlugin:\n\t\t\tplugins.Logging = append(plugins.Logging, pconfig)\n\t\t}\n\t}\n\tc.Plugins = plugins\n\treturn &c, nil\n}", "func (m *APILoader) RegisterAPI(def *api.Definition) {\n\tlogger := log.WithField(\"api_name\", def.Name)\n\tlogger.Debug(\"Starting RegisterAPI\")\n\n\tactive, err := def.Validate()\n\tif false == active && err != nil {\n\t\tlogger.WithError(err).Error(\"Validation errors\")\n\t}\n\n\tif false == def.Active {\n\t\tlogger.Warn(\"API is not active, skipping...\")\n\t\tactive = false\n\t}\n\n\tif active {\n\t\trouterDefinition := proxy.NewRouterDefinition(def.Proxy)\n\n\t\tfor _, plg := range def.Plugins {\n\t\t\tl := logger.WithField(\"name\", plg.Name)\n\n\t\t\tisValid, err := plugin.ValidateConfig(plg.Name, plg.Config)\n\t\t\tif !isValid || err != nil {\n\t\t\t\tl.WithError(err).Error(\"Plugin configuration is invalid\")\n\t\t\t}\n\n\t\t\tif plg.Enabled {\n\t\t\t\tl.Debug(\"Plugin enabled\")\n\n\t\t\t\tsetup, err := plugin.DirectiveAction(plg.Name)\n\t\t\t\tif err != nil {\n\t\t\t\t\tl.WithError(err).Error(\"Error loading plugin\")\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\terr = setup(routerDefinition, plg.Config)\n\t\t\t\tif err != nil {\n\t\t\t\t\tl.WithError(err).Error(\"Error executing plugin\")\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tl.Debug(\"Plugin not enabled\")\n\t\t\t}\n\t\t}\n\n\t\tif len(def.Proxy.Hosts) > 0 {\n\t\t\trouterDefinition.AddMiddleware(middleware.NewHostMatcher(def.Proxy.Hosts).Handler)\n\t\t}\n\n\t\t// Add middleware to insert tags to context\n\t\ttags := []tag.Mutator{\n\t\t\ttag.Insert(obs.KeyListenPath, def.Proxy.ListenPath),\n\t\t}\n\t\trouterDefinition.AddMiddleware(middleware.NewStatsTagger(tags).Handler)\n\n\t\tm.register.Add(routerDefinition)\n\t\tlogger.Debug(\"API registered\")\n\t} else {\n\t\tlogger.WithError(err).Warn(\"API URI is invalid or not active, skipping...\")\n\t}\n}", "func NewAPI(root string, pg *ast.CXProgram) *API {\n\tif root == \"\" {\n\t\troot = \"/\"\n\t}\n\treturn &API{root: root, pg: pg}\n}", "func NewAPI(s *session.Session, server *server.Server) *API {\n\tapi := API{\n\t\tSession: s,\n\t\tServer: server,\n\t\tdb: server.Db,\n\t\tIPPort: s.GetAPIIPPort(),\n\t\tVersion: Version,\n\t\tPrefix: \"/api/\" + Version,\n\t}\n\n\tapi.Routes()\n\tapi.setupSignal()\n\treturn &api\n}", "func NewAPI() *API {\n\treturn &API{}\n}", "func NewAPI(ctx context.Context, client client.Client, component *rainbondv1alpha1.RbdComponent, cluster *rainbondv1alpha1.RainbondCluster) ComponentHandler {\n\treturn &api{\n\t\tctx: ctx,\n\t\tclient: client,\n\t\tcomponent: component,\n\t\tcluster: cluster,\n\t\tlabels: LabelsForRainbondComponent(component),\n\t\tpvcName: \"rbd-api\",\n\t\tdataStorageRequest: getStorageRequest(\"API_DATA_STORAGE_REQUEST\", 1),\n\t\tgrdataStorageRequest: getStorageRequest(\"GRDATA_STORAGE_REQUEST\", 40),\n\t}\n}", "func New() *API {\n\treturn &API{\n\t\tprefix: \"/\",\n\t\tresources: make(map[string]Rest),\n\t\troutes: &node{},\n\t}\n}", "func New(token string) *API {\n\treturn &API{\n\t\tToken: token,\n\t\thasHandlers: false,\n\t}\n}", "func (agr *apiGatewayResource) Create(request *http.Request) (string, restful.Attributes, error) {\n\tctx := request.Context()\n\tapiGatewayInfo, err := agr.getAPIGatewayInfoFromRequest(request)\n\tif err != nil {\n\t\tagr.Logger.WarnWithCtx(ctx, \"Failed to get api gateway config and status from body\", \"err\", err)\n\t\treturn \"\", nil, err\n\t}\n\n\treturn agr.createAPIGateway(request, apiGatewayInfo)\n}", "func DefaultRestAPIConfig() RestAPIConfig {\n\treturn RestAPIConfig{\n\t\tBindAddress: \"0.0.0.0:14265\",\n\t\tPermittedRoutes: []string{\n\t\t\t\"/health\",\n\t\t\t\"/mqtt\",\n\t\t\t\"/api/v1/info\",\n\t\t\t\"/api/v1/tips\",\n\t\t\t\"/api/v1/messages/:messageID\",\n\t\t\t\"/api/v1/messages/:messageID/metadata\",\n\t\t\t\"/api/v1/messages/:messageID/raw\",\n\t\t\t\"/api/v1/messages/:messageID/children\",\n\t\t\t\"/api/v1/messages\",\n\t\t\t\"/api/v1/transactions/:transactionID/included-message\",\n\t\t\t\"/api/v1/milestones/:milestoneIndex\",\n\t\t\t\"/api/v1/outputs/:outputID\",\n\t\t\t\"/api/v1/addresses/:address\",\n\t\t\t\"/api/v1/addresses/:address/outputs\",\n\t\t\t\"/api/v1/addresses/ed25519/:address\",\n\t\t\t\"/api/v1/addresses/ed25519/:address/outputs\",\n\t\t\t\"/api/v1/treasury\",\n\t\t\t\"/api/v1/receipts\",\n\t\t\t\"/api/v1/receipts/:milestoneIndex\",\n\t\t\t\"/api/v1/peers/:peerID\",\n\t\t\t\"/api/v1/peers\",\n\t\t\t\"/api/plugins/*\",\n\t\t},\n\t\tPoWEnabled: true,\n\t}\n}", "func getAPIResourceConfig(options configapi.MasterConfig) genericapiserver.APIResourceConfigSource {\n\tresourceConfig := genericapiserver.NewResourceConfig()\n\n\tfor group := range configapi.KnownKubeAPIGroups {\n\t\tfor _, version := range configapi.GetEnabledAPIVersionsForGroup(*options.KubernetesMasterConfig, group) {\n\t\t\tgv := unversioned.GroupVersion{Group: group, Version: version}\n\t\t\tresourceConfig.EnableVersions(gv)\n\t\t}\n\t}\n\n\tfor group := range options.KubernetesMasterConfig.DisabledAPIGroupVersions {\n\t\tfor _, version := range configapi.GetDisabledAPIVersionsForGroup(*options.KubernetesMasterConfig, group) {\n\t\t\tgv := unversioned.GroupVersion{Group: group, Version: version}\n\t\t\tresourceConfig.DisableVersions(gv)\n\t\t}\n\t}\n\n\treturn resourceConfig\n}", "func SetupAPI() error {\n\terr := CheckAPI()\n\tif err != nil {\n\t\treturn err\n\t}\n\tflags := other.Flags\n\tfmt.Println(\"setting up the api...\")\n\tif len(flags.WGKey) == 0 {\n\t\treturn errors.New(\"No wargaming api key defined use `./wotnlclans -help` to get more info\")\n\t}\n\tfmt.Println(\"Running api...\")\n\tBuzzy = true\n\tGetIcons()\n\tclanIds := db.GetClanIDs()\n\tif len(clanIds) == 0 || flags.ForceStartupIndexing {\n\t\terr := SearchForClanIds(flags, true)\n\t\tif err != nil {\n\t\t\tfmt.Println(\"ERROR: [SearchForClanIds]:\", err.Error())\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\tGetClanData(clanIds)\n\t}\n\tGetIcons()\n\tBuzzy = false\n\n\tRunSchedule()\n\n\treturn nil\n}", "func NewAPI(ctx context.Context, client client.Client, component *rainbondv1alpha1.RbdComponent, cluster *rainbondv1alpha1.RainbondCluster) ComponentHandler {\n\treturn &api{\n\t\tctx: ctx,\n\t\tclient: client,\n\t\tcomponent: component,\n\t\tcluster: cluster,\n\t\tlabels: LabelsForRainbondComponent(component),\n\t\tpvcName: \"rbd-api\",\n\t\tdataStorageRequest: getStorageRequest(\"API_DATA_STORAGE_REQUEST\", 1),\n\t\tgrdataStorageRequest: getStorageRequest(\"GRDATA_STORAGE_REQUEST\", 20),\n\t}\n}", "func SetupAPI() *gin.Engine {\n\trouter := gin.Default()\n\n\trouter.GET(\"/\", getTodos)\n\trouter.GET(\"/todo/:id\", getTodo)\n\trouter.POST(\"/create\", createTodo)\n\trouter.POST(\"/update\", updateTodo)\n\trouter.GET(\"/delete/:id\", deleteTodo)\n\n\treturn router\n}", "func CreateHTTPAPIHandler(iManager integration.IntegrationManager, cManager clientapi.DevOpsClientManager,\n\tsManager settings.SettingsManager,\n\tsbManager systembanner.SystemBannerManager,\n\ttpManager thirdpartyapi.ThirdPartyManager) (http.Handler, error) {\n\n\tmw := NewLicenseMiddlewareFactory(false)\n\tapiHandler := APIHandler{iManager: iManager, cManager: cManager, sManager: &sManager, groupValidator: mw}\n\twsContainer := restful.NewContainer()\n\twsContainer.EnableContentEncoding(true)\n\n\tapiV1Ws := new(restful.WebService)\n\n\tInstallFilters(apiV1Ws, cManager, mw)\n\n\tapiV1Ws.Path(\"/api/v1\").\n\t\tConsumes(restful.MIME_JSON).\n\t\tProduces(restful.MIME_JSON).\n\t\tParam(restful.HeaderParameter(\"Authorization\", \"Given Bearer token will use this as authorization for the API\"))\n\n\twsContainer.Add(apiV1Ws)\n\n\tapiV2Ws := new(restful.WebService)\n\tapiV2Ws.Path(\"\").\n\t\tConsumes(restful.MIME_JSON).\n\t\tProduces(restful.MIME_JSON).\n\t\tParam(restful.HeaderParameter(\"Authorization\", \"Given Bearer token will use this as authorization for the API\"))\n\twsContainer.Add(apiV2Ws)\n\n\tintegrationHandler := integration.NewIntegrationHandler(iManager)\n\tintegrationHandler.Install(apiV1Ws)\n\tintegrationHandler.Install(apiV2Ws)\n\n\tsettingsHandler := settings.NewSettingsHandler(sManager)\n\tsettingsHandler.Install(apiV1Ws)\n\tsettingsHandler.Install(apiV2Ws)\n\n\tsystemBannerHandler := systembanner.NewSystemBannerHandler(sbManager)\n\tsystemBannerHandler.Install(apiV1Ws)\n\tsystemBannerHandler.Install(apiV2Ws)\n\n\tthirPartyHandler := thirdparty.NewThirdPartyHandler(&sManager, cManager, tpManager)\n\tthirPartyHandler.Install(apiV1Ws)\n\tthirPartyHandler.Install(apiV2Ws)\n\n\tconfigurationHandler := thandler.NewAPIHandler(\"configuration\")\n\tconfigurationHandler.Install(apiV1Ws)\n\tconfigurationHandler.Install(apiV2Ws)\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/namespaces\").\n\t\t\tTo(apiHandler.handleGetNamespaces).\n\t\t\tWrites(ns.NamespaceList{}).\n\t\t\tDoc(\"get namespaces list\").\n\t\t\tReturns(200, \"OK\", ns.NamespaceList{}))\n\n\tapiV2Ws.Route(\n\t\tapiV2Ws.GET(\"/apis/v1/projects/{name}/clusters/{cluster}/namespaces\").\n\t\t\tTo(apiHandler.handleNewGetNamespaces).\n\t\t\tWrites(v1.NamespaceList{}).\n\t\t\tDoc(\"new get project list\").\n\t\t\tReturns(200, \"OK\", v1.NamespaceList{}))\n\n\tapiV2Ws.Route(\n\t\tapiV2Ws.GET(\"/project/v1/projects/{name}/clusters/{cluster}/namespaces\").\n\t\t\tTo(apiHandler.handleNewGetNamespaces).\n\t\t\tWrites(v1.NamespaceList{}).\n\t\t\tDoc(\"new get project list\").\n\t\t\tReturns(200, \"OK\", v1.NamespaceList{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/appdeployment\").\n\t\t\tTo(apiHandler.handleDeploy).\n\t\t\tReads(deployment.AppDeploymentSpec{}).\n\t\t\tWrites(deployment.AppDeploymentSpec{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/configuration\").\n\t\t\tTo(apiHandler.handleGetPlatformConfiguration).\n\t\t\tWrites(configmap.ConfigMapDetail{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/cani\").\n\t\t\tTo(apiHandler.handleCanI).\n\t\t\tReads(authv1.SelfSubjectAccessReviewSpec{}).\n\t\t\tWrites(common.CanIResponse{}).\n\t\t\tDoc(\"Validates access for user\").\n\t\t\tReturns(200, \"OK\", common.CanIResponse{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/caniadmin\").\n\t\t\tTo(apiHandler.handleCanIAdmin).\n\t\t\tWrites(common.CanIResponse{}).\n\t\t\tDoc(\"Validates access for admin user\").\n\t\t\tReturns(200, \"OK\", common.CanIResponse{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/apis\").\n\t\t\tTo(apiHandler.handleGetAPIGroups).\n\t\t\tWrites(metav1.APIGroupList{}).\n\t\t\tDoc(\"Fetches a list of API groups available\").\n\t\t\tReturns(200, \"OK\", metav1.APIGroupList{}))\n\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.POST(\"/appdeployment/validate/name\").\n\t//\t\tTo(apiHandler.handleNameValidity).\n\t//\t\tReads(validation.AppNameAppNameValiditySpecValiditySpec{}).\n\t//\t\tWrites(validation.AppNameValidity{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.POST(\"/appdeployment/validate/imagereference\").\n\t//\t\tTo(apiHandler.handleImageReferenceValidity).\n\t//\t\tReads(validation.ImageReferenceValiditySpec{}).\n\t//\t\tWrites(validation.ImageReferenceValidity{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.POST(\"/appdeployment/validate/protocol\").\n\t//\t\tTo(apiHandler.handleProtocolValidity).\n\t//\t\tReads(validation.ProtocolValiditySpec{}).\n\t//\t\tWrites(validation.ProtocolValidity{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/appdeployment/protocols\").\n\t//\t\tTo(apiHandler.handleGetAvailableProcotols).\n\t//\t\tWrites(deployment.Protocols{}))\n\t//\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.POST(\"/appdeploymentfromfile\").\n\t//\t\tTo(apiHandler.handleDeployFromFile).\n\t//\t\tReads(deployment.AppDeploymentFromFileSpec{}).\n\t//\t\tWrites(deployment.AppDeploymentFromFileResponse{}))\n\t//\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/replicationcontroller\").\n\t//\t\tTo(apiHandler.handleGetReplicationControllerList).\n\t//\t\tWrites(replicationcontroller.ReplicationControllerList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/replicationcontroller/{namespace}\").\n\t//\t\tTo(apiHandler.handleGetReplicationControllerList).\n\t//\t\tWrites(replicationcontroller.ReplicationControllerList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/replicationcontroller/{namespace}/{replicationController}\").\n\t//\t\tTo(apiHandler.handleGetReplicationControllerDetail).\n\t//\t\tWrites(replicationcontroller.ReplicationControllerDetail{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.POST(\"/replicationcontroller/{namespace}/{replicationController}/update/pod\").\n\t//\t\tTo(apiHandler.handleUpdateReplicasCount).\n\t//\t\tReads(replicationcontroller.ReplicationControllerSpec{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/replicationcontroller/{namespace}/{replicationController}/pod\").\n\t//\t\tTo(apiHandler.handleGetReplicationControllerPods).\n\t//\t\tWrites(pod.PodList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/replicationcontroller/{namespace}/{replicationController}/event\").\n\t//\t\tTo(apiHandler.handleGetReplicationControllerEvents).\n\t//\t\tWrites(common.EventList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/replicationcontroller/{namespace}/{replicationController}/service\").\n\t//\t\tTo(apiHandler.handleGetReplicationControllerServices).\n\t//\t\tWrites(resourceService.ServiceList{}))\n\t//\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/workload\").\n\t//\t\tTo(apiHandler.handleGetWorkloads).\n\t//\t\tWrites(workload.Workloads{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/workload/{namespace}\").\n\t//\t\tTo(apiHandler.handleGetWorkloads).\n\t//\t\tWrites(workload.Workloads{}))\n\t//\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/cluster\").\n\t//\t\tTo(apiHandler.handleGetCluster).\n\t//\t\tWrites(cluster.Cluster{}))\n\t//\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/discovery\").\n\t//\t\tTo(apiHandler.handleGetDiscovery).\n\t//\t\tWrites(discovery.Discovery{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/discovery/{namespace}\").\n\t//\t\tTo(apiHandler.handleGetDiscovery).\n\t//\t\tWrites(discovery.Discovery{}))\n\t//\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/config\").\n\t//\t\tTo(apiHandler.handleGetConfig).\n\t//\t\tWrites(config.Config{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/config/{namespace}\").\n\t//\t\tTo(apiHandler.handleGetConfig).\n\t//\t\tWrites(config.Config{}))\n\t//\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/replicaset\").\n\t//\t\tTo(apiHandler.handleGetReplicaSets).\n\t//\t\tWrites(replicaset.ReplicaSetList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/replicaset/{namespace}\").\n\t//\t\tTo(apiHandler.handleGetReplicaSets).\n\t//\t\tWrites(replicaset.ReplicaSetList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/replicaset/{namespace}/{replicaSet}\").\n\t//\t\tTo(apiHandler.handleGetReplicaSetDetail).\n\t//\t\tWrites(replicaset.ReplicaSetDetail{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/replicaset/{namespace}/{replicaSet}/pod\").\n\t//\t\tTo(apiHandler.handleGetReplicaSetPods).\n\t//\t\tWrites(pod.PodList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/replicaset/{namespace}/{replicaSet}/event\").\n\t//\t\tTo(apiHandler.handleGetReplicaSetEvents).\n\t//\t\tWrites(common.EventList{}))\n\t//\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/pod\").\n\t//\t\tTo(apiHandler.handleGetPods).\n\t//\t\tWrites(pod.PodList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/pod/{namespace}\").\n\t//\t\tTo(apiHandler.handleGetPods).\n\t//\t\tWrites(pod.PodList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/pod/{namespace}/{pod}\").\n\t//\t\tTo(apiHandler.handleGetPodDetail).\n\t//\t\tWrites(pod.PodDetail{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/pod/{namespace}/{pod}/container\").\n\t\t\tTo(apiHandler.handleGetPodContainers).\n\t\t\tWrites(pod.PodDetail{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/pod/{namespace}/{pod}/event\").\n\t//\t\tTo(apiHandler.handleGetPodEvents).\n\t//\t\tWrites(common.EventList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/pod/{namespace}/{pod}/shell/{container}\").\n\t\t\tTo(apiHandler.handleExecShell).\n\t\t\tWrites(TerminalResponse{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/pod/{namespace}/{pod}/persistentvolumeclaim\").\n\t//\t\tTo(apiHandler.handleGetPodPersistentVolumeClaims).\n\t//\t\tWrites(persistentvolumeclaim.PersistentVolumeClaimList{}))\n\t//\n\n\t// region Deployment\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/deployment\").\n\t\t\tTo(apiHandler.handleGetDeployments).\n\t\t\tWrites(deployment.DeploymentList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/deployment/{namespace}\").\n\t\t\tTo(apiHandler.handleGetDeployments).\n\t\t\tWrites(deployment.DeploymentList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/deployment/{namespace}/{deployment}\").\n\t\t\tTo(apiHandler.handleGetDeploymentDetail).\n\t\t\tWrites(deployment.DeploymentDetail{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/deployment/{namespace}/{deployment}\").\n\t\t\tTo(apiHandler.handleUpdateDeploymentDetail).\n\t\t\tWrites(appsv1.Deployment{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/deployment/{namespace}/{deployment}/start\").\n\t\t\tTo(apiHandler.handleStartStopDeployment).\n\t\t\tDoc(\"start deployment\").\n\t\t\tReturns(http.StatusNoContent, \"OK\", struct{}{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/deployment/{namespace}/{deployment}/stop\").\n\t\t\tTo(apiHandler.handleStartStopDeployment).\n\t\t\tDoc(\"stop deployment\").\n\t\t\tReturns(http.StatusNoContent, \"OK\", struct{}{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/deployment/{namespace}/{deployment}/yaml\").\n\t\t\tTo(apiHandler.handleUpdateDeploymentDetailYaml).\n\t\t\tWrites(appsv1.Deployment{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/deployment/{namespace}/{deployment}/replicas\").\n\t\t\tTo(apiHandler.handleUpdateDeploymentReplicas).\n\t\t\tWrites(deployment.DeploymentReplica{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/deployment/{namespace}/{deployment}/network\").\n\t\t\tTo(apiHandler.handleUpdateDeploymentNetwork).\n\t\t\tWrites(appsv1.Deployment{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/deployment/{namespace}/{deployment}/container/{container}/\").\n\t\t\tTo(apiHandler.handlePutDeploymentContainer).\n\t\t\tWrites(appsv1.Deployment{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/deployment/{namespace}/{deployment}/container/{container}/image\").\n\t\t\tTo(apiHandler.handleUpdateDeploymentContainerImage).\n\t\t\tWrites(appsv1.Deployment{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/deployment/{namespace}/{deployment}/container/{container}/resources\").\n\t\t\tTo(apiHandler.handleUpdateDeploymentContainerResources).\n\t\t\tWrites(appsv1.Deployment{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/deployment/{namespace}/{deployment}/container/{container}/env\").\n\t\t\tTo(apiHandler.handleUpdateDeploymentContainerEnv).\n\t\t\tWrites(appsv1.Deployment{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/deployment/{namespace}/{deployment}/container/{container}/volumeMount/\").\n\t\t\tTo(apiHandler.handleCreateDeploymentVolumeMount).\n\t\t\tWrites(appsv1.Deployment{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/deployment/{namespace}/{deployment}/event\").\n\t\t\tTo(apiHandler.handleGetDeploymentEvents).\n\t\t\tWrites(common.EventList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/deployment/{namespace}/{deployment}/pods\").\n\t\t\tTo(apiHandler.handleGetDeploymentPods).\n\t\t\tWrites(pod.PodList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/deployment/{namespace}/{deployment}/oldreplicaset\").\n\t\t\tTo(apiHandler.handleGetDeploymentOldReplicaSets).\n\t\t\tWrites(replicaset.ReplicaSetList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/deployment/{namespace}/{deployment}/actions/rollback\").\n\t\t\tTo(apiHandler.handleRollBackDeploymentToRevision).\n\t\t\tReads(common.RevisionDetail{}).\n\t\t\tWrites(appsv1.Deployment{}).\n\t\t\tDoc(\"rollback deployment to special revision\").\n\t\t\tReturns(200, \"OK\", appsv1.Deployment{}))\n\n\t// endregion\n\n\t// region Scale\n\t//\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.PUT(\"/scale/{kind}/{namespace}/{name}/\").\n\t//\t\tTo(apiHandler.handleScaleResource).\n\t//\t\tWrites(scaling.ReplicaCounts{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/scale/{kind}/{namespace}/{name}\").\n\t//\t\tTo(apiHandler.handleGetReplicaCount).\n\t//\t\tWrites(scaling.ReplicaCounts{}))\n\t// endregion\n\n\t// region Deamonset\n\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/daemonset\").\n\t//\t\tTo(apiHandler.handleGetDaemonSetList).\n\t//\t\tWrites(daemonset.DaemonSetList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/daemonset/{namespace}\").\n\t//\t\tTo(apiHandler.handleGetDaemonSetList).\n\t//\t\tWrites(daemonset.DaemonSetList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/daemonset/{namespace}/{daemonset}\").\n\t\t\tTo(apiHandler.handleGetDaemonSetDetail).\n\t\t\tWrites(appsv1.DaemonSet{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/daemonset/{namespace}/{daemonset}\").\n\t\t\tTo(apiHandler.handleUpdateDaemonSetDetail).\n\t\t\tWrites(appsv1.DaemonSet{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/daemonset/{namespace}/{daemonset}/yaml\").\n\t\t\tTo(apiHandler.handleUpdateDaemonSetDetail).\n\t\t\tWrites(appsv1.Deployment{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/daemonset/{namespace}/{daemonset}/pods\").\n\t\t\tTo(apiHandler.handleGetDaemonSetPods).\n\t\t\tWrites(pod.PodList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/daemonset/{namespace}/{daemonset}/container/{container}/\").\n\t\t\tTo(apiHandler.handlePutDaemonSetContainer).\n\t\t\tWrites(appsv1.DaemonSet{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/daemonset/{namespace}/{daemonset}/container/{container}/image\").\n\t\t\tTo(apiHandler.handleUpdateDaemonSetContainerImage).\n\t\t\tWrites(appsv1.DaemonSet{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/daemonset/{namespace}/{daemonset}/container/{container}/env\").\n\t\t\tTo(apiHandler.handleUpdateDaemonSetContainerEnv).\n\t\t\tWrites(appsv1.DaemonSet{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/daemonset/{namespace}/{daemonset}/container/{container}/resources\").\n\t\t\tTo(apiHandler.handleUpdateDaemonSetContainerResource).\n\t\t\tWrites(appsv1.DaemonSet{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/daemonset/{namespace}/{daemonset}/container/{container}/volumeMount/\").\n\t\t\tTo(apiHandler.handleCreateDaemonSetVolumeMount).\n\t\t\tWrites(appsv1.DaemonSet{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/daemonset/{namespace}/{daemonSet}/service\").\n\t//\t\tTo(apiHandler.handleGetDaemonSetServices).\n\t//\t\tWrites(resourceService.ServiceList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/daemonset/{namespace}/{daemonSet}/event\").\n\t//\t\tTo(apiHandler.handleGetDaemonSetEvents).\n\t//\t\tWrites(common.EventList{}))\n\n\t// endregion\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/horizontalpodautoscaler\").\n\t\t\tTo(apiHandler.handleGetHorizontalPodAutoscalerList).\n\t\t\tWrites(horizontalpodautoscaler.HorizontalPodAutoscalerList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/horizontalpodautoscaler/{namespace}\").\n\t\t\tTo(apiHandler.handleGetHorizontalPodAutoscalerList).\n\t\t\tWrites(horizontalpodautoscaler.HorizontalPodAutoscalerList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/horizontalpodautoscaler/{namespace}/{horizontalpodautoscaler}\").\n\t\t\tTo(apiHandler.handleGetHorizontalPodAutoscalerDetail).\n\t\t\tWrites(horizontalpodautoscaler.HorizontalPodAutoscalerDetail{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/horizontalpodautoscaler/{namespace}\").\n\t\t\tTo(apiHandler.handleCreateHorizontalPodAutoscaler).\n\t\t\tReads(horizontalpodautoscaler.HorizontalPodAutoscalerDetail{}).\n\t\t\tWrites(horizontalpodautoscaler.HorizontalPodAutoscalerDetail{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/horizontalpodautoscaler/{namespace}/{horizontalpodautoscaler}\").\n\t\t\tTo(apiHandler.handleUpdateHorizontalPodAutoscaler).\n\t\t\tWrites(horizontalpodautoscaler.HorizontalPodAutoscalerDetail{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/horizontalpodautoscaler/{namespace}/{horizontalpodautoscaler}\").\n\t\t\tTo(apiHandler.handleDeleteHorizontalPodAutoscaler).\n\t\t\tWrites(horizontalpodautoscaler.HorizontalPodAutoscalerDetail{}))\n\t//\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/job\").\n\t//\t\tTo(apiHandler.handleGetJobList).\n\t//\t\tWrites(job.JobList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/job/{namespace}\").\n\t//\t\tTo(apiHandler.handleGetJobList).\n\t//\t\tWrites(job.JobList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/job/{namespace}/{name}\").\n\t//\t\tTo(apiHandler.handleGetJobDetail).\n\t//\t\tWrites(job.JobDetail{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/job/{namespace}/{name}/pod\").\n\t//\t\tTo(apiHandler.handleGetJobPods).\n\t//\t\tWrites(pod.PodList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/job/{namespace}/{name}/event\").\n\t//\t\tTo(apiHandler.handleGetJobEvents).\n\t//\t\tWrites(common.EventList{}))\n\t//\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/cronjob\").\n\t//\t\tTo(apiHandler.handleGetCronJobList).\n\t//\t\tWrites(cronjob.CronJobList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/cronjob/{namespace}\").\n\t//\t\tTo(apiHandler.handleGetCronJobList).\n\t//\t\tWrites(cronjob.CronJobList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/cronjob/{namespace}/{name}\").\n\t//\t\tTo(apiHandler.handleGetCronJobDetail).\n\t//\t\tWrites(cronjob.CronJobDetail{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/cronjob/{namespace}/{name}/job\").\n\t//\t\tTo(apiHandler.handleGetCronJobJobs).\n\t//\t\tWrites(job.JobList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/cronjob/{namespace}/{name}/event\").\n\t//\t\tTo(apiHandler.handleGetCronJobEvents).\n\t//\t\tWrites(common.EventList{}))\n\t//\n\n\t// region Namespace\n\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.POST(\"/namespace\").\n\t//\t\tTo(apiHandler.handleCreateNamespace).\n\t//\t\tReads(ns.NamespaceSpec{}).\n\t//\t\tWrites(ns.NamespaceSpec{}))\n\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/namespace/{name}\").\n\t//\t\tTo(apiHandler.handleGetNamespaceDetail).\n\t//\t\tWrites(ns.NamespaceDetail{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/namespace/{name}/event\").\n\t//\t\tTo(apiHandler.handleGetNamespaceEvents).\n\t//\t\tWrites(common.EventList{}))\n\t//\n\t// endregion\n\n\t// region Secret\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/secret\").\n\t\t\tTo(apiHandler.handleGetSecretList).\n\t\t\tWrites(secret.SecretList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/secret/{namespace}\").\n\t\t\tTo(apiHandler.handleGetSecretList).\n\t\t\tWrites(secret.SecretList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/secret/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleGetSecretDetail).\n\t\t\tWrites(secret.SecretDetail{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/secret/{name}/resources\").\n\t\t\tTo(apiHandler.handleGetSecretRelatedResources).\n\t\t\tWrites(secret.ResourceList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/secret/{namespace}/{name}/resources\").\n\t\t\tTo(apiHandler.handleGetSecretRelatedResources).\n\t\t\tWrites(secret.ResourceList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/secret/{name}\").\n\t\t\tTo(apiHandler.handleUpdateSecret).\n\t\t\tWrites(secret.SecretDetail{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/secret/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleUpdateSecret).\n\t\t\tWrites(secret.SecretDetail{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/secret\").\n\t\t\tTo(apiHandler.handleCreateSecret).\n\t\t\tReads(secret.SecretDetail{}).\n\t\t\tWrites(secret.Secret{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/secret/{namespace}\").\n\t\t\tTo(apiHandler.handleCreateSecret).\n\t\t\tReads(secret.SecretDetail{}).\n\t\t\tWrites(secret.Secret{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/secret/{name}\").\n\t\t\tTo(apiHandler.handleDeleteSecret).\n\t\t\tWrites(secret.SecretDetail{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/secret/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleDeleteSecret).\n\t\t\tWrites(secret.SecretDetail{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/secret/{namespace}/{name}/actions/tradeapp\").\n\t\t\tTo(apiHandler.handleUpdateSecretBelongApp).\n\t\t\tReads(common.AppNameDetail{}).\n\t\t\tWrites(secret.SecretDetail{}).\n\t\t\tDoc(\"update secret belongs app\").\n\t\t\tReturns(200, \"OK\", secret.SecretDetail{}))\n\t// endregion\n\n\t// region Configmap\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/configmap\").\n\t\t\tTo(apiHandler.handleGetConfigMapList).\n\t\t\tWrites(configmap.ConfigMapList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/configmap/{namespace}\").\n\t\t\tTo(apiHandler.handleGetConfigMapList).\n\t\t\tWrites(configmap.ConfigMapList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/configmap/{namespace}/{configmap}\").\n\t\t\tTo(apiHandler.handleGetConfigMapDetail).\n\t\t\tWrites(configmap.ConfigMapDetail{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/configmap/{namespace}\").\n\t\t\tTo(apiHandler.handleCreateConfigMap).\n\t\t\tReads(configmap.ConfigMapDetail{}).\n\t\t\tWrites(configmap.ConfigMapDetail{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/configmap/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleUpdateConfigMap).\n\t\t\tWrites(configmap.ConfigMapDetail{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/configmap/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleDeleteConfigMap).\n\t\t\tWrites(configmap.ConfigMapDetail{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/configmap/{namespace}/{name}/actions/tradeapp\").\n\t\t\tTo(apiHandler.handleUpdateConfigMapBelongApp).\n\t\t\tReads(common.AppNameDetail{}).\n\t\t\tWrites(configmap.ConfigMapDetail{}).\n\t\t\tDoc(\"update configmap belongs app\").\n\t\t\tReturns(200, \"OK\", configmap.ConfigMapDetail{}))\n\t// endregion\n\n\t//\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/service\").\n\t//\t\tTo(apiHandler.handleGetServiceList).\n\t//\t\tWrites(resourceService.ServiceList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/service/{namespace}\").\n\t//\t\tTo(apiHandler.handleGetServiceList).\n\t//\t\tWrites(resourceService.ServiceList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/service/{namespace}/{service}\").\n\t\t\tTo(apiHandler.handleGetServiceDetail))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/service/{namespace}/{service}/pod\").\n\t//\t\tTo(apiHandler.handleGetServicePods).\n\t//\t\tWrites(pod.PodList{}))\n\t//\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/ingress\").\n\t//\t\tTo(apiHandler.handleGetIngressList).\n\t//\t\tWrites(ingress.IngressList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/ingress/{namespace}\").\n\t//\t\tTo(apiHandler.handleGetIngressList).\n\t//\t\tWrites(ingress.IngressList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/ingress/{namespace}/{name}\").\n\t//\t\tTo(apiHandler.handleGetIngressDetail).\n\t//\t\tWrites(ingress.IngressDetail{}))\n\t//\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/statefulset\").\n\t//\t\tTo(apiHandler.handleGetStatefulSetList).\n\t//\t\tWrites(statefulset.StatefulSetList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/statefulset/{namespace}\").\n\t//\t\tTo(apiHandler.handleGetStatefulSetList).\n\t//\t\tWrites(statefulset.StatefulSetList{}))\n\n\t// region Statefulset\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/statefulset/{namespace}/{statefulset}\").\n\t\t\tTo(apiHandler.handleGetStatefulSetDetail).\n\t\t\tWrites(appsv1.StatefulSet{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/statefulset/{namespace}/{statefulset}\").\n\t\t\tTo(apiHandler.handleUpdateStatefulSetDetail).\n\t\t\tWrites(appsv1.StatefulSet{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/statefulset/{namespace}/{statefulset}/start\").\n\t\t\tTo(apiHandler.handleStartStopStatefulSet).\n\t\t\tDoc(\"start statefulset\").\n\t\t\tReturns(http.StatusNoContent, \"OK\", struct{}{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/statefulset/{namespace}/{statefulset}/stop\").\n\t\t\tTo(apiHandler.handleStartStopStatefulSet).\n\t\t\tDoc(\"stop statefulset\").\n\t\t\tReturns(http.StatusNoContent, \"OK\", struct{}{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/statefulset/{namespace}/{statefulset}/yaml\").\n\t\t\tTo(apiHandler.handleUpdateStatefulSetDetail).\n\t\t\tWrites(appsv1.StatefulSet{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/statefulset/{namespace}/{statefulset}/replicas\").\n\t\t\tTo(apiHandler.handleUpdateStatefulSetReplicas).\n\t\t\tWrites(appsv1.StatefulSet{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/statefulset/{namespace}/{statefulset}/pods\").\n\t\t\tTo(apiHandler.handleGetStatefulSetPods).\n\t\t\tWrites(pod.PodList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/statefulset/{namespace}/{statefulset}/container/{container}/\").\n\t\t\tTo(apiHandler.handlePutStatefulSetContainer).\n\t\t\tWrites(appsv1.StatefulSet{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/statefulset/{namespace}/{statefulset}/container/{container}/image\").\n\t\t\tTo(apiHandler.handleUpdateStatefulSetContainerImage).\n\t\t\tWrites(appsv1.StatefulSet{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/statefulset/{namespace}/{statefulset}/container/{container}/env\").\n\t\t\tTo(apiHandler.handleUpdateStatefulSetContainerEnv).\n\t\t\tWrites(appsv1.StatefulSet{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/statefulset/{namespace}/{statefulset}/container/{container}/resources\").\n\t\t\tTo(apiHandler.handleUpdateStatefulSetContainerResource).\n\t\t\tWrites(appsv1.StatefulSet{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/statefulset/{namespace}/{statefulset}/container/{container}/volumeMount/\").\n\t\t\tTo(apiHandler.handleCreateStatefulSetVolumeMount).\n\t\t\tWrites(appsv1.StatefulSet{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/statefulset/{namespace}/{statefulset}/event\").\n\t//\t\tTo(apiHandler.handleGetStatefulSetEvents).\n\t//\t\tWrites(common.EventList{}))\n\n\t// endregion\n\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/node\").\n\t//\t\tTo(apiHandler.handleGetNodeList).\n\t//\t\tWrites(node.NodeList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/node/{name}\").\n\t//\t\tTo(apiHandler.handleGetNodeDetail).\n\t//\t\tWrites(node.NodeDetail{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/node/{name}/event\").\n\t//\t\tTo(apiHandler.handleGetNodeEvents).\n\t//\t\tWrites(common.EventList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/node/{name}/pod\").\n\t//\t\tTo(apiHandler.handleGetNodePods).\n\t//\t\tWrites(pod.PodList{}))\n\t//\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.DELETE(\"/_raw/{kind}/namespace/{namespace}/name/{name}\").\n\t//\t\tTo(apiHandler.handleDeleteResource))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/_raw/{kind}/namespace/{namespace}/name/{name}\").\n\t//\t\tTo(apiHandler.handleGetResource))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.PUT(\"/_raw/{kind}/namespace/{namespace}/name/{name}\").\n\t//\t\tTo(apiHandler.handlePutResource))\n\t//\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.DELETE(\"/_raw/{kind}/name/{name}\").\n\t//\t\tTo(apiHandler.handleDeleteResource))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/_raw/{kind}/name/{name}\").\n\t//\t\tTo(apiHandler.handleGetResource))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.PUT(\"/_raw/{kind}/name/{name}\").\n\t//\t\tTo(apiHandler.handlePutResource))\n\t//\n\n\t// region RBAC\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/rbac/role\").\n\t\t\tTo(apiHandler.handleGetRbacRoleList).\n\t\t\tWrites(rbacroles.RbacRoleList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/rbac/rolebinding\").\n\t\t\tTo(apiHandler.handleGetRbacRoleBindingList).\n\t\t\tWrites(rbacrolebindings.RbacRoleBindingList{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/rolebinding/{namespace}\").\n\t\t\tTo(apiHandler.handleListRoleBindingsOriginal).\n\t\t\tWrites(rolebinding.RoleBindingList{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/rolebinding/{namespace}\").\n\t\t\tTo(apiHandler.handleCreateRoleBinding).\n\t\t\tDoc(\"creates a rolebinding\").\n\t\t\tWrites(rbacv1.RoleBinding{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/rolebinding/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleDeleteRoleBindingsOriginal).\n\t\t\tDoc(\"delete a rolebinding\").\n\t\t\tWrites(rbacv1.RoleBinding{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/rbac/status\").\n\t//\t\tTo(apiHandler.handleRbacStatus).\n\t//\t\tWrites(validation.RbacStatus{}))\n\n\t// endregion\n\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/persistentvolume\").\n\t//\t\tTo(apiHandler.handleGetPersistentVolumeList).\n\t//\t\tWrites(persistentvolume.PersistentVolumeList{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/persistentvolume/{persistentvolume}\").\n\t//\t\tTo(apiHandler.handleGetPersistentVolumeDetail).\n\t//\t\tWrites(persistentvolume.PersistentVolumeDetail{}))\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/persistentvolume/namespace/{namespace}/name/{persistentvolume}\").\n\t//\t\tTo(apiHandler.handleGetPersistentVolumeDetail).\n\t//\t\tWrites(persistentvolume.PersistentVolumeDetail{}))\n\t//\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/persistentvolumeclaim/\").\n\t//\t\tTo(apiHandler.handleGetPersistentVolumeClaimList).\n\t//\t\tWrites(persistentvolumeclaim.PersistentVolumeClaimList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/persistentvolumeclaim/{namespace}\").\n\t\t\tTo(apiHandler.handleGetPersistentVolumeClaimList).\n\t\t\tWrites(persistentvolumeclaim.PersistentVolumeClaimList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/persistentvolumeclaim/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleGetPersistentVolumeClaimDetail).\n\t\t\tWrites(persistentvolumeclaim.PersistentVolumeClaimDetail{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/persistentvolumeclaim/{namespace}\").\n\t\t\tTo(apiHandler.handleCreatePersistentVolumeClaim).\n\t\t\tReads(persistentvolumeclaim.PersistentVolumeClaimDetail{}).\n\t\t\tWrites(persistentvolumeclaim.PersistentVolumeClaimDetail{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/persistentvolumeclaim/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleUpdatePersistentVolumeClaim).\n\t\t\tWrites(persistentvolumeclaim.PersistentVolumeClaimDetail{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/persistentvolumeclaim/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleDeletePersistentVolumeClaim).\n\t\t\tWrites(persistentvolumeclaim.PersistentVolumeClaimDetail{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/persistentvolumeclaim/{namespace}/{name}/actions/tradeapp\").\n\t\t\tTo(apiHandler.handleUpdatePersistentVolumeClaimBelongApp).\n\t\t\tReads(common.AppNameDetail{}).\n\t\t\tWrites(persistentvolumeclaim.PersistentVolumeClaimDetail{}).\n\t\t\tDoc(\"update persistentvolumeclaim belongs app\").\n\t\t\tReturns(200, \"OK\", persistentvolumeclaim.PersistentVolumeClaimDetail{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/storageclass\").\n\t\t\tTo(apiHandler.handleGetStorageClassList).\n\t\t\tWrites(storageclass.StorageClassList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/storageclass/{storageclass}\").\n\t\t\tTo(apiHandler.handleGetStorageClass).\n\t\t\tWrites(storageclass.StorageClass{}))\n\n\t// apiV1Ws.Route(\n\t// \tapiV1Ws.GET(\"/storageclass/{storageclass}/persistentvolume\").\n\t// \t\tTo(apiHandler.handleGetStorageClassPersistentVolumes).\n\t// \t\tWrites(persistentvolume.PersistentVolumeList{}))\n\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/log/source/{namespace}/{resourceName}/{resourceType}\").\n\t//\t\tTo(apiHandler.handleLogSource).\n\t//\t\tWrites(controller.LogSources{}))\n\n\t// region log\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/log/{namespace}/{pod}\").\n\t\t\tTo(apiHandler.handleLogs).\n\t\t\tWrites(logs.LogDetails{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/log/{namespace}/{pod}/{container}\").\n\t\t\tTo(apiHandler.handleLogs).\n\t\t\tWrites(logs.LogDetails{}))\n\t//\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/log/file/{namespace}/{pod}/{container}\").\n\t\t\tTo(apiHandler.handleLogFile).\n\t\t\tWrites(logs.LogDetails{}))\n\t// endregion\n\n\t//\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/overview/\").\n\t//\t\tTo(apiHandler.handleOverview).\n\t//\t\tWrites(overview.Overview{}))\n\t//\n\t//apiV1Ws.Route(\n\t//\tapiV1Ws.GET(\"/overview/{namespace}\").\n\t//\t\tTo(apiHandler.handleOverview).\n\t//\t\tWrites(overview.Overview{}))\n\t//\n\n\t// region others\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/others\").\n\t\t\tTo(apiHandler.handleOtherResourcesList).\n\t\t\tWrites(other.ResourceList{}).\n\t\t\tDoc(\"get all resources\").\n\t\t\tParam(restful.QueryParameter(\"filterBy\", \"filter option separated by comma. For example parameter1,value1,parameter2,value2 - means that the data should be filtered by parameter1 equals value1 and parameter2 equals value2\").\n\t\t\t\tDataType(\"string\").\n\t\t\t\tAllowableValues(map[string]string{\n\t\t\t\t\t\"name\": \"search by name partial match\",\n\t\t\t\t\t\"namespace\": \"filter by namespace\",\n\t\t\t\t\t\"kind\": \"filter by kind\",\n\t\t\t\t\t\"scope\": \"allowed value `namespaced` and `clustered` filter by if a resource is namespaced\",\n\t\t\t\t})).\n\t\t\tParam(restful.QueryParameter(\"sortBy\", \"sort option separated by comma. For example a,parameter1,d,parameter2 - means that the data should be sorted by parameter1 (ascending) and later sort by parameter2 (descending)\").\n\t\t\t\tDataType(\"string\").\n\t\t\t\tAllowableValues(map[string]string{\n\t\t\t\t\t\"name\": \"\",\n\t\t\t\t\t\"namespace\": \"\",\n\t\t\t\t\t\"kind\": \"\",\n\t\t\t\t\t\"creationTimestamp\": \"\",\n\t\t\t\t})).\n\t\t\tParam(restful.QueryParameter(\"itemsPerPage\", \"items per page\").\n\t\t\t\tDataType(\"integer\")).\n\t\t\tParam(restful.QueryParameter(\"page\", \"page number\").DataType(\"integer\")).\n\t\t\tReturns(200, \"OK\", other.ResourceList{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/others\").\n\t\t\tTo(apiHandler.handleOtherResourceCreate).\n\t\t\tDoc(\"create a resource\").\n\t\t\tReads([]unstructured.Unstructured{}).\n\t\t\tConsumes(restful.MIME_JSON).\n\t\t\tReturns(200, \"OK\", CreateResponse{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/releases\").\n\t\t\tTo(apiHandler.handleReleaseCreate).\n\t\t\tDoc(\"create a release\").\n\t\t\tReads([]unstructured.Unstructured{}).\n\t\t\tConsumes(restful.MIME_JSON).\n\t\t\tReturns(200, \"OK\", []unstructured.Unstructured{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/releases/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleGetReleaseDetail).\n\t\t\tDoc(\"get a release\").\n\t\t\tReads(release.ReleaseDetails{}).\n\t\t\tConsumes(restful.MIME_JSON).\n\t\t\tReturns(200, \"OK\", release.ReleaseDetails{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/others/{group}/{version}/{kind}/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleOtherResourceDetail).\n\t\t\tWrites(other.OtherResourceDetail{}).\n\t\t\tDoc(\"get a resource detail with events\").\n\t\t\tReturns(200, \"OK\", other.OtherResourceDetail{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/others/{group}/{version}/{kind}/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleOtherResourceDetail).\n\t\t\tDoc(\"delete a resource\"))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/others/{group}/{version}/{kind}/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleOtherResourceDetail).\n\t\t\tDoc(\"update a resource with whole resource json\").\n\t\t\tReads(unstructured.Unstructured{}).\n\t\t\tConsumes(restful.MIME_JSON))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PATCH(\"/others/{group}/{version}/{kind}/{namespace}/{name}/{field}\").\n\t\t\tTo(apiHandler.handleOtherResourcePatch).\n\t\t\tDoc(\"update resource annotations or labels\").\n\t\t\tReads(other.FieldPayload{}).\n\t\t\tConsumes(restful.MIME_JSON))\n\t// endregion\n\n\t// ---- DEVOPS APIS ----\n\n\t// region Jenkins\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/jenkinses\").\n\t\t\tTo(apiHandler.handleGetJenkins).\n\t\t\tWrites(jenkins.JenkinsList{}).\n\t\t\tDoc(\"get jenkins list\").\n\t\t\tReturns(200, \"OK\", jenkins.JenkinsList{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/jenkinses/{name}\").\n\t\t\tTo(apiHandler.handleRetriveJenkins).\n\t\t\tWrites(v1alpha1.Jenkins{}).\n\t\t\tDoc(\"retrieve jenkins config\").\n\t\t\tReturns(200, \"OK\", v1alpha1.Jenkins{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/jenkinses/{name}/resources\").\n\t\t\tTo(apiHandler.handleGetJenkinsResources).\n\t\t\tWrites(common.ResourceList{}).\n\t\t\tDoc(\"retrieve resources associated with jenkins\").\n\t\t\tReturns(200, \"OK\", common.ResourceList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/jenkinses/{name}\").\n\t\t\tTo(apiHandler.handleDeleteJenkins).\n\t\t\tWrites(jenkins.Jenkins{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/jenkinses/{name}\").\n\t\t\tTo(apiHandler.handlePutJenkins).\n\t\t\tWrites(v1alpha1.Jenkins{}).\n\t\t\tDoc(\"update jenkins config\").\n\t\t\tReturns(200, \"OK\", v1alpha1.Jenkins{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/jenkinses\").\n\t\t\tTo(apiHandler.handleCreateJenkins).\n\t\t\tWrites(v1alpha1.Jenkins{}).\n\t\t\tDoc(\"update jenkins config\").\n\t\t\tReturns(200, \"OK\", v1alpha1.Jenkins{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/jenkinsbinding\").\n\t\t\tTo(apiHandler.handleGetJenkinsBindingList).\n\t\t\tWrites(jenkinsbinding.JenkinsBindingList{}).\n\t\t\tDoc(\"get jenkinsbinding list\").\n\t\t\tReturns(200, \"OK\", jenkinsbinding.JenkinsBindingList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/jenkinsbinding/{namespace}\").\n\t\t\tTo(apiHandler.handleGetJenkinsBindingList).\n\t\t\tWrites(jenkinsbinding.JenkinsBindingList{}).\n\t\t\tDoc(\"get namespaced jenkinsbinding list\").\n\t\t\tReturns(200, \"OK\", jenkinsbinding.JenkinsBindingList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/jenkinsbinding/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleGetJenkinsBinding).\n\t\t\tDoc(\"get jenkinsbinding details\").\n\t\t\tWrites(v1alpha1.JenkinsBinding{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/jenkinsbinding/{namespace}/{name}/croncheck\").\n\t\t\tTo(apiHandler.handleCronCheck).\n\t\t\tDoc(\"cron syntax check\").\n\t\t\tWrites(jenkinsbinding.CronCheckResult{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/jenkinsbinding/{namespace}/{name}/resources\").\n\t\t\tTo(apiHandler.handleGetJenkinsBindingResources).\n\t\t\tWrites(common.ResourceList{}).\n\t\t\tDoc(\"retrieve resources associated with jenkinsbinding\").\n\t\t\tReturns(200, \"OK\", common.ResourceList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/jenkinsbinding/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleDeleteJenkinsBinding).\n\t\t\tWrites(v1alpha1.JenkinsBinding{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/jenkinsbinding/{namespace}\").\n\t\t\tTo(apiHandler.handleCreateJenkinsBinding).\n\t\t\tWrites(v1alpha1.JenkinsBinding{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/jenkinsbinding/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleUpdateJenkinsBinding).\n\t\t\tWrites(v1alpha1.JenkinsBinding{}))\n\t// endregion\n\n\t//domain\n\t// region DomainBinding\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/domainbinding\").\n\t\t\tTo(apiHandler.handleGetDomainBindingList).\n\t\t\tWrites(domainbinding.DomainBindingList{}).\n\t\t\tDoc(\"get domianbinding list\"))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/domainbinding\").\n\t\t\tTo(apiHandler.handleCreateDomainBinding).\n\t\t\tWrites(domainbinding.DomainBindingDetail{}).\n\t\t\tDoc(\"create domainbinding\"))\n\tdomainBindDetailURI := \"/domainbinding/{name}\"\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(domainBindDetailURI).\n\t\t\tTo(apiHandler.handleGetDomainBindingDetail).\n\t\t\tWrites(domainbinding.DomainBindingDetail{}).\n\t\t\tDoc(\"get domainbinding detail\"))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(domainBindDetailURI).\n\t\t\tTo(apiHandler.handleUpdateDomainBindingDetail).\n\t\t\tWrites(domainbinding.DomainBindingDetail{}).\n\t\t\tDoc(\"update domainbinding detail\"))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(domainBindDetailURI).\n\t\t\tTo(apiHandler.handleDeleteDomainBindingDetail).\n\t\t\tDoc(\"delete domainbinding detailt\"))\n\t// endregion\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/chart/{name}\").\n\t\t\tTo(apiHandler.handleGetChartDetail).\n\t\t\tWrites(catalog.Chart{}).\n\t\t\tDoc(\"get chart detail\"))\n\n\t// region PipelineTemplate\n\t// PipelineTemplateSync\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/pipelinetemplatesync/{namespace}\").\n\t\t\tTo(apiHandler.handleGetPipelineTemplateSyncList).\n\t\t\tWrites(pipelinetemplatesync.PipelineTemplateSyncList{}).\n\t\t\tDoc(\"get pipelineTemplateSync list\").\n\t\t\tReturns(200, \"OK\", pipelinetemplatesync.PipelineTemplateSyncList{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/pipelinetemplatesync/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleGetPipelineTemplateSync).\n\t\t\tWrites(pipelinetemplatesync.PipelineTemplateSync{}).\n\t\t\tDoc(\"get detail of specific PipelineTemplateSync\").\n\t\t\tReturns(200, \"OK\", pipelinetemplatesync.PipelineTemplateSync{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/pipelinetemplatesync/{namespace}\").\n\t\t\tTo(apiHandler.handleCreatePipelineTemplateSync).\n\t\t\tWrites(pipelinetemplatesync.PipelineTemplateSync{}).\n\t\t\tDoc(\"create a pipelineTemplateSync\").\n\t\t\tReturns(200, \"OK\", pipelinetemplatesync.PipelineTemplateSync{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/pipelinetemplatesync/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleUpdatePipelineTemplateSync).\n\t\t\tWrites(pipelinetemplatesync.PipelineTemplateSync{}).\n\t\t\tDoc(\"update a pipelineTemplateSync\").\n\t\t\tReturns(200, \"OK\", pipelinetemplatesync.PipelineTemplateSync{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/pipelinetemplatesync/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleDeletePipelineTemplateSync).\n\t\t\tWrites(struct{}{}).\n\t\t\tDoc(\"delete a PipelineTemplateSync\").\n\t\t\tReturns(200, \"OK\", struct{}{}))\n\n\t// PipelineTaskTemplate\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/pipelinetasktemplate/{namespace}\").\n\t\t\tTo(apiHandler.handleGetPipelineTaskTemplateList).\n\t\t\tWrites(pipelinetasktemplate.PipelineTaskTemplateList{}).\n\t\t\tDoc(\"get a list of PipelineTaskTemplate\").\n\t\t\tReturns(200, \"OK\", pipelinetasktemplate.PipelineTaskTemplate{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/pipelinetasktemplate/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleGetPipelineTaskTemplate).\n\t\t\tWrites(pipelinetasktemplate.PipelineTaskTemplate{}).\n\t\t\tDoc(\"get a PipelineTaskTemplate\").\n\t\t\tReturns(200, \"OK\", pipelinetasktemplate.PipelineTaskTemplate{}))\n\n\t// ClusterPipelineTemplate\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/clusterpipelinetemplate\").\n\t\t\tTo(apiHandler.handleGetClusterPipelineTemplateList).\n\t\t\tWrites(clusterpipelinetemplate.ClusterPipelineTemplateList{}).\n\t\t\tDoc(\"get a list of ClusterPipelineTemplate\").\n\t\t\tReturns(200, \"OK\", clusterpipelinetemplate.ClusterPipelineTemplateList{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/clusterpipelinetemplate/{name}\").\n\t\t\tTo(apiHandler.handleGetClusterPipelineTemplate).\n\t\t\tWrites(clusterpipelinetemplate.ClusterPipelineTemplate{}).\n\t\t\tDoc(\"get a ClusterPipelineTemplate\").\n\t\t\tReturns(200, \"OK\", clusterpipelinetemplate.ClusterPipelineTemplate{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/clusterpipelinetemplate/{name}/preview\").\n\t\t\tTo(apiHandler.handlePreviewClusterPipelineTemplate).\n\t\t\tWrites(clusterpipelinetemplate.PreviewOptions{}).\n\t\t\tDoc(\"preview a ClusterPipelineTemplate\").\n\t\t\tReturns(200, \"OK\", \"\"))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/clusterpipelinetemplate/{name}/exports\").\n\t\t\tTo(apiHandler.handlerExportsClusterPiplineTemplate).\n\t\t\tWrites(clusterpipelinetemplate.PipelineExportedVariables{}).\n\t\t\tDoc(\"get the exports in clusterpipelinetemplate\").\n\t\t\tReturns(200, \"OK\", clusterpipelinetemplate.PipelineExportedVariables{}))\n\n\t// PipelineTemplateSync\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/pipelinetemplatesync/{namespace}\").\n\t\t\tTo(apiHandler.handleGetPipelineTemplateSyncList).\n\t\t\tWrites(pipelinetemplatesync.PipelineTemplateSyncList{}).\n\t\t\tDoc(\"get pipelineTemplateSync list\").\n\t\t\tReturns(200, \"OK\", pipelinetemplatesync.PipelineTemplateSyncList{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/pipelinetemplatesync/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleGetPipelineTemplateSync).\n\t\t\tWrites(pipelinetemplatesync.PipelineTemplateSync{}).\n\t\t\tDoc(\"get detail of specific PipelineTemplateSync\").\n\t\t\tReturns(200, \"OK\", pipelinetemplatesync.PipelineTemplateSync{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/pipelinetemplatesync/{namespace}\").\n\t\t\tTo(apiHandler.handleCreatePipelineTemplateSync).\n\t\t\tWrites(pipelinetemplatesync.PipelineTemplateSync{}).\n\t\t\tDoc(\"create a pipelineTemplateSync\").\n\t\t\tReturns(200, \"OK\", pipelinetemplatesync.PipelineTemplateSync{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/pipelinetemplatesync/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleUpdatePipelineTemplateSync).\n\t\t\tWrites(pipelinetemplatesync.PipelineTemplateSync{}).\n\t\t\tDoc(\"update a pipelineTemplateSync\").\n\t\t\tReturns(200, \"OK\", pipelinetemplatesync.PipelineTemplateSync{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/pipelinetemplatesync/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleDeletePipelineTemplateSync).\n\t\t\tWrites(struct{}{}).\n\t\t\tDoc(\"delete a PipelineTemplateSync\").\n\t\t\tReturns(200, \"OK\", struct{}{}))\n\n\t// PipelineTaskTemplate\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/pipelinetasktemplate/{namespace}\").\n\t\t\tTo(apiHandler.handleGetPipelineTaskTemplateList).\n\t\t\tWrites(pipelinetasktemplate.PipelineTaskTemplateList{}).\n\t\t\tDoc(\"get a list of PipelineTaskTemplate\").\n\t\t\tReturns(200, \"OK\", pipelinetasktemplate.PipelineTaskTemplate{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/pipelinetasktemplate/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleGetPipelineTaskTemplate).\n\t\t\tWrites(pipelinetasktemplate.PipelineTaskTemplate{}).\n\t\t\tDoc(\"get a PipelineTaskTemplate\").\n\t\t\tReturns(200, \"OK\", pipelinetasktemplate.PipelineTaskTemplate{}))\n\n\t// PipelineTemplate\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/pipelinetemplate/{namespace}\").\n\t\t\tTo(apiHandler.handleGetPipelineTemplateList).\n\t\t\tWrites(pipelinetemplate.PipelineTemplateList{}).\n\t\t\tDoc(\"get a list of PipelineTemplate\").\n\t\t\tReturns(200, \"OK\", pipelinetemplate.PipelineTemplateList{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/pipelinetemplate/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleGetPipelineTemplate).\n\t\t\tWrites(pipelinetemplate.PipelineTemplate{}).\n\t\t\tDoc(\"get a PipelineTemplate\").\n\t\t\tReturns(200, \"OK\", pipelinetemplate.PipelineTemplate{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"pipelinetemplate/{namespace}/{name}/preview\").\n\t\t\tTo(apiHandler.handlePreviewPipelineTemplate).\n\t\t\tWrites(pipelinetemplate.PreviewOptions{}).\n\t\t\tDoc(\"jenkinsfile preview from PipelineTemplate\").\n\t\t\tReturns(200, \"OK\", \"\"))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/pipelinetemplate/{namespace}/{name}/exports\").\n\t\t\tTo(apiHandler.handlerExportsPiplineTemplate).\n\t\t\tWrites(clusterpipelinetemplate.PipelineExportedVariables{}).\n\t\t\tDoc(\"get the exports in pipelinetemplate\").\n\t\t\tReturns(200, \"OK\", clusterpipelinetemplate.PipelineExportedVariables{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/pipelinetemplatecategories/{namespace}\").\n\t\t\tTo(apiHandler.handlePipelinetemplatecategories).\n\t\t\tWrites(pipelinetemplate.PipelineTemplateCategoryList{}).\n\t\t\tDoc(\"get a PipelineTemplate\").\n\t\t\tReturns(200, \"OK\", pipelinetemplate.PipelineTemplateCategoryList{}))\n\n\t// endregion\n\n\t// region Pipeline\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/pipelineconfig/{namespace}\").\n\t\t\tTo(apiHandler.handleGetPipelineConfigList).\n\t\t\tWrites(pipelineconfig.PipelineConfigList{}).\n\t\t\tDoc(\"get namespaced pipelineconfig list\").\n\t\t\tReturns(200, \"OK\", pipelineconfig.PipelineConfigList{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/pipelineconfig/{namespace}\").\n\t\t\tTo(apiHandler.handleCreatePipelineConfig).\n\t\t\tWrites(pipelineconfig.PipelineConfigDetail{}).\n\t\t\tDoc(\"creates namespaced pipelineconfig\").\n\t\t\tReturns(200, \"OK\", pipelineconfig.PipelineConfigDetail{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/pipelineconfig/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleGetPipelineConfigDetail).\n\t\t\tWrites(pipelineconfig.PipelineConfig{}).\n\t\t\tDoc(\"get pipeline config details\").\n\t\t\tReturns(200, \"OK\", pipelineconfig.PipelineConfigDetail{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/pipelineconfig/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleUpdatePipelineConfig).\n\t\t\tWrites(pipelineconfig.PipelineConfig{}).\n\t\t\tDoc(\"update pipeline config\").\n\t\t\tReturns(200, \"OK\", pipelineconfig.PipelineConfigDetail{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/pipelineconfig/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleDeletePipelineConfig).\n\t\t\tWrites(struct{}{}).\n\t\t\tDoc(\"deletes a pipeline config\").\n\t\t\tReturns(200, \"OK\", struct{}{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/pipelineconfig/{namespace}/{name}/trigger\").\n\t\t\tTo(apiHandler.handleTriggerPipelineConfig).\n\t\t\tWrites(pipelineconfig.PipelineConfigTrigger{}).\n\t\t\tDoc(\"triggers pipeline\").\n\t\t\tReturns(200, \"OK\", pipelineconfig.PipelineTriggerResponse{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/pipelineconfig/{namespace}/{name}/preview\").\n\t\t\tTo(apiHandler.handlePreviewPipelineConfig).\n\t\t\tWrites(pipelineconfig.PipelineConfigDetail{}).\n\t\t\tDoc(\"jenkinsfile preview\").\n\t\t\tReturns(200, \"OK\", \"\"))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/pipelineconfig/{namespace}/{name}/scan\").\n\t\t\tTo(apiHandler.handleScanPipelineConfig).\n\t\t\tDoc(\"scan multi-branch\").\n\t\t\tReturns(200, \"OK\", \"\"))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/pipelineconfig/{namespace}/{name}/logs\").\n\t\t\tParam(restful.PathParameter(\"namespace\", \"Namespace to use\")).\n\t\t\tParam(restful.PathParameter(\"name\", \"Pipeline name to filter scope\")).\n\t\t\tParam(restful.QueryParameter(\"start\", \"Start offset to fetch logs\")).\n\t\t\tTo(apiHandler.handlePipelineConfigLogs).\n\t\t\tDoc(\"gets scan logs for multi-branch pipeline\").\n\t\t\tReturns(200, \"OK\", v1alpha1.PipelineConfigLog{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/pipeline/{namespace}\").\n\t\t\tTo(apiHandler.handleGetPipelineList).\n\t\t\tWrites(pipeline.PipelineList{}).\n\t\t\tDoc(\"get namespaced pipeline list\").\n\t\t\tReturns(200, \"OK\", pipeline.PipelineList{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/pipeline/{namespace}/{name}\").\n\t\t\tParam(restful.QueryParameter(\"withFreshStages\", \"Whether to retrieve newest stages from Jenkins\")).\n\t\t\tTo(apiHandler.handleGetPipelineDetail).\n\t\t\tWrites(pipeline.Pipeline{}).\n\t\t\tDoc(\"get pipeline details\").\n\t\t\tReturns(200, \"OK\", pipeline.Pipeline{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/pipeline/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleDeletePipeline).\n\t\t\tWrites(struct{}{}).\n\t\t\tDoc(\"deletes a pipeline\").\n\t\t\tReturns(200, \"OK\", struct{}{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/pipeline/{namespace}/{name}/retry\").\n\t\t\tTo(apiHandler.handleRetryPipelineDetail).\n\t\t\tWrites(pipeline.RetryRequest{}).\n\t\t\tDoc(\"retries a pipeline\").\n\t\t\tReturns(200, \"OK\", v1alpha1.Pipeline{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/pipeline/{namespace}/{name}/abort\").\n\t\t\tTo(apiHandler.handleAbortPipeline).\n\t\t\tWrites(pipeline.AbortRequest{}).\n\t\t\tDoc(\"aborts a pipeline\").\n\t\t\tReturns(200, \"OK\", v1alpha1.Pipeline{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/pipeline/{namespace}/{name}/logs\").\n\t\t\tParam(restful.PathParameter(\"namespace\", \"Namespace to use\")).\n\t\t\tParam(restful.PathParameter(\"name\", \"Pipeline name to filter scope\")).\n\t\t\tParam(restful.QueryParameter(\"start\", \"Start offset to fetch logs\")).\n\t\t\tParam(restful.QueryParameter(\"stage\", \"Stage to fetch logs from\")).\n\t\t\tParam(restful.QueryParameter(\"step\", \"Step to fetch logs from. Can be combined with stage\")).\n\t\t\tTo(apiHandler.handlePipelineLogs).\n\t\t\tDoc(\"gets logs for pipeline\").\n\t\t\tReturns(200, \"OK\", v1alpha1.PipelineLog{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/pipeline/{namespace}/{name}/tasks\").\n\t\t\tParam(restful.PathParameter(\"namespace\", \"Namespace to use\")).\n\t\t\tParam(restful.PathParameter(\"name\", \"Pipeline name to filter scope\")).\n\t\t\tParam(restful.QueryParameter(\"stage\", \"Stage to fetch steps from. If not provided will return all stages\")).\n\t\t\tTo(apiHandler.handlePipelineTasks).\n\t\t\tDoc(\"gets steps for pipeline\").\n\t\t\tReturns(200, \"OK\", v1alpha1.PipelineTask{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/pipeline/{namespace}/{name}/inputs\").\n\t\t\tTo(apiHandler.handlePipelineInput).\n\t\t\tWrites(pipeline.InputOptions{}).\n\t\t\tDoc(\"response a input request which in a pipeline\").\n\t\t\tReturns(200, \"OK\", pipeline.InputResponse{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/pipeline/{namespace}/{name}/testreports\").\n\t\t\tParam(restful.QueryParameter(\"start\", \"Start offset to fetch test report items\")).\n\t\t\tParam(restful.QueryParameter(\"limit\", \"Limit of number to fetch test report items\")).\n\t\t\tTo(apiHandler.handlePipelineTestReports).\n\t\t\tDoc(\"response a input request which in a pipeline\").\n\t\t\tReturns(200, \"OK\", pipeline.PipelineTestReports{}))\n\n\t// endregion\n\n\t// region CodeRepository\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/codereposervice\").\n\t\t\tTo(apiHandler.handleCreateCodeRepoService).\n\t\t\tWrites(codereposervice.CodeRepoServiceList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/codereposervice/{name}\").\n\t\t\tTo(apiHandler.handleDeleteCodeRepoService).\n\t\t\tWrites(codereposervice.CodeRepoService{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/codereposervice/{name}\").\n\t\t\tTo(apiHandler.handleUpdateCodeRepoService).\n\t\t\tWrites(v1alpha1.CodeRepoService{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/codereposervice\").\n\t\t\tTo(apiHandler.handleGetCodeRepoServiceList).\n\t\t\tWrites(codereposervice.CodeRepoServiceList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/codereposervices\").\n\t\t\tTo(apiHandler.handleGetCodeRepoServiceList).\n\t\t\tWrites(codereposervice.CodeRepoServiceList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/codereposervice/{name}\").\n\t\t\tTo(apiHandler.handleGetCodeRepoServiceDetail).\n\t\t\tWrites(v1alpha1.CodeRepoService{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/codereposervice/{name}/resources\").\n\t\t\tTo(apiHandler.handleGetCodeRepoServiceResourceList).\n\t\t\tWrites(common.ResourceList{}).\n\t\t\tDoc(\"retrieve resources associated with codereposervice\").\n\t\t\tReturns(200, \"OK\", common.ResourceList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/codereposervice/{name}/secrets\").\n\t\t\tTo(apiHandler.handleGetCodeRepoServiceSecretList).\n\t\t\tWrites(secret.SecretList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/coderepobinding/{namespace}\").\n\t\t\tTo(apiHandler.handleCreateCodeRepoBinding).\n\t\t\tWrites(v1alpha1.CodeRepoBinding{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/coderepobinding/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleDeleteCodeRepoBinding).\n\t\t\tWrites(v1alpha1.CodeRepoBinding{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/coderepobinding/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleUpdateCodeRepoBinding).\n\t\t\tWrites(struct{}{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/coderepobinding\").\n\t\t\tTo(apiHandler.handleGetCodeRepoBindingList).\n\t\t\tWrites(coderepobinding.CodeRepoBindingList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/coderepobinding/{namespace}\").\n\t\t\tTo(apiHandler.handleGetCodeRepoBindingList).\n\t\t\tWrites(coderepobinding.CodeRepoBindingList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/coderepobinding/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleGetCodeRepoBindingDetail).\n\t\t\tWrites(v1alpha1.CodeRepoBinding{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/coderepobinding/{namespace}/{name}/resources\").\n\t\t\tTo(apiHandler.handleGetCodeRepoBindingResources).\n\t\t\tWrites(common.ResourceList{}).\n\t\t\tDoc(\"retrieve resources associated with coderepobinding\").\n\t\t\tReturns(200, \"OK\", common.ResourceList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/coderepobinding/{namespace}/{name}/secrets\").\n\t\t\tTo(apiHandler.handleGetCodeRepoBindingSecretList).\n\t\t\tWrites(secret.SecretList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/coderepobinding/{namespace}/{name}/repositories\").\n\t\t\tTo(apiHandler.handleGetCodeRepositoryListInBinding).\n\t\t\tWrites(coderepository.CodeRepositoryList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/coderepobinding/{namespace}/{name}/remote-repositories\").\n\t\t\tTo(apiHandler.handleGetRemoteRepositoryList).\n\t\t\tWrites(coderepository.CodeRepositoryList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/coderepository/{namespace}\").\n\t\t\tTo(apiHandler.handleGetCodeRepositoryList).\n\t\t\tWrites(coderepository.CodeRepositoryList{}).\n\t\t\tDoc(\"get namespaced coderepository list\").\n\t\t\tReturns(200, \"OK\", coderepository.CodeRepositoryList{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/coderepository/{namespace}/{name}/branches\").\n\t\t\tParam(restful.PathParameter(\"sortBy\", \"sort option. The choices are creationTime\")).\n\t\t\tParam(restful.PathParameter(\"sortMode\", \"sort option. The choices are desc or asc\")).\n\t\t\tTo(apiHandler.HandleGetCodeRepositoryBranches).\n\t\t\tReturns(200, \"Get coderepo branch Successful\", v1alpha1.CodeRepoBranchResult{}))\n\n\t// endregion\n\n\t// region ToolChain\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/toolchain\").\n\t\t\tTo(apiHandler.handleGetToolChains).\n\t\t\tWrites(toolchain.ToolChainList{}).\n\t\t\tDoc(\"get namespaced coderepository list\").\n\t\t\tReturns(200, \"OK\", coderepository.CodeRepositoryList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/toolchain/bindings\").\n\t\t\tTo(apiHandler.handleGetToolChainBindings).\n\t\t\tWrites(toolchain.ToolChainBindingList{}).\n\t\t\tDoc(\"get toolchain binding list\").\n\t\t\tReturns(200, \"OK\", coderepository.CodeRepositoryList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/toolchain/bindings/{namespace}\").\n\t\t\tTo(apiHandler.handleGetToolChainBindings).\n\t\t\tWrites(toolchain.ToolChainBindingList{}).\n\t\t\tDoc(\"get namespaced toolchain binding list\").\n\t\t\tReturns(200, \"OK\", coderepository.CodeRepositoryList{}))\n\t// endregion\n\n\t// region callback\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/callback/oauth/{namespace}/secret/{secretNamespace}/{secretName}/codereposervice/{serviceName}\").\n\t\t\tTo(apiHandler.handleOAuthCallback).\n\t\t\tWrites(struct{}{}))\n\t// endregion\n\n\t// region ImageRegistry\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/imageregistry\").\n\t\t\tTo(apiHandler.handleCreateImageRegistry).\n\t\t\tWrites(imageregistry.ImageRegistryList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/imageregistry/{name}\").\n\t\t\tTo(apiHandler.handleDeleteImageRegsitry).\n\t\t\tWrites(imageregistry.ImageRegistry{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/imageregistry/{name}\").\n\t\t\tTo(apiHandler.handleUpdateImageRegistry).\n\t\t\tWrites(v1alpha1.ImageRegistry{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/imageregistry\").\n\t\t\tTo(apiHandler.handleGetImageRegistryList).\n\t\t\tWrites(imageregistry.ImageRegistryList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/imageregistry/{name}\").\n\t\t\tTo(apiHandler.handleGetImageRegistryDetail).\n\t\t\tWrites(v1alpha1.ImageRegistry{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/imageregistry/{name}/secrets\").\n\t\t\tTo(apiHandler.handleGetImageRegistrySecretList).\n\t\t\tWrites(secret.SecretList{}))\n\t// endregion\n\n\t// region ImageRegistryBinding\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/imageregistrybinding/{namespace}\").\n\t\t\tTo(apiHandler.handleCreateImageRegistryBinding).\n\t\t\tWrites(v1alpha1.ImageRegistryBinding{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(URLIMAGEREGISTRYBINDINGDETAIL).\n\t\t\tTo(apiHandler.handleUpdateImageRegistryBinding).\n\t\t\tWrites(v1alpha1.ImageRegistryBinding{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(URLIMAGEREGISTRYBINDINGDETAIL).\n\t\t\tTo(apiHandler.handleDeleteImageRegistryBinding).\n\t\t\tWrites(v1alpha1.ImageRegistryBinding{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/imageregistrybinding\").\n\t\t\tTo(apiHandler.handleGetImageRegistryBindingList).\n\t\t\tWrites(imageregistrybinding.ImageRegistryBindingList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/imageregistrybinding/{namespace}\").\n\t\t\tTo(apiHandler.handleGetImageRegistryBindingList).\n\t\t\tWrites(imageregistrybinding.ImageRegistryBindingList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(URLIMAGEREGISTRYBINDINGDETAIL).\n\t\t\tTo(apiHandler.handleGetImageRegistryBindingDetail).\n\t\t\tWrites(v1alpha1.ImageRegistryBinding{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/imageregistrybinding/{namespace}/{name}/secrets\").\n\t\t\tTo(apiHandler.handleGetImageRegistryBindingSecretList).\n\t\t\tWrites(secret.SecretList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/imageregistrybinding/{namespace}/{name}/repositories\").\n\t\t\tTo(apiHandler.handleGetImageRepositoryListInBinding).\n\t\t\tWrites(imagerepository.ImageRepositoryList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/imageregistrybinding/{namespace}/{name}/remote-repositories\").\n\t\t\tTo(apiHandler.handleGetImageOriginRepositoryList).\n\t\t\tWrites(imagerepository.ImageRepositoryList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/imageregistrybinding/{namespace}/{name}/remote-repositories-project\").\n\t\t\tTo(apiHandler.handleGetImageOriginRepositoryProjectList).\n\t\t\tWrites(imagerepository.ImageRepositoryList{}))\n\t// endregion\n\n\t// region ImageRepository\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/imagerepository/{namespace}\").\n\t\t\tTo(apiHandler.handleGetImageRepositoryList).\n\t\t\tWrites(imagerepository.ImageRepositoryList{}).\n\t\t\tDoc(\"get namespaced imagerepository list\").\n\t\t\tReturns(200, \"OK\", imagerepository.ImageRepositoryList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/imagerepositoryproject/{namespace}\").\n\t\t\tTo(apiHandler.handleGetImageRepositoryProjectList).\n\t\t\tWrites(imagerepository.ImageRepositoryList{}).\n\t\t\tDoc(\"get namespaced imagerepository list\").\n\t\t\tReturns(200, \"OK\", imagerepository.ImageRepositoryList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/imagerepository/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleGetImageRepositoryDetail).\n\t\t\tWrites(v1alpha1.ImageRepository{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/imagerepository/{namespace}/{name}/tags\").\n\t\t\tParam(restful.PathParameter(\"sortBy\", \"sort option. The choices are creationTime\")).\n\t\t\tParam(restful.PathParameter(\"sortMode\", \"sort option. The choices are desc or asc\")).\n\t\t\tTo(apiHandler.HandleGetImageTags).\n\t\t\tReturns(200, \"Get Image tags Successful\", v1alpha1.ImageTagResult{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"imagerepository/{namespace}/{name}/security\").\n\t\t\tParam(restful.PathParameter(\"tag\", \"Scan image tag name\")).\n\t\t\tTo(apiHandler.HandleScanImage).\n\t\t\tReturns(200, \"Create Scan Image Job Successful.\", v1alpha1.ImageResult{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"imagerepository/{namespace}/{name}/security\").\n\t\t\tParam(restful.PathParameter(\"tag\", \"Get image vulnerability tag name\")).\n\t\t\tTo(apiHandler.HandleGetVulnerability).\n\t\t\tReturns(200, \"Get Image Vulnerability Successful\", v1alpha1.VulnerabilityList{}))\n\t// endregion\n\n\t// region microservicesenvironments\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/microservicesenvironments\").\n\t\t\tFilter(mw.Product(ACPServiceFramework)).\n\t\t\tTo(apiHandler.handleMicroservicesEnvironmentList).\n\t\t\tWrites(asfClient.MicroservicesEnvironmentList{}).\n\t\t\tDoc(\"get microservicesenvironment list\").\n\t\t\tReturns(200, \"OK\", asfClient.MicroservicesEnvironmentList{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/microservicesenvironments/{name}\").\n\t\t\tFilter(mw.Product(ACPServiceFramework)).\n\t\t\tWrites(microservicesenvironment.MicroservicesEnvironmentDetail{}).\n\t\t\tTo(apiHandler.handleGetMicroservicesEnviromentDetail).\n\t\t\tDoc(\"get microservicesenvironments detail by name\").\n\t\t\tReturns(200, \"OK\", microservicesenvironment.MicroservicesEnvironmentDetail{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/microservicescomponent/{namespace}/{name}\").\n\t\t\tFilter(mw.Product(ACPServiceFramework)).\n\t\t\tTo(apiHandler.handlePutMicroservicesComponent).\n\t\t\tWrites(asfClient.MicroservicesComponent{}).\n\t\t\tDoc(\"install component\").\n\t\t\tReturns(200, \"OK\", asfClient.MicroservicesComponent{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/microservicescomponent/{namespace}/{name}\").\n\t\t\tFilter(mw.Product(ACPServiceFramework)).\n\t\t\tTo(apiHandler.handlePutMicroservicesComponent).\n\t\t\tWrites(asfClient.MicroservicesComponent{}).\n\t\t\tDoc(\"update component\").\n\t\t\tReturns(200, \"OK\", asfClient.MicroservicesComponent{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/microservicescomponent/{namespace}/{name}/start\").\n\t\t\tFilter(mw.Product(ACPServiceFramework)).\n\t\t\tTo(apiHandler.handlePutMicroservicesComponentStart).\n\t\t\tWrites(asfClient.MicroservicesComponent{}).\n\t\t\tDoc(\"start component\").\n\t\t\tReturns(200, \"OK\", asfClient.MicroservicesComponent{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/microservicescomponent/{namespace}/{name}/stop\").\n\t\t\tFilter(mw.Product(ACPServiceFramework)).\n\t\t\tTo(apiHandler.handlePutMicroservicesComponentStop).\n\t\t\tWrites(asfClient.MicroservicesComponentList{}).\n\t\t\tDoc(\"stop component\").\n\t\t\tReturns(200, \"OK\", asfClient.MicroservicesComponent{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/microservicesapps\").\n\t\t\tFilter(mw.Product(ACPServiceFramework)).\n\t\t\tWrites(microservicesapplication.MicroservicesApplicationList{}).\n\t\t\tTo(apiHandler.handleGetMicroservicesApps).\n\t\t\tDoc(\"get microservicesenvironments detail by name\").\n\t\t\tReturns(200, \"OK\", microservicesapplication.MicroservicesApplicationList{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/microservicesconfigs\").\n\t\t\tWrites(microservicesconfiguration.MicroservicesConfigurationList{}).\n\t\t\tTo(apiHandler.handleGetMicroservicesConfigs).\n\t\t\tDoc(\"get microservicesenvironments detail by name\").\n\t\t\tReturns(200, \"OK\", microservicesconfiguration.MicroservicesConfigurationList{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/domains\").\n\t\t\tWrites(domain.DomainList{}).\n\t\t\tTo(apiHandler.handleGetDomainList).\n\t\t\tDoc(\"get microservicesenvironments detail by name\").\n\t\t\tReturns(200, \"OK\", domain.DomainList{}))\n\n\t// endregion\n\n\t// region ProjectManagement\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/projectmanagement\").\n\t\t\tTo(apiHandler.handleCreateProjectManagement).\n\t\t\tWrites(v1alpha1.ProjectManagement{}).\n\t\t\tDoc(\"create a projectmanagement\").\n\t\t\tReturns(200, \"OK\", v1alpha1.ProjectManagement{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(URLProjectManagementDetails).\n\t\t\tTo(apiHandler.handleDeleteProjectManagement).\n\t\t\tWrites(struct{}{}).\n\t\t\tDoc(\"delete a projectmanagement\").\n\t\t\tReturns(200, \"OK\", struct{}{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(URLProjectManagementDetails).\n\t\t\tTo(apiHandler.handleUpdateProjectManagement).\n\t\t\tWrites(v1alpha1.ProjectManagement{}).\n\t\t\tDoc(\"update a projectmanagement\").\n\t\t\tReturns(200, \"OK\", v1alpha1.ProjectManagement{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/projectmanagement\").\n\t\t\tTo(apiHandler.handleGetProjectManagementList).\n\t\t\tWrites(projectmanagement.ProjectManagementList{}).\n\t\t\tDoc(\"get projectmanagement list\").\n\t\t\tReturns(200, \"OK\", projectmanagement.ProjectManagementList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(URLProjectManagementDetails).\n\t\t\tTo(apiHandler.handleGetProjectManagementDetail).\n\t\t\tWrites(v1alpha1.ProjectManagement{}).\n\t\t\tDoc(\"get a projectmanagement\").\n\t\t\tReturns(200, \"OK\", v1alpha1.ProjectManagement{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/projectmanagementbinding/{namespace}\").\n\t\t\tTo(apiHandler.handleCreateProjectManagementBinding).\n\t\t\tWrites(v1alpha1.ProjectManagementBinding{}).\n\t\t\tDoc(\"create a projectmanagementbinding\").\n\t\t\tReturns(200, \"OK\", v1alpha1.ProjectManagementBinding{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(URLProjectManagementBindingDetails).\n\t\t\tTo(apiHandler.handleDeleteProjectManagementBinding).\n\t\t\tWrites(struct{}{}).\n\t\t\tDoc(\"delete a projectmanagementbinding\").\n\t\t\tReturns(200, \"OK\", struct{}{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(URLProjectManagementBindingDetails).\n\t\t\tTo(apiHandler.handleUpdateProjectManagementBinding).\n\t\t\tWrites(v1alpha1.ProjectManagementBinding{}).\n\t\t\tDoc(\"update a projectmanagementbinding\").\n\t\t\tReturns(200, \"OK\", v1alpha1.ProjectManagementBinding{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/projectmanagementbinding\").\n\t\t\tTo(apiHandler.handleGetProjectManagementBindingList).\n\t\t\tWrites(projectmanagementbinding.ProjectManagementBindingList{}).\n\t\t\tDoc(\"get projectmanagementbinding list in all namespaces\").\n\t\t\tReturns(200, \"OK\", projectmanagementbinding.ProjectManagementBindingList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/projectmanagementbinding/{namespace}\").\n\t\t\tTo(apiHandler.handleGetProjectManagementBindingList).\n\t\t\tWrites(projectmanagementbinding.ProjectManagementBindingList{}).\n\t\t\tDoc(\"get projectmanagementbinding list in one namespace\").\n\t\t\tReturns(200, \"OK\", projectmanagementbinding.ProjectManagementBindingList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(URLProjectManagementBindingDetails).\n\t\t\tTo(apiHandler.handleGetProjectManagementBindingDetail).\n\t\t\tWrites(v1alpha1.ProjectManagementBinding{}).\n\t\t\tDoc(\"get a projectmanagementbinding\").\n\t\t\tReturns(200, \"OK\", v1alpha1.ProjectManagementBinding{}))\n\t// endregion\n\n\t// region TestTool\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/testtool\").\n\t\t\tTo(apiHandler.handleCreateTestTool).\n\t\t\tWrites(v1alpha1.TestTool{}).\n\t\t\tDoc(\"create a testtool\").\n\t\t\tReturns(200, \"OK\", v1alpha1.TestTool{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(URLTestToolDetails).\n\t\t\tTo(apiHandler.handleDeleteTestTool).\n\t\t\tWrites(struct{}{}).\n\t\t\tDoc(\"delete a testtool\").\n\t\t\tReturns(200, \"OK\", struct{}{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(URLTestToolDetails).\n\t\t\tTo(apiHandler.handleUpdateTestTool).\n\t\t\tWrites(v1alpha1.TestTool{}).\n\t\t\tDoc(\"update a testtool\").\n\t\t\tReturns(200, \"OK\", v1alpha1.TestTool{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/testtool\").\n\t\t\tTo(apiHandler.handleGetTestToolList).\n\t\t\tWrites(testtool.TestToolList{}).\n\t\t\tDoc(\"get testtool list\").\n\t\t\tReturns(200, \"OK\", testtool.TestToolList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(URLTestToolDetails).\n\t\t\tTo(apiHandler.handleGetTestToolDetail).\n\t\t\tWrites(v1alpha1.TestTool{}).\n\t\t\tDoc(\"get a testtool\").\n\t\t\tReturns(200, \"OK\", v1alpha1.TestTool{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/testtoolbinding/{namespace}\").\n\t\t\tTo(apiHandler.handleCreateTestToolBinding).\n\t\t\tWrites(v1alpha1.TestToolBinding{}).\n\t\t\tDoc(\"create a testtoolbinding\").\n\t\t\tReturns(200, \"OK\", v1alpha1.TestToolBinding{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(URLTestToolBindingDetails).\n\t\t\tTo(apiHandler.handleDeleteTestToolBinding).\n\t\t\tWrites(struct{}{}).\n\t\t\tDoc(\"delete a testtoolbinding\").\n\t\t\tReturns(200, \"OK\", struct{}{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(URLTestToolBindingDetails).\n\t\t\tTo(apiHandler.handleUpdateTestToolBinding).\n\t\t\tWrites(v1alpha1.TestToolBinding{}).\n\t\t\tDoc(\"update a testtoolbinding\").\n\t\t\tReturns(200, \"OK\", v1alpha1.TestToolBinding{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/testtoolbinding\").\n\t\t\tTo(apiHandler.handleGetTestToolBindingList).\n\t\t\tWrites(testtoolbinding.TestToolBindingList{}).\n\t\t\tDoc(\"get testtoolbinding list in all namespaces\").\n\t\t\tReturns(200, \"OK\", testtoolbinding.TestToolBindingList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/testtoolbinding/{namespace}\").\n\t\t\tTo(apiHandler.handleGetTestToolBindingList).\n\t\t\tWrites(testtoolbinding.TestToolBindingList{}).\n\t\t\tDoc(\"get testtoolbinding list in one namespace\").\n\t\t\tReturns(200, \"OK\", testtoolbinding.TestToolBindingList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(URLTestToolBindingDetails).\n\t\t\tTo(apiHandler.handleGetTestToolBindingDetail).\n\t\t\tWrites(v1alpha1.TestToolBinding{}).\n\t\t\tDoc(\"get a testtoolbinding\").\n\t\t\tReturns(200, \"OK\", v1alpha1.TestToolBinding{}))\n\t// endregion\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/microservicesconfigs\").\n\t\t\tFilter(mw.Product(ACPServiceFramework)).\n\t\t\tWrites(microservicesconfiguration.MicroservicesConfigurationList{}).\n\t\t\tTo(apiHandler.handleGetMicroservicesConfigs).\n\t\t\tDoc(\"get microservicesenvironments detail by name\").\n\t\t\tReturns(200, \"OK\", microservicesconfiguration.MicroservicesConfigurationList{}))\n\n\t// region Statistics\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/statistics/pipeline/{namespace}\").\n\t\t\tTo(apiHandler.handleGetPipelineStatistics).\n\t\t\tWrites(struct{}{}).\n\t\t\tDoc(\"get the statistics info of pipeline\").\n\t\t\tReturns(200, \"OK\", struct{}{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/statistics/stage/{namespace}\").\n\t\t\tTo(apiHandler.handleGetStageStatistics).\n\t\t\tWrites(struct{}{}).\n\t\t\tDoc(\"get the statistics info of stage\").\n\t\t\tReturns(200, \"OK\", struct{}{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/statistics/codequality/{namespace}\").\n\t\t\tTo(apiHandler.handleGetCodeQualityStatistics).\n\t\t\tWrites(struct{}{}).\n\t\t\tDoc(\"get the statistics info of stage\").\n\t\t\tReturns(200, \"OK\", struct{}{}))\n\n\t// endregion\n\n\t// region CodeQualityTool\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/codequalitytool\").\n\t\t\tTo(apiHandler.handleCreateCodeQualityTool).\n\t\t\tWrites(v1alpha1.CodeQualityTool{}).\n\t\t\tDoc(\"create a code quality tool\").\n\t\t\tReturns(200, \"OK\", v1alpha1.CodeQualityTool{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(URLCODEQUALITYTOOLDETAIL).\n\t\t\tTo(apiHandler.handleDeleteCodeQualityTool).\n\t\t\tWrites(v1alpha1.CodeQualityTool{}).\n\t\t\tDoc(\"delete a code quality tool with name\").\n\t\t\tReturns(200, \"OK\", v1alpha1.CodeQualityTool{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(URLCODEQUALITYTOOLDETAIL).\n\t\t\tTo(apiHandler.handleUpdateCodeQualityTool).\n\t\t\tWrites(v1alpha1.CodeQualityTool{}).\n\t\t\tDoc(\"update a code quality tool with name\").\n\t\t\tReturns(200, \"OK\", v1alpha1.CodeQualityTool{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(URLCODEQUALITYTOOLDETAIL).\n\t\t\tTo(apiHandler.handleGetCodeQualityTool).\n\t\t\tWrites(v1alpha1.CodeQualityTool{}).\n\t\t\tDoc(\"get a code quality tool with name\").\n\t\t\tReturns(200, \"OK\", v1alpha1.CodeQualityTool{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/codequalitytool\").\n\t\t\tTo(apiHandler.handleListCodeQualityTool).\n\t\t\tWrites(v1alpha1.CodeQualityTool{}).\n\t\t\tDoc(\"list code quality tools\").\n\t\t\tReturns(200, \"OK\", codequalitytool.CodeQualityToolList{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/codequalitybinding/{namespace}\").\n\t\t\tTo(apiHandler.handleCreateCodeQualityBinding).\n\t\t\tWrites(v1alpha1.CodeQualityBinding{}).\n\t\t\tDoc(\"create a code quality binding\").\n\t\t\tReturns(200, \"OK\", v1alpha1.CodeQualityBinding{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/codequalitybinding/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleUpdateCodeQualityBinding).\n\t\t\tDoc(\"update a code quality binding with name\").\n\t\t\tReturns(200, \"OK\", v1alpha1.CodeQualityBinding{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/codequalitybinding/{namespace}\").\n\t\t\tTo(apiHandler.handleGetCodeQualityBindingList).\n\t\t\tDoc(\"get namespaced code quality binding list\").\n\t\t\tReturns(200, \"OK\", codequalitybinding.CodeQualityBindingList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/codequalitybinding\").\n\t\t\tTo(apiHandler.handleGetCodeQualityBindingList).\n\t\t\tDoc(\"get all code quality binding list\").\n\t\t\tReturns(200, \"OK\", codequalitybinding.CodeQualityBindingList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/codequalitybinding/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleGetCodeQualityBindingDetail).\n\t\t\tDoc(\"get code quality binding with name\").\n\t\t\tReturns(200, \"OK\", v1alpha1.CodeQualityBinding{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/codequalitybinding/{namespace}/{name}/projects\").\n\t\t\tTo(apiHandler.handleGetCodeQualityProjectListInBinding).\n\t\t\tDoc(\"get code quality project list in binding\").\n\t\t\tReturns(200, \"OK\", codequalityproject.CodeQualityProjectList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/codequalitybinding/{namespace}/{name}/secrets\").\n\t\t\tTo(apiHandler.handleGetCodeQualityBindingSecretList).\n\t\t\tDoc(\"get bind secret list\").\n\t\t\tReturns(200, \"OK\", secret.SecretList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/codequalitybinding/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleDeleteCodeQualityBinding).\n\t\t\tDoc(\"delete code quality binding with name\").\n\t\t\tReturns(200, \"OK\", common.ResourceList{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/codequalityproject/{namespace}\").\n\t\t\tTo(apiHandler.handleCreateCodeQualityProject).\n\t\t\tDoc(\"create a code quality project\").\n\t\t\tReturns(200, \"OK\", v1alpha1.CodeQualityProject{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/codequalityproject/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleUpdateCodeQualityProject).\n\t\t\tDoc(\"update a code quality project with name\").\n\t\t\tReturns(200, \"OK\", v1alpha1.CodeQualityProject{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/codequalityproject/{namespace}\").\n\t\t\tTo(apiHandler.handleGetCodeQualityProjectList).\n\t\t\tDoc(\"create a code quality project\").\n\t\t\tReturns(200, \"OK\", codequalityproject.CodeQualityProjectList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/codequalityproject/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleGetCodeQualityProjectDetail).\n\t\t\tDoc(\"create a code quality project\").\n\t\t\tReturns(200, \"OK\", v1alpha1.CodeQualityProject{}))\n\n\t//region asm\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/service/{namespace}\").\n\t\t\tTo(apiHandler.handleGetServiceListByProject).\n\t\t\tWrites(resourceService.ServiceNameList{}))\n\t//endregion\n\n\t// region asm\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/servicemesh/graphs/{namespace}\").\n\t\t\tTo(apiHandler.handleGetNamespaceGraph).\n\t\t\tDoc(\"get namespace service graph\").\n\t\t\tReturns(200, \"OK\", servicegraph.Graph{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/servicemesh/metrics\").\n\t\t\tTo(apiHandler.handleGetMetrics).\n\t\t\tDoc(\"get metrics from given options\").\n\t\t\tReturns(200, \"ok\", \"\"))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/servicemesh/nodegraphs\").\n\t\t\tTo(apiHandler.handleGetNodeGraph).\n\t\t\tDoc(\"get namespace service graph\").\n\t\t\tReturns(200, \"OK\", servicegraph.Graph{}))\n\t//endregion\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/microservice/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleGetMicroserviceRelation).\n\t\t\tDoc(\"get microservice deployment and svc relation\"))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/microservice/{namespace}/{name}/service\").\n\t\t\tTo(apiHandler.handleCreateMicroserviceSvc))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/microservice/{namespace}/{name}/service/{servicename}\").\n\t\t\tTo(apiHandler.handleUpdateMicroserviceSvc))\n\n\t// destinationrule\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/destinationrule/{namespace}\").\n\t\t\tTo(apiHandler.handleListDestinationRule).\n\t\t\tDoc(\"get namespace destination rule\"),\n\t)\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/destinationrule/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleGetDestinationRuleDetail),\n\t)\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/destinationruleinfohost/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleGetDestinationRuleInfoHost),\n\t)\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/destinationrule/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleUpdateDestinationRule))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/destinationrule/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleDeleteDestinationRule),\n\t)\n\n\t// virtualservice\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/virtualservice/{namespace}\").\n\t\t\tTo(apiHandler.handleListVirtualService))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/virtualservice/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleGetVirtualService))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/virtualservicehost/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleGetVirtualServiceByHost))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/virtualservice/{namespace}\").\n\t\t\tTo(apiHandler.handleCreateVirtualService),\n\t)\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/virtualservice/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleUpdateVirtualService))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/virtualservice/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleDeleteVirtualService),\n\t)\n\n\t// Policy\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/policy/{namespace}\").\n\t\t\tTo(apiHandler.handleListPolicy))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/policy/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleGetPolicy))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/policy/{namespace}\").\n\t\t\tTo(apiHandler.handleCreatePolicy),\n\t)\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/policy/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleUpdatePolicy))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/policy/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleDeletePolicy),\n\t)\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/gateway/{namespace}\").\n\t\t\tTo(apiHandler.handleListGateways))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/gateway/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleGetGateway))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/gateway/{namespace}\").\n\t\t\tTo(apiHandler.handleCreateGateway))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/gateway/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleUpdateGateway))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/gateway/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleDeleteGateway))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/asmclusterconfig/{name}\").\n\t\t\tTo(apiHandler.handleGetASMClusterConfig))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/asmclusterconfig/{name}\").\n\t\t\tTo(apiHandler.handleUpdateASMClusterConfig))\n\t// endregion\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/artifactregistrymanagers\").\n\t\t\tTo(apiHandler.handleCreateArtifactRegistryManager).\n\t\t\tWrites(artifactregistrymanager.ArtifactRegistryManager{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/artifactregistrymanagers/{name}\").\n\t\t\tTo(apiHandler.handleDeleteArtifactRegistryManager).\n\t\t\tWrites(artifactregistrymanager.ArtifactRegistryManager{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/artifactregistrymanagers/{name}\").\n\t\t\tTo(apiHandler.handleUpdateArtifactRegistryManager).\n\t\t\tWrites(v1alpha1.ArtifactRegistryManager{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/artifactregistrymanagers\").\n\t\t\tTo(apiHandler.handleGetArtifactRegistryManagerList).\n\t\t\tWrites(artifactregistrymanager.ArtifactRegistryManagerList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/artifactregistrymanagers/{name}\").\n\t\t\tTo(apiHandler.handleGetArtifactRegistryManagerDetail).\n\t\t\tWrites(v1alpha1.ArtifactRegistryManager{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/artifactregistries\").\n\t\t\tTo(apiHandler.handleCreateArtifactRegistry).\n\t\t\tWrites(artifactregistry.ArtifactRegistry{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/artifactregistries/{name}\").\n\t\t\tTo(apiHandler.handleDeleteArtifactRegistry).\n\t\t\tWrites(artifactregistry.ArtifactRegistry{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/artifactregistries/{name}\").\n\t\t\tTo(apiHandler.handleUpdateArtifactRegistry).\n\t\t\tWrites(v1alpha1.ArtifactRegistry{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/artifactregistries\").\n\t\t\tTo(apiHandler.handleGetArtifactRegistryList).\n\t\t\tWrites(artifactregistry.ArtifactRegistryList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/artifactregistries/{name}\").\n\t\t\tTo(apiHandler.handleGetArtifactRegistryDetail).\n\t\t\tWrites(v1alpha1.ArtifactRegistry{}))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/artifactregistrybindings\").\n\t\t\tTo(apiHandler.handleCreateArtifactRegistryBinding).\n\t\t\tWrites(artifactregistrybinding.ArtifactRegistryBinding{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/artifactregistrybindings/{namespace}\").\n\t\t\tTo(apiHandler.handleCreateArtifactRegistryBinding).\n\t\t\tWrites(artifactregistrybinding.ArtifactRegistryBinding{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/artifactregistrybindings/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleDeleteArtifactRegistryBinding).\n\t\t\tWrites(artifactregistrybinding.ArtifactRegistryBinding{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/artifactregistrybindings/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleUpdateArtifactRegistryBinding).\n\t\t\tWrites(v1alpha1.ArtifactRegistryBinding{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/artifactregistrybindings/{namespace}\").\n\t\t\tTo(apiHandler.handleGetArtifactRegistryBindingList).\n\t\t\tWrites(artifactregistrybinding.ArtifactRegistryBindingList{}))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/artifactregistrybindings/{namespace}/{name}\").\n\t\t\tTo(apiHandler.handleGetArtifactRegistryBindingDetail).\n\t\t\tWrites(v1alpha1.ArtifactRegistryBinding{}))\n\n\t//common route\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/common/{resource}\").\n\t\t\tTo(apiHandler.handlePostCommonResource).\n\t\t\tWrites(make(map[string]interface{})))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/common/{resource}/{name}\").\n\t\t\tTo(apiHandler.handleDeleteCommonResource).\n\t\t\tWrites(make(map[string]interface{})))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/common/{resource}/{name}\").\n\t\t\tTo(apiHandler.handlePutCommonResource).\n\t\t\tWrites(make(map[string]interface{})))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/common/{resource}\").\n\t\t\tTo(apiHandler.handleGetCommonResourceList).\n\t\t\tWrites(make(map[string]interface{})))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/common/{resource}/{name}\").\n\t\t\tTo(apiHandler.handleGetCommonResource).\n\t\t\tWrites(make(map[string]interface{})))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/common/namespace/{namespace}/{resource}\").\n\t\t\tTo(apiHandler.handlePostCommonResource).\n\t\t\tWrites(make(map[string]interface{})))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.DELETE(\"/common/namespace/{namespace}/{resource}/{name}\").\n\t\t\tTo(apiHandler.handleDeleteCommonResource).\n\t\t\tWrites(make(map[string]interface{})))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.PUT(\"/common/namespace/{namespace}/{resource}/{name}\").\n\t\t\tTo(apiHandler.handlePutCommonResource).\n\t\t\tWrites(make(map[string]interface{})))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/common/namespace/{namespace}/{resource}\").\n\t\t\tTo(apiHandler.handleGetCommonResourceList).\n\t\t\tWrites(make(map[string]interface{})))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/common/namespace/{namespace}/{resource}/{name}\").\n\t\t\tTo(apiHandler.handleGetCommonResource).\n\t\t\tWrites(make(map[string]interface{})))\n\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/common/{resource}/{name}/sub/{sub}\").\n\t\t\tTo(apiHandler.handleGetCommonResourceSub).\n\t\t\tWrites(make(map[string]interface{})))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/common/{resource}/{name}/sub/{sub}\").\n\t\t\tTo(apiHandler.handlePostCommonResourceSub).\n\t\t\tWrites(make(map[string]interface{})))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.GET(\"/common/namespace/{namespace}/{resource}/{name}/sub/{sub}\").\n\t\t\tTo(apiHandler.handleGetCommonResourceSub).\n\t\t\tWrites(make(map[string]interface{})))\n\tapiV1Ws.Route(\n\t\tapiV1Ws.POST(\"/common/namespace/{namespace}/{resource}/{name}/sub/{sub}\").\n\t\t\tTo(apiHandler.handlePostCommonResourceSub).\n\t\t\tWrites(make(map[string]interface{})))\n\n\tAddAppCoreUrl(apiV1Ws, apiHandler)\n\treturn wsContainer, nil\n}", "func ToAPISshConfig(from *pb.SshConfig) *system.SSHConfig {\n\tvar gw *system.SSHConfig\n\tif from.Gateway != nil {\n\t\tgw = ToAPISshConfig(from.Gateway)\n\t}\n\treturn &system.SSHConfig{\n\t\tUser: from.User,\n\t\tHost: from.Host,\n\t\tPrivateKey: from.PrivateKey,\n\t\tPort: int(from.Port),\n\t\tGatewayConfig: gw,\n\t}\n}", "func InitApi(app *iris.Application) {\n\t// var getAPI router.Party\n\n\tif config.Conf.Debug {\n\t\tppApi := app.Party(\"/debug\")\n\t\tppApi.Get(\"/pprof\", pprofHandler(pprof.Index))\n\t\tppApi.Get(\"/cmdline\", pprofHandler(pprof.Cmdline))\n\t\tppApi.Get(\"/profile\", pprofHandler(pprof.Profile))\n\t\tppApi.Post(\"/symbol\", pprofHandler(pprof.Symbol))\n\t\tppApi.Get(\"/symbol\", pprofHandler(pprof.Symbol))\n\t\tppApi.Get(\"/trace\", pprofHandler(pprof.Trace))\n\t\tppApi.Get(\"/block\", pprofHandler(pprof.Handler(\"block\").ServeHTTP))\n\t\tppApi.Get(\"/goroutine\", pprofHandler(pprof.Handler(\"goroutine\").ServeHTTP))\n\t\tppApi.Get(\"/allocs\", pprofHandler(pprof.Handler(\"allocs\").ServeHTTP))\n\t\tppApi.Get(\"/heap\", pprofHandler(pprof.Handler(\"heap\").ServeHTTP))\n\t\tppApi.Get(\"/mutex\", pprofHandler(pprof.Handler(\"mutex\").ServeHTTP))\n\t\tppApi.Get(\"/threadcreate\", pprofHandler(pprof.Handler(\"threadcreate\").ServeHTTP))\n\n\t\tgetAPI := app.Party(\"/get\")\n\t\tgetAPI.Get(\"/crontab\", ShowCrontab)\n\t}\n}", "func defConfig() Config {\n\treturn Config{\n\t\tAPI: api{\n\t\t\tServiceName: \"token-svc\",\n\t\t\tMetricsPort: \"4001\",\n\t\t\tPort: \"4000\",\n\t\t\tShutdownTimeoutSecs: 120,\n\t\t\tIdleTimeOutSecs: 90,\n\t\t\tWriteTimeOutSecs: 30,\n\t\t\tReadTimeOutSecs: 5,\n\t\t\tTimeoutSecs: 30,\n\t\t\tAllowedHeaders: []string{\"X-Requested-With\", \"X-Request-ID\", \"jaeger-debug-id\", \"Content-Type\", \"Authorization\"},\n\t\t\tAllowedOrigins: []string{\"*\"},\n\t\t\tAllowedMethods: []string{\"GET\", \"HEAD\", \"POST\", \"PUT\", \"OPTIONS\", \"DELETE\"},\n\t\t\tOpenEndPoints: []string{\"/login\", \"/health/ping\", \"/register\"},\n\t\t},\n\t\tLogger: logger{\n\t\t\tLevel: \"debug\",\n\t\t\tEncoding: \"json\",\n\t\t\tOutputPaths: []string{\"stdout\", \"/tmp/logs/tokensvc.logs\"},\n\t\t\tErrorOutputPaths: []string{\"stderr\"},\n\t\t},\n\t\tDB: db{\n\t\t\tUser: \"postgres\",\n\t\t\tPass: \"postgres\",\n\t\t\tHost: \"postgres\",\n\t\t\tPort: \"5432\",\n\t\t\tName: \"postgres\",\n\t\t\tTimeout: \"30\",\n\t\t},\n\t\tToken: token{\n\t\t\tAccessTokenLifeSpanMins: 30, // half hour\n\t\t\tRefreshTokenLifeSpanMins: 10080, // 1 week\n\t\t\tFailedLoginAttemptCacheLifeSpanMins: 30,\n\t\t\tFailedLoginAttemptsMax: 5,\n\t\t\tAuthPrivateKeyPath: \"/tmp/certs/app.rsa\", // TODO: Let's read these in from Vault\n\t\t\tAuthPublicKeyPath: \"/tmp/certs/app.rsa.pub\",\n\t\t\tIssuer: \"homerow.tech\",\n\t\t\tAccessCacheKeyID: \"token-access-user\",\n\t\t\tRefreshCacheKeyID: \"token-refresh-user\",\n\t\t\tFailedLoginCacheKeyID: \"failed-login-user\",\n\t\t},\n\t\tCookie: cookie{\n\t\t\tLifeSpanDays: 7,\n\t\t\tHashKey: \"something-that-is-32-byte-secret\",\n\t\t\tBlockKey: \"something-else-16-24-or-32secret\",\n\t\t\tName: \"homerow.tech\",\n\t\t\tDomain: \"dev.homerow.tech\",\n\t\t\tKeyUserID: \"id\",\n\t\t\tKeyEmail: \"email\",\n\t\t\tKeyJWTAccessID: \"jti-access\",\n\t\t\tKeyJWTRefreshID: \"jti-refresh\",\n\t\t},\n\t\tCache: cache{\n\t\t\tHost: \"redis\",\n\t\t\tPort: \"6379\",\n\t\t\tUserAccountLockedLifeSpanMins: 60,\n\t\t\tUserAccountLockedKeyID: \"account-locked-user\",\n\t\t},\n\t}\n}", "func NewAPI(name string) *API {\n\treturn &API{\n\t\tName: name,\n\t\tMethods: []*APIMethod{},\n\t\tAuth: \"\",\n\t}\n}", "func createTargetConfig(c TargetConfigController, recorder events.Recorder, operatorConfig *operatorv1.KubeAPIServer) (bool, error) {\n\terrors := []error{}\n\n\tdirectResourceResults := resourceapply.ApplyDirectly(c.kubeClient, c.eventRecorder, v311_00_assets.Asset,\n\t\t\"v3.11.0/kube-apiserver/ns.yaml\",\n\t\t\"v3.11.0/kube-apiserver/svc.yaml\",\n\t)\n\n\tfor _, currResult := range directResourceResults {\n\t\tif currResult.Error != nil {\n\t\t\terrors = append(errors, fmt.Errorf(\"%q (%T): %v\", currResult.File, currResult.Type, currResult.Error))\n\t\t}\n\t}\n\n\t_, _, err := manageKubeAPIServerConfig(c.kubeClient.CoreV1(), recorder, operatorConfig)\n\tif err != nil {\n\t\terrors = append(errors, fmt.Errorf(\"%q: %v\", \"configmap/config\", err))\n\t}\n\t_, _, err = managePod(c.kubeClient.CoreV1(), recorder, operatorConfig, c.targetImagePullSpec)\n\tif err != nil {\n\t\terrors = append(errors, fmt.Errorf(\"%q: %v\", \"configmap/kube-apiserver-pod\", err))\n\t}\n\t_, _, err = manageClientCABundle(c.configMapLister, c.kubeClient.CoreV1(), recorder)\n\tif err != nil {\n\t\terrors = append(errors, fmt.Errorf(\"%q: %v\", \"configmap/client-ca\", err))\n\t}\n\t_, _, err = manageAggregatorClientCABundle(c.configMapLister, c.kubeClient.CoreV1(), recorder)\n\tif err != nil {\n\t\terrors = append(errors, fmt.Errorf(\"%q: %v\", \"configmap/aggregator-client-ca\", err))\n\t}\n\t_, _, err = manageKubeletServingCABundle(c.configMapLister, c.kubeClient.CoreV1(), recorder)\n\tif err != nil {\n\t\terrors = append(errors, fmt.Errorf(\"%q: %v\", \"configmap/kubelet-serving-ca\", err))\n\t}\n\n\tif len(errors) > 0 {\n\t\tcondition := operatorv1.OperatorCondition{\n\t\t\tType: \"TargetConfigControllerFailing\",\n\t\t\tStatus: operatorv1.ConditionTrue,\n\t\t\tReason: \"SynchronizationError\",\n\t\t\tMessage: v1helpers.NewMultiLineAggregate(errors).Error(),\n\t\t}\n\t\tif _, _, err := v1helpers.UpdateStaticPodStatus(c.operatorClient, v1helpers.UpdateStaticPodConditionFn(condition)); err != nil {\n\t\t\treturn true, err\n\t\t}\n\t\treturn true, nil\n\t}\n\n\tcondition := operatorv1.OperatorCondition{\n\t\tType: \"TargetConfigControllerFailing\",\n\t\tStatus: operatorv1.ConditionFalse,\n\t}\n\tif _, _, err := v1helpers.UpdateStaticPodStatus(c.operatorClient, v1helpers.UpdateStaticPodConditionFn(condition)); err != nil {\n\t\treturn true, err\n\t}\n\n\treturn false, nil\n}" ]
[ "0.6515344", "0.64714587", "0.64564407", "0.63245904", "0.6291978", "0.62325585", "0.62114006", "0.61824024", "0.6093118", "0.60907114", "0.60671043", "0.59729344", "0.59535104", "0.5905515", "0.58848405", "0.5875292", "0.58348083", "0.58194786", "0.5817505", "0.5806318", "0.5780986", "0.5756742", "0.57444376", "0.5731038", "0.570473", "0.567622", "0.5640204", "0.561416", "0.5600705", "0.5599516", "0.55796146", "0.5570862", "0.5564371", "0.5561008", "0.5540011", "0.5524665", "0.54960215", "0.5492027", "0.5487332", "0.5483168", "0.5480619", "0.54779863", "0.54685533", "0.5458009", "0.545301", "0.54454666", "0.54241276", "0.54239607", "0.5391409", "0.5389599", "0.5389138", "0.5386071", "0.5385391", "0.5379343", "0.53744245", "0.536877", "0.536026", "0.534349", "0.53415304", "0.5338087", "0.53347236", "0.5330221", "0.5318139", "0.53165555", "0.5313034", "0.53114694", "0.5299579", "0.5294804", "0.529368", "0.52894205", "0.52864134", "0.52861726", "0.52797645", "0.5279126", "0.52751803", "0.5269655", "0.52690893", "0.5268338", "0.5263865", "0.52475965", "0.5241232", "0.5236948", "0.5230387", "0.5226441", "0.52225256", "0.5220746", "0.5218241", "0.5216905", "0.5208274", "0.52071476", "0.520709", "0.5206667", "0.5206132", "0.52020997", "0.51978177", "0.51951045", "0.5193905", "0.5190229", "0.518999", "0.5189191" ]
0.74036556
0
ParseDubboString parse the dubbo urls dubbo://192.168.3.46:20002/org.apache.dubbo.UserProvider2?anyhost=true&app.version=0.0.1&application=UserInfoServer&bean.name=UserProvider &cluster=failover&environment=dev&export=true&interface=org.apache.dubbo.UserProvider2&ip=192.168.3.46&loadbalance=random&message_size=4 &methods=GetUser&methods.GetUser.loadbalance=random&methods.GetUser.retries=1&methods.GetUser.weight=0&module=dubbogo userinfo server &name=UserInfoServer&organization=dubbo.io&pid=11037&registry.role=3&release=dubbogolang1.5.6 &service.filter=echo,token,accesslog,tps,generic_service,execute,pshutdown&side=provider&sslenabled=false&timestamp=1624716984&warmup=100
func ParseDubboString(urlString string) (config.DubboBackendConfig, []string, error) { url, err := common.NewURL(urlString) if err != nil { return config.DubboBackendConfig{}, nil, errors.WithStack(err) } return config.DubboBackendConfig{ ClusterName: url.GetParam(constant.ClusterKey, ""), ApplicationName: url.GetParam(constant.ApplicationKey, ""), Version: url.GetParam(constant.VersionKey, ""), Protocol: string(config.DubboRequest), Group: url.GetParam(constant.GroupKey, ""), Interface: url.GetParam(constant.InterfaceKey, ""), Retries: url.GetParam(constant.RetriesKey, ""), }, strings.Split(url.GetParam(constant.MethodsKey, ""), constant.StringSeparator), nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (h *MysqlHealthCheck) ParseDSN(dsn string) (cfg *DSN, err error) {\n\t// New config with some default values\n\tcfg = new(DSN)\n\n\t// [user[:password]@][net[(addr)]]/dbname[?param1=value1&paramN=valueN]\n\t// Find the last '/' (since the password or the net addr might contain a '/')\n\tfoundSlash := false\n\tfor i := len(dsn) - 1; i >= 0; i-- {\n\t\tif dsn[i] == '/' {\n\t\t\tfoundSlash = true\n\t\t\tvar j, k int\n\n\t\t\t// left part is empty if i <= 0\n\t\t\tif i > 0 {\n\t\t\t\t// [username[:password]@][protocol[(address)]]\n\t\t\t\t// Find the last '@' in dsn[:i]\n\t\t\t\tfor j = i; j >= 0; j-- {\n\t\t\t\t\tif dsn[j] == '@' {\n\t\t\t\t\t\t// username[:password]\n\t\t\t\t\t\t// Find the first ':' in dsn[:j]\n\t\t\t\t\t\tfor k = 0; k < j; k++ {\n\t\t\t\t\t\t\tif dsn[k] == ':' {\n\t\t\t\t\t\t\t\tcfg.Password = dsn[k+1 : j]\n\t\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t\tcfg.User = dsn[:k]\n\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\t// [protocol[(address)]]\n\t\t\t\t// Find the first '(' in dsn[j+1:i]\n\t\t\t\tfor k = j + 1; k < i; k++ {\n\t\t\t\t\tif dsn[k] == '(' {\n\t\t\t\t\t\t// dsn[i-1] must be == ')' if an address is specified\n\t\t\t\t\t\tif dsn[i-1] != ')' {\n\t\t\t\t\t\t\tif strings.ContainsRune(dsn[k+1:i], ')') {\n\t\t\t\t\t\t\t\treturn nil, errInvalidDSNUnescaped\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\treturn nil, errInvalidDSNAddr\n\t\t\t\t\t\t}\n\t\t\t\t\t\tcfg.Addr = dsn[k+1 : i-1]\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tcfg.Net = dsn[j+1 : k]\n\t\t\t}\n\n\t\t\t// dbname[?param1=value1&...&paramN=valueN]\n\t\t\t// Find the first '?' in dsn[i+1:]\n\t\t\tfor j = i + 1; j < len(dsn); j++ {\n\t\t\t\tif dsn[j] == '?' {\n\t\t\t\t\tif err = parseDSNParams(cfg, dsn[j+1:]); err != nil {\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\tcfg.DBName = dsn[i+1 : j]\n\n\t\t\tbreak\n\t\t}\n\t}\n\tif !foundSlash && len(dsn) > 0 {\n\t\treturn nil, errInvalidDSNNoSlash\n\t}\n\treturn\n}", "func ParseUrl(urlstring string) map[string][]string{\n values,err := url.ParseQuery(urlstring);\n if(err != nil){\n panic(\"boom\")\n }\n return values;\n}", "func unpackDatabaseUrl(url string) (urlComponents map[string]string, err error) {\n\tconnectionString, err := pq.ParseURL(url)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"Invalid database URL: %v\", url)\n\t\treturn\n\t}\n\n\turlComponents = make(map[string]string)\n\tpairs := strings.Split(connectionString, \" \")\n\tfor _, pair := range pairs {\n\t\tpairSlice := strings.Split(pair, \"=\")\n\t\turlComponents[pairSlice[0]] = pairSlice[1]\n\t}\n\n\treturn\n}", "func ParseUdnString(db *sql.DB, udn_schema map[string]interface{}, udn_value_source string) *UdnPart {\n\n\t// First Stage\n\tnext_split := _SplitQuotes(db, udn_schema, udn_value_source)\n\n\t//UdnLog(udn_schema, \"\\nSplit: Quotes: AFTER: %v\\n\\n\", next_split)\n\n\tnext_split = _SplitCompoundStatements(db, udn_schema, next_split)\n\n\t//UdnLog(udn_schema, \"\\nSplit: Compound: AFTER: %v\\n\\n\", next_split)\n\n\tnext_split = _SplitStatementLists(db, udn_schema, next_split)\n\n\t//UdnLog(udn_schema, \"\\nSplit: List: AFTER: %v\\n\\n\", next_split)\n\n\t// Forth Stage\n\tnext_split = _SplitStatementMaps(db, udn_schema, next_split)\n\n\t//UdnLog(udn_schema, \"\\nSplit: Compound: Map: %v\\n\\n\", next_split)\n\n\t// Fifth Stage\n\tnext_split = _SplitStatementMapKeyValues(db, udn_schema, next_split)\n\n\t//UdnLog(udn_schema, \"\\nSplit: Compound: Map Key Values: %v\\n\\n\", next_split)\n\n\t// Put it into a structure now -- UdnPart\n\t//\n\tudn_start := CreateUdnPartsFromSplit_Initial(db, udn_schema, next_split)\n\n\t////output := DescribeUdnPart(&udn_start)\n\t////UdnLog(udn_schema, \"\\n===== 0 - Description of UDN Part:\\n\\n%s\\n===== 0 - END\\n\", output)\n\n\t// Put it into a structure now -- UdnPart\n\t//\n\tFinalParseProcessUdnParts(db, udn_schema, &udn_start)\n\n\t//output := DescribeUdnPart(&udn_start)\n\t//UdnLogLevel(nil, log_trace, \"\\n===== 1 - Description of UDN Part:\\n\\n%s\\n===== 1 - END\\n\", output)\n\n\treturn &udn_start\n}", "func parseDataURL(url string, du *dataURL) error {\n\n\t// Check prefix\n\tif !isDataURL(url) {\n\t\treturn fmt.Errorf(\"specified string is not a data URL\")\n\t}\n\n\t// Separate header from data\n\tbody := url[len(dataURLprefix):]\n\tparts := strings.Split(body, \",\")\n\tif len(parts) != 2 {\n\t\treturn fmt.Errorf(\"data URI contains more than one ','\")\n\t}\n\tdu.Data = parts[1]\n\n\t// Separate media type from optional encoding\n\tres := strings.Split(parts[0], \";\")\n\tdu.MediaType = res[0]\n\tif len(res) < 2 {\n\t\treturn nil\n\t}\n\tif len(res) >= 2 {\n\t\tdu.Encoding = res[1]\n\t}\n\treturn nil\n}", "func SplitConnStr(s string) (driver string, source string, implementation Implementation, err error) {\n\t// consider https://github.com/xo/dburl if this ends up lacking\n\tparts := strings.SplitN(s, \"://\", 2)\n\tif len(parts) != 2 {\n\t\treturn \"\", \"\", Unknown, fmt.Errorf(\"could not parse DB URL %s\", s)\n\t}\n\tdriver = parts[0]\n\tsource = parts[1]\n\timplementation = ImplementationForScheme(parts[0])\n\n\tswitch implementation {\n\tcase Postgres:\n\t\tsource = s // postgres wants full URLS for its DSN\n\t\tdriver = \"pgx\"\n\tcase Cockroach:\n\t\tsource = s // cockroach wants full URLS for its DSN\n\t\tdriver = \"pgxcockroach\"\n\t}\n\treturn driver, source, implementation, nil\n}", "func ParseDSN(dsn string) (cfg *DSN, err error) {\n\t// New config with some default values\n\tcfg = NewDSN()\n\n\t// [user[:password]@][net[(addr)]]/dbname[?param1=value1&paramN=valueN]\n\t// Find the last '/' (since the password or the net addr might contain a '/')\n\tfoundSlash := false\n\tfor i := len(dsn) - 1; i >= 0; i-- {\n\t\tif dsn[i] == '/' {\n\t\t\tfoundSlash = true\n\t\t\tvar j, k int\n\n\t\t\t// left part is empty if i <= 0\n\t\t\tif i > 0 {\n\t\t\t\t// [username[:password]@][protocol[(address)]]\n\t\t\t\t// Find the last '@' in dsn[:i]\n\t\t\t\tfor j = i; j >= 0; j-- {\n\t\t\t\t\tif dsn[j] == '@' {\n\t\t\t\t\t\t// username[:password]\n\t\t\t\t\t\t// Find the first ':' in dsn[:j]\n\t\t\t\t\t\tfor k = 0; k < j; k++ {\n\t\t\t\t\t\t\tif dsn[k] == ':' {\n\t\t\t\t\t\t\t\tcfg.Passwd = dsn[k+1: j]\n\t\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t\tcfg.User = dsn[:k]\n\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\t// [protocol[(address)]]\n\t\t\t\t// Find the first '(' in dsn[j+1:i]\n\t\t\t\tfor k = j + 1; k < i; k++ {\n\t\t\t\t\tif dsn[k] == '(' {\n\t\t\t\t\t\t// dsn[i-1] must be == ')' if an address is specified\n\t\t\t\t\t\tif dsn[i-1] != ')' {\n\t\t\t\t\t\t\tif strings.ContainsRune(dsn[k+1:i], ')') {\n\t\t\t\t\t\t\t\treturn nil, errInvalidDSNUnescaped\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\treturn nil, errInvalidDSNAddr\n\t\t\t\t\t\t}\n\t\t\t\t\t\tcfg.Addr = dsn[k+1: i-1]\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tcfg.Net = dsn[j+1: k]\n\t\t\t}\n\n\t\t\t// dbname[?param1=value1&...&paramN=valueN]\n\t\t\t// Find the first '?' in dsn[i+1:]\n\t\t\tfor j = i + 1; j < len(dsn); j++ {\n\t\t\t\tif dsn[j] == '?' {\n\t\t\t\t\tif err = parseDSNParams(cfg, dsn[j+1:]); err != nil {\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\tcfg.DBName = dsn[i+1: j]\n\n\t\t\tbreak\n\t\t}\n\t}\n\tif !foundSlash && len(dsn) > 0 {\n\t\treturn nil, errInvalidDSNNoSlash\n\t}\n\treturn\n}", "func parseConnectionStringURL(connectionString string) (sqlInstance string, databaseName string) {\n\tdatabaseName = emptyDatabaseName\n\n\tu, err := url.Parse(connectionString)\n\tif err != nil {\n\t\treturn emptySQLInstance, emptyDatabaseName\n\t}\n\n\tsqlInstance = u.Hostname()\n\tif len(u.Path) > 1 {\n\t\t// There was a SQL instance name specified in addition to the host\n\t\t// E.g. \"the.host.com:1234/InstanceName\" or \"the.host.com/InstanceName\"\n\t\tsqlInstance = sqlInstance + \"\\\\\" + u.Path[1:]\n\t}\n\n\tquery := u.Query()\n\tfor key, value := range query {\n\t\tif strings.EqualFold(\"database\", key) {\n\t\t\tdatabaseName = value[0]\n\t\t\tbreak\n\t\t}\n\t}\n\n\treturn sqlInstance, databaseName\n}", "func ParseUname(uname string) (b Bck, objName string) {\n\tvar prev, itemIdx int\n\tfor i := 0; i < len(uname); i++ {\n\t\tif uname[i] != filepath.Separator {\n\t\t\tcontinue\n\t\t}\n\n\t\titem := uname[prev:i]\n\t\tswitch itemIdx {\n\t\tcase 0:\n\t\t\tb.Provider = item\n\t\tcase 1:\n\t\t\tb.Ns = ParseNsUname(item)\n\t\tcase 2:\n\t\t\tb.Name = item\n\t\t\tobjName = uname[i+1:]\n\t\t\treturn\n\t\t}\n\n\t\titemIdx++\n\t\tprev = i + 1\n\t}\n\treturn\n}", "func ParseUname(uname string) (b Bck, objName string) {\n\tvar prev, itemIdx int\n\tfor i := 0; i < len(uname); i++ {\n\t\tif uname[i] != filepath.Separator {\n\t\t\tcontinue\n\t\t}\n\n\t\titem := uname[prev:i]\n\t\tswitch itemIdx {\n\t\tcase 0:\n\t\t\tb.Provider = item\n\t\tcase 1:\n\t\t\tb.Ns = ParseNsUname(item)\n\t\tcase 2:\n\t\t\tb.Name = item\n\t\t\tobjName = uname[i+1:]\n\t\t\treturn\n\t\t}\n\n\t\titemIdx++\n\t\tprev = i + 1\n\t}\n\treturn\n}", "func SmbScriptParse(output *string) *SMBOSDiscovery {\n\tarr := strings.Split(*output, \"\\n \")[1:]\n\tbuf := SMBOSDiscovery{}\n\tfor _, item := range arr {\n\t\tif match, _ := regexp.MatchString(`OS: ([a-zA-Z0-9]+) ([\\(\\)\\a-zA-Z0-9\\.]+)`, item); match {\n\t\t\tmatch, _ := regexp.Compile(`OS: ([a-zA-Z0-9]+) ([\\(\\)\\a-zA-Z0-9\\.]+)`)\n\t\t\tresult := match.FindStringSubmatch(item)\n\t\t\tif len(result) == 3 {\n\t\t\t\tbuf.OS = result[1]\n\t\t\t\tbuf.LanManager = result[2]\n\t\t\t}\n\t\t} else if match, _ := regexp.MatchString(`Computer name: ([a-zA-Z0-9\\.\\-]+)`, item); match {\n\t\t\tmatch, _ := regexp.Compile(`Computer name: ([a-zA-Z0-9\\.\\-]+)`)\n\t\t\tresult := match.FindStringSubmatch(item)\n\t\t\tif len(result) == 2 {\n\t\t\t\tbuf.NetworkName = result[1]\n\t\t\t}\n\t\t} else if match, _ := regexp.MatchString(`NetBIOS computer name: ([a-zA-Z0-9\\.\\-]+)`, item); match {\n\t\t\tmatch, _ := regexp.Compile(`NetBIOS computer name: ([a-zA-Z0-9\\.\\-]+)`)\n\t\t\tresult := match.FindStringSubmatch(item)\n\t\t\tif len(result) == 2 {\n\t\t\t\tbuf.NetBIOS = result[1]\n\t\t\t}\n\t\t} else if match, _ := regexp.MatchString(`FQDN: ([a-zA-Z0-9\\.\\-]+)`, item); match {\n\t\t\tmatch, _ := regexp.Compile(`FQDN: ([a-zA-Z0-9\\.\\-]+)`)\n\t\t\tresult := match.FindStringSubmatch(item)\n\t\t\tif len(result) == 2 {\n\t\t\t\tbuf.FQDN = result[1]\n\t\t\t}\n\t\t} else if match, _ := regexp.MatchString(`OS CPE: ([a-zA-Z0-9\\.\\-\\:\\/]+)`, item); match {\n\t\t\tmatch, _ := regexp.Compile(`OS CPE: ([a-zA-Z0-9\\.\\-\\:\\/\\_]+)`)\n\t\t\tresult := match.FindStringSubmatch(item)\n\t\t\tif len(result) == 2 {\n\t\t\t\tbuf.CPE = result[1]\n\t\t\t}\n\t\t}\n\t}\n\treturn &buf\n}", "func parse(b []byte) *Message {\n var servername, nick, user, host string\n var command, target, msg string\n words := bytes.Split(b, bytes.NewBufferString(\" \").Bytes())\n\n if len(words) >= 4 {\n if match, _ := regexp.Match(\"^:\", words[0]); match {\n if match, _ := regexp.Match(\"!|@\", words[0]); match {\n i := 1\n for words[0][i] != '!' { i++ }\n nick = bytes.NewBuffer(words[0][1:i]).String()\n j := i+1\n for words[0][j] != '@' { j++ }\n var wordstart int = i + 1\n if words[0][i+1] == '~' {\n wordstart = i+2\n }\n\n user = bytes.NewBuffer(words[0][wordstart:j]).String()\n k := j+1\n host = bytes.NewBuffer(words[0][k:len(words[0])]).String()\n } else {\n servername = bytes.NewBuffer(words[0][1:len(words[0])]).String()\n }\n }\n command = bytes.NewBuffer(words[1]).String()\n target = bytes.NewBuffer(words[2]).String()\n str := bytes.Join(words[3:len(words)], bytes.NewBufferString(\" \").Bytes())\n msg = bytes.NewBuffer(str[1:len(str)]).String()\n } else {\n if match, _ := regexp.Match(\"PING\", words[0]); match {\n command = \"PING\"\n host= bytes.NewBuffer(words[1][1:len(words[1])]).String()\n fmt.Println(host)\n }\n }\n\n return &Message{\n Servername: servername,\n Nickname: nick,\n Username: user,\n Hostname: host,\n Command: command,\n Target: target,\n Message: msg,\n }\n}", "func parseNodesString(nodes string, proto string) (parsed map[string]string) {\n\tvar nodeContactLen int\n\tif proto == \"udp4\" {\n\t\tnodeContactLen = v4nodeContactLen\n\t} else if proto == \"udp6\" {\n\t\tnodeContactLen = v6nodeContactLen\n\t} else {\n\t\treturn\n\t}\n\tparsed = make(map[string]string)\n\tif len(nodes)%nodeContactLen > 0 {\n\t\tlogger.Infof(\"DHT: len(NodeString) = %d, INVALID LENGTH, should be a multiple of %d\", len(nodes), nodeContactLen)\n\t\tlogger.Infof(\"%T %#v\\n\", nodes, nodes)\n\t\treturn\n\t} else {\n\t\tlogger.Infof(\"DHT: len(NodeString) = %d, had %d nodes, nodeContactLen=%d\\n\", len(nodes), len(nodes)/nodeContactLen, nodeContactLen)\n\t}\n\tfor i := 0; i < len(nodes); i += nodeContactLen {\n\t\tid := nodes[i : i+nodeIdLen]\n\t\taddress := nettools.BinaryToDottedPort(nodes[i+nodeIdLen : i+nodeContactLen])\n\t\tparsed[id] = address\n\t}\n\treturn\n\n}", "func ParseConnString(s string) (ConnectionParams, error) { return dsn.Parse(s) }", "func url_parser(url_string string) url.Values{\n\tparams, err := url.ParseQuery(url_string)\n\tif err != nil{\n\t\tfmt.Println(err)\n\t}\n\treturn params\n}", "func parseDSNParams(cfg *DSN, params string) (err error) {\n\tfor _, v := range strings.Split(params, \"&\") {\n\t\tparam := strings.SplitN(v, \"=\", 2)\n\t\tif len(param) != 2 {\n\t\t\tcontinue\n\t\t}\n\t\t// lazy init\n\t\tif cfg.Params == nil {\n\t\t\tcfg.Params = make(map[string]string)\n\t\t}\n\t\tvalue := param[1]\n\t\tif cfg.Params[param[0]], err = url.QueryUnescape(value); err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\treturn\n}", "func ParseVideoUrl(video *Video) {\n\tdefer wg.Done()\n\tres, err := http.Get(video.Url)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn\n\t}\n\tdefer res.Body.Close()\n\tbytes, err := ioutil.ReadAll(res.Body)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn\n\t}\n\treg := `videoFlashPlayUrl\\s=\\s\\'(.*)\\'`\n\tsubmatch := regexp.MustCompile(reg).FindAllSubmatch(bytes, -1)\n\tif len(submatch) > 0 && len(submatch[0]) > 0 {\n\t\tvv := submatch[0][1]\n\t\turlInfo, err := url.Parse(string(vv))\n\t\tif err != nil {\n\t\t\tlog.Println(err, string(vv))\n\t\t\treturn\n\t\t}\n\t\tlog.Println(urlInfo.Query()[\"video\"][0])\n\t\tvideo.VideoUrl = urlInfo.Query()[\"video\"][0]\n\t\tgo DownloadVideo(video)\n\t\twg.Add(1)\n\t\t// videoChan <- video\n\t}\n}", "func parseUrl(ignore string, urlPath string) map[string]string {\n\turlMap := make(map[string]string)\n\tfor _, s := range strings.Split(urlPath,\"/\") {\n\t\tif s == \"\" || s == ignore {\n\t\t\tcontinue\n\t\t}\n\t\tspl := strings.Split(s, \"=\")\n\t\tif len(spl) != 2 {\n\t\t\treturn nil\n\t\t}\n\t\turlMap[spl[0]] = spl[1]\n\t}\n\treturn urlMap\n}", "func getBzzAddrFromHost(client *rpc.Client) (string, error) {\n\tvar hive string\n\n\terr := client.Call(&hive, \"bzz_hive\")\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// we make an ugly assumption about the output format of the hive.String() method\n\t// ideally we should replace this with an API call that returns the bzz addr for a given host,\n\t// but this also works for now (provided we don't change the hive.String() method, which we haven't in some time\n\tss := strings.Split(strings.Split(hive, \"\\n\")[3], \" \")\n\treturn ss[len(ss)-1], nil\n}", "func parseIPAddressInPxctlServiceKvdbMembers(kvdbMembersOutput string) ([]string, error) {\n\tkvdbMemberIPs := []string{}\n\tfor _, line := range strings.Split(kvdbMembersOutput, \"\\n\") {\n\t\tif strings.Contains(line, \"http\") {\n\t\t\tcols := strings.Fields(strings.TrimSpace(line))\n\t\t\tif len(cols) >= 2 {\n\t\t\t\t// Parse out <ip>:<port> from URL \"http(s)://<ip>:<port>\" in the line\n\t\t\t\tendPt := kvdbEndPtsRgx.FindSubmatch([]byte(strings.Trim(cols[2], \"[]\")))\n\t\t\t\tip, _, err := net.SplitHostPort(string(bytes.TrimSpace(endPt[1])))\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn kvdbMemberIPs, fmt.Errorf(\"Member parse error %v\", err)\n\t\t\t\t}\n\t\t\t\tkvdbMemberIPs = append(kvdbMemberIPs, ip)\n\t\t\t}\n\t\t}\n\t}\n\treturn kvdbMemberIPs, nil\n}", "func TestUrlParsing(t *testing.T) {\n\n\tu, _ := url.ParseRequestURI(strings.TrimPrefix(\n\t\t\"jdbc:mysql://mysql.default.svc.cluster.local:3306/vamp-${parent}?test=1&test2=2&useSSL=false\", \"jdbc:\"))\n\n\tassert.Equal(t, \"test=1&test2=2&useSSL=false\", u.Query().Encode())\n\n\tu, _ = url.ParseRequestURI(strings.TrimPrefix(\n\t\t\"jdbc:mysql://mysql.default.svc.cluster.local:3306/vamp-${parent}\", \"jdbc:\"))\n\n\tassert.Equal(t, \"\", u.Query().Encode())\n\n}", "func parseBroadcastUrl(broadcastUrl string) (string, error) {\n\tfileURL, err := url.Parse(broadcastUrl)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tpath := fileURL.Path\n\tsegments := strings.Split(path, \"/\")\n\tfileName := segments[len(segments)-1]\n\n\treturn fileName, nil\n}", "func CreateConfigFromURL(url string ) (*Config, error) {\n\targuments, err := plugins.ExtractArguments(url)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif len(arguments) < 3 {\n\t\tfmt.Println(arguments)\n\t\treturn nil, errors.New(string(NotEnoughArguments))\n\t}\n\n\tif len(arguments) < 4 {\n\t\treturn &Config{\n\t\t\tBotName: DefaultUser,\n\t\t\tToken: Token{\n\t\t\t\tA: arguments[0],\n\t\t\t\tB: arguments[1],\n\t\t\t\tC: arguments[2],\n\t\t\t},\n\t\t}, nil\n\t}\n\n\treturn &Config{\n\t\tBotName: arguments[0],\n\t\tToken: Token{\n\t\t\tA: arguments[1],\n\t\t\tB: arguments[2],\n\t\t\tC: arguments[3],\n\t\t},\n\t}, nil\n}", "func getBzzAddrFromHost(client *rpc.Client) (string, error) {\n\tvar hive string\n\n\terr := client.Call(&hive, \"bzz_hive\")\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// we make an ugly assumption about the output format of the hive.String() method\n\t// ideally we should replace this with an API call that returns the bzz addr for a given host,\n\t// but this also works for now (provided we don't change the hive.String() method, which we haven't in some time\n\treturn strings.Split(strings.Split(hive, \"\\n\")[3], \" \")[10], nil\n}", "func GetDubboGoMetadata(dubboGoMetadata string) *structpb.Struct {\n\treturn &structpb.Struct{\n\t\tFields: map[string]*structpb.Value{\n\t\t\tconstant.XDSMetadataClusterIDKey: {\n\t\t\t\t// Set cluster id to Kubernetes to ensure dubbo-go's xds client can get service\n\t\t\t\t// istiod.istio-system.svc.cluster.local's\n\t\t\t\t// pods ip from istiod by eds, to call no-endpoint port of istio like 8080\n\t\t\t\tKind: &structpb.Value_StringValue{StringValue: constant.XDSMetadataDefaultDomainName},\n\t\t\t},\n\t\t\tconstant.XDSMetadataLabelsKey: {\n\t\t\t\tKind: &structpb.Value_StructValue{StructValue: &structpb.Struct{\n\t\t\t\t\tFields: map[string]*structpb.Value{\n\t\t\t\t\t\tconstant.XDSMetadataDubboGoMapperKey: {\n\t\t\t\t\t\t\tKind: &structpb.Value_StringValue{StringValue: dubboGoMetadata},\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t}},\n\t\t\t},\n\t\t},\n\t}\n}", "func parseDialingAddress(ctx *context.T, vaddress string) (network string, address string, tag string, p flow.Protocol, err error) {\n\tparts := strings.SplitN(vaddress, \"/\", 3)\n\tif len(parts) != 3 {\n\t\treturn \"\", \"\", \"\", nil, ErrorfInvalidAddress(ctx, \"invalid vine address %v, address must be of the form 'network/address/tag'\", vaddress)\n\t}\n\tp, _ = flow.RegisteredProtocol(parts[0])\n\tif p == nil {\n\t\treturn \"\", \"\", \"\", nil, ErrorfNoRegisteredProtocol(ctx, \"no registered protocol: %v\", parts[0])\n\t}\n\treturn parts[0], parts[1], parts[2], p, nil\n}", "func (e *extractor) Extract(url string, option types.Options) ([]*types.Data, error) {\n\tif !strings.Contains(url, \"m.weibo.cn\") {\n\t\tif strings.Contains(url, \"weibo.com/tv/show/\") {\n\t\t\treturn downloadWeiboTV(url)\n\t\t} else if strings.Contains(url, \"video.h5.weibo.cn\") {\n\t\t\treturn downloadWeiboVideo(url)\n\t\t}\n\t\turl = strings.Replace(url, \"weibo.com\", \"m.weibo.cn\", 1)\n\t}\n\thtml, err := request.Get(url, url, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ttitles := utils.MatchOneOf(\n\t\thtml, `\"content2\": \"(.+?)\",`, `\"status_title\": \"(.+?)\",`,\n\t)\n\tif titles == nil || len(titles) < 2 {\n\t\treturn nil, types.ErrURLParseFailed\n\t}\n\ttitle := titles[1]\n\n\trealURLs := utils.MatchOneOf(\n\t\thtml, `\"stream_url_hd\": \"(.+?)\"`, `\"stream_url\": \"(.+?)\"`,\n\t)\n\tif realURLs == nil || len(realURLs) < 2 {\n\t\treturn nil, types.ErrURLParseFailed\n\t}\n\trealURL := realURLs[1]\n\n\tsize, err := request.Size(realURL, url)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\turlData := &types.Part{\n\t\tURL: realURL,\n\t\tSize: size,\n\t\tExt: \"mp4\",\n\t}\n\tstreams := map[string]*types.Stream{\n\t\t\"default\": {\n\t\t\tParts: []*types.Part{urlData},\n\t\t\tSize: size,\n\t\t},\n\t}\n\n\treturn []*types.Data{\n\t\t{\n\t\t\tSite: \"微博 weibo.com\",\n\t\t\tTitle: title,\n\t\t\tType: types.DataTypeVideo,\n\t\t\tStreams: streams,\n\t\t\tURL: url,\n\t\t},\n\t}, nil\n}", "func ParseUrl(url string)(ip string,port int,path string,err error){\n\t// ipex := regexp.MustCompile(`http://(.*?)(:(\\d{2,4}))?(/.*)`)\n\tegex := regexp.MustCompile(`http://(\\d+\\.\\d+\\.\\d+\\.\\d+)(:(\\d{2,4}))?(/.*)`)\n\tregex := egex.FindAllStringSubmatch(url,-1)\n\tif len(regex) != 1 {\n\t\treturn ip,port,path,errors.New(\"url is invalid\")\n\t}\n\turlPar := regex[0]\n\tvar oport string\n\tip,oport,path = urlPar[1],urlPar[3],urlPar[4]\n\tlog.Println(urlPar)\n\tif oport == \"\" {\n\t\toport = \"80\"\n\t}\n\tport,err = strconv.Atoi(oport) \n\treturn\n}", "func ParseDeleteaspecificBannedUserResponse(rsp *http.Response) (*DeleteaspecificBannedUserResponse, error) {\n\tbodyBytes, err := ioutil.ReadAll(rsp.Body)\n\tdefer rsp.Body.Close()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresponse := &DeleteaspecificBannedUserResponse{\n\t\tBody: bodyBytes,\n\t\tHTTPResponse: rsp,\n\t}\n\n\tswitch {\n\t}\n\n\treturn response, nil\n}", "func parseHostPort(str string) (string, string) {\n\tvar (\n\t\thost string\n\t\tport string\n\n\t\ti = strings.Index(str, \":\")\n\t)\n\tif i == -1 {\n\t\treturn str, \"\"\n\t}\n\n\thost = str[:i]\n\tport = str[i+1:]\n\n\treturn host, port\n}", "func (c *Config) getBootInfoFromUrl(url string) (bootInfo BootInfo, err error) {\n\tconst ignitionPath = \"/ignition?\"\n\tconst ignitionPxePath = \"/ignition-pxe?\"\n\n\t// split after the base url path to find the parameters\n\t// we have two possible routes set in constant above\n\tquery := strings.SplitAfter(url, ignitionPxePath)\n\tif len(query) != 2 {\n\t\tglog.Warningf(\"incoherent SplitAfter %s, trying %s\", ignitionPxePath, ignitionPath)\n\t\tquery = strings.SplitAfter(url, ignitionPath)\n\t\tif len(query) != 2 {\n\t\t\treturn bootInfo, errors.New(\n\t\t\t\tfmt.Sprintf(\"incoherent SplitAfter(url, %s) || SplitAfter(url, %s)) != 2\", ignitionPath, ignitionPxePath))\n\t\t}\n\t}\n\targs := strings.Split(query[1], \"&\")\n\tfor _, arg := range args {\n\t\tif strings.Contains(arg, uuidField) {\n\t\t\tbootInfo.Uuid = strings.Split(arg, uuidField)[1]\n\t\t\tglog.V(2).Infof(\"uuid: %q\", bootInfo.Uuid)\n\t\t}\n\t\tif strings.Contains(arg, macField) {\n\t\t\tbootInfo.Mac = strings.Split(arg, macField)[1]\n\t\t\tbootInfo.Mac = strings.Replace(bootInfo.Mac, \"-\", \":\", -1)\n\t\t\tglog.V(2).Infof(\"mac: %q\", bootInfo.Mac)\n\t\t}\n\t}\n\tbootInfo.RandomId, err = c.getRandomId()\n\tif err != nil {\n\t\tglog.Errorf(\"fail to get RandomId: %s\", err)\n\t\treturn bootInfo, err\n\t}\n\treturn bootInfo, nil\n}", "func parseIPAddressInPxctlServiceKvdbEndpoints(kvdbEndpointsOutput string) ([]string, error) {\n\t// Parse out all <ip>:<port> strings from URLs \"http(s)://<ip>:<port>\" in the line\n\tkvdbEndPts := kvdbEndPtsRgx.FindAllSubmatch([]byte(kvdbEndpointsOutput), -1)\n\n\tkvdbEndPtsIPs := []string{}\n\tfor _, endPt := range kvdbEndPts {\n\t\tip, _, err := net.SplitHostPort(string(bytes.TrimSpace(endPt[1])))\n\t\tif err != nil {\n\t\t\treturn kvdbEndPtsIPs, fmt.Errorf(\"Endpoint parse error %v\", err)\n\t\t}\n\t\tkvdbEndPtsIPs = append(kvdbEndPtsIPs, ip)\n\t}\n\treturn kvdbEndPtsIPs, nil\n}", "func CBAuthURL(urlStr string) (string, error) {\n\tu, err := url.Parse(urlStr)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tcbUser, cbPasswd, err := cbauth.GetHTTPServiceAuth(u.Host)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tu.User = url.UserPassword(cbUser, cbPasswd)\n\n\treturn u.String(), nil\n}", "func ParseURL(u string) (URLParts, error) {\n\treturn sas.ParseURL(u)\n}", "func ParseDSN(dsn string) (*Config, error) {\n\tsub := reDSN.FindStringSubmatch(dsn)\n\tif len(sub) != 5 {\n\t\treturn nil, fmt.Errorf(`dsn %s doesn't match pop_access_id:pop_access_secret@pop_url?params`, dsn)\n\t}\n\tpid, ps, purl := sub[1], sub[2], sub[3]\n\n\tkvs, err := url.ParseQuery(sub[4])\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\trequiredParameter := []string{\"env\", \"with\", \"curr_project\"}\n\tfor _, k := range requiredParameter {\n\t\tv := kvs.Get(k)\n\t\tif v == \"\" {\n\t\t\treturn nil, fmt.Errorf(`dsn is missing required parameter %s`, k)\n\t\t}\n\t}\n\n\tenv, err := decodeJSONB64(kvs.Get(\"env\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\twith, err := decodeJSONB64(kvs.Get(\"with\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tscheme := kvs.Get(\"scheme\")\n\tif len(scheme) == 0 {\n\t\tscheme = \"http\"\n\t}\n\tverbose := kvs.Get(\"verbose\") == \"true\"\n\tproject := kvs.Get(\"curr_project\")\n\n\treturn &Config{POPAccessID: pid, POPAccessSecret: ps, POPURL: purl, Env: env, With: with, Verbose: verbose, Project: project, POPScheme: scheme}, nil\n}", "func parseConnection(dsn string, tlsInfo tls.Config) (*Config, error) {\n\tconfig := &Config{\n\t\tslowThreshold: defaultSlowMethod,\n\t\trevHistory: defaultRevHistory,\n\t\tbucket: defaultBucket,\n\t\treplicas: defaultReplicas,\n\t}\n\n\t// Parse the first URL in the connection string which contains the\n\t// query parameters.\n\tconnections := strings.Split(dsn, \",\")\n\tu, err := url.Parse(connections[0])\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Extract the host and port if embedded server is used.\n\tconfig.host = u.Hostname()\n\tif u.Port() != \"\" {\n\t\tconfig.port, _ = strconv.Atoi(u.Port())\n\t}\n\n\t// Extract the query parameters to build configuration.\n\tqueryMap, err := url.ParseQuery(u.RawQuery)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif v := queryMap.Get(\"bucket\"); v != \"\" {\n\t\tconfig.bucket = v\n\t}\n\n\tif v := queryMap.Get(\"replicas\"); v != \"\" {\n\t\tif r, err := strconv.ParseUint(v, 10, 8); err == nil {\n\t\t\tif r >= 1 && r <= 5 {\n\t\t\t\tconfig.replicas = int(r)\n\t\t\t} else {\n\t\t\t\treturn nil, fmt.Errorf(\"invalid replicas, must be >= 1 and <= 5\")\n\t\t\t}\n\t\t}\n\t}\n\n\tif d := queryMap.Get(\"slowMethod\"); d != \"\" {\n\t\tif dur, err := time.ParseDuration(d); err == nil {\n\t\t\tconfig.slowThreshold = dur\n\t\t} else {\n\t\t\treturn nil, fmt.Errorf(\"invalid slowMethod duration: %w\", err)\n\t\t}\n\t}\n\n\tif r := queryMap.Get(\"revHistory\"); r != \"\" {\n\t\tif revs, err := strconv.ParseUint(r, 10, 8); err == nil {\n\t\t\tif revs >= 2 && revs <= 64 {\n\t\t\t\tconfig.revHistory = uint8(revs)\n\t\t\t} else {\n\t\t\t\treturn nil, fmt.Errorf(\"invalid revHistory, must be >= 2 and <= 64\")\n\t\t\t}\n\t\t}\n\t}\n\n\tif tlsInfo.KeyFile != \"\" && tlsInfo.CertFile != \"\" {\n\t\tconfig.clientOptions = append(config.clientOptions, nats.ClientCert(tlsInfo.CertFile, tlsInfo.KeyFile))\n\t}\n\n\tif tlsInfo.CAFile != \"\" {\n\t\tconfig.clientOptions = append(config.clientOptions, nats.RootCAs(tlsInfo.CAFile))\n\t}\n\n\tif f := queryMap.Get(\"contextFile\"); f != \"\" {\n\t\tif u.Host != \"\" {\n\t\t\treturn config, fmt.Errorf(\"when using context endpoint no host should be provided\")\n\t\t}\n\n\t\tlogrus.Debugf(\"loading nats context file: %s\", f)\n\n\t\tnatsContext, err := natscontext.NewFromFile(f)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tconnections = strings.Split(natsContext.ServerURL(), \",\")\n\n\t\t// command line options provided to kine will override the file\n\t\t// https://github.com/nats-io/jsm.go/blob/v0.0.29/natscontext/context.go#L257\n\t\t// allows for user, creds, nke, token, certifcate, ca, inboxprefix from the context.json\n\t\tnatsClientOpts, err := natsContext.NATSOptions(config.clientOptions...)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tconfig.clientOptions = natsClientOpts\n\t}\n\n\tconnBuilder := strings.Builder{}\n\tfor idx, c := range connections {\n\t\tif idx > 0 {\n\t\t\tconnBuilder.WriteString(\",\")\n\t\t}\n\n\t\tu, err := url.Parse(c)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tif u.Scheme != \"nats\" {\n\t\t\treturn nil, fmt.Errorf(\"invalid connection string=%s\", c)\n\t\t}\n\n\t\tconnBuilder.WriteString(\"nats://\")\n\n\t\tif u.User != nil && idx == 0 {\n\t\t\tuserInfo := strings.Split(u.User.String(), \":\")\n\t\t\tif len(userInfo) > 1 {\n\t\t\t\tconfig.clientOptions = append(config.clientOptions, nats.UserInfo(userInfo[0], userInfo[1]))\n\t\t\t} else {\n\t\t\t\tconfig.clientOptions = append(config.clientOptions, nats.Token(userInfo[0]))\n\t\t\t}\n\t\t}\n\t\tconnBuilder.WriteString(u.Host)\n\t}\n\n\tconfig.clientURL = connBuilder.String()\n\n\t// Config options only relevant if built with embedded NATS.\n\tif natsserver.Embedded {\n\t\tconfig.noEmbed = queryMap.Has(\"noEmbed\")\n\t\tconfig.serverConfig = queryMap.Get(\"serverConfig\")\n\t\tconfig.stdoutLogging = queryMap.Has(\"stdoutLogging\")\n\t\tconfig.dontListen = queryMap.Has(\"dontListen\")\n\t}\n\n\tlogrus.Debugf(\"using config %#v\", config)\n\n\treturn config, nil\n}", "func QueryStringParser(queryStr string, filters map[string]string) []FilteredResult {\n\t//define custom map type to allowduplicate keys\n\ttype Map struct {\n\t\tKey string\n\t\tValue string\n\t}\n\n\tparams := []Map{}\n\tsearchFilters := []FilteredResult{}\n\n\tparts := strings.Split(queryStr, \"&\")\n\n\t//build a key/value map of the querystring by\n\t//storing the query as key and the fragment as the value\n\tfor _, part := range parts {\n\t\tsplit := strings.Split(part, \"=\")\n\n\t\tif len(split) > 1 && split[1] != \"\" {\n\t\t\tparams = append(params, Map{\n\t\t\t\tKey: split[0],\n\t\t\t\tValue: split[1],\n\t\t\t})\n\t\t} else {\n\t\t\tparams = append(params, Map{\n\t\t\t\tKey: split[0],\n\t\t\t\tValue: \"\",\n\t\t\t})\n\t\t}\n\t}\n\n\t//\n\tfor _, param := range params {\n\t\tfor name, varType := range filters {\n\t\t\tif param.Key == name {\n\t\t\t\tesc, _ := url.QueryUnescape(param.Value)\n\t\t\t\tparseValue, operator, condition := RHSParser(esc, varType)\n\n\t\t\t\tsearchFilters = append(searchFilters, FilteredResult{\n\t\t\t\t\tField: param.Key,\n\t\t\t\t\tType: varType,\n\t\t\t\t\tValue: parseValue,\n\t\t\t\t\tOperator: operator,\n\t\t\t\t\tCondition: condition,\n\t\t\t\t})\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\n\treturn searchFilters\n}", "func parseUpstream(upstream string) (result Upstream, err error) {\n\tif strings.Trim(upstream, \" \") == \"\" {\n\t\treturn Upstream{}, nil\n\t}\n\n\tparts := strings.Split(upstream, \":\")\n\n\tif len(parts) < 2 || len(parts) > 3 {\n\t\terr = fmt.Errorf(\"wrong configuration, couldn't parse input '%s', please enter net:host[:port]\", upstream)\n\t\treturn\n\t}\n\n\tnet := strings.TrimSpace(parts[0])\n\n\tif _, ok := netDefaultPort[net]; !ok {\n\t\terr = fmt.Errorf(\"wrong configuration, couldn't parse net '%s', please user one of %s\",\n\t\t\tnet, reflect.ValueOf(netDefaultPort).MapKeys())\n\t\treturn\n\t}\n\n\tvar port uint16\n\n\thost := strings.TrimSpace(parts[1])\n\n\tif len(parts) == 3 {\n\t\tvar p int\n\t\tp, err = strconv.Atoi(strings.TrimSpace(parts[2]))\n\n\t\tif err != nil {\n\t\t\terr = fmt.Errorf(\"can't convert port to number %v\", err)\n\t\t\treturn\n\t\t}\n\n\t\tif p < 1 || p > 65535 {\n\t\t\terr = fmt.Errorf(\"invalid port %d\", p)\n\t\t\treturn\n\t\t}\n\n\t\tport = uint16(p)\n\t} else {\n\t\tport = netDefaultPort[net]\n\t}\n\n\treturn Upstream{Net: net, Host: host, Port: port}, nil\n}", "func (pool *Pool) ParseURL(rawurl string) (err error) {\n\tif rawurl == \"\" {\n\t\treturn\n\t}\n\tu, err := url.Parse(rawurl)\n\tif err != nil {\n\t\treturn\n\t}\n\tif u.Scheme != \"redis\" {\n\t\terr = fmt.Errorf(`Invalid URL scheme %q`, u.Scheme)\n\t\treturn\n\t}\n\n\tif path := strings.Trim(u.Path, \"/\"); path != \"\" {\n\t\tn, err := strconv.ParseInt(path, 10, 32)\n\t\tif err != nil || n < 0 {\n\t\t\treturn fmt.Errorf(`Invalid URL path %q`, u.Path)\n\t\t}\n\t\tpool.DB = int(n)\n\t}\n\n\tq := u.Query()\n\thost, port := u.Hostname(), u.Port()\n\tif port == \"\" {\n\t\tport = \"6379\"\n\t}\n\tpool.Address = host + \":\" + port\n\n\tif v, ok := q[\"read-timeout\"]; ok && len(v) > 0 {\n\t\tif d, _ := time.ParseDuration(v[0]); d > 0 {\n\t\t\tpool.ReadTimeout = d\n\t\t}\n\t}\n\tif v, ok := q[\"write-timeout\"]; ok && len(v) > 0 {\n\t\tif d, _ := time.ParseDuration(v[0]); d > 0 {\n\t\t\tpool.WriteTimeout = d\n\t\t}\n\t}\n\tif v, ok := q[\"wait-timeout\"]; ok && len(v) > 0 {\n\t\tif d, _ := time.ParseDuration(v[0]); d > 0 {\n\t\t\tpool.WaitTimeout = d\n\t\t}\n\t}\n\tif v, ok := q[\"read-buffer-size\"]; ok && len(v) > 0 {\n\t\tif size, _ := strconv.Atoi(v[0]); size > 0 {\n\t\t\tpool.ReadBufferSize = size\n\t\t}\n\t}\n\n\tif v, ok := q[\"max-conn-age\"]; ok && len(v) > 0 {\n\t\tif d, _ := time.ParseDuration(v[0]); d > 0 {\n\t\t\tpool.MaxConnectionAge = d\n\t\t}\n\t}\n\tif v, ok := q[\"max-idle-time\"]; ok && len(v) > 0 {\n\t\tif d, _ := time.ParseDuration(v[0]); d > 0 {\n\t\t\tpool.MaxIdleTime = d\n\t\t}\n\t}\n\tif v, ok := q[\"check-idle-interval\"]; ok && len(v) > 0 {\n\t\tif d, _ := time.ParseDuration(v[0]); d > 0 {\n\t\t\tpool.CheckIdleInterval = d\n\t\t}\n\t}\n\n\treturn\n}", "func ParseConnString(connString string) (ConnectionParams, error) {\n\tP := ConnectionParams{\n\t\tMinSessions: DefaultPoolMinSessions,\n\t\tMaxSessions: DefaultPoolMaxSessions,\n\t\tPoolIncrement: DefaultPoolIncrement,\n\t\tConnClass: DefaultConnectionClass,\n\t\tMaxLifeTime: DefaultMaxLifeTime,\n\t\tWaitTimeout: DefaultWaitTimeout,\n\t\tSessionTimeout: DefaultSessionTimeout,\n\t}\n\tif !strings.HasPrefix(connString, \"oracle://\") {\n\t\ti := strings.IndexByte(connString, '/')\n\t\tif i < 0 {\n\t\t\treturn P, errors.New(\"no '/' in connection string\")\n\t\t}\n\t\tP.Username, connString = connString[:i], connString[i+1:]\n\n\t\tuSid := strings.ToUpper(connString)\n\t\t//fmt.Printf(\"connString=%q SID=%q\\n\", connString, uSid)\n\t\tif strings.Contains(uSid, \" AS \") {\n\t\t\tif P.IsSysDBA = strings.HasSuffix(uSid, \" AS SYSDBA\"); P.IsSysDBA {\n\t\t\t\tconnString = connString[:len(connString)-10]\n\t\t\t} else if P.IsSysOper = strings.HasSuffix(uSid, \" AS SYSOPER\"); P.IsSysOper {\n\t\t\t\tconnString = connString[:len(connString)-11]\n\t\t\t} else if P.IsSysASM = strings.HasSuffix(uSid, \" AS SYSASM\"); P.IsSysASM {\n\t\t\t\tconnString = connString[:len(connString)-10]\n\t\t\t}\n\t\t}\n\t\tif i = strings.IndexByte(connString, '@'); i >= 0 {\n\t\t\tP.Password, P.SID = connString[:i], connString[i+1:]\n\t\t} else {\n\t\t\tP.Password = connString\n\t\t}\n\t\tif strings.HasSuffix(P.SID, \":POOLED\") {\n\t\t\tP.ConnClass, P.SID = \"POOLED\", P.SID[:len(P.SID)-7]\n\t\t}\n\t\t//fmt.Printf(\"connString=%q params=%s\\n\", connString, P)\n\t\treturn P, nil\n\t}\n\tu, err := url.Parse(connString)\n\tif err != nil {\n\t\treturn P, errors.Errorf(\"%s: %w\", connString, err)\n\t}\n\tif usr := u.User; usr != nil {\n\t\tP.Username = usr.Username()\n\t\tP.Password, _ = usr.Password()\n\t}\n\tP.SID = u.Hostname()\n\tif u.Port() != \"\" {\n\t\tP.SID += \":\" + u.Port()\n\t}\n\tif u.Path != \"\" && u.Path != \"/\" {\n\t\tP.SID += u.Path\n\t}\n\tq := u.Query()\n\tif vv, ok := q[\"connectionClass\"]; ok {\n\t\tP.ConnClass = vv[0]\n\t}\n\tfor _, task := range []struct {\n\t\tDest *bool\n\t\tKey string\n\t}{\n\t\t{&P.IsSysDBA, \"sysdba\"},\n\t\t{&P.IsSysOper, \"sysoper\"},\n\t\t{&P.IsSysASM, \"sysasm\"},\n\t\t{&P.IsPrelim, \"prelim\"},\n\n\t\t{&P.StandaloneConnection, \"standaloneConnection\"},\n\t\t{&P.EnableEvents, \"enableEvents\"},\n\t\t{&P.HeterogeneousPool, \"heterogeneousPool\"},\n\t} {\n\t\t*task.Dest = q.Get(task.Key) == \"1\"\n\t}\n\tif tz := q.Get(\"timezone\"); tz != \"\" {\n\t\tif tz == \"local\" {\n\t\t\tP.Timezone = time.Local\n\t\t} else if strings.Contains(tz, \"/\") {\n\t\t\tif P.Timezone, err = time.LoadLocation(tz); err != nil {\n\t\t\t\treturn P, errors.Errorf(\"%s: %w\", tz, err)\n\t\t\t}\n\t\t} else if off, err := parseTZ(tz); err == nil {\n\t\t\tP.Timezone = time.FixedZone(tz, off)\n\t\t} else {\n\t\t\treturn P, errors.Errorf(\"%s: %w\", tz, err)\n\t\t}\n\t}\n\n\tP.StandaloneConnection = P.StandaloneConnection || P.ConnClass == NoConnectionPoolingConnectionClass\n\tif P.IsPrelim {\n\t\tP.ConnClass = \"\"\n\t}\n\tif P.StandaloneConnection {\n\t\tP.NewPassword = q.Get(\"newPassword\")\n\t}\n\n\tfor _, task := range []struct {\n\t\tDest *int\n\t\tKey string\n\t}{\n\t\t{&P.MinSessions, \"poolMinSessions\"},\n\t\t{&P.MaxSessions, \"poolMaxSessions\"},\n\t\t{&P.PoolIncrement, \"poolIncrement\"},\n\t} {\n\t\ts := q.Get(task.Key)\n\t\tif s == \"\" {\n\t\t\tcontinue\n\t\t}\n\t\tvar err error\n\t\t*task.Dest, err = strconv.Atoi(s)\n\t\tif err != nil {\n\t\t\treturn P, errors.Errorf(\"%s: %w\", task.Key+\"=\"+s, err)\n\t\t}\n\t}\n\tfor _, task := range []struct {\n\t\tDest *time.Duration\n\t\tKey string\n\t}{\n\t\t{&P.SessionTimeout, \"poolSessionTimeout\"},\n\t\t{&P.WaitTimeout, \"poolWaitTimeout\"},\n\t\t{&P.MaxLifeTime, \"poolSessionMaxLifetime\"},\n\t} {\n\t\ts := q.Get(task.Key)\n\t\tif s == \"\" {\n\t\t\tcontinue\n\t\t}\n\t\tvar err error\n\t\t*task.Dest, err = time.ParseDuration(s)\n\t\tif err != nil {\n\t\t\tif !strings.Contains(err.Error(), \"time: missing unit in duration\") {\n\t\t\t\treturn P, errors.Errorf(\"%s: %w\", task.Key+\"=\"+s, err)\n\t\t\t}\n\t\t\ti, err := strconv.Atoi(s)\n\t\t\tif err != nil {\n\t\t\t\treturn P, errors.Errorf(\"%s: %w\", task.Key+\"=\"+s, err)\n\t\t\t}\n\t\t\tbase := time.Second\n\t\t\tif task.Key == \"poolWaitTimeout\" {\n\t\t\t\tbase = time.Millisecond\n\t\t\t}\n\t\t\t*task.Dest = time.Duration(i) * base\n\t\t}\n\t}\n\tif P.MinSessions > P.MaxSessions {\n\t\tP.MinSessions = P.MaxSessions\n\t}\n\tif P.MinSessions == P.MaxSessions {\n\t\tP.PoolIncrement = 0\n\t} else if P.PoolIncrement < 1 {\n\t\tP.PoolIncrement = 1\n\t}\n\treturn P, nil\n}", "func ParseURL(serverFlag string) (*UrlRes, int, string) {\n\n\tpURL := &UrlRes{serverFlag, \"\", \"\"}\n\n\t// the URL golang Parse method has the limitation that when we pass in a host that is a string\n\t// and not an ip, without the protocol scheme, it mis-interprets the url string. For such cases we\n\t// need to explicitely make sure that we are missing a protocol scheme.\n\n\t// If no protocol exists, then append http:// as default protocol.\n\n\tif !strings.HasPrefix(strings.ToLower(serverFlag), \"https://\") &&\n\t\t!strings.HasPrefix(strings.ToLower(serverFlag), \"http://\") &&\n\t\t!strings.HasPrefix(strings.ToLower(serverFlag), \"couchbase://\") &&\n\t\t!strings.HasPrefix(strings.ToLower(serverFlag), \"couchbases://\") {\n\t\t//There is something else wrong and we need to throw an error.\n\t\tserverFlag = \"http://\" + serverFlag\n\t}\n\n\t//Parse the url\n\tparsedURL, err := url.Parse(serverFlag)\n\tif err != nil {\n\t\treturn pURL, errors.INVALID_URL, err.Error()\n\t}\n\n\tif parsedURL.Host == \"\" {\n\t\treturn pURL, errors.INVALID_URL, INVALIDHOST\n\t}\n\n\t// Check if the input url is a DNS SRV\n\t_, addr, err := net.LookupSRV(parsedURL.Scheme, \"tcp\", parsedURL.Hostname())\n\tif err == nil {\n\t\t// It is a DNS SRV .. Has couchbase or couchbases as a scheme\n\t\tparsedURL.Host = addr[0].Target\n\t}\n\n\t// We now have a valid URL. Check if we have a port\n\t_, portNo, err := net.SplitHostPort(parsedURL.Host)\n\n\t// couchbase:// and couchbases:// will represent http:// ... :8091 and\n\t// https:// ... 18091 respectively. If the port is specified along with\n\t// the scheme for this case, we throw an error.\n\n\tif parsedURL.Hostname() != \"\" {\n\t\tparsedURL.Host = parsedURL.Hostname()\n\t}\n\n\tif portNo == \"\" {\n\t\tif strings.ToLower(parsedURL.Scheme) == \"couchbase\" || strings.ToLower(parsedURL.Scheme) == \"couchbases\" {\n\n\t\t\tif strings.ToLower(parsedURL.Scheme) == \"couchbase\" {\n\t\t\t\tparsedURL.Host = net.JoinHostPort(parsedURL.Host, \"8091\")\n\t\t\t\tparsedURL.Scheme = \"http\"\n\n\t\t\t} else {\n\t\t\t\tparsedURL.Scheme = \"https\"\n\t\t\t\tparsedURL.Host = net.JoinHostPort(parsedURL.Host, \"18091\")\n\t\t\t}\n\n\t\t} else {\n\t\t\tif strings.ToLower(parsedURL.Scheme) == \"http\" {\n\t\t\t\tparsedURL.Host = net.JoinHostPort(parsedURL.Host, \"8091\")\n\n\t\t\t} else if strings.ToLower(parsedURL.Scheme) == \"https\" {\n\t\t\t\tparsedURL.Host = net.JoinHostPort(parsedURL.Host, \"18091\")\n\t\t\t}\n\t\t}\n\t} else {\n\t\tparsedURL.Host = net.JoinHostPort(parsedURL.Host, portNo)\n\t\t// Cannot give port with couchbase:// couchbases://\n\t\tif strings.ToLower(parsedURL.Scheme) == \"couchbase\" || strings.ToLower(parsedURL.Scheme) == \"couchbases\" {\n\t\t\treturn pURL, errors.INVALID_URL, INVALIDPORT\n\t\t} else {\n\t\t\tif err != nil {\n\t\t\t\treturn pURL, errors.INVALID_URL, err.Error()\n\t\t\t}\n\t\t}\n\t}\n\n\tpURL.Password, _ = parsedURL.User.Password()\n\tpURL.Username = parsedURL.User.Username()\n\tpURL.ServerFlag = parsedURL.String()\n\n\treturn pURL, 0, \"\"\n}", "func parseAddr(msg string) string {\n\tif strings.HasSuffix(msg, crlf) {\n\t\tmsg = msg + crlf\n\t}\n\n\tresp, err := http.ReadResponse(bufio.NewReader(strings.NewReader(msg)), nil)\n\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn \"\"\n\t}\n\n\tdefer resp.Body.Close()\n\n\treturn strings.TrimPrefix(resp.Header.Get(\"LOCATION\"), \"yeelight://\")\n}", "func ParseDSN(dsn string, hostOverride HostOverride, logf Logf) (*Config, error) {\n\tif logf == nil {\n\t\tlogf = bytehouse.EmptyConnectionContext.GetLogf()\n\t}\n\tif hostOverride == nil {\n\t\thostOverride = bytehouse.EmptyConnectionContext.GetResolveHost()\n\t}\n\n\thost, urlValues, err := parseAndResolveHost(dsn, hostOverride)\n\tif err != nil {\n\t\treturn nil, errors.ErrorfWithCaller(\"host port resolution error = %v\", err)\n\t}\n\n\tconnOptions, err := makeConnConfigs(host, urlValues, logf)\n\tif err != nil {\n\t\treturn nil, errors.ErrorfWithCaller(\"makeConnConfigs error = %v\", err)\n\t}\n\n\tdatabaseName := urlValues.Get(\"database\")\n\n\tauthentication := makeAuthentication(urlValues)\n\n\t//impersonation, err := makeImpersonation(urlValues)\n\tif err != nil {\n\t\treturn nil, errors.ErrorfWithCaller(\"makeImpersonation error = %v\", err)\n\t}\n\n\tcompress, err := parseBool(urlValues.Get(param.COMPRESS))\n\tif err != nil {\n\t\treturn nil, errors.ErrorfWithCaller(\"error parsing compress parameter as bool = %v\", err)\n\t}\n\n\tquerySettings := makeQuerySettings(urlValues)\n\n\treturn &Config{\n\t\tconnConfig: connOptions,\n\t\tdatabaseName: databaseName,\n\t\tauthentication: authentication,\n\t\t//impersonation: impersonation,\n\t\tcompress: compress,\n\t\tquerySettings: querySettings,\n\t}, nil\n}", "func parseHost(host string) string {\n\trealHost, _, _ := net.SplitHostPort(host)\n\tif realHost != \"\" {\n\t\treturn realHost\n\t}\n\treturn host\n}", "func parseRepo(rawURL string) (string, string) {\n\tu, err := url.Parse(rawURL)\n\tif err == nil {\n\t\tp := strings.Split(u.Path, \"/\")\n\t\tif u.Hostname() != \"\" {\n\t\t\treturn p[1], p[2]\n\t\t}\n\t\treturn p[0], p[1]\n\t}\n\t// Not a URL\n\tp := strings.Split(rawURL, \"/\")\n\treturn p[0], p[1]\n}", "func parseEnsAPIAddress(s string) (tld, endpoint string, addr common.Address) {\n\tisAllLetterString := func(s string) bool {\n\t\tfor _, r := range s {\n\t\t\tif !unicode.IsLetter(r) {\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\t\treturn true\n\t}\n\tendpoint = s\n\tif i := strings.Index(endpoint, \":\"); i > 0 {\n\t\tif isAllLetterString(endpoint[:i]) && len(endpoint) > i+2 && endpoint[i+1:i+3] != \"//\" {\n\t\t\ttld = endpoint[:i]\n\t\t\tendpoint = endpoint[i+1:]\n\t\t}\n\t}\n\tif i := strings.Index(endpoint, \"@\"); i > 0 {\n\t\taddr = common.HexToAddress(endpoint[:i])\n\t\tendpoint = endpoint[i+1:]\n\t}\n\treturn\n}", "func parseAddr(addr string) (string, string) {\n\tparsed := strings.SplitN(addr, \":\", 2)\n\treturn parsed[0], parsed[1]\n}", "func DnsDecoder(urlStr string) (*string, *string, error) {\n\tu, err := url.Parse(urlStr)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\thostTmp := u.Host\n\tIP := Dns(u.Host)\n\tif IP != nil {\n\t\tu.Host = IP.String()\n\t\turlStr = u.String()\n\t\treturn &urlStr, &hostTmp, nil\n\t}\n\treturn nil, nil, fmt.Errorf(\"dnsDecoder fail\")\n}", "func (_BaseContentSpace *BaseContentSpaceFilterer) ParseBindUserWallet(log types.Log) (*BaseContentSpaceBindUserWallet, error) {\n\tevent := new(BaseContentSpaceBindUserWallet)\n\tif err := _BaseContentSpace.contract.UnpackLog(event, \"BindUserWallet\", log); err != nil {\n\t\treturn nil, err\n\t}\n\tevent.Raw = log\n\treturn event, nil\n}", "func parseUrl(rawurl string) (string, error) {\n\t_, err := url.Parse(rawurl)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn rawurl, nil\n}", "func parseURLForGithubAPIEndpoint(urlString string) (*url.URL, string) {\n\tu, err := url.Parse(urlString)\n\tif err != nil {\n\t\tlog.Fatalf(\"an error occurred parsing url %s for potential repository: %s\", urlString, err)\n\t}\n\n\t// parse hostname and path for potential github repo api endpoint\n\thostname := u.Hostname()\n\tpath := u.Path\n\trepoURL := github.GetApiEndpoint(hostname, path)\n\treturn u, repoURL\n}", "func RawParser(raw string) Results {\n results := Results{}\n results.Command = toS(find(raw, \"^(httperf .*)\"))\n results.MaxConnectBurstLength = toI(find(raw, \"Maximum connect burst length: ([0-9]*?\\\\.?[0-9]+)$\"))\n results.TotalConnections = toI(find(raw, \"^Total: connections ([0-9]*?\\\\.?[0-9]+) \"))\n results.TotalRequests = toI(find(raw, \"^Total: connections .+ requests ([0-9]*?\\\\.?[0-9]+) \"))\n results.TotalReplies = toI(find(raw, \"^Total: connections .+ replies ([0-9]*?\\\\.?[0-9]+) \"))\n results.TotalTestDuration = toF(find(raw, \"^Total: connections .+ test-duration ([0-9]*?\\\\.?[0-9]+) \"))\n results.ConnectionRatePerSec = toF(find(raw, \"^Connection rate: ([0-9]*?\\\\.?[0-9]+) \"))\n results.ConnectionRateMsConn = toF(find(raw, \"^Connection rate: .+ \\\\(([0-9]*?\\\\.?[0-9]+) ms\"))\n results.ConnectionTimeMin = toF(find(raw, \"^Connection time \\\\[ms\\\\]: min ([0-9]*?\\\\.?[0-9]+) \"))\n results.ConnectionTimeAvg = toF(find(raw, \"^Connection time \\\\[ms\\\\]: min .+ avg ([0-9]*?\\\\.?[0-9]+) \"))\n results.ConnectionTimeMax = toF(find(raw, \"^Connection time \\\\[ms\\\\]: min .+ max ([0-9]*?\\\\.?[0-9]+) \"))\n results.ConnectionTimeMedian = toF(find(raw, \"^Connection time \\\\[ms\\\\]: min .+ median ([0-9]*?\\\\.?[0-9]+) \"))\n results.ConnectionTimeStddev = toF(find(raw, \"^Connection time \\\\[ms\\\\]: min .+ stddev ([0-9]*?\\\\.?[0-9]+)$\"))\n results.ConnectionTimeConnect = toF(find(raw, \"^Connection time \\\\[ms\\\\]: connect ([0-9]*?\\\\.?[0-9]+)$\"))\n results.ConnectionLength = toF(find(raw, \"^Connection length \\\\[replies\\\\/conn\\\\]: ([0-9]*?\\\\.?[0-9]+)$\"))\n results.RequestRatePerSec = toF(find(raw, \"^Request rate: ([0-9]*?\\\\.?[0-9]+) req\"))\n results.RequestRateMsRequest = toF(find(raw, \"^Request rate: .+ \\\\(([0-9]*?\\\\.?[0-9]+) ms\"))\n results.RequestSize = toF(find(raw, \"^Request size \\\\[B\\\\]: ([0-9]*?\\\\.?[0-9]+)$\"))\n results.ReplyRateMin = toF(find(raw, \"^Reply rate \\\\[replies\\\\/s\\\\]: min ([0-9]*?\\\\.?[0-9]+) \"))\n results.ReplyRateAvg = toF(find(raw, \"^Reply rate \\\\[replies\\\\/s\\\\]: min .+ avg ([0-9]*?\\\\.?[0-9]+) \"))\n results.ReplyRateMax = toF(find(raw, \"^Reply rate \\\\[replies\\\\/s\\\\]: min .+ max ([0-9]*?\\\\.?[0-9]+) \"))\n results.ReplyRateStddev = toF(find(raw, \"^Reply rate \\\\[replies\\\\/s\\\\]: min .+ stddev ([0-9]*?\\\\.?[0-9]+) \"))\n results.ReplyRateSamples = toI(find(raw, \"^Reply rate \\\\[replies\\\\/s\\\\]: min .+ \\\\(([0-9]*?\\\\.?[0-9]+) samples\"))\n results.ReplyTimeResponse = toF(find(raw, \"^Reply time \\\\[ms\\\\]: response ([0-9]*?\\\\.?[0-9]+) \"))\n results.ReplyTimeTransfer = toF(find(raw, \"^Reply time \\\\[ms\\\\]: response .+ transfer ([0-9]*?\\\\.?[0-9]+)$\"))\n results.ReplySizeHeader = toF(find(raw, \"^Reply size \\\\[B\\\\]: header ([0-9]*?\\\\.?[0-9]+) \"))\n results.ReplySizeContent = toF(find(raw, \"^Reply size \\\\[B\\\\]: header .+ content ([0-9]*?\\\\.?[0-9]+) \"))\n results.ReplySizeFooter = toF(find(raw, \"^Reply size \\\\[B\\\\]: header .+ footer ([0-9]*?\\\\.?[0-9]+) \"))\n results.ReplySizeTotal = toF(find(raw, \"^Reply size \\\\[B\\\\]: header .+ \\\\(total ([0-9]*?\\\\.?[0-9]+)\\\\)\"))\n results.ReplyStatus1xx = toI(find(raw, \"^Reply status: 1xx=([0-9]*?\\\\.?[0-9]+) \"))\n results.ReplyStatus2xx = toI(find(raw, \"^Reply status: .+ 2xx=([0-9]*?\\\\.?[0-9]+) \"))\n results.ReplyStatus3xx = toI(find(raw, \"^Reply status: .+ 3xx=([0-9]*?\\\\.?[0-9]+) \"))\n results.ReplyStatus4xx = toI(find(raw, \"^Reply status: .+ 4xx=([0-9]*?\\\\.?[0-9]+) \"))\n results.ReplyStatus5xx = toI(find(raw, \"^Reply status: .+ 5xx=([0-9]*?\\\\.?[0-9]+)\"))\n results.CPUTimeUserSec = toF(find(raw, \"^CPU time \\\\[s\\\\]: user ([0-9]*?\\\\.?[0-9]+) \"))\n results.CPUTimeUserPct = toF(find(raw, \"^CPU time \\\\[s\\\\]: .+ \\\\(user ([0-9]*?\\\\.?[0-9]+)\\\\% \"))\n results.CPUTimeSystemSec = toF(find(raw, \"^CPU time \\\\[s\\\\]: .+ system ([0-9]*?\\\\.?[0-9]+) \"))\n results.CPUTimeSystemPct = toF(find(raw, \"^CPU time \\\\[s\\\\]: user .+ system .+ system ([0-9]*?\\\\.?[0-9]+)\\\\% \"))\n results.CPUTimeTotalPct = toF(find(raw, \"^CPU time \\\\[s\\\\]: user .+ total ([0-9]*?\\\\.?[0-9]+)\\\\%\"))\n results.NetIoKbSec = toF(find(raw, \"^Net I\\\\/O: ([0-9]*?\\\\.?[0-9]+) KB\"))\n results.NetIoBps = toS(find(raw, \"^Net I\\\\/O: .+ \\\\((.+) bps\\\\)\"))\n results.ErrorsTotal = toI(find(raw, \"^Errors: total ([0-9]*?\\\\.?[0-9]+) \"))\n results.ErrorsClientTimeout = toI(find(raw, \"^Errors: total .+ client-timo ([0-9]*?\\\\.?[0-9]+) \"))\n results.ErrorsSocketTimeout = toI(find(raw, \"^Errors: total .+ socket-timo ([0-9]*?\\\\.?[0-9]+) \"))\n results.ErrorsConnRefused = toI(find(raw, \"^Errors: total .+ connrefused ([0-9]*?\\\\.?[0-9]+) \"))\n results.ErrorsConnReset = toI(find(raw, \"^Errors: total .+ connreset ([0-9]*?\\\\.?[0-9]+)\"))\n results.ErrorsFdUnavail = toI(find(raw, \"^Errors: fd-unavail ([0-9]*?\\\\.?[0-9]+) \"))\n results.ErrorsAddrUnavail = toI(find(raw, \"^Errors: fd-unavail .+ addrunavail ([0-9]*?\\\\.?[0-9]+) \"))\n results.ErrorsFtabFull = toI(find(raw, \"^Errors: fd-unavail .+ ftab-full ([0-9]*?\\\\.?[0-9]+) \"))\n results.ErrorsOther = toI(find(raw, \"^Errors: fd-unavail .+ other ([0-9]*?\\\\.?[0-9]+)\"))\n results.ConnectionTimes = findConnectionTimes(raw)\n results.calculatePercentiles()\n\n return results\n}", "func (s *HTTPServer) parseDC(req *http.Request, dc *string) {\n\tif other := req.URL.Query().Get(\"dc\"); other != \"\" {\n\t\t*dc = other\n\t} else if *dc == \"\" {\n\t\t*dc = s.agent.config.Datacenter\n\t}\n}", "func (c *Client) BanURL(channels []string, host string, value []string) error {\n\treturn c.Do(channels, Request{Command: \"ban.url\", Host: host, Value: value})\n}", "func (r *repoURL) parse(input string) error {\n\tvar t string\n\tc := strings.Split(input, \"/\")\n\tswitch l := len(c); {\n\t// owner/repo format is not supported for gitlab, it's github-only\n\tcase l == 2:\n\t\treturn sce.WithMessage(sce.ErrScorecardInternal, fmt.Sprintf(\"gitlab repo must specify host: %s\", input))\n\tcase l >= 3:\n\t\tt = input\n\t}\n\n\t// Allow skipping scheme for ease-of-use, default to https.\n\tif !strings.Contains(t, \"://\") {\n\t\tt = \"https://\" + t\n\t}\n\n\tu, e := url.Parse(t)\n\tif e != nil {\n\t\treturn sce.WithMessage(sce.ErrScorecardInternal, fmt.Sprintf(\"url.Parse: %v\", e))\n\t}\n\n\tconst splitLen = 2\n\tsplit := strings.SplitN(strings.Trim(u.Path, \"/\"), \"/\", splitLen)\n\tif len(split) != splitLen {\n\t\treturn sce.WithMessage(sce.ErrorInvalidURL, fmt.Sprintf(\"%v. Expected full repository url\", input))\n\t}\n\n\tr.scheme, r.host, r.owner, r.project = u.Scheme, u.Host, split[0], split[1]\n\treturn nil\n}", "func ParseURL(u string) (*AbtabURL, error) {\n\turl, err := url.Parse(u)\n\tif err != nil {\n\t\treturn nil, AbtabError{Message: fmt.Sprintf(\"Error: invalid url: '%s' :: %s\", u, err), CausedBy: err}\n\t}\n\n\treturn &AbtabURL{\n\t\tOriginalUrl: u,\n\t\tUrl: url,\n\t}, nil\n}", "func parseURL(url *url.URL) (BlockID, string, error) {\n\n\tpath := strings.Split(url.Path, \"/\")\n\tif len(path) != 5 || path[1] != \"block\" || path[2] == \"\" || path[3] != \"txs\" || path[4] == \"\" {\n\t\treturn \"\", \"\", errors.New(\"Invalid request path\")\n\t}\n\n\tblockPattern := path[2]\n\tif blockPattern != \"latest\" {\n\t\tblockNumber, err := strconv.Atoi(path[2])\n\t\tif err != nil || blockNumber < 0 {\n\t\t\treturn \"\", \"\", errors.New(\"Invalid block number: \" + path[2])\n\t\t}\n\t}\n\n\treturn BlockID(blockPattern), path[4], nil\n}", "func ParseNUpDetails(s string, nup *model.NUp) error {\n\tif s == \"\" {\n\t\treturn errInvalidNUpConfig\n\t}\n\n\tss := strings.Split(s, \",\")\n\n\tfor _, s := range ss {\n\n\t\tss1 := strings.Split(s, \":\")\n\t\tif len(ss1) != 2 {\n\t\t\treturn errInvalidNUpConfig\n\t\t}\n\n\t\tparamPrefix := strings.TrimSpace(ss1[0])\n\t\tparamValueStr := strings.TrimSpace(ss1[1])\n\n\t\tif err := nupParamMap.Handle(paramPrefix, paramValueStr, nup); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func TestURLParse(t *testing.T) {\n\ttests := []struct {\n\t\tin string\n\t\twant string\n\t}{\n\t\t{\"ssh://github.com/grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"ssh://[email protected]/grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"ssh://[email protected]:1234/grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"ssh://[email protected]/~user/grdl/git-get.git\", \"github.com/user/grdl/git-get\"},\n\t\t{\"git+ssh://github.com/grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"[email protected]:grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"[email protected]:/~user/grdl/git-get.git\", \"github.com/user/grdl/git-get\"},\n\t\t{\"git://github.com/grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"git://github.com/~user/grdl/git-get.git\", \"github.com/user/grdl/git-get\"},\n\t\t{\"https://github.com/grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"http://github.com/grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"https://github.com/grdl/git-get\", \"github.com/grdl/git-get\"},\n\t\t{\"https://github.com/git-get.git\", \"github.com/git-get\"},\n\t\t{\"https://github.com/git-get\", \"github.com/git-get\"},\n\t\t{\"https://github.com/grdl/sub/path/git-get.git\", \"github.com/grdl/sub/path/git-get\"},\n\t\t{\"https://github.com:1234/grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"https://github.com/grdl/git-get.git/\", \"github.com/grdl/git-get\"},\n\t\t{\"https://github.com/grdl/git-get/\", \"github.com/grdl/git-get\"},\n\t\t{\"https://github.com/grdl/git-get/////\", \"github.com/grdl/git-get\"},\n\t\t{\"https://github.com/grdl/git-get.git/////\", \"github.com/grdl/git-get\"},\n\t\t{\"ftp://github.com/grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"ftps://github.com/grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"rsync://github.com/grdl/git-get.git\", \"github.com/grdl/git-get\"},\n\t\t{\"local/grdl/git-get/\", \"github.com/local/grdl/git-get\"},\n\t\t{\"file://local/grdl/git-get\", \"local/grdl/git-get\"},\n\t}\n\n\tfor _, test := range tests {\n\t\turl, err := ParseURL(test.in, cfg.Defaults[cfg.KeyDefaultHost], cfg.Defaults[cfg.KeyDefaultScheme])\n\t\tassert.NoError(t, err)\n\n\t\tgot := URLToPath(*url, false)\n\t\tassert.Equal(t, test.want, got)\n\t}\n}", "func parseUrlforStu(url string) *gofeed.Feed {\n\tfp := gofeed.NewParser()\n\toutput, _ := fp.ParseURL(url)\n\treturn output\n}", "func parseConnectionStr(connectionStr string) (string, error) {\n\tdbPos := strings.Index(connectionStr, \":\")\n\tif dbPos == -1 {\n\t\treturn \"\", errorList[\"err.db.conf.separator.missing\"]\n\t}\n\tdbName := strings.TrimSpace(connectionStr[:dbPos])\n\n\treturn dbName, nil\n}", "func splitBeanName(bean string) (string, string, error) {\n\tdomainQuery := strings.SplitN(bean, \":\", 2)\n\tif len(domainQuery) != 2 {\n\t\treturn \"\", \"\", fmt.Errorf(\"invalid domain:bean string %s\", bean)\n\t}\n\treturn domainQuery[0], domainQuery[1], nil\n}", "func (d *Downloader) parseURI(keyString string) (string, string, string, string) {\n\tvar region string\n\tss := strings.Split(keyString, \"/\")\n\tbucketSs := strings.Split(ss[2], \".\")\n\tbucket := bucketSs[0]\n\tregionSs := strings.Split(bucketSs[1], \"-\")\n\t// Default to us-east-1 if just <bucket>.s3.amazonaws.com is passed\n\tif len(regionSs) == 1 {\n\t\tregion = \"us-east-1\"\n\t} else {\n\t\tregion = strings.Join(regionSs[1:], \"-\")\n\t}\n\tkey := strings.Join(ss[3:], \"/\")\n\tfilename := ss[len(ss)-1]\n\treturn bucket, region, key, filename\n}", "func parseDatabaseURL(cfg *Prest) {\n\tif cfg.PGURL == \"\" {\n\t\tlog.Debugln(\"no db url found, skipping\")\n\t\treturn\n\t}\n\t// Parser PG URL, get database connection via string URL\n\tu, err := url.Parse(cfg.PGURL)\n\tif err != nil {\n\t\tlog.Errorf(\"cannot parse db url, err: %v\\n\", err)\n\t\treturn\n\t}\n\tcfg.PGHost = u.Hostname()\n\tif u.Port() != \"\" {\n\t\tpgPort, err := strconv.Atoi(u.Port())\n\t\tif err != nil {\n\t\t\tlog.Errorf(\n\t\t\t\t\"cannot parse db url port '%v', falling back to default values\\n\",\n\t\t\t\tu.Port())\n\t\t\treturn\n\t\t}\n\t\tcfg.PGPort = pgPort\n\t}\n\tcfg.PGUser = u.User.Username()\n\tpgPass, pgPassExist := u.User.Password()\n\tif pgPassExist {\n\t\tcfg.PGPass = pgPass\n\t}\n\tcfg.PGDatabase = strings.Replace(u.Path, \"/\", \"\", -1)\n\tif u.Query().Get(\"sslmode\") != \"\" {\n\t\tcfg.SSLMode = u.Query().Get(\"sslmode\")\n\t}\n}", "func GetDaemonEndpointsFromString(dAddr string) (*DaemonEndpoints, error) {\n\tvar daemonAddr string\n\t// Try to get the X-Ray daemon address from an environment variable\n\tif envDaemonAddr := os.Getenv(\"AWS_XRAY_DAEMON_ADDRESS\"); envDaemonAddr != \"\" {\n\t\tdaemonAddr = envDaemonAddr\n\t\tlogger.Infof(\"using daemon endpoints from environment variable AWS_XRAY_DAEMON_ADDRESS: %v\", envDaemonAddr)\n\t} else if dAddr != \"\" {\n\t\tdaemonAddr = dAddr\n\t}\n\tif daemonAddr != \"\" {\n\t\treturn resolveAddress(daemonAddr)\n\t}\n\treturn nil, nil\n}", "func ParseRepo(str string) (user, repo string, err error) {\n\tparts := strings.Split(str, \"/\")\n\tif len(parts) != 2 {\n\t\terr = fmt.Errorf(\"Error: Invalid or missing repository. eg octocat/hello-world\")\n\t\treturn\n\t}\n\tuser = parts[0]\n\trepo = parts[1]\n\treturn\n}", "func ParseParams(urlString string) (Params, error) {\n\tp := NewParams()\n\tif !strings.HasPrefix(urlString, \"http\") {\n\t\turlString = \"https://\" + urlString\n\t}\n\tu, err := url.Parse(urlString)\n\tif err != nil {\n\t\treturn p, err\n\t}\n\tp.Query = u.Query()\n\tif u.Scheme == \"\" {\n\t\tu.Scheme = \"https\"\n\t}\n\tif u.Path != \"\" {\n\t\tp.Prefix = strings.Trim(u.Path, \"/\")\n\t}\n\tu.RawQuery = \"\"\n\tu.Fragment = \"\"\n\tu.Path = \"\"\n\tu.RawPath = \"\"\n\tp.Server = u.String()\n\treturn p, nil\n}", "func parseAdvertiseAddr(advAddr string, port int) (string, int) {\n\treturn advAddr, port\n\n\t// bug: if use domain, always return empty host\n\t/*m, e := regexp.Match(ipv4Pattern, []byte(advAddr))\n\t// if parse error, use serve port and parsed ip address\n\tif e != nil {\n\t\treturn \"\", port\n\t}\n\tif m {\n\t\treturn advAddr, port\n\t}\n\n\tm, e1 := regexp.Match(ipv4WithPortPattern, []byte(advAddr))\n\t// if parse error, use serve port and parsed ip address\n\tif e1 != nil {\n\t\treturn \"\", port\n\t}\n\tif m {\n\t\t// 1 5\n\t\tregxp := regexp.MustCompile(ipv4WithPortPattern)\n\t\tadAddr := regxp.ReplaceAllString(advAddr, \"${1}\")\n\t\tadPort, _ := strconv.Atoi(regxp.ReplaceAllString(advAddr, \"${5}\"))\n\t\treturn adAddr, adPort\n\t}\n\treturn \"\", port*/\n}", "func udpParseSockStr(netStr, sockStr string) udpAddr {\n\tvar s udpAddr\n\ts.net = netStr\n\n\tsockStr = parseNetStr(s.net == \"udp6\", sockStr)\n\ta, err := net.ResolveUDPAddr(netStr, sockStr)\n\tif err != nil {\n\t\tlog.Fatalf(\"Error parsing socket string '%s': %v\", sockStr, err)\n\t}\n\ts.addr = a\n\treturn s\n}", "func getBlockNameFromUrl(rurl string) string {\n\tvar blk string\n\tif strings.Contains(rurl, \"replicas/cms/\") {\n\t\t// url/replicas/cms/blk/datasets\n\t\tparts := strings.Split(rurl, \"replicas/cms/\")\n\t\tif len(parts) > 1 {\n\t\t\tarr := strings.Split(parts[1], \"/datasets\")\n\t\t\tblk = arr[0]\n\t\t}\n\t} else if strings.Contains(rurl, \"dids/cms/\") {\n\t\t// url/dids/cms/blk/dids\n\t\tparts := strings.Split(rurl, \"dids/cms/\")\n\t\tif len(parts) > 1 {\n\t\t\tarr := strings.Split(parts[1], \"/dids\")\n\t\t\tblk = arr[0]\n\t\t}\n\t}\n\tb, err := url.QueryUnescape(blk)\n\tif err == nil {\n\t\treturn b\n\t}\n\treturn blk\n}", "func parseFieldParameters(str string) (params fieldParameters) {}", "func ParseChannelUrl(cUrl string) *LinkUrl {\n\t//logger.Log(fmt.Sprintf(\"Entering LinkUrl:ParseChannelUrl w/ URL string as '%s'\", cUrl))\n\tif cUrl == \"\" {\n\t\treturn DefaultLinkUrl()\n\t}\n\n\t//logger.Log(fmt.Sprintf(\"Returning LinkUrl:ParseChannelUrl w/ URL string as '%s'\", cUrl))\n\treturn NewLinkUrl(cUrl)\n}", "func ParseDomain(bytes []byte) (string, error) {\n\tvar hook BuildHookResponse\n\tif err := json.Unmarshal(bytes, &hook); err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif hook.Domains == nil {\n\t\treturn \"\", fmt.Errorf(\"invalid application domain\")\n\t}\n\n\tif len(hook.Domains) < 1 {\n\t\treturn \"\", fmt.Errorf(\"invalid application domain\")\n\t}\n\n\treturn hook.Domains[0], nil\n}", "func ParseConnStr(c *cli.Context) string {\n\totherParams := \"sslmode=disable connect_timeout=5\"\n\tif c.GlobalBool(\"ssl\") {\n\t\totherParams = \"sslmode=require connect_timeout=5\"\n\t}\n\treturn fmt.Sprintf(\"user=%s dbname=%s password='%s' host=%s port=%s %s\",\n\t\tc.GlobalString(\"username\"),\n\t\tc.GlobalString(\"dbname\"),\n\t\tc.GlobalString(\"pass\"),\n\t\tc.GlobalString(\"host\"),\n\t\tc.GlobalString(\"port\"),\n\t\totherParams,\n\t)\n}", "func parseArg(url string) (string, string) {\n\tif strings.Contains(url, \"spotify.com\") {\n\t\turl = strings.ReplaceAll(url, \"\\\\\", \"/\")\n\t\turl = strings.TrimSuffix(url, \"/\")\n\t\tlist := strings.Split(url, \"/\")\n\t\tid := list[len(list)-1]\n\t\tid = strings.Split(id, \"?\")[0]\n\t\tif strings.Contains(url, \"track\") {\n\t\t\treturn \"track\", id\n\t\t} else if strings.Contains(url, \"album\") {\n\t\t\treturn \"album\", id\n\t\t} else if strings.Contains(url, \"playlist\") {\n\t\t\treturn \"playlist\", id\n\t\t}\n\t}\n\treturn \"query\", url\n}", "func parseListeningAddress(ctx *context.T, laddress string) (network string, address string, p flow.Protocol, err error) {\n\tparts := strings.SplitN(laddress, \"/\", 2)\n\tif len(parts) != 2 {\n\t\treturn \"\", \"\", nil, ErrorfInvalidAddress(ctx, \"invalid vine address %v, address must be of the form 'network/address/tag'\", laddress)\n\t}\n\tp, _ = flow.RegisteredProtocol(parts[0])\n\tif p == nil {\n\t\treturn \"\", \"\", nil, ErrorfNoRegisteredProtocol(ctx, \"no registered protocol: %v\", parts[0])\n\t}\n\treturn parts[0], parts[1], p, nil\n}", "func parseURI(uri string) (string, string, error) {\n\tu, err := url.Parse(uri)\n\tif err != nil {\n\t\treturn \"\", \"\", err\n\t}\n\n\treturn strings.Split(u.Host, \".\")[0], strings.TrimLeft(u.Path, \"/\"), nil\n}", "func ParseGetaspecificBannedUserResponse(rsp *http.Response) (*GetaspecificBannedUserResponse, error) {\n\tbodyBytes, err := ioutil.ReadAll(rsp.Body)\n\tdefer rsp.Body.Close()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresponse := &GetaspecificBannedUserResponse{\n\t\tBody: bodyBytes,\n\t\tHTTPResponse: rsp,\n\t}\n\n\tswitch {\n\tcase strings.Contains(rsp.Header.Get(\"Content-Type\"), \"json\") && rsp.StatusCode == 200:\n\t\tvar dest string\n\t\tif err := json.Unmarshal(bodyBytes, &dest); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tresponse.JSON200 = &dest\n\n\t}\n\n\treturn response, nil\n}", "func parseDtab(dtabStr string) (Dtab, error) {\n\tdtabStr = strings.Replace(dtabStr, \"\\n\", \"\", -1)\n\tif dtabStr == \"\" {\n\t\treturn Dtab([]*Dentry{}), nil\n\t}\n\tdentryStrs := dentrySepRE.Split(dtabStr, -1)\n\tdentries := []*Dentry{}\n\tfor _, dentryStr := range dentryStrs {\n\t\tif dentryStr == \"\" {\n\t\t\tcontinue\n\t\t}\n\t\tparts := strings.SplitN(dentryStr, \"=>\", 2)\n\t\tif len(parts) != 2 || parts[0] == \"\" || parts[1] == \"\" {\n\t\t\treturn nil, fmt.Errorf(\"invalid dentry: '%s'\", dentryStr)\n\t\t}\n\t\tdentries = append(dentries, &Dentry{parts[0], parts[1]})\n\t}\n\tdtab := Dtab(dentries)\n\treturn dtab, nil\n}", "func SABSendURL(guid string, nzblink string, nicename string, category string) string {\n\tlog.Printf(\"SABSendURL:Grabbing:%s:%s:%s\", guid, category, nicename)\n\tnzburl, err := url.Parse(MYSABURL)\n\tif err != nil {\n\t\tlog.Print(\"SABSendURL:Parse:\", err)\n\t\treturn \"\"\n\t}\n\tparams := url.Values{}\n\tparams.Add(\"mode\", \"addurl\")\n\tparams.Add(\"output\", \"json\")\n\tparams.Add(\"apikey\", MYSABAPI)\n\tparams.Add(\"name\", nzblink)\n\tparams.Add(\"nzbname\", nicename)\n\tif category != \"\" {\n\t\tparams.Add(\"cat\", category)\n\t}\n\tnzburl.RawQuery = params.Encode()\n\tSabR := new(SabResponse)\n\terr = JsonFromURL(nzburl.String(), SabR)\n\tif err != nil {\n\t\tlog.Print(\"SABSendURL:JSON:\", err)\n\t\treturn \"\"\n\t}\n\n\tif SabR.Status {\n\t\tif len(SabR.Nzo_ids) > 0 {\n\t\t\t//sleep for 250ms after a positive add\n\t\t\ttime.Sleep(250 * time.Millisecond)\n\t\t\treturn SabR.Nzo_ids[0]\n\t\t} else {\n\t\t\treturn \"\"\n\t\t}\n\t} else {\n\t\tlog.Print(\"SABSendURL:SendReturnedError:\", SabR.SabErr)\n\t\treturn \"\"\n\t}\n}", "func DBSUrl(inst string) string {\n\turl := bareDbsUrl(inst)\n\tif strings.HasPrefix(url, \"http\") {\n\t\treturn url\n\t}\n\treturn utils.AdjustUrl(fmt.Sprintf(\"%s/%s\", FrontendURL, url))\n}", "func parse(target, protocol, port string) string {\n\tline := target\n\n\t// if no protocol, no port was specified\n\tif len(protocol) == 0 && len(port) == 0 {\n\t\t// scope only http/https\n\t\tline = `http(s)?://` + line\n\n\t\t// if port was specified but no protocol\n\t} else if len(protocol) == 0 && len(port) > 0 {\n\t\t// scope any protocol\n\t\tline = `\\w+://` + line\n\t} else {\n\n\t}\n\n\t// escape '.'\n\tline = strings.Replace(line, \".\", `\\.`, -1)\n\t// escape '/'\n\tline = strings.Replace(line, \"/\", `\\/`, -1)\n\t// replace wildcard\n\tline = strings.Replace(line, \"*\", `[\\S]*`, -1)\n\t// Zap needs this to scope URL params\n\tline = `^` + line + `[\\S]*$`\n\n\treturn line\n}", "func parseSink(sink string) (string, string, string) {\n\tparts := strings.SplitN(sink, \":\", 3)\n\tswitch {\n\tcase len(parts) == 1:\n\t\treturn \"ksvc\", parts[0], \"\"\n\tcase parts[0] == \"http\" || parts[0] == \"https\":\n\t\treturn \"\", sink, \"\"\n\tcase len(parts) == 3:\n\t\treturn parts[0], parts[1], parts[2]\n\tdefault:\n\t\treturn parts[0], parts[1], \"\"\n\t}\n}", "func parseDOption(s string) (veth vEth, err error) {\n\tn := strings.Split(s, \":\")\n\tif len(n) != 3 && len(n) != 2 {\n\t\terr = fmt.Errorf(\"failed to parse %s\", s)\n\t\treturn\n\t}\n\n\tveth.nsName, err = getDockerContainerNS(n[0])\n\tif err != nil {\n\t\tlogger(fmt.Sprintf(\"%v\", err))\n\t}\n\n\tveth.linkName = n[1]\n\n\tif len(n) == 3 {\n\t\tip, mask, err2 := net.ParseCIDR(n[2])\n\t\tif err2 != nil {\n\t\t\terr = fmt.Errorf(\"failed to parse IP addr %s: %v\",\n\t\t\t\tn[2], err2)\n\t\t\treturn\n\t\t}\n\t\tveth.ipAddr.IP = ip\n\t\tveth.ipAddr.Mask = mask.Mask\n\t\tveth.withIPAddr = true\n\t} else {\n\t\tveth.withIPAddr = false\n\t}\n\n\treturn\n}", "func parseContainerID(ID string) string {\n\ti := strings.Index(ID, \"://\")\n\tif i < 0 {\n\t\treturn ID\n\t}\n\treturn ID[i+3:]\n}", "func ParseHost(host string) (string, string, string, error) {\n\tprotoAddrParts := strings.SplitN(host, \"://\", 2)\n\tif len(protoAddrParts) == 1 {\n\t\treturn \"\", \"\", \"\", fmt.Errorf(\"unable to parse storm host `%s`\", host)\n\t}\n\n\tvar basePath string\n\tproto, addr := protoAddrParts[0], protoAddrParts[1]\n\tif proto == \"tcp\" {\n\t\tparsed, err := url.Parse(\"tcp://\" + addr)\n\t\tif err != nil {\n\t\t\treturn \"\", \"\", \"\", err\n\t\t}\n\t\taddr = parsed.Host\n\t\tbasePath = parsed.Path\n\t}\n\treturn proto, addr, basePath, nil\n}", "func MakeDBCred(str string) string {\n\tmatches := dbURLRegex.FindStringSubmatch(str)\n\n\tif len(matches) == 10 {\n\t\tmatchIdxs := dbURLRegex.FindStringSubmatchIndex(str)\n\t\tsubstr := str[matchIdxs[0]:matchIdxs[1]]\n\t\tfor idx := 1; idx < len(matches); idx++ {\n\t\t\tif matches[idx] != \"\" {\n\t\t\t\tif strings.Index(matches[idx], \"user=\") == 0 {\n\t\t\t\t\tsubstr = strings.Replace(substr, matches[idx], \"user=****\", 1)\n\t\t\t\t} else if strings.Index(matches[idx], \"password=\") == 0 {\n\t\t\t\t\tsubstr = strings.Replace(substr, matches[idx], \"password=****\", 1)\n\t\t\t\t} else {\n\t\t\t\t\tsubstr = strings.Replace(substr, matches[idx], \"****\", 1)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tstr = str[:matchIdxs[0]] + substr + str[matchIdxs[1]:len(str)]\n\t}\n\treturn str\n}", "func parseNotifyURL(u string) (notifyURL *model.NotifyURL, err error) {\n\tvar parsedURL *url.URL\n\tparsedURL, err = url.Parse(u)\n\tif err != nil {\n\t\treturn\n\t}\n\tnotifyURL = &model.NotifyURL{\n\t\tRawURL: u,\n\t\tSchema: parsedURL.Scheme,\n\t\tHost: parsedURL.Host,\n\t\tPath: parsedURL.Path,\n\t\tQuery: parsedURL.Query(),\n\t}\n\treturn\n}", "func (_BaseAccessWallet *BaseAccessWalletFilterer) ParseDbgAddress(log types.Log) (*BaseAccessWalletDbgAddress, error) {\n\tevent := new(BaseAccessWalletDbgAddress)\n\tif err := _BaseAccessWallet.contract.UnpackLog(event, \"dbgAddress\", log); err != nil {\n\t\treturn nil, err\n\t}\n\tevent.Raw = log\n\treturn event, nil\n}", "func getBaiduPanAddress(downloadAddress string, outputurl chan []string) {\n\n\tce := strings.Split(downloadAddress, \"/\")\n\ta := ce[:len(ce)-2]\n\tai := strings.Join(a, \"/\")\n\tb := ce[len(ce)-1]\n\tdownloadAddress = fmt.Sprintf(\"%s/download/%s\", ai, b)\n\t//\n\tc := colly.NewCollector()\n\t// Find and visit all links\n\tif err := c.SetCookies(downloadAddress, []*http.Cookie{\n\t\t{\n\t\t\tName: \"Nobird_DownLoad\",\n\t\t\tValue: cookieValue,\n\t\t},\n\t}); err != nil {\n\t\tglog.Error(err)\n\t}\n\tpool := make([]string,2)\n\t// 去得到百度的url\n\tc.OnHTML(\".con\", func(e *colly.HTMLElement) {\n\t\tif e.Index == 1 {\n\t\t\tpool[0] = e.DOM.Text()\n\t\t}\n\t})\n\t//去得到歌名\n\tc.OnHTML(\"h1\", func(e *colly.HTMLElement) {\n\t\tre := regexp.MustCompile(\"《.+》\")\n\t\tpool[1] = re.FindString(e.Text)\n\t})\n\tif err := c.Visit(downloadAddress); err != nil {\n\t\tglog.Error(err)\n\t}\n\toutputurl <- pool\n}", "func (d Decoder) DecodeString(dst interface{}, src string) error {\n\tvs, err := url.ParseQuery(src)\n\tif err != nil {\n\t\treturn err\n\t}\n\tv := reflect.ValueOf(dst)\n\treturn d.decodeNode(v, parseValues(d.d, d.e, vs, canIndexOrdinally(v)))\n}", "func parseConnSpec(connStr string) (out connSpec, err error) {\n\tpartMatcher := regexp.MustCompile(`((.*):\\/\\/)?(([^\\/?:]*)(:([^\\/?:@]*))?@)?([^\\/?]*)(\\/([^\\?]*))?(\\?(.*))?`)\n\thostMatcher := regexp.MustCompile(`([^;\\,\\:]+)(:([0-9]*))?(;\\,)?`)\n\tparts := partMatcher.FindStringSubmatch(connStr)\n\n\tif parts[2] != \"\" {\n\t\t(&out.Scheme).load(parts[2])\n\t\tif out.Scheme == csInvalid {\n\t\t\terr = fmt.Errorf(\"Unknown scheme '%s'\", parts[2])\n\t\t\treturn\n\t\t}\n\t\tout.hasExplicitScheme = true\n\t} else {\n\t\tout.Scheme = csPlainMcd\n\t}\n\n\tif parts[7] != \"\" {\n\t\thosts := hostMatcher.FindAllStringSubmatch(parts[7], -1)\n\t\tfor _, hostInfo := range hosts {\n\t\t\tport := 0\n\t\t\tif hostInfo[3] != \"\" {\n\t\t\t\tport, err = strconv.Atoi(hostInfo[3])\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tout.hasExplicitPort = true\n\t\t\t}\n\t\t\terr = out.addRawHost(hostInfo[1], port)\n\t\t\tif err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}\n\n\tif len(out.HttpHosts) == 0 && len(out.MemcachedHosts) == 0 {\n\t\terr = out.addRawHost(\"127.0.0.1\", 0)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\n\tif parts[9] != \"\" {\n\t\tout.Bucket, err = url.QueryUnescape(parts[9])\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\n\tif parts[11] != \"\" {\n\t\tout.Options, err = url.ParseQuery(parts[11])\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\n\treturn\n}", "func splitBlobURI(URI string) (string, string, string, error) {\n\turi, err := url.Parse(URI)\n\tif err != nil {\n\t\treturn \"\", \"\", \"\", err\n\t}\n\n\taccountName := strings.Split(uri.Host, \".\")[0]\n\turlParts := strings.Split(uri.Path, \"/\")\n\n\tcontainerName := urlParts[1]\n\tblobPath := strings.Join(urlParts[2:], \"/\")\n\n\treturn accountName, containerName, blobPath, nil\n}", "func ParseUserString(s string) User {\n\tif s == \"\" {\n\t\treturn User{}\n\t}\n\n\tsplitForHost := strings.Split(s, \"@\")\n\n\tvar host string\n\tif len(splitForHost) > 1 {\n\t\thost = splitForHost[1]\n\t}\n\n\tvar nick string\n\tvar ident string\n\tsplitForIdent := strings.Split(splitForHost[0], \"!\")\n\tif len(splitForIdent) > 1 {\n\t\tident = splitForIdent[1]\n\t}\n\tnick = splitForIdent[0]\n\n\treturn User{\n\t\tNick: nick,\n\t\tIdent: ident,\n\t\tHost: host,\n\t}\n}", "func MakeList(d string) (x []string) {\n\tl := strings.Split(d, \"vmess://\")\n\tfor i, item := range l {\n\t\tvar itemLen int\n\t\titemLen = len(item)\n\t\tif itemLen > 0 {\n\t\t\tvar (\n\t\t\t\tnewstr string\n\t\t\t\tv string\n\t\t\t\tother bool\n\t\t\t)\n\n\t\t\tother = strings.Contains(item, \"?remarks=\")\n\t\t\tif other {\n\t\t\t\tstrsss := strings.Split(item, \"?remarks=\")\n\t\t\t\tcodes, err := base64.RawStdEncoding.DecodeString(strsss[0])\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn x\n\t\t\t\t}\n\n\t\t\t\tnewstr := string(codes)\n\t\t\t\t// fmt.Println(newstr)\n\t\t\t\tblen := len(newstr)\n\t\t\t\ta := strings.Index(newstr, \":\")\n\t\t\t\tb := strings.Index(newstr, \"@\")\n\t\t\t\tc := strings.LastIndex(newstr, \":\")\n\t\t\t\tuuid := newstr[a+1 : b]\n\t\t\t\thost := newstr[b+1 : c]\n\t\t\t\tport := newstr[c+1 : blen]\n\n\t\t\t\tparams := strsss[1]\n\n\t\t\t\t// fmt.Println(params)\n\t\t\t\tf := strings.Index(params, \"obfs=\")\n\n\t\t\t\tvar (\n\t\t\t\t\to int\n\t\t\t\t\te int\n\t\t\t\t\tg int\n\t\t\t\t\tpath string\n\t\t\t\t\ttls string\n\t\t\t\t\tobfs string\n\t\t\t\t\tobfsParam string = \"\"\n\t\t\t\t)\n\t\t\t\tif strings.Contains(params, \"path=\") {\n\t\t\t\t\te = strings.Index(params, \"path=\")\n\t\t\t\t\tpath = params[e+5 : f-1]\n\t\t\t\t}\n\t\t\t\tif strings.Contains(params, \"tls=\") {\n\t\t\t\t\tg = strings.Index(params, \"tls=\")\n\t\t\t\t\ttls = params[g+4:]\n\t\t\t\t\tobfs = params[f+5 : g-1]\n\t\t\t\t} else {\n\t\t\t\t\tobfs = params[f+5:]\n\t\t\t\t}\n\n\t\t\t\tif strings.Contains(params, \"obfsParam=\") {\n\t\t\t\t\to = strings.Index(params, \"obfsParam=\")\n\t\t\t\t\tif strings.Contains(params, \"path=\") {\n\t\t\t\t\t\tobfsParam = params[o+10 : e-1]\n\t\t\t\t\t} else {\n\t\t\t\t\t\tobfsParam = params[o+10 : strings.Index(params, \"&obfs=\")-1]\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\tif obfs == \"websocket\" {\n\t\t\t\t\tobfs = \"ws\"\n\t\t\t\t} else {\n\t\t\t\t\tobfs = \"tcp\"\n\t\t\t\t}\n\n\t\t\t\tif tls == \"1\" {\n\t\t\t\t\ttls = \"tls\"\n\t\t\t\t} else {\n\t\t\t\t\ttls = \"tcp\"\n\t\t\t\t}\n\t\t\t\tif len(obfsParam) <= 0 {\n\t\t\t\t\tobfsParam = host\n\t\t\t\t}\n\t\t\t\tif strings.Index(obfsParam, \"/\") == 0 {\n\t\t\t\t\tpath = obfsParam\n\t\t\t\t\tobfsParam = host\n\t\t\t\t}\n\t\t\t\t// fmt.Println(uuid, host, port, path, obfs, tls, obfsParam)\n\t\t\t\t// fmt.Println(host, obfs, obfsParam)\n\t\t\t\t// log.Println(uuid, host, port, param, path, obfs, tls)\n\t\t\t\tcumv := strconv.Itoa(i)\n\t\t\t\tmyname := strings.Join([]string{\"公益节点\", cumv}, \"-\")\n\t\t\t\tvjson := &Vary{\n\t\t\t\t\tVersion: \"2\",\n\t\t\t\t\tHost: obfsParam,\n\t\t\t\t\tPath: path,\n\t\t\t\t\tTLS: tls,\n\t\t\t\t\tPs: myname,\n\t\t\t\t\tAdd: host,\n\t\t\t\t\tProt: port,\n\t\t\t\t\tID: uuid,\n\t\t\t\t\tAid: \"1\",\n\t\t\t\t\tNet: obfs,\n\t\t\t\t\tType: \"null\",\n\t\t\t\t}\n\t\t\t\tbytes, err := json.Marshal(vjson)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn x\n\t\t\t\t}\n\t\t\t\t// fmt.Println(string(bytes))\n\t\t\t\tv = strings.Join([]string{\"vmess:\", base64.StdEncoding.EncodeToString(bytes)}, \"//\")\n\t\t\t\tx = append(x, v)\n\n\t\t\t} else {\n\t\t\t\tvar str []byte = []byte(item)\n\t\t\t\tdecodeBytes := make([]byte, base64.StdEncoding.DecodedLen(len(str))) // 计算解码后的长度\n\t\t\t\tbase64.StdEncoding.Decode(decodeBytes, str)\n\t\t\t\treg := regexp.MustCompile(`\"ps\" :\"翻墙党fanqiangdang.com\",\"\" :`)\n\t\t\t\tnewstr = reg.ReplaceAllString(string(decodeBytes), `\"ps\":`)\n\t\t\t\tif !strings.Contains(newstr, \"}\") {\n\t\t\t\t\tnewstr = strings.Join([]string{newstr, \"}\"}, \"\")\n\t\t\t\t}\n\t\t\t\tbjson := StrToJsons(newstr)\n\t\t\t\tvar strtobyte []byte = []byte(jsonToStr(bjson))\n\t\t\t\tv = strings.Join([]string{\"vmess:\", base64.StdEncoding.EncodeToString(strtobyte)}, \"//\")\n\t\t\t\tx = append(x, v)\n\t\t\t}\n\t\t}\n\t}\n\treturn\n}", "func parseURL(rawurl string) (scheme string, address string, path string, err error) {\n\turi, err := url.Parse(rawurl)\n\tif err != nil {\n\t\treturn uri.Scheme, uri.Host, uri.Path, err\n\t}\n\n\tscheme = uri.Scheme\n\n\tswitch uri.Scheme {\n\tcase \"unix\":\n\t\tresult := strings.Split(uri.Path, \";\")\n\t\taddress = result[0]\n\t\tif len(result) > 1 {\n\t\t\tpath = result[1]\n\t\t}\n\tdefault:\n\t\taddress = uri.Host\n\t\tpath = uri.Path\n\t}\n\n\treturn\n}", "func parseBwtestParameters(s string) BwtestParameters {\n\t// Since \"-\" is not part of the parse string, all numbers read are positive\n\tre := regexp.MustCompile(\"[0-9]+\")\n\ta := re.FindAllString(s, -1)\n\tif len(a) != 3 {\n\t\tCheck(fmt.Errorf(\"Incorrect number of arguments, need 3 values for bwtestparameters\"))\n\t}\n\n\ta1, err := strconv.Atoi(a[0])\n\tCheck(err)\n\td := time.Second * time.Duration(a1)\n\tif d > MaxDuration {\n\t\tCheck(fmt.Errorf(\"Duration is exceeding MaxDuration:\", a1, \">\", MaxDuration/time.Second))\n\t}\n\ta2, err := strconv.Atoi(a[1])\n\tCheck(err)\n\tif a2 < MinPacketSize {\n\t\ta2 = MinPacketSize\n\t}\n\tif a2 > MaxPacketSize {\n\t\ta2 = MaxPacketSize\n\t}\n\ta3, err := strconv.Atoi(a[2])\n\tCheck(err)\n\tkey := prepareAESKey()\n\treturn BwtestParameters{d, a2, a3, key, 0}\n}", "func ParseURL(redisURL string) (*Pool, error) {\n\tpool := Pool{}\n\tu, err := url.Parse(redisURL)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tpool.Dial, err = dialURL(u)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tq := u.Query()\n\n\tif v, ok := q[\"clock-interval\"]; ok && len(v) > 0 {\n\t\tif d, _ := time.ParseDuration(v[0]); d > 0 {\n\t\t\tpool.ClockInterval = d\n\t\t}\n\t}\n\n\tif v, ok := q[\"max-idle-time\"]; ok && len(v) > 0 {\n\t\tif d, _ := time.ParseDuration(v[0]); d > 0 {\n\t\t\tpool.MaxIdleTime = d\n\t\t}\n\t}\n\tif v, ok := q[\"max-connections\"]; ok && len(v) > 0 {\n\t\tif size, _ := strconv.Atoi(v[0]); size > 0 {\n\t\t\tpool.MaxConnections = size\n\t\t}\n\t}\n\tif v, ok := q[\"min-connections\"]; ok && len(v) > 0 {\n\t\tif size, _ := strconv.Atoi(v[0]); size > 0 {\n\t\t\tpool.MinConnections = size\n\t\t}\n\t}\n\n\treturn &pool, nil\n}", "func ParseEndpoint(endpoint string) (string, string, string, error) {\n\turl, err := url.Parse(endpoint)\n\tif err != nil {\n\t\treturn \"\", \"\", \"\", err\n\t}\n\tsplits := strings.Split(url.Host, \":\")\n\tif len(splits) != 2 {\n\t\treturn \"\", \"\", \"\", fmt.Errorf(\"invalid host: %s\", url.Host)\n\t}\n\treturn url.Scheme, splits[0], splits[1], nil\n}", "func (e *EventParser) ParseBitbucketServerPullEvent(body []byte) (pull models.PullRequest, baseRepo models.Repo, headRepo models.Repo, user models.User, err error) {\n\tvar event bitbucketserver.PullRequestEvent\n\tif err = json.Unmarshal(body, &event); err != nil {\n\t\terr = errors.Wrap(err, \"parsing json\")\n\t\treturn\n\t}\n\tif err = validator.New().Struct(event); err != nil {\n\t\terr = errors.Wrapf(err, \"API response %q was missing fields\", string(body))\n\t\treturn\n\t}\n\tpull, baseRepo, headRepo, user, err = e.parseCommonBitbucketServerEventData(event.CommonEventData)\n\treturn\n}" ]
[ "0.5207556", "0.50693107", "0.5065838", "0.5031698", "0.49781075", "0.4944224", "0.48886964", "0.48665693", "0.479547", "0.479547", "0.47776258", "0.47523627", "0.47295797", "0.46874997", "0.4674082", "0.4671452", "0.4596945", "0.45955494", "0.45829305", "0.45667833", "0.45657474", "0.4553619", "0.45430726", "0.454248", "0.45383453", "0.4534099", "0.4522121", "0.4509148", "0.45079833", "0.450182", "0.45003685", "0.4495407", "0.44865844", "0.44836384", "0.44815394", "0.44797167", "0.44761708", "0.44728798", "0.44722828", "0.44718844", "0.44618368", "0.4456958", "0.44558686", "0.444801", "0.44464195", "0.44405273", "0.4429741", "0.442849", "0.44234833", "0.4419592", "0.44049308", "0.4403096", "0.43974063", "0.43817887", "0.43810806", "0.43699992", "0.43686536", "0.4365564", "0.43655196", "0.43654478", "0.43603006", "0.43488574", "0.4340129", "0.4339433", "0.43393323", "0.43334007", "0.43329576", "0.4332106", "0.43300316", "0.43257985", "0.43109548", "0.43033102", "0.43011448", "0.42981958", "0.4298193", "0.42981187", "0.4296533", "0.4291782", "0.42891696", "0.428801", "0.42820838", "0.4277119", "0.4274874", "0.42718497", "0.42639756", "0.42615527", "0.42613283", "0.42559782", "0.42524973", "0.42456445", "0.42413485", "0.42408517", "0.4240411", "0.42402917", "0.42382413", "0.42369735", "0.42312178", "0.42301613", "0.4221413", "0.421468" ]
0.7711171
0
GetAPIPattern generate the API path pattern. /application/interface/version
func GetAPIPattern(bkConfig config.DubboBackendConfig) string { return strings.Join([]string{"/" + bkConfig.ApplicationName, bkConfig.Interface, bkConfig.Version}, constant.PathSlash) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *scaleClient) apiPathFor(groupVer schema.GroupVersion) string {\n\t// we need to set the API path based on GroupVersion (defaulting to the legacy path if none is set)\n\t// TODO: we \"cheat\" here since the API path really only depends on group ATM, but this should\n\t// *probably* take GroupVersionResource and not GroupVersionKind.\n\tapiPath := c.apiPathResolverFunc(groupVer.WithKind(\"\"))\n\tif apiPath == \"\" {\n\t\tapiPath = \"/api\"\n\t}\n\n\treturn restclient.DefaultVersionedAPIPath(apiPath, groupVer)\n}", "func (c *Client) getAPIPath(ctx context.Context, p string, query url.Values) string {\n\tvar apiPath string\n\tif c.negotiateVersion && !c.negotiated {\n\t\tc.NegotiateAPIVersion(ctx)\n\t}\n\tif c.version != \"\" {\n\t\tv := strings.TrimPrefix(c.version, \"v\")\n\t\tapiPath = path.Join(c.basePath, \"/v\"+v, p)\n\t} else {\n\t\tapiPath = path.Join(c.basePath, p)\n\t}\n\treturn (&url.URL{Path: apiPath, RawQuery: query.Encode()}).String()\n}", "func VersionAPI(path string, version Version) (API, error) {\n\treturn nil, fmt.Errorf(\"not implemented\")\n}", "func HandlerApi(w http.ResponseWriter, r *http.Request) {\n\thttp.Header.Add(w.Header(), \"content-type\", \"application/json\")\n\tparts := strings.Split(r.URL.Path, \"/\")\n\tif len(parts) == 4 && parts[3] == \"\" {\n\t\tapi := _struct.Information{_struct.Uptime(), _struct.Description, _struct.Version}\n\t\tjson.NewEncoder(w).Encode(api)\n\t} else {\n\t\thttp.Error(w, http.StatusText(404), 404)\n\t}\n}", "func (Interface) ApiRoutes() []string {\n\treturn []string{\n\t\t\"/api/objects/interface/bridge/\",\n\t\t\"/api/objects/interface/bridge/{ref}\",\n\t\t\"/api/objects/interface/bridge/{ref}/usedby\",\n\t\t\"/api/objects/interface/ethernet/\",\n\t\t\"/api/objects/interface/ethernet/{ref}\",\n\t\t\"/api/objects/interface/ethernet/{ref}/usedby\",\n\t\t\"/api/objects/interface/group/\",\n\t\t\"/api/objects/interface/group/{ref}\",\n\t\t\"/api/objects/interface/group/{ref}/usedby\",\n\t\t\"/api/objects/interface/ppp3g/\",\n\t\t\"/api/objects/interface/ppp3g/{ref}\",\n\t\t\"/api/objects/interface/ppp3g/{ref}/usedby\",\n\t\t\"/api/objects/interface/pppmodem/\",\n\t\t\"/api/objects/interface/pppmodem/{ref}\",\n\t\t\"/api/objects/interface/pppmodem/{ref}/usedby\",\n\t\t\"/api/objects/interface/pppoa/\",\n\t\t\"/api/objects/interface/pppoa/{ref}\",\n\t\t\"/api/objects/interface/pppoa/{ref}/usedby\",\n\t\t\"/api/objects/interface/pppoe/\",\n\t\t\"/api/objects/interface/pppoe/{ref}\",\n\t\t\"/api/objects/interface/pppoe/{ref}/usedby\",\n\t\t\"/api/objects/interface/tunnel/\",\n\t\t\"/api/objects/interface/tunnel/{ref}\",\n\t\t\"/api/objects/interface/tunnel/{ref}/usedby\",\n\t\t\"/api/objects/interface/vlan/\",\n\t\t\"/api/objects/interface/vlan/{ref}\",\n\t\t\"/api/objects/interface/vlan/{ref}/usedby\",\n\t}\n}", "func (cli *Client) getAPIPath(p string, query url.Values) string {\n\tvar apiPath string\n\tif cli.version != \"\" {\n\t\tv := strings.TrimPrefix(cli.version, \"v\")\n\t\tapiPath = fmt.Sprintf(\"%s/v%s%s\", cli.basePath, v, p)\n\t} else {\n\t\tapiPath = fmt.Sprintf(\"%s%s\", cli.basePath, p)\n\t}\n\n\tu := &url.URL{\n\t\tPath: apiPath,\n\t}\n\tif len(query) > 0 {\n\t\tu.RawQuery = query.Encode()\n\t}\n\treturn u.String()\n}", "func (c *AlphaSOCClient) getAPIPath(path string, query url.Values) string {\n\tif query == nil {\n\t\treturn fmt.Sprintf(\"%s/%s/%s\", c.host, c.version, path)\n\t}\n\treturn fmt.Sprintf(\"%s/%s/%s?%s\", c.host, c.version, path, query.Encode())\n}", "func GenerateBrokerAPIPath(scheme string, host string, apiPath string,\n\treplacer *strings.Replacer) string {\n\tgenPath := url.URL{\n\t\tScheme: scheme,\n\t\tHost: host,\n\t\tPath: apiPath,\n\t}\n\tif replacer != nil {\n\t\treturn replacer.Replace(genPath.String())\n\t}\n\treturn genPath.String()\n}", "func api(api API) []string {\n\turi := api.URI\n\trequire(\"api\", uri)\n\treturn []string{\"api\", uri}\n}", "func (r *Request) GeneratePath(URLPattern string) {\n\tif len(r.Params) == 0 {\n\t\tr.Path = URLPattern\n\t\treturn\n\t}\n\tbuff := []byte(URLPattern)\n\tfor k, v := range r.Params {\n\t\tkey := []byte{}\n\t\tkey = append(key, \"{{.\"...)\n\t\tkey = append(key, k...)\n\t\tkey = append(key, \"}}\"...)\n\t\tbuff = bytes.Replace(buff, key, []byte(v), -1)\n\t}\n\tr.Path = string(buff)\n}", "func getEndpoint(suffix string) string {\n\treturn fmt.Sprintf(endpointFormat, ApiVersion, suffix)\n}", "func VersionDebugAPIPath() string {\n\treturn \"/api/debug/version\"\n}", "func (f Freckle) api(path string, ps Parameters) string {\n\tu := fmt.Sprintf(\"%s%s\", f.base, path)\n\tif ps != nil && len(ps) > 0 {\n\t\tvar v url.Values = make(url.Values)\n\t\tfor key, value := range ps {\n\t\t\tv.Set(key, value)\n\t\t}\n\t\tu = fmt.Sprintf(\"%s?%s\", u, v.Encode())\n\t}\n\treturn u\n}", "func getDesiredPattern(pattern string) string {\n\twant := []string{}\n\tfor _, token := range strings.Split(pattern, \"/\") {\n\t\tif strings.HasPrefix(token, \"{\") && strings.HasSuffix(token, \"}\") {\n\t\t\tvarname := token[1 : len(token)-1]\n\t\t\twant = append(want, fmt.Sprintf(\"{%s}\", strings.TrimSuffix(strcase.SnakeCase(varname), \"_id\")))\n\t\t} else {\n\t\t\twant = append(want, strcase.LowerCamelCase(token))\n\t\t}\n\t}\n\treturn strings.Join(want, \"/\")\n}", "func handleAPI(req typhon.Request) typhon.Response {\n\tparts := reAPI.FindStringSubmatch(req.URL.Path)\n\tif len(parts) != 3 {\n\t\treturn typhon.Response{Error: terrors.NotFound(\"bad_endpoint\", \"Unable to determine API endpoint.\", nil)}\n\t}\n\n\treturn handle(req, \"s-api-\"+parts[1], parts[2])\n}", "func (c *PrivateClient) getAPIPath(query url.Values) string {\n\t// set default query values if not specified\n\tif query.Get(\"api_version\") == \"\" {\n\t\tquery.Set(\"api_version\", \"1.0\")\n\t}\n\tif query.Get(\"input\") == \"\" {\n\t\tquery.Set(\"input\", \"3\")\n\t}\n\tif query.Get(\"api_token\") == \"\" {\n\t\tquery.Set(\"api_token\", \"null\")\n\t}\n\treturn (&url.URL{Path: c.path, RawQuery: query.Encode()}).String()\n}", "func RegisterGroupAPIRoute(basePath string, ginEngine *gin.Engine, controllers []IBaseController) {\n\tif !strings.HasPrefix(basePath, \"/\") {\n\t\tbasePath = \"/\" + basePath\n\t}\n\tg := ginEngine.Group(basePath)\n\t{\n\t\troutesControllerMapping(g, controllers)\n\t}\n}", "func (Http) ApiRoutes() []string {\n\treturn []string{\n\t\t\"/api/objects/http/cff_action/\",\n\t\t\"/api/objects/http/cff_action/{ref}\",\n\t\t\"/api/objects/http/cff_action/{ref}/usedby\",\n\t\t\"/api/objects/http/cff_profile/\",\n\t\t\"/api/objects/http/cff_profile/{ref}\",\n\t\t\"/api/objects/http/cff_profile/{ref}/usedby\",\n\t\t\"/api/objects/http/device_auth/\",\n\t\t\"/api/objects/http/device_auth/{ref}\",\n\t\t\"/api/objects/http/device_auth/{ref}/usedby\",\n\t\t\"/api/objects/http/domain_regex/\",\n\t\t\"/api/objects/http/domain_regex/{ref}\",\n\t\t\"/api/objects/http/domain_regex/{ref}/usedby\",\n\t\t\"/api/objects/http/exception/\",\n\t\t\"/api/objects/http/exception/{ref}\",\n\t\t\"/api/objects/http/exception/{ref}/usedby\",\n\t\t\"/api/objects/http/group/\",\n\t\t\"/api/objects/http/group/{ref}\",\n\t\t\"/api/objects/http/group/{ref}/usedby\",\n\t\t\"/api/objects/http/local_site/\",\n\t\t\"/api/objects/http/local_site/{ref}\",\n\t\t\"/api/objects/http/local_site/{ref}/usedby\",\n\t\t\"/api/objects/http/lsl_tag/\",\n\t\t\"/api/objects/http/lsl_tag/{ref}\",\n\t\t\"/api/objects/http/lsl_tag/{ref}/usedby\",\n\t\t\"/api/objects/http/pac_file/\",\n\t\t\"/api/objects/http/pac_file/{ref}\",\n\t\t\"/api/objects/http/pac_file/{ref}/usedby\",\n\t\t\"/api/objects/http/parent_proxy/\",\n\t\t\"/api/objects/http/parent_proxy/{ref}\",\n\t\t\"/api/objects/http/parent_proxy/{ref}/usedby\",\n\t\t\"/api/objects/http/profile/\",\n\t\t\"/api/objects/http/profile/{ref}\",\n\t\t\"/api/objects/http/profile/{ref}/usedby\",\n\t\t\"/api/objects/http/sp_category/\",\n\t\t\"/api/objects/http/sp_category/{ref}\",\n\t\t\"/api/objects/http/sp_category/{ref}/usedby\",\n\t\t\"/api/objects/http/sp_subcat/\",\n\t\t\"/api/objects/http/sp_subcat/{ref}\",\n\t\t\"/api/objects/http/sp_subcat/{ref}/usedby\",\n\t}\n}", "func (a *App) apiRoute(e *echo.Echo) error {\n\tapiUrl, err := a.conf.Get(\"api-url\")\n\tif err != nil {\n\t\treturn err\n\t}\n\t//General API\n\tg := e.Group(apiUrl.(string))\n\t//auth\n\tg.Post(\"/auth\", a.login)\n\t//oauth\n\ta.fb, err = oauth.NewFacebook(a)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfbg := g.Group(\"/oauth\")\n\tfbg.Get(\"/facebook\", a.facebookAuth)\n\tfbg.Get(\"/facebook/redirect\", a.facebookRedirect)\n\t//tasks\n\ttokenizer := jwt.NewTokenizer(a)\n\ttg := g.Group(\"/tasks\", tokenizer.Check())\n\ttg.Post(\"\", a.create)\n\ttg.Get(\"/:id\", a.retrieve)\n\ttg.Get(\"\", a.retrieveAll)\n\ttg.Put(\"/:id\", a.update)\n\ttg.Delete(\"/:id\", a.delete)\n\treturn nil\n}", "func apiURL(cmd, path string, args []string) string {\n\tif len(args) > 0 {\n\t\tvar arglist string\n\t\tfor i := 0; i < len(args); i++ {\n\t\t\targlist += fmt.Sprintf(\"&%s\", args[i])\n\t\t}\n\t\treturn fmt.Sprintf(\"http://%s/api/v0/%s?arg=%s&%s\", defaultIPFSHost, cmd, path, arglist)\n\t}\n\treturn fmt.Sprintf(\"http://%s/api/v0/%s?arg=%s\", defaultIPFSHost, cmd, path)\n}", "func (o BlobReferenceInputDataSourceResponseOutput) PathPattern() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BlobReferenceInputDataSourceResponse) *string { return v.PathPattern }).(pulumi.StringPtrOutput)\n}", "func (o BlobOutputDataSourceResponseOutput) PathPattern() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BlobOutputDataSourceResponse) *string { return v.PathPattern }).(pulumi.StringPtrOutput)\n}", "func API(shutdown chan os.Signal, log *log.Logger) http.Handler {\n\n\t// Construct the web.App which holds all routes as well as common Middleware.\n\tapp := web.NewApp(shutdown, log, webcontext.Env_Dev, mid.Logger(log))\n\n\tapp.Handle(\"GET\", \"/swagger/\", saasSwagger.WrapHandler)\n\tapp.Handle(\"GET\", \"/swagger/*\", saasSwagger.WrapHandler)\n\n\t/*\n\t\tOr can use SaasWrapHandler func with configurations.\n\t\turl := saasSwagger.URL(\"http://localhost:1323/swagger/doc.json\") //The url pointing to API definition\n\t\te.GET(\"/swagger/*\", saasSwagger.SaasWrapHandler(url))\n\t*/\n\n\treturn app\n}", "func (c *scaleClient) pathAndVersionFor(resource schema.GroupResource) (string, schema.GroupVersionResource, error) {\n\tgvr, err := c.mapper.ResourceFor(resource.WithVersion(\"\"))\n\tif err != nil {\n\t\treturn \"\", gvr, fmt.Errorf(\"unable to get full preferred group-version-resource for %s: %v\", resource.String(), err)\n\t}\n\n\tgroupVer := gvr.GroupVersion()\n\n\treturn c.apiPathFor(groupVer), gvr, nil\n}", "func (o BlobStreamInputDataSourceResponseOutput) PathPattern() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BlobStreamInputDataSourceResponse) *string { return v.PathPattern }).(pulumi.StringPtrOutput)\n}", "func EndpointPattern(urlPattern string, configFunc func()) {\n\tpathRegex := regexp.MustCompile(urlPattern)\n\tEndpointForCondition(predicate.PathMatches(pathRegex), configFunc)\n}", "func RegisterAPIRoutes(router *mux.Router) {\n\trouter.HandleFunc(\"/api/v1/word/today/\", GetWordToday).Methods(\"GET\")\n}", "func (*InterfaceGroups) GetPath() string { return \"/api/objects/interface/group/\" }", "func (Condition) ApiRoutes() []string {\n\treturn []string{\n\t\t\"/api/objects/condition/group/\",\n\t\t\"/api/objects/condition/group/{ref}\",\n\t\t\"/api/objects/condition/group/{ref}/usedby\",\n\t\t\"/api/objects/condition/objref/\",\n\t\t\"/api/objects/condition/objref/{ref}\",\n\t\t\"/api/objects/condition/objref/{ref}/usedby\",\n\t}\n}", "func buildApiURL(sig, protocol string) string {\n\treturn protocol + API_HOST + API_ENDPOIT + SIG_GET_KEY + sig\n}", "func (Itfparams) ApiRoutes() []string {\n\treturn []string{\n\t\t\"/api/objects/itfparams/bridge_port/\",\n\t\t\"/api/objects/itfparams/bridge_port/{ref}\",\n\t\t\"/api/objects/itfparams/bridge_port/{ref}/usedby\",\n\t\t\"/api/objects/itfparams/group/\",\n\t\t\"/api/objects/itfparams/group/{ref}\",\n\t\t\"/api/objects/itfparams/group/{ref}/usedby\",\n\t\t\"/api/objects/itfparams/link_aggregation_group/\",\n\t\t\"/api/objects/itfparams/link_aggregation_group/{ref}\",\n\t\t\"/api/objects/itfparams/link_aggregation_group/{ref}/usedby\",\n\t\t\"/api/objects/itfparams/primary/\",\n\t\t\"/api/objects/itfparams/primary/{ref}\",\n\t\t\"/api/objects/itfparams/primary/{ref}/usedby\",\n\t\t\"/api/objects/itfparams/secondary/\",\n\t\t\"/api/objects/itfparams/secondary/{ref}\",\n\t\t\"/api/objects/itfparams/secondary/{ref}/usedby\",\n\t}\n}", "func (o CustomHttpPatternOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v CustomHttpPattern) *string { return v.Path }).(pulumi.StringPtrOutput)\n}", "func (s *RestServer) AddInterfaceAPIRoutes(r *mux.Router) {\n\n\tr.Methods(\"GET\").Subrouter().HandleFunc(\"/\", httputils.MakeHTTPHandler(s.listInterfaceHandler))\n\n\tr.Methods(\"GET\").Subrouter().HandleFunc(\"/{ObjectMeta.Tenant}/{ObjectMeta.Namespace}/{ObjectMeta.Name}\", httputils.MakeHTTPHandler(s.getInterfaceHandler))\n\n\tr.Methods(\"POST\").Subrouter().HandleFunc(\"/\", httputils.MakeHTTPHandler(s.postInterfaceHandler))\n\n\tr.Methods(\"DELETE\").Subrouter().HandleFunc(\"/{ObjectMeta.Tenant}/{ObjectMeta.Namespace}/{ObjectMeta.Name}\", httputils.MakeHTTPHandler(s.deleteInterfaceHandler))\n\n\tr.Methods(\"PUT\").Subrouter().HandleFunc(\"/{ObjectMeta.Tenant}/{ObjectMeta.Namespace}/{ObjectMeta.Name}\", httputils.MakeHTTPHandler(s.putInterfaceHandler))\n\n}", "func isAPIPath(ctx *macaron.Context) bool {\n\treturn strings.HasPrefix(ctx.Req.URL.Path, \"/api/\")\n}", "func (r *TestSchema) URLPatterns() []Route {\n\treturn []Route{\n\t\t{Method: http.MethodGet, Path: \"/\", ResourceFunc: r.Get,\n\t\t\tReturns: []*Returns{{Code: 200}}},\n\n\t\t{Method: http.MethodPost, Path: \"/sayhello/{userid}\", ResourceFunc: r.Post,\n\t\t\tReturns: []*Returns{{Code: 200}}},\n\n\t\t{Method: http.MethodDelete, Path: \"/sayhi\", ResourceFunc: r.Delete,\n\t\t\tReturns: []*Returns{{Code: 200}}},\n\n\t\t{Method: http.MethodHead, Path: \"/sayjson\", ResourceFunc: r.Head,\n\t\t\tReturns: []*Returns{{Code: 200}}},\n\t\t{Method: http.MethodPatch, Path: \"/sayjson\", ResourceFunc: r.Patch,\n\t\t\tReturns: []*Returns{{Code: 200}}},\n\t\t{Method: http.MethodPut, Path: \"/hi\", ResourceFunc: r.Put,\n\t\t\tReturns: []*Returns{{Code: 200}}},\n\t}\n}", "func (o ProviderOutput) ApiBasePath() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *Provider) pulumi.StringPtrOutput { return v.ApiBasePath }).(pulumi.StringPtrOutput)\n}", "func (*HttpGroups) GetPath() string { return \"/api/objects/http/group/\" }", "func (o MethodOutput) RestApi() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Method) pulumi.StringOutput { return v.RestApi }).(pulumi.StringOutput)\n}", "func GETapi(w http.ResponseWriter, request *http.Request) {\n\tw.Header().Set(\"content-type\", \"application/json\")\n\n\tURLs := mux.Vars(request)\n\tif len(URLs) != 0 {\n\t\thttp.Error(w, \"400 - Bad Request!\", http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tmetaInfo := &MetaInfo{}\n\tmetaInfo.Uptime = FormatSince(startTime)\n\tmetaInfo.Info = \"Service for IGC tracks\"\n\tmetaInfo.Version = \"version 1.0\"\n\n\tjson.NewEncoder(w).Encode(metaInfo)\n}", "func ApiHandler(rw http.ResponseWriter, req *http.Request) {\n\t_, err := rw.Write([]byte(serviceApiVersion))\n\tif err != nil {\n\t\tlog.Error(\"ERROR: Variable <g_api> is not defined properly\")\n\t}\n\trw.WriteHeader(http.StatusOK)\n}", "func (RemoteSyslog) ApiRoutes() []string {\n\treturn []string{\n\t\t\"/api/objects/remote_syslog/group/\",\n\t\t\"/api/objects/remote_syslog/group/{ref}\",\n\t\t\"/api/objects/remote_syslog/group/{ref}/usedby\",\n\t\t\"/api/objects/remote_syslog/server/\",\n\t\t\"/api/objects/remote_syslog/server/{ref}\",\n\t\t\"/api/objects/remote_syslog/server/{ref}/usedby\",\n\t}\n}", "func API(r *gin.Engine) {\n\tapi := r.Group(\"/api\")\n\t{\n\n\t\tapi.GET(\"/\", func(c *gin.Context) {\n\t\t\tc.JSON(http.StatusOK, gin.H{\"message\": \"/api\", \"status\": http.StatusOK})\n\t\t})\n\n\t\tAuthRoutes(api)\n\n\t\t// api.Use(jwt.JWT().MiddlewareFunc())\n\n\t\tUsersRoutes(api)\n\t\tPostsRoutes(api)\n\t}\n}", "func InitApi(app *iris.Application) {\n\t// var getAPI router.Party\n\n\tif config.Conf.Debug {\n\t\tppApi := app.Party(\"/debug\")\n\t\tppApi.Get(\"/pprof\", pprofHandler(pprof.Index))\n\t\tppApi.Get(\"/cmdline\", pprofHandler(pprof.Cmdline))\n\t\tppApi.Get(\"/profile\", pprofHandler(pprof.Profile))\n\t\tppApi.Post(\"/symbol\", pprofHandler(pprof.Symbol))\n\t\tppApi.Get(\"/symbol\", pprofHandler(pprof.Symbol))\n\t\tppApi.Get(\"/trace\", pprofHandler(pprof.Trace))\n\t\tppApi.Get(\"/block\", pprofHandler(pprof.Handler(\"block\").ServeHTTP))\n\t\tppApi.Get(\"/goroutine\", pprofHandler(pprof.Handler(\"goroutine\").ServeHTTP))\n\t\tppApi.Get(\"/allocs\", pprofHandler(pprof.Handler(\"allocs\").ServeHTTP))\n\t\tppApi.Get(\"/heap\", pprofHandler(pprof.Handler(\"heap\").ServeHTTP))\n\t\tppApi.Get(\"/mutex\", pprofHandler(pprof.Handler(\"mutex\").ServeHTTP))\n\t\tppApi.Get(\"/threadcreate\", pprofHandler(pprof.Handler(\"threadcreate\").ServeHTTP))\n\n\t\tgetAPI := app.Party(\"/get\")\n\t\tgetAPI.Get(\"/crontab\", ShowCrontab)\n\t}\n}", "func GetAPILastRoute(router *gin.RouterGroup) {\n\trouter.GET(\"\", GetAPILast)\n}", "func NewAPIWithVersion(config *conf.Configuration, db *gorm.DB, paypal *paypalsdk.Client, mailer *mailer.Mailer, assets assetstores.Store, version string) *API {\n\tapi := &API{\n\t\tlog: logrus.WithField(\"component\", \"api\"),\n\t\tconfig: config,\n\t\tdb: db,\n\t\tpaypal: paypal,\n\t\tmailer: mailer,\n\t\thttpClient: &http.Client{},\n\t\tassets: assets,\n\t\tversion: version,\n\t}\n\n\tmux := kami.New()\n\tmux.Use(\"/\", api.populateContext)\n\tmux.Use(\"/\", api.withToken)\n\tmux.LogHandler = api.logCompleted\n\n\t// endpoints\n\tmux.Get(\"/\", api.Index)\n\n\tmux.Get(\"/orders\", api.OrderList)\n\tmux.Post(\"/orders\", api.OrderCreate)\n\tmux.Get(\"/orders/:id\", api.OrderView)\n\tmux.Put(\"/orders/:id\", api.OrderUpdate)\n\tmux.Get(\"/orders/:order_id/payments\", api.PaymentListForOrder)\n\tmux.Post(\"/orders/:order_id/payments\", api.PaymentCreate)\n\tmux.Post(\"/orders/:order_id/receipt\", api.ResendOrderReceipt)\n\n\tmux.Get(\"/users\", api.UserList)\n\tmux.Get(\"/users/:user_id\", api.UserView)\n\tmux.Get(\"/users/:user_id/payments\", api.PaymentListForUser)\n\tmux.Delete(\"/users/:user_id\", api.UserDelete)\n\tmux.Get(\"/users/:user_id/addresses\", api.AddressList)\n\tmux.Get(\"/users/:user_id/addresses/:addr_id\", api.AddressView)\n\tmux.Delete(\"/users/:user_id/addresses/:addr_id\", api.AddressDelete)\n\tmux.Get(\"/users/:user_id/orders\", api.OrderList)\n\n\tmux.Get(\"/downloads/:id\", api.DownloadURL)\n\tmux.Get(\"/downloads\", api.DownloadList)\n\tmux.Get(\"/orders/:order_id/downloads\", api.DownloadList)\n\n\tmux.Get(\"/vatnumbers/:number\", api.VatnumberLookup)\n\n\tmux.Get(\"/payments\", api.PaymentList)\n\tmux.Get(\"/payments/:pay_id\", api.PaymentView)\n\tmux.Post(\"/payments/:pay_id/refund\", api.PaymentRefund)\n\n\tmux.Post(\"/paypal\", api.PaypalCreatePayment)\n\tmux.Get(\"/paypal/:payment_id\", api.PaypalGetPayment)\n\n\tmux.Get(\"/reports/sales\", api.SalesReport)\n\tmux.Get(\"/reports/products\", api.ProductsReport)\n\n\tmux.Get(\"/coupons/:code\", api.CouponView)\n\n\tmux.Post(\"/claim\", api.ClaimOrders)\n\n\tcorsHandler := cors.New(cors.Options{\n\t\tAllowedMethods: []string{\"GET\", \"POST\", \"PATCH\", \"PUT\", \"DELETE\"},\n\t\tAllowedHeaders: []string{\"Accept\", \"Authorization\", \"Content-Type\"},\n\t\tExposedHeaders: []string{\"Link\", \"X-Total-Count\"},\n\t\tAllowCredentials: true,\n\t})\n\n\tapi.handler = corsHandler.Handler(mux)\n\n\treturn api\n}", "func (module *SdkLibrary) apiDistPath(apiScope *apiScope) string {\n\treturn path.Join(\"apistubs\", module.distGroup(), apiScope.name)\n}", "func (this *RouteServiceProvider) mapApiRoutes() {\n\tFacades.Route().Group(map[string]string{\"prefix\": \"api\"}, Routes.Api)\n}", "func (a *Router) RegisterAPI(app *gin.Engine) {\n\tg := app.Group(\"/api\")\n\n\tg.Use(middleware.UserAuthMiddleware(a.Auth))\n\tg.Use(middleware.RateLimiterMiddleware())\n\n\tv1 := g.Group(\"/v1\")\n\t{\n\t\tgDemo := v1.Group(\"demos\")\n\t\t{\n\t\t\tgDemo.GET(\"\", a.DemoAPI.Query)\n\t\t\tgDemo.GET(\":id\", a.DemoAPI.Get)\n\t\t\tgDemo.POST(\"\", a.DemoAPI.Create)\n\t\t\tgDemo.PUT(\":id\", a.DemoAPI.Update)\n\t\t\tgDemo.DELETE(\":id\", a.DemoAPI.Delete)\n\t\t\tgDemo.PATCH(\":id/enable\", a.DemoAPI.Enable)\n\t\t\tgDemo.PATCH(\":id/disable\", a.DemoAPI.Disable)\n\t\t}\n\n\t} // v1 end\n}", "func (o CustomHttpPatternResponseOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v CustomHttpPatternResponse) string { return v.Path }).(pulumi.StringOutput)\n}", "func API(w http.ResponseWriter, r *http.Request, cfg cli.Config) {\n\tvar f = mux.Vars(r)[\"name\"]\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\n\tswitch f {\n\tcase \"ping\":\n\t\tping(w, r, &cfg)\n\tcase \"init.trace\":\n\t\tinitTrace(w, r, &cfg)\n\tcase \"get.trace\":\n\t\tgetTrace(w, r)\n\tcase \"close.trace\":\n\t\tcloseTrace(w, r)\n\tcase \"geo\":\n\t\tgetGeo(w, r)\n\t}\n}", "func App() http.Handler {\n\n\t// Create a new mux which will process the\n\t// initial requests.\n\tr := http.NewServeMux()\n\n\t// This main handler identifies which version of\n\t// the api to use and uses it.\n\tr.HandleFunc(\"/api/\", func(res http.ResponseWriter, req *http.Request) {\n\n\t\t// Look for a specified version.\n\t\tv := req.Header.Get(\"x-version\")\n\n\t\t// Retrieve that version and validate it exists. If\n\t\t// not use the default version.\n\t\th := apis[v]\n\t\tif h == nil {\n\t\t\th = apis[defaultAPIVersion]\n\t\t}\n\n\t\t// Strip the duplication of the path and process\n\t\t// the route against the api version.\n\t\thttp.StripPrefix(\"/api\", h).ServeHTTP(res, req)\n\t})\n\n\treturn r\n}", "func (*InterfaceBridges) GetPath() string { return \"/api/objects/interface/bridge/\" }", "func renderAPI(templateDir string, api, need map[name]interface{}) ([]byte, error) {\n\ttmpl := template.New(\"\")\n\ttmpl.Funcs(template.FuncMap{\n\t\t\"API\": func() map[name]interface{} { return api },\n\t\t\"typeOf\": func(v interface{}) string {\n\t\t\tif v == nil {\n\t\t\t\treturn \"nil\"\n\t\t\t}\n\t\t\treturn strings.ToLower(reflect.TypeOf(v).Name())\n\t\t},\n\t\t\"unexport\": func(s string) string {\n\t\t\tws := camelcase.Split(s)\n\t\t\tif upperWords[strings.ToLower(ws[0])] {\n\t\t\t\tws[0] = strings.ToLower(ws[0])\n\t\t\t} else {\n\t\t\t\tws[0] = strings.ToLower(ws[0][:1]) + ws[0][1:]\n\t\t\t}\n\t\t\tret := strings.Join(ws, \"\")\n\t\t\tif ret == \"type\" {\n\t\t\t\treturn \"typ\"\n\t\t\t}\n\t\t\treturn ret\n\t\t},\n\t\t\"render\": func(v interface{}) (string, error) {\n\t\t\tn := strings.ToLower(reflect.TypeOf(v).Name())\n\t\t\tvar b bytes.Buffer\n\t\t\terr := tmpl.ExecuteTemplate(&b, n, v)\n\t\t\treturn b.String(), err\n\t\t},\n\t\t\"abort\": func(s string) (string, error) { return \"\", errors.New(s) },\n\t\t\"last\": func(fs fields, i int) bool { return i == len(fs)-1 },\n\t})\n\ttmpl, err := tmpl.ParseGlob(filepath.Join(templateDir, \"*\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar out bytes.Buffer\n\tif err := tmpl.ExecuteTemplate(&out, \"main\", need); err != nil {\n\t\treturn nil, err\n\t}\n\treturn out.Bytes(), nil\n}", "func (*HttpDomainRegexs) GetPath() string { return \"/api/objects/http/domain_regex/\" }", "func (r *RestFulHello) URLPatterns() []rf.Route {\n\treturn []rf.Route{\n\t\t{Method: http.MethodGet, Path: \"/sayhello/{userid}\", ResourceFuncName: \"Sayhello\"},\n\t\t{Method: http.MethodPost, Path: \"/sayhi\", ResourceFuncName: \"Sayhi\"},\n\t}\n}", "func ToAPI(api *Api) *v1.Api {\n\tcrd := &v1.Api{}\n\tcrd.TypeMeta.Kind = \"Api\"\n\tcrd.TypeMeta.APIVersion = v1.GroupVersion.Group + \"/\" + v1.GroupVersion.Version\n\n\tcrd.ObjectMeta.Name = api.ID\n\tcrd.ObjectMeta.Namespace = api.Namespace\n\tcrd.ObjectMeta.Labels = make(map[string]string)\n\tcrd.ObjectMeta.Labels[v1.ServiceunitLabel] = api.Serviceunit.ID\n\tcrd.Spec = v1.ApiSpec{\n\t\tName: api.Name,\n\t\tDescription: api.Description,\n\t\tServiceunit: api.Serviceunit,\n\t\tApplications: api.Applications,\n\t\tFrequency: api.Frequency,\n\t\tApiType: api.ApiType,\n\t\tAuthType: api.AuthType,\n\t\tTags: api.Tags,\n\t\tApiBackendType: api.Serviceunit.Type,\n\t\tMethod: api.Method,\n\t\tProtocol: api.Protocol,\n\t\tReturnType: api.ReturnType,\n\t\tRDBQuery: api.RDBQuery,\n\t\tDataWarehouseQuery: api.DataWarehouseQuery,\n\t\tApiDefineInfo: api.ApiDefineInfo,\n\t\tKongApi: api.KongApi,\n\t\tApiQueryInfo: api.ApiQueryInfo,\n\t\tApiReturnInfo: api.ApiReturnInfo,\n\t\tTraffic: api.Traffic,\n\t\tRestriction: api.Restriction,\n\t\tPublishInfo: api.PublishInfo,\n\t\tResponseTransformer: api.ResTransformer,\n\t}\n\n\tcrd.Status = v1.ApiStatus{\n\t\tStatus: v1.Init,\n\t\tAction: v1.Create,\n\t\t//create api update status to unreleased\n\t\tPublishStatus: v1.UnRelease,\n\t\tAccessLink: api.AccessLink,\n\t\tUpdatedAt: metav1.Now(),\n\t\tReleasedAt: metav1.Now(),\n\t\tApplicationCount: api.ApplicationCount,\n\t\tCalledCount: api.CalledCount,\n\t\tFailedCount: api.FailedCount,\n\t\tLatencyCount: api.LatencyCount,\n\t\tCallFrequency: api.CallFrequency,\n\t}\n\t// add user labels\n\tcrd.ObjectMeta.Labels = user.AddUsersLabels(api.Users, crd.ObjectMeta.Labels)\n\treturn crd\n}", "func apiVersion(c *gin.Context) {\n\tc.JSON(200, gin.H{\"Version\": aptly.Version})\n}", "func Path(output string) string {\n\turl := \"https://api.ipify.org\"\n\tquery := \"/?format=json\"\n\n\tif output == strings.ToLower(\"json\") {\n\t\turl += query\n\t} else if output != \"text\" {\n\t\treturn \"\"\n\t}\n\n\treturn url\n}", "func (*ItfparamsGroups) GetPath() string { return \"/api/objects/itfparams/group/\" }", "func configMapAPIPath(cm *apiv1.ConfigMap) string {\n\treturn fmt.Sprintf(\"/api/v1/namespaces/%s/configmaps/%s\", cm.Namespace, cm.Name)\n}", "func (*InterfacePppoas) GetPath() string { return \"/api/objects/interface/pppoa/\" }", "func (u URI) GetEndpointPath(path string, params []string) string {\n\tresult := path\n\tif u == ColonRouterPatternBuilder {\n\t\tfor p := range params {\n\t\t\tparts := strings.Split(result, \"?\")\n\t\t\tparts[0] = strings.ReplaceAll(parts[0], \"{\"+params[p]+\"}\", \":\"+params[p])\n\t\t\tresult = strings.Join(parts, \"?\")\n\t\t}\n\t}\n\treturn result\n}", "func (o RequestValidatorOutput) RestApi() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *RequestValidator) pulumi.StringOutput { return v.RestApi }).(pulumi.StringOutput)\n}", "func RegisterAPI(r *mux.Router) {\n\tsr := r.PathPrefix(\"/api\").Subrouter().StrictSlash(true)\n\n\t// Handle invalid paths\n\tsr.NotFoundHandler = setupAPICall(http.HandlerFunc(func(w http.ResponseWriter, req *http.Request) {\n\t\tw.WriteHeader(404)\n\t\tw.Write([]byte(`{\"message\": \"Not Found\"}`))\n\t}))\n\n\t// Root path\n\tsr.Handle(\"\", setupAPICall(http.HandlerFunc(apiHandler))).Methods(\"GET\")\n\tsr.Handle(\"\", setupAPICall(http.HandlerFunc(notAllowed))).Methods(\"POST\", \"PUT\", \"PATCH\", \"DELETE\")\n\n\t// Host\n\tsr.Handle(\"/host\", setupAPICall(http.HandlerFunc(hostHandler))).Methods(\"GET\")\n\tsr.Handle(\"/host\", setupAPICall(http.HandlerFunc(notAllowed))).Methods(\"POST\", \"PUT\", \"PATCH\", \"DELETE\")\n\n\t// Settings\n\tsr.Handle(\"/settings\", setupAPICall(http.HandlerFunc(apiSettingsHandler))).Methods(\"GET\")\n\tsr.Handle(\"/settings\", setupAPICall(http.HandlerFunc(notAllowed))).Methods(\"POST\", \"PUT\", \"PATCH\", \"DELETE\")\n\n\t// Settings\n\tsr.Handle(\"/metrics\", setupAPICall(http.HandlerFunc(metricsHandler))).Methods(\"GET\")\n\tsr.Handle(\"/metrics\", setupAPICall(http.HandlerFunc(notAllowed))).Methods(\"POST\", \"PUT\", \"PATCH\", \"DELETE\")\n\n}", "func (a* ApiEngine) GenerateAPIRoutesForService() {\n\tenv.Output.WriteChDebug(\"(ApiEngine::GenerateAPIRoutesForService)\")\n\ta.AddRoute(ui.GenerateRoute(\"allservices\",\"GET\",\"/api/services\",apiWriter(GetAllServices)))\n\ta.AddRoute(ui.GenerateRoute(\"service\",\"GET\",\"/api/services/{service}\",apiWriter(GetService)))\n}", "func isAPI(path string) bool {\n\treturn strings.HasPrefix(path, \"/api\")\n}", "func (r Repo) ImportPathPattern() string {\n\treturn r.Root + \"/...\"\n}", "func addAPIListingsRoute(r *mux.Router, srv *RestServer) {\n\tr.Methods(\"GET\").Subrouter().HandleFunc(\"/\", MakeHTTPHandler(srv.listInterfaceHandler))\n}", "func (o AuthBackendCustomEndpointOutput) Api() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v AuthBackendCustomEndpoint) *string { return v.Api }).(pulumi.StringPtrOutput)\n}", "func API(g *gin.Engine) {\n\tcatController := controllers.NewCat()\n\tgoodsController := controllers.NewGoods()\n\tgoodsGroupController := controllers.NewGoodsGroup()\n\tgoodsBrandController := controllers.NewGoodsBrand()\n\tgoodsCategoryController := controllers.NewGoodsCategory()\n\tgoodsSkuController := controllers.NewGoodsSku()\n\n\tapi := g.Group(\"/api\")\n\n\tv1 := api.Group(\"/v1\")\n\t{\n\t\tv1.GET(\"/\", catController.Home)\n\n\t\tv1.GET(\"/user/:name/:action\", func(c *gin.Context) {\n\t\t\tname := c.Param(\"name\")\n\t\t\taction := c.Param(\"action\")\n\t\t\tmessage := name + \" is \" + action\n\t\t\tc.String(http.StatusOK, message)\n\t\t})\n\t}\n\n\tgoods := v1.Group(\"/goods\")\n\t{\n\t\tgoods.GET(\"\", goodsController.Home)\n\t}\n\n\tgoodsGroup := v1.Group(\"/goods/group\")\n\t{\n\t\tgoodsGroup.GET(\"\", goodsGroupController.Home)\n\t}\n\n\tgoodsBrand := v1.Group(\"/goods/brand\")\n\t{\n\t\tgoodsBrand.GET(\"\", goodsBrandController.Home)\n\t}\n\n\tgoodsCategory := v1.Group(\"/goods/category\")\n\t{\n\t\tgoodsCategory.GET(\"\", goodsCategoryController.Home)\n\t}\n\n\tgoodsSku := v1.Group(\"/goods/sku\")\n\t{\n\t\tgoodsSku.GET(\"\", goodsSkuController.Home)\n\t}\n\n}", "func (domain Domain) API() string {\n\ttruncateIndex := strings.LastIndexAny(domain.Host, \".\")\n\t// It is already an API without tld.\n\tif truncateIndex == -1 {\n\t\treturn domain.Host\n\t}\n\treturn domain.Host[:truncateIndex]\n}", "func InitAPI() {\n\tBaseRoutes = &Routes{}\n\tBaseRoutes.Root = mux.NewRouter()\n\tBaseRoutes.Root.Handle(\"/\", http.HandlerFunc(indexHandler))\n\tBaseRoutes.Recipe = BaseRoutes.Root.PathPrefix(\"/recipe\").Subrouter()\n\tBaseRoutes.NeedRecipe = BaseRoutes.Recipe.PathPrefix(\"/{recipe-id:[0-9]+}\").Subrouter()\n\tBaseRoutes.Recipes = BaseRoutes.Root.PathPrefix(\"/recipes\").Subrouter()\n\tInitRecipe()\n}", "func (Spx) ApiRoutes() []string {\n\treturn []string{\n\t\t\"/api/objects/spx/group/\",\n\t\t\"/api/objects/spx/group/{ref}\",\n\t\t\"/api/objects/spx/group/{ref}/usedby\",\n\t\t\"/api/objects/spx/template/\",\n\t\t\"/api/objects/spx/template/{ref}\",\n\t\t\"/api/objects/spx/template/{ref}/usedby\",\n\t}\n}", "func (o BuildSpecStrategyOutput) ApiVersion() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BuildSpecStrategy) *string { return v.ApiVersion }).(pulumi.StringPtrOutput)\n}", "func (h *HttpGroup) GetPath() string { return fmt.Sprintf(\"/api/objects/http/group/%s\", h.Reference) }", "func pathPattern(pattern *regexp.Regexp) *regexp.Regexp {\n\treturn suffixPattern(regexp.MustCompile(\"(^|/)\" + pattern.String()))\n}", "func apiHttpHandler(c echo.Context) error {\n\turiPattern := c.Path()\n\tif _, ok := httpRoutingMap[uriPattern]; !ok {\n\t\treturn c.JSON(http.StatusOK, itineris.ResultNotImplemented.ToMap())\n\t}\n\thttpMethod := strings.ToUpper(c.Request().Method)\n\tif _, ok := httpRoutingMap[uriPattern][httpMethod]; !ok {\n\t\treturn c.JSON(http.StatusOK, itineris.ResultNotImplemented.ToMap())\n\t}\n\n\tapiName := httpRoutingMap[uriPattern][httpMethod]\n\tctx, auth, params := _parseRequest(apiName, c)\n\n\tapiResult := ApiRouter.CallApi(ctx, auth, params)\n\treturn c.JSON(http.StatusOK, apiResult.ToMap())\n}", "func createRoute(method string, pattern string, handler http.Handler) (*route, error) {\n\n if pattern[len(pattern)-1] != '/' {\n pattern += \"/\"\n }\n\n err := validatePattern(pattern)\n if err != nil {\n return nil, err\n }\n\n regexpPattern := strings.ToLower(pattern)\n regexpPattern = slashEscaper.ReplaceAllString(regexpPattern, \"\\\\/\")\n regexpPattern = varReplacer.ReplaceAllString(regexpPattern, \"\\\\/[\\\\w\\\\d-]+\\\\/\")\n regexpPattern = wildcardReplacer.ReplaceAllString(regexpPattern, \"\\\\/[\\\\w\\\\d-\\\\/\\\\.]+\\\\/\")\n regexpPattern = \"^\" + regexpPattern + \"?$\"\n\n r := &route{ method: method,\n pattern: pattern,\n handler: handler,\n patternMatcher: regexp.MustCompile(regexpPattern) }\n\n return r, nil\n}", "func (o PodSchedulingOutput) ApiVersion() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *PodScheduling) pulumi.StringOutput { return v.ApiVersion }).(pulumi.StringOutput)\n}", "func Api_version() string {\n\treturn api_version\n}", "func (o ParamKindPatchOutput) ApiVersion() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ParamKindPatch) *string { return v.ApiVersion }).(pulumi.StringPtrOutput)\n}", "func apihandler(w http.ResponseWriter, r *http.Request) {\n\tcatalogMatch := catalogRequestRegex.FindStringSubmatch(r.RequestURI)\n\tcoreMatch := coreRequestRegex.FindStringSubmatch(r.RequestURI)\n\n\tif len(catalogMatch) == 0 && len(coreMatch) == 0 {\n\t\tw.WriteHeader(500)\n\t\tw.Write([]byte(fmt.Sprintf(\"unexpected request %s %s doesn't match %q or %q\", r.Method, r.RequestURI, catalogRequestRegex, coreRequestRegex)))\n\t\treturn\n\t}\n\n\tif r.Method != http.MethodGet {\n\t\t// Anything more interesting than a GET, i.e. it relies upon server behavior\n\t\t// probably should be an integration test instead\n\t\tw.WriteHeader(500)\n\t\tw.Write([]byte(fmt.Sprintf(\"unallowed method for request %s %s\", r.Method, r.RequestURI)))\n\t\treturn\n\t}\n\n\tvar match string\n\tif len(catalogMatch) > 0 {\n\t\tmatch = filepath.Join(\"catalog\", catalogMatch[1])\n\t} else {\n\t\tmatch = filepath.Join(\"core\", coreMatch[1])\n\t}\n\n\trelpath, err := url.PathUnescape(match)\n\tif err != nil {\n\t\tw.WriteHeader(500)\n\t\tw.Write([]byte(fmt.Sprintf(\"could not unescape path %s (%s)\", match, err)))\n\t\treturn\n\t}\n\tresponseFile := filepath.Join(\"responses\", relpath+\".json\")\n\t_, response, err := test.GetTestdata(responseFile)\n\tif err != nil {\n\t\tw.WriteHeader(500)\n\t\tw.Write([]byte(fmt.Sprintf(\"request %s has no matching testdata at %s (%s)\", r.RequestURI, responseFile, err)))\n\t\treturn\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.Write(response)\n}", "func generateSelfLink(apiVersion, namespace, kind, name string) string {\n\tvar b strings.Builder\n\n\t// for any v1 api served objects, they used to be served from /api\n\t// all others are served from /apis\n\tif apiVersion == \"v1\" {\n\t\tb.WriteString(\"/api\")\n\t} else {\n\t\tb.WriteString(\"/apis\")\n\t}\n\n\tif len(apiVersion) != 0 {\n\t\tfmt.Fprintf(&b, \"/%s\", apiVersion)\n\t}\n\tif len(namespace) != 0 {\n\t\tfmt.Fprintf(&b, \"/namespaces/%s\", namespace)\n\t}\n\tif len(kind) != 0 {\n\t\tvar suffix string\n\t\tif strings.HasSuffix(kind, \"s\") {\n\t\t\tsuffix = \"es\"\n\t\t} else {\n\t\t\tsuffix = \"s\"\n\t\t}\n\t\tfmt.Fprintf(&b, \"/%s%s\", strings.ToLower(kind), suffix)\n\t}\n\tif len(name) != 0 {\n\t\tfmt.Fprintf(&b, \"/%s\", name)\n\t}\n\treturn b.String()\n}", "func generateUserAPIRoutes(router *mux.Router) {\n\tusers := router.PathPrefix(\"/users\").Subrouter()\n\tusers.Use(helpers.LoggingMiddleware)\n\tusers.HandleFunc(\"\", user.GetAll).Methods(\"GET\") // GET Request to handle all data present in the Database\n\n\tsub := router.PathPrefix(\"/user\").Subrouter()\n\tsub.Use(helpers.LoggingMiddleware)\n\t\n\tsub.HandleFunc(\"\", user.GetUser).Methods(\"GET\")\n}", "func API(middlewares ...func(http.Handler) http.Handler) *chi.Mux {\n\trouter := chi.NewRouter()\n\tfor _, mid := range middlewares {\n\t\trouter.Use(mid)\n\t}\n\trouter.Use(render.SetContentType(render.ContentTypeJSON))\n\n\trouter.Route(rest.PathTo_Inventory, func(router chi.Router) {\n\t\trouter.Use(rest.NewFilm)\n\t\trouter.With(rest.FilmRequest).Get(\"/\", rest.GetFilm)\n\t\trouter.With(rest.FilmRequest).Post(\"/\", rest.CreateFilm)\n\t\trouter.Route(\"/{FilmPagination}\", func(router chi.Router) {\n\t\t\trouter.Use(rest.Pagination, rest.FilmRequest)\n\t\t\trouter.Get(\"/\", rest.GetFilm)\n\t\t})\n\n\t\trouter.Route(\"/search/{filter}\", func(router chi.Router) {\n\t\t\trouter.With(rest.FilmRequest).Get(\"/\", rest.GetFilm)\n\t\t\trouter.Group(func(router chi.Router) {\n\t\t\t\trouter.Use(rest.Pagination, rest.FilmRequest)\n\t\t\t\trouter.Get(\"/{FilmPagination}\", rest.GetFilm)\n\t\t\t})\n\t\t})\n\n\t\t// github.com/google/re2/wiki/Syntax\n\t\trouter.Route(\"/film/{id:\\\\d+}/{title:[[:alnum:]\\\\p{Latin}%_]+}\", func(router chi.Router) {\n\t\t\trouter.Use(rest.FilmRequest)\n\t\t\trouter.Get(\"/\", rest.GetFilm)\n\t\t\trouter.Put(\"/\", rest.UpdateFilm)\n\t\t\trouter.Delete(\"/\", rest.DeleteFilm)\n\t\t})\n\t})\n\n\trouter.Route(rest.PathTo_Categories, func(router chi.Router) {\n\t\trouter.Use(rest.NewCategory)\n\t\trouter.Use(rest.CategoryRequest)\n\t\trouter.Get(\"/\", rest.GetCategories)\n\t})\n\n\trouter.Route(rest.PathTo_Actors, func(router chi.Router) {\n\t\trouter.Use(rest.NewActor)\n\t\trouter.Use(rest.ActorRequest)\n\t\trouter.Get(\"/\", rest.GetActors)\n\t})\n\n\treturn router\n}", "func (o EndpointsListOutput) ApiVersion() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *EndpointsList) pulumi.StringOutput { return v.ApiVersion }).(pulumi.StringOutput)\n}", "func FormatEndpointMethod(n string) string {\n\tp := strings.Split(n, \"/\")\n\tm := \"\"\n\tfor _, s := range p {\n\t\tm = fmt.Sprintf(\"%s%s\", m, strings.Title(s))\n\t}\n\treturn m\n}", "func makePath(urlPattern string) (regex *regexp.Regexp, pathParams []string, err error) {\n\tpathParams = make([]string, 0)\n\t//patternPeices = make([]string)\n\tparts := strings.Split(urlPattern, \"/\")\n\tfor i, p := range parts {\n\t\tl := len(p)\n\t\tif l > 0 && p[0] == ':' {\n\t\t\tif l < 2 {\n\t\t\t\t//unnamed param\n\t\t\t\terr = NewRouterError(\"Cannot have unnamed path params in route: \", urlPattern)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tpathParams = append(pathParams, p[1:l])\n\t\t\tparts[i] = \"([A-Za-z0-9]+)\" //just support simple strings not url encoded or anything for now\n\t\t}\n\t}\n\n\tregexString := \"^\" + strings.Join(parts, \"/\") + \"$\"\n\tregex, err = regexp.Compile(regexString)\n\n\treturn\n}", "func (a *DocRouter) RegisterAPI(app *gin.Engine) {\n\tapp.GET(\"/swagger/*any\", ginSwagger.WrapHandler(swaggerFiles.Handler))\n}", "func (s *ServerConfig) VersioningEndpoint(path string) string {\n\treturn \"/api/\" + s.ApiVer + \"/\" + path\n}", "func RegisterAPIRoutes(g *echo.Group, bind context.Binder) {\n\t// Store endpoints first since they are the most active\n\t//e.GET(\"/api/store/\", storeGetView)\n\t//e.POST(\"/api/store/\", storePostView)\n\n\t// TODO Can not register same handler for two different routes\n\t//g = g.Group(\"/store\")\n\t//g.GET(\"/\", storeGetView)\n\t//g.POST(\"/\", storePostView)\n\t// :project_id is [\\w_-]+\n\tg = g.Group(\"/:project_id/store\")\n\tg.GET(\"/\", storeGetView)\n\tg.POST(\"/\", bind.Base(storePostView))\n\t// :project_id is \\d+\n\tg = g.Group(\"/:project_id/csp-report\")\n\t// TODO is CspReportGetView needed?\n\tg.GET(\"/\", cspReportGetView)\n\tg.POST(\"/\", cspReportPostView)\n}", "func (o BlobOutputDataSourceOutput) PathPattern() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BlobOutputDataSource) *string { return v.PathPattern }).(pulumi.StringPtrOutput)\n}", "func GetGroupPath(id string) string {\n\treturn fmt.Sprintf(\"/api/group/%v\", id)\n}", "func (h AppServer) BasePath() string {\n\treturn h.Config.ApiBase\n}", "func (Snmp) ApiRoutes() []string {\n\treturn []string{\n\t\t\"/api/objects/snmp/group/\",\n\t\t\"/api/objects/snmp/group/{ref}\",\n\t\t\"/api/objects/snmp/group/{ref}/usedby\",\n\t\t\"/api/objects/snmp/trap/\",\n\t\t\"/api/objects/snmp/trap/{ref}\",\n\t\t\"/api/objects/snmp/trap/{ref}/usedby\",\n\t}\n}", "func (o BlobReferenceInputDataSourceOutput) PathPattern() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BlobReferenceInputDataSource) *string { return v.PathPattern }).(pulumi.StringPtrOutput)\n}", "func routePath(w http.ResponseWriter, r *http.Request, trimURL string) {\n\n\t/***********************************************/\n\t//TODO: add your custom web API here:\n\t/**********************************************/\n\n\tif strings.HasPrefix(trimURL, \"login\") && webServer.IsPOST(r) { //>>>>authentication\n\t\tauthenticateHandler.HandleHTTPLogin(w, r)\n\t} else if strings.HasPrefix(trimURL, \"logout\") && webServer.IsPOST(r) {\n\t\tauthenticateHandler.HandleHTTPLogout(w, r)\n\t} else if strings.Compare(trimURL, \"current-user\") == 0 && webServer.IsGET(r) {\n\t\tauthenticateHandler.HandleCurrentUser(w, r)\n\t} else if strings.Compare(trimURL, \"role\") == 0 && webServer.IsPOST(r) { //>>>>authorization\n\t\tauthorizeHandler.HandleAddRole(w, r)\n\t} else if strings.Compare(trimURL, \"role\") == 0 && webServer.IsGET(r) {\n\t\tauthorizeHandler.HandleGetRole(w, r)\n\t} else if strings.Compare(trimURL, \"role-access\") == 0 && webServer.IsGET(r) {\n\t\tauthorizeHandler.HandleGetAccessRole(w, r)\n\t} else if strings.Compare(trimURL, \"role-access-count\") == 0 && webServer.IsGET(r) {\n\t\tauthorizeHandler.HandleGetAccessRoleCount(w, r)\n\t} else if strings.Compare(trimURL, \"access\") == 0 && webServer.IsGET(r) {\n\t\tauthorizeHandler.HandleGetAccess(w, r)\n\t} else if strings.HasPrefix(trimURL, \"meals\") { //>>>>sample return JSON\n\t\tw.Header().Set(\"Content-Type\", \"application/json\") //MIME to application/json\n\t\tw.WriteHeader(http.StatusOK) //status code 200, OK\n\t\tw.Write([]byte(\"{ \\\"msg\\\": \\\"this is meal A \\\" }\")) //body text\n\t\treturn\n\t} else if strings.HasPrefix(trimURL, \"img/\") { //>>>>sample return virtual JPG file to client\n\t\tlogicalFilePath := \"./logic-files/\"\n\t\tphysicalFileName := \"neon.jpg\"\n\n\t\t// try read file\n\t\tdata, err := ioutil.ReadFile(logicalFilePath + physicalFileName)\n\t\tif err != nil {\n\t\t\t// show error page if failed to read file\n\t\t\thandleErrorCode(500, \"Unable to retrieve image file\", w)\n\t\t} else {\n\t\t\t//w.Header().Set(\"Content-Type\", \"image/jpg\") // #optional HTTP header info\n\n\t\t\t// uncomment if image file is meant to download instead of display on web browser\n\t\t\t// clientDisplayFileName = \"customName.jpg\"\n\t\t\t//w.header().Set(\"Content-Disposition\", \"attachment; filename=\\\"\" + clientDisplayFileName + \"\\\"\")\n\n\t\t\t// write file (in binary format) direct into HTTP return content\n\t\t\tw.Write(data)\n\t\t}\n\t} else {\n\t\t// show error code 404 not found\n\t\t//(since the requested URL doesn't match any of it)\n\t\thandleErrorCode(404, \"Path not found.\", w)\n\t}\n\n}", "func (c Client) api(endpoint string, params *url.Values) (fullURL *url.URL, err error) {\n\tif endpoint == \"\" {\n\t\treturn nil, errors.New(\"API endpoint has not been provided\")\n\t}\n\n\t// if no additional params has been provided, we have to construct one\n\tif params == nil {\n\t\tparams = &url.Values{}\n\t}\n\t// always add auth token to list of values\n\tparams.Add(\"auth_token\", c.apikey)\n\n\t// adding the API endpoint to path\n\tfullURL, err = url.Parse(c.baseURL.String())\n\tif err != nil {\n\t\treturn\n\t}\n\tfullURL.Path = path.Join(fullURL.Path, endpoint)\n\n\t// building the full query\n\tfullURL.RawQuery = params.Encode()\n\treturn\n}", "func (a *Api) Path(r *http.Request) string {\n\tif a.Prefix != \"\" {\n\t\tif strings.HasSuffix(a.Prefix, \"/\") {\n\t\t\ta.Prefix = strings.TrimSuffix(a.Prefix, \"/\")\n\t\t}\n\n\t\treturn strings.TrimPrefix(r.URL.Path, a.Prefix)\n\t}\n\treturn r.URL.Path\n}", "func (s *fixedService) URLPatterns() []rf.RouteSpec {\n\treturn []rf.RouteSpec {\n\t\t{http.MethodGet, \"/login\", \"LoginEndpoint\"},\n\t\t{http.MethodPost,\"/register\", \"RegisterEndpoint\"},\n\t\t{http.MethodGet, \"/customers\", \"UserGetEndpoint\"},\n\t\t{http.MethodGet, \"/customers/{id}\", \"UserWithIDGetEndpoint\"},\n\t\t{http.MethodGet, \"/customers/{id}/cards\", \"UserWithIDAndCardGetEndpoint\"},\n\t\t{http.MethodGet, \"/customers/{id}/addresses\", \"UserWithIDAndAddressesGetEndpoint\"},\n\t\t{http.MethodGet, \"/cards\", \"CardGetEndpoint\"},\n\t\t{http.MethodGet, \"/cards/{id}\", \"CardWithIDGetEndpoint\"},\n\t\t{http.MethodGet, \"/addresses\", \"AddressGetEndpoint\"},\n\t\t{http.MethodGet, \"/addresses/{id}\", \"AddressWithIDGetEndpoint\"},\n\t\t{http.MethodPost,\"/customers\", \"UserPostEndpoint\"},\n\t\t{http.MethodPost,\"/addresses\", \"AddressPostEndpoint\"},\n\t\t{http.MethodPost,\"/cards\", \"CardPostEndpoint\"},\n\t\t{http.MethodDelete,\"/addresses/{id}\", \"DeleteEndpoint\"},\n\t\t{http.MethodDelete,\"/customers/{id}\", \"DeleteEndpoint\"},\n\t\t{http.MethodDelete,\"/cards/{id}\", \"DeleteEndpoint\"},\n\t\t{http.MethodGet, \"/health\", \"HealthEndpoint\"},\n\n\t}\n}" ]
[ "0.61918443", "0.5917411", "0.5805477", "0.5766767", "0.57138604", "0.5678968", "0.5629477", "0.5625729", "0.55880946", "0.5572507", "0.55685854", "0.5567476", "0.55370563", "0.55348223", "0.5507013", "0.5478041", "0.5453999", "0.54439384", "0.54416347", "0.5429762", "0.54210246", "0.5416878", "0.5364165", "0.5345944", "0.53294724", "0.5307237", "0.5286054", "0.5257021", "0.52037936", "0.5196204", "0.518955", "0.5185704", "0.51849025", "0.5183398", "0.5162742", "0.5153804", "0.5151933", "0.5150341", "0.5148621", "0.5133717", "0.5131405", "0.5130211", "0.5129282", "0.51180726", "0.5111967", "0.5105766", "0.50888395", "0.50877416", "0.50734675", "0.506067", "0.5057459", "0.5055562", "0.5052572", "0.5050302", "0.5029001", "0.50249845", "0.50243574", "0.50190264", "0.50182015", "0.5000331", "0.49973753", "0.4990974", "0.49880812", "0.4980011", "0.4971373", "0.49688533", "0.49612555", "0.49592394", "0.49459085", "0.49428818", "0.49290016", "0.49249095", "0.4924212", "0.4923104", "0.4920616", "0.49166638", "0.49151352", "0.4910714", "0.4901922", "0.4901629", "0.48998496", "0.48907965", "0.4889847", "0.48860255", "0.4873304", "0.48717684", "0.48670292", "0.48645002", "0.48611245", "0.4857048", "0.48569992", "0.48515144", "0.48485005", "0.48483852", "0.48425144", "0.48411503", "0.48403212", "0.48337245", "0.48336518", "0.48290062" ]
0.84291905
0
RegisterAggregator register the aggregator.
func (c *SQLiteConn) RegisterAggregator(name string, impl interface{}, pure bool) error { return errors.New("This feature is not implemented") }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewAggregator() *Aggregator {\n\t// create logger\n\tlogger := common.Logger.With(\"module\", AggregatingService)\n\tLoadedBazooka, err := core.NewPreLoadedBazooka()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\taggregator := &Aggregator{}\n\taggregator.BaseService = *core.NewBaseService(logger, AggregatingService, aggregator)\n\tDB, err := core.NewDB()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\taggregator.DB = DB\n\taggregator.LoadedBazooka = LoadedBazooka\n\treturn aggregator\n}", "func (ag *Aggregator) Register(name string, rep Reporter) {\n\tag.mu.Lock()\n\tdefer ag.mu.Unlock()\n\tif _, ok := ag.reporters[name]; ok {\n\t\tpanic(\"reporter named \" + name + \" is already registered\")\n\t}\n\tag.reporters[name] = rep\n\n}", "func NewAggregator(cfg config.Configuration) *Aggregator {\n\t// create logger\n\tlogger := log.Logger.With(\"module\", AggregatingService)\n\taggregator := &Aggregator{}\n\taggregator.BaseService = *core.NewBaseService(logger, AggregatingService, aggregator)\n\tDB, err := db.NewDB(cfg)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tbz, err := bazooka.NewPreLoadedBazooka(cfg)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\taggregator.Bazooka = bz\n\taggregator.DB = DB\n\n\taggregator.cfg = cfg\n\n\treturn aggregator\n}", "func (ag *Aggregator) Register(name string, rep Reporter) {\n\tag.mu.Lock()\n\tdefer ag.mu.Unlock()\n\tif _, ok := ag.reporters[name]; ok {\n\t\tpanic(\"reporter named \" + name + \" is already registered\")\n\t}\n\tag.reporters[name] = rep\n}", "func NewAggregator() *Aggregator {\n\treturn &Aggregator{\n\t\treporters: make(map[string]Reporter),\n\t}\n}", "func NewAggregator() *Aggregator {\n\treturn &Aggregator{\n\t\treporters: make(map[string]Reporter),\n\t}\n}", "func NewAggregator() *Aggregator {\n\t// TODO: tune the client\n\treturn &Aggregator{\n\t\tAggregatorMap: make(map[string]*GenericMessageSender),\n\t\tMessageClient: messages.NewPubsubClient(&http.Client{}),\n\t}\n}", "func NewAggregator() *Aggregator {\n\treturn &Aggregator{Data: make(map[string]int)}\n}", "func NewAggregator(env *environment.Environment, config *config.Config, resourceID string) *Aggregator {\n\treturn &Aggregator{\n\t\tEnv: env,\n\t\tConfig: config,\n\t\tResourceID: resourceID,\n\t}\n}", "func NewAggregator(address common.Address, backend bind.ContractBackend) (*Aggregator, error) {\n\tcontract, err := bindAggregator(address, backend, backend, backend)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Aggregator{AggregatorCaller: AggregatorCaller{contract: contract}, AggregatorTransactor: AggregatorTransactor{contract: contract}, AggregatorFilterer: AggregatorFilterer{contract: contract}}, nil\n}", "func Register(router *gin.RouterGroup) {\n\trouter.POST(\"/\", ArticleCreate)\n}", "func NewAggregator(config AggregatorConfig) (*Aggregator, error) {\n\tif config.Logger == nil {\n\t\treturn nil, microerror.Maskf(invalidConfigError, \"%T.Logger must not be empty\", config)\n\t}\n\n\ta := &Aggregator{\n\t\tlogger: config.Logger,\n\t}\n\n\treturn a, nil\n}", "func bindAggregator(address common.Address, caller bind.ContractCaller, transactor bind.ContractTransactor, filterer bind.ContractFilterer) (*bind.BoundContract, error) {\n\tparsed, err := abi.JSON(strings.NewReader(AggregatorABI))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn bind.NewBoundContract(address, parsed, caller, transactor, filterer), nil\n}", "func (r *Router) Register(path string, v view.RestViewer, middleware ...gin.HandlerFunc) {\n\n\t// initialize Router\n\tr.prepare()\n\n\t// private middleware use\n\thandlers := append(middleware, http.AsHandlerFunc(v, r.restViewHandlerFunc))\n\n\t// Get formatPath\n\t// Call SetAutoPrefix to change value\n\tpath = r.formatPath(path)\n\n\t// register done\n\tr.group.Any(path, handlers...)\n}", "func (me *Container) Register(r Registries) *Container {\n\tfor _, v := range r {\n\t\tme.bag[v.Key] = v.Value\n\t}\n\n\treturn me\n}", "func RegisterMunger(munger PRMunger) error {\n\tif _, found := mungerMap[munger.Name()]; found {\n\t\treturn fmt.Errorf(\"a munger with that name (%s) already exists\", munger.Name())\n\t}\n\tmungerMap[munger.Name()] = munger\n\tglog.Infof(\"Registered %#v at %s\", munger, munger.Name())\n\treturn nil\n}", "func RegisterMerger(name MergerType, merger NewMerger) {\n\tif _, ok := mergers[name]; ok {\n\t\tpanic(\"merger already register\")\n\t}\n\tmergers[name] = merger\n}", "func NewAggregator(done chan struct{}) (*Aggregator, error) {\n\tagr := Aggregator{\n\t\twait: make(chan struct{}),\n\t\ttickMsgQueue: make(chan *TickMsg, 32),\n\t\tTick: make(chan *AgrTickMsg, 32),\n\t\tregFeed: make(chan FeedInfo),\n\t\tderegFeed: make(chan FeedInfo),\n\t\tForceTick: make(chan ProductType),\n\t\tDone: done,\n\t\tfeeds: make(map[ProductType]map[string]*TickMsg),\n\t\ttotalSources: make(map[ProductType]int),\n\t\tactiveSources: make(map[ProductType]int),\n\t\tretryInterval: 2 * time.Second,\n\t}\n\treturn &agr, nil\n}", "func Register(r *gin.RouterGroup) {\n\tg := r.Group(\"v1\")\n\tinput.Register(g)\n\toutput.Register(g)\n}", "func NewAggregatorTransactor(address common.Address, transactor bind.ContractTransactor) (*AggregatorTransactor, error) {\n\tcontract, err := bindAggregator(address, nil, transactor, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &AggregatorTransactor{contract: contract}, nil\n}", "func (me *GlobalContainer) Register(r Registries) *GlobalContainer {\n\tfor _, v := range r {\n\t\tglobalContainerInstance.Container.bag[v.Key] = v.Value\n\t}\n\n\treturn me\n}", "func (r *router) Register(gin *gin.RouterGroup) {\n\tgin.POST(\"/login\", r.loginHandler)\n}", "func NewAggregator(\n\tin chan *pb.Metric,\n\tout chan *pb.Metric,\n\tsubmissionCadence time.Duration,\n) *aggregator {\n\treturn &aggregator{\n\t\tin: in,\n\t\tout: out,\n\t\tstate: cmap.New(),\n\t\tcadence: submissionCadence,\n\t}\n}", "func NewAggregator(handler *Handler) *Aggregator {\n\treturn &Aggregator{\n\t\thandler: handler,\n\t\tvoteSets: make(map[string]struct {\n\t\t\t*message.StepVotes\n\t\t\tsortedset.Cluster\n\t\t}),\n\t}\n}", "func (a *GaugeMetric) register() {\n\tmetrics := gauges[a.ServiceID]\n\tif metrics == nil {\n\t\tmetrics = make(map[MetricID]*GaugeMetric)\n\t\tgauges[a.ServiceID] = metrics\n\t}\n\tif _, exists := metrics[a.MetricID]; !exists {\n\t\tmetricsRegistry.MustRegister(a.Gauge)\n\t\tmetrics[a.MetricID] = a\n\t}\n}", "func (m *manager) Register(name string, item interface{}, tags map[string]string) error {\n\tif err := dtos.ValidateMetricName(name, \"metric\"); err != nil {\n\t\treturn err\n\t}\n\n\tif len(tags) > 0 {\n\t\tif err := m.setMetricTags(name, tags); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif err := m.registry.Register(name, item); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func Register(m map[string]transformer.Transformer) {\n\tm[operationName] = &Add{}\n}", "func (s *Flattener) Register(pc plugins.SFPluginCache) {\n\tpc.AddHandler(handlerName, NewFlattener)\n\tpc.AddChannel(channelName, NewFlattenerChan)\n}", "func (cb *CommandBus) Register(k string, h cqrs.CommandHandler) {\n\tcb.RegisterFunc(k, h.Handle)\n}", "func (m *autoScalingGroups) Register(asg *Asg) {\n\tm.cacheMutex.Lock()\n\tdefer m.cacheMutex.Unlock()\n\n\tm.registeredAsgs = append(m.registeredAsgs, &asgInformation{\n\t\tconfig: asg,\n\t})\n}", "func (m Metrics) register(m metric) error {\n\t// Register with Prometheus.\n\tprometheus.MustRegister(m.definition())\n}", "func Register(r *gin.RouterGroup) {\n\tg := r.Group(\"input\")\n\tproduction.Register(g)\n\tweather.Register(g)\n}", "func (group *RouterGroup) register(method string, subpattern string, handler HandlerFunc) {\n\tpattern := path.Join(group.prefix + subpattern)\n\tgroup.engine.router.addRoute(method, pattern, handler)\n}", "func NewSumAggregator() SumAggregator {\n\treturn SumAggregator{}\n}", "func Register(r adapter.Registrar) {\n\tr.RegisterMetricsBuilder(newBuilder())\n}", "func (p *PromMetrics) register() {\n\tp.registry.MustRegister(p.nodeDrainTotal)\n\tp.registry.MustRegister(p.nodeDrainFail)\n\tp.registry.MustRegister(p.nodeReapTotal)\n\tp.registry.MustRegister(p.nodeReapFail)\n}", "func NewAggregator(t testing.TB) *Aggregator {\n\tmock := &Aggregator{}\n\n\tt.Cleanup(func() { mock.AssertExpectations(t) })\n\n\treturn mock\n}", "func (c *Component) Register() {}", "func (b *Bucket) Register(r *route.Router, ins extpromhttp.InstrumentationMiddleware) {\n\tinstrf := func(name string, next func(w http.ResponseWriter, r *http.Request)) http.HandlerFunc {\n\t\treturn ins.NewHandler(name, http.HandlerFunc(next))\n\t}\n\n\tr.Get(\"/\", instrf(\"root\", b.root))\n\tr.Get(\"/static/*filepath\", instrf(\"static\", b.serveStaticAsset))\n}", "func (this *Sum) Register(store *core.Store) {\n\tallvars := this.GetAllVars()\n\n\tvar domains map[core.VarId]core.Domain\n\n\tthis.inCh, domains, this.outCh =\n\t\tstore.RegisterPropagatorMap(allvars, this.id)\n\n\tthis.varidToDomainMap = core.GetVaridToIntervalDomains(domains)\n\n\tthis.store = store\n\n\tthis.iColl = indexical.CreateIndexicalCollection()\n\n\tthis.iColl.AddIndexicalsAtPrio(MakeSumBoundsIndexicals(this.pseudoProps, this.varidToDomainMap), indexical.HIGHEST)\n\tthis.iColl.AddIndexicalsAtPrio(MakeSumArcIndexicals(this.pseudoProps, this.varidToDomainMap), indexical.HIGH)\n}", "func Register(r *gin.RouterGroup, s *Service) {\n\tendpoint := r.Group(\"/experiments\")\n\n\t// TODO: add more api handlers\n\tendpoint.GET(\"\", s.listExperiments)\n\tendpoint.POST(\"/new\", s.createExperiment)\n\tendpoint.DELETE(\"/detail/:ns/:name\", s.deleteExperiment)\n\tendpoint.GET(\"/delete/:ns/:name\", s.getExperimentDetail)\n\tendpoint.GET(\"/state\", s.state)\n}", "func NewEventAggregator() *EventAggregator {\n\treturn &EventAggregator{\n\t\tlisteners: make(map[string][]func(interface{})),\n\t}\n}", "func (r *Resolver) Register(name string, e Extension) bool {\n\tr.Execers[name] = e\n\treturn true\n}", "func (c *ContainerBuilder) WithStatsAggregator(aggregators []workloads.StatsAggregator) *ContainerBuilder {\n\tc.Container.StatsAggregator = aggregators\n\treturn c\n}", "func (runner *runnerImpl) Register(g process.GetPID) error {\n\treturn runner.manager.Register(func() (int, error) {\n\t\tpid, err := g()\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\n\t\tfor _, hook := range runner.hooks {\n\t\t\tgo hook.PID(pid)\n\t\t}\n\n\t\treturn pid, err\n\t})\n}", "func Register(c client.Client, logger logr.Logger) {\n\trsh := &appPipelineHandler{\n\t\tclient: c,\n\t\tlogger: logger.WithName(\"appPipelineHandler\"),\n\t}\n\n\twebhook.Register(\"appPipeline\", rsh.filter, rsh.operation)\n}", "func (gd *GrokData) Register(args map[string]interface{}) (error) {\n\n\tgd.parseConfig(args)\n\t\n\terr := gd.grok.AddPattern(\"WORD\", \"\\\\w+\")\t\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = gd.grok.Compile(gd.pattern)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func Register(g *echo.Group, apps *app.Container, m *middleware.Middleware) {\n\th := &handler{\n\t\tapps: apps,\n\t}\n\n\tg.GET(\"\", h.getAllAcompanhamento, m.Auth.Public)\n\tg.GET(\"/:acompanhamento_id\", h.getAcompanhamentoById, m.Auth.Public)\n\tg.GET(\"/byProcedimento/:procedimento_id\", h.getAcompanhamentoByIdProcedimento, m.Auth.Public)\n\tg.POST(\"/anything\", h.getAcompanhamentoByAnything, m.Auth.Public)\n\tg.POST(\"\", h.setAcompanhamento, m.Auth.Public)\n\tg.PUT(\"\", h.updateAcompanhamento, m.Auth.Public)\n\tg.DELETE(\"/:acompanhamento_id\", h.deleteAcompanhamento, m.Auth.Public)\n}", "func RegisterAgent(a *Agent) {\n\t// AgentSotre()[a.Name] = a\n\tif len(a.ID) == 0 {\n\t\ta.ID = genID()\n\t}\n\tGetAgentStoreInstance().Add(a)\n}", "func RegisterAgentAndScale(a *Agent, scale int) {\n\tfor i := 0; i < scale; i++ {\n\t\tagt := *a\n\t\tname := fmt.Sprintf(\"%v_%v\", a.Name, i)\n\t\tagt.Name = name\n\t\tGetAgentStoreInstance().Add(&agt)\n\t}\n}", "func RegisterMarshaller(mime string, marshaller Marshaller) {\n\tmarshallers[mime] = marshaller\n}", "func Register(i interface{}) {\n\tregister(i)\n}", "func (t *targetrunner) register() error {\n\tjsbytes, err := json.Marshal(t.si)\n\tif err != nil {\n\t\tglog.Errorf(\"Unexpected failure to json-marshal %+v, err: %v\", t.si, err)\n\t\treturn err\n\t}\n\turl := ctx.config.Proxy.URL + \"/\" + Rversion + \"/\" + Rcluster\n\t_, err = t.call(url, http.MethodPost, jsbytes)\n\treturn err\n}", "func NewAggregator(c RateCounter, opts ...AggregatorOption) *Aggregator {\n\ta := &Aggregator{\n\t\tcounter: c,\n\t\tpollingInterval: time.Minute,\n\t\tmaxRateBuckets: 10,\n\t}\n\n\tfor _, o := range opts {\n\t\to(a)\n\t}\n\n\ta.data = ring.New(a.maxRateBuckets)\n\treturn a\n}", "func (bus *EventBus) Register(h Handler) {\n\tbus.register <- h\n}", "func (ctl *MealController) register() {\n\tmeals := ctl.router.Group(\"/meals\")\n\n\tmeals.GET(\"\", ctl.ListMeal)\n\n\t// CRUD\n\tmeals.POST(\"\", ctl.CreateMeal)\n\tmeals.GET(\":id\", ctl.GetMeal)\n\tmeals.PUT(\":id\", ctl.UpdateMeal)\n\tmeals.DELETE(\":id\", ctl.DeleteMeal)\n}", "func (c *consulCoordinator) RegisterAsPublisher(compID StreamID) error {\n\treturn c.registerAsPublisher(c.sessionID, compID)\n}", "func Register(router *mux.Router, log logging.Logger) {\n\n\tloggingInterceptor := logging.NewLoggingInterceptor(log)\n\n\trouter.HandleFunc(\"/health-check\", healthCheck).Name(\"health-check\")\n\n\ttestRouter := router.PathPrefix(\"/test\").Subrouter()\n\ttestRouter.Handle(\"\", newTestHandler(log)).Methods(http.MethodPost).Name(\"test\")\n\ttestRouter.Use(loggingInterceptor.LogIntercept)\n}", "func (r *DefaultTaggedRegistry) Register(name string, tags Tags, i interface{}) error {\n\tr.mutex.Lock()\n\tdefer r.mutex.Unlock()\n\treturn r.register(r.metrics, name, tags, i)\n}", "func NewAggregatorFilterer(address common.Address, filterer bind.ContractFilterer) (*AggregatorFilterer, error) {\n\tcontract, err := bindAggregator(address, nil, nil, filterer)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &AggregatorFilterer{contract: contract}, nil\n}", "func register(reg *prometheus.Registry) {\n\treg.MustRegister(\n\t\tdroppedBatchVec,\n\t\tdroppedBytesVec,\n\t\trotateVec,\n\t\tputVec,\n\t\tgetVec,\n\t\tputBytesVec,\n\t\twakeupVec,\n\t\tgetBytesVec,\n\n\t\tcapVec,\n\t\tbatchSizeVec,\n\t\tmaxDataVec,\n\t\tsizeVec,\n\t\tdatafilesVec,\n\n\t\tgetLatencyVec,\n\t\tputLatencyVec)\n}", "func Register() {\n\tregisterConstMetrics()\n\t{{[- if .API.Enabled ]}}\n\tregisterGRPCMetrics()\n\t{{[- end ]}}\n\t{{[- if .Storage.Enabled ]}}\n\tregisterDatabaseMetrics()\n\t{{[- end ]}}\n\tregisterBusinessMetrics()\n}", "func RegisterCompactor(topic string, compactor interface{}) {\n\tpanic(\"not implemented\")\n}", "func (t *Transformer) Register(p ETHProxy) error {\n\tif t.transformers == nil {\n\t\tt.transformers = make(map[string]ETHProxy)\n\t}\n\n\tm := p.Method()\n\tif _, ok := t.transformers[m]; ok {\n\t\treturn errors.Errorf(\"method already exist: %s \", m)\n\t}\n\n\tt.transformers[m] = p\n\n\treturn nil\n}", "func (a *AccessControlCommand) Register(app *kingpin.Application) {\n\taclCmd := app.Command(\"acl\", \"Generate and view ACL options in GrafanaCloud.\")\n\n\tgenerateHeaderCmd := aclCmd.Command(\"generate-header\", \"Generate the header that needs to be passed to the datasource for setting ACLs.\").Action(a.generateHeader)\n\tgenerateHeaderCmd.Flag(\"id\", \"Cortex tenant ID, alternatively set CORTEX_TENANT_ID.\").Envar(\"CORTEX_TENANT_ID\").Required().StringVar(&a.InstanceID)\n\tgenerateHeaderCmd.Flag(\"rule\", \"The access control rules (Prometheus selectors). Set it multiple times to set multiple rules.\").Required().StringsVar(&a.ACLs)\n}", "func Register(name string, obj interface{}) {\n\ti.Register(name, obj)\n}", "func (app *manager) Register(bridge Bridge) {\n\t// hydrated:\n\thydrated := bridge.Hydrated()\n\n\t// hydrated pointer:\n\thydratedPtrType := reflect.Indirect(reflect.ValueOf(hydrated.Pointer())).Type()\n\thydratedPtrName := fmt.Sprintf(doubleStringPattern, hydratedPtrType.PkgPath(), hydratedPtrType.Name())\n\tapp.mp[hydratedPtrName] = bridge\n\n\t// dehydrated:\n\tdehyrated := bridge.Dehydrated()\n\n\t// dehydrated interface:\n\tdehyratedInterfaceType := reflect.TypeOf(dehyrated.Interface())\n\tdehyratedInterfaceName := fmt.Sprintf(doubleStringPattern, dehyratedInterfaceType.PkgPath(), dehyratedInterfaceType.Name())\n\tapp.mp[dehyratedInterfaceName] = bridge\n\n\t// dehydrated pointer:\n\tdehydratedPtrType := reflect.Indirect(reflect.ValueOf(dehyrated.Pointer())).Type()\n\tdehydratedPtrName := fmt.Sprintf(doubleStringPattern, dehydratedPtrType.PkgPath(), dehydratedPtrType.Name())\n\tapp.mp[dehydratedPtrName] = bridge\n}", "func RegisterFormatter() {\n\tformatterRegister.Do(func() {\n\t\tif ResultsPath == \"\" {\n\t\t\tResultsPath = \"./allure-results\"\n\t\t}\n\n\t\tgodog.Format(\"allure\", \"Allure formatter.\",\n\t\t\tfunc(suite string, writer io.Writer) formatters.Formatter {\n\t\t\t\tif suite == \"\" {\n\t\t\t\t\tsuite = \"Features\"\n\t\t\t\t}\n\n\t\t\t\treturn &formatter{\n\t\t\t\t\tresultsPath: strings.TrimSuffix(ResultsPath, \"/\"),\n\t\t\t\t\tcontainer: &Container{\n\t\t\t\t\t\tUUID: uuid.New().String(),\n\t\t\t\t\t\tStart: getTimestampMs(),\n\t\t\t\t\t\tName: suite,\n\t\t\t\t\t},\n\t\t\t\t\tBaseFmt: godog.NewBaseFmt(suite, writer),\n\t\t\t\t}\n\t\t\t})\n\t})\n}", "func RegisterMarshaller(mime string, m *Marshaller) {\n\tif marshallers == nil {\n\t\tmarshallers = make(map[string]*Marshaller, 0)\n\t}\n\tif _, found := marshallers[mime]; !found {\n\t\tmarshallers[mime] = m\n\t}\n}", "func RegisterCompression(compression Compression) {\n\tcompressions[compression.EncodeCompression()] = compression\n}", "func (m Mediator) Register(handler interface{}) error {\n\ttypeOf := reflect.TypeOf(handler)\n\targKind := typeOf.In(0)\n\n\tif typeOf.NumIn() > 1 {\n\t\tif argIsContext(argKind) {\n\t\t\targKind = typeOf.In(1)\n\t\t}\n\t}\n\n\t_, exist := m.registrations[argKind]\n\tif exist {\n\t\treturn fmt.Errorf(\"handler already registered for command %T\", argKind)\n\t}\n\n\tm.registrations[argKind] = reflect.ValueOf(handler)\n\treturn nil\n}", "func Register(f func(), name string) {\n\tsubscriber = append(subscriber, subscriberStruct{\n\t\tf: f,\n\t\tname: name})\n}", "func (r *CollectionRegistry) RegisterWriter(m pcolh.CollectionElem, f func() (CollectionWriter, error)) {\n\tr.mu.Lock()\n\tdefer r.mu.Unlock()\n\n\tk := r.Key(m)\n\n\tif _, ok := r.Writers[k]; ok {\n\t\tgrpclog.Println(\"Ignored duplicate collection writer registration of \", k)\n\t\treturn\n\t}\n\n\tr.Writers[k] = f\n\n}", "func (h *Handler) Register(e *echo.Echo) {\n\t// Index Handler\n\te.GET(\"/\", h.index)\n\n\t// Groups\n\tapi := e.Group(\"/api\")\n\tadmin := e.Group(\"/admin\")\n\tu := e.Group(\"/utils\")\n\n\t// middlewares\n\tapi.Use(h.fireAuthMWare)\n\n\t// Register the routes\n\th.registerUtils(u)\n\th.registerAPI(api)\n\th.registerAdmin(admin)\n}", "func (v *View) Register(g *echo.Group) {\n\tg.Use(wrapResponse, v.wrapSyncStores, v.logVisit, v.extractLocale)\n\tg.GET(\"/ping\", v.ping)\n\tg.GET(\"/health\", v.health)\n\tv.registerUserHandlers(g)\n\tv.registerScopeHandlers(g)\n\tv.registerRoleHandlers(g)\n\tv.registerSessionHandlers(g)\n\tv.registerContestHandlers(g)\n\tv.registerContestStandingsHandlers(g)\n\tv.registerContestMessageHandlers(g)\n\tv.registerProblemHandlers(g)\n\tv.registerSolutionHandlers(g)\n\tv.registerCompilerHandlers(g)\n\tv.registerSettingHandlers(g)\n\tv.registerLocaleHandlers(g)\n\tv.registerFileHandlers(g)\n}", "func (b *Builder) Register(br *sous.BuildResult) error {\n\tfor _, prod := range br.Products {\n\t\tif prod.Advisories.Contains(sous.IsBuilder) {\n\t\t\tmessages.ReportLogFieldsMessage(\"not pushing builder image\", logging.DebugLevel, b.log, prod)\n\t\t\tcontinue\n\t\t}\n\t\terr := b.pushToRegistry(prod)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\terr = b.recordName(prod)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (client *StdoutClient) Register(subscribeBuilder *SubscribeBuilder) error {\n\tsubscribe := struct {\n\t\tTopic string `json:\"topic\"`\n\t\tEventName string `json:\"name\"`\n\t\tTimestamp time.Time `json:\"timestamp\"`\n\t\tVersion int `json:\"version\"`\n\t}{\n\t\tTopic: subscribeBuilder.topic,\n\t\tEventName: subscribeBuilder.eventName,\n\t\tVersion: defaultVersion,\n\t\tTimestamp: time.Now().UTC(),\n\t}\n\n\teventByte, err := json.Marshal(&subscribe)\n\tif err != nil {\n\t\tlogrus.Errorf(\"unable to marshal event : %s , error : %v\", subscribe.EventName, err)\n\t}\n\n\tfmt.Println(string(eventByte))\n\n\treturn nil\n}", "func Register(name string, builder Builder) {\n\tregistry[name] = builder\n}", "func Register(\n\tlifecycle fx.Lifecycle,\n\tbq *bigquery.Client,\n\tdatabase *firestore.Client,\n\tpostgres *sql.DB,\n\tevents *pubsub.Client,\n\tlogger *zap.SugaredLogger,\n\trouter *mux.Router,\n) {\n\tbq, cfg, database, discord, events, logger, postgres, router := common.Register(\n\t\tlifecycle,\n\t\tbq,\n\t\tdatabase,\n\t\tpostgres,\n\t\tevents,\n\t\tlogger,\n\t\trouter,\n\t)\n\n\tNew(bq, cfg, database, discord, events, logger, postgres, router)\n}", "func (o *Group) Register() int {\n\to.wait_lock.Lock()\n\tdefer o.wait_lock.Unlock()\n\to.wg().Add(1)\n\to.wait_index++\n\to.wait_register[o.wait_index] = true\n\treturn o.wait_index\n}", "func (c *mockMediatorClient) Register(connectionID string) error {\n\tif c.RegisterErr != nil {\n\t\treturn c.RegisterErr\n\t}\n\n\treturn nil\n}", "func (c *Consumer) register() error {\n\tif err := c.zoo.RegisterGroup(c.group, c.topic); err != nil {\n\t\treturn err\n\t}\n\tif err := c.zoo.RegisterConsumer(c.group, c.id, c.topic); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (o GetAggregatorsAggregatorOutput) AggregatorId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetAggregatorsAggregator) string { return v.AggregatorId }).(pulumi.StringOutput)\n}", "func (pc *NginxProcessesMetricsCollector) Register(registry *prometheus.Registry) error {\n\treturn registry.Register(pc)\n}", "func (a *API) RegisterDistributor(d *distributor.Distributor, pushConfig distributor.Config) {\n\ta.RegisterRoute(\"/api/v1/push\", push.Handler(pushConfig, d.Push), true)\n\ta.RegisterRoute(\"/distributor/all_user_stats\", http.HandlerFunc(d.AllUserStatsHandler), false)\n\ta.RegisterRoute(\"/distributor/ha_tracker\", d.HATracker, false)\n\n\t// Legacy Routes\n\ta.RegisterRoute(a.cfg.LegacyHTTPPrefix+\"/push\", push.Handler(pushConfig, d.Push), true)\n\ta.RegisterRoute(\"/all_user_stats\", http.HandlerFunc(d.AllUserStatsHandler), false)\n\ta.RegisterRoute(\"/ha-tracker\", d.HATracker, false)\n}", "func Register(mainRouter *mux.Router) {\n\n\tmoviesService = &MoviesService{\n\t\tDao: &dao.JSON{},\n\t}\n\n\tmainRouter.HandleFunc(\"/movies\", moviesService.HandleGetMovies).Methods(\"GET\").Name(\"get-movies\")\n\tmainRouter.HandleFunc(\"/movies/{id}\", moviesService.HandleGetMovie).Methods(\"GET\").Name(\"get-movie\")\n\tmainRouter.HandleFunc(\"/comments\", moviesService.HandleMostComments).Methods(\"GET\").Name(\"get-most-comments\")\n\tmainRouter.HandleFunc(\"/likes\", moviesService.HandleMostLikes).Methods(\"GET\").Name(\"get-most-likes\")\n}", "func (r *MetricRegistry) RegisterDistribution(\n\tID string,\n\ttags ...string,\n) core.MetricSampleListener {\n\tif strings.HasPrefix(ID, \".\") {\n\t\tID = strings.TrimPrefix(ID, \".\")\n\t}\n\n\t// only add once\n\tif l, ok := r.registeredListeners[ID]; ok {\n\t\treturn l\n\t}\n\n\tr.registeredListeners[ID] = &metricSampleListener{\n\t\tclient: r.client,\n\t\tmetricType: 0,\n\t\tid: r.prefix + ID,\n\t}\n\n\treturn r.registeredListeners[ID]\n}", "func (ctl *AuthorController) register() {\n\tAuthors := ctl.router.Group(\"/Authors\")\n\n\tAuthors.GET(\"\", ctl.ListAuthor)\n\n\t// CRUD\n\tAuthors.POST(\"\", ctl.CreateAuthor)\n\tAuthors.GET(\":id\", ctl.GetAuthor)\n\tAuthors.PUT(\":id\", ctl.UpdateAuthor)\n\tAuthors.DELETE(\":id\", ctl.DeleteAuthor)\n}", "func (f *contentSpecMgr) Reg(contentType string, spec ContentResolver, unitTest ...bool) {\n\terr := f._reg(contentType, spec)\n\tif err != nil && len(unitTest) == 0 {\n\t\tdebug.Assert(false)\n\t\tcos.ExitLog(err)\n\t}\n}", "func (r *RepoStruct) Register() {\n\tif r == nil {\n\t\treturn\n\t}\n\tif r.forge == nil {\n\t\treturn\n\t}\n\tr.forge.ForjCore.Repos[r.name] = r\n}", "func (n *NetworkBuilder) WithStatsAggregator(aggregators []workloads.StatsAggregator) *NetworkBuilder {\n\tn.Network.StatsAggregator = aggregators\n\treturn n\n}", "func Register() {\n\tglobalLock.Lock()\n\tdefer globalLock.Unlock()\n\tregister()\n}", "func Register(fn interface{}) error {\n\t// Validate that its a function\n\tfnType := reflect.TypeOf(fn)\n\tif err := validateFnFormat(fnType); err != nil {\n\t\treturn err\n\t}\n\t// Check if already registered\n\tfnName := getFunctionName(fn)\n\t_, ok := fnLookup.getFn(fnName)\n\tif ok {\n\t\treturn nil\n\t}\n\tfor i := 0; i < fnType.NumIn(); i++ {\n\t\targType := fnType.In(i)\n\t\t// Interfaces cannot be registered, their implementations should be\n\t\t// https://golang.org/pkg/encoding/gob/#Register\n\t\tif argType.Kind() != reflect.Interface {\n\t\t\targ := reflect.Zero(argType).Interface()\n\t\t\tif err := GlobalBackend().Encoder().Register(arg); err != nil {\n\t\t\t\treturn errors.Wrap(err, \"unable to register the message for encoding\")\n\t\t\t}\n\t\t}\n\t}\n\tfnLookup.addFn(fnName, fn)\n\treturn nil\n}", "func register() {\n\tapimetrics.Register()\n\tcachermetrics.Register()\n\tetcd3metrics.Register()\n\tflowcontrolmetrics.Register()\n\tpeerproxymetrics.Register()\n}", "func (ctl *ManagerController) register() {\n\tmanagers := ctl.router.Group(\"/managers\")\n\n\tmanagers.GET(\"\", ctl.ListManager)\n\n\t// CRUD\n\tmanagers.POST(\"\", ctl.CreateManager)\n\tmanagers.GET(\":id\", ctl.GetManager)\n\tmanagers.PUT(\":id\", ctl.UpdateManager)\n\tmanagers.DELETE(\":id\", ctl.DeleteManager)\n}", "func (a *GaugeVectorMetric) register() {\n\tmetrics := gaugeVectors[a.ServiceID]\n\tif metrics == nil {\n\t\tmetrics = make(map[MetricID]*GaugeVectorMetric)\n\t\tgaugeVectors[a.ServiceID] = metrics\n\t}\n\tif _, exists := metrics[a.MetricID]; !exists {\n\t\tmetricsRegistry.MustRegister(a.GaugeVec)\n\t\tmetrics[a.MetricID] = a\n\t}\n}", "func Register(m map[string]transformer.Transformer) {\n\tm[operationName] = &Store{}\n}", "func (b *QueryBus) Register(queries map[query.Type]query.Handler) {\n\tb.handlers = queries\n}", "func (c *Aggregator) Aggregation() aggregation.Aggregation {\n\treturn c\n}", "func (c *Aggregator) Aggregation() aggregation.Aggregation {\n\treturn c\n}" ]
[ "0.62071866", "0.61488783", "0.60762906", "0.5992745", "0.5945457", "0.5945457", "0.5665146", "0.56404746", "0.5592253", "0.5574909", "0.5513508", "0.54995465", "0.5498623", "0.53810024", "0.53262913", "0.5318158", "0.53082496", "0.52558357", "0.5239182", "0.5219471", "0.52051175", "0.5197365", "0.519475", "0.5187674", "0.5187442", "0.5185735", "0.5184191", "0.51463926", "0.5136662", "0.51100445", "0.51093256", "0.5100269", "0.50823206", "0.50582546", "0.50557154", "0.5055654", "0.50463134", "0.50437653", "0.5041797", "0.50350755", "0.50089025", "0.4987676", "0.4969934", "0.49603897", "0.49552923", "0.4938995", "0.49216783", "0.49022946", "0.4881599", "0.4860054", "0.48536572", "0.48504034", "0.48491707", "0.48399875", "0.48338634", "0.4830872", "0.48136798", "0.48006836", "0.47933152", "0.47920945", "0.47834218", "0.47782955", "0.47776163", "0.47772655", "0.4772378", "0.47672412", "0.47635275", "0.47585258", "0.47539273", "0.4750679", "0.4749609", "0.47494522", "0.47434768", "0.4741316", "0.4733116", "0.47319624", "0.47299823", "0.47264853", "0.47253382", "0.47222888", "0.47222352", "0.47189957", "0.47154757", "0.47150746", "0.47106647", "0.47083166", "0.47058564", "0.4704633", "0.4697394", "0.46957445", "0.46923852", "0.46831888", "0.46813247", "0.46795547", "0.46792737", "0.4677031", "0.46757132", "0.46642053", "0.46587688", "0.46587688" ]
0.73674434
0
Creates a cli.Command for tenant Subcommands: create; Usage: tenant create delete; Usage: tenant delete show; Usage: tenant show list; Usage: tenant list set; Usage: tenant set get; Usage: tenant get tasks; Usage: tenant tasks []
func GetTenantsCommand() cli.Command { command := cli.Command{ Name: "tenant", Usage: "options for tenant", Subcommands: []cli.Command{ { Name: "create", Usage: "Create a new tenant", ArgsUsage: "<tenant-name>", Flags: []cli.Flag{ cli.StringFlag{ Name: "security-groups, s", Usage: "Comma-separated Lightwave group names, to specify the tenant administrators", }, }, Action: func(c *cli.Context) { err := createTenant(c, os.Stdout) if err != nil { log.Fatal("Error: ", err) } }, }, { Name: "delete", Usage: "Delete a tenant", ArgsUsage: "<tenant-id>", Action: func(c *cli.Context) { err := deleteTenant(c) if err != nil { log.Fatal("Error: ", err) } }, }, { Name: "show", Usage: "Show detailed tenant info with specified id", ArgsUsage: "<tenant-id>", Action: func(c *cli.Context) { err := showTenant(c, os.Stdout) if err != nil { log.Fatal(err) } }, }, { Name: "list", Usage: "List all tenants", ArgsUsage: " ", Action: func(c *cli.Context) { err := listTenants(c, os.Stdout) if err != nil { log.Fatal("Error: ", err) } }, }, { Name: "set", Usage: "Set default tenant", ArgsUsage: "<tenant-name>", Description: "Set the default project that will be used for all photon CLI commands that need a project.\n" + " Most commands allow you to override the default.", Action: func(c *cli.Context) { err := setTenant(c) if err != nil { log.Fatal("Error: ", err) } }, }, { Name: "get", Usage: "Get default tenant", ArgsUsage: " ", Description: "Show default project in use for photon CLI commands. Most command allow you to either\n" + " use this default or specify a specific project to use.", Action: func(c *cli.Context) { err := getTenant(c, os.Stdout) if err != nil { log.Fatal("Error: ", err) } }, }, { Name: "tasks", Usage: "Show tenant tasks", ArgsUsage: "<tenant-id>", Flags: []cli.Flag{ cli.StringFlag{ Name: "state, s", Usage: "Filter by task sate", }, }, Action: func(c *cli.Context) { err := getTenantTasks(c, os.Stdout) if err != nil { log.Fatal("Error: ", err) } }, }, { Name: "set-security-groups", Usage: "Set security groups for a tenant", ArgsUsage: "<tenant-id> <comma separated list of groups>", Description: "Set the list of Lightwave groups that can administer this tenant. This may only be\n" + " be set by a member of the tenant. Be cautious--you can remove your own access if you specify\n" + " the wrong set of groups.", Action: func(c *cli.Context) { err := setSecurityGroups(c) if err != nil { log.Fatal("Error: ", err) } }, }, { Hidden: true, Name: "set_security_groups", Usage: "Set security groups for a tenant", ArgsUsage: "<tenant-id> <comma separated list of groups>", Description: "Deprecated, use set-security-groups instead", Action: func(c *cli.Context) { err := setSecurityGroups(c) if err != nil { log.Fatal("Error: ", err) } }, }, }, } return command }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewTenantCreateCmd() *cobra.Command {\n\ttenantCreateCmd := &cobra.Command{\n\t\tUse: \"create\",\n\t\tTraverseChildren: true,\n\t\tShort: \"Create a tenant resource within Karavi\",\n\t\tLong: `Creates a tenant resource within Karavi`,\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\taddr, err := cmd.Flags().GetString(\"addr\")\n\t\t\tif err != nil {\n\t\t\t\treportErrorAndExit(JSONOutput, cmd.ErrOrStderr(), err)\n\t\t\t}\n\n\t\t\tinsecure, err := cmd.Flags().GetBool(\"insecure\")\n\t\t\tif err != nil {\n\t\t\t\treportErrorAndExit(JSONOutput, cmd.ErrOrStderr(), err)\n\t\t\t}\n\n\t\t\ttenantClient, conn, err := CreateTenantServiceClient(addr, insecure)\n\t\t\tif err != nil {\n\t\t\t\treportErrorAndExit(JSONOutput, cmd.ErrOrStderr(), err)\n\t\t\t}\n\t\t\tdefer conn.Close()\n\n\t\t\tname, err := cmd.Flags().GetString(\"name\")\n\t\t\tif err != nil {\n\t\t\t\treportErrorAndExit(JSONOutput, cmd.ErrOrStderr(), err)\n\t\t\t}\n\t\t\tif strings.TrimSpace(name) == \"\" {\n\t\t\t\treportErrorAndExit(JSONOutput, cmd.ErrOrStderr(), errors.New(\"empty name not allowed\"))\n\t\t\t}\n\n\t\t\t_, err = tenantClient.CreateTenant(context.Background(), &pb.CreateTenantRequest{\n\t\t\t\tTenant: &pb.Tenant{\n\t\t\t\t\tName: name,\n\t\t\t\t},\n\t\t\t})\n\t\t\tif err != nil {\n\t\t\t\treportErrorAndExit(JSONOutput, cmd.ErrOrStderr(), err)\n\t\t\t}\n\t\t},\n\t}\n\n\ttenantCreateCmd.Flags().StringP(\"name\", \"n\", \"\", \"Tenant name\")\n\treturn tenantCreateCmd\n}", "func createTenant(c *cli.Context, w io.Writer) error {\n\tif len(c.Args()) > 1 {\n\t\treturn fmt.Errorf(\"Unknown argument: %v\", c.Args()[1:])\n\t}\n\tname := c.Args().First()\n\tsecurityGroups := c.String(\"security-groups\")\n\n\tif !c.GlobalIsSet(\"non-interactive\") {\n\t\tvar err error\n\t\tname, err = askForInput(\"Tenant name: \", name)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tsecurityGroups, err =\n\t\t\taskForInput(\"Comma-separated security group names, or hit enter for no security groups): \",\n\t\t\t\tsecurityGroups)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif len(name) == 0 {\n\t\treturn fmt.Errorf(\"Please provide tenant name\")\n\t}\n\tsecurityGroupList := []string{}\n\tif securityGroups != \"\" {\n\t\tsecurityGroupList = regexp.MustCompile(`\\s*,\\s*`).Split(securityGroups, -1)\n\t}\n\n\ttenantSpec := &photon.TenantCreateSpec{\n\t\tName: name,\n\t\tSecurityGroups: securityGroupList,\n\t}\n\n\tvar err error\n\tclient.Photonclient, err = client.GetClient(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tcreateTask, err := client.Photonclient.Tenants.Create(tenantSpec)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tid, err := waitOnTaskOperation(createTask.ID, c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif utils.NeedsFormatting(c) {\n\t\ttenant, err := client.Photonclient.Tenants.Get(id)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tutils.FormatObject(tenant, w, c)\n\t}\n\n\treturn nil\n}", "func TenantCreate(tenantID, userEmail, tenantName, UserName string) {\n\tRunCmd(fmt.Sprintf(\"%s tenant -op=add -id=%s -name=\\\"%s\\\" -user-email=%s -user-name=\\\"%s\\\"\",\n\t\tActlPath, tenantID, tenantName, userEmail, UserName))\n}", "func NewCmdCleanUpTenant(f cmdutil.Factory) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"tenant\",\n\t\tShort: \"Hard delete of your tenant pipelines, apps, jobs and releases\",\n\t\tLong: `Hard delete of your tenant pipelines, apps, jobs and releases`,\n\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\tp := cleanUpTenantFlags{}\n\t\t\tif cmd.Flags().Lookup(yesFlag).Value.String() == \"true\" {\n\t\t\t\tp.confirm = true\n\t\t\t}\n\t\t\terr := p.cleanTenant(f)\n\t\t\tif err != nil {\n\t\t\t\tutil.Fatalf(\"%s\\n\", err)\n\t\t\t}\n\t\t\treturn\n\t\t},\n\t}\n\treturn cmd\n}", "func createCommand(t *runner.Task, actionFunc func(*cli.Context) error) *cli.Command {\n\tcommand := &cli.Command{\n\t\tName: t.Name,\n\t\tUsage: strings.TrimSpace(t.Usage),\n\t\tDescription: strings.TrimSpace(t.Description),\n\t\tAction: actionFunc,\n\t}\n\n\tfor _, arg := range t.Args {\n\t\tcommand.ArgsUsage += fmt.Sprintf(\"<%s> \", arg.Name)\n\t}\n\n\tcommand.CustomHelpTemplate = createCommandHelp(t)\n\n\treturn command\n}", "func createCommand() cli.Command {\n\treturn cli.Command{\n\t\tName: \"create\",\n\t\tBefore: survey.RequireGlobalFlagsFunc(requiredFlags...),\n\t\tAction: create(),\n\t}\n}", "func newDeleteCmd(out io.Writer, errOut io.Writer) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"delete\",\n\t\tShort: \"deletes various entities managed by the Ziti Edge Controller\",\n\t\tLong: \"deletes various entities managed by the Ziti Edge Controller\",\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\terr := cmd.Help()\n\t\t\tcmdhelper.CheckErr(err)\n\t\t},\n\t}\n\n\tnewOptions := func() *edgeOptions {\n\t\treturn &edgeOptions{\n\t\t\tCommonOptions: common.CommonOptions{\n\t\t\t\tOut: out,\n\t\t\t\tErr: errOut,\n\t\t\t},\n\t\t}\n\t}\n\n\tcmd.AddCommand(newDeleteCmdForEntityType(\"api-session\", newOptions()))\n\tcmd.AddCommand(newDeleteAuthenticatorCmd(\"authenticator\", newOptions()))\n\tcmd.AddCommand(newDeleteCmdForEntityType(\"ca\", newOptions()))\n\tcmd.AddCommand(newDeleteCmdForEntityType(\"config\", newOptions()))\n\tcmd.AddCommand(newDeleteCmdForEntityType(\"config-type\", newOptions()))\n\tcmd.AddCommand(newDeleteCmdForEntityType(\"edge-router\", newOptions(), \"er\", \"ers\"))\n\tcmd.AddCommand(newDeleteCmdForEntityType(\"edge-router-policy\", newOptions(), \"erp\", \"erps\"))\n\tcmd.AddCommand(newDeleteCmdForEntityType(\"identity\", newOptions()))\n\tcmd.AddCommand(newDeleteCmdForEntityType(\"service\", newOptions()))\n\tcmd.AddCommand(newDeleteCmdForEntityType(\"service-edge-router-policy\", newOptions(), \"serp\", \"serps\"))\n\tcmd.AddCommand(newDeleteCmdForEntityType(\"service-policy\", newOptions(), \"sp\", \"sps\"))\n\tcmd.AddCommand(newDeleteCmdForEntityType(\"session\", newOptions()))\n\tcmd.AddCommand(newDeleteCmdForEntityType(\"terminator\", newOptions()))\n\tcmd.AddCommand(newDeleteCmdForEntityType(\"posture-check\", newOptions()))\n\n\treturn cmd\n}", "func NewTxCmd() *cobra.Command {\n\ttxCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: \"merkledrop transaction subcommands\",\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\ttxCmd.AddCommand(\n\t\tGetCmdCreate(),\n\t\tGetCmdClaim(),\n\t)\n\n\treturn txCmd\n}", "func newCreateCmd() *cobra.Command {\n\tcreateCmd := cobra.Command{\n\t\tUse: \"create\",\n\t\tShort: `Create a new verless object`,\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\treturn cmd.Help()\n\t\t},\n\t}\n\n\tcreateCmd.AddCommand(newCreateProjectCmd())\n\tcreateCmd.AddCommand(newCreateThemeCmd())\n\tcreateCmd.AddCommand(newCreateFile())\n\n\treturn &createCmd\n}", "func newDestroyCmd() *cobra.Command {\n\tdc := &destroyCmd{}\n\n\tdestroyCmd := &cobra.Command{\n\t\tUse: \"destroy [-f SORUCE]\",\n\t\tShort: \"Destroy tau managed infrastructure\",\n\t\tLong: destroyLong,\n\t\tExample: destroyExample,\n\t\tDisableFlagsInUseLine: true,\n\t\tSilenceUsage: true,\n\t\tSilenceErrors: true,\n\t\tArgs: cobra.MaximumNArgs(0),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tif err := dc.meta.init(args); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn dc.run(args)\n\t\t},\n\t}\n\n\tf := destroyCmd.Flags()\n\tf.BoolVar(&dc.autoApprove, \"auto-approve\", false, \"auto approve destruction\")\n\n\tdc.addMetaFlags(destroyCmd)\n\n\treturn destroyCmd\n}", "func newGenerateCmd() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"generate\",\n\t\tShort: \"bulk generate key pairs\",\n\t\tLong: `\nThe generate sub command is intended only for test nets. It generates a \nnumber of key pairs and places them in the current monet keystore. The \naccounts are names <prefix><suffix> where prefix is set by --prefix (default\n\"Account\") and suffix is a number between --min-suffix and --max-suffix \ninclusive. The defaults are 1 and 5.\n`,\n\t\tArgs: cobra.ArbitraryArgs,\n\t\tRunE: generateKey,\n\t}\n\n\tcmd.Flags().StringVar(&prefix, \"prefix\", prefix, \"prefix for account monikers\")\n\tcmd.Flags().IntVar(&minSuffix, \"min-suffix\", minSuffix, \"minimum suffix for account monikers\")\n\tcmd.Flags().IntVar(&maxSuffix, \"max-suffix\", maxSuffix, \"maximum suffix for account monikers\")\n\n\tviper.BindPFlags(cmd.Flags())\n\n\treturn cmd\n}", "func SubCommands() []cli.Command {\n\treturn []cli.Command{\n\t\t{\n\t\t\tName: \"list\",\n\t\t\tUsage: \"Lists the domains of the account group.\",\n\t\t\tAction: cmd.DomainList,\n\t\t},\n\t\t{\n\t\t\tName: \"show\",\n\t\t\tUsage: \"Shows information about a specific domain.\",\n\t\t\tAction: cmd.DomainShow,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Domain Id\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"create\",\n\t\t\tUsage: \"Creates a new domain.\",\n\t\t\tAction: cmd.DomainCreate,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"name\",\n\t\t\t\t\tUsage: \"Fully-qualified domain name (FQDN)\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"ttl\",\n\t\t\t\t\tUsage: \"Time to live (TTL) of the Start of Authority (SOA) record\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"contact\",\n\t\t\t\t\tUsage: \"Contact e-mail\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"minimum\",\n\t\t\t\t\tUsage: \"The minimum TTL of the SOA record\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"update\",\n\t\t\tUsage: \"Updates an existing domain\",\n\t\t\tAction: cmd.DomainUpdate,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Domain Id\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"ttl\",\n\t\t\t\t\tUsage: \"Time to live (TTL) of the Start of Authority (SOA) record\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"contact\",\n\t\t\t\t\tUsage: \"Contact e-mail\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"minimum\",\n\t\t\t\t\tUsage: \"The minimum TTL of the SOA record\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"delete\",\n\t\t\tUsage: \"Deletes a domain\",\n\t\t\tAction: cmd.DomainDelete,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Domain Id\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"list_domain_records\",\n\t\t\tUsage: \"Lists the DNS records of a domain.\",\n\t\t\tAction: cmd.DomainRecordList,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"domain_id\",\n\t\t\t\t\tUsage: \"Domain Id\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"show_domain_record\",\n\t\t\tUsage: \"Shows information about a specific DNS record.\",\n\t\t\tAction: cmd.DomainRecordShow,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"domain_id\",\n\t\t\t\t\tUsage: \"Domain Id\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Record Id\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"create_domain_record\",\n\t\t\tUsage: \"Creates a new DNS record.\",\n\t\t\tAction: cmd.DomainRecordCreate,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"domain_id\",\n\t\t\t\t\tUsage: \"Domain Id\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"type\",\n\t\t\t\t\tUsage: \"Type of record (A, AAAA, CNAME, MX, TXT)\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"name\",\n\t\t\t\t\tUsage: \"Record name\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"content\",\n\t\t\t\t\tUsage: \"Record content\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"ttl\",\n\t\t\t\t\tUsage: \"Time to live (TTL)\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"prio\",\n\t\t\t\t\tUsage: \"Priority (only MX records)\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"server_id\",\n\t\t\t\t\tUsage: \"Identifier of the associated server (only A and AAAA records)\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"update_domain_record\",\n\t\t\tUsage: \"Updates an existing DNS record.\",\n\t\t\tAction: cmd.DomainRecordUpdate,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"domain_id\",\n\t\t\t\t\tUsage: \"Domain Id\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Record Id\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"name\",\n\t\t\t\t\tUsage: \"Record name\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"content\",\n\t\t\t\t\tUsage: \"Record content\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"ttl\",\n\t\t\t\t\tUsage: \"Time to live (TTL)\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"prio\",\n\t\t\t\t\tUsage: \"Priority (only MX records)\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"server_id\",\n\t\t\t\t\tUsage: \"Identifier of the associated server (only A and AAAA records)\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"delete_domain_record\",\n\t\t\tUsage: \"Deletes a DNS record\",\n\t\t\tAction: cmd.DomainRecordDelete,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Record Id\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"domain_id\",\n\t\t\t\t\tUsage: \"Record Id\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n}", "func newCreateIdentityCmd(f cmdutil.Factory, out io.Writer, errOut io.Writer) *cobra.Command {\n\tnewOptions := func() *createIdentityOptions {\n\t\treturn &createIdentityOptions{\n\t\t\tedgeOptions: edgeOptions{\n\t\t\t\tCommonOptions: common.CommonOptions{Factory: f, Out: out, Err: errOut},\n\t\t\t},\n\t\t}\n\t}\n\n\tcmd := &cobra.Command{\n\t\tUse: \"identity\",\n\t\tShort: \"creates a new identity managed by the Ziti Edge Controller\",\n\t\tLong: \"creates a new identity managed by the Ziti Edge Controller\",\n\t}\n\n\tcmd.AddCommand(newCreateIdentityOfTypeCmd(\"device\", newOptions()))\n\tcmd.AddCommand(newCreateIdentityOfTypeCmd(\"user\", newOptions()))\n\tcmd.AddCommand(newCreateIdentityOfTypeCmd(\"service\", newOptions()))\n\n\treturn cmd\n}", "func CreateTenant(tenant string) {\n\tconn, err := grpc.Dial(\"localhost:8885\", grpc.WithInsecure())\n\tif err != nil {\n\t\tfmt.Printf(\"Cannot open gRPC: %s\", err)\n\t\t// TODO error\n\t}\n\t\n\tdefer conn.Close()\n\n\tclient := schema.NewTenantServiceClient(conn)\n\n\tnewTenant, err := client.Create(context.Background(), &wrapperspb.StringValue{Value: tenant})\n\tif err != nil {\n\t\tfmt.Errorf(\"Cannot create a tenant: %s\", err)\n\t}\n\n\tfmt.Printf(\"Tenant created successfully: %s\", newTenant)\n}", "func SubCommands() []cli.Command {\n\treturn []cli.Command{\n\t\t{\n\t\t\tName: \"list\",\n\t\t\tUsage: \"Lists all available templates\",\n\t\t\tAction: cmd.TemplateList,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"labels\",\n\t\t\t\t\tUsage: \"A list of comma separated label as a query filter\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"show\",\n\t\t\tUsage: \"Shows information about a specific template\",\n\t\t\tAction: cmd.TemplateShow,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Template Id\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"create\",\n\t\t\tUsage: \"Creates a new template.\",\n\t\t\tAction: cmd.TemplateCreate,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"name\",\n\t\t\t\t\tUsage: \"Name of the template\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"generic-image-id\",\n\t\t\t\t\tUsage: \"Identifier of the OS image that the template builds on\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"run-list\",\n\t\t\t\t\tUsage: \"A list of comma separated cookbook recipes that is run on the servers at start-up, i.e: --run-list imco::client,1password,joomla\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"cookbook-versions\",\n\t\t\t\t\tUsage: \"The cookbook versions used to configure the service recipes in the run-list, i.e: --cookbook-versions \\\"imco:3.0.3,1password~>1.3.0,joomla:0.11.0\\\" \\n\\tCookbook version format: [NAME<OPERATOR>VERSION] \\n\\tSupported Operators:\\n\\t\\tChef supermarket cookbook '~>','=','>=','>','<','<='\\n\\t\\tUploaded cookbook ':'\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"configuration-attributes\",\n\t\t\t\t\tUsage: \"The attributes used to configure the service recipes in the run-list, as a json formatted parameter. i.e: --configuration-attributes '{\\\"joomla\\\":{\\\"db\\\":{\\\"password\\\":\\\"my_pass\\\"},\\\"port\\\":\\\"8080\\\"}}'\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"configuration-attributes-from-file\",\n\t\t\t\t\tUsage: \"The attributes used to configure the service recipes in the run-list, from file or STDIN, as a json formatted parameter. \\n\\tFrom file: --configuration-attributes-from-file attrs.json \\n\\tFrom STDIN: --configuration-attributes-from-file -\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"labels\",\n\t\t\t\t\tUsage: \"A list of comma separated label names to be associated with template\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"update\",\n\t\t\tUsage: \"Updates an existing template\",\n\t\t\tAction: cmd.TemplateUpdate,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Template Id\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"name\",\n\t\t\t\t\tUsage: \"Name of the template\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"run-list\",\n\t\t\t\t\tUsage: \"A list of comma separated cookbook recipes that is run on the servers at start-up, i.e: --run-list imco::client,1password,joomla\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"cookbook-versions\",\n\t\t\t\t\tUsage: \"The cookbook versions used to configure the service recipes in the run-list, i.e: --cookbook-versions \\\"imco:3.0.3,1password~>1.3.0,joomla:0.11.0\\\" \\n\\tCookbook version format: [NAME<OPERATOR>VERSION] \\n\\tSupported Operators:\\n\\t\\tChef supermarket cookbook '~>','=','>=','>','<','<='\\n\\t\\tUploaded cookbook ':'\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"configuration-attributes\",\n\t\t\t\t\tUsage: \"The attributes used to configure the service recipes in the run-list, as a json formatted parameter. i.e: --configuration-attributes '{\\\"joomla\\\":{\\\"db\\\":{\\\"password\\\":\\\"my_pass\\\"},\\\"port\\\":\\\"8080\\\"}}'\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"configuration-attributes-from-file\",\n\t\t\t\t\tUsage: \"The attributes used to configure the service recipes in the run-list, from file or STDIN, as a json formatted parameter. \\n\\tFrom file: --configuration-attributes-from-file attrs.json \\n\\tFrom STDIN: --configuration-attributes-from-file -\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"compile\",\n\t\t\tUsage: \"Compiles an existing template\",\n\t\t\tAction: cmd.TemplateCompile,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Template Id\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"delete\",\n\t\t\tUsage: \"Deletes a template\",\n\t\t\tAction: cmd.TemplateDelete,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Template Id\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"list-template-scripts\",\n\t\t\tUsage: \"Shows the script characterisations of a template\",\n\t\t\tAction: cmd.TemplateScriptList,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"template-id\",\n\t\t\t\t\tUsage: \"Template Id\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"type\",\n\t\t\t\t\tUsage: \"Must be \\\"operational\\\", \\\"boot\\\" or \\\"shutdown\\\"\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"show-template-script\",\n\t\t\tUsage: \"Shows information about a specific script characterisation\",\n\t\t\tAction: cmd.TemplateScriptShow,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"template-id\",\n\t\t\t\t\tUsage: \"Template Id\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Script Id\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"create-template-script\",\n\t\t\tUsage: \"Creates a new script characterisation for a template and appends it to the list of script characterisations of the same type.\",\n\t\t\tAction: cmd.TemplateScriptCreate,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"template-id\",\n\t\t\t\t\tUsage: \"Template Id\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"type\",\n\t\t\t\t\tUsage: \"Must be \\\"operational\\\", \\\"boot\\\" or \\\"shutdown\\\"\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"script-id\",\n\t\t\t\t\tUsage: \"Identifier for the script that is parameterised by the script characterisation\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"parameter-values\",\n\t\t\t\t\tUsage: \"A map that assigns a value to each script parameter, as a json formatted parameter; i.e: '{\\\"param1\\\":\\\"val1\\\",\\\"param2\\\":\\\"val2\\\"}'\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"parameter-values-from-file\",\n\t\t\t\t\tUsage: \"A map that assigns a value to each script parameter, from file or STDIN, as a json formatted parameter. \\n\\tFrom file: --parameter-values-from-file params.json \\n\\tFrom STDIN: --parameter-values-from-file -\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"update-template-script\",\n\t\t\tUsage: \"Updates an existing script characterisation for a template.\",\n\t\t\tAction: cmd.TemplateScriptUpdate,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"template-id\",\n\t\t\t\t\tUsage: \"Template Id\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Identifier for the template-script that is parameterised by the script characterisation\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"parameter-values\",\n\t\t\t\t\tUsage: \"A map that assigns a value to each script parameter, as a json formatted parameter; i.e: '{\\\"param1\\\":\\\"val1\\\",\\\"param2\\\":\\\"val2\\\"}'\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"parameter-values-from-file\",\n\t\t\t\t\tUsage: \"A map that assigns a value to each script parameter, from file or STDIN, as a json formatted parameter. \\n\\tFrom file: --parameter-values-from-file params.json \\n\\tFrom STDIN: --parameter-values-from-file -\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"reorder-template-scripts\",\n\t\t\tUsage: \"Reorders the scripts of the template and type specified according to the provided order, changing their execution order as corresponds.\",\n\t\t\tAction: cmd.TemplateScriptReorder,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"template-id\",\n\t\t\t\t\tUsage: \"Template Id\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"type\",\n\t\t\t\t\tUsage: \"Must be \\\"operational\\\", \\\"boot\\\", or \\\"shutdown\\\"\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"script-ids\",\n\t\t\t\t\tUsage: \"A list of comma separated scripts ids that must contain all the ids of scripts of the given template and type in the desired execution order\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"delete-template-script\",\n\t\t\tUsage: \"Removes a parametrized script from a template\",\n\t\t\tAction: cmd.TemplateScriptDelete,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"template-id\",\n\t\t\t\t\tUsage: \"Template Id\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Identifier for the template-script that is parameterised by the script characterisation\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"list-template-servers\",\n\t\t\tUsage: \"Returns information about the servers that use a specific template. \",\n\t\t\tAction: cmd.TemplateServersList,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"template-id\",\n\t\t\t\t\tUsage: \"Template Id\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"add-label\",\n\t\t\tUsage: \"This action assigns a single label from a single labelable resource\",\n\t\t\tAction: cmd.LabelAdd,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Template Id\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"label\",\n\t\t\t\t\tUsage: \"Label name\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"resource-type\",\n\t\t\t\t\tUsage: \"Resource Type\",\n\t\t\t\t\tValue: \"template\",\n\t\t\t\t\tHidden: true,\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"remove-label\",\n\t\t\tUsage: \"This action unassigns a single label from a single labelable resource\",\n\t\t\tAction: cmd.LabelRemove,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Template Id\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"label\",\n\t\t\t\t\tUsage: \"Label name\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"resource-type\",\n\t\t\t\t\tUsage: \"Resource Type\",\n\t\t\t\t\tValue: \"template\",\n\t\t\t\t\tHidden: true,\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n}", "func NewTxCmd(m codec.Marshaler, txg tx.Generator, ar tx.AccountRetriever) *cobra.Command {\n\tslashingTxCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: \"Slashing transaction subcommands\",\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\tslashingTxCmd.AddCommand(NewUnjailTxCmd(m, txg, ar))\n\treturn slashingTxCmd\n}", "func NewTxCmd() *cobra.Command {\n\ttxCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: \"Token transaction subcommands\",\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\ttxCmd.AddCommand(\n\t\tGetCmdIssueToken(),\n\t\tGetCmdEditToken(),\n\t\tGetCmdMintToken(),\n\t\tGetCmdBurnToken(),\n\t\tGetCmdUnlockToken(),\n\t\tGetCmdTransferTokenOwner(),\n\t)\n\n\treturn txCmd\n}", "func NewTxCmd(coinParser chainsdk.CoinParser) *cobra.Command {\n\ttxCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: \"Bank transaction subcommands\",\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\ttxCmd.AddCommand(NewSendTxCmd(coinParser))\n\ttxCmd.AddCommand(NewMultiSend1ToManyTxCmd(coinParser))\n\n\treturn txCmd\n}", "func SubCommands() []cli.Command {\n\treturn []cli.Command{\n\t\t{\n\t\t\tName: \"list\",\n\t\t\tUsage: \"Lists all existing node pools in a cluster\",\n\t\t\tAction: cmd.NodePoolList,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Cluster Id\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"show\",\n\t\t\tUsage: \"Shows information about the node pool identified by the given id\",\n\t\t\tAction: cmd.NodePoolShow,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Node pool Id\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"create\",\n\t\t\tUsage: \"Creates a new node pool\",\n\t\t\tAction: cmd.NodePoolCreate,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Cluster Id\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"name\",\n\t\t\t\t\tUsage: \"Logical name of the node pool\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"subnet-id\",\n\t\t\t\t\tUsage: \"Identifier of the subnet where this node pool is deployed\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"node-pool-plan-id\",\n\t\t\t\t\tUsage: \"Identifier of the node pool plan that this node pool is based\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"cpu-type\",\n\t\t\t\t\tUsage: \"Type of CPU each node of the node pools will have. \" +\n\t\t\t\t\t\t\"Can be nil only if the node pool plan does not have any cpu types\",\n\t\t\t\t},\n\t\t\t\tcli.IntFlag{\n\t\t\t\t\tName: \"disk-size\",\n\t\t\t\t\tUsage: \"Size of the disk each node of the node pool will have, expressed in Gigabytes (GB)\",\n\t\t\t\t},\n\t\t\t\tcli.IntFlag{\n\t\t\t\t\tName: \"min-nodes\",\n\t\t\t\t\tUsage: \"Minimum number of nodes the node pool will have\",\n\t\t\t\t},\n\t\t\t\tcli.IntFlag{\n\t\t\t\t\tName: \"max-nodes\",\n\t\t\t\t\tUsage: \"Maximum number of nodes the node pool will have\",\n\t\t\t\t},\n\t\t\t\tcli.IntFlag{\n\t\t\t\t\tName: \"desired-nodes\",\n\t\t\t\t\tUsage: \"Amount of nodes the node pool will tend to have if the node pool does not have autoscaling\",\n\t\t\t\t},\n\t\t\t\tcli.IntFlag{\n\t\t\t\t\tName: \"pods-per-node\",\n\t\t\t\t\tUsage: \"Amount of pods each node of the node pool will have if the node pool plan supports it\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"update\",\n\t\t\tUsage: \"Updates an existing node pool identified by the given id\",\n\t\t\tAction: cmd.NodePoolUpdate,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Node pool Id\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"name\",\n\t\t\t\t\tUsage: \"Logical name of the node pool\",\n\t\t\t\t},\n\t\t\t\tcli.IntFlag{\n\t\t\t\t\tName: \"min-nodes\",\n\t\t\t\t\tUsage: \"Minimum number of nodes the node pool will have\",\n\t\t\t\t},\n\t\t\t\tcli.IntFlag{\n\t\t\t\t\tName: \"max-nodes\",\n\t\t\t\t\tUsage: \"Maximum number of nodes the node pool will have\",\n\t\t\t\t},\n\t\t\t\tcli.IntFlag{\n\t\t\t\t\tName: \"desired-nodes\",\n\t\t\t\t\tUsage: \"Amount of nodes the node pool will tend to have if the node pool does not have autoscaling\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"delete\",\n\t\t\tUsage: \"Deletes a node pool\",\n\t\t\tAction: cmd.NodePoolDelete,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Node pool Id\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"retry\",\n\t\t\tUsage: \"Retries the application of node pool identified by the given id\",\n\t\t\tAction: cmd.NodePoolRetry,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Node pool Id\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"show-plan\",\n\t\t\tUsage: \"Shows information about a specific node pool plan identified by the given id\",\n\t\t\tAction: cmd.NodePoolPlanShow,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Node pool plan Id\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n}", "func SubCommands() []cli.Command {\n\treturn []cli.Command{\n\t\t{\n\t\t\tName: \"list\",\n\t\t\tUsage: \"lists policy assignments for a given cloud account\",\n\t\t\tAction: cmd.PolicyAssignmentList,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"cloud-account-id\",\n\t\t\t\t\tUsage: \"Identifier of the cloud account\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"show\",\n\t\t\tUsage: \"Shows policy assignments\",\n\t\t\tAction: cmd.PolicyAssignmentShow,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Assignment Id\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"create\",\n\t\t\tUsage: \"Creates policy assignment\",\n\t\t\tAction: cmd.PolicyAssignmentCreate,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"name\",\n\t\t\t\t\tUsage: \"Name of the policy assignment\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"description\",\n\t\t\t\t\tUsage: \"Description of the policy assignment\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"cloud-account-id\",\n\t\t\t\t\tUsage: \"Identifier of the cloud account where the assignment belongs\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"definition-id\",\n\t\t\t\t\tUsage: \"Identifier of the policy definition to be assigned\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"parameters\",\n\t\t\t\t\tUsage: \"The parameters used to configure the policy assignment, as a json formatted parameter. \\n\\t\" +\n\t\t\t\t\t\t\"i.e: --parameters '{\\\"param1\\\":\\\"val1\\\",\\\"param2\\\":\\\"val2\\\",\\\"param3\\\":{\\\"id\\\":\\\"val3\\\"},\\\"param4\\\":true}'\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"parameters-from-file\",\n\t\t\t\t\tUsage: \"The parameters used to configure the policy assignment, from file or STDIN, \" +\n\t\t\t\t\t\t\"as a json formatted parameter. \\n\\t\" +\n\t\t\t\t\t\t\"From file: --parameters-from-file params.json \\n\\t\" +\n\t\t\t\t\t\t\"From STDIN: --parameters-from-file -\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"update\",\n\t\t\tUsage: \"Updates an existing policy assignment identified by the given id\",\n\t\t\tAction: cmd.PolicyAssignmentUpdate,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Assignment Id\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"name\",\n\t\t\t\t\tUsage: \"Name of the policy assignment\",\n\t\t\t\t},\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"description\",\n\t\t\t\t\tUsage: \"Description of the policy assignment\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"delete\",\n\t\t\tUsage: \"Deletes a policy assignment\",\n\t\t\tAction: cmd.PolicyAssignmentDelete,\n\t\t\tFlags: []cli.Flag{\n\t\t\t\tcli.StringFlag{\n\t\t\t\t\tName: \"id\",\n\t\t\t\t\tUsage: \"Assignment Id\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n}", "func (api *tenantAPI) Create(obj *cluster.Tenant) error {\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = apicl.ClusterV1().Tenant().Create(context.Background(), obj)\n\t\tif err != nil && strings.Contains(err.Error(), \"AlreadyExists\") {\n\t\t\t_, err = apicl.ClusterV1().Tenant().Update(context.Background(), obj)\n\n\t\t}\n\t\treturn err\n\t}\n\n\tapi.ct.handleTenantEvent(&kvstore.WatchEvent{Object: obj, Type: kvstore.Created})\n\treturn nil\n}", "func newCreateTerminatorCmd(f cmdutil.Factory, out io.Writer, errOut io.Writer) *cobra.Command {\n\toptions := &createTerminatorOptions{\n\t\tedgeOptions: edgeOptions{\n\t\t\tCommonOptions: common.CommonOptions{\n\t\t\t\tFactory: f,\n\t\t\t\tOut: out,\n\t\t\t\tErr: errOut,\n\t\t\t},\n\t\t},\n\t}\n\n\tcmd := &cobra.Command{\n\t\tUse: \"terminator service router address\",\n\t\tShort: \"creates a service terminator managed by the Ziti Edge Controller\",\n\t\tLong: \"creates a service terminator managed by the Ziti Edge Controller\",\n\t\tArgs: cobra.ExactArgs(3),\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\toptions.Cmd = cmd\n\t\t\toptions.Args = args\n\t\t\terr := runCreateTerminator(options)\n\t\t\tcmdhelper.CheckErr(err)\n\t\t},\n\t\tSuggestFor: []string{},\n\t}\n\n\t// allow interspersing positional args and flags\n\tcmd.Flags().SetInterspersed(true)\n\tcmd.Flags().StringVar(&options.binding, \"binding\", xgress_edge_transport.BindingName, \"Set the terminator binding\")\n\tcmd.Flags().Int32VarP(&options.cost, \"cost\", \"c\", 0, \"Set the terminator cost\")\n\tcmd.Flags().StringVarP(&options.precedence, \"precedence\", \"p\", \"\", \"Set the terminator precedence ('default', 'required' or 'failed')\")\n\tcmd.Flags().StringVar(&options.identity, \"identity\", \"\", \"Set the terminator identity\")\n\toptions.AddCommonFlags(cmd)\n\n\treturn cmd\n}", "func New() *cobra.Command {\n\tcommand := cobra.Command{\n\t\tUse: \"%template%\",\n\t\tShort: \"%template%\",\n\t\tLong: \"%template%\",\n\n\t\tArgs: cobra.NoArgs,\n\n\t\tSilenceErrors: false,\n\t\tSilenceUsage: true,\n\t}\n\n\t/* configure instance */\n\tcommand.AddCommand(\n\t\tdemo.Panic(),\n\t\tdemo.Stderr(),\n\t\tdemo.Stdout(),\n\t)\n\n\treturn &command\n}", "func CreateCommand() *cobra.Command {\n\treturn base.CreateCommand(\"template\", \"Create vt template from xml\", New())\n}", "func newDeleteCmd(clientset *client.ConfigSet) *cobra.Command {\n\tvar file string\n\tdeleteCmd := &cobra.Command{\n\t\tUse: \"delete\",\n\t\tShort: \"Delete knative resource\",\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\ts.Namespace = client.Namespace\n\t\t\tif err := s.DeleteYAML(file, args, concurrency, clientset); err != nil {\n\t\t\t\tlog.Fatal(err)\n\t\t\t}\n\t\t},\n\t}\n\n\tdeleteCmd.Flags().StringVarP(&file, \"file\", \"f\", \"serverless.yaml\", \"Delete functions defined in yaml\")\n\tdeleteCmd.Flags().IntVarP(&concurrency, \"concurrency\", \"c\", 3, \"Number of concurrent deletion threads\")\n\tdeleteCmd.AddCommand(cmdDeleteConfiguration(clientset))\n\tdeleteCmd.AddCommand(cmdDeleteRevision(clientset))\n\tdeleteCmd.AddCommand(cmdDeleteService(clientset))\n\tdeleteCmd.AddCommand(cmdDeleteRoute(clientset))\n\tdeleteCmd.AddCommand(cmdDeleteChannel(clientset))\n\tdeleteCmd.AddCommand(cmdDeleteTask(clientset))\n\tdeleteCmd.AddCommand(cmdDeleteTaskRun(clientset))\n\tdeleteCmd.AddCommand(cmdDeletePipelineResource(clientset))\n\n\treturn deleteCmd\n}", "func newTenantMutation(c config, op Op, opts ...tenantOption) *TenantMutation {\n\tm := &TenantMutation{\n\t\tconfig: c,\n\t\top: op,\n\t\ttyp: TypeTenant,\n\t\tclearedFields: make(map[string]struct{}),\n\t}\n\tfor _, opt := range opts {\n\t\topt(m)\n\t}\n\treturn m\n}", "func GetDeploymentsCommand() cli.Command {\n\tcommand := cli.Command{\n\t\tName: \"deployment\",\n\t\tUsage: \"options for deployment\",\n\t\tSubcommands: []cli.Command{\n\t\t\t{\n\t\t\t\tName: \"list\",\n\t\t\t\tUsage: \"Lists all the deployments\",\n\t\t\t\tArgsUsage: \" \",\n\t\t\t\tDescription: \"[Deprecated] List the current deployment.\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := listDeployments(c, os.Stdout)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"show\",\n\t\t\t\tUsage: \"Show deployment info\",\n\t\t\t\tArgsUsage: \" \",\n\t\t\t\tDescription: \"Show detailed information about the current deployment.\\n\" +\n\t\t\t\t\t\" Requires system administrator access,\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := showDeployment(c, os.Stdout)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"list-hosts\",\n\t\t\t\tUsage: \"Lists all ESXi hosts\",\n\t\t\t\tArgsUsage: \" \",\n\t\t\t\tDescription: \"List information about all ESXi hosts used in the deployment.\\n\" +\n\t\t\t\t\t\" For each host, the ID, the current state, the IP, and the type (MGMT and/or CLOUD)\\n\" +\n\t\t\t\t\t\" Requires system administrator access.\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := listDeploymentHosts(c, os.Stdout)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"list-vms\",\n\t\t\t\tUsage: \"Lists all VMs\",\n\t\t\t\tArgsUsage: \" \",\n\t\t\t\tDescription: \"List all VMs associated with all tenants and projects.\\n\" +\n\t\t\t\t\t\" Requires system administrator access.\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := listDeploymentVms(c, os.Stdout)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"enable-cluster-type\",\n\t\t\t\tUsage: \"Enable cluster type for deployment\",\n\t\t\t\tArgsUsage: \" \",\n\t\t\t\tDescription: \"Enable a cluster type (e.g. Kubernetes) and specify the image to be used\\n\" +\n\t\t\t\t\t\" when creating the cluster.\\n\" +\n\t\t\t\t\t\" Requires system administrator access.\",\n\t\t\t\tFlags: []cli.Flag{\n\t\t\t\t\tcli.StringFlag{\n\t\t\t\t\t\tName: \"type, k\",\n\t\t\t\t\t\tUsage: \"Cluster type (accepted values are KUBERNETES, MESOS, or SWARM)\",\n\t\t\t\t\t},\n\t\t\t\t\tcli.StringFlag{\n\t\t\t\t\t\tName: \"image-id, i\",\n\t\t\t\t\t\tUsage: \"ID of the cluster image\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := enableClusterType(c)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"disable-cluster-type\",\n\t\t\t\tUsage: \"Disable cluster type for deployment\",\n\t\t\t\tArgsUsage: \" \",\n\t\t\t\tDescription: \"Disable a cluster type (e.g. Kubernetes). Users will no longer be able\\n\" +\n\t\t\t\t\t\" to deploy clusters of that type, but existing clusters will be unaffected.\\n\" +\n\t\t\t\t\t\" Requires system administrator access.\",\n\t\t\t\tFlags: []cli.Flag{\n\t\t\t\t\tcli.StringFlag{\n\t\t\t\t\t\tName: \"type, k\",\n\t\t\t\t\t\tUsage: \"Cluster type (accepted values are KUBERNETES, MESOS, or SWARM)\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := disableClusterType(c)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"update-image-datastores\",\n\t\t\t\tUsage: \"Updates the list of image datastores\",\n\t\t\t\tArgsUsage: \" \",\n\t\t\t\tDescription: \"Update the list of allowed image datastores.\\n\" +\n\t\t\t\t\t\" Requires system administrator access.\",\n\t\t\t\tFlags: []cli.Flag{\n\t\t\t\t\tcli.StringFlag{\n\t\t\t\t\t\tName: \"datastores, d\",\n\t\t\t\t\t\tUsage: \"Comma separated name of datastore names\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := updateImageDatastores(c)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"sync-hosts-config\",\n\t\t\t\tUsage: \"Synchronizes hosts configurations\",\n\t\t\t\tArgsUsage: \" \",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := syncHostsConfig(c)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"pause\",\n\t\t\t\tUsage: \"Pause system under the deployment\",\n\t\t\t\tArgsUsage: \" \",\n\t\t\t\tDescription: \"Pause Photon Controller. All incoming requests that modify the system\\n\" +\n\t\t\t\t\t\" state (other than resume) will be refused. This implies pause-background-states\" +\n\t\t\t\t\t\" Requires system administrator access.\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := pauseSystem(c)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"pause-background-tasks\",\n\t\t\t\tUsage: \"Pause background tasks\",\n\t\t\t\tArgsUsage: \" \",\n\t\t\t\tDescription: \"Pause all background tasks in Photon Controller, such as image replication.\" +\n\t\t\t\t\t\" Incoming requests from users will continue to work\\n\" +\n\t\t\t\t\t\" Requires system administrator access.\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := pauseBackgroundTasks(c)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"resume\",\n\t\t\t\tUsage: \"Resume system under the deployment\",\n\t\t\t\tArgsUsage: \" \",\n\t\t\t\tDescription: \"Resume Photon Controller after it has been paused.\\n\" +\n\t\t\t\t\t\" Requires system administrator access.\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := resumeSystem(c)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"set-security-groups\",\n\t\t\t\tUsage: \"Set security groups for a deployment\",\n\t\t\t\tArgsUsage: \"<security-groups>\",\n\t\t\t\tDescription: \"Provide the list of Lightwave groups that contain the people who are\\n\" +\n\t\t\t\t\t\" allowed to be system administrators. Be careful: providing the wrong group could remove\\n\" +\n\t\t\t\t\t\" your access.\" +\n\t\t\t\t\t\" Requires system administrator access.\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := setDeploymentSecurityGroups(c)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"migration\",\n\t\t\t\tUsage: \"migrates state and hosts between photon controller deployments\",\n\t\t\t\tSubcommands: []cli.Command{\n\t\t\t\t\t{\n\t\t\t\t\t\tName: \"prepare\",\n\t\t\t\t\t\tUsage: \"initializes the migration\",\n\t\t\t\t\t\tFlags: []cli.Flag{\n\t\t\t\t\t\t\tcli.StringFlag{\n\t\t\t\t\t\t\t\tName: \"endpoint, e\",\n\t\t\t\t\t\t\t\tUsage: \"API endpoint of the old management plane\",\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t},\n\t\t\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\t\t\terr := deploymentMigrationPrepare(c)\n\t\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t\t{\n\t\t\t\t\t\tName: \"finalize\",\n\t\t\t\t\t\tUsage: \"finalizes the migration\",\n\t\t\t\t\t\tFlags: []cli.Flag{\n\t\t\t\t\t\t\tcli.StringFlag{\n\t\t\t\t\t\t\t\tName: \"endpoint, e\",\n\t\t\t\t\t\t\t\tUsage: \"API endpoint of the old management plane\",\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t},\n\t\t\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\t\t\terr := deploymentMigrationFinalize(c)\n\t\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t\t{\n\t\t\t\t\t\tName: \"status\",\n\t\t\t\t\t\tUsage: \"shows the status of the current migration\",\n\t\t\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\t\t\terr := showMigrationStatus(c)\n\t\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\treturn command\n}", "func TenantDelete(tenantID string) {\n\tRunCmd(fmt.Sprintf(\"%s tenant -op=del -id=%s\", ActlPath, tenantID))\n}", "func CreateTenant(tenantName string, tenantParams *TenantParams) error {\n\t_, err := handler.GetDefaultHandlerSet().GetNamespaceHandler().GetNamespace(tenantName)\n\tif err != nil {\n\t\tif adaptor.IsNotFoundErr(err) {\n\t\t\ttenantLabel := make(map[string]string, 0)\n\t\t\tfor k, v := range tenantParams.TenantLabels {\n\t\t\t\ttenantLabel[k] = v\n\t\t\t}\n\t\t\ttenantLabel[\"multi-tenant\"] = fmt.Sprintf(\"tenant-tiller-%s\", tenantName)\n\t\t\tnamespace := corev1.Namespace{\n\t\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\t\tNamespace: tenantName,\n\t\t\t\t\tName: tenantName,\n\t\t\t\t\tLabels: tenantLabel,\n\t\t\t\t\tAnnotations: tenantParams.TenantAnnotations,\n\t\t\t\t},\n\t\t\t}\n\t\t\t_, err = handler.GetDefaultHandlerSet().GetNamespaceHandler().CreateNamespace(&namespace)\n\t\t\tif err != nil {\n\t\t\t\tlogrus.Errorf(\"failed to create namespace %s : %s\", tenantName, err.Error())\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\terr = doCreateTenant(tenantName, tenantParams)\n\t\t\tif err != nil {\n\t\t\t\t// rollback\n\t\t\t\thandler.GetDefaultHandlerSet().GetNamespaceHandler().DeleteNamespace(tenantName)\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tlogrus.Infof(\"succeed to create tenant %s\", tenantName)\n\t\t\treturn nil\n\t\t}\n\t\tlogrus.Errorf(\"failed to get namespace : %s\", err.Error())\n\t\treturn err\n\n\t} else {\n\t\tlogrus.Warnf(\"namespace %s exists\", tenantName)\n\t\treturn nil\n\t}\n}", "func NewClient() *cobra.Command {\n\tcommand := cobra.Command{\n\t\tUse: \"tabloctl\",\n\t\tShort: \"the one point of view to all your task boards\",\n\t\tLong: \"The one point of view to all your task boards.\",\n\n\t\tSilenceErrors: false,\n\t\tSilenceUsage: true,\n\t}\n\t/* configure instance */\n\tcommand.AddCommand( /* related commands */ )\n\treturn &command\n}", "func newTransfer() *cobra.Command {\n\tvar cluster []string\n\tvar timeout time.Duration\n\n\tcmd := &cobra.Command{\n\t\tUse: \"transfer <id>\",\n\t\tShort: \"transfer leadership to a new node.\",\n\t\tArgs: cobra.ExactArgs(1),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tid, err := strconv.ParseUint(args[0], 10, 64)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tctx, cancel := context.WithTimeout(context.Background(), timeout)\n\t\t\tdefer cancel()\n\t\t\tif err := Transfer(ctx, &globalKeys, id, cluster); err != nil {\n\t\t\t\tfmt.Printf(\"transfer to node: %d failed: %v\\n\", id, err)\n\t\t\t\tos.Exit(1)\n\t\t\t}\n\t\t\treturn nil\n\t\t},\n\t}\n\tflags := cmd.Flags()\n\tflags.StringSliceVarP(&cluster, \"cluster\", \"c\", clusterList(), \"addresses of existing cluster nodes\")\n\tflags.DurationVarP(&timeout, \"timeout\", \"t\", time.Second*60, \"time to wait for transfer to complete\")\n\treturn cmd\n}", "func newSubListCmdForEntityType(entityType string, subType string, outputF outputFunction, options *edgeOptions) *cobra.Command {\n\tdesc := fmt.Sprintf(\"lists %v related to a %v instanced managed by the Ziti Edge Controller\", subType, entityType)\n\tcmd := &cobra.Command{\n\t\tUse: fmt.Sprintf(\"%v <id or name>\", subType),\n\t\tShort: desc,\n\t\tLong: desc,\n\t\tArgs: cobra.RangeArgs(1, 2),\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\toptions.Cmd = cmd\n\t\t\toptions.Args = args\n\t\t\terr := runListChilden(entityType, subType, options, outputF)\n\t\t\tcmdhelper.CheckErr(err)\n\t\t},\n\t\tSuggestFor: []string{},\n\t}\n\n\t// allow interspersing positional args and flags\n\tcmd.Flags().SetInterspersed(true)\n\toptions.AddCommonFlags(cmd)\n\n\treturn cmd\n}", "func newActivateCmd(clientFn func() (*fic.ServiceClient, error), out io.Writer) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"activate <id>\",\n\t\tShort: \"Activate port\",\n\t\tExample: \"fic ports activate F012000000168\",\n\t\tArgs: cobra.ExactArgs(1),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tclient, err := clientFn()\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"creating FIC client: %w\", err)\n\t\t\t}\n\n\t\t\tport, err := ports.Activate(client, args[0]).Extract()\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"calling Activate port API: %w\", err)\n\t\t\t}\n\n\t\t\tt := utils.NewTabby(out)\n\t\t\tt.AddHeader(\"id\", \"name\", \"operationStatus\", \"isActivated\", \"vlanRanges\", \"tenantID\", \"switchName\",\n\t\t\t\t\"portType\", \"location\", \"area\", \"operationID\")\n\t\t\tt.AddLine(port.ID, port.Name, port.OperationStatus, port.IsActivated, port.VLANRanges, port.TenantID, port.SwitchName,\n\t\t\t\tport.PortType, port.Location, port.Area, port.OperationID)\n\t\t\tt.Print()\n\n\t\t\treturn nil\n\t\t},\n\t}\n\n\treturn cmd\n}", "func newCommand(tb DirCleaner, opts ...server.CommandOption) *Command {\n\tpath := tb.TempDir()\n\n\t// Set aggressive close timeout by default to avoid hanging tests. This was\n\t// a problem with PDK tests which used pilosa/client as well. We put it at the\n\t// beginning of the option slice so that it can be overridden by user-passed\n\t// options.\n\topts = append([]server.CommandOption{\n\t\tserver.OptCommandCloseTimeout(time.Millisecond * 2),\n\t}, opts...)\n\n\tm := &Command{commandOptions: opts}\n\toutput := io.Discard\n\tif testing.Verbose() {\n\t\toutput = os.Stderr\n\t}\n\tm.Command = server.NewCommand(output, opts...)\n\t// pick etcd ports using a socket rather than a real port\n\terr := GetPortsGenConfigs(tb, []*Command{m})\n\tif err != nil {\n\t\ttb.Fatalf(\"generating config: %v\", err)\n\t}\n\tm.Config.DataDir = path\n\tdefaultConf := server.NewConfig()\n\n\tif m.Config.Bind == defaultConf.Bind {\n\t\tm.Config.Bind = \"http://localhost:0\"\n\t}\n\n\tif m.Config.BindGRPC == defaultConf.BindGRPC {\n\t\tm.Config.BindGRPC = \"http://localhost:0\"\n\t}\n\n\tm.Config.Translation.MapSize = 140000\n\tm.Config.WorkerPoolSize = 2\n\n\treturn m\n}", "func CreateTenants(num int) {\n\tconnStr := viper.GetString(\"database_url\")\n\tfmt.Printf(\"Conn string is %s\\n\", connStr)\n\tdb, err := sql.Open(\"postgres\", connStr)\n\tif err != nil {\n\t\tfmt.Println(\"Failed to open DB: \", err)\n\t}\n\n\tdefer db.Close()\n\n\ttablesNum := viper.GetInt(\"tables\")\n\trecordsNum := viper.GetInt(\"records\")\n\n\tfor i := 1; i <= num; i++ {\n\t\ttenantName := fmt.Sprintf(\"client%d\", i)\n\t\tif err := createTenant(tenantName, db, tablesNum, recordsNum); err != nil {\n\t\t\tfmt.Printf(\"Failed to create client %d: %s\\n\", i, err)\n\t\t\treturn\n\t\t}\n\t\tfmt.Printf(\"Tenant %s created\\n\", tenantName)\n\t}\n}", "func newCreateCmd(clientFn func() (*fic.ServiceClient, error), out io.Writer) *cobra.Command {\n\tvar (\n\t\tsrcRouterID string\n\t\tsrcGroupName string\n\t\tsrcPrimary string\n\t\tsrcSecondary string\n\t\tsrcRouteFilter string\n\t\tdestPrimary string\n\t\tdestSecondary string\n\t\tbandwidth string\n\t)\n\n\tr := regexp.MustCompile(`^[\\w&()-]{1,64}$`)\n\tvalidateSrc := func(splitSrc []string, isPrimary, hasSecondary bool) error {\n\t\t_, ipNet, err := net.ParseCIDR(splitSrc[0])\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"ipAddress must be CIDR whose subnet mask is 30, e.g. 10.0.0.1/30: received %s\", splitSrc[0])\n\t\t}\n\t\tsubNetMaskLength, _ := ipNet.Mask.Size()\n\t\tif subNetMaskLength != 30 {\n\t\t\treturn fmt.Errorf(\"subnet mask of ipAddress must be 30, e.g. 10.0.0.1/30: received %s\", splitSrc[0])\n\t\t}\n\n\t\tif !utils.StringInSlice(splitSrc[1], validPrepends) {\n\t\t\treturn fmt.Errorf(\"asPathPrepend.in must be one of %s: received %s\", validPrepends, splitSrc[1])\n\t\t}\n\n\t\tif !utils.StringInSlice(splitSrc[2], validPrepends) {\n\t\t\treturn fmt.Errorf(\"asPathPrepend.out must be one of %s: received %s\", validPrepends, splitSrc[2])\n\t\t}\n\n\t\tif isPrimary {\n\t\t\tmed, err := strconv.Atoi(splitSrc[3])\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"med.out must be numeric value: received %s\", splitSrc[3])\n\t\t\t}\n\n\t\t\tif hasSecondary {\n\t\t\t\tif !utils.IntInSlice(med, validPrimaryPairedMEDs) {\n\t\t\t\t\treturn fmt.Errorf(\"med.out in paired connection must be one of %v: received %d\", validPrimaryPairedMEDs, med)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tif !utils.IntInSlice(med, validPrimarySingleMEDs) {\n\t\t\t\t\treturn fmt.Errorf(\"med.out in single connection must be one of %v: received %d\", validPrimarySingleMEDs, med)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\treturn nil\n\t}\n\tvalidateDest := func(splitDest []string) error {\n\t\tvlan, err := strconv.Atoi(splitDest[1])\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"vlan must be numeric value: received %s\", splitDest[1])\n\t\t}\n\t\tif vlan < 101 || vlan > 3300 {\n\t\t\treturn fmt.Errorf(\"vlan must be range of 101 to 3300: received %s\", splitDest[1])\n\t\t}\n\n\t\t_, ipNet, err := net.ParseCIDR(splitDest[2])\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"ipAddress must be CIDR whose subnet mask is 30, e.g. 10.0.0.2/30: received %s\", splitDest[2])\n\t\t}\n\t\tsubNetMaskLength, _ := ipNet.Mask.Size()\n\t\tif subNetMaskLength != 30 {\n\t\t\treturn fmt.Errorf(\"subnet mask of ipAddress must be 30, e.g. 10.0.0.2/30: received %s\", splitDest[2])\n\t\t}\n\n\t\tasn, err := strconv.Atoi(splitDest[3])\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"asn must be numeric value: received %s\", splitDest[3])\n\t\t}\n\t\tif asn < 1 || asn > 65535 {\n\t\t\treturn fmt.Errorf(\"asn must be range of 1 to 65535: received %s\", splitDest[3])\n\t\t}\n\t\treturn nil\n\t}\n\n\tcmd := &cobra.Command{\n\t\tUse: \"create <name>\",\n\t\tShort: \"Create router to port connection\",\n\t\tExample: \"# In case of non paired-connection \\n\" +\n\t\t\t\"fic router-to-port-connections create YourConnectionName \" +\n\t\t\t\"--source-router F020123456789 \" +\n\t\t\t\"--source-group group_1 \" +\n\t\t\t\"--source-primary 10.0.0.1/30,4,4,10 \" +\n\t\t\t\"--source-route-filter fullRoute,fullRouteWithDefaultRoute \" +\n\t\t\t\"--destination-primary F010123456789,101,10.0.0.2/30,65000 \" +\n\t\t\t\"--bandwidth 10M \\n\\n\" +\n\t\t\t\"# In case of paired-connection \\n\" +\n\t\t\t\"fic router-to-port-connections create YourConnectionName \" +\n\t\t\t\"--source-router F020123456789 \" +\n\t\t\t\"--source-group group_1 \" +\n\t\t\t\"--source-primary 10.0.0.1/30,4,4,10 \" +\n\t\t\t\"--source-secondary 10.0.0.5/30,2,1 \" +\n\t\t\t\"--source-route-filter fullRoute,fullRouteWithDefaultRoute \" +\n\t\t\t\"--destination-primary F010123456789,101,10.0.0.2/30,65000 \" +\n\t\t\t\"--destination-secondary F019876543210,102,10.0.0.6/30,65000 \" +\n\t\t\t\"--bandwidth 10M\",\n\t\tArgs: cobra.ExactArgs(1),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\n\t\t\tif !r.MatchString(args[0]) {\n\t\t\t\treturn fmt.Errorf(\"name of router to port connection must be composed of alpha-numeric \"+\n\t\t\t\t\t\"characters and & ( ) - _, and must have maximum length of 64 as well: received %s\", args[0])\n\t\t\t}\n\n\t\t\tif !utils.StringInSlice(srcGroupName, validGroupNames) {\n\t\t\t\treturn fmt.Errorf(\"source-group must be one of %s: received %s\", validGroupNames, srcGroupName)\n\t\t\t}\n\n\t\t\tsplitSrcPrimary := strings.Split(strings.TrimSpace(srcPrimary), \",\")\n\t\t\tsplitSrcSecondary := strings.Split(strings.TrimSpace(srcSecondary), \",\")\n\t\t\tsplitSrcRouteFilter := strings.Split(strings.TrimSpace(srcRouteFilter), \",\")\n\t\t\tsplitDestPrimary := strings.Split(strings.TrimSpace(destPrimary), \",\")\n\t\t\tsplitDestSecondary := strings.Split(strings.TrimSpace(destSecondary), \",\")\n\n\t\t\tif len(splitSrcPrimary) != 4 {\n\t\t\t\treturn fmt.Errorf(\"source-primary must have format like \"+\n\t\t\t\t\t\"<ipAddress>,<asPathPrepend.in>,<asPathPrepend.out>,<med.out>: received %s\", srcPrimary)\n\t\t\t}\n\t\t\tif err := validateSrc(splitSrcPrimary, true, srcSecondary != \"\"); err != nil {\n\t\t\t\treturn fmt.Errorf(\"in source-primary, %w\", err)\n\t\t\t}\n\n\t\t\tif len(splitDestPrimary) != 4 {\n\t\t\t\treturn fmt.Errorf(\"destination-primary must have format like \"+\n\t\t\t\t\t\"<portId>,<vlan>,<ipAddress>,<asn>: received %s\", destPrimary)\n\t\t\t}\n\t\t\tif err := validateDest(splitDestPrimary); err != nil {\n\t\t\t\treturn fmt.Errorf(\"in destination-primary, %w\", err)\n\t\t\t}\n\n\t\t\tif srcSecondary != \"\" {\n\t\t\t\tif len(splitSrcSecondary) != 3 {\n\t\t\t\t\treturn fmt.Errorf(\"source-secondary must have format like \"+\n\t\t\t\t\t\t\"<ipAddress>,<asPathPrepend.in>,<asPathPrepend.out>: received %s\", srcSecondary)\n\t\t\t\t}\n\t\t\t\tif err := validateSrc(splitSrcSecondary, false, true); err != nil {\n\t\t\t\t\treturn fmt.Errorf(\"in source-secondary, %w\", err)\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tredundant := false\n\t\t\tif destSecondary != \"\" {\n\t\t\t\tif len(splitDestSecondary) != 4 {\n\t\t\t\t\treturn fmt.Errorf(\"destination-secondary must have format like \"+\n\t\t\t\t\t\t\"<portId>,<vlan>,<ipAddress>,<asn>: received %s\", destSecondary)\n\t\t\t\t}\n\t\t\t\tif err := validateDest(splitDestSecondary); err != nil {\n\t\t\t\t\treturn fmt.Errorf(\"in destination-secondary, %w\", err)\n\t\t\t\t}\n\t\t\t\tredundant = true\n\t\t\t}\n\n\t\t\tif len(splitSrcRouteFilter) != 2 {\n\t\t\t\treturn fmt.Errorf(\"source-route-filter must have format like \"+\n\t\t\t\t\t\"<routeFilter.in>,<routeFilter.out>: received %s\", srcRouteFilter)\n\t\t\t}\n\t\t\tif !utils.StringInSlice(splitSrcRouteFilter[0], validRouteFilterIns) {\n\t\t\t\treturn fmt.Errorf(\"routeFilter.in must be one of %s: received %s\", validRouteFilterIns, splitSrcRouteFilter[0])\n\t\t\t}\n\t\t\tif !utils.StringInSlice(splitSrcRouteFilter[1], validRouteFilterOuts) {\n\t\t\t\treturn fmt.Errorf(\"routeFilter.out must be one of %s: received %s\", validRouteFilterOuts, splitSrcRouteFilter[1])\n\t\t\t}\n\n\t\t\tif !utils.StringInSlice(bandwidth, validBandwidths) {\n\t\t\t\treturn fmt.Errorf(\"bandwidth must be one of %s: received %s\", validBandwidths, bandwidth)\n\t\t\t}\n\n\t\t\tclient, err := clientFn()\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"creating FIC client: %w\", err)\n\t\t\t}\n\n\t\t\tprependIn, prependOut := interface{}(convPrepend(splitSrcPrimary[1])), interface{}(convPrepend(splitSrcPrimary[2]))\n\t\t\tmed, _ := strconv.Atoi(splitSrcPrimary[3])\n\t\t\tvlan, _ := strconv.Atoi(splitDestPrimary[1])\n\n\t\t\tt := utils.NewTabby(out)\n\t\t\tt.AddHeader(\"id\", \"name\", \"redundant\", \"tenantId\", \"area\", \"operationStatus\", \"bandwidth\", \"operationId\")\n\n\t\t\tif !redundant {\n\t\t\t\topts := singleConn.CreateOpts{\n\t\t\t\t\tName: args[0],\n\t\t\t\t\tSource: singleConn.Source{\n\t\t\t\t\t\tRouterID: srcRouterID,\n\t\t\t\t\t\tGroupName: srcGroupName,\n\t\t\t\t\t\tRouteFilter: singleConn.RouteFilter{\n\t\t\t\t\t\t\tIn: splitSrcRouteFilter[0],\n\t\t\t\t\t\t\tOut: splitSrcRouteFilter[1],\n\t\t\t\t\t\t},\n\t\t\t\t\t\tPrimary: singleConn.SourceHAInfo{\n\t\t\t\t\t\t\tIPAddress: splitSrcPrimary[0],\n\t\t\t\t\t\t\tASPathPrepend: singleConn.ASPathPrepend{\n\t\t\t\t\t\t\t\tIn: &prependIn,\n\t\t\t\t\t\t\t\tOut: &prependOut,\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\tMED: &singleConn.MED{\n\t\t\t\t\t\t\t\tOut: med,\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t\tDestination: singleConn.Destination{\n\t\t\t\t\t\tPrimary: singleConn.DestinationHAInfo{\n\t\t\t\t\t\t\tPortID: splitDestPrimary[0],\n\t\t\t\t\t\t\tVLAN: vlan,\n\t\t\t\t\t\t\tIPAddress: splitDestPrimary[2],\n\t\t\t\t\t\t\tASN: splitDestPrimary[3],\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t\tBandwidth: bandwidth,\n\t\t\t\t}\n\n\t\t\t\tc, err := singleConn.Create(client, opts).Extract()\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn fmt.Errorf(\"calling Create router to port connection API: %w\", err)\n\t\t\t\t}\n\n\t\t\t\tt.AddLine(c.ID, c.Name, c.Redundant, c.TenantID, c.Area, c.OperationID, c.OperationStatus, c.Bandwidth)\n\t\t\t\tt.Print()\n\n\t\t\t\treturn nil\n\t\t\t}\n\n\t\t\tsecondPrependIn, secondPrependOut := interface{}(convPrepend(splitSrcSecondary[1])), interface{}(convPrepend(splitSrcSecondary[2]))\n\t\t\tsecondVlan, _ := strconv.Atoi(splitDestSecondary[1])\n\n\t\t\topts := pairedConn.CreateOpts{\n\t\t\t\tName: args[0],\n\t\t\t\tSource: pairedConn.Source{\n\t\t\t\t\tRouterID: srcRouterID,\n\t\t\t\t\tGroupName: srcGroupName,\n\t\t\t\t\tRouteFilter: pairedConn.RouteFilter{\n\t\t\t\t\t\tIn: splitSrcRouteFilter[0],\n\t\t\t\t\t\tOut: splitSrcRouteFilter[1],\n\t\t\t\t\t},\n\t\t\t\t\tPrimary: pairedConn.SourceHAInfo{\n\t\t\t\t\t\tIPAddress: splitSrcPrimary[0],\n\t\t\t\t\t\tASPathPrepend: pairedConn.ASPathPrepend{\n\t\t\t\t\t\t\tIn: &prependIn,\n\t\t\t\t\t\t\tOut: &prependOut,\n\t\t\t\t\t\t},\n\t\t\t\t\t\tMED: &pairedConn.MED{\n\t\t\t\t\t\t\tOut: med,\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t\tSecondary: pairedConn.SourceHAInfo{\n\t\t\t\t\t\tIPAddress: splitSrcSecondary[0],\n\t\t\t\t\t\tASPathPrepend: pairedConn.ASPathPrepend{\n\t\t\t\t\t\t\tIn: &secondPrependIn,\n\t\t\t\t\t\t\tOut: &secondPrependOut,\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tDestination: pairedConn.Destination{\n\t\t\t\t\tPrimary: pairedConn.DestinationHAInfo{\n\t\t\t\t\t\tPortID: splitDestPrimary[0],\n\t\t\t\t\t\tVLAN: vlan,\n\t\t\t\t\t\tIPAddress: splitDestPrimary[2],\n\t\t\t\t\t\tASN: splitDestPrimary[3],\n\t\t\t\t\t},\n\t\t\t\t\tSecondary: pairedConn.DestinationHAInfo{\n\t\t\t\t\t\tPortID: splitDestSecondary[0],\n\t\t\t\t\t\tVLAN: secondVlan,\n\t\t\t\t\t\tIPAddress: splitDestSecondary[2],\n\t\t\t\t\t\tASN: splitDestSecondary[3],\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tBandwidth: bandwidth,\n\t\t\t}\n\n\t\t\tc, err := pairedConn.Create(client, opts).Extract()\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"calling Create router to port connection API: %w\", err)\n\t\t\t}\n\n\t\t\tt.AddLine(c.ID, c.Name, c.Redundant, c.TenantID, c.Area, c.OperationID, c.OperationStatus, c.Bandwidth)\n\t\t\tt.Print()\n\n\t\t\treturn nil\n\t\t},\n\t}\n\n\tcmd.Flags().StringVar(&srcRouterID, \"source-router\", \"\", \"(Required) Router ID belonging to source\")\n\tcmd.Flags().StringVar(&srcGroupName, \"source-group\", \"\", \"(Required) Group Name belonging to source\")\n\tcmd.Flags().StringVar(\n\t\t&srcPrimary,\n\t\t\"source-primary\",\n\t\t\"\",\n\t\t\"(Required) Source Primary Info specified in the format <ipAddress>,<asPathPrepend.in>,<asPathPrepend.out>,<med.out>\")\n\tcmd.Flags().StringVar(\n\t\t&srcSecondary,\n\t\t\"source-secondary\",\n\t\t\"\",\n\t\t\"Source Secondary Info specified in the format <ipAddress>,<asPathPrepend.in>,<asPathPrepend.out>\")\n\tcmd.Flags().StringVar(\n\t\t&srcRouteFilter,\n\t\t\"source-route-filter\",\n\t\t\"\",\n\t\t\"(Required) Set of BGP Filter Ingress and Egress specified in the format <routeFilter.in>,<routeFilter.out>\")\n\tcmd.Flags().StringVar(\n\t\t&destPrimary,\n\t\t\"destination-primary\",\n\t\t\"\",\n\t\t\"(Required) Destination Primary Info specified in the format <portId>,<vlan>,<ipAddress>,<asn>\")\n\tcmd.Flags().StringVar(\n\t\t&destSecondary,\n\t\t\"destination-secondary\",\n\t\t\"\",\n\t\t\"Destination Secondary Info specified in the format <portId>,<vlan>,<ipAddress>,<asn>\")\n\tcmd.Flags().StringVar(&bandwidth, \"bandwidth\", \"\", \"(Required) Bandwidth of router to port connection\")\n\n\tcmd.MarkFlagRequired(\"source-router\")\n\tcmd.MarkFlagRequired(\"source-group\")\n\tcmd.MarkFlagRequired(\"source-primary\")\n\tcmd.MarkFlagRequired(\"source-rout-filter\")\n\tcmd.MarkFlagRequired(\"destination-primary\")\n\tcmd.MarkFlagRequired(\"bandwidth\")\n\n\treturn cmd\n}", "func newCreateCmd(pather command.Pather) *cobra.Command {\n\tnow := time.Now().UTC()\n\tvar flags struct {\n\t\tcsr bool\n\t\tprofile string\n\t\tcommonName string\n\t\tnotBefore flag.Time\n\t\tnotAfter flag.Time\n\t\tca string\n\t\tcaKey string\n\t\texistingKey string\n\t\tcurve string\n\t\tbundle bool\n\t\tforce bool\n\t}\n\tflags.notBefore = flag.Time{\n\t\tTime: now,\n\t\tCurrent: now,\n\t}\n\tflags.notAfter = flag.Time{\n\t\tCurrent: now,\n\t\tDefault: \"depends on profile\",\n\t}\n\n\tvar cmd = &cobra.Command{\n\t\tUse: \"create [flags] <subject-template> <cert-file> <key-file>\",\n\t\tShort: \"Create a certificate or certificate signing request\",\n\t\tExample: fmt.Sprintf(` %[1]s create --profile cp-root subject.tmpl cp-root.crt cp-root.key\n %[1]s create --ca cp-ca.crt --ca-key cp-ca.key subject.tmpl chain.pem cp-as.key\n %[1]s create --csr subject.tmpl chain.csr cp-as.key`,\n\t\t\tpather.CommandPath(),\n\t\t),\n\t\tLong: `'create' generates a certificate or a certificate signing request (CSR).\n\nThe command takes the following positional arguments:\n- <subject-template> is the template for the certificate subject distinguished name.\n- <crt-file> is the file path where the certificate or certificate requests is\n written to. The parent directory must exist and must be writable.\n- <key-file> is the file path where the fresh private key is written to. The\n parent directory must exist and must be writable.\n\nBy default, the command creates a SCION control-plane PKI AS certificate. Another\ncertificate type can be selected by providing the --profile flag. If a certificate\nchain is desired, specify the --bundle flag.\n\nA fresh key is created in the provided <key-file>, unless the --key flag is set.\nIf the --key flag is set, an existing private key is used and the <key-file> is\nignored.\n\nThe --ca and --ca-key flags are required if a AS certificate or CA certificate\nis being created. Otherwise, they are not allowed.\n\nThe --not-before and --not-after flags can either be a timestamp or a relative\ntime offset from the current time.\n\nA timestamp can be provided in two different formats: unix timestamp and\nRFC 3339 timestamp. For example, 2021-06-24T12:01:02Z represents 1 minute and 2\nseconds after the 12th hour of June 26th, 2021 in UTC.\n\nThe relative time offset can be formated as a time duration string with the\nfollowing units: y, w, d, h, m, s. Negative offsets are also allowed. For\nexample, -1h indicates the time of tool invocation minus one hour. Note that\n--not-after is relative to the current time if a relative time offset is used,\nand not to --not-before.\n\nThe <subject-template> is the template for the distinguished name of the\nrequested certificate and must either be a x.509 certificate or a JSON file.\nThe common name can be overridden by supplying the --common-name flag.\n\nIf it is a x.509 certificate, the subject of the template is used as the subject\nof the created certificate or certificate chain request.\n\nA valid example for a JSON formatted template:\n` + subjectHelp,\n\t\tArgs: cobra.RangeArgs(2, 3),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tif len(args) == 2 && flags.existingKey == \"\" {\n\t\t\t\treturn serrors.New(\"positional key file is required\")\n\t\t\t}\n\t\t\tct, err := parseCertType(flags.profile)\n\t\t\tif err != nil {\n\t\t\t\treturn serrors.WrapStr(\"parsing profile\", err)\n\t\t\t}\n\t\t\tsubject, err := createSubject(args[0], flags.commonName)\n\t\t\tif err != nil {\n\t\t\t\treturn serrors.WrapStr(\"creating subject\", err)\n\t\t\t}\n\n\t\t\t// Only check that the flags are set appropriately here.\n\t\t\t// Do the actual parsing after the usage help message is silenced.\n\t\t\tvar loadCA bool\n\t\t\tisSelfSigned := (ct == cppki.Root || ct == cppki.Regular || ct == cppki.Sensitive)\n\t\t\twithCA := (flags.ca != \"\" || flags.caKey != \"\")\n\t\t\tswitch {\n\t\t\tcase flags.csr && withCA:\n\t\t\t\treturn serrors.New(\"CA information set for CSR\")\n\t\t\tcase !flags.csr && isSelfSigned && withCA:\n\t\t\t\treturn serrors.New(\"CA information set for self-signed certificate\")\n\t\t\tdefault:\n\t\t\t\tloadCA = !isSelfSigned && !flags.csr\n\t\t\t}\n\n\t\t\tcmd.SilenceUsage = true\n\n\t\t\tvar privKey key.PrivateKey\n\t\t\tvar encodedKey []byte\n\t\t\tif flags.existingKey != \"\" {\n\t\t\t\tif privKey, err = key.LoadPrivateKey(flags.existingKey); err != nil {\n\t\t\t\t\treturn serrors.WrapStr(\"loading existing private key\", err)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tif privKey, err = key.GeneratePrivateKey(flags.curve); err != nil {\n\t\t\t\t\treturn serrors.WrapStr(\"creating fresh private key\", err)\n\t\t\t\t}\n\t\t\t\tif encodedKey, err = key.EncodePEMPrivateKey(privKey); err != nil {\n\t\t\t\t\treturn serrors.WrapStr(\"encoding fresh private key\", err)\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tvar caCertRaw []byte\n\t\t\tvar caCert *x509.Certificate\n\t\t\tvar caKey key.PrivateKey\n\t\t\tif loadCA {\n\t\t\t\tif caCertRaw, err = os.ReadFile(flags.ca); err != nil {\n\t\t\t\t\treturn serrors.WrapStr(\"read CA certificate\", err)\n\t\t\t\t}\n\t\t\t\tif caCert, err = parseCertificate(caCertRaw); err != nil {\n\t\t\t\t\treturn serrors.WrapStr(\"parsing CA certificate\", err)\n\t\t\t\t}\n\t\t\t\tif caKey, err = key.LoadPrivateKey(flags.caKey); err != nil {\n\t\t\t\t\treturn serrors.WrapStr(\"loading CA private key\", err)\n\t\t\t\t}\n\t\t\t}\n\t\t\tif isSelfSigned {\n\t\t\t\tcaKey = privKey\n\t\t\t}\n\n\t\t\tif flags.csr {\n\t\t\t\tcsr, err := CreateCSR(ct, subject, privKey)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn serrors.WrapStr(\"creating CSR\", err)\n\t\t\t\t}\n\t\t\t\tencodedCSR := pem.EncodeToMemory(&pem.Block{\n\t\t\t\t\tType: \"CERTIFICATE REQUEST\",\n\t\t\t\t\tBytes: csr,\n\t\t\t\t})\n\t\t\t\tif encodedCSR == nil {\n\t\t\t\t\tpanic(\"failed to encode CSR\")\n\t\t\t\t}\n\t\t\t\tcsrFile := args[1]\n\t\t\t\terr = file.WriteFile(csrFile, encodedCSR, 0644, file.WithForce(flags.force))\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn serrors.WrapStr(\"writing CSR\", err)\n\t\t\t\t}\n\t\t\t\tfmt.Printf(\"CSR successfully written to %q\\n\", csrFile)\n\t\t\t} else {\n\t\t\t\tcert, err := CreateCertificate(CertParams{\n\t\t\t\t\tType: ct,\n\t\t\t\t\tSubject: subject,\n\t\t\t\t\tPubKey: privKey.Public(),\n\t\t\t\t\tNotBefore: flags.notBefore.Time,\n\t\t\t\t\tNotAfter: notAfterFromFlags(ct, flags.notBefore, flags.notAfter),\n\t\t\t\t\tCAKey: caKey,\n\t\t\t\t\tCACert: caCert,\n\t\t\t\t})\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn serrors.WrapStr(\"creating certificate\", err)\n\t\t\t\t}\n\t\t\t\tencodedCert := pem.EncodeToMemory(&pem.Block{\n\t\t\t\t\tType: \"CERTIFICATE\",\n\t\t\t\t\tBytes: cert,\n\t\t\t\t})\n\t\t\t\tif encodedCert == nil {\n\t\t\t\t\tpanic(\"failed to encode CSR\")\n\t\t\t\t}\n\t\t\t\tif flags.bundle {\n\t\t\t\t\tfmt.Println(\"Bundling certificate as certificate chain\")\n\t\t\t\t\tencodedCert = append(encodedCert, caCertRaw...)\n\t\t\t\t}\n\t\t\t\tcertFile := args[1]\n\t\t\t\terr = file.WriteFile(certFile, encodedCert, 0644, file.WithForce(flags.force))\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn serrors.WrapStr(\"writing certificate\", err)\n\t\t\t\t}\n\t\t\t\tfmt.Printf(\"Certificate successfully written to %q\\n\", certFile)\n\t\t\t}\n\n\t\t\tif encodedKey != nil {\n\t\t\t\tkeyFile := args[2]\n\t\t\t\tif err := file.CheckDirExists(filepath.Dir(keyFile)); err != nil {\n\t\t\t\t\treturn serrors.WrapStr(\"checking that directory of private key exists\", err)\n\t\t\t\t}\n\t\t\t\terr := file.WriteFile(keyFile, encodedKey, 0600, file.WithForce(flags.force))\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn serrors.WrapStr(\"writing private key\", err)\n\t\t\t\t}\n\t\t\t\tfmt.Printf(\"Private key successfully written to %q\\n\", keyFile)\n\t\t\t}\n\t\t\treturn nil\n\t\t},\n\t}\n\n\tcmd.Flags().BoolVar(&flags.csr, \"csr\", false,\n\t\t\"Generate a certificate signign request instead of a certificate\",\n\t)\n\tcmd.Flags().StringVar(&flags.profile, \"profile\", \"cp-as\",\n\t\t\"The type of certificate to generate (cp-as|cp-ca|cp-root|sensitive-voting|regular-voting)\",\n\t)\n\tcmd.Flags().Var(&flags.notBefore, \"not-before\",\n\t\t`The NotBefore time of the certificate. Can either be a timestamp or an offset.\n\nIf the value is a timestamp, it is expected to either be an RFC 3339 formatted\ntimestamp or a unix timestamp. If the value is a duration, it is used as the\noffset from the current time.`,\n\t)\n\tcmd.Flags().Var(&flags.notAfter, \"not-after\",\n\t\t`The NotAfter time of the certificate. Can either be a timestamp or an offset.\n\nIf the value is a timestamp, it is expected to either be an RFC 3339 formatted\ntimestamp or a unix timestamp. If the value is a duration, it is used as the\noffset from the current time.`,\n\t)\n\tcmd.Flags().StringVar(&flags.commonName, \"common-name\", \"\",\n\t\t\"The common name that replaces the common name in the subject template\",\n\t)\n\tcmd.Flags().StringVar(&flags.ca, \"ca\", \"\",\n\t\t\"The path to the issuer certificate\",\n\t)\n\tcmd.Flags().StringVar(&flags.caKey, \"ca-key\", \"\",\n\t\t\"The path to the issuer private key used to sign the new certificate\",\n\t)\n\tcmd.Flags().StringVar(&flags.existingKey, \"key\", \"\",\n\t\t\"The path to the existing private key to use instead of creating a new one\",\n\t)\n\tcmd.Flags().StringVar(&flags.curve, \"curve\", \"P-256\",\n\t\t\"The elliptic curve to use (P-256|P-384|P-521)\",\n\t)\n\tcmd.Flags().BoolVar(&flags.bundle, \"bundle\", false,\n\t\t\"Bundle the certificate with the issuer certificate as a certificate chain\",\n\t)\n\tcmd.Flags().BoolVar(&flags.force, \"force\", false,\n\t\t\"Force overwritting existing files\",\n\t)\n\n\treturn cmd\n}", "func New() *cobra.Command {\n\t/*\n\t\t\tconst (\n\t\t\t\tlong = `flyctl is a command line interface to the Fly.io platform.\n\n\t\tIt allows users to manage authentication, application launch,\n\t\tdeployment, network configuration, logging and more with just the\n\t\tone command.\n\n\t\tLaunch an app with the launch command\n\t\tDeploy an app with the deploy command\n\t\tView a deployed web application with the open command\n\t\tCheck the status of an application with the status command\n\n\t\tTo read more, use the docs command to view Fly's help on the web.\n\t\t`\n\t\t\t\tshort = \"The Fly CLI\"\n\t\t\t\tusage = \"flyctl\"\n\t\t\t)\n\n\t\t\troot := command.New(usage, short, long, nil)\n\t\t\troot.SilenceUsage = true\n\t\t\troot.SilenceErrors = true\n\n\t\t\tfs := root.PersistentFlags()\n\n\t\t\t_ = fs.StringP(flag.AccessTokenName, \"t\", \"\", \"Fly API Access Token\")\n\t\t\t_ = fs.BoolP(flag.JSONOutputName, \"j\", false, \"JSON output\")\n\t\t\t_ = fs.BoolP(flag.VerboseName, \"v\", false, \"Verbose output\")\n\n\t\t\troot.AddCommand(\n\t\t\t\tversion.New(),\n\t\t\t\tapps.New(),\n\t\t\t\tcreate.New(), // TODO: deprecate\n\t\t\t\tdestroy.New(), // TODO: deprecate\n\t\t\t\tmove.New(), // TODO: deprecate\n\t\t\t\tsuspend.New(), // TODO: deprecate\n\t\t\t\tresume.New(), // TODO: deprecate\n\t\t\t\trestart.New(), // TODO: deprecate\n\t\t\t\torgs.New(),\n\t\t\t\tauth.New(),\n\t\t\t\tbuilds.New(),\n\t\t\t\topen.New(), // TODO: deprecate\n\t\t\t\tcurl.New(),\n\t\t\t\tplatform.New(),\n\t\t\t\tdocs.New(),\n\t\t\t\treleases.New(),\n\t\t\t\tdeploy.New(),\n\t\t\t\thistory.New(),\n\t\t\t\tstatus.New(),\n\t\t\t\tlogs.New(),\n\t\t\t\tdoctor.New(),\n\t\t\t\tdig.New(),\n\t\t\t\tvolumes.New(),\n\t\t\t\tagent.New(),\n\t\t\t)\n\n\t\t\tif os.Getenv(\"DEV\") != \"\" {\n\t\t\t\troot.AddCommand(services.New())\n\t\t\t}\n\n\t\t\treturn root\n\t*/\n\n\tflyctl.InitConfig()\n\n\t// what follows is a hack in order to achieve compatibility with what exists\n\t// already. the commented out code above, is what should remain after the\n\t// migration is complete.\n\n\t// newCommands is the set of commands which work with the new way\n\tnewCommands := []*cobra.Command{\n\t\tversion.New(),\n\t\tapps.New(),\n\t\tcreate.New(), // TODO: deprecate\n\t\tdestroy.New(), // TODO: deprecate\n\t\tmove.New(), // TODO: deprecate\n\t\tsuspend.New(), // TODO: deprecate\n\t\tresume.New(), // TODO: deprecate\n\t\trestart.New(), // TODO: deprecate\n\t\torgs.New(),\n\t\tauth.New(),\n\t\tbuilds.New(),\n\t\topen.New(), // TODO: deprecate\n\t\tcurl.New(),\n\t\tplatform.New(),\n\t\tdocs.New(),\n\t\treleases.New(),\n\t\tdeploy.New(),\n\t\thistory.New(),\n\t\tstatus.New(),\n\t\tlogs.New(),\n\t\tdoctor.New(),\n\t\tdig.New(),\n\t\tvolumes.New(),\n\t\tagent.New(),\n\t\timage.New(),\n\t\tping.New(),\n\t\tproxy.New(),\n\t\tmachine.New(),\n\t}\n\n\tif os.Getenv(\"DEV\") != \"\" {\n\t\tnewCommands = append(newCommands, services.New())\n\t}\n\n\t// newCommandNames is the set of the names of the above commands\n\tnewCommandNames := make(map[string]struct{}, len(newCommands))\n\tfor _, cmd := range newCommands {\n\t\tnewCommandNames[cmd.Name()] = struct{}{}\n\t}\n\n\t// instead of root being constructed like in the commented out snippet, we\n\t// rebuild it the old way.\n\troot := cmd.NewRootCmd(client.New())\n\n\t// gather the slice of commands which must be replaced with their new\n\t// iterations\n\tvar commandsToReplace []*cobra.Command\n\tfor _, cmd := range root.Commands() {\n\t\tif _, exists := newCommandNames[cmd.Name()]; exists {\n\t\t\tcommandsToReplace = append(commandsToReplace, cmd)\n\t\t}\n\t}\n\n\t// remove them\n\troot.RemoveCommand(commandsToReplace...)\n\n\t// make sure the remaining old commands run the preparers\n\t// TODO: remove when migration is done\n\twrapRunE(root)\n\n\t// and finally, add the new commands\n\troot.AddCommand(newCommands...)\n\n\treturn root\n}", "func TestCreateTenant(t *testing.T) {\n\tctx, cancelFunc := context.WithTimeout(context.Background(), standardTimeout)\n\tdefer cancelFunc()\n\n\tconfig := bat.TenantConfig{\n\t\tName: \"CreateTenantTest\",\n\t\tSubnetBits: 20,\n\t}\n\n\ttenant, err := bat.CreateTenant(ctx, config)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to create tenant : %v\", err)\n\t}\n\n\tdefer func() {\n\t\terr = bat.DeleteTenant(ctx, tenant.ID)\n\t\tif err != nil {\n\t\t\tt.Fatalf(\"Failed to delete tenant: %v\", err)\n\t\t}\n\t}()\n\n\tif tenant.Name != config.Name {\n\t\tt.Fatalf(\"Failed to create tenant\")\n\t}\n\n\ttenants, err := bat.GetAllTenants(ctx)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to retrieve tenant list : %v\", err)\n\t}\n\n\tfor _, tt := range tenants {\n\t\tif tt.Name == config.Name {\n\t\t\treturn\n\t\t}\n\t}\n\n\tt.Fatal(\"did not find new tenant in tenants list\")\n}", "func CreateCmd() *cobra.Command {\n\tvar cmd = cobra.Command{\n\t\tUse: \"ch.viac\",\n\t\tShort: \"Import VIAC values from JSON files\",\n\t\tLong: `Open app.viac.ch, choose a portfolio, and select \"From start\" in the overview dash. In the Chrome dev tools, save the response from the \"performance\" XHR call, and pass the resulting file to this importer.`,\n\n\t\tArgs: cobra.ExactValidArgs(1),\n\n\t\tRunE: run,\n\t}\n\tcmd.Flags().StringP(\"from\", \"f\", \"0001-01-01\", \"YYYY-MM-DD - ignore entries before this date\")\n\tcmd.Flags().StringP(\"account\", \"a\", \"\", \"account name\")\n\treturn &cmd\n}", "func CreateCommand(ca *common.CommonArgs) *cobra.Command {\n\tvar c command\n\tc = command{\n\t\tCommand: cobra.Command{\n\t\t\tUse: \"run\",\n\t\t\tShort: \"run observes the state of tugboat.lauches\",\n\t\t\tArgs: cobra.NoArgs,\n\t\t\tPreRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\t\treturn c.preexecute(cmd, args)\n\t\t\t},\n\t\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\t\treturn c.execute(cmd, args)\n\t\t\t},\n\t\t},\n\t\tCommonArgs: ca,\n\t\thttpFlagMgr: httpcliflags.New(),\n\t\tk8sFlagMgr: cliflags.New(),\n\t\tnotificationsFlagMgr: notificationscliflags.New(),\n\t}\n\n\tflags := c.Flags()\n\n\tc.httpFlagMgr.ConfigureHttpFlag(flags)\n\tc.k8sFlagMgr.ConfigureKubernetesConfig(flags)\n\tc.notificationsFlagMgr.ConfigureListenersFlag(flags)\n\n\treturn common.TraverseRunHooks(&c.Command)\n}", "func NewCommand() *cli.Command { //nolint:funlen\n\tconst (\n\t\tapiKeyMinLength = 8\n\t)\n\n\treturn &cli.Command{\n\t\tName: \"quota\",\n\t\tAliases: []string{\"q\"},\n\t\tUsage: \"Get currently used quota\",\n\t\tAction: func(c *cli.Context) error {\n\t\t\tvar apiKeys = c.StringSlice(shared.APIKeyFlag.Name)\n\n\t\t\tif len(apiKeys) == 0 {\n\t\t\t\treturn errors.New(\"API key(s) was not provided\")\n\t\t\t}\n\n\t\t\tvar (\n\t\t\t\tctx, cancel = context.WithCancel(c.Context) // main context creation\n\t\t\t\toss = breaker.NewOSSignals(ctx) // OS signals listener\n\t\t\t)\n\n\t\t\toss.Subscribe(func(os.Signal) { cancel() })\n\n\t\t\tdefer func() {\n\t\t\t\tcancel() // call cancellation function after all for \"service\" goroutines stopping\n\t\t\t\toss.Stop() // stop system signals listening\n\t\t\t}()\n\n\t\t\tvar (\n\t\t\t\twg sync.WaitGroup\n\t\t\t\terrColor = text.Colors{text.FgRed, text.Bold}\n\t\t\t)\n\n\t\t\tfor _, key := range apiKeys {\n\t\t\t\tif len(key) <= apiKeyMinLength {\n\t\t\t\t\t_, _ = fmt.Fprint(os.Stderr, errColor.Sprintf(\"API key (%s) is too short\\n\", key))\n\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\twg.Add(1)\n\n\t\t\t\tgo func(key string) {\n\t\t\t\t\tdefer wg.Done()\n\n\t\t\t\t\tif ctx.Err() != nil { // check if context was canceled\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\n\t\t\t\t\tif count, err := tinypng.NewClient(key).UsedQuota(ctx); err != nil {\n\t\t\t\t\t\t_, _ = fmt.Fprint(os.Stderr, errColor.Sprintf(\"Key %s error: %s\\n\", maskString(key), err))\n\n\t\t\t\t\t\treturn\n\t\t\t\t\t} else {\n\t\t\t\t\t\tvar color = text.FgRed\n\n\t\t\t\t\t\tswitch {\n\t\t\t\t\t\tcase count <= 300: //nolint:gomnd\n\t\t\t\t\t\t\tcolor = text.FgGreen\n\n\t\t\t\t\t\tcase count <= 400: //nolint:gomnd\n\t\t\t\t\t\t\tcolor = text.FgYellow\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\t_, _ = fmt.Fprintf(os.Stdout,\n\t\t\t\t\t\t\t\"Used quota (key %s) is: %s\\n\",\n\t\t\t\t\t\t\ttext.Colors{text.FgHiBlue}.Sprint(maskString(key)),\n\t\t\t\t\t\t\ttext.Colors{color, text.Bold}.Sprintf(\"%d\", count),\n\t\t\t\t\t\t)\n\t\t\t\t\t}\n\t\t\t\t}(key)\n\t\t\t}\n\n\t\t\twg.Wait()\n\n\t\t\treturn nil\n\t\t},\n\t\tFlags: []cli.Flag{\n\t\t\tshared.APIKeyFlag,\n\t\t},\n\t}\n}", "func NewTasksCommand() *cobra.Command {\n\tcommand := &cobra.Command{\n\t\tUse: \"tasks\",\n\t\tShort: \"Task related option\",\n\t\tLong: \"Task related option, e.g. list tasks\",\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\tcmd.Usage()\n\t\t},\n\t}\n\n\tcommand.AddCommand(listTasksCommand())\n\n\treturn command\n}", "func createCommand(ctx context.Context, cmdline []string, env env) commander {\n\tlogrus.Debugf(\"command: %v (env: %s)\", cmdline, env)\n\tcmd := exec.CommandContext(ctx, cmdline[0], cmdline[1:]...)\n\tcmd.Env = env.AsPairs()\n\treturn cmd\n}", "func newDeleteCmdForEntityType(entityType string, options *edgeOptions, aliases ...string) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: entityType + \" <id>\",\n\t\tShort: \"deletes \" + getPlural(entityType) + \" managed by the Ziti Edge Controller\",\n\t\tArgs: cobra.MinimumNArgs(1),\n\t\tAliases: append(aliases, getPlural(entityType)),\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\toptions.Cmd = cmd\n\t\t\toptions.Args = args\n\t\t\terr := runDeleteEntityOfType(options, getPlural(entityType))\n\t\t\tcmdhelper.CheckErr(err)\n\t\t},\n\t\tSuggestFor: []string{},\n\t}\n\n\t// allow interspersing positional args and flags\n\tcmd.Flags().SetInterspersed(true)\n\toptions.AddCommonFlags(cmd)\n\n\tcmd.AddCommand(newDeleteWhereCmdForEntityType(entityType, options))\n\n\treturn cmd\n}", "func newListCmd(out io.Writer, errOut io.Writer) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"list\",\n\t\tShort: \"Lists various entities managed by the Ziti Edge Controller\",\n\t\tLong: \"Lists various entities managed by the Ziti Edge Controller\",\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\terr := cmd.Help()\n\t\t\tcmdhelper.CheckErr(err)\n\t\t},\n\t}\n\n\tnewOptions := func() *edgeOptions {\n\t\treturn &edgeOptions{\n\t\t\tCommonOptions: common.CommonOptions{Out: out, Err: errOut},\n\t\t}\n\t}\n\n\tcmd.AddCommand(newListCmdForEntityType(\"api-sessions\", runListApiSessions, newOptions()))\n\tcmd.AddCommand(newListCmdForEntityType(\"cas\", runListCAs, newOptions()))\n\tcmd.AddCommand(newListCmdForEntityType(\"config-types\", runListConfigTypes, newOptions()))\n\tcmd.AddCommand(newListCmdForEntityType(\"configs\", runListConfigs, newOptions()))\n\tcmd.AddCommand(newListEdgeRoutersCmd(newOptions()))\n\tcmd.AddCommand(newListCmdForEntityType(\"edge-router-policies\", runListEdgeRouterPolicies, newOptions(), \"erps\"))\n\tcmd.AddCommand(newListCmdForEntityType(\"terminators\", runListTerminators, newOptions()))\n\tcmd.AddCommand(newListIdentitiesCmd(newOptions()))\n\tcmd.AddCommand(newListServicesCmd(newOptions()))\n\tcmd.AddCommand(newListCmdForEntityType(\"service-edge-router-policies\", runListServiceEdgeRouterPolices, newOptions(), \"serps\"))\n\tcmd.AddCommand(newListCmdForEntityType(\"service-policies\", runListServicePolices, newOptions(), \"sps\"))\n\tcmd.AddCommand(newListCmdForEntityType(\"sessions\", runListSessions, newOptions()))\n\tcmd.AddCommand(newListCmdForEntityType(\"transit-routers\", runListTransitRouters, newOptions()))\n\n\tcmd.AddCommand(newListCmdForEntityType(\"edge-router-role-attributes\", runListEdgeRouterRoleAttributes, newOptions()))\n\tcmd.AddCommand(newListCmdForEntityType(\"identity-role-attributes\", runListIdentityRoleAttributes, newOptions()))\n\tcmd.AddCommand(newListCmdForEntityType(\"service-role-attributes\", runListServiceRoleAttributes, newOptions()))\n\n\tcmd.AddCommand(newListCmdForEntityType(\"posture-checks\", runListPostureChecks, newOptions()))\n\tcmd.AddCommand(newListCmdForEntityType(\"posture-check-types\", runListPostureCheckTypes, newOptions()))\n\n\tconfigTypeListRootCmd := newEntityListRootCmd(\"config-type\")\n\tconfigTypeListRootCmd.AddCommand(newSubListCmdForEntityType(\"config-type\", \"configs\", outputConfigs, newOptions()))\n\n\tedgeRouterListRootCmd := newEntityListRootCmd(\"edge-router\", \"er\")\n\tedgeRouterListRootCmd.AddCommand(newSubListCmdForEntityType(\"edge-routers\", \"edge-router-policies\", outputEdgeRouterPolicies, newOptions()))\n\tedgeRouterListRootCmd.AddCommand(newSubListCmdForEntityType(\"edge-routers\", \"service-edge-router-policies\", outputServiceEdgeRouterPolicies, newOptions()))\n\tedgeRouterListRootCmd.AddCommand(newSubListCmdForEntityType(\"edge-routers\", \"identities\", outputIdentities, newOptions()))\n\tedgeRouterListRootCmd.AddCommand(newSubListCmdForEntityType(\"edge-routers\", \"services\", outputServices, newOptions()))\n\n\tedgeRouterPolicyListRootCmd := newEntityListRootCmd(\"edge-router-policy\", \"erp\")\n\tedgeRouterPolicyListRootCmd.AddCommand(newSubListCmdForEntityType(\"edge-router-policies\", \"edge-routers\", outputEdgeRouters, newOptions()))\n\tedgeRouterPolicyListRootCmd.AddCommand(newSubListCmdForEntityType(\"edge-router-policies\", \"identities\", outputIdentities, newOptions()))\n\n\tidentityListRootCmd := newEntityListRootCmd(\"identity\")\n\tidentityListRootCmd.AddCommand(newSubListCmdForEntityType(\"identities\", \"edge-router-policies\", outputEdgeRouterPolicies, newOptions()))\n\tidentityListRootCmd.AddCommand(newSubListCmdForEntityType(\"identities\", \"edge-routers\", outputEdgeRouters, newOptions()))\n\tidentityListRootCmd.AddCommand(newSubListCmdForEntityType(\"identities\", \"service-policies\", outputServicePolicies, newOptions()))\n\tidentityListRootCmd.AddCommand(newSubListCmdForEntityType(\"identities\", \"services\", outputServices, newOptions()))\n\tidentityListRootCmd.AddCommand(newSubListCmdForEntityType(\"identities\", \"service-configs\", outputServiceConfigs, newOptions()))\n\n\tserviceListRootCmd := newEntityListRootCmd(\"service\")\n\tserviceListRootCmd.AddCommand(newSubListCmdForEntityType(\"services\", \"configs\", outputConfigs, newOptions()))\n\tserviceListRootCmd.AddCommand(newSubListCmdForEntityType(\"services\", \"service-policies\", outputServicePolicies, newOptions()))\n\tserviceListRootCmd.AddCommand(newSubListCmdForEntityType(\"services\", \"service-edge-router-policies\", outputServiceEdgeRouterPolicies, newOptions()))\n\tserviceListRootCmd.AddCommand(newSubListCmdForEntityType(\"services\", \"terminators\", outputTerminators, newOptions()))\n\tserviceListRootCmd.AddCommand(newSubListCmdForEntityType(\"services\", \"identities\", outputIdentities, newOptions()))\n\tserviceListRootCmd.AddCommand(newSubListCmdForEntityType(\"services\", \"edge-routers\", outputEdgeRouters, newOptions()))\n\n\tserviceEdgeRouterPolicyListRootCmd := newEntityListRootCmd(\"service-edge-router-policy\", \"serp\")\n\tserviceEdgeRouterPolicyListRootCmd.AddCommand(newSubListCmdForEntityType(\"service-edge-router-policies\", \"services\", outputServices, newOptions()))\n\tserviceEdgeRouterPolicyListRootCmd.AddCommand(newSubListCmdForEntityType(\"service-edge-router-policies\", \"edge-routers\", outputEdgeRouters, newOptions()))\n\n\tservicePolicyListRootCmd := newEntityListRootCmd(\"service-policy\", \"sp\")\n\tservicePolicyListRootCmd.AddCommand(newSubListCmdForEntityType(\"service-policies\", \"services\", outputServices, newOptions()))\n\tservicePolicyListRootCmd.AddCommand(newSubListCmdForEntityType(\"service-policies\", \"identities\", outputIdentities, newOptions()))\n\tservicePolicyListRootCmd.AddCommand(newSubListCmdForEntityType(\"service-policies\", \"posture-checks\", outputPostureChecks, newOptions()))\n\n\tcmd.AddCommand(newListCmdForEntityType(\"summary\", runListSummary, newOptions()))\n\n\tcmd.AddCommand(configTypeListRootCmd,\n\t\tedgeRouterListRootCmd,\n\t\tedgeRouterPolicyListRootCmd,\n\t\tidentityListRootCmd,\n\t\tserviceEdgeRouterPolicyListRootCmd,\n\t\tserviceListRootCmd,\n\t\tservicePolicyListRootCmd,\n\t)\n\n\treturn cmd\n}", "func Create(client *gophercloud.ServiceClient, opts CreateOptsBuilder) (r CreateResult) {\r\n\tb, err := opts.ToTenantCreateMap()\r\n\tif err != nil {\r\n\t\tr.Err = err\r\n\t\treturn\r\n\t}\r\n\t_, r.Err = client.Post(createURL(client), b, &r.Body, &gophercloud.RequestOpts{\r\n\t\tOkCodes: []int{200, 201},\r\n\t})\r\n\treturn\r\n}", "func Builder() *cobra.Command {\n\topts := &Opts{}\n\tcmd := &cobra.Command{\n\t\tUse: \"quickstart\",\n\t\tExample: `Skip setting cluster name, provider or database username by using the command options\n $ mongocli atlas quickstart --clusterName Test --provider GCP --username dbuserTest\n`,\n\t\tShort: \"Create and access an Atlas Cluster.\",\n\t\tLong: \"This command creates a new cluster, adds your public IP to the atlas access list and creates a db user to access your new MongoDB instance.\",\n\t\tPreRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tif config.PublicAPIKey() == \"\" || config.PrivateAPIKey() == \"\" {\n\t\t\t\t// no profile set\n\t\t\t\treturn askAtlasAccountAndProfile()\n\t\t\t}\n\t\t\topts.setTier()\n\t\t\treturn opts.PreRunE(\n\t\t\t\topts.ValidateProjectID,\n\t\t\t\topts.initStore,\n\t\t\t\topts.InitOutput(cmd.OutOrStdout(), \"\"),\n\t\t\t)\n\t\t},\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tconst base10 = 10\n\t\t\topts.defaultName = \"Quickstart-\" + strconv.FormatInt(time.Now().Unix(), base10)\n\t\t\topts.providerAndRegionToConstant()\n\n\t\t\tif err := opts.defaultValues(); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn opts.Run()\n\t\t},\n\t}\n\n\tcmd.Flags().StringVar(&opts.ClusterName, flag.ClusterName, \"\", usage.ClusterName)\n\tcmd.Flags().StringVar(&opts.tier, flag.Tier, atlasM2, usage.Tier)\n\tcmd.Flags().StringVar(&opts.Provider, flag.Provider, \"\", usage.Provider)\n\tcmd.Flags().StringVarP(&opts.Region, flag.Region, flag.RegionShort, \"\", usage.Region)\n\tcmd.Flags().StringSliceVar(&opts.IPAddresses, flag.AccessListIP, []string{}, usage.NetworkAccessListIPEntry)\n\tcmd.Flags().StringVar(&opts.DBUsername, flag.Username, \"\", usage.DBUsername)\n\tcmd.Flags().StringVar(&opts.DBUserPassword, flag.Password, \"\", usage.Password)\n\tcmd.Flags().BoolVar(&opts.SkipSampleData, flag.SkipSampleData, false, usage.SkipSampleData)\n\tcmd.Flags().BoolVar(&opts.SkipMongosh, flag.SkipMongosh, false, usage.SkipMongosh)\n\tcmd.Flags().BoolVarP(&opts.defaultValue, flag.Default, \"Y\", false, usage.QuickstartDefault)\n\n\tcmd.Flags().StringVar(&opts.ProjectID, flag.ProjectID, \"\", usage.ProjectID)\n\n\treturn cmd\n}", "func New() *cobra.Command {\n\tconst (\n\t\tlong = `Commands for managing Fly organizations. list, create, show and\ndestroy organizations.\nOrganization admins can also invite or remove users from Organizations.\n`\n\t\tshort = \"Commands for managing Fly organizations\"\n\t)\n\n\t// TODO: list should also accept the --org param\n\n\torgs := command.New(\"orgs\", short, long, nil)\n\n\torgs.AddCommand(\n\t\tnewList(),\n\t\tnewShow(),\n\t\tnewInvite(),\n\t\tnewRemove(),\n\t\tnewCreate(),\n\t\tnewDelete(),\n\t)\n\n\treturn orgs\n}", "func CreateWalletCmds(ccli *client.CommandLineClient) error {\n\tbc, err := client.NewLazyBaseClientFromCommandLineClient(ccli)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\twalletCmd := &walletCmd{\n\t\tcli: ccli,\n\t\twalletClient: rivinecli.NewWalletClient(bc),\n\t\ttxPoolClient: rivinecli.NewTransactionPoolClient(bc),\n\t\ttbClient: NewPluginExplorerClient(bc),\n\t}\n\n\t// define commands\n\tvar (\n\t\tsendBotRegistrationTxCmd = &cobra.Command{\n\t\t\tUse: \"botregistration\",\n\t\t\tShort: \"Create, sign and send a new 3bot registration transaction\",\n\t\t\tLong: `Create, sign and send a new 3bot registration transaction, prepaying 1 month by default.\nThe coin inputs are funded and signed using the wallet of this daemon.\nBy default a public key is generated from this wallet's primary seed as well,\nhowever, it is also allowed for you to give a public key that is already loaded in this wallet,\nfor the creation of the 3bot.\n\nAddresses and names are added as flags, and at least one of both is required.\nMultiple addresses and names are allowed as well, of course.\n\nShould you want to prepay more than 1 month, this has to be specified as a flag as well.\nOne might want to do this, as the ThreefoldFoundation gives 30% discount for 12+ (bot) months,\nand 50% discount for 24 (bot) months (the maximum).\n\nAll fees are automatically added.\n\nIf this command returns without errors, the Tx is signed and sent,\nand you'll receive the TxID and PublicKey which will allow you to look it up in an explorer.\nThe public key is to be used to get to know the unique ID assigned to your registered bot (if succesfull).\n`,\n\t\t\tRun: rivinecli.Wrap(walletCmd.sendBotRegistrationTxCmd),\n\t\t}\n\n\t\tsendBotRecordUpdateTxCmd = &cobra.Command{\n\t\t\tUse: \"botupdate (id|publickey)\",\n\t\t\tShort: \"Create, sign and send a 3bot record update transaction\",\n\t\t\tLong: `Create, sign and send a 3bot record update transaction, updating an existing 3bot.\nThe coin inputs are funded and signed using the wallet of this daemon.\nThe Public key linked to the 3bot has to be loaded into the wallet in order to be able to sign.\n\nAddresses and names to be removed/added are defined as flags, and at least one\nupdate is required (defining NrOfMonths to add (and pay) to the 3bot record counts as an update as well).\n\n> NOTE: a name can only be removed if owned (which implies the 3bot has to be active at the point of the update).\n\nShould you want to prepay more than 1 month at once, this is possible and\nthe ThreefoldFoundation gives 30% discount for 12+ (bot) months,\nand 50% discount for 24 (bot) months (the maximum).\n\nAll fees are automatically added.\n\nIf this command returns without errors, the Tx is signed and sent,\nand you'll receive the TxID which will allow you to look it up in an explorer.\n`,\n\t\t\tRun: rivinecli.Wrap(walletCmd.sendBotRecordUpdateTxCmd),\n\t\t}\n\n\t\tcreateBotNameTransferTxCmd = &cobra.Command{\n\t\t\tUse: \"botnametransfer (id|publickey) (id|publickey) names...\",\n\t\t\tArgs: cobra.MinimumNArgs(3),\n\t\t\tShort: \"Create and optionally sign a 3bot name transfer transaction\",\n\t\t\tLong: `Create and optionally sign a 3bot name transfer transaction, involving two active 3bots.\nThe coin inputs are funded and signed using the wallet of this daemon.\nThe Public key linked to the 3bot has to be loaded into the wallet in order to be able to sign.\n\nThe first positional argument identifies the sender, nad the second positional argument identifies the receiver.\nAll other positional arguments (at least one more is required) define the names to be transfered.\nAt least one name has to be transferred.\n\nAll fees are automatically added.\n\nIf this command returns without errors, the Tx (optionally signed)\nis printed to the STDOUT.\n`,\n\t\t\tRun: walletCmd.createBotNameTransferTxCmd,\n\t\t}\n\t)\n\n\t// add commands as wallet sub commands\n\tccli.WalletCmd.RootCmdCreate.AddCommand(\n\t\tcreateBotNameTransferTxCmd,\n\t)\n\tccli.WalletCmd.RootCmdSend.AddCommand(\n\t\tsendBotRegistrationTxCmd,\n\t\tsendBotRecordUpdateTxCmd,\n\t)\n\n\t// register flags\n\tNetworkAddressArrayFlagVar(\n\t\tsendBotRegistrationTxCmd.Flags(),\n\t\t&walletCmd.sendBotRegistrationTxCfg.Addresses,\n\t\t\"address\",\n\t\t\"add one or multiple addresses, each address defined as seperate flag arguments\",\n\t)\n\tBotNameArrayFlagVar(\n\t\tsendBotRegistrationTxCmd.Flags(),\n\t\t&walletCmd.sendBotRegistrationTxCfg.Names,\n\t\t\"name\",\n\t\t\"add one or multiple names, each name defined as seperate flag arguments\",\n\t)\n\tsendBotRegistrationTxCmd.Flags().Uint8VarP(\n\t\t&walletCmd.sendBotRegistrationTxCfg.NrOfMonths, \"months\", \"m\", 1,\n\t\t\"the amount of months to prepay, required to be in the inclusive interval [1, 24]\")\n\tPublicKeyFlagVar(\n\t\tsendBotRegistrationTxCmd.Flags(),\n\t\t&walletCmd.sendBotRegistrationTxCfg.PublicKey,\n\t\t\"public-key\",\n\t\t\"define a public key to use (of which the private key is loaded in this daemon's wallet)\",\n\t)\n\tsendBotRegistrationTxCmd.Flags().Var(\n\t\tcli.NewEncodingTypeFlag(0, &walletCmd.sendBotRegistrationTxCfg.EncodingType, cli.EncodingTypeHuman|cli.EncodingTypeJSON), \"encoding\",\n\t\tcli.EncodingTypeFlagDescription(cli.EncodingTypeHuman|cli.EncodingTypeJSON))\n\n\tNetworkAddressArrayFlagVar(\n\t\tsendBotRecordUpdateTxCmd.Flags(),\n\t\t&walletCmd.sendBotRecordUpdateTxCfg.AddressesToAdd,\n\t\t\"add-address\",\n\t\t\"add one or multiple addresses, each address defined as seperate flag arguments\",\n\t)\n\tNetworkAddressArrayFlagVar(\n\t\tsendBotRecordUpdateTxCmd.Flags(),\n\t\t&walletCmd.sendBotRecordUpdateTxCfg.AddressesToRemove,\n\t\t\"remove-address\",\n\t\t\"remove one or multiple addresses, each address defined as seperate flag arguments\",\n\t)\n\tBotNameArrayFlagVar(\n\t\tsendBotRecordUpdateTxCmd.Flags(),\n\t\t&walletCmd.sendBotRecordUpdateTxCfg.NamesToAdd,\n\t\t\"add-name\",\n\t\t\"add one or multiple names, each name defined as seperate flag arguments\",\n\t)\n\tBotNameArrayFlagVar(\n\t\tsendBotRecordUpdateTxCmd.Flags(),\n\t\t&walletCmd.sendBotRecordUpdateTxCfg.NamesToRemove,\n\t\t\"remove-name\",\n\t\t\"remove one or multiple names owned, each name defined as seperate flag arguments\",\n\t)\n\tsendBotRecordUpdateTxCmd.Flags().Uint8VarP(\n\t\t&walletCmd.sendBotRecordUpdateTxCfg.NrOfMonthsToAdd, \"add-months\", \"m\", 0,\n\t\t\"the amount of months to add and pay, required to be in the inclusive interval [0, 24]\")\n\tsendBotRecordUpdateTxCmd.Flags().Var(\n\t\tcli.NewEncodingTypeFlag(0, &walletCmd.sendBotRecordUpdateTxCfg.EncodingType, cli.EncodingTypeHuman|cli.EncodingTypeJSON), \"encoding\",\n\t\tcli.EncodingTypeFlagDescription(cli.EncodingTypeHuman|cli.EncodingTypeJSON))\n\n\tcreateBotNameTransferTxCmd.Flags().Var(\n\t\tcli.NewEncodingTypeFlag(0, &walletCmd.createBotNameTransferTxCfg.EncodingType, cli.EncodingTypeHuman|cli.EncodingTypeJSON), \"encoding\",\n\t\tcli.EncodingTypeFlagDescription(cli.EncodingTypeHuman|cli.EncodingTypeJSON))\n\tcreateBotNameTransferTxCmd.Flags().BoolVar(\n\t\t&walletCmd.createBotNameTransferTxCfg.Sign, \"sign\", false,\n\t\t\"optionally sign the transaction (as sender/receiver) prior to printing it\")\n\n\treturn nil\n}", "func NewSeedAdmissionControllerCommand() *cobra.Command {\n\topts := &Options{}\n\n\tcmd := &cobra.Command{\n\t\tUse: Name,\n\t\tShort: \"Launch the \" + Name,\n\t\tLong: Name + \" serves validating and mutating webhook endpoints for resources in seed clusters.\",\n\t\tArgs: cobra.NoArgs,\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tverflag.PrintAndExitIfRequested()\n\n\t\t\tif err := opts.validate(); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tcmd.SilenceUsage = true\n\n\t\t\tlog.Info(\"Starting \"+Name+\"...\", \"version\", version.Get())\n\t\t\tcmd.Flags().VisitAll(func(flag *pflag.Flag) {\n\t\t\t\tlog.Info(fmt.Sprintf(\"FLAG: --%s=%s\", flag.Name, flag.Value))\n\t\t\t})\n\n\t\t\treturn opts.Run(cmd.Context())\n\t\t},\n\t}\n\n\tflags := cmd.Flags()\n\tflags.AddGoFlagSet(flag.CommandLine)\n\tverflag.AddFlags(flags)\n\topts.AddFlags(flags)\n\treturn cmd\n}", "func NewStorageCreateCmd() *cobra.Command {\n\tstorageCreateCmd := &cobra.Command{\n\t\tUse: \"create\",\n\t\tShort: \"Create and register a storage system.\",\n\t\tLong: `Creates and registers a storage system.`,\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\tctx, cancel := context.WithCancel(context.Background())\n\t\t\tdefer cancel()\n\n\t\t\terrAndExit := func(err error) {\n\t\t\t\tfmt.Fprintf(cmd.ErrOrStderr(), \"error: %+v\\n\", err)\n\t\t\t\tosExit(1)\n\t\t\t}\n\n\t\t\t// Convenience functions for ignoring errors whilst\n\t\t\t// getting flag values.\n\t\t\tflagStringValue := func(v string, err error) string {\n\t\t\t\tif err != nil {\n\t\t\t\t\terrAndExit(err)\n\t\t\t\t}\n\t\t\t\treturn v\n\t\t\t}\n\t\t\tflagBoolValue := func(v bool, err error) bool {\n\t\t\t\tif err != nil {\n\t\t\t\t\terrAndExit(err)\n\t\t\t\t}\n\t\t\t\treturn v\n\t\t\t}\n\t\t\tverifyInput := func(v string) string {\n\t\t\t\tinputText := flagStringValue(cmd.Flags().GetString(v))\n\t\t\t\tif strings.TrimSpace(inputText) == \"\" {\n\t\t\t\t\terrAndExit(fmt.Errorf(\"no input provided: %s\", v))\n\t\t\t\t}\n\t\t\t\treturn inputText\n\t\t\t}\n\n\t\t\t// Gather the inputs\n\t\t\tvar input = struct {\n\t\t\t\tType string\n\t\t\t\tEndpoint string\n\t\t\t\tSystemID string\n\t\t\t\tUser string\n\t\t\t\tPassword string\n\t\t\t\tInsecure bool\n\t\t\t}{\n\t\t\t\tType: verifyInput(\"type\"),\n\t\t\t\tEndpoint: verifyInput(\"endpoint\"),\n\t\t\t\tSystemID: flagStringValue(cmd.Flags().GetString(\"system-id\")),\n\t\t\t\tUser: verifyInput(\"user\"),\n\t\t\t\tPassword: flagStringValue(cmd.Flags().GetString(\"password\")),\n\t\t\t\tInsecure: flagBoolValue(cmd.Flags().GetBool(\"insecure\")),\n\t\t\t}\n\n\t\t\t// Parse the URL and prepare for a password prompt.\n\t\t\turlWithUser, err := url.Parse(input.Endpoint)\n\t\t\tif err != nil {\n\t\t\t\terrAndExit(err)\n\t\t\t}\n\n\t\t\turlWithUser.Scheme = \"https\"\n\t\t\turlWithUser.User = url.User(input.User)\n\n\t\t\t// If the password was not provided...\n\t\t\tprompt := fmt.Sprintf(\"Enter password for %v: \", urlWithUser)\n\t\t\t// If the password was not provided...\n\t\t\tif pf := cmd.Flags().Lookup(\"password\"); !pf.Changed {\n\t\t\t\t// Get password from stdin\n\t\t\t\treadPassword(cmd.ErrOrStderr(), prompt, &input.Password)\n\t\t\t}\n\n\t\t\t// Sanitize the endpoint\n\t\t\tepURL, err := url.Parse(input.Endpoint)\n\t\t\tif err != nil {\n\t\t\t\terrAndExit(err)\n\t\t\t}\n\t\t\tepURL.Scheme = \"https\"\n\n\t\t\t// Get the current list of registered storage systems\n\t\t\tk3sCmd := execCommandContext(ctx, K3sPath, \"kubectl\", \"get\",\n\t\t\t\t\"--namespace=karavi\",\n\t\t\t\t\"--output=json\",\n\t\t\t\t\"secret/karavi-storage-secret\")\n\n\t\t\tb, err := k3sCmd.Output()\n\t\t\tif err != nil {\n\t\t\t\terrAndExit(err)\n\t\t\t}\n\t\t\tbase64Systems := struct {\n\t\t\t\tData map[string]string\n\t\t\t}{}\n\t\t\tif err := json.Unmarshal(b, &base64Systems); err != nil {\n\t\t\t\terrAndExit(err)\n\t\t\t}\n\t\t\tdecodedSystems, err := base64.StdEncoding.DecodeString(base64Systems.Data[\"storage-systems.yaml\"])\n\t\t\tif err != nil {\n\t\t\t\terrAndExit(err)\n\t\t\t}\n\n\t\t\tvar listData map[string]Storage\n\t\t\tif err := yaml.Unmarshal(decodedSystems, &listData); err != nil {\n\t\t\t\terrAndExit(err)\n\t\t\t}\n\t\t\tif listData == nil || listData[\"storage\"] == nil {\n\t\t\t\tlistData = make(map[string]Storage)\n\t\t\t\tlistData[\"storage\"] = make(Storage)\n\t\t\t}\n\t\t\tvar storage = listData[\"storage\"]\n\t\t\t// Check that we are not duplicating, no errors, etc.\n\n\t\t\tif _, ok := supportedStorageTypes[input.Type]; !ok {\n\t\t\t\terrAndExit(fmt.Errorf(\"unsupported type: %s\", input.Type))\n\t\t\t}\n\n\t\t\tsysIDs := strings.Split(input.SystemID, \",\")\n\t\t\tisDuplicate := func() (string, bool) {\n\t\t\t\tstorType, ok := storage[input.Type]\n\t\t\t\tif !ok {\n\t\t\t\t\tstorage[input.Type] = make(map[string]System)\n\t\t\t\t\treturn \"\", false\n\t\t\t\t}\n\t\t\t\tfor _, id := range sysIDs {\n\t\t\t\t\tif _, ok = storType[fmt.Sprintf(id)]; ok {\n\t\t\t\t\t\treturn id, true\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\treturn \"\", false\n\t\t\t}\n\n\t\t\tif id, result := isDuplicate(); result {\n\t\t\t\tfmt.Fprintf(cmd.ErrOrStderr(), \"error: %s system with ID %s is already registered\\n\", input.Type, id)\n\t\t\t\tosExit(1)\n\t\t\t}\n\n\t\t\t// Attempt to connect to the storage using the provided details.\n\t\t\t// TODO(ian): This logic should ideally be performed remotely, not\n\t\t\t// in the client.\n\n\t\t\tvar tempStorage SystemType\n\n\t\t\tswitch input.Type {\n\t\t\tcase powerflex:\n\t\t\t\ttempStorage = storage[powerflex]\n\t\t\t\tif tempStorage == nil {\n\t\t\t\t\ttempStorage = make(map[string]System)\n\t\t\t\t}\n\n\t\t\t\tsioClient, err := goscaleio.NewClientWithArgs(epURL.String(), \"\", true, false)\n\t\t\t\tif err != nil {\n\t\t\t\t\terrAndExit(err)\n\t\t\t\t}\n\n\t\t\t\t_, err = sioClient.Authenticate(&goscaleio.ConfigConnect{\n\t\t\t\t\tUsername: input.User,\n\t\t\t\t\tPassword: input.Password,\n\t\t\t\t})\n\t\t\t\tif err != nil {\n\t\t\t\t\terrAndExit(err)\n\t\t\t\t}\n\n\t\t\t\tresp, err := sioClient.FindSystem(input.SystemID, \"\", \"\")\n\t\t\t\tif err != nil {\n\t\t\t\t\terrAndExit(err)\n\t\t\t\t}\n\t\t\t\tif resp.System.ID != input.SystemID {\n\t\t\t\t\tfmt.Fprintf(cmd.ErrOrStderr(), \"system id %q not found\", input.SystemID)\n\t\t\t\t\tosExit(1)\n\t\t\t\t}\n\n\t\t\t\tstorageID := strings.Trim(SystemID{Value: input.SystemID}.String(), \"\\\"\")\n\t\t\t\ttempStorage[storageID] = System{\n\t\t\t\t\tUser: input.User,\n\t\t\t\t\tPassword: input.Password,\n\t\t\t\t\tEndpoint: input.Endpoint,\n\t\t\t\t\tInsecure: input.Insecure,\n\t\t\t\t}\n\n\t\t\tcase powermax:\n\t\t\t\ttempStorage = storage[powermax]\n\t\t\t\tif tempStorage == nil {\n\t\t\t\t\ttempStorage = make(map[string]System)\n\t\t\t\t}\n\n\t\t\t\tpmClient, err := pmax.NewClientWithArgs(epURL.String(), \"\", \"karavi-auth\", true, false)\n\t\t\t\tif err != nil {\n\t\t\t\t\terrAndExit(err)\n\t\t\t\t}\n\n\t\t\t\tconfigConnect := &pmax.ConfigConnect{\n\t\t\t\t\tEndpoint: input.Endpoint,\n\t\t\t\t\tVersion: \"\",\n\t\t\t\t\tUsername: input.User,\n\t\t\t\t\tPassword: input.Password,\n\t\t\t\t}\n\t\t\t\terr = pmClient.Authenticate(ctx, configConnect)\n\t\t\t\tif err != nil {\n\t\t\t\t\terrAndExit(err)\n\t\t\t\t}\n\n\t\t\t\tvar powermaxSymmetrix []*types.Symmetrix\n\n\t\t\t\tsymmetrixIDList, err := pmClient.GetSymmetrixIDList(ctx)\n\t\t\t\tif err != nil {\n\t\t\t\t\terrAndExit(err)\n\t\t\t\t}\n\t\t\t\tfor _, s := range symmetrixIDList.SymmetrixIDs {\n\t\t\t\t\tsymmetrix, err := pmClient.GetSymmetrixByID(ctx, s)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\terrAndExit(err)\n\t\t\t\t\t}\n\t\t\t\t\tif strings.Contains(symmetrix.Model, \"PowerMax\") || strings.Contains(symmetrix.Model, \"VMAX\") {\n\t\t\t\t\t\tpowermaxSymmetrix = append(powermaxSymmetrix, symmetrix)\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\tcreateStorageFunc := func(id string) {\n\t\t\t\t\ttempStorage[id] = System{\n\t\t\t\t\t\tUser: input.User,\n\t\t\t\t\t\tPassword: input.Password,\n\t\t\t\t\t\tEndpoint: input.Endpoint,\n\t\t\t\t\t\tInsecure: input.Insecure,\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\tfor _, p := range powermaxSymmetrix {\n\t\t\t\t\tstorageID := strings.Trim(SystemID{Value: p.SymmetrixID}.String(), \"\\\"\")\n\t\t\t\t\tif input.SystemID != \"\" {\n\t\t\t\t\t\tif len(sysIDs) > 0 {\n\t\t\t\t\t\t\tif contains(p.SymmetrixID, sysIDs) {\n\t\t\t\t\t\t\t\tcreateStorageFunc(storageID)\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\tcontinue\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t\tcreateStorageFunc(storageID)\n\t\t\t\t}\n\n\t\t\tcase powerscale:\n\t\t\t\ttempStorage = storage[powerscale]\n\t\t\t\tif tempStorage == nil {\n\t\t\t\t\ttempStorage = make(map[string]System)\n\t\t\t\t}\n\n\t\t\t\tpsClient, err := pscale.NewClientWithArgs(context.Background(), epURL.String(), input.Insecure, 1, input.User, \"Administrators\", input.Password, \"\", \"777\")\n\t\t\t\tif err != nil {\n\t\t\t\t\terrAndExit(err)\n\t\t\t\t}\n\n\t\t\t\tclusterConfig, err := psClient.GetClusterConfig(context.Background())\n\t\t\t\tif err != nil {\n\t\t\t\t\terrAndExit(err)\n\t\t\t\t}\n\n\t\t\t\tif clusterConfig.Name != input.SystemID {\n\t\t\t\t\tfmt.Fprintf(cmd.ErrOrStderr(), \"cluster name %q not found\", input.SystemID)\n\t\t\t\t\tosExit(1)\n\t\t\t\t}\n\n\t\t\t\ttempStorage[input.SystemID] = System{\n\t\t\t\t\tUser: input.User,\n\t\t\t\t\tPassword: input.Password,\n\t\t\t\t\tEndpoint: input.Endpoint,\n\t\t\t\t\tInsecure: input.Insecure,\n\t\t\t\t}\n\n\t\t\tdefault:\n\t\t\t\terrAndExit(fmt.Errorf(\"invalid storage array type given\"))\n\t\t\t}\n\n\t\t\t// Merge the new connection details and apply them.\n\n\t\t\tstorage[input.Type] = tempStorage\n\t\t\tlistData[\"storage\"] = storage\n\n\t\t\tb, err = yaml.Marshal(&listData)\n\t\t\tif err != nil {\n\t\t\t\terrAndExit(err)\n\t\t\t}\n\n\t\t\ttmpFile, err := ioutil.TempFile(\"\", \"karavi\")\n\t\t\tif err != nil {\n\t\t\t\terrAndExit(err)\n\t\t\t}\n\t\t\tdefer func() {\n\t\t\t\tif err := tmpFile.Close(); err != nil {\n\t\t\t\t\tfmt.Fprintf(os.Stderr, \"error: %+v\\n\", err)\n\t\t\t\t}\n\t\t\t\tif err := os.Remove(tmpFile.Name()); err != nil {\n\t\t\t\t\tfmt.Fprintf(os.Stderr, \"error: %+v\\n\", err)\n\t\t\t\t}\n\t\t\t}()\n\t\t\t_, err = tmpFile.WriteString(string(b))\n\t\t\tif err != nil {\n\t\t\t\terrAndExit(err)\n\t\t\t}\n\n\t\t\tcrtCmd := execCommandContext(ctx, K3sPath, \"kubectl\", \"create\",\n\t\t\t\t\"--namespace=karavi\",\n\t\t\t\t\"secret\", \"generic\", \"karavi-storage-secret\",\n\t\t\t\tfmt.Sprintf(\"--from-file=storage-systems.yaml=%s\", tmpFile.Name()),\n\t\t\t\t\"--output=yaml\",\n\t\t\t\t\"--dry-run=client\")\n\t\t\tappCmd := execCommandContext(ctx, K3sPath, \"kubectl\", \"apply\", \"-f\", \"-\")\n\n\t\t\tif err := pipeCommands(crtCmd, appCmd); err != nil {\n\t\t\t\terrAndExit(err)\n\t\t\t}\n\t\t},\n\t}\n\n\tstorageCreateCmd.Flags().StringP(\"type\", \"t\", \"\", \"Type of storage system\")\n\terr := storageCreateCmd.MarkFlagRequired(\"type\")\n\tif err != nil {\n\t\treportErrorAndExit(JSONOutput, storageCreateCmd.ErrOrStderr(), err)\n\t}\n\tstorageCreateCmd.Flags().StringP(\"endpoint\", \"e\", \"\", \"Endpoint of REST API gateway\")\n\terr = storageCreateCmd.MarkFlagRequired(\"endpoint\")\n\tif err != nil {\n\t\treportErrorAndExit(JSONOutput, storageCreateCmd.ErrOrStderr(), err)\n\t}\n\tstorageCreateCmd.Flags().StringP(\"user\", \"u\", \"\", \"Username\")\n\terr = storageCreateCmd.MarkFlagRequired(\"user\")\n\tif err != nil {\n\t\treportErrorAndExit(JSONOutput, storageCreateCmd.ErrOrStderr(), err)\n\t}\n\tstorageCreateCmd.Flags().StringP(\"system-id\", \"s\", \"\", \"System identifier\")\n\tstorageCreateCmd.Flags().StringP(\"password\", \"p\", \"\", \"Specify password, or omit to use stdin\")\n\tstorageCreateCmd.Flags().BoolP(\"insecure\", \"i\", false, \"Insecure skip verify\")\n\n\treturn storageCreateCmd\n}", "func CommandCreate(appName string) error {\n\tif err := common.IsValidAppName(appName); err != nil {\n\t\treturn err\n\t}\n\n\treturn createApp(appName)\n}", "func New(o *Options) *cobra.Command {\n\tcmd := &cobra.Command{}\n\tcmd.Flags().StringVar(&o.KubeConfigPath, \"kubeconfig\", o.KubeConfigPath, \"absolute path of kubeconfig\")\n\tcmd.Flags().StringVar(&o.Provider, \"provider\", o.Provider, \"provider name of ACME\")\n\tcmd.Flags().StringVar(&o.BackendURL, \"backend\", o.BackendURL, \"URL for backend\")\n\tcmd.Flags().StringVar(&o.CaddyHost, \"caddy-host\", o.CaddyHost, \"hostname or ip of caddy\")\n\n\t// cloudflare provider\n\tcmd.Flags().StringVar(&o.CloudFlareEmail, \"cloudflare-email\", o.CloudFlareEmail, \"[cloudflare] Email\")\n\tcmd.Flags().StringVar(&o.CloudFlareAPIToken, \"cloudflare-api-token\", o.CloudFlareAPIToken, \"[cloudflare] API Token\")\n\treturn cmd\n}", "func NewCommand(\n\tname string,\n\tdescription string,\n\tversion string,\n\tnameConvertFunc spec.AttributeNameConverterFunc,\n\ttypeConvertFunc spec.AttributeTypeConverterFunc,\n\ttypeMappingName string,\n\tgeneratorFunc func([]spec.SpecificationSet, string) error,\n) *cobra.Command {\n\n\tcobra.OnInitialize(func() {\n\t\tviper.SetEnvPrefix(name)\n\t\tviper.AutomaticEnv()\n\t\tviper.SetEnvKeyReplacer(strings.NewReplacer(\"-\", \"_\"))\n\t})\n\n\tvar rootCmd = &cobra.Command{\n\t\tUse: name,\n\t\tShort: description,\n\t}\n\n\trootCmd.PersistentFlags().StringP(\"out\", \"o\", \"codegen\", \"Default output path.\")\n\n\tvar versionCmd = &cobra.Command{\n\t\tUse: \"version\",\n\t\tShort: \"Prints the version and exit.\",\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\tfmt.Println(version)\n\t\t},\n\t}\n\n\tvar cmdFolderGen = &cobra.Command{\n\t\tUse: \"folder\",\n\t\tShort: \"Generate the model using a local directory containing the specs.\",\n\t\tSilenceUsage: true,\n\t\tSilenceErrors: true,\n\t\tPreRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\treturn viper.BindPFlags(cmd.Flags())\n\t\t},\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\n\t\t\tif len(viper.GetStringSlice(\"dir\")) == 0 {\n\t\t\t\treturn errors.New(\"--dir is required\")\n\t\t\t}\n\n\t\t\tvar specSets []spec.SpecificationSet\n\n\t\t\tfor _, dir := range viper.GetStringSlice(\"dir\") {\n\t\t\t\tset, err := spec.LoadSpecificationSet(\n\t\t\t\t\tdir,\n\t\t\t\t\tnameConvertFunc,\n\t\t\t\t\ttypeConvertFunc,\n\t\t\t\t\ttypeMappingName,\n\t\t\t\t)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\tspecSets = append(specSets, set)\n\t\t\t}\n\n\t\t\treturn generatorFunc(specSets, viper.GetString(\"out\"))\n\t\t},\n\t}\n\tcmdFolderGen.Flags().StringSliceP(\"dir\", \"d\", nil, \"Path of the specifications folder.\")\n\n\tvar githubGen = &cobra.Command{\n\t\tUse: \"github\",\n\t\tShort: \"Generate the model using a remote github repository.\",\n\t\tSilenceUsage: true,\n\t\tSilenceErrors: true,\n\t\tPreRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\treturn viper.BindPFlags(cmd.Flags())\n\t\t},\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\n\t\t\tspecSet, err := spec.LoadSpecificationSetFromGithub(\n\t\t\t\tviper.GetString(\"token\"),\n\t\t\t\tviper.GetString(\"repo\"),\n\t\t\t\tviper.GetString(\"ref\"),\n\t\t\t\tviper.GetString(\"path\"),\n\t\t\t\tnameConvertFunc,\n\t\t\t\ttypeConvertFunc,\n\t\t\t\ttypeMappingName,\n\t\t\t)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn generatorFunc([]spec.SpecificationSet{specSet}, viper.GetString(\"out\"))\n\t\t},\n\t}\n\tgithubGen.Flags().StringP(\"repo\", \"r\", \"\", \"Endpoint for the github api.\")\n\tgithubGen.Flags().StringP(\"path\", \"p\", \"\", \"Internal path to a directory in the repo if not in the root.\")\n\tgithubGen.Flags().StringP(\"ref\", \"R\", \"master\", \"Branch or tag to use.\")\n\tgithubGen.Flags().StringP(\"token\", \"t\", \"\", \"The api token to use.\")\n\n\trootCmd.AddCommand(\n\t\tversionCmd,\n\t\tcmdFolderGen,\n\t\tgithubGen,\n\t)\n\n\treturn rootCmd\n}", "func CreateCommand() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"create\",\n\t\tShort: \"create server\",\n\t\tRun: runCreateCommand,\n\t}\n\tcmd.Flags().AddFlagSet(createParam.getFlagSet())\n\treturn cmd\n}", "func NewCreateCommand(c2ClientFactory cli.APIClientFactory) cli.Command {\n\tcreateCmd := &createCommand{\n\t\tc2ClientFactory: c2ClientFactory,\n\t}\n\n\tcobraCmd := &cobra.Command{\n\t\tUse: \"create\",\n\t\tShort: \"Creates a new client\",\n\t\tLong: fmt.Sprintf(\"Creates a new client, require an unique name, and a file containing either a password or a %d bytes key\", e4crypto.KeyLen),\n\t\tRunE: createCmd.run,\n\t}\n\n\tcobraCmd.Flags().SortFlags = false\n\tcobraCmd.Flags().StringVar(&createCmd.flags.Name, \"name\", \"\", \"The client name\")\n\tcobraCmd.Flags().StringVar(&createCmd.flags.KeyPath, \"key\", \"\", fmt.Sprintf(\"Filepath to a %d bytes key\", e4crypto.KeyLen))\n\tcobraCmd.Flags().StringVar(&createCmd.flags.PasswordPath, \"password\", \"\", \"Filepath to a plaintext password file\")\n\n\tcreateCmd.cobraCmd = cobraCmd\n\n\treturn createCmd\n}", "func newListServicesCmd(options *edgeOptions) *cobra.Command {\n\tvar asIdentity string\n\tvar configTypes []string\n\tvar roleFilters []string\n\tvar roleSemantic string\n\n\tcmd := &cobra.Command{\n\t\tUse: \"services <filter>?\",\n\t\tShort: \"lists services managed by the Ziti Edge Controller\",\n\t\tLong: \"lists services managed by the Ziti Edge Controller\",\n\t\tArgs: cobra.MaximumNArgs(1),\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\toptions.Cmd = cmd\n\t\t\toptions.Args = args\n\t\t\terr := runListServices(asIdentity, configTypes, roleFilters, roleSemantic, options)\n\t\t\tcmdhelper.CheckErr(err)\n\t\t},\n\t\tSuggestFor: []string{},\n\t}\n\n\t// allow interspersing positional args and flags\n\tcmd.Flags().SetInterspersed(true)\n\tcmd.Flags().StringVar(&asIdentity, \"as-identity\", \"\", \"Allow admins to see services as they would be seen by a different identity\")\n\tcmd.Flags().StringSliceVar(&configTypes, \"config-types\", nil, \"Override which config types to view on services\")\n\tcmd.Flags().StringSliceVar(&roleFilters, \"role-filters\", nil, \"Allow filtering by roles\")\n\tcmd.Flags().StringVar(&roleSemantic, \"role-semantic\", \"\", \"Specify which roles semantic to use \")\n\toptions.AddCommonFlags(cmd)\n\n\treturn cmd\n}", "func NewRootCommand() *cobra.Command {\n\n\tcmd := &cobra.Command{\n\t\tUse: \"katib-cli\",\n\t\tShort: \"katib cli\",\n\t\tLong: `This is katib cli client using cobra framework`,\n\t}\n\n\t//initialize config\n\tinitFlag(cmd)\n\n\t//add command\n\tcmd.AddCommand(NewCommandCreate())\n\tcmd.AddCommand(NewCommandGet())\n\tcmd.AddCommand(NewCommandStop())\n\tcmd.AddCommand(NewCommandPush())\n\tcmd.AddCommand(NewCommandPull())\n\n\t//\tcmd.AddCommand(NewCommandModel())\n\n\t//MISC\n\t//cmd.AddCommand(NewCommandVersion())\n\n\t//Generate bash completion file\n\t//cmd.AddCommand(NewCommandBashCmp())\n\n\treturn cmd\n}", "func CreateQuota(f cmdutil.Factory, cmdOut io.Writer, cmd *cobra.Command, args []string) error {\n\tname, err := NameFromCommandArgs(cmd, args)\n\tif err != nil {\n\t\treturn err\n\t}\n\tvar generator kubectl.StructuredGenerator\n\tswitch generatorName := cmdutil.GetFlagString(cmd, \"generator\"); generatorName {\n\tcase cmdutil.ResourceQuotaV1GeneratorName:\n\t\tgenerator = &kubectl.ResourceQuotaGeneratorV1{\n\t\t\tName: name,\n\t\t\tHard: cmdutil.GetFlagString(cmd, \"hard\"),\n\t\t\tScopes: cmdutil.GetFlagString(cmd, \"scopes\"),\n\t\t}\n\tdefault:\n\t\treturn cmdutil.UsageError(cmd, fmt.Sprintf(\"Generator: %s not supported.\", generatorName))\n\t}\n\treturn RunCreateSubcommand(f, cmd, cmdOut, &CreateSubcommandOptions{\n\t\tName: name,\n\t\tStructuredGenerator: generator,\n\t\tDryRun: cmdutil.GetFlagBool(cmd, \"dry-run\"),\n\t\tOutputFormat: cmdutil.GetFlagString(cmd, \"output\"),\n\t})\n}", "func ExampleClient_CreateCommand() {\n\tclient := &Client{}\n\n\tnewCommand := &Command{\n\t\tName: \"my-command\",\n\t\tDescription: \"my command\",\n\t\tArgs: \"[@username]\",\n\t\tSet: \"custom_cmd_set\",\n\t}\n\n\t_, _ = client.CreateCommand(newCommand)\n}", "func ScaffoldCommand() cli.Command {\n return cli.Command{\n\t\tName: \"scaff\",\n\t\tUsage: \"Order and orient contigs using Hi-C contact frequency data.\",\n\t\tSubcommands: []cli.Command{\n\t cli.Command{\n\t\t\t\tName: \"infer\",\n\t\t\t\tUsage: \"Infer a scaffolding from link data, e.g. lxy scaff infer --links data/test/GM.1mbp.links --output data/test/testscaffolding.txt --subset X --key data/test/testkey.txt --viz data/test/testorderfig.png\",\n\t\t\t\tFlags: []cli.Flag{\n\t\t\t\t\tcli.BoolFlag{\n\t\t\t\t\t\tName: \"debug\",\n\t\t\t\t\t\tUsage: \"Whether to print detailed debugging information.\",\n\t\t\t\t\t},\n\t\t\t\t \tcli.StringFlag{\n\t\t\t\t \t\tName: \"links\", \n\t\t\t\t \t\tValue: \"\", \n\t\t\t\t \t\tUsage: \"Path to the Hi-C links file.\",\n\t\t\t\t \t},\n\t\t\t\t \tcli.StringFlag{\n\t\t\t\t \t\tName: \"subset\", \n\t\t\t\t \t\tValue: \"\", \n\t\t\t\t \t\tUsage: \"Tag on basis of which to subset links.\",\n\t\t\t\t \t},\n\t\t\t\t \tcli.StringFlag{\n\t\t\t\t \t\tName: \"output\", \n\t\t\t\t \t\tValue: \"\", \n\t\t\t\t \t\tUsage: \"Output path for inferred phasing.\",\n\t\t\t\t \t},\n\t\t\t\t \tcli.StringFlag{\n\t\t\t\t \t\tName: \"key\", \n\t\t\t\t \t\tValue: \"\", \n\t\t\t\t \t\tUsage: \"Path to scaffolding key file.\",\n\t\t\t\t \t},\n\t\t\t\t \tcli.StringFlag{\n\t\t\t\t \t\tName: \"viz\", \n\t\t\t\t \t\tValue: \"\", \n\t\t\t\t \t\tUsage: \"Path to output the ordering viz, requires --key.\",\n\t\t\t\t \t},\n\t\t\t\t \tcli.IntFlag{\n\t\t\t\t \t\tName: \"iterations\", \n\t\t\t\t \t\tValue: 1000, \n\t\t\t\t \t\tUsage: \"Number of iterations or GA 'generations' to perform.\",\n\t\t\t\t \t},\n\t\t\t\t \tcli.IntFlag{\n\t\t\t\t \t\tName: \"popSize\", \n\t\t\t\t \t\tValue: 50, \n\t\t\t\t \t\tUsage: \"Population size for genetic algorithm.\",\n\t\t\t\t \t},\n\t\t\t\t \tcli.Float64Flag{\n\t\t\t\t \t\tName: \"breedProb\", \n\t\t\t\t \t\tValue: 0.8, \n\t\t\t\t \t\tUsage: \"Probability of breeding between two members of population.\",\n\t\t\t\t \t},\t\n\t\t\t\t \tcli.Float64Flag{\n\t\t\t\t \t\tName: \"mutProb\", \n\t\t\t\t \t\tValue: 0.8, \n\t\t\t\t \t\tUsage: \"Probability of a mutation occurring.\",\n\t\t\t\t \t},\t\t \t \t\n\t\t\t\t},\n\t\t\t\tAction: scaffoldInferCommand,\n \t\t},\n\t cli.Command{\n\t\t\t\tName: \"eval\",\n\t\t\t\tUsage: \"Evaluate a scaffolding using a key ordering.\",\n\t\t\t\tFlags: []cli.Flag{\n\t\t\t\t \tcli.StringFlag{\n\t\t\t\t \t\tName: \"scaffolding\", \n\t\t\t\t \t\tValue: \"\", \n\t\t\t\t \t\tUsage: \"Path to scaffolding file.\",\n\t\t\t\t \t},\n\t\t\t\t \tcli.StringFlag{\n\t\t\t\t \t\tName: \"key\", \n\t\t\t\t \t\tValue: \"\", \n\t\t\t\t \t\tUsage: \"Path to scaffolding key file.\",\n\t\t\t\t \t},\n\t\t\t\t \tcli.StringFlag{\n\t\t\t\t \t\tName: \"output\", \n\t\t\t\t \t\tValue: \"\", \n\t\t\t\t \t\tUsage: \"Output path for evaluation stats file.\",\n\t\t\t\t \t},\n\t\t\t\t},\n\t\t\t\tAction: evalScaffoldingCommand,\n \t\t},\n\t cli.Command{\n\t\t\t\tName: \"prep\",\n\t\t\t\tUsage: \"Generate a links file from a set of aligned Hi-C reads.\",\n\t\t\t\tFlags: []cli.Flag{\n\t\t\t\t \tcli.StringFlag{\n\t\t\t\t \t\tName: \"sam\", \n\t\t\t\t \t\tValue: \"\",\n\t\t\t\t \t\tUsage: \"Path to sam file.\",\n\t\t\t\t \t},\n\t\t\t\t \tcli.StringFlag{\n\t\t\t\t \t\tName: \"output\", \n\t\t\t\t \t\tValue: \"\", \n\t\t\t\t \t\tUsage: \"Output path for links file.\",\n\t\t\t\t \t},\n\t\t\t\t},\n\t\t\t\tAction: prepScaffoldingCommand,\n \t\t},\n \t},\n }\n}", "func NewCmdServiceCreate(name, fullName string) *cobra.Command {\n\to := NewServiceCreateOptions()\n\to.CmdFullName = fullName\n\tserviceCreateCmd := &cobra.Command{\n\t\tUse: name + \" <service_type> --plan <plan_name> [service_name]\",\n\t\tShort: createShortDesc,\n\t\tLong: createLongDesc,\n\t\tExample: fmt.Sprintf(createExample, fullName),\n\t\tArgs: cobra.RangeArgs(0, 2),\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\tgenericclioptions.GenericRun(o, cmd, args)\n\t\t},\n\t}\n\n\tif experimental.IsExperimentalModeEnabled() {\n\t\tserviceCreateCmd.Use += fmt.Sprintf(\" [flags]\\n %s <operator_type> --crd <crd_name> [service_name] [flags]\", o.CmdFullName)\n\t\tserviceCreateCmd.Example += fmt.Sprintf(\"\\n\\n\") + fmt.Sprintf(createOperatorExample, fullName)\n\t\tserviceCreateCmd.Flags().StringVar(&o.CustomResource, \"crd\", \"\", \"The name of the CRD of the operator to be used to create the service\")\n\t\tserviceCreateCmd.Flags().BoolVar(&o.DryRun, \"dry-run\", false, \"Print the yaml specificiation that will be used to create the service\")\n\t\t// remove this feature after enabling service create interactive mode for operator backed services\n\t\tserviceCreateCmd.Flags().StringVar(&o.fromFile, \"from-file\", \"\", \"Path to the file containing yaml specification to use to start operator backed service\")\n\t}\n\n\tserviceCreateCmd.Flags().StringVar(&o.Plan, \"plan\", \"\", \"The name of the plan of the service to be created\")\n\tserviceCreateCmd.Flags().StringArrayVarP(&o.parameters, \"parameters\", \"p\", []string{}, \"Parameters of the plan where a parameter is expressed as <key>=<value\")\n\tserviceCreateCmd.Flags().BoolVarP(&o.wait, \"wait\", \"w\", false, \"Wait until the service is ready\")\n\tgenericclioptions.AddContextFlag(serviceCreateCmd, &o.componentContext)\n\tcompletion.RegisterCommandHandler(serviceCreateCmd, completion.ServiceClassCompletionHandler)\n\tcompletion.RegisterCommandFlagHandler(serviceCreateCmd, \"plan\", completion.ServicePlanCompletionHandler)\n\tcompletion.RegisterCommandFlagHandler(serviceCreateCmd, \"parameters\", completion.ServiceParameterCompletionHandler)\n\treturn serviceCreateCmd\n}", "func (*CreateTenantRequest) Descriptor() ([]byte, []int) {\n\treturn file_packetbroker_api_iam_v1_service_proto_rawDescGZIP(), []int{11}\n}", "func NewCreateCmd() *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"create SUBCOMMAND\",\n\t\tShort: \"Create objects\",\n\t\tLong: `Create objects like users, etc.`,\n\t}\n}", "func New(buildRevision, buildTime string) *cobra.Command {\n\trootCmd := &cobra.Command{\n\t\tUse: \"iris-cli\",\n\t\tShort: \"Command Line Interface for Iris\",\n\t\tLong: `Iris CLI is a tool for Iris Web Framework.\nIt can be used to install starter kits and project structures \nComplete documentation is available at https://github.com/kataras/iris-cli`,\n\t\tSilenceErrors: true,\n\t\tSilenceUsage: true,\n\t\tTraverseChildren: true,\n\t\tSuggestionsMinimumDistance: 1,\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t},\n\t}\n\n\thelpTemplate := HelpTemplate{\n\t\tBuildRevision: buildRevision,\n\t\tBuildTime: buildTime,\n\t\tShowGoRuntimeVersion: true,\n\t}\n\trootCmd.SetHelpTemplate(helpTemplate.String())\n\n\t// Commands.\n\trootCmd.AddCommand(newCommand())\n\trootCmd.AddCommand(runCommand())\n\trootCmd.AddCommand(addCommand())\n\n\treturn rootCmd\n}", "func NewWithSubCmds(\n\tdescription string,\n\tversion string,\n\tcommand1 *command.Cmd,\n\tcommandsN ...*command.Cmd,\n) CLI {\n\tapp := app.App{\n\t\tName: path.Base(os.Args[0]),\n\t\tDescription: description,\n\t\tVersionString: version,\n\t\tHasSubCmds: true,\n\t}\n\treturn mkNew(app, append([]*command.Cmd{command1}, commandsN...)...)\n}", "func NewCreateCmd(globalFlags *flags.GlobalFlags) *cobra.Command {\n\tcmd := &CreateCmd{\n\t\tGlobalFlags: globalFlags,\n\t\tlog: log.GetInstance(),\n\t}\n\n\tcobraCmd := &cobra.Command{\n\t\tUse: \"create\",\n\t\tShort: \"Create a new virtual cluster\",\n\t\tLong: `\n#######################################################\n################### vcluster create ###################\n#######################################################\nCreates a new virtual cluster\n\nExample:\nvcluster create test --namespace test\n#######################################################\n\t`,\n\t\tArgs: cobra.ExactArgs(1),\n\t\tRunE: func(cobraCmd *cobra.Command, args []string) error {\n\t\t\t// Check for newer version\n\t\t\tupgrade.PrintNewerVersionWarning()\n\n\t\t\treturn cmd.Run(cobraCmd, args)\n\t\t},\n\t}\n\n\tcobraCmd.Flags().StringVar(&cmd.ChartVersion, \"chart-version\", upgrade.GetVersion(), \"The virtual cluster chart version to use\")\n\tcobraCmd.Flags().StringVar(&cmd.ChartName, \"chart-name\", \"vcluster\", \"The virtual cluster chart name to use\")\n\tcobraCmd.Flags().StringVar(&cmd.ChartRepo, \"chart-repo\", \"https://charts.loft.sh\", \"The virtual cluster chart repo to use\")\n\tcobraCmd.Flags().StringVar(&cmd.ReleaseValues, \"release-values\", \"\", \"Path where to load the virtual cluster helm release values from\")\n\tcobraCmd.Flags().StringVar(&cmd.K3SImage, \"k3s-image\", \"\", \"If specified, use this k3s image version\")\n\tcobraCmd.Flags().StringSliceVarP(&cmd.ExtraValues, \"extra-values\", \"f\", []string{}, \"Path where to load extra helm values from\")\n\tcobraCmd.Flags().BoolVar(&cmd.CreateNamespace, \"create-namespace\", true, \"If true the namespace will be created if it does not exist\")\n\tcobraCmd.Flags().BoolVar(&cmd.DisableIngressSync, \"disable-ingress-sync\", false, \"If true the virtual cluster will not sync any ingresses\")\n\tcobraCmd.Flags().BoolVar(&cmd.CreateClusterRole, \"create-cluster-role\", false, \"If true a cluster role will be created to access nodes, storageclasses and priorityclasses\")\n\tcobraCmd.Flags().BoolVar(&cmd.Expose, \"expose\", false, \"If true will create a load balancer service to expose the vcluster endpoint\")\n\tcobraCmd.Flags().BoolVar(&cmd.Connect, \"connect\", false, \"If true will run vcluster connect directly after the vcluster was created\")\n\treturn cobraCmd\n}", "func Command() cli.Command {\n\treturn cli.Command{\n\t\tName: \"token\",\n\t\tUsage: \"create tokens for connecting to the Smallstep API\",\n\t\tUsageText: \"step api token <subcommand> [arguments] [global-flags] [subcommand-flags]\",\n\t\tSubcommands: cli.Commands{\n\t\t\tcreateCommand(),\n\t\t},\n\t\tDescription: `**step api token** command group provides commands for creating the\ntokens required to connect to the Smallstep API.\n`,\n\t}\n}", "func newDestinationCmd(clientFn func() (*fic.ServiceClient, error), out io.Writer) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"destination <connection_id>\",\n\t\tShort: \"Show destination belonging to connection\",\n\t\tExample: \"fic router-to-port-connections destination F030123456789\",\n\t\tArgs: cobra.ExactArgs(1),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tclient, err := clientFn()\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"creating FIC client: %w\", err)\n\t\t\t}\n\n\t\t\tc, err := conn.Get(client, args[0]).Extract()\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"calling Show router to port connection API: %w\", err)\n\t\t\t}\n\n\t\t\tt := utils.NewTabby(out)\n\t\t\tif !c.Redundant {\n\t\t\t\tt.AddHeader(\"tenantId\", \"primary.portId\", \"primary.vlan\", \"primary.ipAddress\", \"primary.asn\")\n\t\t\t\tt.AddLine(\n\t\t\t\t\tc.TenantID,\n\t\t\t\t\tc.Destination.Primary.PortID, c.Destination.Primary.VLAN, c.Destination.Primary.IPAddress, c.Destination.Primary.ASN,\n\t\t\t\t)\n\t\t\t\tt.Print()\n\n\t\t\t\treturn nil\n\t\t\t}\n\n\t\t\tt.AddHeader(\n\t\t\t\t\"tenantId\",\n\t\t\t\t\"primary.portId\", \"primary.vlan\", \"primary.ipAddress\", \"primary.asn\",\n\t\t\t\t\"secondary.portId\", \"secondary.vlan\", \"secondary.ipAddress\", \"secondary.asn\",\n\t\t\t)\n\t\t\tt.AddLine(\n\t\t\t\tc.TenantID,\n\t\t\t\tc.Destination.Primary.PortID, c.Destination.Primary.VLAN, c.Destination.Primary.IPAddress, c.Destination.Primary.ASN,\n\t\t\t\tc.Destination.Secondary.PortID, c.Destination.Secondary.VLAN, c.Destination.Secondary.IPAddress, c.Destination.Secondary.ASN,\n\t\t\t)\n\t\t\tt.Print()\n\n\t\t\treturn nil\n\t\t},\n\t}\n\n\treturn cmd\n}", "func (c *Command) New() *cobra.Command {\n\tnewCommand := &cobra.Command{\n\t\tUse: \"annactl\",\n\t\tShort: \"Manage the API of the anna project. For more information see https://github.com/the-anna-project/annactl.\",\n\t\tLong: \"Manage the API of the anna project. For more information see https://github.com/the-anna-project/annactl.\",\n\t\tRun: c.Execute,\n\t}\n\n\tnewCommand.AddCommand(c.endpointCommand.New())\n\tnewCommand.AddCommand(c.versionCommand.New())\n\n\treturn newCommand\n}", "func GetCmdCreatePDV(cdc *codec.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"create [type] [pdv]\",\n\t\tShort: \"create PDV\",\n\t\tArgs: cobra.ExactArgs(2),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\t\t\tinBuf := bufio.NewReader(cmd.InOrStdin())\n\t\t\ttxBldr := auth.NewTxBuilderFromCLI(inBuf).WithTxEncoder(utils.GetTxEncoder(cdc))\n\n\t\t\tif hex.EncodeToString(cliCtx.GetFromAddress()) != strings.Split(args[0], \"-\")[0] { // Checks if the the msg sender is the same as the current owner\n\t\t\t\treturn fmt.Errorf(\"invalid owner\")\n\t\t\t}\n\n\t\t\tcaddr, _, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/cerberus-addr\", types.QuerierRoute), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"failed to get cerberus addr: %w\", err)\n\t\t\t}\n\n\t\t\texists, err := cerberusapi.NewClient(string(caddr), secp256k1.PrivKeySecp256k1{}).DoesPDVExist(cmd.Context(), args[0])\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"failed to check pdv existence: %w\", err)\n\t\t\t}\n\n\t\t\tif !exists {\n\t\t\t\treturn fmt.Errorf(\"pdv does not exist\")\n\t\t\t}\n\n\t\t\tt, err := strconv.Atoi(args[0])\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"failed to parse type: %w\", err)\n\t\t\t}\n\n\t\t\tmsg := types.NewMsgCreatePDV(uint64(time.Now().Unix()), args[1], types.PDVType(t), cliCtx.GetFromAddress())\n\t\t\tif err := msg.ValidateBasic(); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn utils.GenerateOrBroadcastMsgs(cliCtx, txBldr, []sdk.Msg{msg})\n\t\t},\n\t}\n}", "func (*CreateTenantResponse) Descriptor() ([]byte, []int) {\n\treturn file_packetbroker_api_iam_v1_service_proto_rawDescGZIP(), []int{12}\n}", "func setTenant(c *cli.Context) error {\n\terr := checkArgCount(c, 1)\n\tif err != nil {\n\t\treturn err\n\t}\n\tname := c.Args().First()\n\n\tclient.Photonclient, err = client.GetClient(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Ensure tenant exists\n\tid, err := findTenantID(name)\n\tif len(id) == 0 || err != nil {\n\t\treturn err\n\t}\n\n\tconfig, err := cf.LoadConfig()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tconfig.Tenant = &cf.TenantConfiguration{Name: name, ID: id}\n\terr = cf.SaveConfig(config)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = clearConfigProject(\"\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif !c.GlobalIsSet(\"non-interactive\") {\n\t\tfmt.Printf(\"Tenant set to '%s'\\n\", name)\n\t}\n\treturn nil\n}", "func newPlanCmd() *cobra.Command {\n\tpc := &planCmd{}\n\n\tplanCmd := &cobra.Command{\n\t\tUse: \"plan [-f SORUCE]\",\n\t\tShort: \"Generate and show an execution plan\",\n\t\tLong: planLong,\n\t\tExample: planExample,\n\t\tDisableFlagsInUseLine: true,\n\t\tSilenceUsage: true,\n\t\tSilenceErrors: true,\n\t\tArgs: cobra.MaximumNArgs(0),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tif err := pc.meta.init(args); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn pc.run(args)\n\t\t},\n\t}\n\n\tf := planCmd.Flags()\n\tf.BoolVar(&pc.destroy, \"destroy\", false, \"create plan to destroy resources\")\n\n\tpc.addMetaFlags(planCmd)\n\n\treturn planCmd\n}", "func newTenantMutation(c config, op Op) *TenantMutation {\n\treturn &TenantMutation{\n\t\tconfig: c,\n\t\top: op,\n\t\ttyp: TypeTenant,\n\t\tclearedFields: make(map[string]struct{}),\n\t}\n}", "func New(settings *environment.Settings) *cobra.Command {\n\treturn newCmd(settings, httpclient.New())\n}", "func CreateCommands(app *cli.Cli) {\n\n\t/* Functions to implement...\n\t/api/v1/go-kms/listkeys\n\t/api/v1/go-kms/createkey\n\t/api/v1/go-kms/generatedatakey\n\t/api/v1/go-kms/enablekey\n\t/api/v1/go-kms/disablekey\n\t/api/v1/go-kms/decrypt\n\t/api/v1/go-kms/encrypt*/\n\n\tapp.Command(\"keys\", \"Key functions\", func(commandCmd *cli.Cmd) {\n\t\tcommandCmd.Command(\"list\", \"List all available keys\", func(listKeysCmd *cli.Cmd) {\n\t\t\tlistKeysCmd.Action = func() {\n\n\t\t\t\t// List the key available...\n\t\t\t\tlistKeyRequest := kms.ListKeysRequest{}\n\n\t\t\t\tlistKeyResponse := &kms.ListKeysResponse{}\n\t\t\t\terr := Client.Do(\"POST\", \"/api/v1/go-kms/listkeys\", &listKeyRequest, listKeyResponse)\n\t\t\t\tif err != nil {\n\t\t\t\t\tExit(err.Error(), 1)\n\t\t\t\t}\n\n\t\t\t\tOutputMetadata(listKeyResponse.KeyMetadata)\n\t\t\t}\n\t\t})\n\t\tcommandCmd.Command(\"create\", \"Create a new key\", func(createKeyCmd *cli.Cmd) {\n\n\t\t\tdescription := createKeyCmd.StringOpt(\"d description\", \"\", \"Description for the new key\")\n\n\t\t\tcreateKeyCmd.Action = func() {\n\t\t\t\tcreateKeyRequest := kms.CreateKeyRequest{Description: *(description)}\n\n\t\t\t\tcreateKeyResponse := &kms.CreateKeyResponse{}\n\t\t\t\terr := Client.Do(\"POST\", \"/api/v1/go-kms/createkey\", &createKeyRequest, createKeyResponse)\n\t\t\t\tif err != nil {\n\t\t\t\t\tExit(err.Error(), 1)\n\t\t\t\t}\n\n\t\t\t\tOutputMetadata([]kms.KeyMetadata{createKeyResponse.KeyMetadata})\n\t\t\t}\n\t\t})\n\t\tcommandCmd.Command(\"disable\", \"Disable a key\", func(disableKeyCmd *cli.Cmd) {\n\n\t\t\tkeyID := disableKeyCmd.StringArg(\"KEYID\", \"\", \"The KeyID of the key to be disabled\")\n\n\t\t\tdisableKeyCmd.Action = func() {\n\t\t\t\tdisableKeyRequest := kms.DisableKeyRequest{KeyID: *(keyID)}\n\n\t\t\t\tdisableKeyResponse := &kms.DisableKeyResponse{}\n\t\t\t\terr := Client.Do(\"POST\", \"/api/v1/go-kms/disablekey\", &disableKeyRequest, disableKeyResponse)\n\t\t\t\tif err != nil {\n\t\t\t\t\tExit(err.Error(), 1)\n\t\t\t\t}\n\n\t\t\t\tOutputMetadata([]kms.KeyMetadata{disableKeyResponse.KeyMetadata})\n\n\t\t\t}\n\t\t})\n\t\tcommandCmd.Command(\"enable\", \"Enable a key\", func(enableKeyCmd *cli.Cmd) {\n\n\t\t\tkeyID := enableKeyCmd.StringArg(\"KEYID\", \"\", \"The KeyID of the key to be enabled\")\n\n\t\t\tenableKeyCmd.Action = func() {\n\t\t\t\tenableKeyRequest := kms.EnableKeyRequest{KeyID: *(keyID)}\n\n\t\t\t\tenableKeyResponse := &kms.EnableKeyResponse{}\n\t\t\t\terr := Client.Do(\"POST\", \"/api/v1/go-kms/enablekey\", &enableKeyRequest, enableKeyResponse)\n\t\t\t\tif err != nil {\n\t\t\t\t\tExit(err.Error(), 1)\n\t\t\t\t}\n\n\t\t\t\tOutputMetadata([]kms.KeyMetadata{enableKeyResponse.KeyMetadata})\n\t\t\t}\n\t\t})\n\n\t})\n}", "func Describe() cli.Command {\n\tcommand := cli.Command{\n\t\tName: \"describe\",\n\t\tAliases: []string{\"d\"},\n\t\tUsage: \"Describe a Supergiant resource.\",\n\t\tSubcommands: []cli.Command{\n\n\t\t\t// Describe Subcommands\n\t\t\t//List Apps\n\t\t\t{\n\t\t\t\tName: \"application\",\n\t\t\t\tAliases: []string{\"apps\", \"app\", \"applications\"},\n\t\t\t\tUsage: \"Describes Supergaint applications. Aliases: \\\"apps\\\", \\\"app\\\", \\\"applications\\\"\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\tsg, err := apictl.NewClient(\"\", \"\", \"\")\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tfmt.Println(\"ERROR:\", err)\n\t\t\t\t\t\tos.Exit(5)\n\t\t\t\t\t}\n\n\t\t\t\t\terr = sg.ListApps(\"\")\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tfmt.Println(\"ERROR:\", err)\n\t\t\t\t\t\tos.Exit(5)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t//List Components\n\t\t\t{\n\t\t\t\tName: \"component\",\n\t\t\t\tAliases: []string{\"components\", \"comp\"},\n\t\t\t\tUsage: \"Describes Supergiant application components. Aliases: \\\"components\\\", \\\"comp\\\"\",\n\t\t\t\tFlags: []cli.Flag{\n\t\t\t\t\tcli.StringFlag{\n\t\t\t\t\t\tName: \"app\",\n\t\t\t\t\t\tValue: \"\",\n\t\t\t\t\t\tUsage: \"Application context.\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\tsg, err := apictl.NewClient(\"\", \"\", \"\")\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tfmt.Println(\"ERROR:\", err)\n\t\t\t\t\t\tos.Exit(5)\n\t\t\t\t\t}\n\n\t\t\t\t\terr = sg.ComponentDetails(\n\t\t\t\t\t\tgetApp(c),\n\t\t\t\t\t\tc.Args().First(),\n\t\t\t\t\t)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tfmt.Println(\"ERROR:\", err)\n\t\t\t\t\t\tos.Exit(5)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t//List Entrypoints\n\t\t\t{\n\t\t\t\tName: \"entrypoint\",\n\t\t\t\tAliases: []string{\"entrypoints\", \"entry\", \"loadbalancer\", \"lb\"},\n\t\t\t\tUsage: \"Describes Supergiant entrypoints. Aliases: \\\"entrypoints\\\", \\\"entry\\\", \\\"loadbalancer\\\", \\\"lb\\\"\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\tsg, err := apictl.NewClient(\"\", \"\", \"\")\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tfmt.Println(\"ERROR:\", err)\n\t\t\t\t\t\tos.Exit(5)\n\t\t\t\t\t}\n\n\t\t\t\t\terr = sg.ListEntryPoints(c.Args().First())\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tfmt.Println(\"ERROR:\", err)\n\t\t\t\t\t\tos.Exit(5)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t// End spacetime actions.\n\t\t},\n\t}\n\n\treturn command\n}", "func NewServiceAccountCommand(f *factory.Factory) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: f.Localizer.LocalizeByID(\"serviceAccount.cmd.use\"),\n\t\tShort: f.Localizer.LocalizeByID(\"serviceAccount.cmd.shortDescription\"),\n\t\tLong: f.Localizer.LocalizeByID(\"serviceAccount.cmd.longDescription\"),\n\t\tArgs: cobra.ExactArgs(1),\n\t}\n\n\tcmd.AddCommand(\n\t\tcreate.NewCreateCommand(f),\n\t\tlist.NewListCommand(f),\n\t\tdelete.NewDeleteCommand(f),\n\t\tresetcredentials.NewResetCredentialsCommand(f),\n\t\tdescribe.NewDescribeCommand(f),\n\t)\n\n\treturn cmd\n}", "func newGetCmd(out io.Writer) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"get\",\n\t\tShort: \"The command to get choerodon resource\",\n\t\tLong: `The command to get choerodon resource.such as organization, project, app, instance.`,\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\tc7nclient.InitClient(&clientConfig, &clientPlatformConfig)\n\t\t\terror := c7nclient.Client.CheckIsLogin()\n\t\t\tif error != nil {\n\t\t\t\tfmt.Println(error)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif len(args) > 0 {\n\t\t\t\tfmt.Printf(\"don't have the resource %s, you can user c7nctl get --help to see the resource you can use!\", args[0])\n\t\t\t} else {\n\t\t\t\tcmd.Help()\n\t\t\t}\n\t\t},\n\t}\n\n\tcmd.AddCommand(newGetEnvCmd(out))\n\tcmd.AddCommand(newGetInstanceCmd(out))\n\tcmd.AddCommand(newGetProCmd(out))\n\tcmd.AddCommand(newGetOrgCmd(out))\n\tcmd.AddCommand(newGetServiceCmd(out))\n\tcmd.AddCommand(newGetIngressCmd(out))\n\tcmd.AddCommand(NewGetAppVersionCmd(out))\n\tcmd.AddCommand(newGetAppCmd(out))\n\tcmd.AddCommand(newGetClusterNodeCmd(out))\n\tcmd.AddCommand(newGetClusterCmd(out))\n\tcmd.AddCommand(newGetValueCmd(out))\n\tcmd.AddCommand(newGetCertCmd(out))\n\tcmd.AddCommand(newGetConfigMapCmd(out))\n\tcmd.AddCommand(newGetSecretCmd(out))\n\tcmd.AddCommand(newGetCustomCmd(out))\n\tcmd.AddCommand(newGetPvcCmd(out))\n\tcmd.AddCommand(newGetPvCmd(out))\n\n\treturn cmd\n}", "func NewCommand() *cobra.Command {\n\tkp := &KubebuilderProject{}\n\n\treturn framework.TemplateCommand{\n\t\tAPI: kp,\n\n\t\tMergeResources: true, // apply additional inputs as patches\n\n\t\t// these are run before the templates\n\t\tPreProcessFilters: []kio.Filter{\n\t\t\t// run controller-gen libraries to generate configuration from code\n\t\t\tControllerGenFilter{KubebuilderProject: kp},\n\t\t\t// inject generated certificates\n\t\t\tCertFilter{KubebuilderProject: kp},\n\t\t},\n\n\t\t// generate resources\n\t\tTemplatesFn: framework.TemplatesFromDir(pkger.Dir(\"/config-gen/templates/resources\")),\n\n\t\t// patch resources\n\t\tPatchTemplatesFn: framework.PatchTemplatesFromDir(\n\t\t\tCRDPatchTemplate(kp),\n\t\t\tCertManagerPatchTemplate(kp),\n\t\t\tControllerManagerPatchTemplate(kp),\n\t\t),\n\n\t\t// perform final modifications\n\t\tPostProcessFilters: []kio.Filter{\n\t\t\t// sort the resources\n\t\t\tComponentFilter{KubebuilderProject: kp},\n\t\t\tSortFilter{KubebuilderProject: kp},\n\t\t},\n\t}.GetCommand()\n}", "func NewCmdDiskCreate() *cobra.Command {\n\tvar async *bool\n\tvar count *int\n\treq := base.BizClient.NewCreateUDiskRequest()\n\tenableDataArk := sdk.String(\"false\")\n\tcmd := &cobra.Command{\n\t\tUse: \"create\",\n\t\tShort: \"Create udisk instance\",\n\t\tLong: \"Create udisk instance\",\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\tif *count > 10 || *count < 1 {\n\t\t\t\tbase.Cxt.Printf(\"Error, count should be between 1 and 10\\n\")\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif *enableDataArk == \"true\" {\n\t\t\t\treq.UDataArkMode = sdk.String(\"Yes\")\n\t\t\t} else {\n\t\t\t\treq.UDataArkMode = sdk.String(\"No\")\n\t\t\t}\n\n\t\t\tif *req.DiskType == \"Oridinary\" {\n\t\t\t\t*req.DiskType = \"DataDisk\"\n\t\t\t} else if *req.DiskType == \"SSD\" {\n\t\t\t\t*req.DiskType = \"SSDDataDisk\"\n\t\t\t}\n\t\t\tfor i := 0; i < *count; i++ {\n\t\t\t\tresp, err := base.BizClient.CreateUDisk(req)\n\t\t\t\tif err != nil {\n\t\t\t\t\tbase.HandleError(err)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tif count := len(resp.UDiskId); count == 1 {\n\t\t\t\t\ttext := fmt.Sprintf(\"udisk:%v is initializing\", resp.UDiskId)\n\t\t\t\t\tif *async {\n\t\t\t\t\t\tbase.Cxt.Println(text)\n\t\t\t\t\t} else {\n\t\t\t\t\t\tpollDisk(resp.UDiskId[0], *req.ProjectId, *req.Region, *req.Zone, text, []string{status.DISK_AVAILABLE, status.DISK_FAILED})\n\t\t\t\t\t}\n\t\t\t\t} else if count > 1 {\n\t\t\t\t\tbase.Cxt.Printf(\"udisk:%v created\\n\", resp.UDiskId)\n\t\t\t\t} else {\n\t\t\t\t\tbase.Cxt.PrintErr(fmt.Errorf(\"none udisk created\"))\n\t\t\t\t}\n\t\t\t}\n\t\t},\n\t}\n\tflags := cmd.Flags()\n\tflags.SortFlags = false\n\treq.Name = flags.String(\"name\", \"\", \"Required. Name of the udisk to create\")\n\treq.Size = flags.Int(\"size-gb\", 10, \"Required. Size of the udisk to create. Unit:GB. Normal udisk [1,8000]; SSD udisk [1,4000] \")\n\treq.ProjectId = flags.String(\"project-id\", base.ConfigInstance.ProjectID, \"Optional. Assign project-id\")\n\treq.Region = flags.String(\"region\", base.ConfigInstance.Region, \"Optional. Assign region\")\n\treq.Zone = flags.String(\"zone\", base.ConfigInstance.Zone, \"Optional. Assign availability zone\")\n\treq.ChargeType = flags.String(\"charge-type\", \"Dynamic\", \"Optional.'Year',pay yearly;'Month',pay monthly;'Dynamic', pay hourly\")\n\treq.Quantity = flags.Int(\"quantity\", 1, \"Optional. The duration of the instance. N years/months.\")\n\tenableDataArk = flags.String(\"enable-data-ark\", \"false\", \"Optional. DataArk supports real-time backup, which can restore the udisk back to any moment within the last 12 hours.\")\n\treq.Tag = flags.String(\"group\", \"Default\", \"Optional. Business group\")\n\treq.DiskType = flags.String(\"udisk-type\", \"Oridinary\", \"Optional. 'Ordinary' or 'SSD'\")\n\treq.CouponId = flags.String(\"coupon-id\", \"\", \"Optional. Coupon ID, The Coupon can deduct part of the payment.See https://accountv2.ucloud.cn\")\n\tasync = flags.Bool(\"async\", false, \"Optional. Do not wait for the long-running operation to finish.\")\n\tcount = flags.Int(\"count\", 1, \"Optional. The count of udisk to create. Range [1,10]\")\n\n\tflags.SetFlagValues(\"charge-type\", \"Month\", \"Year\", \"Dynamic\", \"Trial\")\n\tflags.SetFlagValues(\"enable-data-ark\", \"true\", \"false\")\n\tflags.SetFlagValues(\"udisk-type\", \"Oridinary\", \"SSD\")\n\n\tcmd.MarkFlagRequired(\"size-gb\")\n\tcmd.MarkFlagRequired(\"name\")\n\n\treturn cmd\n}", "func ExampleClient_CreateCommand() {\n\tclient := &Client{}\n\tctx := context.Background()\n\n\tnewCommand := &Command{\n\t\tName: \"my-command\",\n\t\tDescription: \"my command\",\n\t\tArgs: \"[@username]\",\n\t\tSet: \"custom_cmd_set\",\n\t}\n\n\t_, _ = client.CreateCommand(ctx, newCommand)\n}", "func topLevelCmd(use, short string) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: use,\n\t\tShort: short,\n\t\tDisableFlagParsing: false,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: validateCmd,\n\t}\n}", "func NewCommand() (cmd *cobra.Command) {\n\tcmd = &cobra.Command{Use: \"binctl\", Run: func(cmd *cobra.Command, args []string) {\n\t\t//fmt.Println(\"Operate pipeline\")\n\t\t//fmt.Println(\"init binctl\")\n\t}}\n\tcmd.AddCommand(cmdMemory())\n\tcmd.AddCommand(cmdPipeline())\n\tcmd.PersistentFlags().String(\"config\", \"./configs/binlogo.yaml\", \"config file default is ./configs/binlogo.yaml\")\n\terr := viper.BindPFlag(\"config\", cmd.PersistentFlags().Lookup(\"config\"))\n\t// configs.InitViperFromFile(viper.GetString(\"config\"))\n\tconfigs.Init(viper.GetString(\"config\"))\n\t//etcd2.DefaultETCD()\n\tblog.Env(configs.Env(viper.GetString(\"env\")))\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\treturn\n}", "func chainCommand[C chains.Config, R presenters.ChainResource[C], P TableRenderer, P2 ~[]P](\n\ttyp string, client ChainClient[C, R, P, P2], chainID cli.Flag) cli.Command {\n\tif flagName := chainID.GetName(); flagName != \"id\" {\n\t\tpanic(fmt.Errorf(\"chainID flag name must be 'id', got: %s\", flagName))\n\t}\n\tlower := strings.ToLower(typ)\n\treturn cli.Command{\n\t\tName: lower,\n\t\tUsage: fmt.Sprintf(\"Commands for handling %s chains\", typ),\n\t\tSubcommands: cli.Commands{\n\t\t\t{\n\t\t\t\tName: \"create\",\n\t\t\t\tUsage: fmt.Sprintf(\"Create a new %s chain\", typ),\n\t\t\t\tAction: client.CreateChain,\n\t\t\t\tFlags: []cli.Flag{chainID},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"delete\",\n\t\t\t\tUsage: fmt.Sprintf(\"Delete an existing %s chain\", typ),\n\t\t\t\tAction: client.RemoveChain,\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"list\",\n\t\t\t\tUsage: fmt.Sprintf(\"List all existing %s chains\", typ),\n\t\t\t\tAction: client.IndexChains,\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"configure\",\n\t\t\t\tUsage: fmt.Sprintf(\"Configure an existing %s chain\", typ),\n\t\t\t\tAction: client.ConfigureChain,\n\t\t\t\tFlags: []cli.Flag{chainID},\n\t\t\t},\n\t\t},\n\t}\n}", "func CreateTxCmd(cdc *wire.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"create <ticker> <name> <decimals> <reserve_ratio> <initial_supply> <initial_rune_amount>\",\n\t\tShort: \"Create a token with CLP\",\n\t\tArgs: cobra.ExactArgs(6),\n\t\tRunE: func(_ *cobra.Command, args []string) error {\n\t\t\ttxCtx := authctx.NewTxContextFromCLI().WithCodec(cdc)\n\t\t\tcliCtx := context.NewCLIContext().\n\t\t\t\tWithCodec(cdc).\n\t\t\t\tWithLogger(os.Stdout).\n\t\t\t\tWithAccountDecoder(authcmd.GetAccountDecoder(cdc))\n\n\t\t\t// get the from address from the name flag\n\t\t\tfrom, err := cliCtx.GetFromAddress()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\t// create the message\n\t\t\tticker := args[0]\n\t\t\tname := args[1]\n\t\t\tdecimalsInt, _ := strconv.Atoi(args[2])\n\t\t\treserveRatio, _ := strconv.Atoi(args[3])\n\n\t\t\tif decimalsInt < 0 || decimalsInt > 255 {\n\t\t\t\treturn clp.ErrInvalidDecimals(clp.DefaultCodespace)\n\t\t\t}\n\n\t\t\tdecimals := uint8(decimalsInt)\n\n\t\t\tinitialSupply, _ := strconv.Atoi(args[4])\n\t\t\tinitialBaseCoinAmount, _ := strconv.Atoi(args[5])\n\t\t\tmsg := clpTypes.NewMsgCreate(from, ticker, name, decimals, reserveRatio, int64(initialSupply), int64(initialBaseCoinAmount))\n\n\t\t\t// Build and sign the transaction, then broadcast to a Tendermint\n\t\t\t// node.\n\t\t\treturn utils.SendTx(txCtx, cliCtx, []sdk.Msg{msg})\n\t\t},\n\t}\n}", "func NewCmdNamespaceCreate(name, fullName string, testClientset clientset.Clientset) *cobra.Command {\n\to := NewNamespaceCreateOptions()\n\t// To help the UI messages deal better with namespace vs project\n\to.commandName = name\n\tif len(os.Args) > 2 {\n\t\to.commandName = os.Args[2]\n\t}\n\n\tnamespaceCreateCmd := &cobra.Command{\n\t\tUse: name,\n\t\tShort: createShortDesc,\n\t\tLong: createLongDesc,\n\t\tExample: fmt.Sprintf(createExample, fullName),\n\t\tArgs: cobra.ExactArgs(1),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\treturn genericclioptions.GenericRun(o, testClientset, cmd, args)\n\t\t},\n\t\tAliases: []string{\"project\"},\n\t}\n\n\tnamespaceCreateCmd.Flags().BoolVarP(&o.waitFlag, \"wait\", \"w\", false, \"Wait until the namespace is ready\")\n\n\tclientset.Add(namespaceCreateCmd, clientset.KUBERNETES, clientset.PROJECT, clientset.PREFERENCE)\n\tutil.SetCommandGroup(namespaceCreateCmd, util.MainGroup)\n\n\treturn namespaceCreateCmd\n}", "func CreateCommand(s string) Command {\n\tfmt.Println(\"Creating command\")\n\n\treturn &ConsoleOutput{\n\t\tmessage: s,\n\t}\n}", "func GetTxCmd(cdc *codec.Codec) *cobra.Command {\n\tmyappTxCmd := &cobra.Command{\n\t\tUse: types.ModuleName,\n\t\tShort: fmt.Sprintf(\"%s transactions subcommands\", types.ModuleName),\n\t\tDisableFlagParsing: true,\n\t\tSuggestionsMinimumDistance: 2,\n\t\tRunE: client.ValidateCmd,\n\t}\n\n\tmyappTxCmd.AddCommand(flags.PostCommands(\n // this line is used by starport scaffolding # 1\n\t\tGetCmdCreateStudent(cdc),\n\t\tGetCmdSetStudent(cdc),\n\t\tGetCmdDeleteStudent(cdc),\n\t\tGetCmdCreatePost(cdc),\n\t\tGetCmdSetPost(cdc),\n\t\tGetCmdDeletePost(cdc),\n\t)...)\n\n\treturn myappTxCmd\n}", "func RegisterCommands(app *cobra.Command, c *client.Client) {\n\tvar command, sub *cobra.Command\n\tcommand = &cobra.Command{\n\t\tUse: \"create\",\n\t\tShort: `Create new hero`,\n\t}\n\ttmp1 := new(CreateHeroCommand)\n\tsub = &cobra.Command{\n\t\tUse: `hero [\"/heroes\"]`,\n\t\tShort: ``,\n\t\tLong: `\n\nPayload example:\n\n{\n \"name\": \"Voluptatem illum.\"\n}`,\n\t\tRunE: func(cmd *cobra.Command, args []string) error { return tmp1.Run(c, args) },\n\t}\n\ttmp1.RegisterFlags(sub, c)\n\tsub.PersistentFlags().BoolVar(&tmp1.PrettyPrint, \"pp\", false, \"Pretty print response body\")\n\tcommand.AddCommand(sub)\n\tapp.AddCommand(command)\n\tcommand = &cobra.Command{\n\t\tUse: \"list\",\n\t\tShort: `Retrieve all heroes`,\n\t}\n\ttmp2 := new(ListHeroCommand)\n\tsub = &cobra.Command{\n\t\tUse: `hero [\"/heroes\"]`,\n\t\tShort: ``,\n\t\tRunE: func(cmd *cobra.Command, args []string) error { return tmp2.Run(c, args) },\n\t}\n\ttmp2.RegisterFlags(sub, c)\n\tsub.PersistentFlags().BoolVar(&tmp2.PrettyPrint, \"pp\", false, \"Pretty print response body\")\n\tcommand.AddCommand(sub)\n\tapp.AddCommand(command)\n\tcommand = &cobra.Command{\n\t\tUse: \"show\",\n\t\tShort: `Get hero by id`,\n\t}\n\ttmp3 := new(ShowHeroCommand)\n\tsub = &cobra.Command{\n\t\tUse: `hero [\"/heroes/HEROID\"]`,\n\t\tShort: ``,\n\t\tRunE: func(cmd *cobra.Command, args []string) error { return tmp3.Run(c, args) },\n\t}\n\ttmp3.RegisterFlags(sub, c)\n\tsub.PersistentFlags().BoolVar(&tmp3.PrettyPrint, \"pp\", false, \"Pretty print response body\")\n\tcommand.AddCommand(sub)\n\tapp.AddCommand(command)\n}", "func CreateChannelCmd(cdc *wire.Codec) *cobra.Command {\n\tflagTo := \"to\"\n\tflagCoins := \"amount\"\n\n\tcmd := &cobra.Command{\n\t\tUse: \"create\",\n\t\tShort: \"Create a new payment channel\",\n\t\tLong: \"Create a new unidirectional payment channel from a local address to a remote address, funded with some amount of coins. These coins are removed from the sender account and put into the channel.\",\n\t\tArgs: cobra.NoArgs,\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\n\t\t\t// Create a \"client context\" stuct populated with info from common flags\n\t\t\tctx := context.NewCoreContextFromViper().WithDecoder(authcmd.GetAccountDecoder(cdc))\n\t\t\t// TODO is this needed for channelID\n\t\t\t// ctx.PrintResponse = true\n\n\t\t\t// Get sender adress\n\t\t\tsender, err := ctx.GetFromAddress()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\t// Get receiver address\n\t\t\ttoStr := viper.GetString(flagTo)\n\t\t\treceiver, err := sdk.AccAddressFromBech32(toStr)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\t// Get channel funding amount\n\t\t\tcoinsString := viper.GetString(flagCoins)\n\t\t\tcoins, err := sdk.ParseCoins(coinsString)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\t// Create the create channel msg to send\n\t\t\tmsg := paychan.MsgCreate{\n\t\t\t\tParticipants: [2]sdk.AccAddress{sender, receiver},\n\t\t\t\tCoins: coins,\n\t\t\t}\n\t\t\terr = msg.ValidateBasic()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\t// Build and sign the transaction, then broadcast to the blockchain\n\t\t\terr = ctx.EnsureSignBuildBroadcast(ctx.FromAddressName, []sdk.Msg{msg}, cdc)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn nil\n\t\t},\n\t}\n\tcmd.Flags().String(flagTo, \"\", \"Recipient address of the payment channel.\")\n\tcmd.Flags().String(flagCoins, \"\", \"Amount of coins to fund the payment channel with.\")\n\treturn cmd\n}", "func NewCreateServiceCommand(p *config.KfParams, client servicecatalogclient.Interface) *cobra.Command {\n\tvar (\n\t\tconfigAsJSON string\n\t\tbroker string\n\t)\n\n\tcreateCmd := &cobra.Command{\n\t\tUse: \"create-service SERVICE PLAN SERVICE_INSTANCE [-c PARAMETERS_AS_JSON] [-b service-broker]\",\n\t\tAliases: []string{\"cs\"},\n\t\tShort: \"Create a service instance\",\n\t\tExample: `\n # Creates a new instance of a db-service with the name mydb, plan silver, and provisioning configuration\n kf create-service db-service silver mydb -c '{\"ram_gb\":4}'\n\n # Creates a new instance of a db-service from the broker named local-broker\n kf create-service db-service silver mydb -c ~/workspace/tmp/instance_config.json -b local-broker`,\n\t\tArgs: cobra.ExactArgs(3),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tserviceName := args[0]\n\t\t\tplanName := args[1]\n\t\t\tinstanceName := args[2]\n\n\t\t\tcmd.SilenceUsage = true\n\n\t\t\tif err := utils.ValidateNamespace(p); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tparams, err := services.ParseJSONOrFile(configAsJSON)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tparamBytes, err := json.Marshal(params)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\trawParams := &runtime.RawExtension{\n\t\t\t\tRaw: paramBytes,\n\t\t\t}\n\n\t\t\tmatchingClusterPlans, err := findMatchingClusterPlans(client, planName, serviceName, broker)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tif len(matchingClusterPlans) != 0 {\n\n\t\t\t\t// plan found\n\t\t\t\tcreated, err := createServiceInstance(client, p, rawParams, serviceName, planName, instanceName, broker, false)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tfmt.Fprintf(cmd.OutOrStdout(), \"Creating service instance %q %s\", instanceName, utils.AsyncLogSuffix)\n\t\t\t\tdescribe.ServiceInstance(cmd.OutOrStdout(), created)\n\t\t\t\treturn nil\n\t\t\t}\n\n\t\t\tmatchingNamespacedPlans, err := findMatchingNamespacedPlans(client, p.Namespace, planName, serviceName, broker)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tif len(matchingNamespacedPlans) != 0 {\n\n\t\t\t\t// plan found\n\t\t\t\tcreated, err := createServiceInstance(client, p, rawParams, serviceName, planName, instanceName, broker, true)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\tfmt.Fprintf(cmd.OutOrStdout(), \"Creating service instance %q %s\", instanceName, utils.AsyncLogSuffix)\n\t\t\t\tdescribe.ServiceInstance(cmd.OutOrStdout(), created)\n\t\t\t\treturn nil\n\t\t\t}\n\n\t\t\tif broker != \"\" {\n\t\t\t\treturn fmt.Errorf(\"no plan %s found for class %s for the service-broker %s\", planName, serviceName, broker)\n\t\t\t} else {\n\t\t\t\treturn fmt.Errorf(\"no plan %s found for class %s for all service-brokers\", planName, serviceName)\n\t\t\t}\n\t\t},\n\t}\n\n\tcreateCmd.Flags().StringVarP(\n\t\t&configAsJSON,\n\t\t\"config\",\n\t\t\"c\",\n\t\t\"{}\",\n\t\t\"Valid JSON object containing service-specific configuration parameters, provided in-line or in a file.\")\n\n\tcreateCmd.Flags().StringVarP(\n\t\t&broker,\n\t\t\"broker\",\n\t\t\"b\",\n\t\t\"\",\n\t\t\"Service broker to use.\")\n\n\treturn createCmd\n}", "func Command() *cobra.Command {\n\trootCmd := &cobra.Command{\n\t\tUse: \"generator\",\n\t\tPersistentPreRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tif err := os.Setenv(\"NODE_OPTIONS\", \"--max-old-space-size=8192\"); err != nil {\n\t\t\t\treturn fmt.Errorf(\"failed to set environment variable: %v\", err)\n\t\t\t}\n\t\t\tlog.SetFlags(0) // remove the time stamp prefix\n\t\t\treturn nil\n\t\t},\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\treturn fmt.Errorf(\"please specify a subcommand\")\n\t\t},\n\t\tHidden: true,\n\t}\n\n\t//bind global flags\n\tcommon.BindGlobalFlags(rootCmd.PersistentFlags())\n\n\trootCmd.AddCommand(\n\t\tautomation.Command(),\n\t\tissue.Command(),\n\t\ttemplate.Command(),\n\t\trefresh.Command(),\n\t\trelease.Command(),\n\t\tautomation_v2.Command(),\n\t\trelease_v2.Command(),\n\t\trefresh_v2.Command(),\n\t\treadme.Command(),\n\t)\n\n\treturn rootCmd\n}", "func newServer() *cobra.Command {\n\tvar cluster []string\n\tvar dir string\n\tvar address string\n\tvar dbName string\n\tvar role string\n\tvar id, port int\n\tvar skip bool\n\tvar timeout time.Duration\n\n\tcmd := &cobra.Command{\n\t\tUse: \"server\",\n\t\tShort: \"Start a server with web api.\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tctx, cancel := context.WithTimeout(context.Background(), timeout)\n\t\t\tdefer cancel()\n\t\t\t//func StartServer(ctx context.Context, id int, dir, address, web string, cluster []string) error {\n\t\t\t//err := StartServer(ctx, id, port, skip, dbName, dir, address, role, cluster)\n\t\t\tcluster = omit(address, cluster)\n\t\t\tkp := &globalKeys\n\t\t\terr := StartServer(ctx, id, port, kp, dir, address, cluster)\n\t\t\tlog.Println(\"server is done serving:\", err)\n\t\t\treturn nil\n\t\t},\n\t}\n\n\tflags := cmd.Flags()\n\tflags.StringVarP(&address, \"address\", \"a\", envy.StringDefault(\"DQLITED_ADDRESS\", \"127.0.0.1:9181\"), \"address of the node (default is 127.0.0.1:918<ID>)\")\n\tflags.StringSliceVarP(&cluster, \"cluster\", \"c\", clusterList(), \"addresses of existing cluster nodes\")\n\tflags.StringVarP(&dbName, \"database\", \"d\", envy.StringDefault(\"DQLITED_DB\", defaultDatabase), \"name of database to use\")\n\tflags.StringVarP(&dir, \"dir\", \"l\", envy.StringDefault(\"DQLITED_TMP\", \"/tmp/dqlited\"), \"database working directory\")\n\tflags.StringVarP(&role, \"role\", \"r\", envy.StringDefault(\"DQLITED_ROLE\", \"voter\"), \"node role, must be one of: 'voter', 'standby', or 'spare'\")\n\tflags.IntVarP(&id, \"id\", \"i\", envy.IntDefault(\"DQLITED_ID\", 1), \"server id\")\n\tflags.IntVarP(&port, \"port\", \"p\", envy.IntDefault(\"DQLITED_PORT\", 4001), \"port to serve traffic on\")\n\tflags.BoolVarP(&skip, \"skip\", \"s\", envy.Bool(\"DQLITED_SKIP\"), \"do NOT add server to cluster\")\n\tflags.DurationVarP(&timeout, \"timeout\", \"t\", time.Minute*5, \"time to wait for connection to complete\")\n\n\treturn cmd\n}", "func New(appName string) (*ffcli.Command, *Config) {\n\tcfg := Config{\n\t\tAppName: appName,\n\t}\n\n\tfs := flag.NewFlagSet(appName, flag.ExitOnError)\n\tcfg.RegisterFlags(fs)\n\n\treturn &ffcli.Command{\n\t\tShortUsage: appName + \" [flags] <subcommand> [flags] [<arg>...]\",\n\t\tFlagSet: fs,\n\t\tExec: cfg.Exec,\n\t}, &cfg\n}", "func (s *Start) CreateStartCommand() *cobra.Command {\n\tstartCmd := &cobra.Command{\n\t\tUse: \"start\",\n\t\tShort: \"Start horusec-cli\",\n\t\tLong: \"Start the Horusec' analysis in the current path\",\n\t\tExample: \"horusec start\",\n\t\tPersistentPreRunE: s.configs.PersistentPreRun,\n\t\tRunE: s.runE,\n\t}\n\n\tstartCmd.PersistentFlags().\n\t\tInt64P(\n\t\t\t\"monitor-retry-count\", \"m\",\n\t\t\ts.configs.MonitorRetryInSeconds,\n\t\t\t\"The number of retries for the monitor.\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tStringP(\n\t\t\t\"output-format\", \"o\",\n\t\t\ts.configs.PrintOutputType,\n\t\t\t\"The format for the output to be shown. Options are: text (stdout), json, sonarqube\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tStringSliceP(\n\t\t\t\"ignore-severity\", \"s\",\n\t\t\ts.configs.SeveritiesToIgnore,\n\t\t\t\"The level of vulnerabilities to ignore in the output. Example: -s=\\\"LOW, MEDIUM, HIGH\\\"\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tStringP(\n\t\t\t\"json-output-file\", \"O\",\n\t\t\ts.configs.JSONOutputFilePath,\n\t\t\t\"If your pass output-format you can configure the output JSON location. Example: -O=\\\"/tmp/output.json\\\"\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tStringSliceP(\n\t\t\t\"ignore\", \"i\",\n\t\t\ts.configs.FilesOrPathsToIgnore,\n\t\t\t\"Paths to ignore in the analysis. Example: -i=\\\"/home/user/project/assets, /home/user/project/deployments\\\"\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tStringP(\n\t\t\t\"horusec-url\", \"u\",\n\t\t\ts.configs.HorusecAPIUri,\n\t\t\t\"The Horusec API address to access the analysis engine\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tInt64P(\n\t\t\t\"request-timeout\", \"r\",\n\t\t\ts.configs.TimeoutInSecondsRequest,\n\t\t\t\"The timeout threshold for the request to the Horusec API\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tInt64P(\n\t\t\t\"analysis-timeout\", \"t\",\n\t\t\ts.configs.TimeoutInSecondsAnalysis,\n\t\t\t\"The timeout threshold for the Horusec CLI wait for the analysis to complete.\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tStringP(\n\t\t\t\"authorization\", \"a\",\n\t\t\ts.configs.RepositoryAuthorization,\n\t\t\t\"The authorization token for the Horusec API\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tStringToString(\n\t\t\t\"headers\",\n\t\t\ts.configs.Headers,\n\t\t\t\"The headers dynamic to send on request in Horusec API. Example --headers=\\\"{\\\"X-Auth-Service\\\": \\\"my-value\\\"}\\\"\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tBoolP(\n\t\t\t\"return-error\", \"e\",\n\t\t\ts.configs.ReturnErrorIfFoundVulnerability,\n\t\t\t\"The return-error is the option to check if you can return \\\"exit(1)\\\" if found vulnerabilities. Example -e=\\\"true\\\"\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tStringP(\n\t\t\t\"project-path\", \"p\",\n\t\t\ts.configs.ProjectPath,\n\t\t\t\"Path to run an analysis in your project\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tBool(\n\t\t\t\"enable-git-history\",\n\t\t\ts.configs.EnableGitHistoryAnalysis,\n\t\t\t\"When this value is \\\"true\\\" we will run tool gitleaks and search vulnerability in all git history of the project. Example --enable-git-history=\\\"true\\\"\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tBoolP(\n\t\t\t\"insecure-skip-verify\", \"S\",\n\t\t\ts.configs.CertInsecureSkipVerify,\n\t\t\t\"Insecure skip verify cert authority. PLEASE, try not to use it. Example -S=\\\"true\\\"\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tStringP(\n\t\t\t\"certificate-path\", \"C\",\n\t\t\ts.configs.CertPath,\n\t\t\t\"Path to certificate of authority. Example -C=\\\"/example/ca.crt\\\"\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tBoolP(\n\t\t\t\"enable-commit-author\", \"G\",\n\t\t\ts.configs.EnableCommitAuthor,\n\t\t\t\"Used to enable or disable search with vulnerability author. Example -G=\\\"true\\\"\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tStringP(\n\t\t\t\"repository-name\", \"n\",\n\t\t\ts.configs.RepositoryName,\n\t\t\t\"Used to send repository name to horus server. Example -n=\\\"horus\\\"\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tStringSliceP(\n\t\t\t\"false-positive\", \"F\",\n\t\t\ts.configs.FalsePositiveHashes,\n\t\t\t\"Used to ignore a vulnerability by hash and setting it to be of the false positive type. Example -F=\\\"hash1, hash2\\\"\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tStringSliceP(\n\t\t\t\"risk-accept\", \"R\",\n\t\t\ts.configs.RiskAcceptHashes,\n\t\t\t\"Used to ignore a vulnerability by hash and setting it to be of the risk accept type. Example -R=\\\"hash3, hash4\\\"\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tStringP(\n\t\t\t\"container-bind-project-path\", \"P\",\n\t\t\ts.configs.ContainerBindProjectPath,\n\t\t\t\"Used to pass project path in host when running horusec cli inside a container.\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tStringP(\n\t\t\t\"custom-rules-path\", \"c\",\n\t\t\ts.configs.CustomRulesPath,\n\t\t\t\"Used to pass the path to the horusec custom rules file. Example: -c=\\\"./horusec/horusec-custom-rules.json\\\".\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tBoolP(\n\t\t\t\"information-severity\", \"I\",\n\t\t\ts.configs.EnableInformationSeverity,\n\t\t\t\"Used to enable or disable information severity vulnerabilities, information vulnerabilities can contain a lot of false positives. Example: -I=\\\"true\\\"\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tStringSlice(\n\t\t\t\"show-vulnerabilities-types\",\n\t\t\ts.configs.ShowVulnerabilitiesTypes,\n\t\t\t\"Used to show in the output vulnerabilities of types: Vulnerability, Risk Accepted, False Positive, Corrected. Example --show-vulnerabilities-types=\\\"Vulnerability, Risk Accepted\\\"\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tBoolP(\n\t\t\t\"enable-owasp-dependency-check\", \"w\",\n\t\t\ts.configs.EnableOwaspDependencyCheck,\n\t\t\t\"Enable owasp dependency check. Example -w=\\\"true\\\". Default: false\",\n\t\t)\n\n\tstartCmd.PersistentFlags().\n\t\tBoolP(\n\t\t\t\"enable-shellcheck\", \"j\",\n\t\t\ts.configs.EnableShellCheck,\n\t\t\t\"Enable shellcheck. Example -h=\\\"true\\\". Default: false\",\n\t\t)\n\n\tif !dist.IsStandAlone() {\n\t\tstartCmd.PersistentFlags().\n\t\t\tBoolP(\n\t\t\t\t\"disable-docker\", \"D\",\n\t\t\t\ts.configs.DisableDocker,\n\t\t\t\t\"Used to run horusec without docker if enabled it will only run the following tools: horusec-csharp, horusec-kotlin, horusec-java, horusec-kubernetes, horusec-leaks, horusec-nodejs, horusec-dart, horusec-nginx. Example: -D=\\\"true\\\"\",\n\t\t\t)\n\t}\n\n\treturn startCmd\n}", "func InitCommand(client *apiclient.Client, config config.Provider, globalFlags *pflag.FlagSet, version semver.Version) *cobra.Command {\n\tapp := &cobra.Command{\n\t\tUse: \"storageos <command>\",\n\t\tShort: `Storage for Cloud Native Applications.\n\nBy using this product, you are agreeing to the terms of the the StorageOS Ltd. End\nUser Subscription Agreement (EUSA) found at: https://storageos.com/legal/#eusa\n\nTo be notified about stable releases and latest features, sign up at https://my.storageos.com.\n`,\n\n\t\tPersistentPreRunE: func(_ *cobra.Command, _ []string) error {\n\t\t\tuserAgent := strings.Join([]string{UserAgentPrefix, version.String()}, \"/\")\n\n\t\t\tdisabledAuthCache, err := config.AuthCacheDisabled()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\ttransport, err := func() (apiclient.Transport, error) {\n\t\t\t\ttransport, err := openapi.NewOpenAPI(config, userAgent)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn nil, err\n\t\t\t\t}\n\n\t\t\t\tif disabledAuthCache {\n\t\t\t\t\treturn transport, nil\n\t\t\t\t}\n\n\t\t\t\t// Ensure that the cache dir exists\n\t\t\t\tcacheDir, err := config.CacheDir()\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn transport, nil\n\t\t\t\t}\n\n\t\t\t\terr = os.Mkdir(cacheDir, 0o700)\n\t\t\t\tswitch {\n\t\t\t\tcase err == nil, os.IsExist(err):\n\t\t\t\t\t// This is ok\n\t\t\t\tdefault:\n\t\t\t\t\treturn nil, err\n\t\t\t\t}\n\n\t\t\t\t// Only wrap with caching if desired\n\t\t\t\treturn apiclient.NewAuthCachedTransport(\n\t\t\t\t\ttransport,\n\t\t\t\t\tcache.NewSessionCache(config, time.Now, 5*time.Second),\n\t\t\t\t), nil\n\t\t\t}()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\t// Wrap the transport implementation in a re-authenticate layer.\n\t\t\treturn client.ConfigureTransport(\n\t\t\t\tapiclient.NewTransportWithReauth(transport, config),\n\t\t\t)\n\t\t},\n\n\t\tSilenceErrors: true,\n\t}\n\n\t// Register the generic CLI commands that don't do any API interaction.\n\tversionCommand := &cobra.Command{\n\t\tUse: \"version\",\n\t\tShort: \"View version information for the StorageOS CLI\",\n\t\tRun: func(_ *cobra.Command, _ []string) {\n\t\t\tfmt.Printf(\"StorageOS CLI version: %v\\n\", version.String())\n\t\t},\n\t}\n\n\tapp.AddCommand(\n\t\tapply.NewCommand(client, config),\n\t\tupdate.NewCommand(client, config),\n\t\tcreate.NewCommand(client, config),\n\t\tget.NewCommand(client, config),\n\t\tdescribe.NewCommand(client, config),\n\t\tdelete.NewCommand(client, config),\n\t\tattach.NewCommand(client, config),\n\t\tdetach.NewCommand(client, config),\n\t\tnfs.NewCommand(client, config),\n\t\tcordon.NewCordonCommand(os.Stdout, client, config),\n\t\tcordon.NewUncordonCommand(os.Stdout, client, config),\n\t\tversionCommand,\n\t\tevict.NewCommand(client, config),\n\t\tpromote.NewCommand(client, config),\n\t)\n\n\t// Cobra subcommands which are not runnable and do not themselves have\n\t// subcommands are added as additional help topics.\n\tapp.AddCommand(\n\t\tnewConfigFileHelpTopic(config),\n\t\tnewEnvConfigHelpTopic(),\n\t\tnewExitCodeHelpTopic(),\n\t)\n\n\tapp.PersistentFlags().AddFlagSet(globalFlags)\n\n\treturn app\n}", "func NewCommand() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tArgs: cobra.NoArgs,\n\t\tUse: \"get\",\n\t\tShort: \"get shows info about specific object\",\n\t\tLong: \"get shows info about specific objects\",\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\treturn cmd.Help()\n\t\t},\n\t}\n\t// add subcommands\n\tcmd.AddCommand(pods.NewCommand())\n\treturn cmd\n}" ]
[ "0.7556033", "0.69892746", "0.6661952", "0.66520774", "0.6465044", "0.60784453", "0.5981338", "0.59450066", "0.5934187", "0.5931092", "0.59059626", "0.5853058", "0.58109266", "0.5786375", "0.57806927", "0.57552415", "0.567265", "0.5668012", "0.56584144", "0.56572723", "0.5648297", "0.56259537", "0.56039566", "0.5599265", "0.55976415", "0.55896705", "0.5578038", "0.5572141", "0.5571731", "0.556211", "0.5561028", "0.55598915", "0.5559696", "0.5556877", "0.5533198", "0.55230623", "0.5521308", "0.5511829", "0.55086815", "0.5494634", "0.54852116", "0.5475875", "0.5475453", "0.54601425", "0.5457711", "0.5457526", "0.54570884", "0.5456116", "0.5445327", "0.5444429", "0.5433864", "0.5433468", "0.5429532", "0.5427643", "0.5421043", "0.5410432", "0.5409262", "0.5401659", "0.5401245", "0.5387119", "0.53811735", "0.5380936", "0.5380313", "0.5379513", "0.5378421", "0.5372852", "0.5363265", "0.5362035", "0.53572756", "0.53488773", "0.5345879", "0.53370976", "0.53317636", "0.53248537", "0.53177166", "0.531413", "0.5311377", "0.5307623", "0.5306014", "0.53058356", "0.5303435", "0.53029406", "0.53022015", "0.5301619", "0.5300861", "0.52853346", "0.5284511", "0.52842337", "0.52815586", "0.52808994", "0.52717954", "0.5261062", "0.52601326", "0.5260126", "0.5256543", "0.52528137", "0.5251067", "0.5241272", "0.52342623", "0.52218133" ]
0.810491
0
Sends a create tenant task to client based on the cli.Context Returns an error if one occurred
func createTenant(c *cli.Context, w io.Writer) error { if len(c.Args()) > 1 { return fmt.Errorf("Unknown argument: %v", c.Args()[1:]) } name := c.Args().First() securityGroups := c.String("security-groups") if !c.GlobalIsSet("non-interactive") { var err error name, err = askForInput("Tenant name: ", name) if err != nil { return err } securityGroups, err = askForInput("Comma-separated security group names, or hit enter for no security groups): ", securityGroups) if err != nil { return err } } if len(name) == 0 { return fmt.Errorf("Please provide tenant name") } securityGroupList := []string{} if securityGroups != "" { securityGroupList = regexp.MustCompile(`\s*,\s*`).Split(securityGroups, -1) } tenantSpec := &photon.TenantCreateSpec{ Name: name, SecurityGroups: securityGroupList, } var err error client.Photonclient, err = client.GetClient(c) if err != nil { return err } createTask, err := client.Photonclient.Tenants.Create(tenantSpec) if err != nil { return err } id, err := waitOnTaskOperation(createTask.ID, c) if err != nil { return err } if utils.NeedsFormatting(c) { tenant, err := client.Photonclient.Tenants.Get(id) if err != nil { return err } utils.FormatObject(tenant, w, c) } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func TenantCreate(tenantID, userEmail, tenantName, UserName string) {\n\tRunCmd(fmt.Sprintf(\"%s tenant -op=add -id=%s -name=\\\"%s\\\" -user-email=%s -user-name=\\\"%s\\\"\",\n\t\tActlPath, tenantID, tenantName, userEmail, UserName))\n}", "func NewTenantCreateCmd() *cobra.Command {\n\ttenantCreateCmd := &cobra.Command{\n\t\tUse: \"create\",\n\t\tTraverseChildren: true,\n\t\tShort: \"Create a tenant resource within Karavi\",\n\t\tLong: `Creates a tenant resource within Karavi`,\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\taddr, err := cmd.Flags().GetString(\"addr\")\n\t\t\tif err != nil {\n\t\t\t\treportErrorAndExit(JSONOutput, cmd.ErrOrStderr(), err)\n\t\t\t}\n\n\t\t\tinsecure, err := cmd.Flags().GetBool(\"insecure\")\n\t\t\tif err != nil {\n\t\t\t\treportErrorAndExit(JSONOutput, cmd.ErrOrStderr(), err)\n\t\t\t}\n\n\t\t\ttenantClient, conn, err := CreateTenantServiceClient(addr, insecure)\n\t\t\tif err != nil {\n\t\t\t\treportErrorAndExit(JSONOutput, cmd.ErrOrStderr(), err)\n\t\t\t}\n\t\t\tdefer conn.Close()\n\n\t\t\tname, err := cmd.Flags().GetString(\"name\")\n\t\t\tif err != nil {\n\t\t\t\treportErrorAndExit(JSONOutput, cmd.ErrOrStderr(), err)\n\t\t\t}\n\t\t\tif strings.TrimSpace(name) == \"\" {\n\t\t\t\treportErrorAndExit(JSONOutput, cmd.ErrOrStderr(), errors.New(\"empty name not allowed\"))\n\t\t\t}\n\n\t\t\t_, err = tenantClient.CreateTenant(context.Background(), &pb.CreateTenantRequest{\n\t\t\t\tTenant: &pb.Tenant{\n\t\t\t\t\tName: name,\n\t\t\t\t},\n\t\t\t})\n\t\t\tif err != nil {\n\t\t\t\treportErrorAndExit(JSONOutput, cmd.ErrOrStderr(), err)\n\t\t\t}\n\t\t},\n\t}\n\n\ttenantCreateCmd.Flags().StringP(\"name\", \"n\", \"\", \"Tenant name\")\n\treturn tenantCreateCmd\n}", "func CreateTenant(tenant string) {\n\tconn, err := grpc.Dial(\"localhost:8885\", grpc.WithInsecure())\n\tif err != nil {\n\t\tfmt.Printf(\"Cannot open gRPC: %s\", err)\n\t\t// TODO error\n\t}\n\t\n\tdefer conn.Close()\n\n\tclient := schema.NewTenantServiceClient(conn)\n\n\tnewTenant, err := client.Create(context.Background(), &wrapperspb.StringValue{Value: tenant})\n\tif err != nil {\n\t\tfmt.Errorf(\"Cannot create a tenant: %s\", err)\n\t}\n\n\tfmt.Printf(\"Tenant created successfully: %s\", newTenant)\n}", "func (api *tenantAPI) Create(obj *cluster.Tenant) error {\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = apicl.ClusterV1().Tenant().Create(context.Background(), obj)\n\t\tif err != nil && strings.Contains(err.Error(), \"AlreadyExists\") {\n\t\t\t_, err = apicl.ClusterV1().Tenant().Update(context.Background(), obj)\n\n\t\t}\n\t\treturn err\n\t}\n\n\tapi.ct.handleTenantEvent(&kvstore.WatchEvent{Object: obj, Type: kvstore.Created})\n\treturn nil\n}", "func Create(client *gophercloud.ServiceClient, opts CreateOptsBuilder) (r CreateResult) {\r\n\tb, err := opts.ToTenantCreateMap()\r\n\tif err != nil {\r\n\t\tr.Err = err\r\n\t\treturn\r\n\t}\r\n\t_, r.Err = client.Post(createURL(client), b, &r.Body, &gophercloud.RequestOpts{\r\n\t\tOkCodes: []int{200, 201},\r\n\t})\r\n\treturn\r\n}", "func TestCreateTenant(t *testing.T) {\n\tctx, cancelFunc := context.WithTimeout(context.Background(), standardTimeout)\n\tdefer cancelFunc()\n\n\tconfig := bat.TenantConfig{\n\t\tName: \"CreateTenantTest\",\n\t\tSubnetBits: 20,\n\t}\n\n\ttenant, err := bat.CreateTenant(ctx, config)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to create tenant : %v\", err)\n\t}\n\n\tdefer func() {\n\t\terr = bat.DeleteTenant(ctx, tenant.ID)\n\t\tif err != nil {\n\t\t\tt.Fatalf(\"Failed to delete tenant: %v\", err)\n\t\t}\n\t}()\n\n\tif tenant.Name != config.Name {\n\t\tt.Fatalf(\"Failed to create tenant\")\n\t}\n\n\ttenants, err := bat.GetAllTenants(ctx)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to retrieve tenant list : %v\", err)\n\t}\n\n\tfor _, tt := range tenants {\n\t\tif tt.Name == config.Name {\n\t\t\treturn\n\t\t}\n\t}\n\n\tt.Fatal(\"did not find new tenant in tenants list\")\n}", "func CreateTenant(tenantName string, tenantParams *TenantParams) error {\n\t_, err := handler.GetDefaultHandlerSet().GetNamespaceHandler().GetNamespace(tenantName)\n\tif err != nil {\n\t\tif adaptor.IsNotFoundErr(err) {\n\t\t\ttenantLabel := make(map[string]string, 0)\n\t\t\tfor k, v := range tenantParams.TenantLabels {\n\t\t\t\ttenantLabel[k] = v\n\t\t\t}\n\t\t\ttenantLabel[\"multi-tenant\"] = fmt.Sprintf(\"tenant-tiller-%s\", tenantName)\n\t\t\tnamespace := corev1.Namespace{\n\t\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\t\tNamespace: tenantName,\n\t\t\t\t\tName: tenantName,\n\t\t\t\t\tLabels: tenantLabel,\n\t\t\t\t\tAnnotations: tenantParams.TenantAnnotations,\n\t\t\t\t},\n\t\t\t}\n\t\t\t_, err = handler.GetDefaultHandlerSet().GetNamespaceHandler().CreateNamespace(&namespace)\n\t\t\tif err != nil {\n\t\t\t\tlogrus.Errorf(\"failed to create namespace %s : %s\", tenantName, err.Error())\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\terr = doCreateTenant(tenantName, tenantParams)\n\t\t\tif err != nil {\n\t\t\t\t// rollback\n\t\t\t\thandler.GetDefaultHandlerSet().GetNamespaceHandler().DeleteNamespace(tenantName)\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tlogrus.Infof(\"succeed to create tenant %s\", tenantName)\n\t\t\treturn nil\n\t\t}\n\t\tlogrus.Errorf(\"failed to get namespace : %s\", err.Error())\n\t\treturn err\n\n\t} else {\n\t\tlogrus.Warnf(\"namespace %s exists\", tenantName)\n\t\treturn nil\n\t}\n}", "func TestCreateTenantWorkload(t *testing.T) {\n\ttestCreateWorkload(t, false)\n}", "func CreateTenant(name string) error {\n\treturn util.ErrUnimplementedFeature\n}", "func (ts *TaskService) Create(requestCtx context.Context, req *taskAPI.CreateTaskRequest) (*taskAPI.CreateTaskResponse, error) {\n\tdefer logPanicAndDie(log.G(requestCtx))\n\ttaskID := req.ID\n\texecID := \"\" // the exec ID of the initial process in a task is an empty string by containerd convention\n\n\tlogger := log.G(requestCtx).WithField(\"TaskID\", taskID).WithField(\"ExecID\", execID)\n\tlogger.Info(\"create\")\n\n\textraData, err := unmarshalExtraData(req.Options)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to unmarshal extra data\")\n\t}\n\n\t// Just provide runc the options it knows about, not our wrapper\n\treq.Options = extraData.RuncOptions\n\n\t// Override the bundle dir and rootfs paths, which were set on the Host and thus not valid here in the Guest\n\tbundleDir := bundle.Dir(filepath.Join(containerRootDir, taskID))\n\terr = bundleDir.Create()\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to create bundle dir\")\n\t}\n\n\tdefer func() {\n\t\tif err != nil {\n\t\t\tremoveErr := os.RemoveAll(bundleDir.RootPath())\n\t\t\tif removeErr != nil {\n\t\t\t\tlogger.WithError(removeErr).Error(\"failed to cleanup bundle dir\")\n\t\t\t}\n\t\t}\n\t}()\n\n\terr = bundleDir.OCIConfig().Write(extraData.JsonSpec)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to write oci config file\")\n\t}\n\n\tdriveID := strings.TrimSpace(extraData.DriveID)\n\tdrive, ok := ts.driveHandler.GetDrive(driveID)\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"Drive %q could not be found\", driveID)\n\t}\n\n\tconst (\n\t\tmaxRetries = 100\n\t\tretryDelay = 10 * time.Millisecond\n\t)\n\n\t// We retry here due to guest kernel needing some time to populate the guest\n\t// drive.\n\t// https://github.com/firecracker-microvm/firecracker/issues/1159\n\tfor i := 0; i < maxRetries; i++ {\n\t\tif err = bundleDir.MountRootfs(drive.Path(), \"ext4\", nil); isRetryableMountError(err) {\n\t\t\tlogger.WithError(err).Warnf(\"retrying to mount rootfs %q\", drive.Path())\n\n\t\t\ttime.Sleep(retryDelay)\n\t\t\tcontinue\n\t\t}\n\n\t\tbreak\n\t}\n\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"failed to mount rootfs %q\", drive.Path())\n\t}\n\n\treq.Bundle = bundleDir.RootPath()\n\treq.Rootfs = nil\n\n\tvar ioConnectorSet vm.IOProxy\n\n\tif vm.IsAgentOnlyIO(req.Stdout, logger) {\n\t\tioConnectorSet = vm.NewNullIOProxy()\n\t} else {\n\t\t// Override the incoming stdio FIFOs, which have paths from the host that we can't use\n\t\tfifoSet, err := cio.NewFIFOSetInDir(bundleDir.RootPath(), fifoName(taskID, execID), req.Terminal)\n\t\tif err != nil {\n\t\t\tlogger.WithError(err).Error(\"failed opening stdio FIFOs\")\n\t\t\treturn nil, errors.Wrap(err, \"failed to open stdio FIFOs\")\n\t\t}\n\n\t\tvar stdinConnectorPair *vm.IOConnectorPair\n\t\tif req.Stdin != \"\" {\n\t\t\treq.Stdin = fifoSet.Stdin\n\t\t\tstdinConnectorPair = &vm.IOConnectorPair{\n\t\t\t\tReadConnector: vm.VSockAcceptConnector(extraData.StdinPort),\n\t\t\t\tWriteConnector: vm.FIFOConnector(fifoSet.Stdin),\n\t\t\t}\n\t\t}\n\n\t\tvar stdoutConnectorPair *vm.IOConnectorPair\n\t\tif req.Stdout != \"\" {\n\t\t\treq.Stdout = fifoSet.Stdout\n\t\t\tstdoutConnectorPair = &vm.IOConnectorPair{\n\t\t\t\tReadConnector: vm.FIFOConnector(fifoSet.Stdout),\n\t\t\t\tWriteConnector: vm.VSockAcceptConnector(extraData.StdoutPort),\n\t\t\t}\n\t\t}\n\n\t\tvar stderrConnectorPair *vm.IOConnectorPair\n\t\tif req.Stderr != \"\" {\n\t\t\treq.Stderr = fifoSet.Stderr\n\t\t\tstderrConnectorPair = &vm.IOConnectorPair{\n\t\t\t\tReadConnector: vm.FIFOConnector(fifoSet.Stderr),\n\t\t\t\tWriteConnector: vm.VSockAcceptConnector(extraData.StderrPort),\n\t\t\t}\n\t\t}\n\n\t\tioConnectorSet = vm.NewIOConnectorProxy(stdinConnectorPair, stdoutConnectorPair, stderrConnectorPair)\n\t}\n\n\tresp, err := ts.taskManager.CreateTask(requestCtx, req, ts.runcService, ioConnectorSet)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to create runc shim for task\")\n\t}\n\n\tlogger.WithField(\"pid\", resp.Pid).Debugf(\"create succeeded\")\n\treturn resp, nil\n}", "func (a *Client) CreateTenant(params *CreateTenantParams) (*CreateTenantOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewCreateTenantParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"createTenant\",\n\t\tMethod: \"POST\",\n\t\tPathPattern: \"/tenants\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\"},\n\t\tParams: params,\n\t\tReader: &CreateTenantReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*CreateTenantOK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\t// safeguard: normally, absent a default response, unknown success responses return an error above: so this is a codegen issue\n\tmsg := fmt.Sprintf(\"unexpected success response for createTenant: API contract not enforced by server. Client expected to get an error, but got: %T\", result)\n\tpanic(msg)\n}", "func InitTenant(ctx context.Context, masterURL string, service tenant.Service, currentTenant *tenant.Tenant) Callback {\n\tvar maxResourceQuotaStatusCheck int32 = 50 // technically a global retry count across all ResourceQuota on all Tenant Namespaces\n\tvar currentResourceQuotaStatusCheck int32 // default is 0\n\treturn func(statusCode int, method string, request, response map[interface{}]interface{}, versionMapping map[env.Type]string) (string, map[interface{}]interface{}, error) {\n\t\tlog.Info(ctx, map[string]interface{}{\n\t\t\t\"status\": statusCode,\n\t\t\t\"method\": method,\n\t\t\t\"cluster_url\": masterURL,\n\t\t\t\"namespace\": env.GetNamespace(request),\n\t\t\t\"name\": env.GetName(request),\n\t\t\t\"kind\": env.GetKind(request),\n\t\t\t\"request\": yamlString(request),\n\t\t\t\"response\": yamlString(response),\n\t\t}, \"resource requested\")\n\t\tif statusCode == http.StatusConflict || statusCode == http.StatusForbidden {\n\t\t\tif env.GetKind(request) == env.ValKindNamespace || env.GetKind(request) == env.ValKindProjectRequest ||\n\t\t\t\tenv.GetKind(request) == env.ValKindPersistenceVolumeClaim {\n\t\t\t\treturn \"\", nil, fmt.Errorf(\"unable to create %s - should create with other base-name\", env.GetNamespace(request))\n\t\t\t}\n\t\t\treturn \"DELETE\", request, nil\n\t\t} else if statusCode == http.StatusCreated {\n\t\t\tif env.GetKind(request) == env.ValKindProjectRequest {\n\t\t\t\tname := env.GetName(request)\n\t\t\t\tenvType := tenant.GetNamespaceType(name, currentTenant.NsBaseName)\n\t\t\t\ttemplatesVersion := versionMapping[envType]\n\t\t\t\tservice.SaveNamespace(&tenant.Namespace{\n\t\t\t\t\tTenantID: currentTenant.ID,\n\t\t\t\t\tName: name,\n\t\t\t\t\tState: tenant.Ready,\n\t\t\t\t\tVersion: templatesVersion,\n\t\t\t\t\tType: envType,\n\t\t\t\t\tMasterURL: masterURL,\n\t\t\t\t\tUpdatedBy: configuration.Commit,\n\t\t\t\t})\n\n\t\t\t\t// HACK to workaround osio applying some dsaas-user permissions async\n\t\t\t\t// Should loop on a Check if allowed type of call instead\n\t\t\t\ttime.Sleep(time.Second * 5)\n\n\t\t\t} else if env.GetKind(request) == env.ValKindNamespace {\n\t\t\t\tname := env.GetName(request)\n\t\t\t\tenvType := tenant.GetNamespaceType(name, currentTenant.NsBaseName)\n\t\t\t\ttemplatesVersion := versionMapping[envType]\n\t\t\t\tservice.SaveNamespace(&tenant.Namespace{\n\t\t\t\t\tTenantID: currentTenant.ID,\n\t\t\t\t\tName: name,\n\t\t\t\t\tState: tenant.Ready,\n\t\t\t\t\tVersion: templatesVersion,\n\t\t\t\t\tType: envType,\n\t\t\t\t\tMasterURL: masterURL,\n\t\t\t\t\tUpdatedBy: configuration.Commit,\n\t\t\t\t})\n\t\t\t} else if env.GetKind(request) == env.ValKindResourceQuota {\n\t\t\t\t// trigger a check status loop\n\t\t\t\ttime.Sleep(time.Millisecond * 50)\n\t\t\t\treturn \"GET\", response, nil\n\t\t\t}\n\t\t\treturn \"\", nil, nil\n\t\t} else if statusCode == http.StatusOK {\n\t\t\tif method == \"DELETE\" {\n\t\t\t\treturn \"POST\", request, nil\n\t\t\t} else if method == \"GET\" {\n\t\t\t\tif env.GetKind(request) == env.ValKindResourceQuota {\n\n\t\t\t\t\tif env.HasValidStatus(response) || atomic.LoadInt32(&currentResourceQuotaStatusCheck) >= maxResourceQuotaStatusCheck {\n\t\t\t\t\t\treturn \"\", nil, nil\n\t\t\t\t\t}\n\t\t\t\t\tatomic.AddInt32(&currentResourceQuotaStatusCheck, 1)\n\t\t\t\t\ttime.Sleep(time.Millisecond * 50)\n\t\t\t\t\treturn \"GET\", response, nil\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn \"\", nil, nil\n\t\t}\n\t\tlog.Info(ctx, map[string]interface{}{\n\t\t\t\"status\": statusCode,\n\t\t\t\"method\": method,\n\t\t\t\"namespace\": env.GetNamespace(request),\n\t\t\t\"cluster_url\": masterURL,\n\t\t\t\"name\": env.GetName(request),\n\t\t\t\"kind\": env.GetKind(request),\n\t\t\t\"request\": yamlString(request),\n\t\t\t\"response\": yamlString(response),\n\t\t}, \"unhandled resource response\")\n\t\treturn \"\", nil, nil\n\t}\n}", "func (c *TenantController) Setup(ctx *app.SetupTenantContext) error {\n\tuserToken := goajwt.ContextJWT(ctx)\n\tif userToken == nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewUnauthorizedError(\"Missing JWT token\"))\n\t}\n\tttoken := &auth.TenantToken{Token: userToken}\n\texists := c.tenantService.Exists(ttoken.Subject())\n\tif exists {\n\t\treturn ctx.Conflict()\n\t}\n\n\t// fetch the cluster the user belongs to\n\tuser, err := c.authClientService.GetUser(ctx)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, err)\n\t}\n\n\tif user.UserData.Cluster == nil {\n\t\tlog.Error(ctx, nil, \"no cluster defined for tenant\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewInternalError(ctx, fmt.Errorf(\"unable to provision to undefined cluster\")))\n\t}\n\n\t// fetch the cluster info\n\tcluster, err := c.clusterService.GetCluster(ctx, *user.UserData.Cluster)\n\tif err != nil {\n\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\"err\": err,\n\t\t\t\"cluster_url\": *user.UserData.Cluster,\n\t\t}, \"unable to fetch cluster\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewInternalError(ctx, err))\n\t}\n\n\tnsBaseName, err := tenant.ConstructNsBaseName(c.tenantService, env.RetrieveUserName(user.OpenShiftUsername))\n\tif err != nil {\n\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\"err\": err,\n\t\t\t\"os_username\": user.OpenShiftUsername,\n\t\t}, \"unable to construct namespace base name\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewInternalError(ctx, err))\n\t}\n\n\t// create openshift config\n\topenshiftConfig := openshift.NewConfig(c.config, user.UserData, cluster.User, cluster.Token, cluster.APIURL)\n\ttenant := &tenant.Tenant{\n\t\tID: ttoken.Subject(),\n\t\tEmail: ttoken.Email(),\n\t\tOSUsername: user.OpenShiftUsername,\n\t\tNsBaseName: nsBaseName,\n\t}\n\terr = c.tenantService.CreateTenant(tenant)\n\tif err != nil {\n\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\"err\": err,\n\t\t}, \"unable to store tenant configuration\")\n\t\treturn ctx.Conflict()\n\t}\n\n\tgo func() {\n\t\tctx := ctx\n\t\tt := tenant\n\t\terr = openshift.RawInitTenant(\n\t\t\tctx,\n\t\t\topenshiftConfig,\n\t\t\tInitTenant(ctx, openshiftConfig.MasterURL, c.tenantService, t),\n\t\t\tuser.OpenShiftUsername,\n\t\t\tnsBaseName,\n\t\t\tuser.OpenShiftUserToken)\n\n\t\tif err != nil {\n\t\t\tsentry.LogError(ctx, map[string]interface{}{\n\t\t\t\t\"os_user\": user.OpenShiftUsername,\n\t\t\t}, err, \"unable initialize tenant\")\n\t\t}\n\t}()\n\n\tctx.ResponseData.Header().Set(\"Location\", rest.AbsoluteURL(ctx.RequestData.Request, app.TenantHref()))\n\treturn ctx.Accepted()\n}", "func (client LabClient) CreateEnvironmentSender(req *http.Request) (future LabCreateEnvironmentFuture, err error) {\n\tvar resp *http.Response\n\tresp, err = autorest.SendWithSender(client, req,\n\t\tazure.DoRetryWithRegistration(client.Client))\n\tif err != nil {\n\t\treturn\n\t}\n\tfuture.Future, err = azure.NewFutureFromResponse(resp)\n\treturn\n}", "func (api *tenantAPI) SyncCreate(obj *cluster.Tenant) error {\n\tnewObj := obj\n\tevtType := kvstore.Created\n\tvar writeErr error\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\tnewObj, writeErr = apicl.ClusterV1().Tenant().Create(context.Background(), obj)\n\t\tif writeErr != nil && strings.Contains(writeErr.Error(), \"AlreadyExists\") {\n\t\t\tnewObj, writeErr = apicl.ClusterV1().Tenant().Update(context.Background(), obj)\n\t\t\tevtType = kvstore.Updated\n\t\t}\n\t}\n\n\tif writeErr == nil {\n\t\tapi.ct.handleTenantEvent(&kvstore.WatchEvent{Object: newObj, Type: evtType})\n\t}\n\treturn writeErr\n}", "func GetTenantsCommand() cli.Command {\n\tcommand := cli.Command{\n\t\tName: \"tenant\",\n\t\tUsage: \"options for tenant\",\n\t\tSubcommands: []cli.Command{\n\t\t\t{\n\t\t\t\tName: \"create\",\n\t\t\t\tUsage: \"Create a new tenant\",\n\t\t\t\tArgsUsage: \"<tenant-name>\",\n\t\t\t\tFlags: []cli.Flag{\n\t\t\t\t\tcli.StringFlag{\n\t\t\t\t\t\tName: \"security-groups, s\",\n\t\t\t\t\t\tUsage: \"Comma-separated Lightwave group names, to specify the tenant administrators\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := createTenant(c, os.Stdout)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"delete\",\n\t\t\t\tUsage: \"Delete a tenant\",\n\t\t\t\tArgsUsage: \"<tenant-id>\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := deleteTenant(c)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"show\",\n\t\t\t\tUsage: \"Show detailed tenant info with specified id\",\n\t\t\t\tArgsUsage: \"<tenant-id>\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := showTenant(c, os.Stdout)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"list\",\n\t\t\t\tUsage: \"List all tenants\",\n\t\t\t\tArgsUsage: \" \",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := listTenants(c, os.Stdout)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"set\",\n\t\t\t\tUsage: \"Set default tenant\",\n\t\t\t\tArgsUsage: \"<tenant-name>\",\n\t\t\t\tDescription: \"Set the default project that will be used for all photon CLI commands that need a project.\\n\" +\n\t\t\t\t\t\" Most commands allow you to override the default.\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := setTenant(c)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"get\",\n\t\t\t\tUsage: \"Get default tenant\",\n\t\t\t\tArgsUsage: \" \",\n\t\t\t\tDescription: \"Show default project in use for photon CLI commands. Most command allow you to either\\n\" +\n\t\t\t\t\t\" use this default or specify a specific project to use.\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := getTenant(c, os.Stdout)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"tasks\",\n\t\t\t\tUsage: \"Show tenant tasks\",\n\t\t\t\tArgsUsage: \"<tenant-id>\",\n\t\t\t\tFlags: []cli.Flag{\n\t\t\t\t\tcli.StringFlag{\n\t\t\t\t\t\tName: \"state, s\",\n\t\t\t\t\t\tUsage: \"Filter by task sate\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := getTenantTasks(c, os.Stdout)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"set-security-groups\",\n\t\t\t\tUsage: \"Set security groups for a tenant\",\n\t\t\t\tArgsUsage: \"<tenant-id> <comma separated list of groups>\",\n\t\t\t\tDescription: \"Set the list of Lightwave groups that can administer this tenant. This may only be\\n\" +\n\t\t\t\t\t\" be set by a member of the tenant. Be cautious--you can remove your own access if you specify\\n\" +\n\t\t\t\t\t\" the wrong set of groups.\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := setSecurityGroups(c)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tHidden: true,\n\t\t\t\tName: \"set_security_groups\",\n\t\t\t\tUsage: \"Set security groups for a tenant\",\n\t\t\t\tArgsUsage: \"<tenant-id> <comma separated list of groups>\",\n\t\t\t\tDescription: \"Deprecated, use set-security-groups instead\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := setSecurityGroups(c)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\treturn command\n}", "func CreateTask(c *gin.Context) {\n\tsrv := server.GetServer()\n\n\tvar param TaskParams\n\tc.BindJSON(&param)\n\n\targs := make([]tasks.Arg, len(param.Args))\n\tfor idx, arg := range param.Args {\n\t\targs[idx] = tasks.Arg{\n\t\t\tType: \"int64\",\n\t\t\tValue: arg,\n\t\t}\n\t}\n\tsignature := &tasks.Signature{\n\t\tName: param.TaskName,\n\t\tArgs: args,\n\t\tRetryCount: 3,\n\t}\n\n\tasyncResult, err := srv.SendTask(signature)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tlog.Println(\"asyncResult:\", asyncResult)\n\n\tc.JSON(http.StatusOK, gin.H{\"Status\": \"In progress\", \"Job\": asyncResult})\n}", "func (c *TenantController) Setup(ctx *app.SetupTenantContext) error {\n\tuserToken := goajwt.ContextJWT(ctx)\n\tif userToken == nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewUnauthorizedError(\"Missing JWT token\"))\n\t}\n\tttoken := &TenantToken{token: userToken}\n\texists := c.tenantService.Exists(ttoken.Subject())\n\tif exists {\n\t\treturn ctx.Conflict()\n\t}\n\n\t// fetch the cluster the user belongs to\n\tuser, err := c.userService.GetUser(ctx, ttoken.Subject())\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, err)\n\t}\n\n\tif user.Cluster == nil {\n\t\tlog.Error(ctx, nil, \"no cluster defined for tenant\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewInternalError(ctx, fmt.Errorf(\"unable to provision to undefined cluster\")))\n\t}\n\n\t// fetch the users cluster token\n\topenshiftUsername, openshiftUserToken, err := c.resolveTenant(ctx, *user.Cluster, userToken.Raw)\n\tif err != nil {\n\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\"err\": err,\n\t\t\t\"cluster_url\": *user.Cluster,\n\t\t}, \"unable to fetch tenant token from auth\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewUnauthorizedError(\"Could not resolve user token\"))\n\t}\n\n\t// fetch the cluster info\n\tcluster, err := c.resolveCluster(ctx, *user.Cluster)\n\tif err != nil {\n\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\"err\": err,\n\t\t\t\"cluster_url\": *user.Cluster,\n\t\t}, \"unable to fetch cluster\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewInternalError(ctx, err))\n\t}\n\n\t// create openshift config\n\topenshiftConfig := openshift.NewConfig(c.defaultOpenshiftConfig, user, cluster.User, cluster.Token, cluster.APIURL)\n\ttenant := &tenant.Tenant{ID: ttoken.Subject(), Email: ttoken.Email()}\n\terr = c.tenantService.CreateTenant(tenant)\n\tif err != nil {\n\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\"err\": err,\n\t\t}, \"unable to store tenant configuration\")\n\t\treturn ctx.Conflict()\n\t}\n\n\tgo func() {\n\t\tctx := ctx\n\t\tt := tenant\n\t\terr = openshift.RawInitTenant(\n\t\t\tctx,\n\t\t\topenshiftConfig,\n\t\t\tInitTenant(ctx, openshiftConfig.MasterURL, c.tenantService, t),\n\t\t\topenshiftUsername,\n\t\t\topenshiftUserToken,\n\t\t\tc.templateVars)\n\n\t\tif err != nil {\n\t\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\t\"err\": err,\n\t\t\t\t\"os_user\": openshiftUsername,\n\t\t\t}, \"unable initialize tenant\")\n\t\t}\n\t}()\n\n\tctx.ResponseData.Header().Set(\"Location\", rest.AbsoluteURL(ctx.RequestData.Request, app.TenantHref()))\n\treturn ctx.Accepted()\n}", "func InitTenant(ctx context.Context, masterURL string, service tenant.Service, currentTenant *tenant.Tenant) openshift.Callback {\n\tvar maxResourceQuotaStatusCheck int32 = 50 // technically a global retry count across all ResourceQuota on all Tenant Namespaces\n\tvar currentResourceQuotaStatusCheck int32 // default is 0\n\treturn func(statusCode int, method string, request, response map[interface{}]interface{}) (string, map[interface{}]interface{}) {\n\t\tlog.Info(ctx, map[string]interface{}{\n\t\t\t\"status\": statusCode,\n\t\t\t\"method\": method,\n\t\t\t\"cluster_url\": masterURL,\n\t\t\t\"namespace\": env.GetNamespace(request),\n\t\t\t\"name\": env.GetName(request),\n\t\t\t\"kind\": env.GetKind(request),\n\t\t\t\"request\": yamlString(request),\n\t\t\t\"response\": yamlString(response),\n\t\t}, \"resource requested\")\n\t\tif statusCode == http.StatusConflict {\n\t\t\tif env.GetKind(request) == env.ValKindNamespace {\n\t\t\t\treturn \"\", nil\n\t\t\t}\n\t\t\tif env.GetKind(request) == env.ValKindProjectRequest {\n\t\t\t\treturn \"\", nil\n\t\t\t}\n\t\t\tif env.GetKind(request) == env.ValKindPersistenceVolumeClaim {\n\t\t\t\treturn \"\", nil\n\t\t\t}\n\t\t\tif env.GetKind(request) == env.ValKindServiceAccount {\n\t\t\t\treturn \"\", nil\n\t\t\t}\n\t\t\treturn \"DELETE\", request\n\t\t} else if statusCode == http.StatusCreated {\n\t\t\tif env.GetKind(request) == env.ValKindProjectRequest {\n\t\t\t\tname := env.GetName(request)\n\t\t\t\tservice.SaveNamespace(&tenant.Namespace{\n\t\t\t\t\tTenantID: currentTenant.ID,\n\t\t\t\t\tName: name,\n\t\t\t\t\tState: \"created\",\n\t\t\t\t\tVersion: env.GetLabelVersion(request),\n\t\t\t\t\tType: tenant.GetNamespaceType(name, currentTenant.NsBaseName),\n\t\t\t\t\tMasterURL: masterURL,\n\t\t\t\t})\n\n\t\t\t\t// HACK to workaround osio applying some dsaas-user permissions async\n\t\t\t\t// Should loop on a Check if allowed type of call instead\n\t\t\t\ttime.Sleep(time.Second * 5)\n\n\t\t\t} else if env.GetKind(request) == env.ValKindNamespace {\n\t\t\t\tname := env.GetName(request)\n\t\t\t\tservice.SaveNamespace(&tenant.Namespace{\n\t\t\t\t\tTenantID: currentTenant.ID,\n\t\t\t\t\tName: name,\n\t\t\t\t\tState: \"created\",\n\t\t\t\t\tVersion: env.GetLabelVersion(request),\n\t\t\t\t\tType: tenant.GetNamespaceType(name, currentTenant.NsBaseName),\n\t\t\t\t\tMasterURL: masterURL,\n\t\t\t\t})\n\t\t\t} else if env.GetKind(request) == env.ValKindResourceQuota {\n\t\t\t\t// trigger a check status loop\n\t\t\t\ttime.Sleep(time.Millisecond * 50)\n\t\t\t\treturn \"GET\", response\n\t\t\t}\n\t\t\treturn \"\", nil\n\t\t} else if statusCode == http.StatusOK {\n\t\t\tif method == \"DELETE\" {\n\t\t\t\treturn \"POST\", request\n\t\t\t} else if method == \"GET\" {\n\t\t\t\tif env.GetKind(request) == env.ValKindResourceQuota {\n\n\t\t\t\t\tif env.HasValidStatus(response) || atomic.LoadInt32(&currentResourceQuotaStatusCheck) >= maxResourceQuotaStatusCheck {\n\t\t\t\t\t\treturn \"\", nil\n\t\t\t\t\t}\n\t\t\t\t\tatomic.AddInt32(&currentResourceQuotaStatusCheck, 1)\n\t\t\t\t\ttime.Sleep(time.Millisecond * 50)\n\t\t\t\t\treturn \"GET\", response\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn \"\", nil\n\t\t}\n\t\tlog.Info(ctx, map[string]interface{}{\n\t\t\t\"status\": statusCode,\n\t\t\t\"method\": method,\n\t\t\t\"namespace\": env.GetNamespace(request),\n\t\t\t\"cluster_url\": masterURL,\n\t\t\t\"name\": env.GetName(request),\n\t\t\t\"kind\": env.GetKind(request),\n\t\t\t\"request\": yamlString(request),\n\t\t\t\"response\": yamlString(response),\n\t\t}, \"unhandled resource response\")\n\t\treturn \"\", nil\n\t}\n}", "func CreateTenants(num int) {\n\tconnStr := viper.GetString(\"database_url\")\n\tfmt.Printf(\"Conn string is %s\\n\", connStr)\n\tdb, err := sql.Open(\"postgres\", connStr)\n\tif err != nil {\n\t\tfmt.Println(\"Failed to open DB: \", err)\n\t}\n\n\tdefer db.Close()\n\n\ttablesNum := viper.GetInt(\"tables\")\n\trecordsNum := viper.GetInt(\"records\")\n\n\tfor i := 1; i <= num; i++ {\n\t\ttenantName := fmt.Sprintf(\"client%d\", i)\n\t\tif err := createTenant(tenantName, db, tablesNum, recordsNum); err != nil {\n\t\t\tfmt.Printf(\"Failed to create client %d: %s\\n\", i, err)\n\t\t\treturn\n\t\t}\n\t\tfmt.Printf(\"Tenant %s created\\n\", tenantName)\n\t}\n}", "func (ctrler CtrlDefReactor) OnTenantCreate(obj *Tenant) error {\n\tlog.Info(\"OnTenantCreate is not implemented\")\n\treturn nil\n}", "func setTenant(c *cli.Context) error {\n\terr := checkArgCount(c, 1)\n\tif err != nil {\n\t\treturn err\n\t}\n\tname := c.Args().First()\n\n\tclient.Photonclient, err = client.GetClient(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Ensure tenant exists\n\tid, err := findTenantID(name)\n\tif len(id) == 0 || err != nil {\n\t\treturn err\n\t}\n\n\tconfig, err := cf.LoadConfig()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tconfig.Tenant = &cf.TenantConfiguration{Name: name, ID: id}\n\terr = cf.SaveConfig(config)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = clearConfigProject(\"\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif !c.GlobalIsSet(\"non-interactive\") {\n\t\tfmt.Printf(\"Tenant set to '%s'\\n\", name)\n\t}\n\treturn nil\n}", "func InitTenant(ctx context.Context, masterURL string, service tenant.Service, currentTenant *tenant.Tenant) openshift.Callback {\n\tvar maxResourceQuotaStatusCheck int32 = 50 // technically a global retry count across all ResourceQuota on all Tenant Namespaces\n\tvar currentResourceQuotaStatusCheck int32 = 0\n\treturn func(statusCode int, method string, request, response map[interface{}]interface{}) (string, map[interface{}]interface{}) {\n\t\tlog.Info(ctx, map[string]interface{}{\n\t\t\t\"status\": statusCode,\n\t\t\t\"method\": method,\n\t\t\t\"cluster_url\": masterURL,\n\t\t\t\"namespace\": openshift.GetNamespace(request),\n\t\t\t\"name\": openshift.GetName(request),\n\t\t\t\"kind\": openshift.GetKind(request),\n\t\t\t\"request\": yamlString(request),\n\t\t\t\"response\": yamlString(response),\n\t\t}, \"resource requested\")\n\t\tif statusCode == http.StatusConflict {\n\t\t\tif openshift.GetKind(request) == openshift.ValKindNamespace {\n\t\t\t\treturn \"\", nil\n\t\t\t}\n\t\t\tif openshift.GetKind(request) == openshift.ValKindProjectRequest {\n\t\t\t\treturn \"\", nil\n\t\t\t}\n\t\t\tif openshift.GetKind(request) == openshift.ValKindPersistenceVolumeClaim {\n\t\t\t\treturn \"\", nil\n\t\t\t}\n\t\t\tif openshift.GetKind(request) == openshift.ValKindServiceAccount {\n\t\t\t\treturn \"\", nil\n\t\t\t}\n\t\t\treturn \"DELETE\", request\n\t\t} else if statusCode == http.StatusCreated {\n\t\t\tif openshift.GetKind(request) == openshift.ValKindProjectRequest {\n\t\t\t\tname := openshift.GetName(request)\n\t\t\t\tservice.SaveNamespace(&tenant.Namespace{\n\t\t\t\t\tTenantID: currentTenant.ID,\n\t\t\t\t\tName: name,\n\t\t\t\t\tState: \"created\",\n\t\t\t\t\tVersion: openshift.GetLabelVersion(request),\n\t\t\t\t\tType: tenant.GetNamespaceType(name),\n\t\t\t\t\tMasterURL: masterURL,\n\t\t\t\t})\n\n\t\t\t\t// HACK to workaround osio applying some dsaas-user permissions async\n\t\t\t\t// Should loop on a Check if allowed type of call instead\n\t\t\t\ttime.Sleep(time.Second * 5)\n\n\t\t\t} else if openshift.GetKind(request) == openshift.ValKindNamespace {\n\t\t\t\tname := openshift.GetName(request)\n\t\t\t\tservice.SaveNamespace(&tenant.Namespace{\n\t\t\t\t\tTenantID: currentTenant.ID,\n\t\t\t\t\tName: name,\n\t\t\t\t\tState: \"created\",\n\t\t\t\t\tVersion: openshift.GetLabelVersion(request),\n\t\t\t\t\tType: tenant.GetNamespaceType(name),\n\t\t\t\t\tMasterURL: masterURL,\n\t\t\t\t})\n\t\t\t} else if openshift.GetKind(request) == openshift.ValKindResourceQuota {\n\t\t\t\t// trigger a check status loop\n\t\t\t\ttime.Sleep(time.Millisecond * 50)\n\t\t\t\treturn \"GET\", response\n\t\t\t}\n\t\t\treturn \"\", nil\n\t\t} else if statusCode == http.StatusOK {\n\t\t\tif method == \"DELETE\" {\n\t\t\t\treturn \"POST\", request\n\t\t\t} else if method == \"GET\" {\n\t\t\t\tif openshift.GetKind(request) == openshift.ValKindResourceQuota {\n\n\t\t\t\t\tif openshift.HasValidStatus(response) || atomic.LoadInt32(&currentResourceQuotaStatusCheck) >= maxResourceQuotaStatusCheck {\n\t\t\t\t\t\treturn \"\", nil\n\t\t\t\t\t}\n\t\t\t\t\tatomic.AddInt32(&currentResourceQuotaStatusCheck, 1)\n\t\t\t\t\ttime.Sleep(time.Millisecond * 50)\n\t\t\t\t\treturn \"GET\", response\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn \"\", nil\n\t\t}\n\t\tlog.Info(ctx, map[string]interface{}{\n\t\t\t\"status\": statusCode,\n\t\t\t\"method\": method,\n\t\t\t\"namespace\": openshift.GetNamespace(request),\n\t\t\t\"cluster_url\": masterURL,\n\t\t\t\"name\": openshift.GetName(request),\n\t\t\t\"kind\": openshift.GetKind(request),\n\t\t\t\"request\": yamlString(request),\n\t\t\t\"response\": yamlString(response),\n\t\t}, \"unhandled resource response\")\n\t\treturn \"\", nil\n\t}\n}", "func (c *Client) CreateTask(ctx context.Context, in *todopb.TaskCreateRequest, opts ...grpc.CallOption) (*todopb.TaskResponse, error) {\n\treturn c.client.CreateTask(ctx, in, opts...)\n}", "func (adminOrg *AdminOrg) CreateVdc(vdcConfiguration *types.VdcConfiguration) (Task, error) {\n\terr := validateVdcConfiguration(vdcConfiguration)\n\tif err != nil {\n\t\treturn Task{}, err\n\t}\n\n\tvdcConfiguration.Xmlns = types.XMLNamespaceVCloud\n\n\tvdcCreateHREF, err := url.ParseRequestURI(adminOrg.AdminOrg.HREF)\n\tif err != nil {\n\t\treturn Task{}, fmt.Errorf(\"error parsing admin org url: %s\", err)\n\t}\n\tvdcCreateHREF.Path += \"/vdcsparams\"\n\n\tadminVdc := NewAdminVdc(adminOrg.client)\n\n\t_, err = adminOrg.client.ExecuteRequest(vdcCreateHREF.String(), http.MethodPost,\n\t\t\"application/vnd.vmware.admin.createVdcParams+xml\", \"error creating VDC: %s\", vdcConfiguration, adminVdc.AdminVdc)\n\tif err != nil {\n\t\treturn Task{}, err\n\t}\n\n\t// Return the task\n\ttask := NewTask(adminOrg.client)\n\ttask.Task = adminVdc.AdminVdc.Tasks.Task[0]\n\treturn *task, nil\n}", "func (a *Client) CreateTask(params *CreateTaskParams, authInfo runtime.ClientAuthInfoWriter, opts ...ClientOption) (*CreateTaskCreated, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewCreateTaskParams()\n\t}\n\top := &runtime.ClientOperation{\n\t\tID: \"createTask\",\n\t\tMethod: \"POST\",\n\t\tPathPattern: \"/tasks\",\n\t\tProducesMediaTypes: []string{\"application/vnd.goswagger.examples.task-tracker.v1+json\"},\n\t\tConsumesMediaTypes: []string{\"application/vnd.goswagger.examples.task-tracker.v1+json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &CreateTaskReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t}\n\tfor _, opt := range opts {\n\t\topt(op)\n\t}\n\n\tresult, err := a.transport.Submit(op)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*CreateTaskCreated)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\tunexpectedSuccess := result.(*CreateTaskDefault)\n\treturn nil, runtime.NewAPIError(\"unexpected success response: content available as default response in error\", unexpectedSuccess, unexpectedSuccess.Code())\n}", "func (t *TaskService) Create(w http.ResponseWriter, r *http.Request) {\n\tvar ctx = r.Context()\n\tcurrentUser, err := common.UserFromCtx(ctx)\n\tif err != nil {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorCantDecodeInputData, \"TaskService.Create: %s\", \"can't get current user. err: %v\", err)\n\t\tcommon.SendBadRequest(w, r, errorcode.ErrorCantDecodeInputData)\n\t\treturn\n\t}\n\tlogger.Log.InfofCtx(r.Context(), currentUserFmt, currentUser)\n\n\ttask, err := t.extractPostTaskPayload(r, w)\n\tif err != nil {\n\t\treturn\n\t}\n\n\ttask.CreatedBy = currentUser.UID\n\ttask.PartnerID = currentUser.PartnerID\n\ttask.IsRequireNOCAccess = currentUser.IsNOCAccess\n\ttask.Schedule.StartRunTime = task.Schedule.StartRunTime.Truncate(time.Minute)\n\ttask.Schedule.EndRunTime = task.Schedule.EndRunTime.Truncate(time.Minute)\n\ttask.ID = gocql.TimeUUID()\n\tif task.TargetsByType == nil {\n\t\ttask.TargetsByType = make(models.TargetsByType)\n\t}\n\n\tif task.Targets.Type != 0 {\n\t\ttask.TargetsByType[task.Targets.Type] = task.Targets.IDs\n\t}\n\n\tfor targetType, targets := range task.TargetsByType {\n\t\ttask.Targets.Type = targetType\n\t\ttask.Targets.IDs = targets\n\t}\n\n\tsiteIDs, err := sites.GetSiteIDs(ctx, t.httpClient, currentUser.PartnerID, config.Config.SitesMsURL, currentUser.Token)\n\tif err != nil {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorCantCreateNewTask, \"TaskService.Create: User-Site restrictions for dynamic groups, err: %v\", err)\n\t\tcommon.SendInternalServerError(w, r, errorcode.ErrorCantCreateNewTask)\n\t\treturn\n\t}\n\n\tif ids, ok := task.TargetsByType[models.Site]; ok && len(ids) > 0 && !currentUser.IsNOCAccess {\n\t\tsite, ok := isUserSites(ids, siteIDs)\n\t\tif !ok {\n\t\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorCantCreateNewTask, \"TaskService.Create: User doesn't have access to site with id: %s\", site)\n\t\t\tcommon.SendForbidden(w, r, errorcode.ErrorCantCreateNewTask)\n\t\t\treturn\n\t\t}\n\t}\n\n\tif ids, ok := task.TargetsByType[models.DynamicSite]; ok && len(ids) > 0 && !currentUser.IsNOCAccess {\n\t\tsite, ok := isUserSites(ids, siteIDs)\n\t\tif !ok {\n\t\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorCantCreateNewTask, \"TaskService.Create: User doesn't have access to site with id: %s\", site)\n\t\t\tcommon.SendForbidden(w, r, errorcode.ErrorCantCreateNewTask)\n\t\t\treturn\n\t\t}\n\t}\n\n\terr = t.userSitesPersistence.InsertUserSites(r.Context(), currentUser.PartnerID, currentUser.UID, siteIDs)\n\tif err != nil {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorCantCreateNewTask, \"TaskService.Create: User-Site restrictions for dynamic groups, err: %v\", err)\n\t\tcommon.SendInternalServerError(w, r, errorcode.ErrorCantCreateNewTask)\n\t\treturn\n\t}\n\n\tif task.ExternalTask {\n\t\tgo t.createTaskBackground(ctx, task, currentUser.IsNOCAccess, r, siteIDs)\n\t\tcommon.RenderJSONCreated(w, task)\n\t\treturn\n\t}\n\n\tt.createTaskFlow(ctx, task, currentUser, r, w, siteIDs)\n}", "func (ten *Tenant) Commit() error {\n\tif ten.HasError() {\n\t\treturn ten.err\n\t}\n\n\terr := ten.Client.CreateTenant(ten.Tenant)\n\tif err != nil {\n\t\tten.err = err\n\t\tlog.Infof(\"Creating or updating Tenant failed %v\", err)\n\t\treturn err\n\t}\n\n\tlog.Debugf(\"Created VPC : %#v\", ten)\n\n\treturn nil\n}", "func (client BaseClient) CreateSystemSender(req *http.Request) (*http.Response, error) {\n\treturn autorest.SendWithSender(client, req,\n\t\tautorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n}", "func (suite *TenantTestSuite) TestCreateTenant() {\n\n\t// create json input data for the request\n\tpostData := `\n {\n \"info\":{\n\t\t\t\t\"name\":\"mutants\",\n\t\t\t\t\"email\":\"yo@yo\",\n\t\t\t\t\"website\":\"website\"\n\t\t\t},\n \"db_conf\": [\n {\n \"store\":\"ar\",\n \"server\":\"localhost\",\n \"port\":27017,\n \"database\":\"ar_db\",\n \"username\":\"admin\",\n \"password\":\"3NCRYPT3D\"\n },\n {\n \"store\":\"status\",\n \"server\":\"localhost\",\n \"port\":27017,\n \"database\":\"status_db\",\n \"username\":\"admin\",\n \"password\":\"3NCRYPT3D\"\n }],\n \"users\": [\n {\n \"name\":\"xavier\",\n \"email\":\"[email protected]\",\n \"api_key\":\"X4V13R\"\n },\n {\n \"name\":\"magneto\",\n \"email\":\"[email protected]\",\n \"api_key\":\"M4GN3T0\"\n }]\n }`\n\n\tjsonOutput := `{\n \"status\": {\n \"message\": \"Tenant was succesfully created\",\n \"code\": \"201\"\n },\n \"data\": {\n \"id\": \"{{ID}}\",\n \"links\": {\n \"self\": \"https:///api/v2/admin/tenants/{{ID}}\"\n }\n }\n}`\n\n\tjsonCreated := `{\n \"status\": {\n \"message\": \"Success\",\n \"code\": \"200\"\n },\n \"data\": [\n {\n \"id\": \"{{ID}}\",\n \"info\": {\n \"name\": \"mutants\",\n \"email\": \"yo@yo\",\n \"website\": \"website\",\n \"created\": \"{{TIMESTAMP}}\",\n \"updated\": \"{{TIMESTAMP}}\"\n },\n \"db_conf\": [\n {\n \"store\": \"ar\",\n \"server\": \"localhost\",\n \"port\": 27017,\n \"database\": \"ar_db\",\n \"username\": \"admin\",\n \"password\": \"3NCRYPT3D\"\n },\n {\n \"store\": \"status\",\n \"server\": \"localhost\",\n \"port\": 27017,\n \"database\": \"status_db\",\n \"username\": \"admin\",\n \"password\": \"3NCRYPT3D\"\n }\n ],\n \"users\": [\n {\n \"name\": \"xavier\",\n \"email\": \"[email protected]\",\n \"api_key\": \"X4V13R\"\n },\n {\n \"name\": \"magneto\",\n \"email\": \"[email protected]\",\n \"api_key\": \"M4GN3T0\"\n }\n ]\n }\n ]\n}`\n\n\trequest, _ := http.NewRequest(\"POST\", \"/api/v2/admin/tenants\", strings.NewReader(postData))\n\trequest.Header.Set(\"x-api-key\", suite.clientkey)\n\trequest.Header.Set(\"Accept\", \"application/json\")\n\tresponse := httptest.NewRecorder()\n\n\tsuite.router.ServeHTTP(response, request)\n\n\t// Grab ID from mongodb\n\tsession, err := mgo.Dial(suite.cfg.MongoDB.Host)\n\tdefer session.Close()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\t// Retrieve id from database\n\tvar result map[string]interface{}\n\tc := session.DB(suite.cfg.MongoDB.Db).C(\"tenants\")\n\n\tc.Find(bson.M{\"info.name\": \"mutants\"}).One(&result)\n\tid := result[\"id\"].(string)\n\tinfo := result[\"info\"].(map[string]interface{})\n\ttimestamp := info[\"created\"].(string)\n\n\tcode := response.Code\n\toutput := response.Body.String()\n\n\tsuite.Equal(201, code, \"Internal Server Error\")\n\t// Apply id to output template and check\n\tsuite.Equal(strings.Replace(jsonOutput, \"{{ID}}\", id, 2), output, \"Response body mismatch\")\n\n\t// Check that actually the item has been created\n\t// Call List one with the specific ID\n\trequest2, _ := http.NewRequest(\"GET\", \"/api/v2/admin/tenants/\"+id, strings.NewReader(\"\"))\n\trequest2.Header.Set(\"x-api-key\", suite.clientkey)\n\trequest2.Header.Set(\"Accept\", \"application/json\")\n\tresponse2 := httptest.NewRecorder()\n\n\tsuite.router.ServeHTTP(response2, request2)\n\n\tcode2 := response2.Code\n\toutput2 := response2.Body.String()\n\t// Check that we must have a 200 ok code\n\tsuite.Equal(200, code2, \"Internal Server Error\")\n\n\tjsonCreated = strings.Replace(jsonCreated, \"{{ID}}\", id, 1)\n\tjsonCreated = strings.Replace(jsonCreated, \"{{TIMESTAMP}}\", timestamp, 2)\n\t// Compare the expected and actual json response\n\tsuite.Equal(jsonCreated, output2, \"Response body mismatch\")\n\n}", "func createDefaultTenant() {\n\t// tenant params\n\ttenant := client.Tenant{\n\t\tTenantName: \"default\",\n\t}\n\n\t// create a tenant\n\terr := contivClient.TenantPost(&tenant)\n\tif err != nil {\n\t\tlog.Fatalf(\"Error creating default tenant. Err: %v\", err)\n\t}\n\n\t// Get the tenant and verify it exists\n\tgotTenant, err := contivClient.TenantGet(\"default\")\n\tif err != nil {\n\t\tlog.Fatalf(\"Error getting default tenant. Err: %v\", err)\n\t}\n\n\tif gotTenant.TenantName != tenant.TenantName {\n\t\tlog.Fatalf(\"Got invalid tenant name. expecting %s. Got %s\", tenant.TenantName, gotTenant.TenantName)\n\t}\n}", "func (cmd *CreateTechHyTechCommand) Run(c *client.Client, args []string) error {\n\tvar path string\n\tif len(args) > 0 {\n\t\tpath = args[0]\n\t} else {\n\t\tpath = \"/api/tech\"\n\t}\n\tvar payload client.CreateTechHyTechPayload\n\tif cmd.Payload != \"\" {\n\t\terr := json.Unmarshal([]byte(cmd.Payload), &payload)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"failed to deserialize payload: %s\", err)\n\t\t}\n\t}\n\tlogger := goa.NewLogger(log.New(os.Stderr, \"\", log.LstdFlags))\n\tctx := goa.WithLogger(context.Background(), logger)\n\tresp, err := c.CreateTechHyTech(ctx, path, &payload, cmd.ContentType)\n\tif err != nil {\n\t\tgoa.LogError(ctx, \"failed\", \"err\", err)\n\t\treturn err\n\t}\n\n\tgoaclient.HandleResponse(c.Client, resp, cmd.PrettyPrint)\n\treturn nil\n}", "func (c *BulkDeletesController) Create(ctx *gin.Context) {\n\trequest := models.BulkDeleteRunRequest{}\n\tif err := ctx.ShouldBindJSON(&request); err != nil {\n\t\tctx.AbortWithError(422, err)\n\t} else if task, err := models.NewBulkDeleteRunTask(request); err != nil {\n\t\tctx.AbortWithError(422, err)\n\t} else if err := c.App.GetStore().Save(task); err != nil {\n\t\tctx.AbortWithError(500, err)\n\t} else if doc, err := jsonapi.Marshal(task); err != nil {\n\t\tctx.AbortWithError(500, err)\n\t} else {\n\t\tc.App.WakeBulkRunDeleter()\n\t\tctx.Data(201, MediaType, doc)\n\t}\n}", "func (c *Client) CreateBkOpsTask(url string, paras *CreateTaskPathParas,\n\trequest *CreateTaskRequest) (*CreateTaskResponse, error) {\n\tif c == nil {\n\t\treturn nil, ErrServerNotInit\n\t}\n\n\tvar (\n\t\treqURL = fmt.Sprintf(\"/create_task/%s/%s/\", paras.TemplateID, paras.BkBizID)\n\t\trespData = &CreateTaskResponse{}\n\t)\n\n\tuserAuth, err := c.generateGateWayAuth(paras.Operator)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"bksops CreateBkOpsTask generateGateWayAuth failed: %v\", err)\n\t}\n\trequest.FlowType = string(CommonFlow)\n\t// TemplateSource 模版来源, 默认是业务流程; 可由用户自定义\n\tif request.TemplateSource == \"\" {\n\t\trequest.TemplateSource = string(BusinessTpl)\n\t}\n\n\t_, _, errs := gorequest.New().\n\t\tTimeout(defaultTimeOut).\n\t\tPost(c.server+reqURL).\n\t\tSet(\"Content-Type\", \"application/json\").\n\t\tSet(\"Accept\", \"application/json\").\n\t\tSet(\"X-Bkapi-Authorization\", userAuth).\n\t\tSetDebug(c.serverDebug).\n\t\tSend(request).\n\t\tEndStruct(&respData)\n\tif len(errs) > 0 {\n\t\tblog.Errorf(\"call api CreateBkOpsTask failed: %v\", errs[0])\n\t\treturn nil, errs[0]\n\t}\n\n\tif !respData.Result {\n\t\tblog.Errorf(\"call api CreateBkOpsTask failed: %v\", respData.Message)\n\t\treturn nil, fmt.Errorf(respData.Message)\n\t}\n\t//successfully request\n\tblog.Infof(\"call api CreateBkOpsTask with url(%s) successfully\", reqURL)\n\treturn respData, nil\n}", "func (*CreateTenantRequest) Descriptor() ([]byte, []int) {\n\treturn file_packetbroker_api_iam_v1_service_proto_rawDescGZIP(), []int{11}\n}", "func deleteTenant(c *cli.Context) error {\n\terr := checkArgCount(c, 1)\n\tif err != nil {\n\t\treturn err\n\t}\n\tid := c.Args().First()\n\n\tclient.Photonclient, err = client.GetClient(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdeleteTask, err := client.Photonclient.Tenants.Delete(id)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = waitOnTaskOperation(deleteTask.ID, c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = clearConfigTenant(id)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func resourceNetboxOrgTenantGroupCreate(d *schema.ResourceData, meta interface{}) error {\n\tnetboxClient := meta.(*ProviderNetboxClient).client\n\n\tname := d.Get(\"name\").(string)\n\tslug := d.Get(\"slug\").(string)\n\n\tvar parm = tenancy.NewTenancyTenantGroupsCreateParams().WithData(\n\t\t&models.TenantGroup{\n\t\t\tName: &name,\n\t\t\tSlug: &slug,\n\t\t},\n\t)\n\n\tlog.Debugf(\"Executing TenancyTenantGroupsCreate against Netbox: %v\", parm)\n\n\tout, err := netboxClient.Tenancy.TenancyTenantGroupsCreate(parm, nil)\n\n\tif err != nil {\n\t\tlog.Debugf(\"Failed to execute TenancyTenantGroupsCreate: %v\", err)\n\n\t\treturn err\n\t}\n\n\t// TODO Probably a better way to parse this ID\n\td.SetId(fmt.Sprintf(\"org/tenant-group/%d\", out.Payload.ID))\n\td.Set(\"tenant_group_id\", out.Payload.ID)\n\n\tlog.Debugf(\"Done Executing TenancyTenantGroupsCreate: %v\", out)\n\n\treturn nil\n}", "func CreateTask(w http.ResponseWriter, r *http.Request, _ httprouter.Params) {\n\tt := Task{}\n\tdecoder := json.NewDecoder(r.Body)\n\terr := decoder.Decode(&t)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\terr = store.CreateTask(&t)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tw.WriteHeader(http.StatusCreated)\n}", "func (c Client) CreateApplication(name string, email string, accounts string, cloudProviders string, instancePort string, description string) (types.TaskRef, error) {\n var taskRef types.TaskRef\n var jobs []types.Job\n\n a := strings.Split(accounts, \",\")\n\n for _, account := range a {\n jobs = append(jobs, types.Job {\n Type: \"createApplication\",\n Account: account,\n User: \"\",\n Application: types.CreateApplication {\n Name: name,\n Description: description,\n Accounts: accounts,\n CloudProviders: cloudProviders,\n Email: email,\n InstancePort: instancePort,\n },\n })\n }\n\n task := types.Task {\n Application: name,\n Description: \"Create Application: \" + name,\n Job : jobs,\n }\n\n resp, err := c.post(\"/applications/\" + name + \"/tasks\", task)\n defer ensureReaderClosed(resp)\n if err != nil {\n return taskRef, err\n }\n\n err = json.NewDecoder(resp.body).Decode(&taskRef)\n return taskRef, err\n}", "func MythicRPCTaskCreateSubtask(input MythicRPCTaskCreateSubtaskMessage) MythicRPCTaskCreateSubtaskMessageResponse {\n\tresponse := MythicRPCTaskCreateSubtaskMessageResponse{\n\t\tSuccess: false,\n\t}\n\ttaskingLocation := \"mythic_rpc\"\n\tcreateTaskInput := CreateTaskInput{\n\t\tParentTaskID: &input.TaskID,\n\t\tCommandName: input.CommandName,\n\t\tParams: input.Params,\n\t\tToken: input.Token,\n\t\tParameterGroupName: input.ParameterGroupName,\n\t\tSubtaskCallbackFunction: input.SubtaskCallbackFunction,\n\t\tTaskingLocation: &taskingLocation,\n\t}\n\ttask := databaseStructs.Task{}\n\toperatorOperation := databaseStructs.Operatoroperation{}\n\tif err := database.DB.Get(&task, `SELECT \n\tcallback.id \"callback.id\",\n\tcallback.display_id \"callback.display_id\",\n\tcallback.operation_id \"callback.operation_id\",\n\toperator.id \"operator.id\",\n\toperator.admin \"operator.admin\" \n\tFROM task\n\tJOIN callback ON task.callback_id = callback.id \n\tJOIN operator ON task.operator_id = operator.id\n\tWHERE task.id=$1`, input.TaskID); err != nil {\n\t\tresponse.Error = err.Error()\n\t\tlogging.LogError(err, \"Failed to fetch task/callback information when creating subtask\")\n\t\treturn response\n\t} else if err := database.DB.Get(&operatorOperation, `SELECT\n\tbase_disabled_commands_id\n\tFROM operatoroperation\n\tWHERE operator_id = $1 AND operation_id = $2\n\t`, task.Operator.ID, task.Callback.OperationID); err != nil {\n\t\tlogging.LogError(err, \"Failed to get operation information when creating subtask\")\n\t\tresponse.Error = err.Error()\n\t\treturn response\n\t} else {\n\t\tcreateTaskInput.IsOperatorAdmin = task.Operator.Admin\n\t\tcreateTaskInput.CallbackDisplayID = task.Callback.DisplayID\n\t\tcreateTaskInput.CurrentOperationID = task.Callback.OperationID\n\t\tif operatorOperation.BaseDisabledCommandsID.Valid {\n\t\t\tbaseDisabledCommandsID := int(operatorOperation.BaseDisabledCommandsID.Int64)\n\t\t\tcreateTaskInput.DisabledCommandID = &baseDisabledCommandsID\n\t\t}\n\t\tcreateTaskInput.OperatorID = task.Operator.ID\n\t\t// create a subtask of this task\n\t\tcreationResponse := CreateTask(createTaskInput)\n\t\tif creationResponse.Status == \"success\" {\n\t\t\tresponse.Success = true\n\t\t\tresponse.TaskID = creationResponse.TaskID\n\t\t} else {\n\t\t\tresponse.Error = creationResponse.Error\n\t\t}\n\t\treturn response\n\t}\n\n}", "func CmdCreateDomain(c *cli.Context) error {\n\tctx := c.Context\n\tsess := edgegrid.GetSession(ctx)\n\tclient := gtm.Client(sess)\n\n\t// tfWorkPath is a target directory for generated terraform resources\n\tvar tfWorkPath = \"./\"\n\tif c.IsSet(\"tfworkpath\") {\n\t\ttfWorkPath = c.String(\"tfworkpath\")\n\t}\n\n\tdatacentersPath := filepath.Join(tfWorkPath, \"datacenters.tf\")\n\tdomainPath := filepath.Join(tfWorkPath, \"domain.tf\")\n\timportPath := filepath.Join(tfWorkPath, \"import.sh\")\n\tmapsPath := filepath.Join(tfWorkPath, \"maps.tf\")\n\tpropertiesPath := filepath.Join(tfWorkPath, \"properties.tf\")\n\tresourcesPath := filepath.Join(tfWorkPath, \"resources.tf\")\n\tvariablesPath := filepath.Join(tfWorkPath, \"variables.tf\")\n\n\ttemplateToFile := map[string]string{\n\t\t\"datacenters.tmpl\": datacentersPath,\n\t\t\"domain.tmpl\": domainPath,\n\t\t\"imports.tmpl\": importPath,\n\t\t\"maps.tmpl\": mapsPath,\n\t\t\"properties.tmpl\": propertiesPath,\n\t\t\"resources.tmpl\": resourcesPath,\n\t\t\"variables.tmpl\": variablesPath,\n\t}\n\n\terr := tools.CheckFiles(datacentersPath, domainPath, importPath, mapsPath, propertiesPath, resourcesPath, variablesPath)\n\tif err != nil {\n\t\treturn cli.Exit(color.RedString(err.Error()), 1)\n\t}\n\n\tprocessor := templates.FSTemplateProcessor{\n\t\tTemplatesFS: templateFiles,\n\t\tTemplateTargets: templateToFile,\n\t\tAdditionalFuncs: template.FuncMap{\n\t\t\t\"normalize\": normalizeResourceName,\n\t\t\t\"toUpper\": strings.ToUpper,\n\t\t\t\"isDefaultDC\": isDefaultDatacenter,\n\t\t\t\"escapeString\": tools.EscapeQuotedStringLit,\n\t\t},\n\t}\n\n\tdomainName := c.Args().First()\n\tsection := edgegrid.GetEdgercSection(c)\n\tif err := createDomain(ctx, client, domainName, section, processor); err != nil {\n\t\treturn cli.Exit(color.RedString(fmt.Sprintf(\"Error exporting domain HCL: %s\", err)), 1)\n\t}\n\treturn nil\n}", "func (client DatasetClient) CreateSender(req *http.Request) (future DatasetCreateFuture, err error) {\n var resp *http.Response\n resp, err = client.Send(req, autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n if err != nil {\n return\n }\n var azf azure.Future\n azf, err = azure.NewFutureFromResponse(resp)\n future.FutureAPI = &azf\n future.Result = future.result\n return\n }", "func CreateCreateTagTaskRequest() (request *CreateTagTaskRequest) {\n\trequest = &CreateTagTaskRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"viapi-regen\", \"2021-11-19\", \"CreateTagTask\", \"selflearning\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func (*TenantCreateRequest) Descriptor() ([]byte, []int) {\n\treturn file_v1_tenant_proto_rawDescGZIP(), []int{1}\n}", "func CreateTenant(stateDriver core.StateDriver, tenant *intent.ConfigTenant) error {\n\treturn validateTenantConfig(tenant)\n}", "func (a *App) create(c *echo.Context) error {\n\ttask := &model.Task{}\n\tvar err error\n\n\terr = c.Bind(task)\n\tif err != nil {\n\t\tc.JSON(http.StatusInternalServerError, ErrorMsg{Msg: err.Error()})\n\t\treturn err\n\t}\n\tstatus, err := task.Create(a.GetDB())\n\tc.JSON(status, ErrorMsg{Msg: err.Error()})\n\treturn err\n}", "func (ts *TaskService) Create(ctx context.Context, req *taskAPI.CreateTaskRequest) (*taskAPI.CreateTaskResponse, error) {\n\tdefer logPanicAndDie(log.G(ctx))\n\tlogger := log.G(ctx).WithFields(logrus.Fields{\"id\": req.ID, \"bundle\": req.Bundle})\n\tlogger.Info(\"create\")\n\n\textraData, err := unmarshalExtraData(req.Options)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbundleDir := bundle.Dir(filepath.Join(containerRootDir, req.ID))\n\terr = bundleDir.Create()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = bundleDir.OCIConfig().Write(extraData.JsonSpec)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// TODO replace with proper drive mounting once that PR is merged. Right now, all containers in\n\t// this VM start up with the same rootfs image no matter their configuration\n\terr = bundleDir.MountRootfs(\"/dev/vdb\", \"ext4\", nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Create a runc shim to manage this task\n\t// TODO if we update to the v2 runc implementation in containerd, we can use a single\n\t// runc service instance to manage all tasks instead of creating a new one for each\n\truncService, err := runc.New(ctx, req.ID, ts.eventExchange)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Override the incoming stdio FIFOs, which have paths from the host that we can't use\n\tfifoSet, err := cio.NewFIFOSetInDir(bundleDir.RootPath(), req.ID, req.Terminal)\n\tif err != nil {\n\t\tlog.G(ctx).WithError(err).Error(\"failed opening stdio FIFOs\")\n\t\treturn nil, err\n\t}\n\n\t// Don't try to connect any io streams that weren't requested by the client\n\tif req.Stdin == \"\" {\n\t\tfifoSet.Stdin = \"\"\n\t}\n\n\tif req.Stdout == \"\" {\n\t\tfifoSet.Stdout = \"\"\n\t}\n\n\tif req.Stderr == \"\" {\n\t\tfifoSet.Stderr = \"\"\n\t}\n\n\ttask, err := ts.taskManager.AddTask(ctx, req.ID, runcService, bundleDir, extraData, fifoSet)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlogger.Debug(\"calling runc create\")\n\n\t// Override some of the incoming paths, which were set on the Host and thus not valid here in the Guest\n\treq.Bundle = bundleDir.RootPath()\n\treq.Rootfs = nil\n\treq.Stdin = fifoSet.Stdin\n\treq.Stdout = fifoSet.Stdout\n\treq.Stderr = fifoSet.Stderr\n\n\t// Just provide runc the options it knows about, not our wrapper\n\treq.Options = task.ExtraData().GetRuncOptions()\n\n\t// Start the io proxy and wait for initialization to complete before starting\n\t// the task to ensure we capture all task output\n\terr = <-task.StartStdioProxy(ctx, vm.VSockToFIFO, acceptVSock)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresp, err := task.Create(ctx, req)\n\tif err != nil {\n\t\tlogger.WithError(err).Error(\"error creating container\")\n\t\treturn nil, err\n\t}\n\n\tlogger.WithField(\"pid\", resp.Pid).Debugf(\"create succeeded\")\n\treturn resp, nil\n}", "func InitTenant(ctx context.Context, config tenantConfig) error {\n\n\tc, err := createClient(ctx, config)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Ignore response for now\n\t_, err = c.SetupTenant(goasupport.ForwardContextRequestID(ctx), tenant.SetupTenantPath())\n\n\treturn err\n}", "func (client DeviceClient) CreateSender(req *http.Request) (*http.Response, error) {\n\treturn client.Send(req, autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n}", "func TestTenantAddDelete(t *testing.T) {\n\t// tenant params\n\ttenant := client.Tenant{\n\t\tTenantName: \"tenant1\",\n\t}\n\n\t// create a tenant\n\terr := contivClient.TenantPost(&tenant)\n\tcheckError(t, \"create tenant\", err)\n\n\t// Get the tenant and verify it exists\n\tgotTenant, err := contivClient.TenantGet(\"tenant1\")\n\tcheckError(t, \"get tenant\", err)\n\n\tif gotTenant.TenantName != tenant.TenantName {\n\t\tt.Fatalf(\"Got invalid tenant name. expecting %s. Got %s\", tenant.TenantName, gotTenant.TenantName)\n\t}\n\n\t// delete tenant\n\terr = contivClient.TenantDelete(\"tenant1\")\n\tcheckError(t, \"delete tenant\", err)\n\n\t// find again and make sure its gone\n\t_, err = contivClient.TenantGet(\"tenant1\")\n\tif err == nil {\n\t\tt.Fatalf(\"Tenant was not deleted\")\n\t}\n}", "func CreateTaskController(w http.ResponseWriter, r *http.Request) {\n\tvar task Task\n\n\terr := json.NewDecoder(r.Body).Decode(&task)\n\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\ttask.Status = \"P\"\n\terr = task.Create()\n\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tresponse, _ := json.Marshal(task)\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(http.StatusOK)\n\tw.Write(response)\n}", "func InitiateRakeTask(taskName string, settings *models.Settings) {\n\trakeTask := map[string]string{}\n\tb, err := json.Marshal(rakeTask)\n\tif err != nil {\n\t\tfmt.Println(err.Error())\n\t\tos.Exit(1)\n\t}\n\tencodedTaskName, err := url.Parse(taskName)\n\tif err != nil {\n\t\tfmt.Println(err.Error())\n\t\tos.Exit(1)\n\t}\n\thttpclient.Post(b, fmt.Sprintf(\"%s/v1/environments/%s/services/%s/rake/%s\", settings.PaasHost, settings.EnvironmentID, settings.ServiceID, encodedTaskName), true, settings)\n}", "func TaskCreate(w http.ResponseWriter, r *http.Request) {\n\ttask := task.NewTask()\n\tbody, err := ioutil.ReadAll(io.LimitReader(r.Body, 2<<19))\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tif err := r.Body.Close(); err != nil {\n\t\tpanic(err)\n\t}\n\tif err := json.Unmarshal(body, task); err != nil {\n\t\tglog.Error(err)\n\t\tw.Header().Set(\"Content-Type\", \"application/json; charset=UTF-8\")\n\t\tw.WriteHeader(http.StatusUnprocessableEntity) // unprocessable entity\n\t\tif err := json.NewEncoder(w).Encode(err); err != nil {\n\t\t\tglog.Error(err)\n\t\t}\n\t\treturn\n\t}\n\n\tif len(task.Destinations) == 0 {\n\t\tglog.Error(\"List of destinations are empty!\")\n\t\tw.WriteHeader(http.StatusUnprocessableEntity)\n\t\treturn\n\t}\n\n\tfor _, v := range task.Destinations {\n\t\tif len(*v) == 0 {\n\t\t\tglog.Error(\"List of destinations could not be empty!\")\n\t\t\tw.WriteHeader(http.StatusUnprocessableEntity)\n\t\t\treturn\n\t\t}\n\t}\n\n\tglog.Info(string(body))\n\tgo workers.Handler(task)\n}", "func NewCmdCleanUpTenant(f cmdutil.Factory) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"tenant\",\n\t\tShort: \"Hard delete of your tenant pipelines, apps, jobs and releases\",\n\t\tLong: `Hard delete of your tenant pipelines, apps, jobs and releases`,\n\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\tp := cleanUpTenantFlags{}\n\t\t\tif cmd.Flags().Lookup(yesFlag).Value.String() == \"true\" {\n\t\t\t\tp.confirm = true\n\t\t\t}\n\t\t\terr := p.cleanTenant(f)\n\t\t\tif err != nil {\n\t\t\t\tutil.Fatalf(\"%s\\n\", err)\n\t\t\t}\n\t\t\treturn\n\t\t},\n\t}\n\treturn cmd\n}", "func (api *tenantAPI) Delete(obj *cluster.Tenant) error {\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = apicl.ClusterV1().Tenant().Delete(context.Background(), &obj.ObjectMeta)\n\t\treturn err\n\t}\n\n\tapi.ct.handleTenantEvent(&kvstore.WatchEvent{Object: obj, Type: kvstore.Deleted})\n\treturn nil\n}", "func getTenantTasks(c *cli.Context, w io.Writer) error {\n\terr := checkArgCount(c, 1)\n\tif err != nil {\n\t\treturn err\n\t}\n\tid := c.Args().First()\n\n\tstate := c.String(\"state\")\n\toptions := &photon.TaskGetOptions{\n\t\tState: state,\n\t}\n\n\tclient.Photonclient, err = client.GetClient(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttaskList, err := client.Photonclient.Tenants.GetTasks(id, options)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = printTaskList(taskList.Items, c)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (oo *OmciCC) SendCreateTDVar(ctx context.Context, timeout int, highPrio bool, rxChan chan Message, params ...me.ParamData) (*me.ManagedEntity, error) {\n\ttid := oo.GetNextTid(highPrio)\n\tlogger.Debugw(ctx, \"send TD-Create-msg:\", log.Fields{\"device-id\": oo.deviceID,\n\t\t\"SequNo\": strconv.FormatInt(int64(tid), 16),\n\t\t\"InstId\": strconv.FormatInt(int64(params[0].EntityID), 16)})\n\tmeInstance, omciErr := me.NewTrafficDescriptor(params[0])\n\tif omciErr.GetError() == nil {\n\t\tomciLayer, msgLayer, err := oframe.EncodeFrame(meInstance, omci.CreateRequestType, oframe.TransactionID(tid))\n\t\tif err != nil {\n\t\t\tlogger.Errorw(ctx, \"Cannot encode TD for create\", log.Fields{\"Err\": err, \"device-id\": oo.deviceID})\n\t\t\treturn nil, err\n\t\t}\n\t\tpkt, err := SerializeOmciLayer(ctx, omciLayer, msgLayer)\n\t\tif err != nil {\n\t\t\tlogger.Errorw(ctx, \"Cannot serialize TD create\", log.Fields{\"Err\": err, \"device-id\": oo.deviceID})\n\t\t\treturn nil, err\n\t\t}\n\t\tomciRxCallbackPair := CallbackPair{\n\t\t\tCbKey: tid,\n\t\t\tCbEntry: CallbackPairEntry{rxChan, oo.receiveOmciResponse, true},\n\t\t}\n\t\terr = oo.Send(ctx, pkt, timeout, CDefaultRetries, highPrio, omciRxCallbackPair)\n\t\tif err != nil {\n\t\t\tlogger.Errorw(ctx, \"Cannot send TD create\", log.Fields{\"Err\": err, \"device-id\": oo.deviceID})\n\t\t\treturn nil, err\n\t\t}\n\t\tlogger.Debug(ctx, \"send TD-Create-msg done\")\n\t\treturn meInstance, nil\n\t}\n\tlogger.Errorw(ctx, \"Cannot generate TD Instance\", log.Fields{\"Err\": omciErr.GetError(), \"device-id\": oo.deviceID})\n\treturn nil, omciErr.GetError()\n}", "func TenantDelete(tenantID string) {\n\tRunCmd(fmt.Sprintf(\"%s tenant -op=del -id=%s\", ActlPath, tenantID))\n}", "func CreateTxCmd(cdc *wire.Codec) *cobra.Command {\n\treturn &cobra.Command{\n\t\tUse: \"create <ticker> <name> <decimals> <reserve_ratio> <initial_supply> <initial_rune_amount>\",\n\t\tShort: \"Create a token with CLP\",\n\t\tArgs: cobra.ExactArgs(6),\n\t\tRunE: func(_ *cobra.Command, args []string) error {\n\t\t\ttxCtx := authctx.NewTxContextFromCLI().WithCodec(cdc)\n\t\t\tcliCtx := context.NewCLIContext().\n\t\t\t\tWithCodec(cdc).\n\t\t\t\tWithLogger(os.Stdout).\n\t\t\t\tWithAccountDecoder(authcmd.GetAccountDecoder(cdc))\n\n\t\t\t// get the from address from the name flag\n\t\t\tfrom, err := cliCtx.GetFromAddress()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\t// create the message\n\t\t\tticker := args[0]\n\t\t\tname := args[1]\n\t\t\tdecimalsInt, _ := strconv.Atoi(args[2])\n\t\t\treserveRatio, _ := strconv.Atoi(args[3])\n\n\t\t\tif decimalsInt < 0 || decimalsInt > 255 {\n\t\t\t\treturn clp.ErrInvalidDecimals(clp.DefaultCodespace)\n\t\t\t}\n\n\t\t\tdecimals := uint8(decimalsInt)\n\n\t\t\tinitialSupply, _ := strconv.Atoi(args[4])\n\t\t\tinitialBaseCoinAmount, _ := strconv.Atoi(args[5])\n\t\t\tmsg := clpTypes.NewMsgCreate(from, ticker, name, decimals, reserveRatio, int64(initialSupply), int64(initialBaseCoinAmount))\n\n\t\t\t// Build and sign the transaction, then broadcast to a Tendermint\n\t\t\t// node.\n\t\t\treturn utils.SendTx(txCtx, cliCtx, []sdk.Msg{msg})\n\t\t},\n\t}\n}", "func (*CreateTenantResponse) Descriptor() ([]byte, []int) {\n\treturn file_packetbroker_api_iam_v1_service_proto_rawDescGZIP(), []int{12}\n}", "func HandleCreateTask(w http.ResponseWriter, r *http.Request) {\n\tlog.Root.Info(\"HandleCreateTask BEGIN\")\n\n\tif r.Method != http.MethodPost {\n\t\tw.WriteHeader(http.StatusNotFound)\n\t\tHttpResponseError(w, ErrNotFound)\n\t\treturn\n\t}\n\n\tdefer r.Body.Close()\n\tbody, _ := ioutil.ReadAll(r.Body)\n\n\tdata := make(map[string]interface{})\n\terr := json.Unmarshal(body, &data)\n\tif err != nil {\n\t\tlog.Root.Error(\"HandleCreateTask Parse HTTP request body error\")\n\t\tHttpResponseError(w, ErrForm)\n\t\treturn\n\t}\n\n\telem, ok := data[\"taskPara\"]\n\tif !ok {\n\t\tlog.Root.Error(\"HandleCreateTask HTTP form data error\")\n\t\tHttpResponseError(w, ErrForm)\n\t\treturn\n\t}\n\n\ttaskPara := elem.(string)\n\ttaskID, err := node.CreateTask(taskPara)\n\tif err != nil {\n\t\tlog.Root.Error(\"HandleCreateTask Create task error\")\n\t\tHttpResponseError(w, ErrServer)\n\t\treturn\n\t}\n\n\tlog.Root.Info(\"HandleCreateTask END\")\n\tHttpResponseData(w, H{\n\t\t\"taskID\": taskID,\n\t})\n\treturn\n}", "func (client QuotaRequestClient) CreateSender(req *http.Request) (future QuotaRequestCreateFuture, err error) {\n\tvar resp *http.Response\n\tresp, err = client.Send(req, azure.DoRetryWithRegistration(client.Client))\n\tif err != nil {\n\t\treturn\n\t}\n\tfuture.Future, err = azure.NewFutureFromResponse(resp)\n\treturn\n}", "func (c *Client) CreateTask(tr TaskRequest) (task Task, err error) {\n\tbodyReader, err := createReader(tr)\n\tif err != nil {\n\t\treturn task, err\n\t}\n\n\trequest := fmt.Sprintf(\"/v3/apps/%s/tasks\", tr.DropletGUID)\n\treq := c.NewRequestWithBody(\"POST\", request, bodyReader)\n\n\tresp, err := c.DoRequest(req)\n\tif err != nil {\n\t\treturn task, errors.Wrap(err, \"Error creating task\")\n\t}\n\tdefer resp.Body.Close()\n\n\tbody, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn task, errors.Wrap(err, \"Error reading task after creation\")\n\t}\n\n\terr = json.Unmarshal(body, &task)\n\tif err != nil {\n\t\treturn task, errors.Wrap(err, \"Error unmarshaling task\")\n\t}\n\treturn task, err\n}", "func (client JobClient) CreateSender(req *http.Request) (*http.Response, error) {\n\treturn client.Send(req, azure.DoRetryWithRegistration(client.Client))\n}", "func CreateTask(w http.ResponseWriter, r *http.Request) {\n\tvar task models.Task\n\terr := json.NewDecoder(r.Body).Decode(&task)\n\n\tif err != nil {\n\t\thttp.Error(w, http.StatusText(400), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tnewTask, err := repository.CreateTask(task)\n\n\tif err != nil {\n\t\thttp.Error(w, http.StatusText(500), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(http.StatusCreated)\n\n\tjson.NewEncoder(w).Encode(newTask)\n}", "func (tm *TenantsManager) CreateTenants(ctx context.Context, tenants []model.BusinessTenantMappingInput) error {\n\ttenantsToCreateGQL := tm.tenantConverter.MultipleInputToGraphQLInput(tenants)\n\treturn runInChunks(ctx, tm.config.TenantOperationChunkSize, tenantsToCreateGQL, func(ctx context.Context, chunk []graphql.BusinessTenantMappingInput) error {\n\t\treturn tm.gqlClient.WriteTenants(ctx, chunk)\n\t})\n}", "func (s *TaskTypesEndpoint) Create(ctx context.Context, division int, entity *TaskTypes) (*TaskTypes, error) {\n\tu, _ := s.client.ResolvePathWithDivision(\"/api/v1/{division}/accountancy/TaskTypes\", division) // #nosec\n\te := &TaskTypes{}\n\t_, _, err := s.client.NewRequestAndDo(ctx, \"POST\", u.String(), entity, e)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn e, nil\n}", "func (client IngestionSettingsClient) CreateSender(req *http.Request) (*http.Response, error) {\n\treturn client.Send(req, azure.DoRetryWithRegistration(client.Client))\n}", "func (opts CreateOpts) ToTenantCreateMap() (map[string]interface{}, error) {\r\n\treturn gophercloud.BuildRequestBody(opts, \"tenant\")\r\n}", "func (t TaskService) CreateTask(ctx context.Context, tc taskmodel.TaskCreate) (*taskmodel.Task, error) {\n\tspan, _ := tracing.StartSpanFromContext(ctx)\n\tdefer span.Finish()\n\tvar tr taskResponse\n\n\terr := t.Client.\n\t\tPostJSON(tc, prefixTasks).\n\t\tDecodeJSON(&tr).\n\t\tDo(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn convertTask(tr.Task), nil\n}", "func TenantID() string {\n\treturn tenantID\n}", "func TenantID() string {\n\treturn tenantID\n}", "func TenantID() string {\n\treturn tenantID\n}", "func (client BaseClient) CreateSystem(ctx context.Context, body *SystemInputs) (result System, err error) {\n\tif err := validation.Validate([]validation.Validation{\n\t\t{TargetValue: body,\n\t\t\tConstraints: []validation.Constraint{{Target: \"body\", Name: validation.Null, Rule: false,\n\t\t\t\tChain: []validation.Constraint{{Target: \"body.Name\", Name: validation.Null, Rule: true, Chain: nil},\n\t\t\t\t\t{Target: \"body.Tenant\", Name: validation.Null, Rule: true, Chain: nil},\n\t\t\t\t}}}}}); err != nil {\n\t\treturn result, validation.NewError(\"beacon.BaseClient\", \"CreateSystem\", err.Error())\n\t}\n\n\treq, err := client.CreateSystemPreparer(ctx, body)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"CreateSystem\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresp, err := client.CreateSystemSender(req)\n\tif err != nil {\n\t\tresult.Response = autorest.Response{Response: resp}\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"CreateSystem\", resp, \"Failure sending request\")\n\t\treturn\n\t}\n\n\tresult, err = client.CreateSystemResponder(resp)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"CreateSystem\", resp, \"Failure responding to request\")\n\t}\n\n\treturn\n}", "func (client ThreatIntelligenceIndicatorClient) CreateSender(req *http.Request) (*http.Response, error) {\n\treturn client.Send(req, azure.DoRetryWithRegistration(client.Client))\n}", "func (client GroupClient) CreateSecretSender(req *http.Request) (*http.Response, error) {\n\treturn autorest.SendWithSender(client,\n\t\treq,\n\t\tautorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n}", "func GenTxCmd(ctx *server.Context, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"gentx\",\n\t\tShort: \"Generate a genesis tx carrying a self delegation\",\n\t\tLong: fmt.Sprintf(`This command is an alias of the 'gaiad tx create-validator' command'.\n\nIt creates a genesis piece carrying a self delegation with the\nfollowing delegation and commission default parameters:\n\n\tdelegation amount: %s\n\tcommission rate: %s\n\tcommission max rate: %s\n\tcommission max change rate: %s\n`, defaultAmount, defaultCommissionRate, defaultCommissionMaxRate, defaultCommissionMaxChangeRate),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\n\t\t\tconfig := ctx.Config\n\t\t\tconfig.SetRoot(viper.GetString(tmcli.HomeFlag))\n\t\t\tnodeID, valPubKey, err := InitializeNodeValidatorFiles(ctx.Config)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tip, err := server.ExternalIP()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tgenDoc, err := loadGenesisDoc(cdc, config.GenesisFile())\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\t// Read --pubkey, if empty take it from priv_validator.json\n\t\t\tif valPubKeyString := viper.GetString(cli.FlagPubKey); valPubKeyString != \"\" {\n\t\t\t\tvalPubKey, err = sdk.GetConsPubKeyBech32(valPubKeyString)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t}\n\t\t\t// Run gaiad tx create-validator\n\t\t\tprepareFlagsForTxCreateValidator(config, nodeID, ip, genDoc.ChainID, valPubKey)\n\t\t\tcreateValidatorCmd := cli.GetCmdCreateValidator(cdc)\n\n\t\t\tw, err := ioutil.TempFile(\"\", \"gentx\")\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tunsignedGenTxFilename := w.Name()\n\t\t\tdefer os.Remove(unsignedGenTxFilename)\n\t\t\tos.Stdout = w\n\t\t\tif err = createValidatorCmd.RunE(nil, args); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tw.Close()\n\n\t\t\tprepareFlagsForTxSign()\n\t\t\tsignCmd := authcmd.GetSignCommand(cdc, authcmd.GetAccountDecoder(cdc))\n\t\t\tif w, err = prepareOutputFile(config.RootDir, nodeID); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tos.Stdout = w\n\t\t\treturn signCmd.RunE(nil, []string{unsignedGenTxFilename})\n\t\t},\n\t}\n\n\tcmd.Flags().String(tmcli.HomeFlag, app.DefaultNodeHome, \"node's home directory\")\n\tcmd.Flags().String(flagClientHome, app.DefaultCLIHome, \"client's home directory\")\n\tcmd.Flags().String(client.FlagName, \"\", \"name of private key with which to sign the gentx\")\n\tcmd.Flags().AddFlagSet(cli.FsCommissionCreate)\n\tcmd.Flags().AddFlagSet(cli.FsAmount)\n\tcmd.Flags().AddFlagSet(cli.FsPk)\n\tcmd.MarkFlagRequired(client.FlagName)\n\treturn cmd\n}", "func (c *restClient) CreateTask(ctx context.Context, req *cloudtaskspb.CreateTaskRequest, opts ...gax.CallOption) (*cloudtaskspb.Task, error) {\n\tm := protojson.MarshalOptions{AllowPartial: true, UseEnumNumbers: true}\n\tjsonReq, err := m.Marshal(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v2beta3/%v/tasks\", req.GetParent())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"parent\", url.QueryEscape(req.GetParent()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\topts = append((*c.CallOptions).CreateTask[0:len((*c.CallOptions).CreateTask):len((*c.CallOptions).CreateTask)], opts...)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tresp := &cloudtaskspb.Task{}\n\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"POST\", baseUrl.String(), bytes.NewReader(jsonReq))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}, opts...)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\treturn resp, nil\n}", "func (client Client) CreateSender(req *http.Request) (*http.Response, error) {\n\treturn client.Send(req, autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n}", "func (c *RestClient) DeleteTenant(tenantCluster string) (*models.TcaTask, error) {\n\n\tglog.Infof(\"Deleting tenant cluster %v\", tenantCluster)\n\n\tc.GetClient()\n\tresp, err := c.Client.R().Delete(c.BaseURL + fmt.Sprintf(TcaDeleteTenant, tenantCluster))\n\tif err != nil {\n\t\tglog.Error(err)\n\t\treturn nil, err\n\t}\n\n\tif c.isTrace && resp != nil {\n\t\tfmt.Println(string(resp.Body()))\n\t}\n\n\tif !resp.IsSuccess() {\n\t\treturn nil, c.checkErrors(resp)\n\t}\n\n\tvar task models.TcaTask\n\tif err := json.Unmarshal(resp.Body(), &task); err != nil {\n\t\tglog.Error(\"Failed parse server respond.\")\n\t\treturn nil, err\n\t}\n\n\treturn &task, nil\n}", "func (ctc *ClinicalTrialCreate) Exec(ctx context.Context) error {\n\t_, err := ctc.Save(ctx)\n\treturn err\n}", "func (client *Client) CreateTagTask(request *CreateTagTaskRequest) (response *CreateTagTaskResponse, err error) {\n\tresponse = CreateCreateTagTaskResponse()\n\terr = client.DoAction(request, response)\n\treturn\n}", "func createVdcAsyncV97(adminOrg *AdminOrg, vdcConfiguration *types.VdcConfiguration) (Task, error) {\n\tutil.Logger.Printf(\"[TRACE] createVdcAsyncV97 called %#v\", *vdcConfiguration)\n\terr := validateVdcConfigurationV97(*vdcConfiguration)\n\tif err != nil {\n\t\treturn Task{}, err\n\t}\n\n\tvdcConfiguration.Xmlns = types.XMLNamespaceVCloud\n\n\tvdcCreateHREF, err := url.ParseRequestURI(adminOrg.AdminOrg.HREF)\n\tif err != nil {\n\t\treturn Task{}, fmt.Errorf(\"error parsing admin org url: %s\", err)\n\t}\n\tvdcCreateHREF.Path += \"/vdcsparams\"\n\n\tadminVdc := NewAdminVdc(adminOrg.client)\n\n\t_, err = adminOrg.client.ExecuteRequest(vdcCreateHREF.String(), http.MethodPost,\n\t\t\"application/vnd.vmware.admin.createVdcParams+xml\", \"error creating VDC: %s\",\n\t\tvdcConfiguration, adminVdc.AdminVdc)\n\tif err != nil {\n\t\treturn Task{}, err\n\t}\n\n\t// Return the task\n\ttask := NewTask(adminOrg.client)\n\ttask.Task = adminVdc.AdminVdc.Tasks.Task[0]\n\treturn *task, nil\n}", "func (client CloudEndpointsClient) CreateSender(req *http.Request) (future CloudEndpointsCreateFuture, err error) {\n\tvar resp *http.Response\n\tfuture.FutureAPI = &azure.Future{}\n\tresp, err = client.Send(req, azure.DoRetryWithRegistration(client.Client))\n\tif err != nil {\n\t\treturn\n\t}\n\tvar azf azure.Future\n\tazf, err = azure.NewFutureFromResponse(resp)\n\tfuture.FutureAPI = &azf\n\tfuture.Result = future.result\n\treturn\n}", "func (*CreateTenantRequest) Descriptor() ([]byte, []int) {\n\treturn file_global_proto_rawDescGZIP(), []int{3}\n}", "func (a *DefaultApiService) ShowTenantExecute(r ApiShowTenantRequest) (Tenant, *_nethttp.Response, GenericOpenAPIError) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodGet\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\texecutionError GenericOpenAPIError\n\t\tlocalVarReturnValue Tenant\n\t)\n\n\tlocalBasePath, err := a.client.cfg.ServerURLWithContext(r.ctx, \"DefaultApiService.ShowTenant\")\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarReturnValue, nil, executionError\n\t}\n\n\tlocalVarPath := localBasePath + \"/tenants/{id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"id\"+\"}\", _neturl.PathEscape(parameterToString(r.id, \"\")), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\treq, err := a.client.prepareRequest(r.ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarReturnValue, nil, executionError\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(req)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarReturnValue, localVarHTTPResponse, executionError\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tlocalVarHTTPResponse.Body = _ioutil.NopCloser(bytes.NewBuffer(localVarBody))\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarReturnValue, localVarHTTPResponse, executionError\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v ErrorNotFound\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, executionError\n}", "func NewTaskCreateRequest(flux string) *TaskCreateRequest {\n\tthis := TaskCreateRequest{}\n\tthis.Flux = flux\n\treturn &this\n}", "func (client ListManagementTermListsClient) CreateSender(req *http.Request) (*http.Response, error) {\n\treturn autorest.SendWithSender(client, req,\n\t\tautorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n}", "func (client AccountClient) CreateOrUpdateSender(req *http.Request) (future AccountCreateOrUpdateFuture, err error) {\n var resp *http.Response\n resp, err = client.Send(req, azure.DoRetryWithRegistration(client.Client))\n if err != nil {\n return\n }\n future.Future, err = azure.NewFutureFromResponse(resp)\n return\n }", "func (o NetworkOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Network) pulumi.StringOutput { return v.TenantId }).(pulumi.StringOutput)\n}", "func fnTenant(ctx Context, doc *JDoc, params []string) interface{} {\n\tstats := ctx.Value(EelTotalStats).(*ServiceStats)\n\tif params == nil || params[0] != \"\" {\n\t\tctx.Log().Error(\"error_type\", \"func_tenant\", \"op\", \"tenant\", \"cause\", \"no_parameters_expected\", \"params\", params)\n\t\tstats.IncErrors()\n\t\tAddError(ctx, SyntaxError{fmt.Sprintf(\"no parameters expected in call to tenant function\"), \"tenant\", params})\n\t\treturn \"\"\n\t}\n\th := GetCurrentHandlerConfig(ctx)\n\tif h == nil {\n\t\tctx.Log().Error(\"error_type\", \"func_tenant\", \"op\", \"tenant\", \"cause\", \"no_handler\", \"params\", params)\n\t\tstats.IncErrors()\n\t\tAddError(ctx, RuntimeError{fmt.Sprintf(\"current handler not found in call to tenant function\"), \"tenant\", params})\n\t\treturn \"\"\n\t}\n\t//Add logic to separating tenant and partner\n\tallowPartner := GetConfig(ctx).AllowPartner\n\tif h.TenantId == \"_default\" {\n\t\t//The default tenant handler is used. We need to drill-down into context to find out the actual tenant\n\t\ttenantHeaderKey := GetConfig(ctx).HttpTenantHeader\n\t\tif ctx.Value(tenantHeaderKey) != nil {\n\t\t\tcombinedTenant := ctx.Value(tenantHeaderKey).(string)\n\t\t\ttenant := \"\"\n\t\t\tif allowPartner && strings.LastIndex(combinedTenant, \"_\") > 0 && strings.LastIndex(combinedTenant, \"_\") < len(combinedTenant)-1 {\n\t\t\t\ttenant = combinedTenant[:strings.LastIndex(combinedTenant, \"_\")]\n\t\t\t} else {\n\t\t\t\ttenant = combinedTenant\n\t\t\t}\n\t\t\treturn tenant\n\t\t}\n\t}\n\tcombinedTenant := h.TenantId\n\ttenant := \"\"\n\tif allowPartner && strings.LastIndex(combinedTenant, \"_\") > 0 && strings.LastIndex(combinedTenant, \"_\") < len(combinedTenant)-1 {\n\t\ttenant = combinedTenant[:strings.LastIndex(combinedTenant, \"_\")]\n\t} else {\n\t\ttenant = combinedTenant\n\t}\n\treturn tenant\n}", "func (client BaseClient) CreateFeatureInstanceSender(req *http.Request) (*http.Response, error) {\n\treturn autorest.SendWithSender(client, req,\n\t\tautorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n}", "func CreateHydrationTask(userID string) (*datastore.Key, error) {\n\tctx := context.Background()\n\tclient, err := datastore.NewClient(ctx, datastore.DetectProjectID)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer client.Close()\n\ttask := &HydrationTask{\n\t\tSlackUserID: userID,\n\t}\n\tkey := datastore.IncompleteKey(kindHydrationTask, nil)\n\treturn client.Put(ctx, key, task)\n}", "func (ctrl *TaskController) CreateTask(w http.ResponseWriter, r *http.Request) {\n\ttask := &model.Task{}\n\tlogrus.Println(r.Body)\n\terr := GetJSONContent(task, r)\n\tif err != nil {\n\t\tlogrus.Error(err)\n\t\tSendJSONError(w, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tlogrus.Println(\"create task\")\n\n\ttask.CreationDate = time.Now()\n\ttask.ModificationDate = time.Now()\n\ttask.Status = 0\n\n\ttask, err = ctrl.taskDao.Upsert(task)\n\tif err != nil {\n\t\tlogrus.Error(err)\n\t\tSendJSONError(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tlogrus.Println(\"task : \", task)\n\tSendJSONWithHTTPCode(w, task, http.StatusCreated)\n}", "func (cmd *CreateCompanyHyCompanyCommand) Run(c *client.Client, args []string) error {\n\tvar path string\n\tif len(args) > 0 {\n\t\tpath = args[0]\n\t} else {\n\t\tpath = \"/api/company\"\n\t}\n\tvar payload client.CreateCompanyHyCompanyPayload\n\tif cmd.Payload != \"\" {\n\t\terr := json.Unmarshal([]byte(cmd.Payload), &payload)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"failed to deserialize payload: %s\", err)\n\t\t}\n\t}\n\tlogger := goa.NewLogger(log.New(os.Stderr, \"\", log.LstdFlags))\n\tctx := goa.WithLogger(context.Background(), logger)\n\tresp, err := c.CreateCompanyHyCompany(ctx, path, &payload, cmd.ContentType)\n\tif err != nil {\n\t\tgoa.LogError(ctx, \"failed\", \"err\", err)\n\t\treturn err\n\t}\n\n\tgoaclient.HandleResponse(c.Client, resp, cmd.PrettyPrint)\n\treturn nil\n}", "func ExampleIscsiTargetsClient_BeginCreateOrUpdate() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclientFactory, err := armstoragepool.NewClientFactory(\"<subscription-id>\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tpoller, err := clientFactory.NewIscsiTargetsClient().BeginCreateOrUpdate(ctx, \"myResourceGroup\", \"myDiskPool\", \"myIscsiTarget\", armstoragepool.IscsiTargetCreate{\n\t\tProperties: &armstoragepool.IscsiTargetCreateProperties{\n\t\t\tACLMode: to.Ptr(armstoragepool.IscsiTargetACLModeDynamic),\n\t\t\tLuns: []*armstoragepool.IscsiLun{\n\t\t\t\t{\n\t\t\t\t\tName: to.Ptr(\"lun0\"),\n\t\t\t\t\tManagedDiskAzureResourceID: to.Ptr(\"/subscriptions/11111111-1111-1111-1111-111111111111/resourceGroups/myResourceGroup/providers/Microsoft.Compute/disks/vm-name_DataDisk_1\"),\n\t\t\t\t}},\n\t\t\tTargetIqn: to.Ptr(\"iqn.2005-03.org.iscsi:server1\"),\n\t\t},\n\t}, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\tres, err := poller.PollUntilDone(ctx, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to pull the result: %v\", err)\n\t}\n\t// You could use response here. We use blank identifier for just demo purposes.\n\t_ = res\n\t// If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes.\n\t// res.IscsiTarget = armstoragepool.IscsiTarget{\n\t// \tName: to.Ptr(\"myIscsiTarget\"),\n\t// \tType: to.Ptr(\"Microsoft.StoragePool/diskPools/iscsiTargets\"),\n\t// \tID: to.Ptr(\"/subscriptions/00000000-0000-0000-0000-000000000000/resourceGroups/myResourceGroup/providers/Microsoft.StoragePool/diskPools/myDiskPool/iscsiTargets/myIscsiTarget\"),\n\t// \tManagedBy: to.Ptr(\"/subscriptions/11111111-1111-1111-1111-111111111111/resourceGroups/myResourceGroup/providers/Microsoft.AVS/privateClouds/myPrivateCloud/clusters/Cluster-1/datastores/datastore1\"),\n\t// \tManagedByExtended: []*string{\n\t// \t\tto.Ptr(\"/subscriptions/11111111-1111-1111-1111-111111111111/resourceGroups/myResourceGroup/providers/Microsoft.AVS/privateClouds/myPrivateCloud/clusters/Cluster-1/datastores/datastore1\"),\n\t// \t\tto.Ptr(\"/subscriptions/11111111-1111-1111-1111-111111111111/resourceGroups/myResourceGroup/providers/Microsoft.AVS/privateClouds/myPrivateCloud/clusters/Cluster-1/datastores/datastore2\")},\n\t// \t\tProperties: &armstoragepool.IscsiTargetProperties{\n\t// \t\t\tACLMode: to.Ptr(armstoragepool.IscsiTargetACLModeDynamic),\n\t// \t\t\tEndpoints: []*string{\n\t// \t\t\t\tto.Ptr(\"10.0.0.1:3260\")},\n\t// \t\t\t\tLuns: []*armstoragepool.IscsiLun{\n\t// \t\t\t\t\t{\n\t// \t\t\t\t\t\tName: to.Ptr(\"lun0\"),\n\t// \t\t\t\t\t\tLun: to.Ptr[int32](3),\n\t// \t\t\t\t\t\tManagedDiskAzureResourceID: to.Ptr(\"/subscriptions/11111111-1111-1111-1111-111111111111/resourceGroups/myResourceGroup/providers/Microsoft.Compute/disks/vm-name_DataDisk_1\"),\n\t// \t\t\t\t}},\n\t// \t\t\t\tPort: to.Ptr[int32](3260),\n\t// \t\t\t\tProvisioningState: to.Ptr(armstoragepool.ProvisioningStatesSucceeded),\n\t// \t\t\t\tSessions: []*string{\n\t// \t\t\t\t\tto.Ptr(\"iqn.2005-03.org.iscsi:client\")},\n\t// \t\t\t\t\tStatus: to.Ptr(armstoragepool.OperationalStatusUnknown),\n\t// \t\t\t\t\tTargetIqn: to.Ptr(\"iqn.2005-03.org.iscsi:server1\"),\n\t// \t\t\t\t},\n\t// \t\t\t\tSystemData: &armstoragepool.SystemMetadata{\n\t// \t\t\t\t\tCreatedAt: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2020-06-24T06:53:57+00:00\"); return t}()),\n\t// \t\t\t\t\tCreatedBy: to.Ptr(\"alias\"),\n\t// \t\t\t\t\tCreatedByType: to.Ptr(armstoragepool.CreatedByTypeUser),\n\t// \t\t\t\t\tLastModifiedAt: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2020-06-24T06:53:57+00:00\"); return t}()),\n\t// \t\t\t\t\tLastModifiedBy: to.Ptr(\"alias\"),\n\t// \t\t\t\t\tLastModifiedByType: to.Ptr(armstoragepool.CreatedByTypeUser),\n\t// \t\t\t\t},\n\t// \t\t\t}\n}", "func (e CreateTenantV1RequestValidationError) Cause() error { return e.cause }", "func (client ApplicationsClient) CreateSender(req *http.Request) (*http.Response, error) {\n\treturn client.Send(req, autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n}", "func (client DatasetClient) Create(ctx context.Context, conversionID string, datasetID string, descriptionDataset string) (result DatasetCreateFuture, err error) {\n if tracing.IsEnabled() {\n ctx = tracing.StartSpan(ctx, fqdn + \"/DatasetClient.Create\")\n defer func() {\n sc := -1\n if result.FutureAPI != nil && result.FutureAPI.Response() != nil {\n sc = result.FutureAPI.Response().StatusCode\n }\n tracing.EndSpan(ctx, sc, err)\n }()\n }\n req, err := client.CreatePreparer(ctx, conversionID, datasetID, descriptionDataset)\n if err != nil {\n err = autorest.NewErrorWithError(err, \"creator.DatasetClient\", \"Create\", nil , \"Failure preparing request\")\n return\n }\n\n result, err = client.CreateSender(req)\n if err != nil {\n err = autorest.NewErrorWithError(err, \"creator.DatasetClient\", \"Create\", nil , \"Failure sending request\")\n return\n }\n\n return\n}", "func HandleCreateTOTPAccount(w http.ResponseWriter, r *http.Request) {\n\n\tvar body CreateAccountRequest\n\terr := api.UnmarshalJSONFromRequest(r, &body)\n\tif err != nil {\n\t\tapi.WriteError(w, http.StatusBadRequest, err, false, nil)\n\t\treturn\n\t}\n\n\tvar req = totp.CreateAccountRequest{\n\t\tName: body.Name,\n\t\tPrivateKey: []byte(body.PrivateKey),\n\t}\n\n\t// Attempt login and get the token\n\ta, err := totp.CreateAccount(req)\n\tif err != nil {\n\t\tapi.WriteError(w, http.StatusBadRequest, err, false, nil)\n\t\treturn\n\t}\n\n\tapi.WriteResponse(w, http.StatusCreated, a.ID)\n\n}" ]
[ "0.6968292", "0.650007", "0.6387776", "0.6244856", "0.6144715", "0.61194825", "0.59259975", "0.58894646", "0.5825773", "0.5821312", "0.580582", "0.579278", "0.57073945", "0.5685259", "0.56761265", "0.56399065", "0.55743116", "0.55708957", "0.5554511", "0.5541497", "0.5490196", "0.5448725", "0.5404259", "0.5396189", "0.53691995", "0.53169996", "0.5305854", "0.5278828", "0.5259096", "0.52428025", "0.5240993", "0.5192392", "0.51920706", "0.5181863", "0.5179813", "0.51622534", "0.5160744", "0.51602304", "0.5158016", "0.515456", "0.5137254", "0.5130304", "0.5119904", "0.51124114", "0.51052654", "0.509331", "0.5087275", "0.5061353", "0.5048126", "0.5036209", "0.50274676", "0.49733838", "0.49712497", "0.49704474", "0.49540806", "0.49476272", "0.49433333", "0.49360195", "0.4923951", "0.49233606", "0.4908926", "0.48995075", "0.48926422", "0.4886507", "0.48810494", "0.48753697", "0.48667702", "0.48621553", "0.48620299", "0.48559347", "0.4854782", "0.4854782", "0.4854782", "0.484878", "0.48448536", "0.48189163", "0.47895318", "0.4777673", "0.47731435", "0.47646463", "0.47639686", "0.47635573", "0.47515592", "0.47498173", "0.47495747", "0.47491902", "0.47441262", "0.47434193", "0.47394913", "0.47364882", "0.47347578", "0.4728521", "0.47269213", "0.47239736", "0.4721817", "0.47135198", "0.47119477", "0.47076616", "0.4705708", "0.47034132" ]
0.7233564
0
Retrieves a list of tenants, returns an error if one occurred
func listTenants(c *cli.Context, w io.Writer) error { err := checkArgCount(c, 0) if err != nil { return err } client.Photonclient, err = client.GetClient(c) if err != nil { return err } tenants, err := client.Photonclient.Tenants.GetAll() if err != nil { return err } if c.GlobalIsSet("non-interactive") { for _, tenant := range tenants.Items { fmt.Printf("%s\t%s\n", tenant.ID, tenant.Name) } } else if utils.NeedsFormatting(c) { utils.FormatObjects(tenants.Items, w, c) } else { w := new(tabwriter.Writer) w.Init(os.Stdout, 4, 4, 2, ' ', 0) fmt.Fprintf(w, "ID\tName\n") for _, tenant := range tenants.Items { fmt.Fprintf(w, "%s\t%s\n", tenant.ID, tenant.Name) } err = w.Flush() if err != nil { return err } fmt.Printf("\nTotal: %d\n", len(tenants.Items)) } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *Client) GetTenants(ctx context.Context, limit string, cursor string) (*ListTenantsResponse, error) {\n\turl := fmt.Sprintf(c.API.BaseURL + \"/tenants\")\n\n\treq, err := http.NewRequestWithContext(ctx, \"GET\", url, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tq := req.URL.Query()\n\tif limit != \"\" {\n\t\tq.Add(\"limit\", limit)\n\t}\n\tif cursor != \"\" {\n\t\tq.Add(\"cursor\", cursor)\n\t}\n\treq.URL.RawQuery = q.Encode()\n\n\tbytes, err := c.API.ExecuteRequest(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar data ListTenantsResponse\n\terr = json.Unmarshal(bytes, &data)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &data, nil\n}", "func GetTenants(db *database.DB) (Tenants, error) {\n\n\tvar tenantList Tenants\n\n\tresult := db.Client.Find(&tenantList)\n\tif result.Error != nil {\n\t\treturn nil, result.Error\n\t}\n\n\treturn tenantList, nil\n}", "func (client *Client) ListTenants() (types.TenantsListResponse, error) {\n\tvar tenants types.TenantsListResponse\n\n\turl, err := client.getCiaoTenantsResource()\n\tif err != nil {\n\t\treturn tenants, err\n\t}\n\n\terr = client.getResource(url, api.TenantsV1, nil, &tenants)\n\n\treturn tenants, err\n}", "func (a *Client) ListTenants(params *ListTenantsParams) (*ListTenantsOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewListTenantsParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"listTenants\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/tenants\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\"},\n\t\tParams: params,\n\t\tReader: &ListTenantsReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*ListTenantsOK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\t// safeguard: normally, absent a default response, unknown success responses return an error above: so this is a codegen issue\n\tmsg := fmt.Sprintf(\"unexpected success response for listTenants: API contract not enforced by server. Client expected to get an error, but got: %T\", result)\n\tpanic(msg)\n}", "func (c *RestClient) GetTenantsQuery(f *specs.TenantsNfFilter) (*response.Tenants, error) {\n\n\tc.GetClient()\n\tresp, err := c.Client.R().SetBody(f).SetQueryString(apiTenantAction).Post(c.BaseURL + apiTenants)\n\tif err != nil {\n\t\tglog.Error(err)\n\t\treturn nil, err\n\t}\n\n\tif c.isTrace && resp != nil {\n\t\tfmt.Println(string(resp.Body()))\n\t}\n\n\tif !resp.IsSuccess() {\n\t\treturn nil, c.checkError(resp)\n\t}\n\n\tvar tenants response.Tenants\n\tif err := json.Unmarshal(resp.Body(), &tenants); err != nil {\n\t\tglog.Error(\"Failed parse server respond.\")\n\t\treturn nil, err\n\t}\n\n\treturn &tenants, nil\n}", "func TestGetTenants(t *testing.T) {\n\tctx, cancelFunc := context.WithTimeout(context.Background(), standardTimeout)\n\t_, err := bat.GetAllTenants(ctx)\n\tcancelFunc()\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to retrieve tenant list : %v\", err)\n\t}\n}", "func TestGetTenants(t *testing.T) {\n\tctx, cancelFunc := context.WithTimeout(context.Background(), standardTimeout)\n\t_, err := bat.GetAllTenants(ctx)\n\tcancelFunc()\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to retrieve tenant list : %v\", err)\n\t}\n}", "func (a *DefaultApiService) ListTenants(ctx _context.Context) ApiListTenantsRequest {\n\treturn ApiListTenantsRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t}\n}", "func GetTenantList() ([]tenants.Tenant, error) {\n\tc, err := getIdentityClient()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\topts := tenants.ListOpts{}\n\tpager := tenants.List(c, &opts)\n\tpage, err := pager.AllPages()\n\tif err == nil {\n\t\treturn tenants.ExtractTenants(page)\n\t}\n\treturn nil, err\n}", "func (api *tenantAPI) List(ctx context.Context, opts *api.ListWatchOptions) ([]*Tenant, error) {\n\tvar objlist []*Tenant\n\tobjs, err := api.ct.List(\"Tenant\", ctx, opts)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfor _, obj := range objs {\n\t\tswitch tp := obj.(type) {\n\t\tcase *Tenant:\n\t\t\teobj := obj.(*Tenant)\n\t\t\tobjlist = append(objlist, eobj)\n\t\tdefault:\n\t\t\tlog.Fatalf(\"Got invalid object type %v while looking for Tenant\", tp)\n\t\t}\n\t}\n\n\treturn objlist, nil\n}", "func (c *RestClient) GetVimTenants(ctx context.Context) (*response.Tenants, error) {\n\n\tglog.Infof(\"Retrieving vim tenants\")\n\n\tc.GetClient()\n\tresp, err := c.Client.R().SetContext(ctx).Get(c.BaseURL + apiTenants)\n\tif err != nil {\n\t\tglog.Error(err)\n\t\treturn nil, err\n\t}\n\n\tif c.isTrace && resp != nil {\n\t\tfmt.Println(string(resp.Body()))\n\t}\n\n\tif !resp.IsSuccess() {\n\t\treturn nil, c.checkError(resp)\n\t}\n\n\tvar tenants response.Tenants\n\tif err := json.Unmarshal(resp.Body(), &tenants); err != nil {\n\t\tglog.Error(\"Failed parse server respond. %v\", err)\n\t\treturn nil, err\n\t}\n\n\treturn &tenants, nil\n}", "func (db *MySQLDB) ListTenant(ctx context.Context, request *helper.PageRequest) ([]*Tenant, *helper.Page, error) {\n\tfLog := mysqlLog.WithField(\"func\", \"GetUserByRecID\").WithField(\"RequestID\", ctx.Value(constants.RequestID))\n\tq := \"SELECT COUNT(*) AS CNT FROM HANSIP_TENANT\"\n\tret := make([]*Tenant, 0)\n\trow := db.instance.QueryRowContext(ctx, q)\n\tcount := 0\n\terr := row.Scan(&count)\n\tif err != nil {\n\t\tif err == sql.ErrNoRows {\n\t\t\treturn ret, helper.NewPage(request, uint(count)), nil\n\t\t}\n\t\tfLog.Errorf(\"db.instance.QueryRowContext got %s\", err.Error())\n\t\treturn nil, nil, &ErrDBQueryError{\n\t\t\tWrapped: err,\n\t\t\tMessage: \"Error ListTenant\",\n\t\t\tSQL: q,\n\t\t}\n\t}\n\n\tvar OrderBy string\n\tswitch strings.ToUpper(request.OrderBy) {\n\tcase \"TENANT_NAME\":\n\t\tOrderBy = \"TENANT_NAME\"\n\tcase \"TENANT_DOMAIN\":\n\t\tOrderBy = \"TENANT_DOMAIN\"\n\tdefault:\n\t\tOrderBy = \"TENANT_NAME\"\n\t}\n\n\tpage := helper.NewPage(request, uint(count))\n\tq = fmt.Sprintf(\"SELECT REC_ID, TENANT_NAME, TENANT_DOMAIN, DESCRIPTION FROM HANSIP_TENANT ORDER BY %s %s LIMIT %d, %d\", OrderBy, request.Sort, page.OffsetStart, page.OffsetEnd-page.OffsetStart)\n\trows, err := db.instance.QueryContext(ctx, q)\n\tif err != nil {\n\t\tfLog.Errorf(\"db.instance.QueryContext got %s. SQL = %s\", err.Error(), q)\n\t\treturn nil, nil, &ErrDBQueryError{\n\t\t\tWrapped: err,\n\t\t\tMessage: \"Error ListTenant\",\n\t\t\tSQL: q,\n\t\t}\n\t}\n\tdefer rows.Close()\n\tfor rows.Next() {\n\t\tt := &Tenant{}\n\t\terr := rows.Scan(&t.RecID, &t.Name, &t.Domain, &t.Description)\n\t\tif err != nil {\n\t\t\tfLog.Warnf(\"rows.Scan got %s\", err.Error())\n\t\t\treturn nil, nil, &ErrDBScanError{\n\t\t\t\tWrapped: err,\n\t\t\t\tMessage: \"Error ListTenant\",\n\t\t\t\tSQL: q,\n\t\t\t}\n\t\t} else {\n\t\t\tret = append(ret, t)\n\t\t}\n\t}\n\treturn ret, page, nil\n}", "func (api *tenantAPI) ApisrvList(ctx context.Context, opts *api.ListWatchOptions) ([]*cluster.Tenant, error) {\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn nil, err\n\t\t}\n\n\t\treturn apicl.ClusterV1().Tenant().List(context.Background(), opts)\n\t}\n\n\t// List from local cache\n\tctkitObjs, err := api.List(ctx, opts)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar ret []*cluster.Tenant\n\tfor _, obj := range ctkitObjs {\n\t\tret = append(ret, &obj.Tenant)\n\t}\n\treturn ret, nil\n}", "func (a *Client) ListOpenstackTenants(params *ListOpenstackTenantsParams, authInfo runtime.ClientAuthInfoWriter, opts ...ClientOption) (*ListOpenstackTenantsOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewListOpenstackTenantsParams()\n\t}\n\top := &runtime.ClientOperation{\n\t\tID: \"listOpenstackTenants\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/api/v1/providers/openstack/tenants\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &ListOpenstackTenantsReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t}\n\tfor _, opt := range opts {\n\t\topt(op)\n\t}\n\n\tresult, err := a.transport.Submit(op)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*ListOpenstackTenantsOK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\tunexpectedSuccess := result.(*ListOpenstackTenantsDefault)\n\treturn nil, runtime.NewAPIError(\"unexpected success response: content available as default response in error\", unexpectedSuccess, unexpectedSuccess.Code())\n}", "func ExtractTenants(r pagination.Page) ([]Tenant, error) {\n\tvar s struct {\n\t\tContractID string `json:\"contract_id\"`\n\t\tTenants []Tenant `json:\"tenants\"`\n\t}\n\n\t// In list response case, each json element does not have contract_id.\n\t// It is set at out layer of each element.\n\t// So following logic set contract_id into inside of tenants slice forcibly.\n\t// In \"show(get with ID of tennat)\" case, this does not occur.\n\terr := (r.(TenantPage)).ExtractInto(&s)\n\tcontractID := s.ContractID\n\n\tfor i := 0; i < len(s.Tenants); i++ {\n\t\ts.Tenants[i].ContractID = contractID\n\t}\n\treturn s.Tenants, err\n}", "func List(client *gophercloud.ServiceClient, opts *ListOpts) pagination.Pager {\r\n\turl := listURL(client)\r\n\tif opts != nil {\r\n\t\tq, err := gophercloud.BuildQueryString(opts)\r\n\t\tif err != nil {\r\n\t\t\treturn pagination.Pager{Err: err}\r\n\t\t}\r\n\t\turl += q.String()\r\n\t}\r\n\treturn pagination.NewPager(client, url, func(r pagination.PageResult) pagination.Page {\r\n\t\treturn TenantPage{pagination.LinkedPageBase{PageResult: r}}\r\n\t})\r\n}", "func (suite *TenantTestSuite) TestListTenants() {\n\n\trequest, _ := http.NewRequest(\"GET\", \"/api/v2/admin/tenants\", strings.NewReader(\"\"))\n\trequest.Header.Set(\"x-api-key\", suite.clientkey)\n\trequest.Header.Set(\"Accept\", \"application/json\")\n\tresponse := httptest.NewRecorder()\n\n\tsuite.router.ServeHTTP(response, request)\n\n\tcode := response.Code\n\toutput := response.Body.String()\n\n\tprofileJSON := `{\n \"status\": {\n \"message\": \"Success\",\n \"code\": \"200\"\n },\n \"data\": [\n {\n \"id\": \"6ac7d684-1f8e-4a02-a502-720e8f11e50b\",\n \"info\": {\n \"name\": \"AVENGERS\",\n \"email\": \"email@something\",\n \"website\": \"www.avengers.com\",\n \"created\": \"2015-10-20 02:08:04\",\n \"updated\": \"2015-10-20 02:08:04\"\n },\n \"db_conf\": [\n {\n \"store\": \"ar\",\n \"server\": \"a.mongodb.org\",\n \"port\": 27017,\n \"database\": \"ar_db\",\n \"username\": \"admin\",\n \"password\": \"3NCRYPT3D\"\n },\n {\n \"store\": \"status\",\n \"server\": \"b.mongodb.org\",\n \"port\": 27017,\n \"database\": \"status_db\",\n \"username\": \"admin\",\n \"password\": \"3NCRYPT3D\"\n }\n ],\n \"users\": [\n {\n \"name\": \"cap\",\n \"email\": \"[email protected]\",\n \"api_key\": \"C4PK3Y\"\n },\n {\n \"name\": \"thor\",\n \"email\": \"[email protected]\",\n \"api_key\": \"TH0RK3Y\"\n }\n ]\n },\n {\n \"id\": \"6ac7d684-1f8e-4a02-a502-720e8f11e50c\",\n \"info\": {\n \"name\": \"GUARDIANS\",\n \"email\": \"email@something2\",\n \"website\": \"www.gotg.com\",\n \"created\": \"2015-10-20 02:08:04\",\n \"updated\": \"2015-10-20 02:08:04\"\n },\n \"db_conf\": [\n {\n \"store\": \"ar\",\n \"server\": \"a.mongodb.org\",\n \"port\": 27017,\n \"database\": \"ar_db\",\n \"username\": \"admin\",\n \"password\": \"3NCRYPT3D\"\n },\n {\n \"store\": \"status\",\n \"server\": \"b.mongodb.org\",\n \"port\": 27017,\n \"database\": \"status_db\",\n \"username\": \"admin\",\n \"password\": \"3NCRYPT3D\"\n }\n ],\n \"users\": [\n {\n \"name\": \"groot\",\n \"email\": \"[email protected]\",\n \"api_key\": \"GR00TK3Y\"\n },\n {\n \"name\": \"starlord\",\n \"email\": \"[email protected]\",\n \"api_key\": \"ST4RL0RDK3Y\"\n }\n ]\n }\n ]\n}`\n\t// Check that we must have a 200 ok code\n\tsuite.Equal(200, code, \"Internal Server Error\")\n\t// Compare the expected and actual json response\n\tsuite.Equal(profileJSON, output, \"Response body mismatch\")\n\n}", "func (r *queryResolver) Tenants(ctx context.Context, first *int, after *graphql.PageCursor, searchTerm *string) (*graphql.TenantPage, error) {\n\treturn r.tenant.Tenants(ctx, first, after, searchTerm)\n}", "func (a *appRuntimeStore) GetTenantResourceList() []TenantResource {\n\treturn a.resourceCache.GetAllTenantResource()\n}", "func (a *Client) GrantsList(params *GrantsListParams, authInfo runtime.ClientAuthInfoWriter) (*GrantsListOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGrantsListParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"grants_list\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/v3/customer/core/grants/\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &GrantsListReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*GrantsListOK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\t// safeguard: normally, absent a default response, unknown success responses return an error above: so this is a codegen issue\n\tmsg := fmt.Sprintf(\"unexpected success response for grants_list: API contract not enforced by server. Client expected to get an error, but got: %T\", result)\n\tpanic(msg)\n}", "func (s *MultiElasticsearch) AllTenants() []string {\n\ttenants := make([]string, len(s.stores))\n\ti := 0\n\tfor tenant := range s.stores {\n\t\ttenants[i] = tenant\n\t\ti++\n\t}\n\tsort.Strings(tenants)\n\treturn tenants\n}", "func Tenants(mods ...qm.QueryMod) tenantQuery {\n\tmods = append(mods, qm.From(\"`tenants`\"))\n\tq := NewQuery(mods...)\n\tif len(queries.GetSelect(q)) == 0 {\n\t\tqueries.SetSelect(q, []string{\"`tenants`.*\"})\n\t}\n\n\treturn tenantQuery{q}\n}", "func FetchAllTenant() ([]string, error) {\n\n\tvar accounts []model.CloudAccount\n\tvar tenant []string\n\terr := db.From(\"user\").All(&accounts)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfor i, _ := range accounts {\n\t\ttenant = append(tenant, accounts[i].Tenant)\n\t}\n\treturn tenant, nil\n}", "func tenantAllHandler(formatter *render.Render) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, req *http.Request) {\n\t\tsetupResponse(&w, req)\n\t\n\t\tif (*req).Method == \"OPTIONS\" {\n\t\t\tfmt.Println(\"PREFLIGHT Request\")\n\t\t\treturn\n\t\t}\n\n\t\tsession, err := mgo.Dial(mongodb_server)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\tdefer session.Close()\n\t\tsession.SetMode(mgo.Monotonic, true)\n\t\tvar result []Tenant\n\t\tc := session.DB(mongodb_database).C(mongodb_collection)\n\t\terr = c.Find(bson.M{}).All(&result)\n\t\tif err != nil {\n\t\t\tfmt.Println(\" Error: \", err)\n\t\t\tformatter.JSON(w, http.StatusBadRequest, \"Not Found\")\n\t\t}\n\t\tfmt.Println(\"All Tenants:\", result)\n\t\tformatter.JSON(w, http.StatusOK, result)\n\t\t\n\t}\n}", "func (a *DefaultApiService) ListTenantsExecute(r ApiListTenantsRequest) (TenantsCollection, *_nethttp.Response, GenericOpenAPIError) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodGet\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\texecutionError GenericOpenAPIError\n\t\tlocalVarReturnValue TenantsCollection\n\t)\n\n\tlocalBasePath, err := a.client.cfg.ServerURLWithContext(r.ctx, \"DefaultApiService.ListTenants\")\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarReturnValue, nil, executionError\n\t}\n\n\tlocalVarPath := localBasePath + \"/tenants\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tif r.limit != nil {\n\t\tlocalVarQueryParams.Add(\"limit\", parameterToString(*r.limit, \"\"))\n\t}\n\tif r.offset != nil {\n\t\tlocalVarQueryParams.Add(\"offset\", parameterToString(*r.offset, \"\"))\n\t}\n\tif r.filter != nil {\n\t\tlocalVarQueryParams.Add(\"filter\", parameterToString(*r.filter, \"\"))\n\t}\n\tif r.sortBy != nil {\n\t\tlocalVarQueryParams.Add(\"sort_by\", parameterToString(*r.sortBy, \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\treq, err := a.client.prepareRequest(r.ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarReturnValue, nil, executionError\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(req)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarReturnValue, localVarHTTPResponse, executionError\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tlocalVarHTTPResponse.Body = _ioutil.NopCloser(bytes.NewBuffer(localVarBody))\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarReturnValue, localVarHTTPResponse, executionError\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, executionError\n}", "func (m *MockTenantDao) GetALLTenants(query string) ([]*model.Tenants, error) {\n\tret := m.ctrl.Call(m, \"GetALLTenants\", query)\n\tret0, _ := ret[0].([]*model.Tenants)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func LoadDBTenants(config *config.Config, createConnector func(*config.Config) db.Connector) []string {\n\tconnector := createConnector(config)\n\tconnector.Init()\n\tdefer connector.Dispose()\n\tdbTenants := connector.GetTenants()\n\tlog.Printf(\"Read [%d] DB tenants ==> OK\", len(dbTenants))\n\treturn dbTenants\n}", "func (m *MockTenantDao) GetPagedTenants(offset, len int) ([]*model.Tenants, error) {\n\tret := m.ctrl.Call(m, \"GetPagedTenants\", offset, len)\n\tret0, _ := ret[0].([]*model.Tenants)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func NewTenants(s store.Store) *Tenants {\n\treturn &Tenants{store: s}\n}", "func (a *Client) GetAzureTenantIDs(params *GetAzureTenantIDsParams, opts ...ClientOption) (*GetAzureTenantIDsOK, *GetAzureTenantIDsMultiStatus, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetAzureTenantIDsParams()\n\t}\n\top := &runtime.ClientOperation{\n\t\tID: \"GetAzureTenantIDs\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/kubernetes-protection/entities/tenants/azure/v1\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\", \"application/octet-stream\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &GetAzureTenantIDsReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t}\n\tfor _, opt := range opts {\n\t\topt(op)\n\t}\n\n\tresult, err := a.transport.Submit(op)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tswitch value := result.(type) {\n\tcase *GetAzureTenantIDsOK:\n\t\treturn value, nil, nil\n\tcase *GetAzureTenantIDsMultiStatus:\n\t\treturn nil, value, nil\n\t}\n\t// safeguard: normally, absent a default response, unknown success responses return an error above: so this is a codegen issue\n\tmsg := fmt.Sprintf(\"unexpected success response for kubernetes_protection: API contract not enforced by server. Client expected to get an error, but got: %T\", result)\n\tpanic(msg)\n}", "func (r *RuntimeServer) GetTenantResources(context.Context, *pb.Empty) (*pb.TenantResourceList, error) {\n\tres := r.store.GetTenantResourceList()\n\tvar trs = make(map[string]*pb.TenantResource)\n\tfor _, re := range res {\n\t\tvar tr pb.TenantResource\n\t\trunningApps := r.store.GetTenantRunningApp(re.Namespace)\n\t\tfor _, app := range runningApps {\n\t\t\tif app.ServiceKind == model.ServiceKindThirdParty {\n\t\t\t\ttr.RunningAppThirdNum++\n\t\t\t} else if app.ServiceKind == model.ServiceKindInternal {\n\t\t\t\ttr.RunningAppInternalNum++\n\t\t\t}\n\t\t}\n\t\ttr.RunningAppNum = int64(len(runningApps))\n\t\ttr.CpuLimit = re.CPULimit\n\t\ttr.CpuRequest = re.CPURequest\n\t\ttr.MemoryLimit = re.MemoryLimit / 1024 / 1024\n\t\ttr.MemoryRequest = re.MemoryRequest / 1024 / 1024\n\t\ttrs[re.Namespace] = &tr\n\t}\n\treturn &pb.TenantResourceList{Resources: trs}, nil\n}", "func (m *MockEnterpriseDao) GetEnterpriseTenants(enterpriseID string) ([]*model.Tenants, error) {\n\tret := m.ctrl.Call(m, \"GetEnterpriseTenants\", enterpriseID)\n\tret0, _ := ret[0].([]*model.Tenants)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (uc MerchantController) GetAllMerchants(w http.ResponseWriter, r *http.Request, p httprouter.Params) {\n\n\tlog.Println(\"Getting all Merchants...\")\n\n\tresults := []models.Merchant{}\n\n\tif err := uc.session.DB(common.AppSettings.DBName).C(\"Merchant\").Find(nil).All(&results); err != nil {\n\t\tw.WriteHeader(404)\n\t\treturn\n\t}\n\n\tMerchantsJSON, _ := json.Marshal(results)\n\n\t// Write content-type, statuscode, payload\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(http.StatusOK)\n\tfmt.Fprintf(w, \"%s\", MerchantsJSON)\n\t// } else {\n\t// \t//If Token was reject\n\t// \tw.WriteHeader(http.StatusUnauthorized)\n\t// \tfmt.Fprint(w, msgError)\n\t// }\n\n\tlog.Println(\"retriving all Merchants...\")\n}", "func (a *Client) ListOpenstackTenantsNoCredentials(params *ListOpenstackTenantsNoCredentialsParams, authInfo runtime.ClientAuthInfoWriter, opts ...ClientOption) (*ListOpenstackTenantsNoCredentialsOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewListOpenstackTenantsNoCredentialsParams()\n\t}\n\top := &runtime.ClientOperation{\n\t\tID: \"listOpenstackTenantsNoCredentials\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/api/v1/projects/{project_id}/dc/{dc}/clusters/{cluster_id}/providers/openstack/tenants\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &ListOpenstackTenantsNoCredentialsReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t}\n\tfor _, opt := range opts {\n\t\topt(op)\n\t}\n\n\tresult, err := a.transport.Submit(op)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*ListOpenstackTenantsNoCredentialsOK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\tunexpectedSuccess := result.(*ListOpenstackTenantsNoCredentialsDefault)\n\treturn nil, runtime.NewAPIError(\"unexpected success response: content available as default response in error\", unexpectedSuccess, unexpectedSuccess.Code())\n}", "func getTenantTasks(c *cli.Context, w io.Writer) error {\n\terr := checkArgCount(c, 1)\n\tif err != nil {\n\t\treturn err\n\t}\n\tid := c.Args().First()\n\n\tstate := c.String(\"state\")\n\toptions := &photon.TaskGetOptions{\n\t\tState: state,\n\t}\n\n\tclient.Photonclient, err = client.GetClient(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttaskList, err := client.Photonclient.Tenants.GetTasks(id, options)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = printTaskList(taskList.Items, c)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func CreateTenants(num int) {\n\tconnStr := viper.GetString(\"database_url\")\n\tfmt.Printf(\"Conn string is %s\\n\", connStr)\n\tdb, err := sql.Open(\"postgres\", connStr)\n\tif err != nil {\n\t\tfmt.Println(\"Failed to open DB: \", err)\n\t}\n\n\tdefer db.Close()\n\n\ttablesNum := viper.GetInt(\"tables\")\n\trecordsNum := viper.GetInt(\"records\")\n\n\tfor i := 1; i <= num; i++ {\n\t\ttenantName := fmt.Sprintf(\"client%d\", i)\n\t\tif err := createTenant(tenantName, db, tablesNum, recordsNum); err != nil {\n\t\t\tfmt.Printf(\"Failed to create client %d: %s\\n\", i, err)\n\t\t\treturn\n\t\t}\n\t\tfmt.Printf(\"Tenant %s created\\n\", tenantName)\n\t}\n}", "func GetTenantsCommand() cli.Command {\n\tcommand := cli.Command{\n\t\tName: \"tenant\",\n\t\tUsage: \"options for tenant\",\n\t\tSubcommands: []cli.Command{\n\t\t\t{\n\t\t\t\tName: \"create\",\n\t\t\t\tUsage: \"Create a new tenant\",\n\t\t\t\tArgsUsage: \"<tenant-name>\",\n\t\t\t\tFlags: []cli.Flag{\n\t\t\t\t\tcli.StringFlag{\n\t\t\t\t\t\tName: \"security-groups, s\",\n\t\t\t\t\t\tUsage: \"Comma-separated Lightwave group names, to specify the tenant administrators\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := createTenant(c, os.Stdout)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"delete\",\n\t\t\t\tUsage: \"Delete a tenant\",\n\t\t\t\tArgsUsage: \"<tenant-id>\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := deleteTenant(c)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"show\",\n\t\t\t\tUsage: \"Show detailed tenant info with specified id\",\n\t\t\t\tArgsUsage: \"<tenant-id>\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := showTenant(c, os.Stdout)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"list\",\n\t\t\t\tUsage: \"List all tenants\",\n\t\t\t\tArgsUsage: \" \",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := listTenants(c, os.Stdout)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"set\",\n\t\t\t\tUsage: \"Set default tenant\",\n\t\t\t\tArgsUsage: \"<tenant-name>\",\n\t\t\t\tDescription: \"Set the default project that will be used for all photon CLI commands that need a project.\\n\" +\n\t\t\t\t\t\" Most commands allow you to override the default.\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := setTenant(c)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"get\",\n\t\t\t\tUsage: \"Get default tenant\",\n\t\t\t\tArgsUsage: \" \",\n\t\t\t\tDescription: \"Show default project in use for photon CLI commands. Most command allow you to either\\n\" +\n\t\t\t\t\t\" use this default or specify a specific project to use.\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := getTenant(c, os.Stdout)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"tasks\",\n\t\t\t\tUsage: \"Show tenant tasks\",\n\t\t\t\tArgsUsage: \"<tenant-id>\",\n\t\t\t\tFlags: []cli.Flag{\n\t\t\t\t\tcli.StringFlag{\n\t\t\t\t\t\tName: \"state, s\",\n\t\t\t\t\t\tUsage: \"Filter by task sate\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := getTenantTasks(c, os.Stdout)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"set-security-groups\",\n\t\t\t\tUsage: \"Set security groups for a tenant\",\n\t\t\t\tArgsUsage: \"<tenant-id> <comma separated list of groups>\",\n\t\t\t\tDescription: \"Set the list of Lightwave groups that can administer this tenant. This may only be\\n\" +\n\t\t\t\t\t\" be set by a member of the tenant. Be cautious--you can remove your own access if you specify\\n\" +\n\t\t\t\t\t\" the wrong set of groups.\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := setSecurityGroups(c)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tHidden: true,\n\t\t\t\tName: \"set_security_groups\",\n\t\t\t\tUsage: \"Set security groups for a tenant\",\n\t\t\t\tArgsUsage: \"<tenant-id> <comma separated list of groups>\",\n\t\t\t\tDescription: \"Deprecated, use set-security-groups instead\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := setSecurityGroups(c)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\treturn command\n}", "func (m *ListTenantsV1Response) Validate() error {\n\tif m == nil {\n\t\treturn nil\n\t}\n\n\tfor idx, item := range m.GetTenants() {\n\t\t_, _ = idx, item\n\n\t\tif v, ok := interface{}(item).(interface{ Validate() error }); ok {\n\t\t\tif err := v.Validate(); err != nil {\n\t\t\t\treturn ListTenantsV1ResponseValidationError{\n\t\t\t\t\tfield: fmt.Sprintf(\"Tenants[%v]\", idx),\n\t\t\t\t\treason: \"embedded message failed validation\",\n\t\t\t\t\tcause: err,\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\treturn nil\n}", "func (c *TenantController) Show(ctx *app.ShowTenantContext) error {\n\tuserToken := goajwt.ContextJWT(ctx)\n\tif userToken == nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewUnauthorizedError(\"Missing JWT token\"))\n\t}\n\n\tttoken := &auth.TenantToken{Token: userToken}\n\ttenantID := ttoken.Subject()\n\ttenant, err := c.tenantService.GetTenant(tenantID)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewNotFoundError(\"tenants\", tenantID.String()))\n\t}\n\n\tnamespaces, err := c.tenantService.GetNamespaces(tenantID)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, err)\n\t}\n\n\tresult := &app.TenantSingle{Data: convertTenant(ctx, tenant, namespaces, c.clusterService.GetCluster)}\n\treturn ctx.OK(result)\n}", "func (a *AccountClient) List(paging PagingParams) (*Resources, error) {\n\n\tr := a.client.R().SetResult(&Resources{})\n\n\tif paging.number != \"\" {\n\t\tr.SetQueryParam(\"page[number]\", paging.number)\n\t}\n\n\tif paging.size != \"\" {\n\t\tr.SetQueryParam(\"page[size]\", paging.size)\n\t}\n\tresp, err := r.Get(\"/v1/organisation/accounts\")\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"list accounts failed: %s\", err)\n\t}\n\n\tif resp.Error() != nil {\n\t\treturn nil, getAPIError(resp)\n\t}\n\n\treturn resp.Result().(*Resources), nil\n}", "func main() {\n c, err := identity.NewIdentityClientWithConfigurationProvider(common.DefaultConfigProvider())\n if err != nil {\n fmt.Println(\"Error:\", err)\n return\n }\n\n // The OCID of the tenancy containing the compartment.\n tenancyID, err := common.DefaultConfigProvider().TenancyOCID()\n if err != nil {\n fmt.Println(\"Error:\", err)\n return\n }\n\n request := identity.ListAvailabilityDomainsRequest{\n CompartmentId: &tenancyID,\n }\n\n r, err := c.ListAvailabilityDomains(context.Background(), request)\n if err != nil {\n fmt.Println(\"Error:\", err)\n return\n }\n\n fmt.Printf(\"List of available domains: %v\\n\", r.Items)\n\n return\n}", "func (c *TenantController) Show(ctx *app.ShowTenantContext) error {\n\ttoken := goajwt.ContextJWT(ctx)\n\tif token == nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewUnauthorizedError(\"Missing JWT token\"))\n\t}\n\n\tttoken := &TenantToken{token: token}\n\ttenantID := ttoken.Subject()\n\ttenant, err := c.tenantService.GetTenant(tenantID)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewNotFoundError(\"tenants\", tenantID.String()))\n\t}\n\n\tnamespaces, err := c.tenantService.GetNamespaces(tenantID)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, err)\n\t}\n\n\tresult := &app.TenantSingle{Data: convertTenant(ctx, tenant, namespaces, c.resolveCluster)}\n\treturn ctx.OK(result)\n}", "func (q tenantQuery) All(ctx context.Context, exec boil.ContextExecutor) (TenantSlice, error) {\n\tvar o []*Tenant\n\n\terr := q.Bind(ctx, exec, &o)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"dbmodel: failed to assign all query results to Tenant slice\")\n\t}\n\n\tif len(tenantAfterSelectHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterSelectHooks(ctx, exec); err != nil {\n\t\t\t\treturn o, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn o, nil\n}", "func getAccounts() ([]string, error) {\n\tout, err := exec.Command(\"ykman\", \"oath\", \"accounts\", \"list\").Output()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t//fmt.Printf(\"Cmd out:\\n%s\\n\", out)\n\treturn strings.Split(strings.ReplaceAll(string(out), \"\\r\\n\", \"\\n\"), \"\\n\"), nil\n}", "func (r *mutationResolver) DeleteTenants(ctx context.Context, in []string) (int, error) {\n\treturn r.tenant.Delete(ctx, in)\n}", "func (c *Client) ListGrants(ctx context.Context, params *ListGrantsInput, optFns ...func(*Options)) (*ListGrantsOutput, error) {\n\tif params == nil {\n\t\tparams = &ListGrantsInput{}\n\t}\n\n\tresult, metadata, err := c.invokeOperation(ctx, \"ListGrants\", params, optFns, c.addOperationListGrantsMiddlewares)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tout := result.(*ListGrantsOutput)\n\tout.ResultMetadata = metadata\n\treturn out, nil\n}", "func (t *Tenants) Get(id string) (*Tenant, error) {\n\tvalue, err := t.store.Get(id)\n\tif err != nil {\n\t\treturn &Tenant{}, err\n\t}\n\tr := bytes.NewReader([]byte(value))\n\treturn Decode(r)\n}", "func (tq *TenantQuery) All(ctx context.Context) ([]*Tenant, error) {\n\tctx = setContextOp(ctx, tq.ctx, \"All\")\n\tif err := tq.prepareQuery(ctx); err != nil {\n\t\treturn nil, err\n\t}\n\tqr := querierAll[[]*Tenant, *TenantQuery]()\n\treturn withInterceptors[[]*Tenant](ctx, tq, qr, tq.inters)\n}", "func getTenant(c *cli.Context, w io.Writer) error {\n\terr := checkArgCount(c, 0)\n\tif err != nil {\n\t\treturn err\n\t}\n\tconfig, err := cf.LoadConfig()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttenant := config.Tenant\n\tif tenant == nil {\n\t\tfmt.Printf(\"No tenant selected\\n\")\n\t} else {\n\t\tif c.GlobalIsSet(\"non-interactive\") {\n\t\t\tfmt.Printf(\"%s\\t%s\\n\", tenant.ID, tenant.Name)\n\t\t} else if utils.NeedsFormatting(c) {\n\t\t\tutils.FormatObject(tenant, w, c)\n\t\t} else {\n\t\t\tfmt.Printf(\"Current tenant is '%s' with ID %s\\n\", tenant.Name, tenant.ID)\n\t\t}\n\t}\n\treturn nil\n}", "func ShowTenant(ctx context.Context, config tenantConfig) (*tenant.TenantSingle, error) {\n\n\tc, err := createClient(ctx, config)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tres, err := c.ShowTenant(goasupport.ForwardContextRequestID(ctx), tenant.ShowTenantPath())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif res.StatusCode == http.StatusOK {\n\t\ttenant, err := c.DecodeTenantSingle(res)\n\t\tif err != nil {\n\t\t\treturn nil, errors.NewInternalError(ctx, err)\n\t\t}\n\t\treturn tenant, nil\n\t} else if res.StatusCode > 400 {\n\t\tjsonErr, err := c.DecodeJSONAPIErrors(res)\n\t\tif err == nil {\n\t\t\tif len(jsonErr.Errors) > 0 {\n\t\t\t\treturn nil, errors.NewInternalError(ctx, fmt.Errorf(jsonErr.Errors[0].Detail))\n\t\t\t}\n\t\t}\n\t}\n\treturn nil, errors.NewInternalError(ctx, fmt.Errorf(\"Unknown response \"+res.Status))\n}", "func List(r *http.Request, cfg config.Config) (int, http.Header, []byte, error) {\n\n\t//STANDARD DECLARATIONS START\n\tcode := http.StatusOK\n\th := http.Header{}\n\toutput := []byte(\"\")\n\terr := error(nil)\n\tcharset := \"utf-8\"\n\t//STANDARD DECLARATIONS END\n\n\t// Set Content-Type response Header value\n\tcontentType := r.Header.Get(\"Accept\")\n\th.Set(\"Content-Type\", fmt.Sprintf(\"%s; charset=%s\", contentType, charset))\n\n\t// Try to open the mongo session\n\tsession, err := mongo.OpenSession(cfg.MongoDB)\n\tdefer session.Close()\n\n\tif err != nil {\n\t\tcode = http.StatusInternalServerError\n\t\treturn code, h, output, err\n\t}\n\n\t// Create structure for storing query results\n\tresults := []Tenant{}\n\t// Query tenant collection for all available documents.\n\t// nil query param == match everything\n\terr = mongo.Find(session, cfg.MongoDB.Db, \"tenants\", nil, \"name\", &results)\n\n\tif err != nil {\n\t\tcode = http.StatusInternalServerError\n\t\treturn code, h, output, err\n\t}\n\t// After successfully retrieving the db results\n\t// call the createView function to render them into idented xml\n\toutput, err = createListView(results, \"Success\", code)\n\n\tif err != nil {\n\t\tcode = http.StatusInternalServerError\n\t\treturn code, h, output, err\n\t}\n\n\th.Set(\"Content-Type\", fmt.Sprintf(\"%s; charset=%s\", contentType, charset))\n\treturn code, h, output, err\n}", "func (tm *TenantsManager) FetchTenant(ctx context.Context, externalTenantID string) (*model.BusinessTenantMappingInput, error) {\n\tadditionalFields := map[string]string{\n\t\ttm.config.QueryConfig.EntityField: externalTenantID,\n\t}\n\tconfigProvider := eventsQueryConfigProviderWithAdditionalFields(tm.config, additionalFields)\n\n\tfetchedTenants, err := fetchCreatedTenantsWithRetries(ctx, tm.eventAPIClient, tm.config.RetryAttempts, tm.supportedEventTypes, configProvider)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif len(fetchedTenants) >= 1 {\n\t\tlog.C(ctx).Infof(\"Tenant found from central region with universal client\")\n\t\treturn &fetchedTenants[0], err\n\t}\n\n\tlog.C(ctx).Infof(\"Tenant not found from central region, checking regional APIs\")\n\n\ttenantChan := make(chan *model.BusinessTenantMappingInput, len(tm.regionalClients))\n\tfor region, regionalClient := range tm.regionalClients {\n\t\tgo func(ctx context.Context, region string, regionalClient EventAPIClient, ch chan *model.BusinessTenantMappingInput) {\n\t\t\tctx = context.WithValue(ctx, TenantRegionCtxKey, region)\n\t\t\tcreatedRegionalTenants, err := fetchCreatedTenantsWithRetries(ctx, regionalClient, tm.config.RetryAttempts, tm.supportedEventTypes, configProvider)\n\t\t\tif err != nil {\n\t\t\t\tlog.C(ctx).WithError(err).Errorf(\"Failed to fetch created tenants from region %s: %v\", region, err)\n\t\t\t}\n\n\t\t\tif len(createdRegionalTenants) == 1 {\n\t\t\t\tlog.C(ctx).Infof(\"Tenant found in region %s\", region)\n\t\t\t\tif createdRegionalTenants[0].Region == \"\" {\n\t\t\t\t\tcreatedRegionalTenants[0].Region = region\n\t\t\t\t}\n\t\t\t\tch <- &createdRegionalTenants[0]\n\t\t\t} else {\n\t\t\t\tlog.C(ctx).Warnf(\"Tenant not found in region %s\", region)\n\t\t\t\tch <- nil\n\t\t\t}\n\t\t}(ctx, region, regionalClient, tenantChan)\n\t}\n\n\tpendingRegionalInfo := len(tm.regionalClients)\n\tif pendingRegionalInfo == 0 {\n\t\t// TODO return error when lazy store is reverted\n\t\tlog.C(ctx).Error(\"no regions are configured\")\n\t\treturn nil, nil\n\t}\n\n\tvar tenant *model.BusinessTenantMappingInput\n\tfor result := range tenantChan {\n\t\tif result != nil {\n\t\t\ttenant = result\n\t\t\tbreak\n\t\t}\n\t\tpendingRegionalInfo--\n\t\tif pendingRegionalInfo == 0 {\n\t\t\t// TODO return error when lazy store is reverted\n\t\t\tlog.C(ctx).Error(\"tenant not found in all configured regions\")\n\t\t\treturn nil, nil\n\t\t}\n\t}\n\n\treturn tenant, nil\n}", "func Accounts(client *ticketmatic.Client) ([]*ticketmatic.AccountInfo, error) {\n\tr := client.NewRequest(\"GET\", \"/_/tools/accounts\", \"json\")\n\n\tvar obj []*ticketmatic.AccountInfo\n\terr := r.Run(&obj)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn obj, nil\n}", "func (a *AuthorizationsService) All() (auths []Authorization, result *Result) {\n\tresult = a.client.get(a.URL, &auths)\n\treturn\n}", "func (client AppsClient) ListCortanaEndpoints(ctx context.Context) (result PersonalAssistantsResponse, err error) {\n\treq, err := client.ListCortanaEndpointsPreparer(ctx)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"programmatic.AppsClient\", \"ListCortanaEndpoints\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresp, err := client.ListCortanaEndpointsSender(req)\n\tif err != nil {\n\t\tresult.Response = autorest.Response{Response: resp}\n\t\terr = autorest.NewErrorWithError(err, \"programmatic.AppsClient\", \"ListCortanaEndpoints\", resp, \"Failure sending request\")\n\t\treturn\n\t}\n\n\tresult, err = client.ListCortanaEndpointsResponder(resp)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"programmatic.AppsClient\", \"ListCortanaEndpoints\", resp, \"Failure responding to request\")\n\t}\n\n\treturn\n}", "func (client IdentityClient) getTenancy(ctx context.Context, request common.OCIRequest, binaryReqBody *common.OCIReadSeekCloser, extraHeaders map[string]string) (common.OCIResponse, error) {\n\n\thttpRequest, err := request.HTTPRequest(http.MethodGet, \"/tenancies/{tenancyId}\", binaryReqBody, extraHeaders)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar response GetTenancyResponse\n\tvar httpResponse *http.Response\n\thttpResponse, err = client.Call(ctx, &httpRequest)\n\tdefer common.CloseBodyIfValid(httpResponse)\n\tresponse.RawResponse = httpResponse\n\tif err != nil {\n\t\treturn response, err\n\t}\n\n\terr = common.UnmarshalResponse(httpResponse, &response)\n\treturn response, err\n}", "func (am *AS3Manager) getTenantObjects(partitions []string) string {\n\tvar as3Config map[string]interface{}\n\tbaseAS3ConfigEmpty := fmt.Sprintf(baseAS3Config, am.as3Version, am.as3Release, am.as3SchemaVersion)\n\t_ = json.Unmarshal([]byte(baseAS3ConfigEmpty), &as3Config)\n\tdecl := as3Config[\"declaration\"].(map[string]interface{})\n\tfor _, partition := range partitions {\n\t\tdecl[partition] = map[string]string{\"class\": \"Tenant\"}\n\t}\n\tdata, _ := json.Marshal(as3Config)\n\treturn string(data)\n}", "func (m *ManagementTemplateStepTenantSummary) GetIneligibleTenantsCount()(*int32) {\n val, err := m.GetBackingStore().Get(\"ineligibleTenantsCount\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*int32)\n }\n return nil\n}", "func getCurrentTenant() *Tenant {\n\tif currentTenant == nil {\n\t\ttenants, err := providers.Tenants()\n\t\tif err != nil || len(tenants) != 1 {\n\t\t\treturn nil\n\t\t}\n\t\t// Set unqiue tenant as selected\n\t\tlog.Println(\"Unique tenant set\")\n\t\tfor name := range tenants {\n\t\t\tservice, err := providers.GetService(name)\n\t\t\tif err != nil {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\tcurrentTenant = &Tenant{name: name, Service: service}\n\t\t}\n\t}\n\treturn currentTenant\n}", "func (a *Client) GetTenant(params *GetTenantParams) (*GetTenantOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetTenantParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"getTenant\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/tenants/{id}\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\"},\n\t\tParams: params,\n\t\tReader: &GetTenantReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*GetTenantOK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\t// safeguard: normally, absent a default response, unknown success responses return an error above: so this is a codegen issue\n\tmsg := fmt.Sprintf(\"unexpected success response for getTenant: API contract not enforced by server. Client expected to get an error, but got: %T\", result)\n\tpanic(msg)\n}", "func (t Client) GetTenantInfo(tenantID string) (ti Info, err error) {\n\tif len(tenantID) == 0 {\n\t\terr = errors.New(\"tenant ID cannot be empty string\")\n\t\treturn\n\t}\n\treq, err := http.NewRequest(\"GET\", fmt.Sprintf(\"%s/api/tenants/%s\", t.tenantServiceURL, tenantID), nil)\n\tif err != nil {\n\t\treturn\n\t}\n\treq.Header.Set(\"Authorization\", fmt.Sprintf(\"Bearer %s\", t.authToken))\n\n\tt.logger.WithFields(log.Fields{\n\t\t\"type\": \"id\",\n\t\t\"id\": tenantID,\n\t}).Info(\"Tenant by id\")\n\n\tresp, err := t.client.Do(req)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tdefer resp.Body.Close()\n\tbody, err := ioutil.ReadAll(resp.Body)\n\terr = json.Unmarshal(body, &ti)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tif len(ti.Errors) > 0 {\n\t\terr = fmt.Errorf(\"%+v\", ti.Errors)\n\t}\n\n\treturn\n}", "func (repository *GormRepository) GetAllForTenant(uow *UnitOfWork, out interface{}, tenantID uuid.UUID, queryProcessors []QueryProcessor) microappError.DatabaseError {\n\tqueryProcessors = append([]QueryProcessor{Filter(\"tenantID = ?\", tenantID)}, queryProcessors...)\n\treturn repository.GetAll(uow, out, queryProcessors)\n}", "func (am *AS3Manager) getTenantObjects(partitions []string) string {\n\tvar as3Config map[string]interface{}\n\tbaseAS3ConfigEmpty := fmt.Sprintf(baseAS3Config, am.as3Version, am.as3Release, am.as3Version)\n\t_ = json.Unmarshal([]byte(baseAS3ConfigEmpty), &as3Config)\n\tdecl := as3Config[\"declaration\"].(map[string]interface{})\n\tfor _, partition := range partitions {\n\n\t\tdecl[partition] = map[string]string{\"class\": \"Tenant\"}\n\t}\n\tdata, _ := json.Marshal(as3Config)\n\treturn string(data)\n}", "func (s *Server) GetAllPlants(ctx context.Context, e *Empty) (*Plants, error) {\n\tplants, err := getAllPlantsFromDB()\n\tif err != nil {\n\t\tlog.Fatalf(\"Failure fetching all the plants from the DB: %s\\n\", err)\n\t}\n\n\treturn plants, nil\n}", "func GetAll(ctx context.Context, client *selvpcclient.ServiceClient) ([]*Quota, *selvpcclient.ResponseResult, error) {\n\turl := strings.Join([]string{client.Endpoint, resourceURL}, \"/\")\n\tresponseResult, err := client.DoRequest(ctx, http.MethodGet, url, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tif responseResult.Err != nil {\n\t\treturn nil, responseResult, responseResult.Err\n\t}\n\n\t// Extract quotas from the response body.\n\tvar result ResourcesQuotas\n\terr = responseResult.ExtractResult(&result)\n\tif err != nil {\n\t\treturn nil, responseResult, err\n\t}\n\n\treturn result.Quotas, responseResult, nil\n}", "func GetAllTenant(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(Tenant))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tif strings.Contains(k, \"isnull\") {\n\t\t\tqs = qs.Filter(k, (v == \"true\" || v == \"1\"))\n\t\t} else {\n\t\t\tqs = qs.Filter(k, v)\n\t\t}\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []Tenant\n\tqs = qs.OrderBy(sortFields...)\n\tif _, err = qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func List(params ...int) (DataList, error) {\n\tqueryParams := map[string]string{}\n\n\tif len(params) == 2 {\n\t\tqueryParams[\"page[number]\"] = strconv.Itoa(params[0])\n\t\tqueryParams[\"page[size]\"] = strconv.Itoa(params[1])\n\t}\n\n\tresponseStatus, responsePayload, err := doRequest(&request{\n\t\tmethod: \"GET\",\n\t\tresource: \"v1/organisation/accounts/\",\n\t\tqueryParams: queryParams,\n\t})\n\n\tif err != nil {\n\t\treturn DataList{}, err\n\t}\n\treturn handleResponseDataList(responsePayload, http.StatusOK, responseStatus)\n}", "func GetAccounts(w http.ResponseWriter, r *http.Request) {\n\n\t// Add header so that received knows they're receiving JSON\n\tw.Header().Add(\"Content-Type\", \"application/json\")\n\n\t// Retrieving name of node from query request\n\tnodeName := r.URL.Query().Get(\"name\")\n\tconfirmation, socket := checkNodeName(nodeName)\n\tif confirmation == false {\n\n\t\t// Stop code here no need to establish connection and reply\n\t\tjson.NewEncoder(w).Encode(responses.ErrorResponse{\n\t\t\tError: \"Node name requested doesn't exist\"})\n\t\treturn\n\t}\n\n\t// Retrieving height from query request\n\trecvHeight := r.URL.Query().Get(\"height\")\n\theight := checkHeight(recvHeight)\n\tif height == -1 {\n\n\t\t// Stop code here no need to establish connection and reply\n\t\tjson.NewEncoder(w).Encode(responses.ErrorResponse{\n\t\t\tError: \"Unexepcted value found, height needs to be string of int!\"})\n\t\treturn\n\t}\n\n\t// Attempt to load connection with staking client\n\tconnection, so := loadStakingClient(socket)\n\n\t// Close connection once code underneath executes\n\tdefer connection.Close()\n\n\t// If null object was retrieved send response\n\tif so == nil {\n\n\t\t// Stop code here faild to establish connection and reply\n\t\tjson.NewEncoder(w).Encode(responses.ErrorResponse{\n\t\t\tError: \"Failed to establish connection using socket : \" + socket})\n\t\treturn\n\t}\n\n\t// Return accounts from staking client\n\taccounts, err := so.Addresses(context.Background(), height)\n\tif err != nil {\n\t\tjson.NewEncoder(w).Encode(responses.ErrorResponse{\n\t\t\tError: \"Failed to get Accounts!\"})\n\t\tlgr.Error.Println(\n\t\t\t\"Request at /api/staking/accounts failed to retrieve Accounts : \",\n\t\t\terr)\n\t\treturn\n\t}\n\n\t// Respond with array of all accounts\n\tlgr.Info.Println(\"Request at /api/staking/accounts responding with \" +\n\t\t\"Accounts!\")\n\tjson.NewEncoder(w).Encode(responses.AllAccountsResponse{AllAccounts: accounts})\n}", "func (s *AppsServiceOp) ListTiers(ctx context.Context) ([]*AppTier, *Response, error) {\n\tpath := fmt.Sprintf(\"%s/tiers\", appsBasePath)\n\treq, err := s.client.NewRequest(ctx, http.MethodGet, path, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\troot := new(appTiersRoot)\n\tresp, err := s.client.Do(ctx, req, root)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\treturn root.Tiers, resp, nil\n}", "func CheckTenantAccess(c echo.Context) *echo.HTTPError {\n\tif !TLS {\n\t\treturn nil\n\t}\n\n\tcert := getCert(c)\n\tif cert == nil {\n\t\terr := errors.New(\"Client certificate with valid SANs is required for tenant access\")\n\t\treturn echo.NewHTTPError(http.StatusForbidden, err.Error())\n\t}\n\n\tif cert.Subject.CommonName == wildcard || cert.Subject.CommonName == networkWildcard {\n\t\treturn nil\n\t}\n\tfor _, san := range cert.DNSNames {\n\t\tif san == wildcard || san == networkWildcard {\n\t\t\treturn nil\n\t\t}\n\t}\n\tglog.Infof(\"Client cert %s does not have wildcard access\", util.FormatPkixSubject(&cert.Subject))\n\treturn echo.NewHTTPError(http.StatusForbidden, \"Client certificate is not authorized\")\n}", "func (t Tenant) GetTenantInfo(tenantId string) (ti TenantInfo, err error) {\n\treq, err := http.NewRequest(\"GET\", fmt.Sprintf(\"%s/api/tenants/%s\", t.tenantServiceURL, tenantId), nil)\n\tif err != nil {\n\t\treturn\n\t}\n\treq.Header.Set(\"Authorization\", fmt.Sprintf(\"Bearer %s\", t.authToken))\n\n\tclient := &http.Client{}\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tdefer resp.Body.Close()\n\tbody, err := ioutil.ReadAll(resp.Body)\n\terr = json.Unmarshal(body, &ti)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tif len(ti.Errors) > 0 {\n\t\terr = errors.New(fmt.Sprintf(\"%+v\", ti.Errors))\n\t}\n\n\treturn\n}", "func (a *Client) GrantsRead(params *GrantsReadParams, authInfo runtime.ClientAuthInfoWriter) (*GrantsReadOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGrantsReadParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"grants_read\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/v3/customer/core/grants/{id}/\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &GrantsReadReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*GrantsReadOK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\t// safeguard: normally, absent a default response, unknown success responses return an error above: so this is a codegen issue\n\tmsg := fmt.Sprintf(\"unexpected success response for grants_read: API contract not enforced by server. Client expected to get an error, but got: %T\", result)\n\tpanic(msg)\n}", "func GetCustomerList() (err error) {\n\n\treturn err\n}", "func (client IdentityClient) GetTenancy(ctx context.Context, request GetTenancyRequest) (response GetTenancyResponse, err error) {\n\tvar ociResponse common.OCIResponse\n\tpolicy := common.NoRetryPolicy()\n\tif client.RetryPolicy() != nil {\n\t\tpolicy = *client.RetryPolicy()\n\t}\n\tif request.RetryPolicy() != nil {\n\t\tpolicy = *request.RetryPolicy()\n\t}\n\tociResponse, err = common.Retry(ctx, request, client.getTenancy, policy)\n\tif err != nil {\n\t\tif ociResponse != nil {\n\t\t\tif httpResponse := ociResponse.HTTPResponse(); httpResponse != nil {\n\t\t\t\topcRequestId := httpResponse.Header.Get(\"opc-request-id\")\n\t\t\t\tresponse = GetTenancyResponse{RawResponse: httpResponse, OpcRequestId: &opcRequestId}\n\t\t\t} else {\n\t\t\t\tresponse = GetTenancyResponse{}\n\t\t\t}\n\t\t}\n\t\treturn\n\t}\n\tif convertedResponse, ok := ociResponse.(GetTenancyResponse); ok {\n\t\tresponse = convertedResponse\n\t} else {\n\t\terr = fmt.Errorf(\"failed to convert OCIResponse into GetTenancyResponse\")\n\t}\n\treturn\n}", "func TenantExists(name string) bool {\n\tvar tenant bool\n\n\tc, err := getIdentityClient()\n\tif err != nil {\n\t\tlog.Println(\"Error getting Identity Client: \", err)\n\t\treturn false\n\t}\n\n\topts := tenants.ListOpts{Limit: 20}\n\tpager := tenants.List(c, &opts)\n\t// brute force the whole tenant list to get tenant details?\n\tpager.EachPage(\n\t\tfunc(page pagination.Page) (bool, error) {\n\t\t\ttenantList, _ := tenants.ExtractTenants(page)\n\t\t\tfor _, t := range tenantList {\n\t\t\t\t// \"t\" is tenants.Tenant\n\t\t\t\tif t.ID == name || t.Name == name {\n\t\t\t\t\ttenant = true\n\t\t\t\t\t// stop iterating and return tenant\n\t\t\t\t\treturn false, nil\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn true, nil\n\t\t},\n\t)\n\treturn tenant\n}", "func (c Client) FetchMerchants(appID string, limit int) ([]*Merchant, error) {\n\treturn c.fetchMerchants(appID, limit)\n}", "func (t *ManageMerchant) getAllMerchants(stub shim.ChaincodeStubInterface, args []string) ([]byte, error) {\r\n\tvar jsonResp, errResp string\r\n\tvar merchantIndex []string\r\n\tvar err error\r\n\tfmt.Println(\"start getAllMerchants\")\r\n\t\t\r\n\tmerchantAsBytes, err := stub.GetState(MerchantIndexStr)\r\n\tif err != nil {\r\n\t\treturn nil, errors.New(\"Failed to get Merchant index\")\r\n\t}\r\n\tjson.Unmarshal(merchantAsBytes, &merchantIndex)\t\t\t\t\t\t\t\t//un stringify it aka JSON.parse()\r\n\tjsonResp = \"{\"\r\n\tfor i,val := range merchantIndex{\r\n\t\tfmt.Println(strconv.Itoa(i) + \" - looking at \" + val + \" for all Merchant\")\r\n\t\tvalueAsBytes, err := stub.GetState(val)\r\n\t\tif err != nil {\r\n\t\t\terrResp = \"{\\\"Error\\\":\\\"Failed to get state for \" + val + \"\\\"}\"\r\n\t\t\treturn nil, errors.New(errResp)\r\n\t\t}\r\n\t\tfmt.Print(\"valueAsBytes : \")\r\n\t\tfmt.Println(valueAsBytes)\r\n\t\tjsonResp = jsonResp + \"\\\"\"+ val + \"\\\":\" + string(valueAsBytes[:])\r\n\t\tif i < len(merchantIndex)-1 {\r\n\t\t\tjsonResp = jsonResp + \",\"\r\n\t\t}\r\n\t}\r\n\tjsonResp = jsonResp + \"}\"\r\n\tif strings.Contains(jsonResp, \"},}\"){\r\n\t\tfmt.Println(\"in if for jsonResp contains wrong json\")\t\r\n\t\tjsonResp = strings.Replace(jsonResp, \"},}\", \"}}\", -1)\r\n\t}\r\n\tfmt.Println(\"jsonResp in getAllMerchants::\")\r\n\tfmt.Println(jsonResp)\r\n\r\n\tfmt.Println(\"end getAllMerchants\")\r\n\treturn []byte(jsonResp), nil\t\t\t//send it onward\r\n}", "func ListEndpoint(projectProvider provider.ProjectProvider, privilegedProjectProvider provider.PrivilegedProjectProvider, serviceAccountProvider provider.ServiceAccountProvider, privilegedServiceAccount provider.PrivilegedServiceAccountProvider, memberMapper provider.ProjectMemberMapper, userInfoGetter provider.UserInfoGetter) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\t\treq, ok := request.(common.GetProjectRq)\n\t\tif !ok {\n\t\t\treturn nil, errors.NewBadRequest(\"invalid request\")\n\t\t}\n\n\t\tif len(req.ProjectID) == 0 {\n\t\t\treturn nil, errors.NewBadRequest(\"the name of the project cannot be empty\")\n\t\t}\n\n\t\tproject, err := common.GetProject(ctx, userInfoGetter, projectProvider, privilegedProjectProvider, req.ProjectID)\n\t\tif err != nil {\n\t\t\treturn nil, common.KubernetesErrorToHTTPError(err)\n\t\t}\n\t\tsaList, err := listSA(ctx, serviceAccountProvider, privilegedServiceAccount, userInfoGetter, project, nil)\n\t\tif err != nil {\n\t\t\treturn nil, common.KubernetesErrorToHTTPError(err)\n\t\t}\n\n\t\tvar errorList []string\n\t\tresponse := make([]*apiv1.ServiceAccount, 0)\n\t\tfor _, sa := range saList {\n\t\t\texternalSA := convertInternalServiceAccountToExternal(sa)\n\t\t\tif apiv1.ServiceAccountInactive == externalSA.Status {\n\t\t\t\tresponse = append(response, externalSA)\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tgroup, err := memberMapper.MapUserToGroup(sa.Spec.Email, project.Name)\n\t\t\tif err != nil {\n\t\t\t\terrorList = append(errorList, err.Error())\n\t\t\t} else {\n\t\t\t\texternalSA.Group = group\n\t\t\t\tresponse = append(response, externalSA)\n\t\t\t}\n\t\t}\n\t\tif len(errorList) > 0 {\n\t\t\treturn response, errors.NewWithDetails(http.StatusInternalServerError, \"failed to get some service accounts, please examine details field for more info\", errorList)\n\t\t}\n\n\t\treturn response, nil\n\t}\n}", "func List(helper Helper, writer io.Writer) error {\n\taccts, err := helper.List()\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn json.NewEncoder(writer).Encode(accts)\n}", "func (r *mutationResolver) WriteTenants(ctx context.Context, in []*graphql.BusinessTenantMappingInput) ([]string, error) {\n\treturn r.tenant.Write(ctx, in)\n}", "func getCustomerFullList(c *gin.Context) {\n\tmdb := db.MongoSession\n\n\tcustomers, err := mdb.GetCustomerList(10, \"\", \"\")\n\n\tif err != nil {\n\t\tc.String(500, \"{\\\"code\\\": -1, \\\"message\\\": \\\"An unexpected error occurred\\\"}\")\n\t} else {\n\t\tc.JSON(200, customers)\n\t}\n}", "func getCurrentTenant() *Tenant {\n\tif currentTenant == nil {\n\t\ttenants, err := iaas.GetTenantNames()\n\t\tif err != nil || len(tenants) != 1 {\n\t\t\treturn nil\n\t\t}\n\t\t// Set unique tenant as selected\n\t\tlog.Println(\"Unique tenant set\")\n\t\tfor name := range tenants {\n\t\t\tservice, err := iaas.UseService(name)\n\t\t\tif err != nil {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\tcurrentTenant = &Tenant{name: name, Service: service}\n\t\t}\n\t}\n\treturn currentTenant\n}", "func (a *Client) ListOpenstackTenantsNoCredentialsV2(params *ListOpenstackTenantsNoCredentialsV2Params, authInfo runtime.ClientAuthInfoWriter, opts ...ClientOption) (*ListOpenstackTenantsNoCredentialsV2OK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewListOpenstackTenantsNoCredentialsV2Params()\n\t}\n\top := &runtime.ClientOperation{\n\t\tID: \"listOpenstackTenantsNoCredentialsV2\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/api/v2/projects/{project_id}/clusters/{cluster_id}/providers/openstack/tenants\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &ListOpenstackTenantsNoCredentialsV2Reader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t}\n\tfor _, opt := range opts {\n\t\topt(op)\n\t}\n\n\tresult, err := a.transport.Submit(op)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*ListOpenstackTenantsNoCredentialsV2OK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\tunexpectedSuccess := result.(*ListOpenstackTenantsNoCredentialsV2Default)\n\treturn nil, runtime.NewAPIError(\"unexpected success response: content available as default response in error\", unexpectedSuccess, unexpectedSuccess.Code())\n}", "func (c *Client) WeeklyRants() ([]RantModel, error) {\n\turl := fmt.Sprintf(WEEKLY_PATH, API, APP_VERSION)\n\tres, err := http.Get(url)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar data RantsResponse\n\tjson.NewDecoder(res.Body).Decode(&data)\n\tif !data.Success && data.Error != \"\" {\n\t\treturn nil, errors.New(data.Error)\n\t}\n\treturn data.Rants, nil\n}", "func TestTenantsSuite(t *testing.T) {\n\tsuite.Run(t, new(TenantTestSuite))\n}", "func (owner *WalletOwnerAPI) RetrieveTxs(refreshFromNode bool, txID *uint32, txSlateID *uuid.UUID) (bool, *[]libwallet.TxLogEntry, error) {\n\tparams := struct {\n\t\tToken string `json:\"token\"`\n\t\tRefreshFromNode bool `json:\"refresh_from_node\"`\n\t\tTxID *uint32 `json:\"tx_id\"`\n\t\tTxSlateID *uuid.UUID `json:\"tx_slate_id\"`\n\t}{\n\t\tToken: owner.token,\n\t\tRefreshFromNode: refreshFromNode,\n\t\tTxID: txID,\n\t\tTxSlateID: txSlateID,\n\t}\n\n\tparamsBytes, err := json.Marshal(params)\n\tif err != nil {\n\t\treturn false, nil, err\n\t}\n\tenvl, err := owner.client.EncryptedRequest(\"retrieve_txs\", paramsBytes, owner.sharedSecret)\n\tif err != nil {\n\t\treturn false, nil, err\n\t}\n\tif envl == nil {\n\t\treturn false, nil, errors.New(\"WalletOwnerAPI: Empty RPC Response from grin-wallet\")\n\t}\n\tif envl.Error != nil {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"code\": envl.Error.Code,\n\t\t\t\"message\": envl.Error.Message,\n\t\t}).Error(\"WalletOwnerAPI: RPC Error during RetrieveTxs\")\n\t\treturn false, nil, errors.New(string(envl.Error.Code) + \"\" + envl.Error.Message)\n\t}\n\tvar result Result\n\tif err = json.Unmarshal(envl.Result, &result); err != nil {\n\t\treturn false, nil, err\n\t}\n\tif result.Err != nil {\n\t\treturn false, nil, errors.New(string(result.Err))\n\t}\n\n\tvar okArray []json.RawMessage\n\tif err = json.Unmarshal(result.Ok, &okArray); err != nil {\n\t\treturn false, nil, err\n\t}\n\tif len(okArray) < 2 {\n\t\treturn false, nil, errors.New(\"Wrong okArray length\")\n\t}\n\tvar refreshedFromNode bool\n\tif err = json.Unmarshal(okArray[0], &refreshedFromNode); err != nil {\n\t\treturn false, nil, err\n\t}\n\tvar txLogEntries []libwallet.TxLogEntry\n\tif err := json.Unmarshal(okArray[1], &txLogEntries); err != nil {\n\t\treturn false, nil, err\n\t}\n\n\treturn refreshedFromNode, &txLogEntries, nil\n}", "func (c *GethClient) Accounts(ctx context.Context) ([]string, error) {\n\tvar result []string\n\terr := c.rpcCli.CallContext(ctx, &result, \"personal_listAccounts\")\n\treturn result, err\n}", "func AllCustomersEndpoint(w http.ResponseWriter, r *http.Request) {\n\tcustomers, err := dao.FindAllCustomers()\n\n\tif err != nil {\n\t\trespondWithError(w, http.StatusInternalServerError, err.Error())\n\t\treturn\n\t}\n\trespondWithJson(w, http.StatusOK, customers)\n}", "func ListOne(r *http.Request, cfg config.Config) (int, http.Header, []byte, error) {\n\n\t//STANDARD DECLARATIONS START\n\tcode := http.StatusOK\n\th := http.Header{}\n\toutput := []byte(\"\")\n\terr := error(nil)\n\tcharset := \"utf-8\"\n\t//STANDARD DECLARATIONS END\n\n\tvars := mux.Vars(r)\n\n\t// Set Content-Type response Header value\n\tcontentType := r.Header.Get(\"Accept\")\n\th.Set(\"Content-Type\", fmt.Sprintf(\"%s; charset=%s\", contentType, charset))\n\n\t// Try to open the mongo session\n\tsession, err := mongo.OpenSession(cfg.MongoDB)\n\tdefer session.Close()\n\n\tif err != nil {\n\t\tcode = http.StatusInternalServerError\n\t\treturn code, h, output, err\n\t}\n\n\t// Create structure to hold query results\n\tresults := []Tenant{}\n\n\t// Create a simple query object to query by id\n\tquery := bson.M{\"id\": vars[\"ID\"]}\n\t// Query collection tenants for the specific tenant id\n\terr = mongo.Find(session, cfg.MongoDB.Db, \"tenants\", query, \"name\", &results)\n\n\tif err != nil {\n\t\tcode = http.StatusInternalServerError\n\t\treturn code, h, output, err\n\t}\n\n\t// Check if nothing found\n\tif len(results) < 1 {\n\t\toutput, _ = respond.MarshalContent(respond.NotFound, contentType, \"\", \" \")\n\t\tcode = http.StatusNotFound\n\t\treturn code, h, output, err\n\t}\n\n\t// After successfully retrieving the db results\n\t// call the createView function to render them into idented xml\n\toutput, err = createListView(results, \"Success\", code)\n\n\tif err != nil {\n\t\tcode = http.StatusInternalServerError\n\t\treturn code, h, output, err\n\t}\n\n\th.Set(\"Content-Type\", fmt.Sprintf(\"%s; charset=%s\", contentType, charset))\n\treturn code, h, output, err\n}", "func (v *Kounta) GetSites(token string, company string) (Sites, error) {\n\tclient := &http.Client{}\n\tclient.CheckRedirect = checkRedirectFunc\n\n\tu, _ := url.ParseRequestURI(baseURL)\n\tu.Path = fmt.Sprintf(sitesURL, company)\n\turlStr := fmt.Sprintf(\"%v\", u)\n\n\tr, err := http.NewRequest(\"GET\", urlStr, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tr.Header = http.Header(make(map[string][]string))\n\tr.Header.Set(\"Accept\", \"application/json\")\n\tr.Header.Set(\"Authorization\", \"Bearer \"+token)\n\n\tres, err := client.Do(r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\trawResBody, err := ioutil.ReadAll(res.Body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfmt.Println(\"GetSites Body\", string(rawResBody))\n\n\tif res.StatusCode == 200 {\n\t\tvar resp Sites\n\n\t\terr = json.Unmarshal(rawResBody, &resp)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn resp, nil\n\t}\n\treturn nil, fmt.Errorf(\"Failed to get Kounta Company %s\", res.Status)\n\n}", "func (c Client) fetchMerchants(appID string, limit int) ([]*Merchant, error) {\n\tpath := fmt.Sprintf(\"/identities/%s/merchants?limit=%d\", appID, limit)\n\treq, err := http.NewRequest(\"GET\", c.getURL(path), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar marshalled []*Merchant\n\treturn marshalled, c.executeRequestAndMarshal(req, &marshalled)\n}", "func (m *TenantManager) Read(opts ...RequestOption) (t *Tenant, err error) {\n\terr = m.Request(\"GET\", m.URI(\"tenants\", \"settings\"), &t, opts...)\n\treturn\n}", "func (t *Token) All(customerID string, queryParams map[string]interface{}, extraHeaders map[string]string) (map[string]interface{}, error) {\n\turl := fmt.Sprintf(\"/%s%s/%s/tokens\", constants.VERSION_V1, constants.CUSTOMER_URL, customerID)\n\treturn t.Request.Get(url, queryParams, extraHeaders)\n}", "func (oauthClient *OauthClient) GetWithGrants(db *gorm.DB) error {\n\tif err := db.Find(&oauthClient).Association(\"OauthGrants\").Find(&oauthClient.OauthGrants).Error; err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (s *Identity) AccountsGET(w http.ResponseWriter, r *http.Request) {\n\ts.mu.RLock()\n\tdefer s.mu.RUnlock()\n\twriteResponse(s.addresses, w, r)\n}", "func (client *Client) ListQuotas(tenantID string) ([]types.QuotaDetails, error) {\n\tvar result types.QuotaListResponse\n\n\turl, err := client.getCiaoQuotasResource()\n\tif err != nil {\n\t\treturn result.Quotas, errors.Wrap(err, \"Error getting quotas resource\")\n\t}\n\n\tif tenantID != \"\" {\n\t\turl = fmt.Sprintf(\"%s/%s/quotas\", url, tenantID)\n\t} else {\n\t\turl = fmt.Sprintf(\"%s/quotas\", url)\n\t}\n\n\terr = client.getResource(url, api.TenantsV1, nil, &result)\n\n\treturn result.Quotas, err\n}", "func (p *Poloniex) GetActiveLoans(ctx context.Context) (ActiveLoans, error) {\n\tresult := ActiveLoans{}\n\treturn result, p.SendAuthenticatedHTTPRequest(ctx, exchange.RestSpot, http.MethodPost, poloniexActiveLoans, url.Values{}, &result)\n}", "func GetPlants(svc *dynamodb.DynamoDB, vals ...map[string][]string) (*[]PlantInfo, int, error) {\n\t// Note: occurrence is spelt wrong\n\tvar filt expression.ConditionBuilder\n\n\tif len(vals) <= 1 && len(vals[0]) == 0 {\n\t\tfilt = expression.Name(\"PlantID\").GreaterThan(expression.Value(-1))\n\t} else {\n\t\tfor _, val := range vals {\n\t\t\tid, prs := val[\"common\"]\n\t\t\tif prs {\n\t\t\t\tfilt = expression.Name(\"Common\").Equal(expression.Value(id[0]))\n\t\t\t}\n\t\t\tid, prs = val[\"scientific\"]\n\t\t\tif prs {\n\t\t\t\tfilt = expression.Name(\"Scientific\").Equal(expression.Value(id[0]))\n\t\t\t}\n\t\t\tid, prs = val[\"id\"]\n\t\t\tif prs {\n\t\t\t\tidInt, err := strconv.Atoi(id[0])\n\t\t\t\tif err != nil {\n\t\t\t\t\tfmt.Println(\"Failed int to string conversion\")\n\t\t\t\t\tfmt.Println((err.Error()))\n\t\t\t\t\treturn nil, -1, err\n\t\t\t\t}\n\t\t\t\tfilt = expression.Name(\"PlantID\").Equal(expression.Value(idInt))\n\t\t\t}\n\t\t}\n\t}\n\n\tproj := expression.NamesList(\n\t\texpression.Name(\"PlantID\"),\n\t\texpression.Name(\"Common\"),\n\t\texpression.Name(\"Scientific\"),\n\t)\n\n\texpr, err := expression.NewBuilder().WithFilter(filt).WithProjection(proj).Build()\n\tif err != nil {\n\t\tfmt.Println(\"Got error building expression:\")\n\t\tfmt.Println(err.Error())\n\t\treturn nil, -1, err\n\t}\n\n\tparams := &dynamodb.ScanInput{\n\t\tExpressionAttributeNames: expr.Names(),\n\t\tExpressionAttributeValues: expr.Values(),\n\t\tFilterExpression: expr.Filter(),\n\t\tProjectionExpression: expr.Projection(),\n\t\tTableName: aws.String(\"Plants\"),\n\t}\n\n\tresult, err := svc.Scan(params)\n\tif err != nil {\n\t\tfmt.Println(\"Query API call failed:\")\n\t\tfmt.Println((err.Error()))\n\t\treturn nil, -1, err\n\t}\n\n\tnumItems := 0\n\tresultItems := make([]PlantInfo, 0)\n\tfor _, i := range result.Items {\n\t\titem := PlantInfo{}\n\t\terr = dynamodbattribute.UnmarshalMap(i, &item)\n\n\t\tif err != nil {\n\t\t\tfmt.Println(\"Got error unmarshalling:\")\n\t\t\tfmt.Println(err.Error())\n\t\t\treturn nil, -1, err\n\t\t}\n\n\t\tnumItems++\n\t\t//fmt.Printf(\"%+v\\n\", item)\n\t\tresultItems = append(resultItems, item)\n\t}\n\t//fmt.Printf(\"Found %d items\\n\", numItems)\n\t//fmt.Println(\"-----------------\")\n\t/*\n\t\tfor _, item := range resultItems {\n\t\t\tfmt.Printf(\"%+v\\n\", item)\n\t\t}\n\t\tfmt.Println(\"-----------------\")\n\t*/\n\treturn &resultItems, numItems, nil\n}", "func GetZones(full bool, tenant string) []Zone {\n\ttenantStr := func() string {\n\t\tif len(tenant) == 0 {\n\t\t\treturn \"\"\n\t\t}\n\t\treturn \"-tenant=\" + tenant\n\t}()\n\tfullStr := func() string {\n\t\tif full {\n\t\t\treturn \"-full\"\n\t\t}\n\t\treturn \"\"\n\t}()\n\n\toutput := RunCmd(fmt.Sprintf(\"%s api -fetch-zone-apps %s %s\", ActlPath, fullStr, tenantStr))\n\tlistOfZones := []Zone{}\n\tyaml.Unmarshal([]byte(output), &listOfZones)\n\treturn listOfZones\n}", "func getAdmins(e echo.Context) error {\n\tdb := e.Get(\"database\").(*mgo.Database)\n\tif db == nil {\n\t\treturn fmt.Errorf(\"Bad database session\")\n\t}\n\n\tuuid, err := uuid.FromString(e.QueryParam(\"siteUuid\"))\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Bad parameters\")\n\t}\n\n\ta := models.Admins{\n\t\tAdmins: []models.Admin{},\n\t}\n\n\terr = db.C(\"Admins\").Find(bson.M{\"adminSiteUuid\": uuid}).All(&a.Admins)\n\tif err != nil {\n\t\treturn e.NoContent(http.StatusNotFound)\n\t}\n\treturn e.JSON(http.StatusOK, a)\n}" ]
[ "0.7893181", "0.74823755", "0.7339581", "0.7297483", "0.72763216", "0.7272475", "0.7272475", "0.69063103", "0.6845236", "0.66029674", "0.65769786", "0.65024614", "0.64386296", "0.6414358", "0.6391792", "0.6340095", "0.6315212", "0.6247335", "0.6185607", "0.6183421", "0.59682983", "0.5962859", "0.5954854", "0.58791834", "0.5821252", "0.57587725", "0.57391757", "0.56399727", "0.5584261", "0.5572324", "0.55670565", "0.556021", "0.55551267", "0.5524279", "0.5518725", "0.5509407", "0.5500424", "0.54814965", "0.5458679", "0.54551786", "0.54263574", "0.5405864", "0.5349865", "0.5343744", "0.5339502", "0.5320658", "0.5294172", "0.52787614", "0.5153558", "0.5116916", "0.51162565", "0.50997496", "0.509751", "0.5094397", "0.5077372", "0.50742286", "0.5074123", "0.50657314", "0.5064425", "0.50585634", "0.5052486", "0.5046168", "0.5044865", "0.50355697", "0.50239825", "0.5019011", "0.5007643", "0.50043386", "0.4999658", "0.4995818", "0.49912542", "0.49874738", "0.49789968", "0.49760577", "0.49702895", "0.49555913", "0.4930433", "0.49235058", "0.4909607", "0.49065572", "0.4905514", "0.4903283", "0.489858", "0.4896308", "0.48950276", "0.48933083", "0.48864147", "0.48842207", "0.4864412", "0.4862642", "0.485947", "0.4852509", "0.4847189", "0.4846792", "0.48429233", "0.4840798", "0.48330817", "0.47945005", "0.47914016", "0.4790546" ]
0.7416126
2
Sends a delete tenant task to client based on the cli.Context Returns an error if one occurred
func deleteTenant(c *cli.Context) error { err := checkArgCount(c, 1) if err != nil { return err } id := c.Args().First() client.Photonclient, err = client.GetClient(c) if err != nil { return err } deleteTask, err := client.Photonclient.Tenants.Delete(id) if err != nil { return err } _, err = waitOnTaskOperation(deleteTask.ID, c) if err != nil { return err } err = clearConfigTenant(id) if err != nil { return err } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func TenantDelete(tenantID string) {\n\tRunCmd(fmt.Sprintf(\"%s tenant -op=del -id=%s\", ActlPath, tenantID))\n}", "func (c *RestClient) DeleteTenant(tenantCluster string) (*models.TcaTask, error) {\n\n\tglog.Infof(\"Deleting tenant cluster %v\", tenantCluster)\n\n\tc.GetClient()\n\tresp, err := c.Client.R().Delete(c.BaseURL + fmt.Sprintf(TcaDeleteTenant, tenantCluster))\n\tif err != nil {\n\t\tglog.Error(err)\n\t\treturn nil, err\n\t}\n\n\tif c.isTrace && resp != nil {\n\t\tfmt.Println(string(resp.Body()))\n\t}\n\n\tif !resp.IsSuccess() {\n\t\treturn nil, c.checkErrors(resp)\n\t}\n\n\tvar task models.TcaTask\n\tif err := json.Unmarshal(resp.Body(), &task); err != nil {\n\t\tglog.Error(\"Failed parse server respond.\")\n\t\treturn nil, err\n\t}\n\n\treturn &task, nil\n}", "func (api *tenantAPI) Delete(obj *cluster.Tenant) error {\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = apicl.ClusterV1().Tenant().Delete(context.Background(), &obj.ObjectMeta)\n\t\treturn err\n\t}\n\n\tapi.ct.handleTenantEvent(&kvstore.WatchEvent{Object: obj, Type: kvstore.Deleted})\n\treturn nil\n}", "func (api *tenantAPI) SyncDelete(obj *cluster.Tenant) error {\n\tvar writeErr error\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t_, writeErr = apicl.ClusterV1().Tenant().Delete(context.Background(), &obj.ObjectMeta)\n\t}\n\n\tif writeErr == nil {\n\t\tapi.ct.handleTenantEvent(&kvstore.WatchEvent{Object: obj, Type: kvstore.Deleted})\n\t}\n\n\treturn writeErr\n}", "func DeleteTask(w http.ResponseWriter, r *http.Request, _ httprouter.Params) {\n\tfmt.Fprint(w, \"DeleteTask\\n\")\n}", "func (t *TaskService) Delete(w http.ResponseWriter, r *http.Request) {\n\tvar (\n\t\temptyUUID gocql.UUID\n\t\ttaskIDStr = mux.Vars(r)[\"taskID\"]\n\t\tpartnerID = mux.Vars(r)[\"partnerID\"]\n\t\tctx = r.Context()\n\t\tcurrentUser = t.userService.GetUser(r, t.httpClient)\n\t)\n\n\ttaskID, err := gocql.ParseUUID(taskIDStr)\n\tif err != nil || taskID == emptyUUID {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorTaskIDHasBadFormat, \"TaskService.Delete: task ID(UUID=%s) has bad format or empty. err=%v\", taskIDStr, err)\n\t\tcommon.SendBadRequest(w, r, errorcode.ErrorTaskIDHasBadFormat)\n\t\treturn\n\t}\n\n\tinternalTasks, err := t.taskPersistence.GetByIDs(ctx, nil, partnerID, false, taskID)\n\tif err != nil {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorCantGetTaskByTaskID, \"TaskService.Delete: can't get internal tasks by task ID %v. err=%v\", taskID, err)\n\t\tcommon.SendInternalServerError(w, r, errorcode.ErrorCantGetTaskByTaskID)\n\t\treturn\n\t}\n\n\tif len(internalTasks) == 0 {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorTaskIsNotFoundByTaskID, \"TaskService.Delete: task with ID %v not found.\", taskID)\n\t\tcommon.SendBadRequest(w, r, errorcode.ErrorTaskIsNotFoundByTaskID)\n\t\treturn\n\t}\n\n\tcommonTaskData := internalTasks[0]\n\tif currentUser.HasNOCAccess() != commonTaskData.IsRequireNOCAccess {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorAccessDenied, \"TaskService.Delete: current user %s is not authorized to delete task with ID %v for partnerID %v\", currentUser.UID(), commonTaskData.ID, commonTaskData.PartnerID)\n\t\tcommon.SendForbidden(w, r, errorcode.ErrorAccessDenied)\n\t\treturn\n\t}\n\n\tdto, err := t.getDataToDelete(ctx, taskID, r, w, partnerID)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tdto.tasks = internalTasks\n\tif err = t.executeDeleting(dto); err != nil {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorCantDeleteTask, \"TaskService.Delete: can't process deleting of the task. err=%v\", err)\n\t\tcommon.SendInternalServerError(w, r, errorcode.ErrorCantDeleteTask)\n\t\treturn\n\t}\n\n\tif !currentUser.HasNOCAccess() {\n\t\t// update counters for tasks in separate goroutine\n\t\tgo func(ctx context.Context, iTasks []models.Task) {\n\t\t\tcounters := getCountersForInternalTasks(iTasks)\n\t\t\tif len(counters) == 0 {\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\terr := t.taskCounterRepo.DecreaseCounter(commonTaskData.PartnerID, counters, false)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Log.ErrfCtx(ctx, errorcode.ErrorCantProcessData, \"Delete: error while trying to increase counter: \", err)\n\t\t\t}\n\t\t}(ctx, internalTasks)\n\t}\n\n\tlogger.Log.InfofCtx(r.Context(), \"TaskService.Delete: successfully deleted task with ID = %v\", taskID)\n\tcommon.SendNoContent(w)\n}", "func (r *DeviceAppManagementTaskRequest) Delete(ctx context.Context) error {\n\treturn r.JSONRequest(ctx, \"DELETE\", \"\", nil, nil)\n}", "func DeleteTask(c *gin.Context) {\n\tfmt.Println(\"deleteTask\")\n\ttask := c.Param(\"id\")\n\tfmt.Println(\"task_id: \", task)\n\tdeleteOneTask(task)\n\tc.JSON(http.StatusOK, task)\n\t// json.NewEncoder(w).Encode(\"Task not found\")\n\n}", "func (dtm *DfgetTaskManager) Delete(ctx context.Context, clientID, taskID string) error {\n\tkey, err := generateKey(clientID, taskID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn dtm.dfgetTaskStore.Delete(key)\n}", "func (oo *OmciCC) SendDeleteTD(ctx context.Context, timeout int, highPrio bool,\n\trxChan chan Message, aInstID uint16) (*me.ManagedEntity, error) {\n\ttid := oo.GetNextTid(highPrio)\n\tlogger.Debugw(ctx, \"send TD-Delete-msg:\", log.Fields{\"device-id\": oo.deviceID,\n\t\t\"SequNo\": strconv.FormatInt(int64(tid), 16),\n\t\t\"InstId\": strconv.FormatInt(int64(aInstID), 16)})\n\n\tmeParams := me.ParamData{EntityID: aInstID}\n\tmeInstance, omciErr := me.NewTrafficDescriptor(meParams)\n\tif omciErr.GetError() == nil {\n\t\tomciLayer, msgLayer, err := oframe.EncodeFrame(meInstance, omci.DeleteRequestType, oframe.TransactionID(tid))\n\t\tif err != nil {\n\t\t\tlogger.Errorw(ctx, \"Cannot encode TD for delete\", log.Fields{\"Err\": err, \"device-id\": oo.deviceID})\n\t\t\treturn nil, err\n\t\t}\n\t\tpkt, err := SerializeOmciLayer(ctx, omciLayer, msgLayer)\n\t\tif err != nil {\n\t\t\tlogger.Errorw(ctx, \"Cannot serialize TD delete\", log.Fields{\"Err\": err, \"device-id\": oo.deviceID})\n\t\t\treturn nil, err\n\t\t}\n\t\tomciRxCallbackPair := CallbackPair{\n\t\t\tCbKey: tid,\n\t\t\tCbEntry: CallbackPairEntry{rxChan, oo.receiveOmciResponse, true},\n\t\t}\n\t\terr = oo.Send(ctx, pkt, timeout, CDefaultRetries, highPrio, omciRxCallbackPair)\n\t\tif err != nil {\n\t\t\tlogger.Errorw(ctx, \"Cannot send TD delete\", log.Fields{\"Err\": err, \"device-id\": oo.deviceID})\n\t\t\treturn nil, err\n\t\t}\n\t\tlogger.Debug(ctx, \"send TD-Delete-msg done\")\n\t\treturn meInstance, nil\n\t}\n\tlogger.Errorw(ctx, \"Cannot generate TD Instance\", log.Fields{\"Err\": omciErr.GetError(), \"device-id\": oo.deviceID})\n\treturn nil, omciErr.GetError()\n\n}", "func (c *DeleteCommand) Exec(_ io.Reader, out io.Writer) error {\n\tinput := c.constructInput()\n\n\terr := c.Globals.APIClient.DeleteTLSSubscription(input)\n\tif err != nil {\n\t\tc.Globals.ErrLog.AddWithContext(err, map[string]any{\n\t\t\t\"TLS Subscription ID\": c.id,\n\t\t\t\"Force\": c.force.Value,\n\t\t})\n\t\treturn err\n\t}\n\n\ttext.Success(out, \"Deleted TLS Subscription '%s' (force: %t)\", c.id, c.force.Value)\n\treturn nil\n}", "func (n *NodeClient) Delete(twin, deployment uint32) (err error) {\n\turl := n.url(\"deployment\", fmt.Sprint(twin), fmt.Sprint(deployment))\n\n\trequest, err := http.NewRequest(http.MethodDelete, url, nil)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to build request\")\n\t}\n\n\tif err := n.client.authorize(request); err != nil {\n\t\treturn errors.Wrap(err, \"failed to sign request\")\n\t}\n\n\tresponse, err := http.DefaultClient.Do(request)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err := n.response(response, nil, http.StatusAccepted); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (suite *TenantTestSuite) TestDeleteTenant() {\n\n\trequest, _ := http.NewRequest(\"DELETE\", \"/api/v2/admin/tenants/6ac7d684-1f8e-4a02-a502-720e8f11e50b\", strings.NewReader(\"\"))\n\trequest.Header.Set(\"x-api-key\", suite.clientkey)\n\trequest.Header.Set(\"Accept\", \"application/json\")\n\tresponse := httptest.NewRecorder()\n\n\tsuite.router.ServeHTTP(response, request)\n\n\tcode := response.Code\n\toutput := response.Body.String()\n\n\tmetricProfileJSON := `{\n \"status\": {\n \"message\": \"Tenant Successfully Deleted\",\n \"code\": \"200\"\n }\n}`\n\t// Check that we must have a 200 ok code\n\tsuite.Equal(200, code, \"Internal Server Error\")\n\t// Compare the expected and actual json response\n\tsuite.Equal(metricProfileJSON, output, \"Response body mismatch\")\n\n\t// check that the element has actually been Deleted\n\t// connect to mongodb\n\tsession, err := mgo.Dial(suite.cfg.MongoDB.Host)\n\tdefer session.Close()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\t// try to retrieve item\n\tvar result map[string]interface{}\n\tc := session.DB(suite.cfg.MongoDB.Db).C(\"tenants\")\n\terr = c.Find(bson.M{\"id\": \"6ac7d684-1f8e-4a02-a502-720e8f11e50b\"}).One(&result)\n\n\tsuite.NotEqual(err, nil, \"No not found error\")\n\tsuite.Equal(err.Error(), \"not found\", \"No not found error\")\n}", "func (m *TenantStatusRequestBuilder) Delete(ctx context.Context, requestConfiguration *TenantStatusRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func (db *MySQLDB) DeleteTenant(ctx context.Context, tenant *Tenant) error {\n\tfLog := mysqlLog.WithField(\"func\", \"DeleteTenant\").WithField(\"RequestID\", ctx.Value(constants.RequestID))\n\tq := \"DELETE FROM HANSIP_TENANT WHERE REC_ID=?\"\n\t_, err := db.instance.ExecContext(ctx, q, tenant.RecID)\n\tif err != nil {\n\t\tfLog.Errorf(\"db.instance.ExecContext got %s. SQL = %s\", err.Error(), q)\n\t\treturn &ErrDBExecuteError{\n\t\t\tWrapped: err,\n\t\t\tMessage: \"Error DeleteTenant\",\n\t\t\tSQL: q,\n\t\t}\n\t}\n\n\tdomainToDelete := tenant.Domain\n\n\t// delete all user-roles ...\n\tq = \"DELETE FROM HANSIP_USER_ROLE WHERE HANSIP_USER_ROLE.ROLE_REC_ID = HANSIP_ROLE.REC_ID AND HANSIP_ROLE.ROLE_DOMAIN = ?\"\n\t_, err = db.instance.ExecContext(ctx, q, domainToDelete)\n\tif err != nil {\n\t\tfLog.Errorf(\"db.instance.ExecContext got %s. SQL = %s\", err.Error(), q)\n\t\treturn &ErrDBExecuteError{\n\t\t\tWrapped: err,\n\t\t\tMessage: \"Error DeleteTenant\",\n\t\t\tSQL: q,\n\t\t}\n\t}\n\n\t// delete all group-roles ...\n\tq = \"DELETE FROM HANSIP_GROUP_ROLE WHERE HANSIP_GROUP_ROLE.GROUP_REC_ID = HANSIP_GROUP.REC_ID AND HANSIP_GROUP.GROUP_DOMAIN = ?\"\n\t_, err = db.instance.ExecContext(ctx, q, domainToDelete)\n\tif err != nil {\n\t\tfLog.Errorf(\"db.instance.ExecContext got %s. SQL = %s\", err.Error(), q)\n\t\treturn &ErrDBExecuteError{\n\t\t\tWrapped: err,\n\t\t\tMessage: \"Error DeleteTenant\",\n\t\t\tSQL: q,\n\t\t}\n\t}\n\n\t// delete all user-groups ...\n\tq = \"DELETE FROM HANSIP_USER_GROUP WHERE HANSIP_USER_GROUP.GROUP_REC_ID = HANSIP_GROUP.REC_ID AND HANSIP_GROUP.GROUP_DOMAIN = ?\"\n\t_, err = db.instance.ExecContext(ctx, q, domainToDelete)\n\tif err != nil {\n\t\tfLog.Errorf(\"db.instance.ExecContext got %s. SQL = %s\", err.Error(), q)\n\t\treturn &ErrDBExecuteError{\n\t\t\tWrapped: err,\n\t\t\tMessage: \"Error DeleteTenant\",\n\t\t\tSQL: q,\n\t\t}\n\t}\n\n\t// delete all groups ...\n\tq = \"DELETE FROM HANSIP_GROUP WHERE HANSIP_GROUP.GROUP_DOMAIN = ?\"\n\t_, err = db.instance.ExecContext(ctx, q, domainToDelete)\n\tif err != nil {\n\t\tfLog.Errorf(\"db.instance.ExecContext got %s. SQL = %s\", err.Error(), q)\n\t\treturn &ErrDBExecuteError{\n\t\t\tWrapped: err,\n\t\t\tMessage: \"Error DeleteTenant\",\n\t\t\tSQL: q,\n\t\t}\n\t}\n\n\t// delete all roles ...\n\tq = \"DELETE FROM HANSIP_ROLE WHERE HANSIP_ROLE.ROLE_DOMAIN = ?\"\n\t_, err = db.instance.ExecContext(ctx, q, domainToDelete)\n\tif err != nil {\n\t\tfLog.Errorf(\"db.instance.ExecContext got %s. SQL = %s\", err.Error(), q)\n\t\treturn &ErrDBExecuteError{\n\t\t\tWrapped: err,\n\t\t\tMessage: \"Error DeleteTenant\",\n\t\t\tSQL: q,\n\t\t}\n\t}\n\n\treturn err\n}", "func (ctrler CtrlDefReactor) OnTenantDelete(obj *Tenant) error {\n\tlog.Info(\"OnTenantDelete is not implemented\")\n\treturn nil\n}", "func (c *TenantController) Clean(ctx *app.CleanTenantContext) error {\n\tuserToken := goajwt.ContextJWT(ctx)\n\tif userToken == nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewUnauthorizedError(\"Missing JWT token\"))\n\t}\n\tttoken := &auth.TenantToken{Token: userToken}\n\n\t// fetch the cluster the user belongs to\n\tuser, err := c.authClientService.GetUser(ctx)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, err)\n\t}\n\n\ttenant, err := c.tenantService.GetTenant(ttoken.Subject())\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewNotFoundError(\"tenants\", ttoken.Subject().String()))\n\t}\n\n\t// restrict deprovision from cluster to internal users only\n\tremoveFromCluster := false\n\tif user.UserData.FeatureLevel != nil && *user.UserData.FeatureLevel == \"internal\" {\n\t\tremoveFromCluster = ctx.Remove\n\t}\n\n\tcluster, err := c.clusterService.GetCluster(ctx, *user.UserData.Cluster)\n\tif err != nil {\n\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\"err\": err,\n\t\t\t\"cluster_url\": *user.UserData.Cluster,\n\t\t}, \"unable to fetch cluster\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewInternalError(ctx, err))\n\t}\n\n\t// create openshift config\n\topenshiftConfig := openshift.NewConfig(c.config, user.UserData, cluster.User, cluster.Token, cluster.APIURL)\n\n\tnsBaseName := tenant.NsBaseName\n\tif nsBaseName == \"\" {\n\t\tnsBaseName = env.RetrieveUserName(user.OpenShiftUsername)\n\t}\n\n\terr = openshift.CleanTenant(ctx, openshiftConfig, user.OpenShiftUsername, nsBaseName, removeFromCluster)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, err)\n\t}\n\tif removeFromCluster {\n\t\terr = c.tenantService.DeleteAll(ttoken.Subject())\n\t\tif err != nil {\n\t\t\treturn jsonapi.JSONErrorResponse(ctx, err)\n\t\t}\n\t}\n\treturn ctx.NoContent()\n}", "func (a *Client) DeleteTenant(params *DeleteTenantParams) (*DeleteTenantCreated, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewDeleteTenantParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"deleteTenant\",\n\t\tMethod: \"DELETE\",\n\t\tPathPattern: \"/tenants/{id}\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\"},\n\t\tParams: params,\n\t\tReader: &DeleteTenantReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*DeleteTenantCreated)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\t// safeguard: normally, absent a default response, unknown success responses return an error above: so this is a codegen issue\n\tmsg := fmt.Sprintf(\"unexpected success response for deleteTenant: API contract not enforced by server. Client expected to get an error, but got: %T\", result)\n\tpanic(msg)\n}", "func (o *Tenant) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"dbmodel: no Tenant provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), tenantPrimaryKeyMapping)\n\tsql := \"DELETE FROM `tenants` WHERE `id`=?\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dbmodel: unable to delete from tenants\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dbmodel: failed to get rows affected by delete for tenants\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn rowsAff, nil\n}", "func (cmd *DeleteTechHyTechCommand) Run(c *client.Client, args []string) error {\n\tvar path string\n\tif len(args) > 0 {\n\t\tpath = args[0]\n\t} else {\n\t\tpath = fmt.Sprintf(\"/api/tech/%v\", cmd.TechID)\n\t}\n\tlogger := goa.NewLogger(log.New(os.Stderr, \"\", log.LstdFlags))\n\tctx := goa.WithLogger(context.Background(), logger)\n\tresp, err := c.DeleteTechHyTech(ctx, path)\n\tif err != nil {\n\t\tgoa.LogError(ctx, \"failed\", \"err\", err)\n\t\treturn err\n\t}\n\n\tgoaclient.HandleResponse(c.Client, resp, cmd.PrettyPrint)\n\treturn nil\n}", "func taskDeleteHandler(w http.ResponseWriter, r *http.Request) {\r\n\r\n\tdn := getRequestParam(r, \"model\")\r\n\ttn := getRequestParam(r, \"task\")\r\n\r\n\t// delete modeling task\r\n\tok, err := theCatalog.DeleteTask(dn, tn)\r\n\tif err != nil {\r\n\t\thttp.Error(w, \"Task delete failed \"+dn+\": \"+tn, http.StatusBadRequest)\r\n\t\treturn\r\n\t}\r\n\tif ok {\r\n\t\tw.Header().Set(\"Content-Location\", \"/api/model/\"+dn+\"/task/\"+tn)\r\n\t\tw.Header().Set(\"Content-Type\", \"text/plain\")\r\n\t}\r\n}", "func (c *TenantController) Clean(ctx *app.CleanTenantContext) error {\n\tuserToken := goajwt.ContextJWT(ctx)\n\tif userToken == nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewUnauthorizedError(\"Missing JWT token\"))\n\t}\n\tttoken := &TenantToken{token: userToken}\n\n\t// fetch the cluster the user belongs to\n\tuser, err := c.userService.GetUser(ctx, ttoken.Subject())\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, err)\n\t}\n\n\t// restrict deprovision from cluster to internal users only\n\tremoveFromCluster := false\n\tif user.FeatureLevel != nil && *user.FeatureLevel == \"internal\" {\n\t\tremoveFromCluster = ctx.Remove\n\t}\n\n\t// fetch the users cluster token\n\topenshiftUsername, _, err := c.resolveTenant(ctx, *user.Cluster, userToken.Raw)\n\tif err != nil {\n\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\"err\": err,\n\t\t\t\"cluster_url\": *user.Cluster,\n\t\t}, \"unable to fetch tenant token from auth\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewUnauthorizedError(\"Could not resolve user token\"))\n\t}\n\n\tcluster, err := c.resolveCluster(ctx, *user.Cluster)\n\tif err != nil {\n\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\"err\": err,\n\t\t\t\"cluster_url\": *user.Cluster,\n\t\t}, \"unable to fetch cluster\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewInternalError(ctx, err))\n\t}\n\n\t// create openshift config\n\topenshiftConfig := openshift.NewConfig(c.defaultOpenshiftConfig, user, cluster.User, cluster.Token, cluster.APIURL)\n\n\terr = openshift.CleanTenant(ctx, openshiftConfig, openshiftUsername, c.templateVars, removeFromCluster)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, err)\n\t}\n\tif removeFromCluster {\n\t\terr = c.tenantService.DeleteAll(ttoken.Subject())\n\t\tif err != nil {\n\t\t\treturn jsonapi.JSONErrorResponse(ctx, err)\n\t\t}\n\t}\n\treturn ctx.NoContent()\n}", "func (m *ManagedTenantsManagedTenantTicketingEndpointsManagedTenantTicketingEndpointItemRequestBuilder) Delete(ctx context.Context, requestConfiguration *ManagedTenantsManagedTenantTicketingEndpointsManagedTenantTicketingEndpointItemRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func TestTenantAddDelete(t *testing.T) {\n\t// tenant params\n\ttenant := client.Tenant{\n\t\tTenantName: \"tenant1\",\n\t}\n\n\t// create a tenant\n\terr := contivClient.TenantPost(&tenant)\n\tcheckError(t, \"create tenant\", err)\n\n\t// Get the tenant and verify it exists\n\tgotTenant, err := contivClient.TenantGet(\"tenant1\")\n\tcheckError(t, \"get tenant\", err)\n\n\tif gotTenant.TenantName != tenant.TenantName {\n\t\tt.Fatalf(\"Got invalid tenant name. expecting %s. Got %s\", tenant.TenantName, gotTenant.TenantName)\n\t}\n\n\t// delete tenant\n\terr = contivClient.TenantDelete(\"tenant1\")\n\tcheckError(t, \"delete tenant\", err)\n\n\t// find again and make sure its gone\n\t_, err = contivClient.TenantGet(\"tenant1\")\n\tif err == nil {\n\t\tt.Fatalf(\"Tenant was not deleted\")\n\t}\n}", "func DeleteTask(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.Header().Set(\"Access-Control-Allow-Origin\", \"*\")\n\tw.Header().Set(\"Access-Control-Allow-Methods\", \"DELETE\")\n\tw.Header().Set(\"Access-Control-Allow-Headers\", \"Content-Type\")\n\tparams := mux.Vars(r)\n\tdeleteOneTask(params[\"id\"])\n\tjson.NewEncoder(w).Encode(params[\"id\"])\n}", "func deleteClusterTask(ctx context.Context, t *testing.T, c *clients, name string) {\n\tt.Logf(\"Deleting clustertask %s\", name)\n\tif err := c.ClusterTaskClient.Delete(ctx, name, metav1.DeleteOptions{}); err != nil {\n\t\tt.Fatalf(\"Failed to delete clustertask: %v\", err)\n\t}\n}", "func (*TenantDeleteRequest) Descriptor() ([]byte, []int) {\n\treturn file_v1_tenant_proto_rawDescGZIP(), []int{3}\n}", "func DeleteTask(w http.ResponseWriter, r *http.Request) {\n\tparams := mux.Vars(r)\n\n\tif params[\"id\"] == \"\" {\n\t\thttp.Error(w, http.StatusText(400), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\ttaskID := bson.ObjectIdHex(params[\"id\"])\n\n\tdeleted, err := repository.DeleteTask(taskID)\n\n\tif err != nil {\n\t\thttp.Error(w, http.StatusText(500), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tif !deleted {\n\t\thttp.Error(w, http.StatusText(500), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tw.WriteHeader(http.StatusOK)\n}", "func (client DeploymentsClient) DeleteSender(req *http.Request) (future DeploymentsDeleteFuture, err error) {\n\tvar resp *http.Response\n\tfuture.FutureAPI = &azure.Future{}\n\tresp, err = client.Send(req, azure.DoRetryWithRegistration(client.Client))\n\tif err != nil {\n\t\treturn\n\t}\n\tvar azf azure.Future\n\tazf, err = azure.NewFutureFromResponse(resp)\n\tfuture.FutureAPI = &azf\n\tfuture.Result = future.result\n\treturn\n}", "func DeleteTask(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Context-Type\", \"application/x-www-form-urlencoded\")\n\tw.Header().Set(\"Access-Control-Allow-Origin\", \"*\")\n\tw.Header().Set(\"Access-Control-Allow-Methods\", \"DELETE\")\n\tw.Header().Set(\"Access-Control-Allow-Headers\", \"Content-Type\")\n\tparams := mux.Vars(r)\n\tdeleteOneTask(params[\"id\"])\n\tjson.NewEncoder(w).Encode(params[\"id\"])\n\t// json.NewEncoder(w).Encode(\"Task not found\")\n\n}", "func (m *ItemTodoListsItemTasksTodoTaskItemRequestBuilder) Delete(ctx context.Context, requestConfiguration *ItemTodoListsItemTasksTodoTaskItemRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": ia572726a95efa92ddd544552cd950653dc691023836923576b2f4bf716cf204a.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": ia572726a95efa92ddd544552cd950653dc691023836923576b2f4bf716cf204a.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func (r DeleteDeploymentRequest) Send(ctx context.Context) (*DeleteDeploymentOutput, error) {\n\tr.Request.SetContext(ctx)\n\terr := r.Request.Send()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn r.Request.Data.(*DeleteDeploymentOutput), nil\n}", "func (client AccountClient) DeleteSender(req *http.Request) (future AccountDeleteFuture, err error) {\n var resp *http.Response\n resp, err = client.Send(req, azure.DoRetryWithRegistration(client.Client))\n if err != nil {\n return\n }\n future.Future, err = azure.NewFutureFromResponse(resp)\n return\n }", "func (h DeviceController) Delete(c *gin.Context) {\n\ttenantID := c.Param(\"tenantId\")\n\tdeviceID := c.Param(\"deviceId\")\n\n\tctx := c.Request.Context()\n\tif err := h.app.DeleteDevice(ctx, tenantID, deviceID); err != nil {\n\t\tc.JSON(http.StatusInternalServerError, gin.H{\n\t\t\t\"error\": errors.Wrap(err, \"error deleting the device\").Error(),\n\t\t})\n\t\treturn\n\t}\n\n\tc.Writer.WriteHeader(http.StatusAccepted)\n}", "func (ts *TaskService) Delete(ctx context.Context, req *taskAPI.DeleteRequest) (*taskAPI.DeleteResponse, error) {\n\tdefer logPanicAndDie(log.G(ctx))\n\n\tlog.G(ctx).WithFields(logrus.Fields{\"id\": req.ID, \"exec_id\": req.ExecID}).Debug(\"delete\")\n\ttask, err := ts.taskManager.Task(req.ID)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tctx = namespaces.WithNamespace(ctx, defaultNamespace)\n\tresp, err := task.Delete(ctx, req)\n\tif err != nil {\n\t\tlog.G(ctx).WithError(err).Error(\"delete failed\")\n\t\treturn nil, err\n\t}\n\n\tlog.G(ctx).WithFields(logrus.Fields{\n\t\t\"pid\": resp.Pid,\n\t\t\"exit_status\": resp.ExitStatus,\n\t}).Debug(\"delete succeeded\")\n\treturn resp, nil\n}", "func (gt GtwyMgr) Delete(ctx context.Context, appcontext, remoteAddress string) error {\n\tif EnvDebugOn {\n\t\tlblog.LogEvent(\"GtwyMgr\", \"Delete\", \"info\", \"start\")\n\t}\n\n\t//check the approval list\n\tq := datastore.NewQuery(gt.bc.GetConfigValue(ctx, \"EnvGtwayDsKind\")).\n\t\tNamespace(gt.bc.GetConfigValue(ctx, \"EnvGtwayDsNamespace\")).\n\t\tFilter(\"appcontext =\", appcontext).\n\t\tFilter(\"remoteaddress =\", remoteAddress).\n\t\tKeysOnly()\n\n\tvar arr []Gateway\n\tkeys, err := gt.ds.GetAll(ctx, q, &arr)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttx, err := gt.ds.NewTransaction(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err := tx.DeleteMulti(keys); err != nil {\n\t\ttx.Rollback()\n\t\treturn err\n\t}\n\n\tif _, err = tx.Commit(); err != nil {\n\t\treturn err\n\t}\n\n\tif EnvDebugOn {\n\t\tlblog.LogEvent(\"GtwyMgr\", \"Delete\", \"info\", \"end\")\n\t}\n\treturn nil\n}", "func (m *TenantAttachRBACRequestBuilder) Delete(ctx context.Context, requestConfiguration *TenantAttachRBACRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func Delete(r *http.Request, cfg config.Config) (int, http.Header, []byte, error) {\n\n\t//STANDARD DECLARATIONS START\n\n\tcode := http.StatusOK\n\th := http.Header{}\n\toutput := []byte(\"\")\n\terr := error(nil)\n\tcharset := \"utf-8\"\n\n\t//STANDARD DECLARATIONS END\n\n\t// Set Content-Type response Header value\n\tcontentType := r.Header.Get(\"Accept\")\n\th.Set(\"Content-Type\", fmt.Sprintf(\"%s; charset=%s\", contentType, charset))\n\n\tvars := mux.Vars(r)\n\n\t// Try to open the mongo session\n\tsession, err := mongo.OpenSession(cfg.MongoDB)\n\tdefer session.Close()\n\n\tif err != nil {\n\t\tcode = http.StatusInternalServerError\n\t\treturn code, h, output, err\n\t}\n\n\tfilter := bson.M{\"id\": vars[\"ID\"]}\n\n\t// Retrieve Results from database\n\tresults := []Tenant{}\n\terr = mongo.Find(session, cfg.MongoDB.Db, \"tenants\", filter, \"name\", &results)\n\n\tif err != nil {\n\t\tcode = http.StatusInternalServerError\n\t\treturn code, h, output, err\n\t}\n\n\t// Check if nothing found\n\tif len(results) < 1 {\n\t\toutput, _ = respond.MarshalContent(respond.NotFound, contentType, \"\", \" \")\n\t\tcode = http.StatusNotFound\n\t\treturn code, h, output, err\n\t}\n\n\tmongo.Remove(session, cfg.MongoDB.Db, \"tenants\", filter)\n\n\tif err != nil {\n\t\tcode = http.StatusInternalServerError\n\t\treturn code, h, output, err\n\t}\n\n\t// Create view of the results\n\toutput, err = createMsgView(\"Tenant Successfully Deleted\", 200) //Render the results into JSON\n\n\tif err != nil {\n\t\tcode = http.StatusInternalServerError\n\t\treturn code, h, output, err\n\t}\n\n\th.Set(\"Content-Type\", fmt.Sprintf(\"%s; charset=%s\", contentType, charset))\n\treturn code, h, output, err\n\n}", "func (client DatasetClient) DeleteSender(req *http.Request) (*http.Response, error) {\n return client.Send(req, autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n }", "func (client DatasetClient) Delete(ctx context.Context, datasetID string) (result autorest.Response, err error) {\n if tracing.IsEnabled() {\n ctx = tracing.StartSpan(ctx, fqdn + \"/DatasetClient.Delete\")\n defer func() {\n sc := -1\n if result.Response != nil {\n sc = result.Response.StatusCode\n }\n tracing.EndSpan(ctx, sc, err)\n }()\n }\n req, err := client.DeletePreparer(ctx, datasetID)\n if err != nil {\n err = autorest.NewErrorWithError(err, \"creator.DatasetClient\", \"Delete\", nil , \"Failure preparing request\")\n return\n }\n\n resp, err := client.DeleteSender(req)\n if err != nil {\n result.Response = resp\n err = autorest.NewErrorWithError(err, \"creator.DatasetClient\", \"Delete\", resp, \"Failure sending request\")\n return\n }\n\n result, err = client.DeleteResponder(resp)\n if err != nil {\n err = autorest.NewErrorWithError(err, \"creator.DatasetClient\", \"Delete\", resp, \"Failure responding to request\")\n return\n }\n\n return\n}", "func deleteTask(msg BackendPayload) {\n\tincomingTaskID := msg.ID // attempt to retreive the id of the task in the case its an update POST\n\t// _, exists := TodoList[incomingTaskID]\n\t// if exists { // only attempt to delete if it exists\n\t// \tdelete(TodoList, incomingTaskID)\n\t// }\n\tTodoList.Delete(incomingTaskID)\n\treturn\n}", "func (ctrl *TaskController) DeleteTask(w http.ResponseWriter, r *http.Request) {\n\ttaskId := ParamAsString(\"id\", r)\n\tlogrus.Println(\"delete task : \", taskId)\n\n\terr := ctrl.taskDao.Delete(taskId)\n\tif err != nil {\n\t\tlogrus.Error(err)\n\t\tSendJSONError(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tlogrus.Println(\"deleted task : \", taskId)\n\tSendJSONWithHTTPCode(w, nil, http.StatusNoContent)\n}", "func (client StorageTargetsClient) DeleteSender(req *http.Request) (future StorageTargetsDeleteFuture, err error) {\n\tvar resp *http.Response\n\tfuture.FutureAPI = &azure.Future{}\n\tresp, err = client.Send(req, azure.DoRetryWithRegistration(client.Client))\n\tif err != nil {\n\t\treturn\n\t}\n\tvar azf azure.Future\n\tazf, err = azure.NewFutureFromResponse(resp)\n\tfuture.FutureAPI = &azf\n\tfuture.Result = future.result\n\treturn\n}", "func (r *mutationResolver) DeleteTenants(ctx context.Context, in []string) (int, error) {\n\treturn r.tenant.Delete(ctx, in)\n}", "func (oo *OmciCC) SendDeleteVtfd(ctx context.Context, timeout int, highPrio bool,\n\trxChan chan Message, aInstID uint16) (*me.ManagedEntity, error) {\n\ttid := oo.GetNextTid(highPrio)\n\tlogger.Debugw(ctx, \"send VTFD-Delete-msg:\", log.Fields{\"device-id\": oo.deviceID,\n\t\t\"SequNo\": strconv.FormatInt(int64(tid), 16),\n\t\t\"InstId\": strconv.FormatInt(int64(aInstID), 16)})\n\n\tmeParams := me.ParamData{EntityID: aInstID}\n\tmeInstance, omciErr := me.NewVlanTaggingFilterData(meParams)\n\tif omciErr.GetError() == nil {\n\t\tomciLayer, msgLayer, err := oframe.EncodeFrame(meInstance, omci.DeleteRequestType,\n\t\t\toframe.TransactionID(tid))\n\t\tif err != nil {\n\t\t\tlogger.Errorw(ctx, \"Cannot encode VTFD for delete\", log.Fields{\n\t\t\t\t\"Err\": err, \"device-id\": oo.deviceID})\n\t\t\t//TODO!!: refactoring improvement requested, here as an example for [VOL-3457]:\n\t\t\t// return (dual format) error code that can be used at caller for immediate error treatment\n\t\t\t// (relevant to all used sendXX() methods and their error conditions)\n\t\t\treturn nil, err\n\t\t}\n\n\t\tpkt, err := SerializeOmciLayer(ctx, omciLayer, msgLayer)\n\t\tif err != nil {\n\t\t\tlogger.Errorw(ctx, \"Cannot serialize VTFD delete\", log.Fields{\n\t\t\t\t\"Err\": err, \"device-id\": oo.deviceID})\n\t\t\treturn nil, err\n\t\t}\n\n\t\tomciRxCallbackPair := CallbackPair{\n\t\t\tCbKey: tid,\n\t\t\tCbEntry: CallbackPairEntry{rxChan, oo.receiveOmciResponse, true},\n\t\t}\n\t\terr = oo.Send(ctx, pkt, timeout, CDefaultRetries, highPrio, omciRxCallbackPair)\n\t\tif err != nil {\n\t\t\tlogger.Errorw(ctx, \"Cannot send VTFD delete\", log.Fields{\n\t\t\t\t\"Err\": err, \"device-id\": oo.deviceID})\n\t\t\treturn nil, err\n\t\t}\n\t\tlogger.Debug(ctx, \"send VTFD-Delete-msg done\")\n\t\treturn meInstance, nil\n\t}\n\tlogger.Errorw(ctx, \"Cannot generate VTFD Instance for delete\", log.Fields{\n\t\t\"Err\": omciErr.GetError(), \"device-id\": oo.deviceID})\n\treturn nil, omciErr.GetError()\n}", "func (r DeleteTaskSetRequest) Send(ctx context.Context) (*DeleteTaskSetResponse, error) {\n\tr.Request.SetContext(ctx)\n\terr := r.Request.Send()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresp := &DeleteTaskSetResponse{\n\t\tDeleteTaskSetOutput: r.Request.Data.(*DeleteTaskSetOutput),\n\t\tresponse: &aws.Response{Request: r.Request},\n\t}\n\n\treturn resp, nil\n}", "func (client BaseClient) DeleteSystemSender(req *http.Request) (*http.Response, error) {\n\treturn autorest.SendWithSender(client, req,\n\t\tautorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n}", "func (client ViewsClient) DeleteSender(req *http.Request) (*http.Response, error) {\n\treturn client.Send(req, autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n}", "func (o *RevokeDeviceCertificateUsingDELETEParams) SetTenantID(tenantID string) {\n\to.TenantID = tenantID\n}", "func (r *ClusterDeleteRequest) SendContext(ctx context.Context) (result *ClusterDeleteResponse, err error) {\n\tquery := helpers.CopyQuery(r.query)\n\tif r.deprovision != nil {\n\t\thelpers.AddValue(&query, \"deprovision\", *r.deprovision)\n\t}\n\theader := helpers.CopyHeader(r.header)\n\turi := &url.URL{\n\t\tPath: r.path,\n\t\tRawQuery: query.Encode(),\n\t}\n\trequest := &http.Request{\n\t\tMethod: \"DELETE\",\n\t\tURL: uri,\n\t\tHeader: header,\n\t}\n\tif ctx != nil {\n\t\trequest = request.WithContext(ctx)\n\t}\n\tresponse, err := r.transport.RoundTrip(request)\n\tif err != nil {\n\t\treturn\n\t}\n\tdefer response.Body.Close()\n\tresult = &ClusterDeleteResponse{}\n\tresult.status = response.StatusCode\n\tresult.header = response.Header\n\tif result.status >= 400 {\n\t\tresult.err, err = errors.UnmarshalError(response.Body)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t\terr = result.err\n\t\treturn\n\t}\n\treturn\n}", "func ExampleIscsiTargetsClient_BeginDelete() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclientFactory, err := armstoragepool.NewClientFactory(\"<subscription-id>\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tpoller, err := clientFactory.NewIscsiTargetsClient().BeginDelete(ctx, \"myResourceGroup\", \"myDiskPool\", \"myIscsiTarget\", nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\t_, err = poller.PollUntilDone(ctx, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to pull the result: %v\", err)\n\t}\n}", "func DeleteTenant(id int) (err error) {\n\to := orm.NewOrm()\n\tv := Tenant{Id: id}\n\t// ascertain id exists in the database\n\tif err = o.Read(&v); err == nil {\n\t\tvar num int64\n\t\tif num, err = o.Delete(&Tenant{Id: id}); err == nil {\n\t\t\tfmt.Println(\"Number of records deleted in database:\", num)\n\t\t}\n\t}\n\treturn\n}", "func (o *DeleteDeviceUsingDELETEParams) SetTenantID(tenantID string) {\n\to.TenantID = tenantID\n}", "func (access *TaskAccess) DeleteTask(id string) error {\n\treturn access.DB.Where(\"id = ?\", id).First(&models.Task{}).Delete(&models.Task{}).Error\n}", "func (t *TargetToExtent) Delete(server *Server) (*http.Response, error) {\n\tendpoint := fmt.Sprintf(\"/api/v1.0/services/iscsi/targettoextent/%d/\", t.ID)\n\tresp, err := server.getSlingConnection().Delete(endpoint).Receive(nil, nil)\n\tif err != nil {\n\t\tglog.Warningln(err)\n\t}\n\treturn resp, err\n}", "func (c Control) ServeDeleteTask(w http.ResponseWriter, r *http.Request) {\n\tid, err := strconv.ParseInt(r.URL.Query().Get(\"id\"), 10, 64)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\tc.Config.Lock()\n\tdefer c.Config.Unlock()\n\tindex, task := c.findTaskById(id)\n\tif task == nil {\n\t\thttp.Error(w, \"Invalid task ID\", http.StatusBadRequest)\n\t\treturn\n\t}\n\ttask.StopLoop()\n\tfor i := index; i < len(c.Config.Tasks)-1; i++ {\n\t\tc.Config.Tasks[i] = c.Config.Tasks[i+1]\n\t}\n\tc.Config.Tasks = c.Config.Tasks[0 : len(c.Config.Tasks)-1]\n\tc.Config.Save()\n\n\thttp.Redirect(w, r, \"/\", http.StatusTemporaryRedirect)\n}", "func (m *ItemOutlookTaskFoldersItemTasksOutlookTaskItemRequestBuilder) Delete(ctx context.Context, requestConfiguration *ItemOutlookTaskFoldersItemTasksOutlookTaskItemRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func (r DeleteHumanTaskUiRequest) Send(ctx context.Context) (*DeleteHumanTaskUiResponse, error) {\n\tr.Request.SetContext(ctx)\n\terr := r.Request.Send()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresp := &DeleteHumanTaskUiResponse{\n\t\tDeleteHumanTaskUiOutput: r.Request.Data.(*DeleteHumanTaskUiOutput),\n\t\tresponse: &aws.Response{Request: r.Request},\n\t}\n\n\treturn resp, nil\n}", "func (p *DeploymentsClientDeleteAtTenantScopePoller) Poll(ctx context.Context) (*http.Response, error) {\n\treturn p.pt.Poll(ctx)\n}", "func DeleteTenant(stateDriver core.StateDriver, tenant *intent.ConfigTenant) error {\n\treturn validateTenantConfig(tenant)\n}", "func DeleteClient(nbmaster string, httpClient *http.Client, jwt string) {\r\n fmt.Printf(\"\\nSending a DELETE request to delete client %s from policy %s...\\n\", testClientName, testPolicyName)\r\n\r\n uri := \"https://\" + nbmaster + \":\" + port + \"/netbackup/\" + policiesUri + testPolicyName + \"/clients/\" + testClientName\r\n\r\n request, _ := http.NewRequest(http.MethodDelete, uri, nil)\r\n request.Header.Add(\"Authorization\", jwt);\r\n\r\n response, err := httpClient.Do(request)\r\n\r\n if err != nil {\r\n fmt.Printf(\"The HTTP request failed with error: %s\\n\", err)\r\n panic(\"Unable to delete client.\\n\")\r\n } else {\r\n if response.StatusCode != 204 {\r\n printErrorResponse(response)\r\n } else {\r\n fmt.Printf(\"%s deleted successfully.\\n\", testClientName);\r\n }\r\n }\r\n}", "func (client JobClient) DeleteSender(req *http.Request) (future JobDeleteFuture, err error) {\n var resp *http.Response\n resp, err = client.Send(req, azure.DoRetryWithRegistration(client.Client))\n if err != nil {\n return\n }\n future.Future, err = azure.NewFutureFromResponse(resp)\n return\n }", "func (c *Client) DeleteTask(ctx context.Context, in *todopb.TaskQuery, opts ...grpc.CallOption) (*empty.Empty, error) {\n\treturn c.client.DeleteTask(ctx, in, opts...)\n}", "func DeleteTaskFunc(w http.ResponseWriter, r *http.Request) {\n\tusername := sessions.GetCurrentUserName(r)\n\tif r.Method != \"GET\" {\n\t\thttp.Redirect(w, r, \"/\", http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tid := r.URL.Path[len(\"/delete/\"):]\n\tif id == \"all\" {\n\t\terr := db.DeleteAll(username)\n\t\tif err != nil {\n\t\t\tmessage = \"Error deleting tasks\"\n\t\t\thttp.Redirect(w, r, \"/\", http.StatusInternalServerError)\n\t\t}\n\t\thttp.Redirect(w, r, \"/\", http.StatusFound)\n\t} else {\n\t\tid, err := strconv.Atoi(id)\n\t\tif err != nil {\n\t\t\tlog.Println(err)\n\t\t\thttp.Redirect(w, r, \"/\", http.StatusBadRequest)\n\t\t} else {\n\t\t\terr = db.DeleteTask(username, id)\n\t\t\tif err != nil {\n\t\t\t\tmessage = \"Error deleting task\"\n\t\t\t} else {\n\t\t\t\tmessage = \"Task deleted\"\n\t\t\t}\n\t\t\thttp.Redirect(w, r, \"/deleted\", http.StatusFound)\n\t\t}\n\t}\n\n}", "func (task SchTask) Delete(taskname string, own, force bool) string {\n\tcmd := &exec.Cmd{}\n\n\tif Debug {\n\t\treturn dbgMessage\n\t}\n\n\tif own {\n\t\ttaskname = task.prefix + taskname\n\t}\n\n\tif !force {\n\t\tcmd = exec.Command(task.bin, _Delete.Command, _Delete.taskname, taskname)\n\t} else {\n\t\tcmd = exec.Command(task.bin, _Delete.Command, _Delete.taskname, taskname, _Delete.force)\n\t}\n\n\toutput, err := cmd.CombinedOutput()\n\tcatch(output, err)\n\n\treturn string(output)\n}", "func HandleDeleteTask(w http.ResponseWriter, r *http.Request) {\n\tlog.Root.Info(\"HandleDeleteTask BEGIN\")\n\n\tif r.Method != http.MethodPost {\n\t\tw.WriteHeader(http.StatusNotFound)\n\t\tHttpResponseError(w, ErrNotFound)\n\t\treturn\n\t}\n\n\tdefer r.Body.Close()\n\tbody, _ := ioutil.ReadAll(r.Body)\n\n\tdata := make(map[string]interface{})\n\terr := json.Unmarshal(body, &data)\n\tif err != nil {\n\t\tlog.Root.Error(\"HandleDeleteTask Parse HTTP request body error\")\n\t\tHttpResponseError(w, ErrForm)\n\t\treturn\n\t}\n\n\telem, ok := data[\"taskID\"]\n\tif !ok {\n\t\tlog.Root.Error(\"HandleDeleteTask HTTP form data error\")\n\t\tHttpResponseError(w, ErrForm)\n\t\treturn\n\t}\n\n\ttaskID := elem.(string)\n\terr = node.DeleteTask(taskID)\n\tif err != nil {\n\t\tlog.Root.Error(\"HandleDeleteTask Delete task error. TaskID: %v\", taskID)\n\t\tHttpResponseError(w, ErrServer)\n\t\treturn\n\t}\n\n\tlog.Root.Info(\"HandleDeleteTask END\")\n\tHttpResponseOk(w)\n\treturn\n}", "func (client DeviceClient) DeleteSender(req *http.Request) (*http.Response, error) {\n\treturn client.Send(req, autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n}", "func DeleteSchedule(nbmaster string, httpClient *http.Client, jwt string) {\r\n fmt.Printf(\"\\nSending a DELETE request to delete schedule %s from policy %s...\\n\", testScheduleName, testPolicyName)\r\n\r\n uri := \"https://\" + nbmaster + \":\" + port + \"/netbackup/\" + policiesUri + testPolicyName + \"/schedules/\" + testScheduleName\r\n\r\n request, _ := http.NewRequest(http.MethodDelete, uri, nil)\r\n request.Header.Add(\"Authorization\", jwt);\r\n\r\n response, err := httpClient.Do(request)\r\n\r\n if err != nil {\r\n fmt.Printf(\"The HTTP request failed with error: %s\\n\", err)\r\n panic(\"Unable to delete schedule.\\n\")\r\n } else {\r\n if response.StatusCode != 204 {\r\n printErrorResponse(response)\r\n } else {\r\n fmt.Printf(\"%s deleted successfully.\\n\", testScheduleName);\r\n }\r\n }\r\n}", "func (s *TaskTypesEndpoint) Delete(ctx context.Context, division int, id *types.GUID) error {\n\tb, _ := s.client.ResolvePathWithDivision(\"/api/v1/{division}/accountancy/TaskTypes\", division) // #nosec\n\tu, err := api.AddOdataKeyToURL(b, id)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, r, requestError := s.client.NewRequestAndDo(ctx, \"DELETE\", u.String(), nil, nil)\n\tif requestError != nil {\n\t\treturn requestError\n\t}\n\n\tif r.StatusCode != http.StatusNoContent {\n\t\tbody, _ := ioutil.ReadAll(r.Body) // #nosec\n\t\treturn fmt.Errorf(\"Failed with status %v and body %v\", r.StatusCode, body)\n\t}\n\n\treturn nil\n}", "func (ts *TaskService) Delete(requestCtx context.Context, req *taskAPI.DeleteRequest) (*taskAPI.DeleteResponse, error) {\n\tdefer logPanicAndDie(log.G(requestCtx))\n\tlog.G(requestCtx).WithFields(logrus.Fields{\"id\": req.ID, \"exec_id\": req.ExecID}).Debug(\"delete\")\n\n\tresp, err := ts.taskManager.DeleteProcess(requestCtx, req, ts.runcService)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlog.G(requestCtx).WithFields(logrus.Fields{\n\t\t\"pid\": resp.Pid,\n\t\t\"exit_status\": resp.ExitStatus,\n\t}).Debug(\"delete succeeded\")\n\treturn resp, nil\n}", "func sendWorkloadDeleted(ninputs *pb.WorkloadInfo) error {\n\tclient := nagent.ClientUds(configuration.NodeAgentManagementAPI)\n\tif client == nil {\n\t\treturn errors.New(\"failed to create Nodeagent client\")\n\t}\n\n\t_, err := client.WorkloadDeleted(ninputs)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tclient.Close()\n\treturn nil\n}", "func deleteTask(w http.ResponseWriter, r *http.Request){\n\t//definimos variable de vars que devuelve las variables de ruta\n\tvars := mux.Vars(r)\n\n\ttaskID, err := strconv.Atoi(vars[\"id\"])\n\tif err != nil{\n\t\tfmt.Fprintf(w, \"Invalid ID\")\n\t\treturn\n\t}\n\n\t//Se elimina la task a la lista, guardando todas las que estan hasta su indice, y la que le sigue en adelante.\n\tfor i, task := range tasks {\n\t\tif task.ID == taskID {\n\t\t\ttasks = append(tasks[:i], tasks[i + 1:] ...)\n\t\t\tfmt.Fprintf(w, \"The task with ID %v has been removed succesfully\", taskID)\n\t\t}\n\t}\n}", "func DeleteTask(w http.ResponseWriter, r *http.Request, repo *tasks.TaskRepository) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tparams := mux.Vars(r)\n\ttaskID, err := strconv.Atoi(params[\"id\"])\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\ttask, err := repo.DeleteTask(taskID)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tjson.NewEncoder(w).Encode(apiIndexTask(task))\n}", "func (client CloudEndpointsClient) DeleteSender(req *http.Request) (future CloudEndpointsDeleteFuture, err error) {\n\tvar resp *http.Response\n\tfuture.FutureAPI = &azure.Future{}\n\tresp, err = client.Send(req, azure.DoRetryWithRegistration(client.Client))\n\tif err != nil {\n\t\treturn\n\t}\n\tvar azf azure.Future\n\tazf, err = azure.NewFutureFromResponse(resp)\n\tfuture.FutureAPI = &azf\n\tfuture.Result = future.result\n\treturn\n}", "func (t Task) Delete() error {\n\tpath := fmt.Sprintf(\"tasks/%d\", t.ID)\n\t_, err := makeRequest(http.MethodDelete, path, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (client RosettaNetProcessConfigurationsClient) DeleteSender(req *http.Request) (*http.Response, error) {\n\treturn client.Send(req, azure.DoRetryWithRegistration(client.Client))\n}", "func Delete(taskName string, repository repository.Repository) string {\n\tif taskName == \"\" {\n\t\treturn \"Empty task nothing is deleted from the repository\"\n\t}\n\n\tdeletetaskInstance := &models.TaskDeleteDomain{\n\t\tTaskName: taskName,\n\t}\n\tresp, err := repository.DeleteTask(*deletetaskInstance)\n\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\treturn resp\n}", "func NewCmdDelete(\n\tf cmdutil.Factory,\n\treport func(io.Writer, *util.OutputFlags, dto.Task) error,\n) *cobra.Command {\n\tof := util.OutputFlags{}\n\tcmd := &cobra.Command{\n\t\tUse: \"delete <task>\",\n\t\tAliases: []string{\"remove\", \"rm\", \"del\"},\n\t\tArgs: cmdutil.RequiredNamedArgs(\"task\"),\n\t\tValidArgsFunction: cmdcompl.CombineSuggestionsToArgs(\n\t\t\tcmdcomplutil.NewTaskAutoComplete(f, false)),\n\t\tShort: \"Deletes a task from a project on Clockify\",\n\t\tLong: heredoc.Doc(`\n\t\t\tDeletes a task from a project on Clockify\n\t\t\tThis action can't be reverted, and all time entries using this task will revert to not having one\n\t\t`),\n\t\tExample: heredoc.Doc(`\n\t\t\t$ clockify-cli task delete -p \"special\" very\n\t\t\t+--------------------------+----------------+--------+\n\t\t\t| ID | NAME | STATUS |\n\t\t\t+--------------------------+----------------+--------+\n\t\t\t| 62aa5d7049445270d7b979d6 | Very Important | ACTIVE |\n\t\t\t+--------------------------+----------------+--------+\n\n\t\t\t$ clockify-cli task delete -p \"special\" 62aa4eed49445270d7b9666c\n\t\t\t+--------------------------+----------+--------+\n\t\t\t| ID | NAME | STATUS |\n\t\t\t+--------------------------+----------+--------+\n\t\t\t| 62aa4eed49445270d7b9666c | Inactive | DONE |\n\t\t\t+--------------------------+----------+--------+\n\n\t\t\t$ clockify-cli task delete -p \"special\" 62aa4eed49445270d7b9666c\n\t\t\tNo task with id or name containing '62aa4eed49445270d7b9666c' was found\n\t\t`),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tproject, _ := cmd.Flags().GetString(\"project\")\n\t\t\ttask := strings.TrimSpace(args[0])\n\t\t\tif project == \"\" || task == \"\" {\n\t\t\t\treturn errors.New(\"project and task id should not be empty\")\n\t\t\t}\n\n\t\t\tw, err := f.GetWorkspaceID()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tc, err := f.Client()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tif f.Config().IsAllowNameForID() {\n\t\t\t\tif project, err = search.GetProjectByName(\n\t\t\t\t\tc, w, project); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\tif task, err = search.GetTaskByName(\n\t\t\t\t\tc,\n\t\t\t\t\tapi.GetTasksParam{Workspace: w, ProjectID: project},\n\t\t\t\t\ttask,\n\t\t\t\t); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tt, err := c.DeleteTask(api.DeleteTaskParam{\n\t\t\t\tWorkspace: w,\n\t\t\t\tProjectID: project,\n\t\t\t\tTaskID: task,\n\t\t\t})\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tif report == nil {\n\t\t\t\treturn util.TaskReport(cmd, of, t)\n\t\t\t}\n\n\t\t\treturn report(cmd.OutOrStdout(), &of, t)\n\t\t},\n\t}\n\n\tcmdutil.AddProjectFlags(cmd, f)\n\tutil.TaskAddReportFlags(cmd, &of)\n\n\treturn cmd\n}", "func (client *VirtualNetworkTapsClient) delete(ctx context.Context, resourceGroupName string, tapName string, options *VirtualNetworkTapsBeginDeleteOptions) (*azcore.Response, error) {\n\treq, err := client.deleteCreateRequest(ctx, resourceGroupName, tapName, options)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresp, err := client.con.Pipeline().Do(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif !resp.HasStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent) {\n\t\treturn nil, client.deleteHandleError(resp)\n\t}\n\treturn resp, nil\n}", "func (w *workerData) actionDelete(task *taskRequest, reply *taskReply) error {\n\tif task.delete == nil {\n\t\treply.err = errors.New(\"missing delete request\")\n\t\treturn errors.New(\"missing delete request\")\n\t}\n\tif w.ldaprw == nil {\n\t\treply.err = errors.New(\"delete request requires ldaprw\")\n\t\treturn errors.New(\"delete request requires ldaprw\")\n\t}\n\tif err := w.ldaprw.Delete(task.delete); err != nil {\n\t\treply.err = err\n\t\treturn err\n\t}\n\treturn nil\n}", "func (oo *OmciCC) SendDeleteEvtocd(ctx context.Context, timeout int, highPrio bool,\n\trxChan chan Message, params ...me.ParamData) (*me.ManagedEntity, error) {\n\ttid := oo.GetNextTid(highPrio)\n\tlogger.Debugw(ctx, \"send EVTOCD-Delete-msg:\", log.Fields{\"device-id\": oo.deviceID,\n\t\t\"SequNo\": strconv.FormatInt(int64(tid), 16),\n\t\t\"InstId\": strconv.FormatInt(int64(params[0].EntityID), 16)})\n\n\tmeInstance, omciErr := me.NewExtendedVlanTaggingOperationConfigurationData(params[0])\n\tif omciErr.GetError() == nil {\n\t\tomciLayer, msgLayer, err := oframe.EncodeFrame(meInstance, omci.DeleteRequestType, oframe.TransactionID(tid))\n\t\tif err != nil {\n\t\t\tlogger.Errorw(ctx, \"Cannot encode EVTOCD for delete\", log.Fields{\n\t\t\t\t\"Err\": err, \"device-id\": oo.deviceID})\n\t\t\treturn nil, err\n\t\t}\n\n\t\tpkt, err := SerializeOmciLayer(ctx, omciLayer, msgLayer)\n\t\tif err != nil {\n\t\t\tlogger.Errorw(ctx, \"Cannot serialize EVTOCD delete\", log.Fields{\n\t\t\t\t\"Err\": err, \"device-id\": oo.deviceID})\n\t\t\treturn nil, err\n\t\t}\n\n\t\tomciRxCallbackPair := CallbackPair{\n\t\t\tCbKey: tid,\n\t\t\tCbEntry: CallbackPairEntry{rxChan, oo.receiveOmciResponse, true},\n\t\t}\n\t\terr = oo.Send(ctx, pkt, timeout, CDefaultRetries, highPrio, omciRxCallbackPair)\n\t\tif err != nil {\n\t\t\tlogger.Errorw(ctx, \"Cannot send EVTOCD delete\", log.Fields{\n\t\t\t\t\"Err\": err, \"device-id\": oo.deviceID})\n\t\t\treturn nil, err\n\t\t}\n\t\tlogger.Debug(ctx, \"send EVTOCD-delete msg done\")\n\t\treturn meInstance, nil\n\t}\n\tlogger.Errorw(ctx, \"Cannot generate EVTOCD Instance\", log.Fields{\n\t\t\"Err\": omciErr.GetError(), \"device-id\": oo.deviceID})\n\treturn nil, omciErr.GetError()\n}", "func (client VersionsClient) DeleteSender(req *http.Request) (*http.Response, error) {\n\treturn client.Send(req, azure.DoRetryWithRegistration(client.Client))\n}", "func resourceNetboxOrgTenantGroupDelete(d *schema.ResourceData, meta interface{}) error {\n\tlog.Debugf(\"Deleting TenantGroup: %v\\n\", d)\n\n\tid := int64(d.Get(\"tenant_group_id\").(int))\n\n\tvar deleteParameters = tenancy.NewTenancyTenantGroupsDeleteParams().WithID(id)\n\n\tc := meta.(*ProviderNetboxClient).client\n\n\tout, err := c.Tenancy.TenancyTenantGroupsDelete(deleteParameters, nil)\n\n\tif err != nil {\n\t\tlog.Debugf(\"Failed to execute OrgTenantGroupsDelete: %v\", err)\n\t}\n\n\tlog.Debugf(\"Done Executing OrgTenantGroupsDelete: %v\", out)\n\n\treturn nil\n}", "func (client *DedicatedHostsClient) delete(ctx context.Context, resourceGroupName string, hostGroupName string, hostName string, options *DedicatedHostsBeginDeleteOptions) (*azcore.Response, error) {\n\treq, err := client.deleteCreateRequest(ctx, resourceGroupName, hostGroupName, hostName, options)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresp, err := client.con.Pipeline().Do(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif !resp.HasStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent) {\n\t\treturn nil, client.deleteHandleError(resp)\n\t}\n\treturn resp, nil\n}", "func (o *StorageDeleteOptions) Run() (err error) {\n\tvar deleteMsg string\n\n\tvar devFile devfileParser.DevfileObj\n\tmPath := \"\"\n\tif o.isDevfile {\n\t\tdevFile, err = devfile.ParseAndValidate(o.devfilePath)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\terr = validate.ValidateDevfileData(devFile.Data)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tmPath, err = devFile.Data.GetVolumeMountPath(o.storageName)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\tmPath = o.LocalConfigInfo.GetMountPath(o.storageName)\n\t}\n\n\tdeleteMsg = fmt.Sprintf(\"Are you sure you want to delete the storage %v mounted to %v in %v component\", o.storageName, mPath, o.componentName)\n\n\tif log.IsJSON() || o.storageForceDeleteFlag || ui.Proceed(deleteMsg) {\n\t\tif o.isDevfile {\n\t\t\terr = devFile.Data.DeleteVolume(o.storageName)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\terr = devFile.WriteYamlDevfile()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t} else {\n\t\t\terr = o.LocalConfigInfo.StorageDelete(o.storageName)\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"failed to delete storage, cause %v\", err)\n\t\t\t}\n\t\t}\n\n\t\tsuccessMessage := fmt.Sprintf(\"Deleted storage %v from %v\", o.storageName, o.componentName)\n\n\t\tif log.IsJSON() {\n\t\t\tstorage.MachineReadableSuccessOutput(o.storageName, successMessage)\n\t\t} else {\n\t\t\tlog.Infof(successMessage)\n\t\t\tlog.Italic(\"\\nPlease use `odo push` command to delete the storage from the cluster\")\n\t\t}\n\t} else {\n\t\treturn fmt.Errorf(\"aborting deletion of storage: %v\", o.storageName)\n\t}\n\n\treturn\n}", "func (client ServicesClient) DeleteSender(req *http.Request) (future ServicesDeleteFuture, err error) {\n\tvar resp *http.Response\n\tfuture.FutureAPI = &azure.Future{}\n\tresp, err = client.Send(req, azure.DoRetryWithRegistration(client.Client))\n\tif err != nil {\n\t\treturn\n\t}\n\tvar azf azure.Future\n\tazf, err = azure.NewFutureFromResponse(resp)\n\tfuture.FutureAPI = &azf\n\tfuture.Result = future.result\n\treturn\n}", "func (client GroupClient) DeleteSecretSender(req *http.Request) (*http.Response, error) {\n\treturn autorest.SendWithSender(client,\n\t\treq,\n\t\tautorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n}", "func (t *SimpleChaincode) delete_account (stub shim.ChaincodeStubInterface, args []string) pb.Response {\n if len(args) != 1 {\n return shim.Error(\"Incorrect number of arguments. Expecting 1\")\n }\n\n // only Admin is allowed to delete accounts\n if !transactor_is_admin(stub) {\n return shim.Error(\"Only admin user is not authorized to delete_account\")\n }\n\n account_name := args[0]\n err := delete_account_(stub, account_name)\n if err != nil {\n return shim.Error(err.Error())\n }\n\n return shim.Success(nil)\n}", "func (m *TeamItemRequestBuilder) Delete(ctx context.Context, requestConfiguration *TeamItemRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.CreateDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": ia572726a95efa92ddd544552cd950653dc691023836923576b2f4bf716cf204a.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": ia572726a95efa92ddd544552cd950653dc691023836923576b2f4bf716cf204a.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.requestAdapter.SendNoContentAsync(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func (c *restClient) DeleteTask(ctx context.Context, req *cloudtaskspb.DeleteTaskRequest, opts ...gax.CallOption) error {\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v2beta3/%v\", req.GetName())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"name\", url.QueryEscape(req.GetName()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\treturn gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"DELETE\", baseUrl.String(), nil)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\t// Returns nil if there is no error, otherwise wraps\n\t\t// the response code and body into a non-nil error\n\t\treturn googleapi.CheckResponse(httpRsp)\n\t}, opts...)\n}", "func (cmd *DeleteAccountsCommand) Run(c *client.Client, args []string) error {\n\tvar path string\n\tif len(args) > 0 {\n\t\tpath = args[0]\n\t} else {\n\t\tpath = fmt.Sprintf(\"/accounts/%v\", url.QueryEscape(cmd.Target))\n\t}\n\tvar payload client.DeleteAccountRequest\n\tif cmd.Payload != \"\" {\n\t\terr := json.Unmarshal([]byte(cmd.Payload), &payload)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"failed to deserialize payload: %s\", err)\n\t\t}\n\t}\n\tlogger := goa.NewLogger(log.New(os.Stderr, \"\", log.LstdFlags))\n\tctx := goa.WithLogger(context.Background(), logger)\n\tresp, err := c.DeleteAccounts(ctx, path, &payload, cmd.ContentType)\n\tif err != nil {\n\t\tgoa.LogError(ctx, \"failed\", \"err\", err)\n\t\treturn err\n\t}\n\n\tgoaclient.HandleResponse(c.Client, resp, cmd.PrettyPrint)\n\treturn nil\n}", "func Delete(ctx context.Context, client *v1.ServiceClient, domainID int) (*v1.ResponseResult, error) {\n\turl := strings.Join([]string{client.Endpoint, strconv.Itoa(domainID)}, \"/\")\n\tresponseResult, err := client.DoRequest(ctx, http.MethodDelete, url, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif responseResult.Err != nil {\n\t\terr = responseResult.Err\n\t}\n\n\treturn responseResult, err\n}", "func (client ThreatIntelligenceIndicatorClient) DeleteSender(req *http.Request) (*http.Response, error) {\n\treturn client.Send(req, azure.DoRetryWithRegistration(client.Client))\n}", "func (suite *TenantTestSuite) TestDeleteUnauthorized() {\n\trequest, _ := http.NewRequest(\"DELETE\", \"/api/v2/admin/tenants/id\", strings.NewReader(\"\"))\n\trequest.Header.Set(\"x-api-key\", \"FOO\")\n\trequest.Header.Set(\"Accept\", \"application/json\")\n\tresponse := httptest.NewRecorder()\n\n\tsuite.router.ServeHTTP(response, request)\n\n\tcode := response.Code\n\toutput := response.Body.String()\n\n\tsuite.Equal(401, code, \"Internal Server Error\")\n\tsuite.Equal(suite.respUnauthorized, output, \"Response body mismatch\")\n}", "func (m *TeamTemplateDefinitionItemRequestBuilder) Delete(ctx context.Context, requestConfiguration *TeamTemplateDefinitionItemRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func (client AppsClient) DeleteSender(req *http.Request) (*http.Response, error) {\n\treturn autorest.SendWithSender(client, req,\n\t\tautorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n}", "func NewCmdCleanUpTenant(f cmdutil.Factory) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"tenant\",\n\t\tShort: \"Hard delete of your tenant pipelines, apps, jobs and releases\",\n\t\tLong: `Hard delete of your tenant pipelines, apps, jobs and releases`,\n\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\tp := cleanUpTenantFlags{}\n\t\t\tif cmd.Flags().Lookup(yesFlag).Value.String() == \"true\" {\n\t\t\t\tp.confirm = true\n\t\t\t}\n\t\t\terr := p.cleanTenant(f)\n\t\t\tif err != nil {\n\t\t\t\tutil.Fatalf(\"%s\\n\", err)\n\t\t\t}\n\t\t\treturn\n\t\t},\n\t}\n\treturn cmd\n}", "func (client ModelClient) DeleteSubListSender(req *http.Request) (*http.Response, error) {\n\treturn autorest.SendWithSender(client, req,\n\t\tautorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n}", "func (t *SimpleChaincode) delete(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n\tfmt.Println(\"- start delete transfer\")\n\n\ttype transferDeleteTransientInput struct {\n\t\tName string `json:\"name\"`\n\t}\n\n\tif len(args) != 0 {\n\t\treturn shim.Error(\"Incorrect number of arguments. Private transfer name must be passed in transient map.\")\n\t}\n\n\ttransMap, err := stub.GetTransient()\n\tif err != nil {\n\t\treturn shim.Error(\"Error getting transient: \" + err.Error())\n\t}\n\n\tif _, ok := transMap[\"transfer_delete\"]; !ok {\n\t\treturn shim.Error(\"transfer_delete must be a key in the transient map\")\n\t}\n\n\tif len(transMap[\"transfer_delete\"]) == 0 {\n\t\treturn shim.Error(\"transfer_delete value in the transient map must be a non-empty JSON string\")\n\t}\n\n\tvar transferDeleteInput transferDeleteTransientInput\n\terr = json.Unmarshal(transMap[\"transfer_delete\"], &transferDeleteInput)\n\tif err != nil {\n\t\treturn shim.Error(\"Failed to decode JSON of: \" + string(transMap[\"transfer_delete\"]))\n\t}\n\n\tif len(transferDeleteInput.Name) == 0 {\n\t\treturn shim.Error(\"name field must be a non-empty string\")\n\t}\n\n\t// to maintain the authorization~name index, we need to read the transfer first and get its authorization\n\tvalAsbytes, err := stub.GetPrivateData(\"collectionFileTransfer\", transferDeleteInput.Name) //get the marble from chaincode state\n\tif err != nil {\n\t\treturn shim.Error(\"Failed to get state for \" + transferDeleteInput.Name)\n\t} else if valAsbytes == nil {\n\t\treturn shim.Error(\"Transfer does not exist: \" + transferDeleteInput.Name)\n\t}\n\n\tvar transferToDelete fileTransfer\n\terr = json.Unmarshal([]byte(valAsbytes), &transferToDelete)\n\tif err != nil {\n\t\treturn shim.Error(\"Failed to decode JSON of: \" + string(valAsbytes))\n\t}\n\n\t// delete the transfer from state\n\terr = stub.DelPrivateData(\"collectionFileTransfer\", transferDeleteInput.Name)\n\tif err != nil {\n\t\treturn shim.Error(\"Failed to delete state:\" + err.Error())\n\t}\n\n\t// Also delete the transfer from the authorization~name index\n\tindexName := \"authorization~name\"\n\tauthorizationNameIndexKey, err := stub.CreateCompositeKey(indexName, []string{transferToDelete.Authorization, transferToDelete.Name})\n\tif err != nil {\n\t\treturn shim.Error(err.Error())\n\t}\n\terr = stub.DelPrivateData(\"collectionFileTransfer\", authorizationNameIndexKey)\n\tif err != nil {\n\t\treturn shim.Error(\"Failed to delete state:\" + err.Error())\n\t}\n\n\t// Finally, delete private details of transfer\n\terr = stub.DelPrivateData(\"collectionFileTransferPrivateDetails\", transferDeleteInput.Name)\n\tif err != nil {\n\t\treturn shim.Error(err.Error())\n\t}\n\n\treturn shim.Success(nil)\n}", "func (client Client) DeleteSender(req *http.Request) (*http.Response, error) {\n\treturn client.Send(req, autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n}" ]
[ "0.6977299", "0.66991556", "0.65932226", "0.63422877", "0.62385774", "0.61525124", "0.59552073", "0.59169596", "0.59117275", "0.59082484", "0.58545953", "0.58142763", "0.5809877", "0.5798616", "0.5741293", "0.57411957", "0.5740928", "0.57350224", "0.5734197", "0.56706774", "0.5669248", "0.5668251", "0.5656252", "0.5648874", "0.5589363", "0.5560481", "0.55582386", "0.5557161", "0.5556107", "0.55534196", "0.5526208", "0.55217093", "0.55150384", "0.5505475", "0.5494945", "0.5486628", "0.5482943", "0.5478352", "0.54769415", "0.54763013", "0.54733497", "0.5463844", "0.5459022", "0.54588425", "0.5457046", "0.5440274", "0.5438917", "0.54375434", "0.5435383", "0.5432418", "0.5430511", "0.5429673", "0.54263645", "0.5415174", "0.54108167", "0.5404686", "0.54022014", "0.54013306", "0.5387452", "0.53849196", "0.53841376", "0.53793585", "0.5370943", "0.53660905", "0.535103", "0.53459674", "0.534343", "0.53333014", "0.53309774", "0.5329153", "0.532465", "0.5324063", "0.53238475", "0.5320504", "0.53168315", "0.5311641", "0.53055924", "0.528362", "0.5280712", "0.52797973", "0.5274784", "0.52736896", "0.5273302", "0.5260011", "0.52428293", "0.5241722", "0.52307844", "0.5214845", "0.520985", "0.5209824", "0.5206834", "0.5206766", "0.5201747", "0.5200997", "0.519492", "0.5190452", "0.5185652", "0.5184987", "0.5177839", "0.5176529" ]
0.6896285
1
Overwrites the tenant in the config file
func setTenant(c *cli.Context) error { err := checkArgCount(c, 1) if err != nil { return err } name := c.Args().First() client.Photonclient, err = client.GetClient(c) if err != nil { return err } // Ensure tenant exists id, err := findTenantID(name) if len(id) == 0 || err != nil { return err } config, err := cf.LoadConfig() if err != nil { return err } config.Tenant = &cf.TenantConfiguration{Name: name, ID: id} err = cf.SaveConfig(config) if err != nil { return err } err = clearConfigProject("") if err != nil { return err } if !c.GlobalIsSet("non-interactive") { fmt.Printf("Tenant set to '%s'\n", name) } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (m *CloudPcConnection) SetTenantId(value *string)() {\n err := m.GetBackingStore().Set(\"tenantId\", value)\n if err != nil {\n panic(err)\n }\n}", "func (c *TenantController) Update(ctx *app.UpdateTenantContext) error {\n\tuserToken := goajwt.ContextJWT(ctx)\n\tif userToken == nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewUnauthorizedError(\"Missing JWT token\"))\n\t}\n\tttoken := &auth.TenantToken{Token: userToken}\n\ttenant, err := c.tenantService.GetTenant(ttoken.Subject())\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewNotFoundError(\"tenants\", ttoken.Subject().String()))\n\t}\n\n\t// fetch the cluster the user belongs to\n\tuser, err := c.authClientService.GetUser(ctx)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, err)\n\t}\n\n\tif user.UserData.Cluster == nil {\n\t\tlog.Error(ctx, nil, \"no cluster defined for tenant\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewInternalError(ctx, fmt.Errorf(\"unable to provision to undefined cluster\")))\n\t}\n\n\tcluster, err := c.clusterService.GetCluster(ctx, *user.UserData.Cluster)\n\tif err != nil {\n\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\"err\": err,\n\t\t\t\"cluster_url\": *user.UserData.Cluster,\n\t\t}, \"unable to fetch cluster\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewInternalError(ctx, err))\n\t}\n\n\t// create openshift config\n\topenshiftConfig := openshift.NewConfig(c.config, user.UserData, cluster.User, cluster.Token, cluster.APIURL)\n\n\t// update tenant config\n\ttenant.OSUsername = user.OpenShiftUsername\n\tif tenant.NsBaseName == \"\" {\n\t\ttenant.NsBaseName = env.RetrieveUserName(user.OpenShiftUsername)\n\t}\n\tif err = c.tenantService.SaveTenant(tenant); err != nil {\n\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\"err\": err,\n\t\t}, \"unable to update tenant configuration\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewInternalError(ctx, fmt.Errorf(\"unable to update tenant configuration: %v\", err)))\n\t}\n\n\tgo func() {\n\t\tctx := ctx\n\t\tt := tenant\n\t\terr = openshift.RawUpdateTenant(\n\t\t\tctx,\n\t\t\topenshiftConfig,\n\t\t\tInitTenant(ctx, openshiftConfig.MasterURL, c.tenantService, t),\n\t\t\tuser.OpenShiftUsername,\n\t\t\ttenant.NsBaseName)\n\n\t\tif err != nil {\n\t\t\tsentry.LogError(ctx, map[string]interface{}{\n\t\t\t\t\"os_user\": user.OpenShiftUsername,\n\t\t\t}, err, \"unable initialize tenant\")\n\t\t}\n\t}()\n\n\tctx.ResponseData.Header().Set(\"Location\", rest.AbsoluteURL(ctx.RequestData.Request, app.TenantHref()))\n\treturn ctx.Accepted()\n}", "func (m *SharepointIds) SetTenantId(value *string)() {\n err := m.GetBackingStore().Set(\"tenantId\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *DeviceCompliancePolicySettingStateSummary) SetTenantId(value *string)() {\n err := m.GetBackingStore().Set(\"tenantId\", value)\n if err != nil {\n panic(err)\n }\n}", "func TestUpdateTenantConfig(t *testing.T) {\n\tctx, cancelFunc := context.WithTimeout(context.Background(), standardTimeout)\n\tdefer cancelFunc()\n\n\tconfig := bat.TenantConfig{\n\t\tName: \"UpdateTenantConfigTest\",\n\t\tSubnetBits: 30,\n\t}\n\n\ttenant, err := bat.CreateTenant(ctx, config)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to create tenant : %v\", err)\n\t}\n\n\tdefer func() {\n\t\terr = bat.DeleteTenant(ctx, tenant.ID)\n\t\tif err != nil {\n\t\t\tt.Fatalf(\"Failed to delete tenant: %v\", err)\n\t\t}\n\t}()\n\n\tif tenant.Name != config.Name {\n\t\tt.Fatalf(\"Failed to create tenant\")\n\t}\n\n\tcfg, err := bat.GetTenantConfig(ctx, tenant.ID)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to retrieve tenant config: %v\", err)\n\t}\n\n\tif cfg.Name != config.Name || cfg.SubnetBits != config.SubnetBits {\n\t\tt.Fatalf(\"Failed to retrieve tenant config\")\n\t}\n\n\tconfig.Name = \"Updated Tenant\"\n\tconfig.SubnetBits = 20\n\n\terr = bat.UpdateTenant(ctx, tenant.ID, config)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to update tenant config: %v\", err)\n\t}\n\n\tcfg, err = bat.GetTenantConfig(ctx, tenant.ID)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to retrieve tenant config: %v\", err)\n\t}\n\n\tif cfg.Name != config.Name || cfg.SubnetBits != config.SubnetBits {\n\t\tt.Fatalf(\"Failed to update tenant config: expected %s %d, got %s %d\", config.Name, config.SubnetBits, cfg.Name, cfg.SubnetBits)\n\t}\n}", "func (m *AadUserConversationMember) SetTenantId(value *string)() {\n m.tenantId = value\n}", "func (p *AzureProvider) Configure(tenant string) {\n\tif tenant == \"\" || tenant == \"common\" {\n\t\t// tenant is empty or default, remain on the default \"common\" tenant\n\t\treturn\n\t}\n\n\t// Specific tennant specified, override the Login and RedeemURLs\n\tp.Tenant = tenant\n\toverrideTenantURL(p.LoginURL, azureDefaultLoginURL, tenant, \"authorize\")\n\toverrideTenantURL(p.RedeemURL, azureDefaultRedeemURL, tenant, \"token\")\n}", "func (c *TenantController) Update(ctx *app.UpdateTenantContext) error {\n\tuserToken := goajwt.ContextJWT(ctx)\n\tif userToken == nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewUnauthorizedError(\"Missing JWT token\"))\n\t}\n\tttoken := &TenantToken{token: userToken}\n\ttenant, err := c.tenantService.GetTenant(ttoken.Subject())\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewNotFoundError(\"tenants\", ttoken.Subject().String()))\n\t}\n\n\t// fetch the cluster the user belongs to\n\tuser, err := c.userService.GetUser(ctx, ttoken.Subject())\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, err)\n\t}\n\n\tif user.Cluster == nil {\n\t\tlog.Error(ctx, nil, \"no cluster defined for tenant\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewInternalError(ctx, fmt.Errorf(\"unable to provision to undefined cluster\")))\n\t}\n\n\t// fetch the users cluster token\n\topenshiftUsername, _, err := c.resolveTenant(ctx, *user.Cluster, userToken.Raw)\n\tif err != nil {\n\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\"err\": err,\n\t\t\t\"cluster_url\": *user.Cluster,\n\t\t}, \"unable to fetch tenant token from auth\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewUnauthorizedError(\"Could not resolve user token\"))\n\t}\n\n\tcluster, err := c.resolveCluster(ctx, *user.Cluster)\n\tif err != nil {\n\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\"err\": err,\n\t\t\t\"cluster_url\": *user.Cluster,\n\t\t}, \"unable to fetch cluster\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewInternalError(ctx, err))\n\t}\n\n\t// create openshift config\n\topenshiftConfig := openshift.NewConfig(c.defaultOpenshiftConfig, user, cluster.User, cluster.Token, cluster.APIURL)\n\n\tgo func() {\n\t\tctx := ctx\n\t\tt := tenant\n\t\terr = openshift.RawUpdateTenant(\n\t\t\tctx,\n\t\t\topenshiftConfig,\n\t\t\tInitTenant(ctx, openshiftConfig.MasterURL, c.tenantService, t),\n\t\t\topenshiftUsername,\n\t\t\tc.templateVars)\n\n\t\tif err != nil {\n\t\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\t\"err\": err,\n\t\t\t\t\"os_user\": openshiftUsername,\n\t\t\t}, \"unable initialize tenant\")\n\t\t}\n\t}()\n\n\tctx.ResponseData.Header().Set(\"Location\", rest.AbsoluteURL(ctx.RequestData.Request, app.TenantHref()))\n\treturn ctx.Accepted()\n}", "func (m *SecureScoreControlProfile) SetAzureTenantId(value *string)() {\n err := m.GetBackingStore().Set(\"azureTenantId\", value)\n if err != nil {\n panic(err)\n }\n}", "func (api *tenantAPI) Update(obj *cluster.Tenant) error {\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = apicl.ClusterV1().Tenant().Update(context.Background(), obj)\n\t\treturn err\n\t}\n\n\tapi.ct.handleTenantEvent(&kvstore.WatchEvent{Object: obj, Type: kvstore.Updated})\n\treturn nil\n}", "func Tenant(tenant string) Modifier {\n\treturn func(r *http.Request) error {\n\t\tr.Header.Set(\"Hawkular-Tenant\", tenant)\n\t\treturn nil\n\t}\n}", "func Tenant(tenant string) Modifier {\n\treturn func(r *http.Request) error {\n\t\tr.Header.Set(\"Hawkular-Tenant\", tenant)\n\t\treturn nil\n\t}\n}", "func (s *TenantServiceListener) Set(ctx context.Context, in *pb.TenantName) (*google_protobuf.Empty, error) {\n\tlog.Printf(\"Tenant Set called '%s'\", in.Name)\n\n\tif currentTenant != nil && currentTenant.name == in.GetName() {\n\t\tlog.Printf(\"Tenant '%s' is already selected\", in.GetName())\n\t\treturn &google_protobuf.Empty{}, nil\n\t}\n\n\tservice, err := providers.GetService(in.GetName())\n\tif err != nil {\n\t\treturn &google_protobuf.Empty{}, fmt.Errorf(\"Unable to set tenant '%s': %s\", in.GetName(), err.Error())\n\t}\n\tcurrentTenant = &Tenant{name: in.GetName(), Service: service}\n\tlog.Printf(\"Current tenant is now '%s'\", in.GetName())\n\treturn &google_protobuf.Empty{}, nil\n}", "func (m *TenantManager) Update(t *Tenant, opts ...RequestOption) (err error) {\n\treturn m.Request(\"PATCH\", m.URI(\"tenants\", \"settings\"), t, opts...)\n}", "func (t *Tenants) Set(tenant *Tenant) error {\n\tw := &bytes.Buffer{}\n\tif err := tenant.Encode(w); err != nil {\n\t\treturn err\n\t}\n\treturn t.store.Set(tenant.ID, w.Bytes())\n}", "func (m *SecureScoreControlProfile) SetAzureTenantId(value *string)() {\n m.azureTenantId = value\n}", "func TestTenantPermissionsUpdate(t *testing.T) {\n\tctx, cancelFunc := context.WithTimeout(context.Background(), standardTimeout)\n\tdefer cancelFunc()\n\n\tconfig := bat.TenantConfig{\n\t\tName: \"TestTenantPermissionsUpdate\",\n\t}\n\n\ttenant, err := bat.CreateTenant(ctx, config)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to create tenant : %v\", err)\n\t}\n\n\tdefer func() {\n\t\terr = bat.DeleteTenant(ctx, tenant.ID)\n\t\tif err != nil {\n\t\t\tt.Fatalf(\"Failed to delete tenant: %v\", err)\n\t\t}\n\t}()\n\n\tif tenant.Name != config.Name {\n\t\tt.Fatalf(\"Failed to create tenant\")\n\t}\n\n\tcfg, err := bat.GetTenantConfig(ctx, tenant.ID)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to retrieve tenant config: %v\", err)\n\t}\n\n\tif cfg.Name != config.Name || cfg.Permissions.PrivilegedContainers != false {\n\t\tt.Fatalf(\"Failed to retrieve tenant config\")\n\t}\n\n\tconfig.Permissions.PrivilegedContainers = true\n\n\terr = bat.UpdateTenant(ctx, tenant.ID, config)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to update tenant: %v\", err)\n\t}\n\n\tcfg, err = bat.GetTenantConfig(ctx, tenant.ID)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to retrieve tenant config: %v\", err)\n\t}\n\n\tif cfg.Name != config.Name || cfg.Permissions.PrivilegedContainers != true {\n\t\tt.Fatalf(\"Failed to retrieve tenant config\")\n\t}\n\n\tconfig.Permissions.PrivilegedContainers = false\n\n\terr = bat.UpdateTenant(ctx, tenant.ID, config)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to update tenant: %v\", err)\n\t}\n\n\tcfg, err = bat.GetTenantConfig(ctx, tenant.ID)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to retrieve tenant config: %v\", err)\n\t}\n\n\tif cfg.Name != config.Name || cfg.Permissions.PrivilegedContainers != false {\n\t\tt.Fatalf(\"Failed to retrieve tenant config\")\n\t}\n}", "func (m *CloudPcConnection) SetTenantDisplayName(value *string)() {\n err := m.GetBackingStore().Set(\"tenantDisplayName\", value)\n if err != nil {\n panic(err)\n }\n}", "func (n *NginxConfigFileTemplete) ClearByTenant(tenant string) error {\n\ttenantConfigFile := path.Join(n.configFileDirPath, \"http\", tenant)\n\tif err := os.RemoveAll(tenantConfigFile); err != nil {\n\t\treturn err\n\t}\n\ttenantStreamConfigFile := path.Join(n.configFileDirPath, \"stream\", tenant)\n\treturn os.RemoveAll(tenantStreamConfigFile)\n}", "func SetTenant(tenant string) func(*AviSession) error {\n\treturn func(sess *AviSession) error {\n\t\treturn sess.setTenant(tenant)\n\t}\n}", "func TenantID() string {\n\treturn tenantID\n}", "func TenantID() string {\n\treturn tenantID\n}", "func TenantID() string {\n\treturn tenantID\n}", "func withTenant(node *Tenant) tenantOption {\n\treturn func(m *TenantMutation) {\n\t\tm.oldValue = func(context.Context) (*Tenant, error) {\n\t\t\treturn node, nil\n\t\t}\n\t\tm.id = &node.ID\n\t}\n}", "func InitTenant(ctx context.Context, masterURL string, service tenant.Service, currentTenant *tenant.Tenant) Callback {\n\tvar maxResourceQuotaStatusCheck int32 = 50 // technically a global retry count across all ResourceQuota on all Tenant Namespaces\n\tvar currentResourceQuotaStatusCheck int32 // default is 0\n\treturn func(statusCode int, method string, request, response map[interface{}]interface{}, versionMapping map[env.Type]string) (string, map[interface{}]interface{}, error) {\n\t\tlog.Info(ctx, map[string]interface{}{\n\t\t\t\"status\": statusCode,\n\t\t\t\"method\": method,\n\t\t\t\"cluster_url\": masterURL,\n\t\t\t\"namespace\": env.GetNamespace(request),\n\t\t\t\"name\": env.GetName(request),\n\t\t\t\"kind\": env.GetKind(request),\n\t\t\t\"request\": yamlString(request),\n\t\t\t\"response\": yamlString(response),\n\t\t}, \"resource requested\")\n\t\tif statusCode == http.StatusConflict || statusCode == http.StatusForbidden {\n\t\t\tif env.GetKind(request) == env.ValKindNamespace || env.GetKind(request) == env.ValKindProjectRequest ||\n\t\t\t\tenv.GetKind(request) == env.ValKindPersistenceVolumeClaim {\n\t\t\t\treturn \"\", nil, fmt.Errorf(\"unable to create %s - should create with other base-name\", env.GetNamespace(request))\n\t\t\t}\n\t\t\treturn \"DELETE\", request, nil\n\t\t} else if statusCode == http.StatusCreated {\n\t\t\tif env.GetKind(request) == env.ValKindProjectRequest {\n\t\t\t\tname := env.GetName(request)\n\t\t\t\tenvType := tenant.GetNamespaceType(name, currentTenant.NsBaseName)\n\t\t\t\ttemplatesVersion := versionMapping[envType]\n\t\t\t\tservice.SaveNamespace(&tenant.Namespace{\n\t\t\t\t\tTenantID: currentTenant.ID,\n\t\t\t\t\tName: name,\n\t\t\t\t\tState: tenant.Ready,\n\t\t\t\t\tVersion: templatesVersion,\n\t\t\t\t\tType: envType,\n\t\t\t\t\tMasterURL: masterURL,\n\t\t\t\t\tUpdatedBy: configuration.Commit,\n\t\t\t\t})\n\n\t\t\t\t// HACK to workaround osio applying some dsaas-user permissions async\n\t\t\t\t// Should loop on a Check if allowed type of call instead\n\t\t\t\ttime.Sleep(time.Second * 5)\n\n\t\t\t} else if env.GetKind(request) == env.ValKindNamespace {\n\t\t\t\tname := env.GetName(request)\n\t\t\t\tenvType := tenant.GetNamespaceType(name, currentTenant.NsBaseName)\n\t\t\t\ttemplatesVersion := versionMapping[envType]\n\t\t\t\tservice.SaveNamespace(&tenant.Namespace{\n\t\t\t\t\tTenantID: currentTenant.ID,\n\t\t\t\t\tName: name,\n\t\t\t\t\tState: tenant.Ready,\n\t\t\t\t\tVersion: templatesVersion,\n\t\t\t\t\tType: envType,\n\t\t\t\t\tMasterURL: masterURL,\n\t\t\t\t\tUpdatedBy: configuration.Commit,\n\t\t\t\t})\n\t\t\t} else if env.GetKind(request) == env.ValKindResourceQuota {\n\t\t\t\t// trigger a check status loop\n\t\t\t\ttime.Sleep(time.Millisecond * 50)\n\t\t\t\treturn \"GET\", response, nil\n\t\t\t}\n\t\t\treturn \"\", nil, nil\n\t\t} else if statusCode == http.StatusOK {\n\t\t\tif method == \"DELETE\" {\n\t\t\t\treturn \"POST\", request, nil\n\t\t\t} else if method == \"GET\" {\n\t\t\t\tif env.GetKind(request) == env.ValKindResourceQuota {\n\n\t\t\t\t\tif env.HasValidStatus(response) || atomic.LoadInt32(&currentResourceQuotaStatusCheck) >= maxResourceQuotaStatusCheck {\n\t\t\t\t\t\treturn \"\", nil, nil\n\t\t\t\t\t}\n\t\t\t\t\tatomic.AddInt32(&currentResourceQuotaStatusCheck, 1)\n\t\t\t\t\ttime.Sleep(time.Millisecond * 50)\n\t\t\t\t\treturn \"GET\", response, nil\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn \"\", nil, nil\n\t\t}\n\t\tlog.Info(ctx, map[string]interface{}{\n\t\t\t\"status\": statusCode,\n\t\t\t\"method\": method,\n\t\t\t\"namespace\": env.GetNamespace(request),\n\t\t\t\"cluster_url\": masterURL,\n\t\t\t\"name\": env.GetName(request),\n\t\t\t\"kind\": env.GetKind(request),\n\t\t\t\"request\": yamlString(request),\n\t\t\t\"response\": yamlString(response),\n\t\t}, \"unhandled resource response\")\n\t\treturn \"\", nil, nil\n\t}\n}", "func UpdateTenant(ctx context.Context, config tenantConfig) error {\n\n\tc, err := createClient(ctx, config)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Ignore response for now\n\t_, err = c.UpdateTenant(goasupport.ForwardContextRequestID(ctx), tenant.UpdateTenantPath())\n\n\treturn err\n}", "func (s *MultiElasticsearch) Tenant(tenant string) Store {\n\tkv, found := s.stores[tenant]\n\tif !found {\n\t\tkv = NewElasticsearch(s.config, s.client, tenant, s.typeName)\n\t\ts.stores[tenant] = kv\n\t}\n\treturn kv\n}", "func (s *TenantListener) Set(ctx context.Context, in *pb.TenantName) (empty *googleprotobuf.Empty, err error) {\n\tempty = &googleprotobuf.Empty{}\n\tif s == nil {\n\t\treturn empty, status.Errorf(codes.FailedPrecondition, fail.InvalidInstanceError().Message())\n\t}\n\tif in == nil {\n\t\treturn empty, status.Errorf(codes.InvalidArgument, fail.InvalidParameterError(\"in\", \"cannot be nil\").Message())\n\t}\n\tname := in.GetName()\n\t// FIXME: validate parameters\n\n\ttracer := debug.NewTracer(nil, fmt.Sprintf(\"('%s')\", name), true).WithStopwatch().GoingIn()\n\tdefer tracer.OnExitTrace()()\n\tdefer fail.OnExitLogError(tracer.TraceMessage(\"\"), &err)()\n\n\tctx, cancelFunc := context.WithCancel(ctx)\n\t// FIXME: handle error\n\tif err := srvutils.JobRegister(ctx, cancelFunc, \"Tenant Set \"+name); err == nil {\n\t\tdefer srvutils.JobDeregister(ctx)\n\t}\n\n\tif currentTenant != nil && currentTenant.name == in.GetName() {\n\t\treturn empty, nil\n\t}\n\n\tservice, err := iaas.UseService(in.GetName())\n\tif err != nil {\n\t\treturn empty, fmt.Errorf(\"unable to set tenant '%s': %s\", name, getUserMessage(err))\n\t}\n\tcurrentTenant = &Tenant{name: in.GetName(), Service: service}\n\tlog.Infof(\"Current tenant is now '%s'\", name)\n\treturn empty, nil\n}", "func (p *AzureProvider) Configure(tenant string) {\n\tp.Tenant = tenant\n\tif tenant == \"\" {\n\t\tp.Tenant = \"common\"\n\t}\n\n\tif p.LoginURL == nil || p.LoginURL.String() == \"\" {\n\t\tp.LoginURL = &url.URL{\n\t\t\tScheme: \"https\",\n\t\t\tHost: \"login.microsoftonline.com\",\n\t\t\tPath: \"/\" + p.Tenant + \"/oauth2/authorize\"}\n\t}\n\tif p.RedeemURL == nil || p.RedeemURL.String() == \"\" {\n\t\tp.RedeemURL = &url.URL{\n\t\t\tScheme: \"https\",\n\t\t\tHost: \"login.microsoftonline.com\",\n\t\t\tPath: \"/\" + p.Tenant + \"/oauth2/token\",\n\t\t}\n\t}\n}", "func (m *DeviceCompliancePolicySettingStateSummary) SetTenantDisplayName(value *string)() {\n err := m.GetBackingStore().Set(\"tenantDisplayName\", value)\n if err != nil {\n panic(err)\n }\n}", "func TestCreateTenant(t *testing.T) {\n\tctx, cancelFunc := context.WithTimeout(context.Background(), standardTimeout)\n\tdefer cancelFunc()\n\n\tconfig := bat.TenantConfig{\n\t\tName: \"CreateTenantTest\",\n\t\tSubnetBits: 20,\n\t}\n\n\ttenant, err := bat.CreateTenant(ctx, config)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to create tenant : %v\", err)\n\t}\n\n\tdefer func() {\n\t\terr = bat.DeleteTenant(ctx, tenant.ID)\n\t\tif err != nil {\n\t\t\tt.Fatalf(\"Failed to delete tenant: %v\", err)\n\t\t}\n\t}()\n\n\tif tenant.Name != config.Name {\n\t\tt.Fatalf(\"Failed to create tenant\")\n\t}\n\n\ttenants, err := bat.GetAllTenants(ctx)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to retrieve tenant list : %v\", err)\n\t}\n\n\tfor _, tt := range tenants {\n\t\tif tt.Name == config.Name {\n\t\t\treturn\n\t\t}\n\t}\n\n\tt.Fatal(\"did not find new tenant in tenants list\")\n}", "func getTenant(c *cli.Context, w io.Writer) error {\n\terr := checkArgCount(c, 0)\n\tif err != nil {\n\t\treturn err\n\t}\n\tconfig, err := cf.LoadConfig()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttenant := config.Tenant\n\tif tenant == nil {\n\t\tfmt.Printf(\"No tenant selected\\n\")\n\t} else {\n\t\tif c.GlobalIsSet(\"non-interactive\") {\n\t\t\tfmt.Printf(\"%s\\t%s\\n\", tenant.ID, tenant.Name)\n\t\t} else if utils.NeedsFormatting(c) {\n\t\t\tutils.FormatObject(tenant, w, c)\n\t\t} else {\n\t\t\tfmt.Printf(\"Current tenant is '%s' with ID %s\\n\", tenant.Name, tenant.ID)\n\t\t}\n\t}\n\treturn nil\n}", "func (o AuthConfigAzureAdOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *AuthConfigAzureAd) pulumi.StringOutput { return v.TenantId }).(pulumi.StringOutput)\n}", "func deleteTenant(c *cli.Context) error {\n\terr := checkArgCount(c, 1)\n\tif err != nil {\n\t\treturn err\n\t}\n\tid := c.Args().First()\n\n\tclient.Photonclient, err = client.GetClient(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdeleteTask, err := client.Photonclient.Tenants.Delete(id)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = waitOnTaskOperation(deleteTask.ID, c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = clearConfigTenant(id)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func refreshAccessToken() {\n var newCfg Config = getNewValues(makeRefreshPost())\n MarshalledCfg, _ := yaml.Marshal(&newCfg)\n err := ioutil.WriteFile(pathToConfig(), MarshalledCfg, 0755)\n if err != nil {\n fmt.Println(\"An error occured during writing the new file, here are your new tokens:\\n - Access Token = \" + newCfg.AccessToken + \"\\n - Refresh Token = \" + newCfg.RefreshToken)\n os.Exit(2)\n }\n fmt.Println(\"New token generated and stored in ~/.larder/config.yml!\")\n}", "func (c *TenantController) Setup(ctx *app.SetupTenantContext) error {\n\tuserToken := goajwt.ContextJWT(ctx)\n\tif userToken == nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewUnauthorizedError(\"Missing JWT token\"))\n\t}\n\tttoken := &auth.TenantToken{Token: userToken}\n\texists := c.tenantService.Exists(ttoken.Subject())\n\tif exists {\n\t\treturn ctx.Conflict()\n\t}\n\n\t// fetch the cluster the user belongs to\n\tuser, err := c.authClientService.GetUser(ctx)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, err)\n\t}\n\n\tif user.UserData.Cluster == nil {\n\t\tlog.Error(ctx, nil, \"no cluster defined for tenant\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewInternalError(ctx, fmt.Errorf(\"unable to provision to undefined cluster\")))\n\t}\n\n\t// fetch the cluster info\n\tcluster, err := c.clusterService.GetCluster(ctx, *user.UserData.Cluster)\n\tif err != nil {\n\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\"err\": err,\n\t\t\t\"cluster_url\": *user.UserData.Cluster,\n\t\t}, \"unable to fetch cluster\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewInternalError(ctx, err))\n\t}\n\n\tnsBaseName, err := tenant.ConstructNsBaseName(c.tenantService, env.RetrieveUserName(user.OpenShiftUsername))\n\tif err != nil {\n\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\"err\": err,\n\t\t\t\"os_username\": user.OpenShiftUsername,\n\t\t}, \"unable to construct namespace base name\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewInternalError(ctx, err))\n\t}\n\n\t// create openshift config\n\topenshiftConfig := openshift.NewConfig(c.config, user.UserData, cluster.User, cluster.Token, cluster.APIURL)\n\ttenant := &tenant.Tenant{\n\t\tID: ttoken.Subject(),\n\t\tEmail: ttoken.Email(),\n\t\tOSUsername: user.OpenShiftUsername,\n\t\tNsBaseName: nsBaseName,\n\t}\n\terr = c.tenantService.CreateTenant(tenant)\n\tif err != nil {\n\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\"err\": err,\n\t\t}, \"unable to store tenant configuration\")\n\t\treturn ctx.Conflict()\n\t}\n\n\tgo func() {\n\t\tctx := ctx\n\t\tt := tenant\n\t\terr = openshift.RawInitTenant(\n\t\t\tctx,\n\t\t\topenshiftConfig,\n\t\t\tInitTenant(ctx, openshiftConfig.MasterURL, c.tenantService, t),\n\t\t\tuser.OpenShiftUsername,\n\t\t\tnsBaseName,\n\t\t\tuser.OpenShiftUserToken)\n\n\t\tif err != nil {\n\t\t\tsentry.LogError(ctx, map[string]interface{}{\n\t\t\t\t\"os_user\": user.OpenShiftUsername,\n\t\t\t}, err, \"unable initialize tenant\")\n\t\t}\n\t}()\n\n\tctx.ResponseData.Header().Set(\"Location\", rest.AbsoluteURL(ctx.RequestData.Request, app.TenantHref()))\n\treturn ctx.Accepted()\n}", "func TestGetTenantConfig(t *testing.T) {\n\tctx, cancelFunc := context.WithTimeout(context.Background(), standardTimeout)\n\tdefer cancelFunc()\n\n\tconfig := bat.TenantConfig{\n\t\tName: \"GetTenantConfigTest\",\n\t\tSubnetBits: 30,\n\t}\n\n\ttenant, err := bat.CreateTenant(ctx, config)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to create tenant : %v\", err)\n\t}\n\n\tdefer func() {\n\t\terr = bat.DeleteTenant(ctx, tenant.ID)\n\t\tif err != nil {\n\t\t\tt.Fatalf(\"Failed to delete tenant: %v\", err)\n\t\t}\n\t}()\n\n\tif tenant.Name != config.Name {\n\t\tt.Fatalf(\"Failed to create tenant\")\n\t}\n\n\tcfg, err := bat.GetTenantConfig(ctx, tenant.ID)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to retrieve tenant config: %v\", err)\n\t}\n\n\tif cfg.Name != config.Name || cfg.SubnetBits != config.SubnetBits {\n\t\tt.Fatalf(\"Failed to retrieve tenant config\")\n\t}\n}", "func (nuo *NodeUpdateOne) SetTenant(s string) *NodeUpdateOne {\n\tnuo.mutation.SetTenant(s)\n\treturn nuo\n}", "func (api *tenantAPI) Create(obj *cluster.Tenant) error {\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = apicl.ClusterV1().Tenant().Create(context.Background(), obj)\n\t\tif err != nil && strings.Contains(err.Error(), \"AlreadyExists\") {\n\t\t\t_, err = apicl.ClusterV1().Tenant().Update(context.Background(), obj)\n\n\t\t}\n\t\treturn err\n\t}\n\n\tapi.ct.handleTenantEvent(&kvstore.WatchEvent{Object: obj, Type: kvstore.Created})\n\treturn nil\n}", "func editEndpointInConfig(t *testing.T, configFilePath, endpoint string) {\n\tvar config cli.Config\n\tb, err := ioutil.ReadFile(configFilePath)\n\trequire.NoError(t, err)\n\n\terr = yaml.Unmarshal(b, &config)\n\trequire.NoError(t, err)\n\n\tconfig.Endpoint = endpoint\n\n\tb, err = yaml.Marshal(&config)\n\trequire.NoError(t, err)\n\n\terr = ioutil.WriteFile(configFilePath, b, 0655)\n\trequire.NoError(t, err)\n\n}", "func UpdateTenant(p Tenant, db *database.DB) error {\n\n\tvar tenant Tenant\n\n\tresult := db.Client.Find(&tenant, p.ID)\n\tif result.Error != nil {\n\t\treturn ErrTenantNotFound\n\t}\n\n\t// update fields\n\ttenant.Name = p.Name\n\ttenant.Description = p.Description\n\n\tupdateresult := db.Client.Save(&tenant)\n\tif updateresult.Error != nil {\n\t\treturn ErrTenantNotFound\n\t}\n\n\treturn nil\n}", "func (s *compassE2EState) GetDefaultTenant() string {\n\treturn s.config.Tenant\n}", "func InitTenant(ctx context.Context, masterURL string, service tenant.Service, currentTenant *tenant.Tenant) openshift.Callback {\n\tvar maxResourceQuotaStatusCheck int32 = 50 // technically a global retry count across all ResourceQuota on all Tenant Namespaces\n\tvar currentResourceQuotaStatusCheck int32 // default is 0\n\treturn func(statusCode int, method string, request, response map[interface{}]interface{}) (string, map[interface{}]interface{}) {\n\t\tlog.Info(ctx, map[string]interface{}{\n\t\t\t\"status\": statusCode,\n\t\t\t\"method\": method,\n\t\t\t\"cluster_url\": masterURL,\n\t\t\t\"namespace\": env.GetNamespace(request),\n\t\t\t\"name\": env.GetName(request),\n\t\t\t\"kind\": env.GetKind(request),\n\t\t\t\"request\": yamlString(request),\n\t\t\t\"response\": yamlString(response),\n\t\t}, \"resource requested\")\n\t\tif statusCode == http.StatusConflict {\n\t\t\tif env.GetKind(request) == env.ValKindNamespace {\n\t\t\t\treturn \"\", nil\n\t\t\t}\n\t\t\tif env.GetKind(request) == env.ValKindProjectRequest {\n\t\t\t\treturn \"\", nil\n\t\t\t}\n\t\t\tif env.GetKind(request) == env.ValKindPersistenceVolumeClaim {\n\t\t\t\treturn \"\", nil\n\t\t\t}\n\t\t\tif env.GetKind(request) == env.ValKindServiceAccount {\n\t\t\t\treturn \"\", nil\n\t\t\t}\n\t\t\treturn \"DELETE\", request\n\t\t} else if statusCode == http.StatusCreated {\n\t\t\tif env.GetKind(request) == env.ValKindProjectRequest {\n\t\t\t\tname := env.GetName(request)\n\t\t\t\tservice.SaveNamespace(&tenant.Namespace{\n\t\t\t\t\tTenantID: currentTenant.ID,\n\t\t\t\t\tName: name,\n\t\t\t\t\tState: \"created\",\n\t\t\t\t\tVersion: env.GetLabelVersion(request),\n\t\t\t\t\tType: tenant.GetNamespaceType(name, currentTenant.NsBaseName),\n\t\t\t\t\tMasterURL: masterURL,\n\t\t\t\t})\n\n\t\t\t\t// HACK to workaround osio applying some dsaas-user permissions async\n\t\t\t\t// Should loop on a Check if allowed type of call instead\n\t\t\t\ttime.Sleep(time.Second * 5)\n\n\t\t\t} else if env.GetKind(request) == env.ValKindNamespace {\n\t\t\t\tname := env.GetName(request)\n\t\t\t\tservice.SaveNamespace(&tenant.Namespace{\n\t\t\t\t\tTenantID: currentTenant.ID,\n\t\t\t\t\tName: name,\n\t\t\t\t\tState: \"created\",\n\t\t\t\t\tVersion: env.GetLabelVersion(request),\n\t\t\t\t\tType: tenant.GetNamespaceType(name, currentTenant.NsBaseName),\n\t\t\t\t\tMasterURL: masterURL,\n\t\t\t\t})\n\t\t\t} else if env.GetKind(request) == env.ValKindResourceQuota {\n\t\t\t\t// trigger a check status loop\n\t\t\t\ttime.Sleep(time.Millisecond * 50)\n\t\t\t\treturn \"GET\", response\n\t\t\t}\n\t\t\treturn \"\", nil\n\t\t} else if statusCode == http.StatusOK {\n\t\t\tif method == \"DELETE\" {\n\t\t\t\treturn \"POST\", request\n\t\t\t} else if method == \"GET\" {\n\t\t\t\tif env.GetKind(request) == env.ValKindResourceQuota {\n\n\t\t\t\t\tif env.HasValidStatus(response) || atomic.LoadInt32(&currentResourceQuotaStatusCheck) >= maxResourceQuotaStatusCheck {\n\t\t\t\t\t\treturn \"\", nil\n\t\t\t\t\t}\n\t\t\t\t\tatomic.AddInt32(&currentResourceQuotaStatusCheck, 1)\n\t\t\t\t\ttime.Sleep(time.Millisecond * 50)\n\t\t\t\t\treturn \"GET\", response\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn \"\", nil\n\t\t}\n\t\tlog.Info(ctx, map[string]interface{}{\n\t\t\t\"status\": statusCode,\n\t\t\t\"method\": method,\n\t\t\t\"namespace\": env.GetNamespace(request),\n\t\t\t\"cluster_url\": masterURL,\n\t\t\t\"name\": env.GetName(request),\n\t\t\t\"kind\": env.GetKind(request),\n\t\t\t\"request\": yamlString(request),\n\t\t\t\"response\": yamlString(response),\n\t\t}, \"unhandled resource response\")\n\t\treturn \"\", nil\n\t}\n}", "func (o FastTcpAppOutput) Tenant() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *FastTcpApp) pulumi.StringOutput { return v.Tenant }).(pulumi.StringOutput)\n}", "func envOverride(config *DefaultConfig) (*DefaultConfig, error) {\n\t// override UpdateTime\n\tupdateTime := os.Getenv(\"XIGNITE_FEEDER_UPDATE_TIME\")\n\tif updateTime != \"\" {\n\t\tt, err := time.Parse(ctLayout, updateTime)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tconfig.UpdateTime = t\n\t}\n\n\t// override APIToken\n\tapiToken := os.Getenv(\"XIGNITE_FEEDER_API_TOKEN\")\n\tif apiToken != \"\" {\n\t\tconfig.APIToken = apiToken\n\t}\n\n\t// override NotQuoteSymbolList\n\tnotQuoteStockList := os.Getenv(\"XIGNITE_FEEDER_NOT_QUOTE_STOCK_LIST\")\n\tif notQuoteStockList != \"\" {\n\t\tconfig.NotQuoteStockList = strings.Split(notQuoteStockList, \",\")\n\t}\n\n\treturn config, nil\n}", "func (nu *NodeUpdate) SetTenant(s string) *NodeUpdate {\n\tnu.mutation.SetTenant(s)\n\treturn nu\n}", "func (db *MySQLDB) UpdateTenant(ctx context.Context, tenant *Tenant) error {\n\tfLog := mysqlLog.WithField(\"func\", \"UpdateTenant\").WithField(\"RequestID\", ctx.Value(constants.RequestID))\n\n\texist, err := db.IsTenantRecIDExist(ctx, tenant.RecID)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif !exist {\n\t\treturn ErrNotFound\n\t}\n\n\torigin, err := db.GetTenantByRecID(ctx, tenant.RecID)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdomainChanged := origin.Domain != tenant.Domain\n\n\tq := \"UPDATE HANSIP_TENANT SET TENANT_NAME=?, TENANT_DOMAIN=?, DESCRIPTION=? WHERE REC_ID=?\"\n\t_, err = db.instance.ExecContext(ctx, q,\n\t\ttenant.Name, tenant.Domain, tenant.Description, tenant.RecID)\n\tif err != nil {\n\t\tfLog.Errorf(\"db.instance.ExecContext got %s. SQL = %s\", err.Error(), q)\n\t\treturn &ErrDBExecuteError{\n\t\t\tWrapped: err,\n\t\t\tMessage: \"Error UpdateTenant\",\n\t\t\tSQL: q,\n\t\t}\n\t}\n\n\tif domainChanged {\n\t\tq = \"UPDATE HANSIP_ROLE SET ROLE_DOMAIN=? WHERE ROLE_DOMAIN=?\"\n\t\t_, err = db.instance.ExecContext(ctx, q,\n\t\t\ttenant.Domain, origin.Domain)\n\t\tif err != nil {\n\t\t\tfLog.Errorf(\"db.instance.ExecContext got %s. SQL = %s\", err.Error(), q)\n\t\t\treturn &ErrDBExecuteError{\n\t\t\t\tWrapped: err,\n\t\t\t\tMessage: \"Error UpdateTenant\",\n\t\t\t\tSQL: q,\n\t\t\t}\n\t\t}\n\n\t\tq = \"UPDATE HANSIP_GROUP SET GROUP_DOMAIN=? WHERE GROUP_DOMAIN=?\"\n\t\t_, err = db.instance.ExecContext(ctx, q,\n\t\t\ttenant.Domain, origin.Domain)\n\t\tif err != nil {\n\t\t\tfLog.Errorf(\"db.instance.ExecContext got %s. SQL = %s\", err.Error(), q)\n\t\t\treturn &ErrDBExecuteError{\n\t\t\t\tWrapped: err,\n\t\t\t\tMessage: \"Error UpdateTenant\",\n\t\t\t\tSQL: q,\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func UpdateTestTenantsWithHeaders(t *testing.T, header http.Header) {\n\topts := client.NewRequestOptions()\n\topts.Header = header\n\n\t// Retrieve the Tenant by name so we can get the id for the Update\n\tname := \"tenant2\"\n\topts.QueryParameters.Set(\"name\", name)\n\tresp, _, err := TOSession.GetTenants(opts)\n\tif err != nil {\n\t\tt.Errorf(\"cannot get Tenants filtered by name '%s': %v - alerts: %+v\", name, err, resp.Alerts)\n\t}\n\tif len(resp.Response) != 1 {\n\t\tt.Fatalf(\"Expected exactly one Tenant to exist with the name 'tenant2', found: %d\", len(resp.Response))\n\t}\n\tmodTenant := resp.Response[0]\n\n\tparentName := \"tenant1\"\n\topts.QueryParameters.Set(\"name\", parentName)\n\tresp, _, err = TOSession.GetTenants(opts)\n\tif err != nil {\n\t\tt.Errorf(\"cannot get Tenants filtered by name '%s': %v - alerts: %+v\", parentName, err, resp.Alerts)\n\t}\n\tif len(resp.Response) != 1 {\n\t\tt.Fatalf(\"Expected exactly one Tenant to exist with the name 'tenant1', found: %d\", len(resp.Response))\n\t}\n\tnewParent := resp.Response[0]\n\n\tmodTenant.ParentID = newParent.ID\n\topts.QueryParameters.Del(\"name\")\n\t_, reqInf, err := TOSession.UpdateTenant(modTenant.ID, modTenant, opts)\n\tif err == nil {\n\t\tt.Fatalf(\"expected a precondition failed error, got none\")\n\t}\n\tif reqInf.StatusCode != http.StatusPreconditionFailed {\n\t\tt.Errorf(\"expected a status 412 Precondition Failed, but got %d\", reqInf.StatusCode)\n\t}\n}", "func (config *Config) SetPw(cmd *cobra.Command) error {\n\n\tfmt.Print(\"Password: \")\n\t// syscall.Stdin is not 0 on windows\n\tpw, err := terminal.ReadPassword(int(syscall.Stdin))\n\t// pw, err := terminal.ReadPassword(0)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"setPw(ReadPassword)\")\n\t}\n\n\ttenants, err := cmd.Flags().GetString(\"tenant\")\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"setPw(GetString)\")\n\t}\n\n\tconfig.Secret, err = config.AddSecret(tenants, pw)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"setPw(newSecret)\")\n\t}\n\n\tviper.Set(\"secret\", config.Secret)\n\terr = viper.WriteConfig()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"setPw(WriteConfig)\")\n\t}\n\n\t// connection test for all tenants\n\tsecretMap, err := config.GetSecretMap()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"prepare(getSecretMap)\")\n\t}\n\tfor i := range config.Tenants {\n\t\tdb := config.getConnection(i, secretMap)\n\t\tif db == nil {\n\t\t\tcontinue\n\t\t}\n\t\tdb.Close()\n\t}\n\n\treturn nil\n}", "func (o DataConnectorDynamics365Output) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *DataConnectorDynamics365) pulumi.StringOutput { return v.TenantId }).(pulumi.StringOutput)\n}", "func InitTenant(ctx context.Context, masterURL string, service tenant.Service, currentTenant *tenant.Tenant) openshift.Callback {\n\tvar maxResourceQuotaStatusCheck int32 = 50 // technically a global retry count across all ResourceQuota on all Tenant Namespaces\n\tvar currentResourceQuotaStatusCheck int32 = 0\n\treturn func(statusCode int, method string, request, response map[interface{}]interface{}) (string, map[interface{}]interface{}) {\n\t\tlog.Info(ctx, map[string]interface{}{\n\t\t\t\"status\": statusCode,\n\t\t\t\"method\": method,\n\t\t\t\"cluster_url\": masterURL,\n\t\t\t\"namespace\": openshift.GetNamespace(request),\n\t\t\t\"name\": openshift.GetName(request),\n\t\t\t\"kind\": openshift.GetKind(request),\n\t\t\t\"request\": yamlString(request),\n\t\t\t\"response\": yamlString(response),\n\t\t}, \"resource requested\")\n\t\tif statusCode == http.StatusConflict {\n\t\t\tif openshift.GetKind(request) == openshift.ValKindNamespace {\n\t\t\t\treturn \"\", nil\n\t\t\t}\n\t\t\tif openshift.GetKind(request) == openshift.ValKindProjectRequest {\n\t\t\t\treturn \"\", nil\n\t\t\t}\n\t\t\tif openshift.GetKind(request) == openshift.ValKindPersistenceVolumeClaim {\n\t\t\t\treturn \"\", nil\n\t\t\t}\n\t\t\tif openshift.GetKind(request) == openshift.ValKindServiceAccount {\n\t\t\t\treturn \"\", nil\n\t\t\t}\n\t\t\treturn \"DELETE\", request\n\t\t} else if statusCode == http.StatusCreated {\n\t\t\tif openshift.GetKind(request) == openshift.ValKindProjectRequest {\n\t\t\t\tname := openshift.GetName(request)\n\t\t\t\tservice.SaveNamespace(&tenant.Namespace{\n\t\t\t\t\tTenantID: currentTenant.ID,\n\t\t\t\t\tName: name,\n\t\t\t\t\tState: \"created\",\n\t\t\t\t\tVersion: openshift.GetLabelVersion(request),\n\t\t\t\t\tType: tenant.GetNamespaceType(name),\n\t\t\t\t\tMasterURL: masterURL,\n\t\t\t\t})\n\n\t\t\t\t// HACK to workaround osio applying some dsaas-user permissions async\n\t\t\t\t// Should loop on a Check if allowed type of call instead\n\t\t\t\ttime.Sleep(time.Second * 5)\n\n\t\t\t} else if openshift.GetKind(request) == openshift.ValKindNamespace {\n\t\t\t\tname := openshift.GetName(request)\n\t\t\t\tservice.SaveNamespace(&tenant.Namespace{\n\t\t\t\t\tTenantID: currentTenant.ID,\n\t\t\t\t\tName: name,\n\t\t\t\t\tState: \"created\",\n\t\t\t\t\tVersion: openshift.GetLabelVersion(request),\n\t\t\t\t\tType: tenant.GetNamespaceType(name),\n\t\t\t\t\tMasterURL: masterURL,\n\t\t\t\t})\n\t\t\t} else if openshift.GetKind(request) == openshift.ValKindResourceQuota {\n\t\t\t\t// trigger a check status loop\n\t\t\t\ttime.Sleep(time.Millisecond * 50)\n\t\t\t\treturn \"GET\", response\n\t\t\t}\n\t\t\treturn \"\", nil\n\t\t} else if statusCode == http.StatusOK {\n\t\t\tif method == \"DELETE\" {\n\t\t\t\treturn \"POST\", request\n\t\t\t} else if method == \"GET\" {\n\t\t\t\tif openshift.GetKind(request) == openshift.ValKindResourceQuota {\n\n\t\t\t\t\tif openshift.HasValidStatus(response) || atomic.LoadInt32(&currentResourceQuotaStatusCheck) >= maxResourceQuotaStatusCheck {\n\t\t\t\t\t\treturn \"\", nil\n\t\t\t\t\t}\n\t\t\t\t\tatomic.AddInt32(&currentResourceQuotaStatusCheck, 1)\n\t\t\t\t\ttime.Sleep(time.Millisecond * 50)\n\t\t\t\t\treturn \"GET\", response\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn \"\", nil\n\t\t}\n\t\tlog.Info(ctx, map[string]interface{}{\n\t\t\t\"status\": statusCode,\n\t\t\t\"method\": method,\n\t\t\t\"namespace\": openshift.GetNamespace(request),\n\t\t\t\"cluster_url\": masterURL,\n\t\t\t\"name\": openshift.GetName(request),\n\t\t\t\"kind\": openshift.GetKind(request),\n\t\t\t\"request\": yamlString(request),\n\t\t\t\"response\": yamlString(response),\n\t\t}, \"unhandled resource response\")\n\t\treturn \"\", nil\n\t}\n}", "func (c *TenantController) Setup(ctx *app.SetupTenantContext) error {\n\tuserToken := goajwt.ContextJWT(ctx)\n\tif userToken == nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewUnauthorizedError(\"Missing JWT token\"))\n\t}\n\tttoken := &TenantToken{token: userToken}\n\texists := c.tenantService.Exists(ttoken.Subject())\n\tif exists {\n\t\treturn ctx.Conflict()\n\t}\n\n\t// fetch the cluster the user belongs to\n\tuser, err := c.userService.GetUser(ctx, ttoken.Subject())\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, err)\n\t}\n\n\tif user.Cluster == nil {\n\t\tlog.Error(ctx, nil, \"no cluster defined for tenant\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewInternalError(ctx, fmt.Errorf(\"unable to provision to undefined cluster\")))\n\t}\n\n\t// fetch the users cluster token\n\topenshiftUsername, openshiftUserToken, err := c.resolveTenant(ctx, *user.Cluster, userToken.Raw)\n\tif err != nil {\n\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\"err\": err,\n\t\t\t\"cluster_url\": *user.Cluster,\n\t\t}, \"unable to fetch tenant token from auth\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewUnauthorizedError(\"Could not resolve user token\"))\n\t}\n\n\t// fetch the cluster info\n\tcluster, err := c.resolveCluster(ctx, *user.Cluster)\n\tif err != nil {\n\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\"err\": err,\n\t\t\t\"cluster_url\": *user.Cluster,\n\t\t}, \"unable to fetch cluster\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewInternalError(ctx, err))\n\t}\n\n\t// create openshift config\n\topenshiftConfig := openshift.NewConfig(c.defaultOpenshiftConfig, user, cluster.User, cluster.Token, cluster.APIURL)\n\ttenant := &tenant.Tenant{ID: ttoken.Subject(), Email: ttoken.Email()}\n\terr = c.tenantService.CreateTenant(tenant)\n\tif err != nil {\n\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\"err\": err,\n\t\t}, \"unable to store tenant configuration\")\n\t\treturn ctx.Conflict()\n\t}\n\n\tgo func() {\n\t\tctx := ctx\n\t\tt := tenant\n\t\terr = openshift.RawInitTenant(\n\t\t\tctx,\n\t\t\topenshiftConfig,\n\t\t\tInitTenant(ctx, openshiftConfig.MasterURL, c.tenantService, t),\n\t\t\topenshiftUsername,\n\t\t\topenshiftUserToken,\n\t\t\tc.templateVars)\n\n\t\tif err != nil {\n\t\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\t\"err\": err,\n\t\t\t\t\"os_user\": openshiftUsername,\n\t\t\t}, \"unable initialize tenant\")\n\t\t}\n\t}()\n\n\tctx.ResponseData.Header().Set(\"Location\", rest.AbsoluteURL(ctx.RequestData.Request, app.TenantHref()))\n\treturn ctx.Accepted()\n}", "func (o ManagedInstanceActiveDirectoryAdministratorOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *ManagedInstanceActiveDirectoryAdministrator) pulumi.StringOutput { return v.TenantId }).(pulumi.StringOutput)\n}", "func (uuo *UserUpdateOne) SetTenant(s string) *UserUpdateOne {\n\tuuo.mutation.SetTenant(s)\n\treturn uuo\n}", "func setTincHostsConf(log *logging.Logger, i ClusterInstance, vpnName, tincName, content string) error {\n\tconfDir := path.Join(\"/etc/tinc\", vpnName, \"hosts\")\n\tconfPath := path.Join(confDir, tincName)\n\tif _, err := i.runRemoteCommand(log, fmt.Sprintf(\"sudo mkdir -p %s\", confDir), \"\", false); err != nil {\n\t\treturn maskAny(err)\n\t}\n\tif _, err := i.runRemoteCommand(log, fmt.Sprintf(\"sudo tee %s\", confPath), content, false); err != nil {\n\t\treturn maskAny(err)\n\t}\n\treturn nil\n}", "func (o AzureDataLakeStoreOutputDataSourceOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v AzureDataLakeStoreOutputDataSource) *string { return v.TenantId }).(pulumi.StringPtrOutput)\n}", "func Set(c *deis.Client, app string, config api.Config) (api.Config, error) {\n\tbody, err := json.Marshal(config)\n\n\tif err != nil {\n\t\treturn api.Config{}, err\n\t}\n\n\tu := fmt.Sprintf(\"/v2/apps/%s/config/\", app)\n\n\tres, reqErr := c.Request(\"POST\", u, body)\n\tif reqErr != nil {\n\t\treturn api.Config{}, reqErr\n\t}\n\tdefer res.Body.Close()\n\n\tnewConfig := api.Config{}\n\tif err = json.NewDecoder(res.Body).Decode(&newConfig); err != nil {\n\t\treturn api.Config{}, err\n\t}\n\n\treturn newConfig, reqErr\n}", "func SetShortPathRequestTenant(req *http.Request) (*http.Request, error) {\n\n\tctx := req.Context()\n\n\trequestor, exists := request.UserFrom(ctx)\n\tif !exists {\n\t\treturn nil, errors.New(\"The user info is missing.\")\n\t}\n\n\tuserTenant := requestor.GetTenant()\n\tif userTenant == metav1.TenantNone {\n\t\t// temporary workaround\n\t\t// tracking issue: https://github.com/futurewei-cloud/arktos/issues/102\n\t\tuserTenant = metav1.TenantSystem\n\t\t//When https://github.com/futurewei-cloud/arktos/issues/102 is done, remove the above line\n\t\t// and enable the following two lines.\n\t\t//responsewriters.InternalError(w, req, errors.New(fmt.Sprintf(\"The tenant in the user info of %s is empty. \", requestor.GetName())))\n\t\t//return\n\t}\n\n\trequestInfo, exists := request.RequestInfoFrom(ctx)\n\tif !exists {\n\t\treturn nil, errors.New(\"The request info is missing.\")\n\t}\n\n\t// for a reqeust from a regular user, if the tenant in the object is empty, use the tenant from user info\n\t// this is what we call \"short-path\", which allows users to use traditional Kubernets API in the multi-tenancy Arktos\n\tresourceTenant := requestInfo.Tenant\n\tif resourceTenant == metav1.TenantNone && userTenant != metav1.TenantSystem {\n\t\trequestInfo.Tenant = userTenant\n\t}\n\n\treq = req.WithContext(request.WithRequestInfo(ctx, requestInfo))\n\n\treturn req, nil\n}", "func (o MonitorV1Output) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *MonitorV1) pulumi.StringOutput { return v.TenantId }).(pulumi.StringOutput)\n}", "func (t *Tenant) SetDefaults() bool {\n\tchanged := false\n\tts := &t.Spec\n\tif ts.TenantSecretRef.Name == \"\" {\n\t\tts.TenantSecretRef.Name = fmt.Sprintf(\"%s-%s\", strings.ToLower(t.Name), strings.ToLower(t.Spec.OrganizationName))\n\t\tchanged = true\n\t}\n\tif ts.TenantSecretRef.Namespace == \"\" {\n\t\tts.TenantSecretRef.Namespace = t.Namespace\n\t\tchanged = true\n\t}\n\treturn changed\n}", "func getCurrentTenant() *Tenant {\n\tif currentTenant == nil {\n\t\ttenants, err := iaas.GetTenantNames()\n\t\tif err != nil || len(tenants) != 1 {\n\t\t\treturn nil\n\t\t}\n\t\t// Set unique tenant as selected\n\t\tlog.Println(\"Unique tenant set\")\n\t\tfor name := range tenants {\n\t\t\tservice, err := iaas.UseService(name)\n\t\t\tif err != nil {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\tcurrentTenant = &Tenant{name: name, Service: service}\n\t\t}\n\t}\n\treturn currentTenant\n}", "func (ten *Tenant) Commit() error {\n\tif ten.HasError() {\n\t\treturn ten.err\n\t}\n\n\terr := ten.Client.CreateTenant(ten.Tenant)\n\tif err != nil {\n\t\tten.err = err\n\t\tlog.Infof(\"Creating or updating Tenant failed %v\", err)\n\t\treturn err\n\t}\n\n\tlog.Debugf(\"Created VPC : %#v\", ten)\n\n\treturn nil\n}", "func InitTenant(ctx context.Context, config tenantConfig) error {\n\n\tc, err := createClient(ctx, config)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Ignore response for now\n\t_, err = c.SetupTenant(goasupport.ForwardContextRequestID(ctx), tenant.SetupTenantPath())\n\n\treturn err\n}", "func overrideConfig(config1 *Config, config2 *Config) {\n\n}", "func (gc *GroupCreate) SetTenant(s string) *GroupCreate {\n\tgc.mutation.SetTenant(s)\n\treturn gc\n}", "func (i *awsIamKubeAuthPlugin) persistConfig() (err error) {\n\tcache := map[string]string{\n\t\tClusterIdName: i.o.ClusterID,\n\t\tRegionName: i.o.Region,\n\t\tRoleName: i.o.AssumeRoleARN,\n\t\tExternalIdName: i.o.AssumeRoleExternalID,\n\t\tForwardSessionName: strconv.FormatBool(i.forwardSession),\n\t\tSessionName: i.o.SessionName,\n\t}\n\n\tif i.t.Token != \"\" && i.t.Expiration.After(time.Now()) {\n\t\tcache[AccessTokenName] = i.t.Token\n\t\tcache[ExpiryName] = i.t.Expiration.Format(time.RFC3339Nano)\n\t}\n\n\tif i.persister != nil {\n\t\terr = i.persister.Persist(cache)\n\t}\n\n\tif err != nil {\n\t\tklog.Errorf(\"failed to persist aws-iam identity: %v\", err)\n\t}\n\n\treturn\n}", "func (m *SmsLogRow) SetTenantCountryCode(value *string)() {\n err := m.GetBackingStore().Set(\"tenantCountryCode\", value)\n if err != nil {\n panic(err)\n }\n}", "func (o DataConnectorOfficePowerBiOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *DataConnectorOfficePowerBi) pulumi.StringOutput { return v.TenantId }).(pulumi.StringOutput)\n}", "func (cfg *Conf) overrideEnvConf() {\n\tif os.Getenv(\"SMGMG_BK_API_KEY\") != \"\" {\n\t\tcfg.ApiKey = os.Getenv(\"SMGMG_BK_API_KEY\")\n\t}\n\n\tif os.Getenv(\"SMGMG_BK_API_SECRET\") != \"\" {\n\t\tcfg.ApiSecret = os.Getenv(\"SMGMG_BK_API_SECRET\")\n\t}\n\n\tif os.Getenv(\"SMGMG_BK_USER_TOKEN\") != \"\" {\n\t\tcfg.UserToken = os.Getenv(\"SMGMG_BK_USER_TOKEN\")\n\t}\n\n\tif os.Getenv(\"SMGMG_BK_USER_SECRET\") != \"\" {\n\t\tcfg.UserSecret = os.Getenv(\"SMGMG_BK_USER_SECRET\")\n\t}\n\n\tif os.Getenv(\"SMGMG_BK_DESTINATION\") != \"\" {\n\t\tcfg.Destination = os.Getenv(\"SMGMG_BK_DESTINATION\")\n\t}\n\n\tif os.Getenv(\"SMGMG_BK_FILE_NAMES\") != \"\" {\n\t\tcfg.Filenames = os.Getenv(\"SMGMG_BK_FILE_NAMES\")\n\t}\n}", "func (d *DB) TenantDB(schema string) *pg.DB {\n\treturn d.tenantDB.WithParam(\"schema\", pg.Ident(schema)).WithContext(context.WithValue(context.Background(), KeySchema, schema))\n}", "func (m *SecurityRequestBuilder) ProviderTenantSettings()(*ProviderTenantSettingsRequestBuilder) {\n return NewProviderTenantSettingsRequestBuilderInternal(m.BaseRequestBuilder.PathParameters, m.BaseRequestBuilder.RequestAdapter)\n}", "func (o AzureDataLakeStoreOutputDataSourceResponseOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v AzureDataLakeStoreOutputDataSourceResponse) *string { return v.TenantId }).(pulumi.StringPtrOutput)\n}", "func (o NetworkManagerScopeConnectionOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *NetworkManagerScopeConnection) pulumi.StringOutput { return v.TenantId }).(pulumi.StringOutput)\n}", "func (api *tenantAPI) SyncCreate(obj *cluster.Tenant) error {\n\tnewObj := obj\n\tevtType := kvstore.Created\n\tvar writeErr error\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\tnewObj, writeErr = apicl.ClusterV1().Tenant().Create(context.Background(), obj)\n\t\tif writeErr != nil && strings.Contains(writeErr.Error(), \"AlreadyExists\") {\n\t\t\tnewObj, writeErr = apicl.ClusterV1().Tenant().Update(context.Background(), obj)\n\t\t\tevtType = kvstore.Updated\n\t\t}\n\t}\n\n\tif writeErr == nil {\n\t\tapi.ct.handleTenantEvent(&kvstore.WatchEvent{Object: newObj, Type: evtType})\n\t}\n\treturn writeErr\n}", "func (ctrler CtrlDefReactor) OnTenantUpdate(oldObj *Tenant, newObj *cluster.Tenant) error {\n\tlog.Info(\"OnTenantUpdate is not implemented\")\n\treturn nil\n}", "func (uu *UserUpdate) SetTenant(s string) *UserUpdate {\n\tuu.mutation.SetTenant(s)\n\treturn uu\n}", "func (client *Client) UpdateTenantConfig(ID string, config types.TenantConfig) error {\n\turl, err := client.getCiaoTenantRef(ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar oldconfig types.TenantConfig\n\terr = client.getResource(url, api.TenantsV1, nil, &oldconfig)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ta, err := json.Marshal(oldconfig)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif config.Name == \"\" {\n\t\tconfig.Name = oldconfig.Name\n\t}\n\n\tif config.SubnetBits == 0 {\n\t\tconfig.SubnetBits = oldconfig.SubnetBits\n\t}\n\n\tb, err := json.Marshal(config)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tmerge, err := jsonpatch.CreateMergePatch(a, b)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tbody := bytes.NewReader(merge)\n\n\tresp, err := client.sendHTTPRequest(\"PATCH\", url, nil, body, \"merge-patch+json\")\n\tdefer resp.Body.Close()\n\n\treturn err\n}", "func getCurrentTenant() *Tenant {\n\tif currentTenant == nil {\n\t\ttenants, err := providers.Tenants()\n\t\tif err != nil || len(tenants) != 1 {\n\t\t\treturn nil\n\t\t}\n\t\t// Set unqiue tenant as selected\n\t\tlog.Println(\"Unique tenant set\")\n\t\tfor name := range tenants {\n\t\t\tservice, err := providers.GetService(name)\n\t\t\tif err != nil {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\tcurrentTenant = &Tenant{name: name, Service: service}\n\t\t}\n\t}\n\treturn currentTenant\n}", "func (api *tenantAPI) Delete(obj *cluster.Tenant) error {\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = apicl.ClusterV1().Tenant().Delete(context.Background(), &obj.ObjectMeta)\n\t\treturn err\n\t}\n\n\tapi.ct.handleTenantEvent(&kvstore.WatchEvent{Object: obj, Type: kvstore.Deleted})\n\treturn nil\n}", "func TenantResourceQuota(tenant string) string {\n\treturn tenant\n}", "func (suite *TenantTestSuite) TestUpdateTenant() {\n\n\t// create json input data for the request\n\tputData := `\n {\n \"info\":{\n\t\t\t\t\"name\":\"new_mutants\",\n\t\t\t\t\"email\":\"yo@yo\",\n\t\t\t\t\"website\":\"website\"\n\t\t\t},\n \"db_conf\": [\n {\n \"store\":\"ar\",\n \"server\":\"localhost\",\n \"port\":27017,\n \"database\":\"ar_db\",\n \"username\":\"admin\",\n \"password\":\"3NCRYPT3D\"\n },\n {\n \"store\":\"status\",\n \"server\":\"localhost\",\n \"port\":27017,\n \"database\":\"status_db\",\n \"username\":\"admin\",\n \"password\":\"3NCRYPT3D\"\n }],\n \"users\": [\n {\n \"name\":\"xavier\",\n \"email\":\"[email protected]\",\n \"api_key\":\"X4V13R\"\n },\n {\n \"name\":\"magneto\",\n \"email\":\"[email protected]\",\n \"api_key\":\"M4GN3T0\"\n }]\n }`\n\n\tjsonOutput := `{\n \"status\": {\n \"message\": \"Tenant successfully updated\",\n \"code\": \"200\"\n }\n}`\n\n\trequest, _ := http.NewRequest(\"PUT\", \"/api/v2/admin/tenants/6ac7d684-1f8e-4a02-a502-720e8f11e50c\", strings.NewReader(putData))\n\trequest.Header.Set(\"x-api-key\", suite.clientkey)\n\trequest.Header.Set(\"Accept\", \"application/json\")\n\tresponse := httptest.NewRecorder()\n\n\tsuite.router.ServeHTTP(response, request)\n\n\tcode := response.Code\n\toutput := response.Body.String()\n\n\tsuite.Equal(200, code, \"Internal Server Error\")\n\t// Compare the expected and actual xml response\n\tsuite.Equal(jsonOutput, output, \"Response body mismatch\")\n\n}", "func (o FluxConfigurationBlobStorageServicePrincipalOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v FluxConfigurationBlobStorageServicePrincipal) string { return v.TenantId }).(pulumi.StringOutput)\n}", "func (o UpdateHistoryPropertyResponseOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v UpdateHistoryPropertyResponse) string { return v.TenantId }).(pulumi.StringOutput)\n}", "func createDefaultTenant() {\n\t// tenant params\n\ttenant := client.Tenant{\n\t\tTenantName: \"default\",\n\t}\n\n\t// create a tenant\n\terr := contivClient.TenantPost(&tenant)\n\tif err != nil {\n\t\tlog.Fatalf(\"Error creating default tenant. Err: %v\", err)\n\t}\n\n\t// Get the tenant and verify it exists\n\tgotTenant, err := contivClient.TenantGet(\"default\")\n\tif err != nil {\n\t\tlog.Fatalf(\"Error getting default tenant. Err: %v\", err)\n\t}\n\n\tif gotTenant.TenantName != tenant.TenantName {\n\t\tlog.Fatalf(\"Got invalid tenant name. expecting %s. Got %s\", tenant.TenantName, gotTenant.TenantName)\n\t}\n}", "func updateHydrant(addr string, hydrant *models.Hydrant) {\n\taddr = strings.ToLower(addr)\n\tif err := hydrant.InsertOrUpdate(); err != nil {\n\t\tmsg := fmt.Sprintf(\"Failed to update account: %s, error: %s\", addr, err.Error())\n\t\tbeego.Error(msg)\n\t}\n}", "func (me *CONFIGURATION_IMPL) SetUserId(userId string) {\r\n me.user-id = userId\r\n}", "func (o NetworkOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Network) pulumi.StringOutput { return v.TenantId }).(pulumi.StringOutput)\n}", "func (policy *PolicySvc) augmentEndpoint(endpoint *common.Endpoint) error {\n\ttenantSvcUrl, err := policy.client.GetServiceUrl(\"tenant\")\n\tif err != nil {\n\t\treturn err\n\t}\n\tif endpoint.Peer == common.Wildcard {\n\t\t// If a wildcard is specfied, there is nothing to augment\n\t\treturn nil\n\t}\n\tlog.Printf(\"Policy: Augmenting %#v\", endpoint)\n\n\t// Code below tries to resolve tenant name into tenant_network_id if possible.\n\t//\n\t// TODO this will have to be changed once we implement\n\t// https://paninetworks.kanbanize.com/ctrl_board/3/cards/319/details\n\tten := &tenant.Tenant{}\n\tif endpoint.TenantNetworkID == nil {\n\t\tif endpoint.TenantID != 0 {\n\t\t\ttenantIDToUse := strconv.FormatUint(endpoint.TenantID, 10)\n\t\t\ttenantsUrl := fmt.Sprintf(\"%s/tenants/%s\", tenantSvcUrl, tenantIDToUse)\n\t\t\tlog.Printf(\"Policy: Looking tenant up at %s\", tenantsUrl)\n\t\t\terr = policy.client.Get(tenantsUrl, ten)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tendpoint.TenantNetworkID = &ten.NetworkID\n\n\t\t} else if endpoint.TenantExternalID != \"\" || endpoint.TenantName != \"\" {\n\t\t\tif endpoint.TenantExternalID != \"\" {\n\t\t\t\tten.ExternalID = endpoint.TenantExternalID\n\t\t\t}\n\t\t\tif endpoint.TenantName != \"\" {\n\t\t\t\tten.Name = endpoint.TenantName\n\t\t\t}\n\t\t\terr = policy.client.Find(ten, common.FindLast)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tendpoint.TenantNetworkID = &ten.NetworkID\n\t\t}\n\t}\n\n\tif endpoint.SegmentNetworkID == nil {\n\t\tif ten == nil && (endpoint.SegmentID != 0 || endpoint.SegmentExternalID != \"\" || endpoint.SegmentName != \"\") {\n\t\t\treturn common.NewError400(\"No tenant information specified, cannot look up segment.\")\n\t\t}\n\t\tsegment := &tenant.Segment{}\n\t\tif endpoint.SegmentID != 0 {\n\t\t\tsegmentIDToUse := strconv.FormatUint(endpoint.SegmentID, 10)\n\t\t\tsegmentsUrl := fmt.Sprintf(\"%s/tenants/%d/segments/%s\", tenantSvcUrl, ten.ID, segmentIDToUse)\n\t\t\tlog.Printf(\"Policy: Looking segment up at %s for %#v\", segmentsUrl, endpoint)\n\t\t\terr = policy.client.Get(segmentsUrl, &segment)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tendpoint.SegmentNetworkID = &segment.NetworkID\n\t\t} else if endpoint.SegmentExternalID != \"\" || endpoint.SegmentName != \"\" {\n\t\t\tsegmentsUrl := fmt.Sprintf(\"%s/findLast/segments?tenant_id=%d&\", tenantSvcUrl, ten.ID)\n\t\t\tif endpoint.SegmentExternalID != \"\" {\n\t\t\t\tsegmentsUrl += \"external_id=\" + endpoint.TenantExternalID + \"&\"\n\t\t\t}\n\t\t\tif endpoint.SegmentName != \"\" {\n\t\t\t\tsegmentsUrl += \"name=\" + endpoint.SegmentName\n\t\t\t}\n\t\t\tlog.Printf(\"Policy: Finding segments at %s for %#v (Tenant %#v %t)\", segmentsUrl, endpoint, ten, ten == nil)\n\t\t\terr = policy.client.Get(segmentsUrl, &segment)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tendpoint.SegmentNetworkID = &segment.NetworkID\n\t\t}\n\t}\n\treturn nil\n}", "func envOverride(config *Config) error {\n\tconst defaultPort = \":3000\"\n\terr := envconfig.Process(\"athens\", config)\n\tif err != nil {\n\t\treturn err\n\t}\n\tportEnv := os.Getenv(\"PORT\")\n\t// ATHENS_PORT takes precedence over PORT\n\tif portEnv != \"\" && os.Getenv(\"ATHENS_PORT\") == \"\" {\n\t\tconfig.Port = portEnv\n\t}\n\tif config.Port == \"\" {\n\t\tconfig.Port = defaultPort\n\t}\n\tconfig.Port = ensurePortFormat(config.Port)\n\treturn nil\n}", "func setReplicas(kogitoApp *v1alpha1.KogitoApp, dc *appsv1.DeploymentConfig) {\n\treplicas := defaultReplicas\n\tif kogitoApp.Spec.Replicas != nil {\n\t\treplicas = *kogitoApp.Spec.Replicas\n\t}\n\tdc.Spec.Replicas = replicas\n}", "func deploySubAccountOverrideToContext(ctx context.Context, subAccount string) context.Context {\n\treturn context.WithValue(ctx, contextSubAccountFlag, subAccount)\n}", "func setConfigFile(etcdConf *embetcd.Config, max int, min int, check int) (configFilePath string, tempDir string) {\n\t// get a temporary directory for the etcd data directory\n\ttempDir, err := ioutil.TempDir(\"\", \"TestProxyCluster\")\n\tSo(err, ShouldBeNil)\n\tSo(os.RemoveAll(tempDir), ShouldBeNil)\n\n\t// get a temporary filename for the config file\n\tfileObj, err := ioutil.TempFile(\"\", \"TestProxyClusterConfig\")\n\tSo(err, ShouldBeNil)\n\tconfigFilePath = fileObj.Name()\n\t// remove the temp file so we can overwrite it\n\tSo(os.Remove(configFilePath), ShouldBeNil)\n\n\t//// remove the temp dir so we can recreate it\n\n\tproxyConf := configEtcd\n\tproxyConf = strings.Replace(proxyConf, \"<<MAX>>\", strconv.FormatInt(int64(max), 10), -1)\n\tproxyConf = strings.Replace(proxyConf, \"<<MIN>>\", strconv.FormatInt(int64(min), 10), -1)\n\tproxyConf = strings.Replace(proxyConf, \"<<CHECK>>\", strconv.FormatInt(int64(check), 10), -1)\n\tproxyConf = strings.Replace(proxyConf, \"<<LPADDRESS>>\", etcdConf.LPUrls[0].String(), -1)\n\tproxyConf = strings.Replace(proxyConf, \"<<APADDRESS>>\", etcdConf.APUrls[0].String(), -1)\n\tproxyConf = strings.Replace(proxyConf, \"<<LCADDRESS>>\", etcdConf.LCUrls[0].String(), -1)\n\tproxyConf = strings.Replace(proxyConf, \"<<ACADDRESS>>\", etcdConf.ACUrls[0].String(), -1)\n\tproxyConf = strings.Replace(proxyConf, \"<<MADDRESS>>\", etcdConf.ListenMetricsUrls[0].String(), -1)\n\tproxyConf = strings.Replace(proxyConf, \"<<UNHEALTHYTTL>>\", etcdConf.UnhealthyTTL.String(), -1)\n\tproxyConf = strings.Replace(proxyConf, \"<<REMOVEMEMBERTIMEOUT>>\", etcdConf.RemoveMemberTimeout.String(), -1)\n\tproxyConf = strings.Replace(proxyConf, \"<<DATADIR>>\", filepath.Join(tempDir, etcdConf.Dir), -1)\n\tproxyConf = strings.Replace(proxyConf, \"<<CLUSTEROP>>\", etcdConf.ClusterState, -1)\n\tproxyConf = strings.Replace(proxyConf, \"<<TARGETADDRESSES>>\", formatTargetAddresses(etcdConf.InitialCluster), -1)\n\tproxyConf = strings.Replace(proxyConf, \"<<SERVERNAME>>\", etcdConf.Name, -1)\n\tproxyConf = strings.Replace(proxyConf, \"<<CLUSTERNAME>>\", etcdConf.ClusterName, -1)\n\n\tSo(ioutil.WriteFile(path.Join(configFilePath), []byte(proxyConf), os.FileMode(0666)), ShouldBeNil)\n\treturn configFilePath, tempDir\n}", "func fnTenant(ctx Context, doc *JDoc, params []string) interface{} {\n\tstats := ctx.Value(EelTotalStats).(*ServiceStats)\n\tif params == nil || params[0] != \"\" {\n\t\tctx.Log().Error(\"error_type\", \"func_tenant\", \"op\", \"tenant\", \"cause\", \"no_parameters_expected\", \"params\", params)\n\t\tstats.IncErrors()\n\t\tAddError(ctx, SyntaxError{fmt.Sprintf(\"no parameters expected in call to tenant function\"), \"tenant\", params})\n\t\treturn \"\"\n\t}\n\th := GetCurrentHandlerConfig(ctx)\n\tif h == nil {\n\t\tctx.Log().Error(\"error_type\", \"func_tenant\", \"op\", \"tenant\", \"cause\", \"no_handler\", \"params\", params)\n\t\tstats.IncErrors()\n\t\tAddError(ctx, RuntimeError{fmt.Sprintf(\"current handler not found in call to tenant function\"), \"tenant\", params})\n\t\treturn \"\"\n\t}\n\t//Add logic to separating tenant and partner\n\tallowPartner := GetConfig(ctx).AllowPartner\n\tif h.TenantId == \"_default\" {\n\t\t//The default tenant handler is used. We need to drill-down into context to find out the actual tenant\n\t\ttenantHeaderKey := GetConfig(ctx).HttpTenantHeader\n\t\tif ctx.Value(tenantHeaderKey) != nil {\n\t\t\tcombinedTenant := ctx.Value(tenantHeaderKey).(string)\n\t\t\ttenant := \"\"\n\t\t\tif allowPartner && strings.LastIndex(combinedTenant, \"_\") > 0 && strings.LastIndex(combinedTenant, \"_\") < len(combinedTenant)-1 {\n\t\t\t\ttenant = combinedTenant[:strings.LastIndex(combinedTenant, \"_\")]\n\t\t\t} else {\n\t\t\t\ttenant = combinedTenant\n\t\t\t}\n\t\t\treturn tenant\n\t\t}\n\t}\n\tcombinedTenant := h.TenantId\n\ttenant := \"\"\n\tif allowPartner && strings.LastIndex(combinedTenant, \"_\") > 0 && strings.LastIndex(combinedTenant, \"_\") < len(combinedTenant)-1 {\n\t\ttenant = combinedTenant[:strings.LastIndex(combinedTenant, \"_\")]\n\t} else {\n\t\ttenant = combinedTenant\n\t}\n\treturn tenant\n}", "func migrateUsersConfigEtcdToV1(isSTS bool) error {\n\tbasePrefix := iamConfigUsersPrefix\n\tif isSTS {\n\t\tbasePrefix = iamConfigSTSPrefix\n\t}\n\n\tctx, cancel := context.WithTimeout(context.Background(), defaultContextTimeout)\n\tdefer cancel()\n\tr, err := globalEtcdClient.Get(ctx, basePrefix, etcd.WithPrefix(), etcd.WithKeysOnly())\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tusers := etcdKvsToSet(basePrefix, r.Kvs)\n\tfor _, user := range users.ToSlice() {\n\t\t{\n\t\t\t// 1. check if there is a policy file in the old loc.\n\t\t\toldPolicyPath := pathJoin(basePrefix, user, iamPolicyFile)\n\t\t\tvar policyName string\n\t\t\terr := loadIAMConfigItemEtcd(ctx, &policyName, oldPolicyPath)\n\t\t\tif err != nil {\n\t\t\t\tswitch err {\n\t\t\t\tcase errConfigNotFound:\n\t\t\t\t\t// No mapped policy or already migrated.\n\t\t\t\tdefault:\n\t\t\t\t\t// corrupt data/read error, etc\n\t\t\t\t}\n\t\t\t\tgoto next\n\t\t\t}\n\n\t\t\t// 2. copy policy to new loc.\n\t\t\tmp := newMappedPolicy(policyName)\n\t\t\tpath := getMappedPolicyPath(user, isSTS)\n\t\t\tif err := saveIAMConfigItemEtcd(ctx, mp, path); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\t// 3. delete policy file in old loc.\n\t\t\tdeleteConfigEtcd(ctx, globalEtcdClient, oldPolicyPath)\n\t\t}\n\n\tnext:\n\t\t// 4. check if user identity has old format.\n\t\tidentityPath := pathJoin(basePrefix, user, iamIdentityFile)\n\t\tvar cred auth.Credentials\n\t\tif err := loadIAMConfigItemEtcd(ctx, &cred, identityPath); err != nil {\n\t\t\tswitch err {\n\t\t\tcase errConfigNotFound:\n\t\t\t\t// This case should not happen.\n\t\t\tdefault:\n\t\t\t\t// corrupt file or read error\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\n\t\t// If the file is already in the new format,\n\t\t// then the parsed auth.Credentials will have\n\t\t// the zero value for the struct.\n\t\tvar zeroCred auth.Credentials\n\t\tif cred == zeroCred {\n\t\t\t// nothing to do\n\t\t\tcontinue\n\t\t}\n\n\t\t// Found a id file in old format. Copy value\n\t\t// into new format and save it.\n\t\tu := newUserIdentity(cred)\n\t\tif err := saveIAMConfigItemEtcd(ctx, u, identityPath); err != nil {\n\t\t\tlogger.LogIf(context.Background(), err)\n\t\t\treturn err\n\t\t}\n\n\t\t// Nothing to delete as identity file location\n\t\t// has not changed.\n\t}\n\treturn nil\n\n}", "func (o ServiceTenantAccessOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ServiceTenantAccess) *string { return v.TenantId }).(pulumi.StringPtrOutput)\n}", "func (o AzureActiveDirectoryOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v AzureActiveDirectory) *string { return v.TenantId }).(pulumi.StringPtrOutput)\n}", "func (config *Config) AddSecret(tenants string, pw []byte) ([]byte, error) {\n\tvar err error\n\n\tsecret, err := config.GetSecretMap()\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"AddSecret(GetSecretMap)\")\n\t}\n\n\t// create secret key once if it doesn't exist\n\tif _, ok := secret.Name[\"secretkey\"]; !ok {\n\n\t\tsecret.Name = make(map[string][]byte)\n\t\tsecret.Name[\"secretkey\"], err = GetSecretKey()\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"AddSecret(GetSecretKey)\")\n\t\t}\n\t}\n\n\t// encrypt password\n\tencPw, err := PwEncrypt(pw, secret.Name[\"secretkey\"])\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"AddSecret(PwEncrypt)\")\n\t}\n\n\tfor _, tenant := range strings.Split(tenants, \",\") {\n\n\t\t// check, if cmd line tenant exists in configfile\n\t\ttInfo := config.FindTenant(low(tenant))\n\t\tif \"\" == tInfo.Name {\n\t\t\tlog.WithFields(log.Fields{\n\t\t\t\t\"tenant\": low(tenant),\n\t\t\t}).Error(\"missing tenant\")\n\t\t\treturn nil, errors.New(\"Did not find tenant in configfile tenants slice.\")\n\t\t}\n\n\t\t// add password to secret map\n\t\tsecret.Name[low(tenant)] = encPw\n\t}\n\n\t// write pw information back to the config file\n\tnewSecret, err := proto.Marshal(&secret)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"AddSecret(Marshal)\")\n\t}\n\n\treturn newSecret, nil\n}", "func Configure(appCfg *AppConfig, tmcfg *config.Config, initializing bool) {\n\tNoColorFormatting = viper.GetBool(\"no-colors\")\n\n\t// Set default version information\n\tappCfg.VersionInfo = &VersionInfo{}\n\n\t// Setup viper and app directories\n\tsetup(appCfg, tmcfg, initializing)\n\n\t// Tendermint config overwrites\n\tchainInfo := setupTendermintCfg(appCfg, tmcfg)\n\n\t// Setup logger\n\tsetupLogger(appCfg, tmcfg)\n\n\t// Add seed peers if .IgnoreSeeds is false\n\tif !appCfg.Node.IgnoreSeeds {\n\t\ttmcfg.P2P.PersistentPeers = appCfg.Node.PersistentPeers + \",\" + strings.Join(chainInfo.ChainSeedPeers, \",\")\n\t\tappCfg.DHT.BootstrapPeers = appCfg.DHT.BootstrapPeers + \",\" + strings.Join(chainInfo.DHTSeedPeers, \",\")\n\t}\n\n\tif appCfg.DHT.Address != \"\" && appCfg.DHT.Address[:1] == \":\" {\n\t\tappCfg.DHT.Address = \"0.0.0.0\" + appCfg.DHT.Address\n\t}\n\n\tif appCfg.RPC.User == \"\" && appCfg.RPC.Password == \"\" {\n\t\tappCfg.RPC.DisableAuth = true\n\t}\n\n\tif tmcfg.P2P.ListenAddress != \"\" && tmcfg.P2P.ListenAddress[:1] == \":\" {\n\t\ttmcfg.P2P.ListenAddress = \"0.0.0.0\" + tmcfg.P2P.ListenAddress\n\t}\n\n\tappCfg.G().Bus = emitter.New(10000)\n\tappCfg.G().TMConfig = tmcfg\n}", "func TenantNamespace(tenant string) string {\n\treturn fmt.Sprintf(\"cyclone--%s\", tenant)\n}", "func CreateConfig(serviceAccount *corev1.ServiceAccount) string {\n\tclientset, err := authorization.CreateClientSet()\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\tpanic(err.Error())\n\t}\n\t// To find out the secret name to use\n\taccountSecretName := \"\"\n\tfor _, accountSecret := range serviceAccount.Secrets {\n\t\tmatch, _ := regexp.MatchString(\"([a-z0-9]+)-token-([a-z0-9]+)\", accountSecret.Name)\n\t\tif match {\n\t\t\taccountSecretName = accountSecret.Name\n\t\t\tbreak\n\t\t}\n\t}\n\t// If there is no matching secret terminate this function as generating kubeconfig file is not possible\n\tif accountSecretName == \"\" {\n\t\tlog.Printf(\"Serviceaccount %s in %s doesn't have a serviceaccount token\", serviceAccount.GetName(), serviceAccount.GetNamespace())\n\t\treturn fmt.Sprintf(\"Serviceaccount %s doesn't have a serviceaccount token\\n\", serviceAccount.GetName())\n\t}\n\tsecret, err := clientset.CoreV1().Secrets(serviceAccount.GetNamespace()).Get(accountSecretName, metav1.GetOptions{})\n\tif errors.IsNotFound(err) {\n\t\tlog.Printf(\"Secret for %s in %s not found\", serviceAccount.GetName(), serviceAccount.GetNamespace())\n\t\treturn fmt.Sprintf(\"Secret %s not found\\n\", serviceAccount.GetName())\n\t} else if statusError, isStatus := err.(*errors.StatusError); isStatus {\n\t\tlog.Printf(\"Error getting secret %s in %s: %v\", serviceAccount.GetName(), serviceAccount.GetNamespace(), statusError.ErrStatus)\n\t\treturn fmt.Sprintf(\"Error getting secret %s: %v\\n\", serviceAccount.GetName(), statusError.ErrStatus)\n\t} else if err != nil {\n\t\tlog.Println(err.Error())\n\t\tpanic(err.Error())\n\t}\n\t// Define the cluster and server by taking advantage of the current config file\n\tcluster, server, err := custconfig.GetClusterServerOfCurrentContext()\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn fmt.Sprintf(\"Err: %s\", err)\n\t}\n\t// Put the collected data into new kubeconfig file\n\tnewKubeConfig := kubeconfigutil.CreateWithToken(server, cluster, serviceAccount.GetName(), secret.Data[\"ca.crt\"], string(secret.Data[\"token\"]))\n\tnewKubeConfig.Contexts[newKubeConfig.CurrentContext].Namespace = serviceAccount.GetNamespace()\n\tkubeconfigutil.WriteToDisk(fmt.Sprintf(\"../../assets/kubeconfigs/edgenet-%s-%s.cfg\", serviceAccount.GetNamespace(), serviceAccount.GetName()), newKubeConfig)\n\t// Check whether the creation process is completed\n\tdat, err := ioutil.ReadFile(fmt.Sprintf(\"../../assets/kubeconfigs/edgenet-%s-%s.cfg\", serviceAccount.GetNamespace(), serviceAccount.GetName()))\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn fmt.Sprintf(\"Err: %s\", err)\n\t}\n\treturn string(dat)\n}" ]
[ "0.6277103", "0.61862326", "0.6160129", "0.5999907", "0.5962999", "0.57921785", "0.5727932", "0.5545833", "0.5529446", "0.5510144", "0.53701234", "0.53701234", "0.532919", "0.5327097", "0.531085", "0.52852225", "0.52685714", "0.52612364", "0.52121156", "0.51724327", "0.5152428", "0.5152428", "0.5152428", "0.51465684", "0.5127652", "0.51060474", "0.51018554", "0.509688", "0.5091053", "0.5085071", "0.5082152", "0.5077791", "0.50309056", "0.5017036", "0.5010476", "0.50046396", "0.5000355", "0.49666476", "0.49615777", "0.49609792", "0.49569413", "0.49368468", "0.49321496", "0.4927263", "0.4922648", "0.49005514", "0.48986307", "0.48846376", "0.4875096", "0.48246858", "0.4822335", "0.47942", "0.4784153", "0.47819975", "0.47647083", "0.47457775", "0.47430736", "0.4742554", "0.47344536", "0.47292578", "0.47141486", "0.47094798", "0.47086546", "0.47078308", "0.47051048", "0.4695654", "0.46931684", "0.46924573", "0.46910366", "0.46657467", "0.46617904", "0.4658283", "0.46579844", "0.4653934", "0.46411887", "0.462393", "0.46211866", "0.4614468", "0.46139428", "0.4601662", "0.45854795", "0.45704797", "0.45655963", "0.45646107", "0.45625794", "0.45521286", "0.45410782", "0.45358813", "0.4535657", "0.45298746", "0.45254737", "0.4525355", "0.45252207", "0.45249146", "0.4524774", "0.45169523", "0.4516858", "0.45109499", "0.45025057", "0.45014304" ]
0.7063303
0
Outputs the set tenant otherwise informs user it is not set
func getTenant(c *cli.Context, w io.Writer) error { err := checkArgCount(c, 0) if err != nil { return err } config, err := cf.LoadConfig() if err != nil { return err } tenant := config.Tenant if tenant == nil { fmt.Printf("No tenant selected\n") } else { if c.GlobalIsSet("non-interactive") { fmt.Printf("%s\t%s\n", tenant.ID, tenant.Name) } else if utils.NeedsFormatting(c) { utils.FormatObject(tenant, w, c) } else { fmt.Printf("Current tenant is '%s' with ID %s\n", tenant.Name, tenant.ID) } } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o FastTcpAppOutput) Tenant() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *FastTcpApp) pulumi.StringOutput { return v.Tenant }).(pulumi.StringOutput)\n}", "func (o DataConnectorDynamics365Output) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *DataConnectorDynamics365) pulumi.StringOutput { return v.TenantId }).(pulumi.StringOutput)\n}", "func TenantID() string {\n\treturn tenantID\n}", "func TenantID() string {\n\treturn tenantID\n}", "func TenantID() string {\n\treturn tenantID\n}", "func (m *UserMutation) Tenant() (r string, exists bool) {\n\tv := m.tenant\n\tif v == nil {\n\t\treturn\n\t}\n\treturn *v, true\n}", "func setTenant(c *cli.Context) error {\n\terr := checkArgCount(c, 1)\n\tif err != nil {\n\t\treturn err\n\t}\n\tname := c.Args().First()\n\n\tclient.Photonclient, err = client.GetClient(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Ensure tenant exists\n\tid, err := findTenantID(name)\n\tif len(id) == 0 || err != nil {\n\t\treturn err\n\t}\n\n\tconfig, err := cf.LoadConfig()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tconfig.Tenant = &cf.TenantConfiguration{Name: name, ID: id}\n\terr = cf.SaveConfig(config)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = clearConfigProject(\"\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif !c.GlobalIsSet(\"non-interactive\") {\n\t\tfmt.Printf(\"Tenant set to '%s'\\n\", name)\n\t}\n\treturn nil\n}", "func (o ManagedInstanceActiveDirectoryAdministratorOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *ManagedInstanceActiveDirectoryAdministrator) pulumi.StringOutput { return v.TenantId }).(pulumi.StringOutput)\n}", "func (o FioSpecVolumeVolumeSourceQuobyteOutput) Tenant() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v FioSpecVolumeVolumeSourceQuobyte) *string { return v.Tenant }).(pulumi.StringPtrOutput)\n}", "func (m *LeaseMutation) Tenant() (r string, exists bool) {\n\tv := m.tenant\n\tif v == nil {\n\t\treturn\n\t}\n\treturn *v, true\n}", "func (c *TenantController) Show(ctx *app.ShowTenantContext) error {\n\tuserToken := goajwt.ContextJWT(ctx)\n\tif userToken == nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewUnauthorizedError(\"Missing JWT token\"))\n\t}\n\n\tttoken := &auth.TenantToken{Token: userToken}\n\ttenantID := ttoken.Subject()\n\ttenant, err := c.tenantService.GetTenant(tenantID)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewNotFoundError(\"tenants\", tenantID.String()))\n\t}\n\n\tnamespaces, err := c.tenantService.GetNamespaces(tenantID)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, err)\n\t}\n\n\tresult := &app.TenantSingle{Data: convertTenant(ctx, tenant, namespaces, c.clusterService.GetCluster)}\n\treturn ctx.OK(result)\n}", "func (o IopingSpecVolumeVolumeSourceQuobyteOutput) Tenant() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceQuobyte) *string { return v.Tenant }).(pulumi.StringPtrOutput)\n}", "func (o ManagedIdentityResponseOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ManagedIdentityResponse) string { return v.TenantId }).(pulumi.StringOutput)\n}", "func (o NetworkOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Network) pulumi.StringOutput { return v.TenantId }).(pulumi.StringOutput)\n}", "func (o GetServiceTenantAccessOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetServiceTenantAccess) string { return v.TenantId }).(pulumi.StringOutput)\n}", "func (o IdentityResponseOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v IdentityResponse) string { return v.TenantId }).(pulumi.StringOutput)\n}", "func (o DataConnectorOfficePowerBiOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *DataConnectorOfficePowerBi) pulumi.StringOutput { return v.TenantId }).(pulumi.StringOutput)\n}", "func (a *DefaultApiService) ShowTenant(ctx _context.Context, id string) ApiShowTenantRequest {\n\treturn ApiShowTenantRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\tid: id,\n\t}\n}", "func (o AzureDataLakeStoreOutputDataSourceResponseOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v AzureDataLakeStoreOutputDataSourceResponse) *string { return v.TenantId }).(pulumi.StringPtrOutput)\n}", "func (c *TenantController) Show(ctx *app.ShowTenantContext) error {\n\ttoken := goajwt.ContextJWT(ctx)\n\tif token == nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewUnauthorizedError(\"Missing JWT token\"))\n\t}\n\n\tttoken := &TenantToken{token: token}\n\ttenantID := ttoken.Subject()\n\ttenant, err := c.tenantService.GetTenant(tenantID)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewNotFoundError(\"tenants\", tenantID.String()))\n\t}\n\n\tnamespaces, err := c.tenantService.GetNamespaces(tenantID)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, err)\n\t}\n\n\tresult := &app.TenantSingle{Data: convertTenant(ctx, tenant, namespaces, c.resolveCluster)}\n\treturn ctx.OK(result)\n}", "func ShowTenant(ctx context.Context, config tenantConfig) (*tenant.TenantSingle, error) {\n\n\tc, err := createClient(ctx, config)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tres, err := c.ShowTenant(goasupport.ForwardContextRequestID(ctx), tenant.ShowTenantPath())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif res.StatusCode == http.StatusOK {\n\t\ttenant, err := c.DecodeTenantSingle(res)\n\t\tif err != nil {\n\t\t\treturn nil, errors.NewInternalError(ctx, err)\n\t\t}\n\t\treturn tenant, nil\n\t} else if res.StatusCode > 400 {\n\t\tjsonErr, err := c.DecodeJSONAPIErrors(res)\n\t\tif err == nil {\n\t\t\tif len(jsonErr.Errors) > 0 {\n\t\t\t\treturn nil, errors.NewInternalError(ctx, fmt.Errorf(jsonErr.Errors[0].Detail))\n\t\t\t}\n\t\t}\n\t}\n\treturn nil, errors.NewInternalError(ctx, fmt.Errorf(\"Unknown response \"+res.Status))\n}", "func (o MonitorV1Output) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *MonitorV1) pulumi.StringOutput { return v.TenantId }).(pulumi.StringOutput)\n}", "func (o AzureDataLakeStoreOutputDataSourceOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v AzureDataLakeStoreOutputDataSource) *string { return v.TenantId }).(pulumi.StringPtrOutput)\n}", "func (o ServiceTenantAccessOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ServiceTenantAccess) *string { return v.TenantId }).(pulumi.StringPtrOutput)\n}", "func (o SubscriptionOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Subscription) pulumi.StringOutput { return v.TenantId }).(pulumi.StringOutput)\n}", "func (o TagPropertyResponseOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v TagPropertyResponse) string { return v.TenantId }).(pulumi.StringOutput)\n}", "func (o NetworkManagerScopeConnectionOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *NetworkManagerScopeConnection) pulumi.StringOutput { return v.TenantId }).(pulumi.StringOutput)\n}", "func (o GetServiceIdentityOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetServiceIdentity) string { return v.TenantId }).(pulumi.StringOutput)\n}", "func (o GetGroupIdentityOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetGroupIdentity) string { return v.TenantId }).(pulumi.StringOutput)\n}", "func (o AuthConfigAzureAdOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *AuthConfigAzureAd) pulumi.StringOutput { return v.TenantId }).(pulumi.StringOutput)\n}", "func (o FioSpecVolumeVolumeSourceQuobytePtrOutput) Tenant() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *FioSpecVolumeVolumeSourceQuobyte) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Tenant\n\t}).(pulumi.StringPtrOutput)\n}", "func (o IopingSpecVolumeVolumeSourceQuobytePtrOutput) Tenant() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *IopingSpecVolumeVolumeSourceQuobyte) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Tenant\n\t}).(pulumi.StringPtrOutput)\n}", "func (o AzureActiveDirectoryOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v AzureActiveDirectory) *string { return v.TenantId }).(pulumi.StringPtrOutput)\n}", "func (o LedgerAzureadBasedServicePrincipalOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LedgerAzureadBasedServicePrincipal) string { return v.TenantId }).(pulumi.StringOutput)\n}", "func (o DppIdentityDetailsResponseOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v DppIdentityDetailsResponse) string { return v.TenantId }).(pulumi.StringOutput)\n}", "func (o UpdateHistoryPropertyResponseOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v UpdateHistoryPropertyResponse) string { return v.TenantId }).(pulumi.StringOutput)\n}", "func (s *TenantServiceListener) Set(ctx context.Context, in *pb.TenantName) (*google_protobuf.Empty, error) {\n\tlog.Printf(\"Tenant Set called '%s'\", in.Name)\n\n\tif currentTenant != nil && currentTenant.name == in.GetName() {\n\t\tlog.Printf(\"Tenant '%s' is already selected\", in.GetName())\n\t\treturn &google_protobuf.Empty{}, nil\n\t}\n\n\tservice, err := providers.GetService(in.GetName())\n\tif err != nil {\n\t\treturn &google_protobuf.Empty{}, fmt.Errorf(\"Unable to set tenant '%s': %s\", in.GetName(), err.Error())\n\t}\n\tcurrentTenant = &Tenant{name: in.GetName(), Service: service}\n\tlog.Printf(\"Current tenant is now '%s'\", in.GetName())\n\treturn &google_protobuf.Empty{}, nil\n}", "func getCurrentTenant() *Tenant {\n\tif currentTenant == nil {\n\t\ttenants, err := iaas.GetTenantNames()\n\t\tif err != nil || len(tenants) != 1 {\n\t\t\treturn nil\n\t\t}\n\t\t// Set unique tenant as selected\n\t\tlog.Println(\"Unique tenant set\")\n\t\tfor name := range tenants {\n\t\t\tservice, err := iaas.UseService(name)\n\t\t\tif err != nil {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\tcurrentTenant = &Tenant{name: name, Service: service}\n\t\t}\n\t}\n\treturn currentTenant\n}", "func (o GetKubernetesClusterIdentityOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetKubernetesClusterIdentity) string { return v.TenantId }).(pulumi.StringOutput)\n}", "func (s Tenant) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (o AzureActiveDirectoryResponseOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v AzureActiveDirectoryResponse) *string { return v.TenantId }).(pulumi.StringPtrOutput)\n}", "func (o GetKubernetesClusterAzureActiveDirectoryRoleBasedAccessControlOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetKubernetesClusterAzureActiveDirectoryRoleBasedAccessControl) string { return v.TenantId }).(pulumi.StringOutput)\n}", "func (o GroupIdentityOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v GroupIdentity) *string { return v.TenantId }).(pulumi.StringPtrOutput)\n}", "func (m *DeviceCompliancePolicySettingStateSummary) GetTenantDisplayName()(*string) {\n val, err := m.GetBackingStore().Get(\"tenantDisplayName\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (o ServiceIdentityOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ServiceIdentity) *string { return v.TenantId }).(pulumi.StringPtrOutput)\n}", "func tenantHandler(formatter *render.Render) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, req *http.Request) {\n\t\tsetupResponse(&w, req)\n\n\t\tif (*req).Method == \"OPTIONS\" {\n\t\t\tfmt.Println(\"PREFLIGHT Request\")\n\t\t\treturn\n\t\t}\n\t\tsession, err := mgo.Dial(mongodb_server)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\tdefer session.Close()\n\t\tsession.SetMode(mgo.Monotonic, true)\n\t\tparams := mux.Vars(req)\n\t\tvar id string = params[\"id\"]\n\t\tfmt.Println(\"Tenant ID: \", id)\n\t\tvar result bson.M\n\t\tif id == \"\" {\n\t\t\tformatter.JSON(w, http.StatusBadRequest, \"Tenant ID Missing\")\n\t\t} else {\n\t\t\tc := session.DB(mongodb_database).C(mongodb_collection)\n\t\t\terr = c.Find(bson.M{\"id\":id}).One(&result)\n\t\t\tif err != nil {\n\t\t\tfmt.Println(\" Tenant: \", err)\n\t\t\tformatter.JSON(w, http.StatusBadRequest, \"Not Found\")\n\t\t\t}\n\t\t\tfmt.Println(\" Tenant: \", result)\n\t\t\tformatter.JSON(w, http.StatusOK, result)\n\t\t}\n\t}\n}", "func getCurrentTenant() *Tenant {\n\tif currentTenant == nil {\n\t\ttenants, err := providers.Tenants()\n\t\tif err != nil || len(tenants) != 1 {\n\t\t\treturn nil\n\t\t}\n\t\t// Set unqiue tenant as selected\n\t\tlog.Println(\"Unique tenant set\")\n\t\tfor name := range tenants {\n\t\t\tservice, err := providers.GetService(name)\n\t\t\tif err != nil {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\tcurrentTenant = &Tenant{name: name, Service: service}\n\t\t}\n\t}\n\treturn currentTenant\n}", "func fnTenant(ctx Context, doc *JDoc, params []string) interface{} {\n\tstats := ctx.Value(EelTotalStats).(*ServiceStats)\n\tif params == nil || params[0] != \"\" {\n\t\tctx.Log().Error(\"error_type\", \"func_tenant\", \"op\", \"tenant\", \"cause\", \"no_parameters_expected\", \"params\", params)\n\t\tstats.IncErrors()\n\t\tAddError(ctx, SyntaxError{fmt.Sprintf(\"no parameters expected in call to tenant function\"), \"tenant\", params})\n\t\treturn \"\"\n\t}\n\th := GetCurrentHandlerConfig(ctx)\n\tif h == nil {\n\t\tctx.Log().Error(\"error_type\", \"func_tenant\", \"op\", \"tenant\", \"cause\", \"no_handler\", \"params\", params)\n\t\tstats.IncErrors()\n\t\tAddError(ctx, RuntimeError{fmt.Sprintf(\"current handler not found in call to tenant function\"), \"tenant\", params})\n\t\treturn \"\"\n\t}\n\t//Add logic to separating tenant and partner\n\tallowPartner := GetConfig(ctx).AllowPartner\n\tif h.TenantId == \"_default\" {\n\t\t//The default tenant handler is used. We need to drill-down into context to find out the actual tenant\n\t\ttenantHeaderKey := GetConfig(ctx).HttpTenantHeader\n\t\tif ctx.Value(tenantHeaderKey) != nil {\n\t\t\tcombinedTenant := ctx.Value(tenantHeaderKey).(string)\n\t\t\ttenant := \"\"\n\t\t\tif allowPartner && strings.LastIndex(combinedTenant, \"_\") > 0 && strings.LastIndex(combinedTenant, \"_\") < len(combinedTenant)-1 {\n\t\t\t\ttenant = combinedTenant[:strings.LastIndex(combinedTenant, \"_\")]\n\t\t\t} else {\n\t\t\t\ttenant = combinedTenant\n\t\t\t}\n\t\t\treturn tenant\n\t\t}\n\t}\n\tcombinedTenant := h.TenantId\n\ttenant := \"\"\n\tif allowPartner && strings.LastIndex(combinedTenant, \"_\") > 0 && strings.LastIndex(combinedTenant, \"_\") < len(combinedTenant)-1 {\n\t\ttenant = combinedTenant[:strings.LastIndex(combinedTenant, \"_\")]\n\t} else {\n\t\ttenant = combinedTenant\n\t}\n\treturn tenant\n}", "func (o GrafanaIdentityOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v GrafanaIdentity) *string { return v.TenantId }).(pulumi.StringPtrOutput)\n}", "func (o RegistryTaskIdentityOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v RegistryTaskIdentity) *string { return v.TenantId }).(pulumi.StringPtrOutput)\n}", "func (o FluxConfigurationBlobStorageServicePrincipalOutput) TenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v FluxConfigurationBlobStorageServicePrincipal) string { return v.TenantId }).(pulumi.StringOutput)\n}", "func (m *DeviceCompliancePolicySettingStateSummary) SetTenantDisplayName(value *string)() {\n err := m.GetBackingStore().Set(\"tenantDisplayName\", value)\n if err != nil {\n panic(err)\n }\n}", "func (o KubernetesClusterAzureActiveDirectoryRoleBasedAccessControlOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v KubernetesClusterAzureActiveDirectoryRoleBasedAccessControl) *string { return v.TenantId }).(pulumi.StringPtrOutput)\n}", "func (o RegistryIdentityOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v RegistryIdentity) *string { return v.TenantId }).(pulumi.StringPtrOutput)\n}", "func (a *DefaultApiService) ShowTenantExecute(r ApiShowTenantRequest) (Tenant, *_nethttp.Response, GenericOpenAPIError) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodGet\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\texecutionError GenericOpenAPIError\n\t\tlocalVarReturnValue Tenant\n\t)\n\n\tlocalBasePath, err := a.client.cfg.ServerURLWithContext(r.ctx, \"DefaultApiService.ShowTenant\")\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarReturnValue, nil, executionError\n\t}\n\n\tlocalVarPath := localBasePath + \"/tenants/{id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"id\"+\"}\", _neturl.PathEscape(parameterToString(r.id, \"\")), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\treq, err := a.client.prepareRequest(r.ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarReturnValue, nil, executionError\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(req)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarReturnValue, localVarHTTPResponse, executionError\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tlocalVarHTTPResponse.Body = _ioutil.NopCloser(bytes.NewBuffer(localVarBody))\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarReturnValue, localVarHTTPResponse, executionError\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v ErrorNotFound\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, executionError\n}", "func (m *AadUserConversationMember) GetTenantId()(*string) {\n return m.tenantId\n}", "func (m *AadUserConversationMember) SetTenantId(value *string)() {\n m.tenantId = value\n}", "func (o ManagedIdentityResponsePtrOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ManagedIdentityResponse) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.TenantId\n\t}).(pulumi.StringPtrOutput)\n}", "func (m *DeviceCompliancePolicySettingStateSummary) GetTenantId()(*string) {\n val, err := m.GetBackingStore().Get(\"tenantId\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (m *CloudPcConnection) SetTenantDisplayName(value *string)() {\n err := m.GetBackingStore().Set(\"tenantDisplayName\", value)\n if err != nil {\n panic(err)\n }\n}", "func (o ConnectionAzureOutput) CustomerTenantId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ConnectionAzure) string { return v.CustomerTenantId }).(pulumi.StringOutput)\n}", "func (o IdentityResponsePtrOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *IdentityResponse) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.TenantId\n\t}).(pulumi.StringPtrOutput)\n}", "func (o KubernetesClusterIdentityOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v KubernetesClusterIdentity) *string { return v.TenantId }).(pulumi.StringPtrOutput)\n}", "func (m *DeviceCompliancePolicySettingStateSummary) SetTenantId(value *string)() {\n err := m.GetBackingStore().Set(\"tenantId\", value)\n if err != nil {\n panic(err)\n }\n}", "func (s *compassE2EState) GetDefaultTenant() string {\n\treturn s.config.Tenant\n}", "func (o AzureActiveDirectoryResponsePtrOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *AzureActiveDirectoryResponse) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.TenantId\n\t}).(pulumi.StringPtrOutput)\n}", "func (t *Tenants) Set(tenant *Tenant) error {\n\tw := &bytes.Buffer{}\n\tif err := tenant.Encode(w); err != nil {\n\t\treturn err\n\t}\n\treturn t.store.Set(tenant.ID, w.Bytes())\n}", "func (s *TenantListener) Set(ctx context.Context, in *pb.TenantName) (empty *googleprotobuf.Empty, err error) {\n\tempty = &googleprotobuf.Empty{}\n\tif s == nil {\n\t\treturn empty, status.Errorf(codes.FailedPrecondition, fail.InvalidInstanceError().Message())\n\t}\n\tif in == nil {\n\t\treturn empty, status.Errorf(codes.InvalidArgument, fail.InvalidParameterError(\"in\", \"cannot be nil\").Message())\n\t}\n\tname := in.GetName()\n\t// FIXME: validate parameters\n\n\ttracer := debug.NewTracer(nil, fmt.Sprintf(\"('%s')\", name), true).WithStopwatch().GoingIn()\n\tdefer tracer.OnExitTrace()()\n\tdefer fail.OnExitLogError(tracer.TraceMessage(\"\"), &err)()\n\n\tctx, cancelFunc := context.WithCancel(ctx)\n\t// FIXME: handle error\n\tif err := srvutils.JobRegister(ctx, cancelFunc, \"Tenant Set \"+name); err == nil {\n\t\tdefer srvutils.JobDeregister(ctx)\n\t}\n\n\tif currentTenant != nil && currentTenant.name == in.GetName() {\n\t\treturn empty, nil\n\t}\n\n\tservice, err := iaas.UseService(in.GetName())\n\tif err != nil {\n\t\treturn empty, fmt.Errorf(\"unable to set tenant '%s': %s\", name, getUserMessage(err))\n\t}\n\tcurrentTenant = &Tenant{name: in.GetName(), Service: service}\n\tlog.Infof(\"Current tenant is now '%s'\", name)\n\treturn empty, nil\n}", "func (o ServiceTenantAccessPtrOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ServiceTenantAccess) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.TenantId\n\t}).(pulumi.StringPtrOutput)\n}", "func (l *labelInfo) tenantLabel() map[string]string {\n\tt := make(map[string]string)\n\tif l.Tenant != \"\" {\n\t\tt[tenantLabelKey] = string(l.Tenant)\n\t}\n\treturn t\n}", "func TenantID(ctx context.Context) string {\n\tif requestID, ok := ctx.Value(ctxTenantKey{}).(string); ok {\n\t\treturn requestID\n\t}\n\n\treturn \"\"\n}", "func (o AzureActiveDirectoryPtrOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *AzureActiveDirectory) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.TenantId\n\t}).(pulumi.StringPtrOutput)\n}", "func Tenant(tenant string) Modifier {\n\treturn func(r *http.Request) error {\n\t\tr.Header.Set(\"Hawkular-Tenant\", tenant)\n\t\treturn nil\n\t}\n}", "func Tenant(tenant string) Modifier {\n\treturn func(r *http.Request) error {\n\t\tr.Header.Set(\"Hawkular-Tenant\", tenant)\n\t\treturn nil\n\t}\n}", "func (p TenantTablePrinter) Print(tenants []*models.V1TenantResponse) {\n\tp.wideHeader = []string{\"ID\", \"Name\", \"Description\", \"Clusters\", \"Machines\", \"IPs\", \"Labels\", \"Annotations\"}\n\tp.shortHeader = p.wideHeader\n\tfor _, tenantResponse := range tenants {\n\t\ttenant := tenantResponse\n\t\tclusterQuota := \"\"\n\t\tmachineQuota := \"\"\n\t\tipQuota := \"\"\n\t\t// FIXME add actual quotas ?\n\t\tif tenant.DefaultQuotas != nil {\n\t\t\tqs := tenant.DefaultQuotas\n\t\t\tif qs.Cluster != nil {\n\t\t\t\tcq := \"∞\"\n\t\t\t\tif qs.Cluster.Quota != 0 {\n\t\t\t\t\tcq = strconv.FormatInt(int64(qs.Cluster.Quota), 10)\n\t\t\t\t}\n\t\t\t\tclusterQuota = fmt.Sprintf(\"%d/%s\", qs.Cluster.Used, cq)\n\t\t\t}\n\t\t\tif qs.Machine != nil {\n\t\t\t\tmq := \"∞\"\n\t\t\t\tif qs.Cluster.Quota != 0 {\n\t\t\t\t\tmq = strconv.FormatInt(int64(qs.Machine.Quota), 10)\n\t\t\t\t}\n\t\t\t\tmachineQuota = fmt.Sprintf(\"%d/%s\", qs.Machine.Used, mq)\n\t\t\t}\n\t\t\tif qs.IP != nil {\n\t\t\t\tiq := \"∞\"\n\t\t\t\tif qs.IP.Quota != 0 {\n\t\t\t\t\tiq = strconv.FormatInt(int64(qs.IP.Quota), 10)\n\t\t\t\t}\n\t\t\t\tipQuota = fmt.Sprintf(\"%d/%s\", qs.IP.Used, iq)\n\t\t\t}\n\t\t}\n\t\tlabels := strings.Join(tenant.Meta.Labels, \"\\n\")\n\t\tas := []string{}\n\t\tfor k, v := range tenant.Meta.Annotations {\n\t\t\tas = append(as, k+\"=\"+v)\n\t\t}\n\t\tannotations := strings.Join(as, \"\\n\")\n\n\t\twide := []string{tenant.Meta.ID, tenant.Name, tenant.Description, clusterQuota, machineQuota, ipQuota, labels, annotations}\n\t\tp.addWideData(wide, tenant)\n\t\tp.addShortData(wide, tenant)\n\t}\n\tp.render()\n}", "func Tenant(v string) predicate.User {\n\treturn predicate.User(func(s *sql.Selector) {\n\t\ts.Where(sql.EQ(s.C(FieldTenant), v))\n\t})\n}", "func (o *NiatelemetryMsoTenantDetailsAllOf) GetTenantIdOk() (*string, bool) {\n\tif o == nil || o.TenantId == nil {\n\t\treturn nil, false\n\t}\n\treturn o.TenantId, true\n}", "func (m *LeaseMutation) Idtenant() (r string, exists bool) {\n\tv := m.idtenant\n\tif v == nil {\n\t\treturn\n\t}\n\treturn *v, true\n}", "func (m *SharepointIds) SetTenantId(value *string)() {\n err := m.GetBackingStore().Set(\"tenantId\", value)\n if err != nil {\n panic(err)\n }\n}", "func (o KubernetesClusterAzureActiveDirectoryRoleBasedAccessControlPtrOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *KubernetesClusterAzureActiveDirectoryRoleBasedAccessControl) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.TenantId\n\t}).(pulumi.StringPtrOutput)\n}", "func (o GrafanaIdentityPtrOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *GrafanaIdentity) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.TenantId\n\t}).(pulumi.StringPtrOutput)\n}", "func (o GroupIdentityPtrOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *GroupIdentity) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.TenantId\n\t}).(pulumi.StringPtrOutput)\n}", "func (o RegistryTaskIdentityPtrOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *RegistryTaskIdentity) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.TenantId\n\t}).(pulumi.StringPtrOutput)\n}", "func (o RegistryIdentityPtrOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *RegistryIdentity) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.TenantId\n\t}).(pulumi.StringPtrOutput)\n}", "func (m *SharepointIds) GetTenantId()(*string) {\n val, err := m.GetBackingStore().Get(\"tenantId\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (c *TenantController) Update(ctx *app.UpdateTenantContext) error {\n\tuserToken := goajwt.ContextJWT(ctx)\n\tif userToken == nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewUnauthorizedError(\"Missing JWT token\"))\n\t}\n\tttoken := &auth.TenantToken{Token: userToken}\n\ttenant, err := c.tenantService.GetTenant(ttoken.Subject())\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewNotFoundError(\"tenants\", ttoken.Subject().String()))\n\t}\n\n\t// fetch the cluster the user belongs to\n\tuser, err := c.authClientService.GetUser(ctx)\n\tif err != nil {\n\t\treturn jsonapi.JSONErrorResponse(ctx, err)\n\t}\n\n\tif user.UserData.Cluster == nil {\n\t\tlog.Error(ctx, nil, \"no cluster defined for tenant\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewInternalError(ctx, fmt.Errorf(\"unable to provision to undefined cluster\")))\n\t}\n\n\tcluster, err := c.clusterService.GetCluster(ctx, *user.UserData.Cluster)\n\tif err != nil {\n\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\"err\": err,\n\t\t\t\"cluster_url\": *user.UserData.Cluster,\n\t\t}, \"unable to fetch cluster\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewInternalError(ctx, err))\n\t}\n\n\t// create openshift config\n\topenshiftConfig := openshift.NewConfig(c.config, user.UserData, cluster.User, cluster.Token, cluster.APIURL)\n\n\t// update tenant config\n\ttenant.OSUsername = user.OpenShiftUsername\n\tif tenant.NsBaseName == \"\" {\n\t\ttenant.NsBaseName = env.RetrieveUserName(user.OpenShiftUsername)\n\t}\n\tif err = c.tenantService.SaveTenant(tenant); err != nil {\n\t\tlog.Error(ctx, map[string]interface{}{\n\t\t\t\"err\": err,\n\t\t}, \"unable to update tenant configuration\")\n\t\treturn jsonapi.JSONErrorResponse(ctx, errors.NewInternalError(ctx, fmt.Errorf(\"unable to update tenant configuration: %v\", err)))\n\t}\n\n\tgo func() {\n\t\tctx := ctx\n\t\tt := tenant\n\t\terr = openshift.RawUpdateTenant(\n\t\t\tctx,\n\t\t\topenshiftConfig,\n\t\t\tInitTenant(ctx, openshiftConfig.MasterURL, c.tenantService, t),\n\t\t\tuser.OpenShiftUsername,\n\t\t\ttenant.NsBaseName)\n\n\t\tif err != nil {\n\t\t\tsentry.LogError(ctx, map[string]interface{}{\n\t\t\t\t\"os_user\": user.OpenShiftUsername,\n\t\t\t}, err, \"unable initialize tenant\")\n\t\t}\n\t}()\n\n\tctx.ResponseData.Header().Set(\"Location\", rest.AbsoluteURL(ctx.RequestData.Request, app.TenantHref()))\n\treturn ctx.Accepted()\n}", "func (l *labelInfo) isSuperTenant() bool {\n\tif l.Tenant == \"\" || strings.ToLower(string(l.Tenant)) == frontend.GetDefaultTenant() {\n\t\treturn true\n\t}\n\treturn false\n}", "func (m *CloudPcConnection) SetTenantId(value *string)() {\n err := m.GetBackingStore().Set(\"tenantId\", value)\n if err != nil {\n panic(err)\n }\n}", "func (o *ResourceVersion) GetTenantIdOk() (*string, bool) {\n\tif o == nil || o.TenantId == nil {\n\t\treturn nil, false\n\t}\n\treturn o.TenantId, true\n}", "func (o *NiatelemetryMsoTenantDetailsAllOf) GetTenantId() string {\n\tif o == nil || o.TenantId == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.TenantId\n}", "func (o KubernetesClusterExtensionAksAssignedIdentityOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v KubernetesClusterExtensionAksAssignedIdentity) *string { return v.TenantId }).(pulumi.StringPtrOutput)\n}", "func (s *MultiElasticsearch) Tenant(tenant string) Store {\n\tkv, found := s.stores[tenant]\n\tif !found {\n\t\tkv = NewElasticsearch(s.config, s.client, tenant, s.typeName)\n\t\ts.stores[tenant] = kv\n\t}\n\treturn kv\n}", "func (o ServiceIdentityPtrOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ServiceIdentity) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.TenantId\n\t}).(pulumi.StringPtrOutput)\n}", "func (m *CloudPcConnection) GetTenantDisplayName()(*string) {\n val, err := m.GetBackingStore().Get(\"tenantDisplayName\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func GetTenantsCommand() cli.Command {\n\tcommand := cli.Command{\n\t\tName: \"tenant\",\n\t\tUsage: \"options for tenant\",\n\t\tSubcommands: []cli.Command{\n\t\t\t{\n\t\t\t\tName: \"create\",\n\t\t\t\tUsage: \"Create a new tenant\",\n\t\t\t\tArgsUsage: \"<tenant-name>\",\n\t\t\t\tFlags: []cli.Flag{\n\t\t\t\t\tcli.StringFlag{\n\t\t\t\t\t\tName: \"security-groups, s\",\n\t\t\t\t\t\tUsage: \"Comma-separated Lightwave group names, to specify the tenant administrators\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := createTenant(c, os.Stdout)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"delete\",\n\t\t\t\tUsage: \"Delete a tenant\",\n\t\t\t\tArgsUsage: \"<tenant-id>\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := deleteTenant(c)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"show\",\n\t\t\t\tUsage: \"Show detailed tenant info with specified id\",\n\t\t\t\tArgsUsage: \"<tenant-id>\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := showTenant(c, os.Stdout)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"list\",\n\t\t\t\tUsage: \"List all tenants\",\n\t\t\t\tArgsUsage: \" \",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := listTenants(c, os.Stdout)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"set\",\n\t\t\t\tUsage: \"Set default tenant\",\n\t\t\t\tArgsUsage: \"<tenant-name>\",\n\t\t\t\tDescription: \"Set the default project that will be used for all photon CLI commands that need a project.\\n\" +\n\t\t\t\t\t\" Most commands allow you to override the default.\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := setTenant(c)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"get\",\n\t\t\t\tUsage: \"Get default tenant\",\n\t\t\t\tArgsUsage: \" \",\n\t\t\t\tDescription: \"Show default project in use for photon CLI commands. Most command allow you to either\\n\" +\n\t\t\t\t\t\" use this default or specify a specific project to use.\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := getTenant(c, os.Stdout)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"tasks\",\n\t\t\t\tUsage: \"Show tenant tasks\",\n\t\t\t\tArgsUsage: \"<tenant-id>\",\n\t\t\t\tFlags: []cli.Flag{\n\t\t\t\t\tcli.StringFlag{\n\t\t\t\t\t\tName: \"state, s\",\n\t\t\t\t\t\tUsage: \"Filter by task sate\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := getTenantTasks(c, os.Stdout)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tName: \"set-security-groups\",\n\t\t\t\tUsage: \"Set security groups for a tenant\",\n\t\t\t\tArgsUsage: \"<tenant-id> <comma separated list of groups>\",\n\t\t\t\tDescription: \"Set the list of Lightwave groups that can administer this tenant. This may only be\\n\" +\n\t\t\t\t\t\" be set by a member of the tenant. Be cautious--you can remove your own access if you specify\\n\" +\n\t\t\t\t\t\" the wrong set of groups.\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := setSecurityGroups(c)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t\t{\n\t\t\t\tHidden: true,\n\t\t\t\tName: \"set_security_groups\",\n\t\t\t\tUsage: \"Set security groups for a tenant\",\n\t\t\t\tArgsUsage: \"<tenant-id> <comma separated list of groups>\",\n\t\t\t\tDescription: \"Deprecated, use set-security-groups instead\",\n\t\t\t\tAction: func(c *cli.Context) {\n\t\t\t\t\terr := setSecurityGroups(c)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t}\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\treturn command\n}", "func (o DppIdentityDetailsResponsePtrOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *DppIdentityDetailsResponse) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.TenantId\n\t}).(pulumi.StringPtrOutput)\n}", "func (r *mutationResolver) WriteTenant(ctx context.Context, in graphql.BusinessTenantMappingInput) (string, error) {\n\treturn r.tenant.WriteSingle(ctx, in)\n}", "func (nuo *NodeUpdateOne) SetTenant(s string) *NodeUpdateOne {\n\tnuo.mutation.SetTenant(s)\n\treturn nuo\n}", "func (o KubernetesClusterIdentityPtrOutput) TenantId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *KubernetesClusterIdentity) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.TenantId\n\t}).(pulumi.StringPtrOutput)\n}", "func (o *GetDeploymentPreview1Params) SetTenant(tenant string) {\n\to.Tenant = tenant\n}" ]
[ "0.70123947", "0.6534614", "0.6488953", "0.6488953", "0.6488953", "0.64799076", "0.6426935", "0.64256495", "0.63903344", "0.6379141", "0.63771296", "0.6361804", "0.63579327", "0.63526416", "0.634557", "0.6340448", "0.63213265", "0.63013923", "0.6283609", "0.62798727", "0.6246797", "0.6246336", "0.6215506", "0.61829346", "0.61552465", "0.61495453", "0.61219853", "0.6076412", "0.6048234", "0.60460466", "0.6030872", "0.60202944", "0.5994091", "0.5958783", "0.59434533", "0.59402955", "0.59314454", "0.5928021", "0.5925996", "0.5911097", "0.5904179", "0.59013236", "0.58894044", "0.587506", "0.5872258", "0.58657897", "0.5853869", "0.58464235", "0.58401936", "0.58265543", "0.58115506", "0.58018184", "0.57985014", "0.57917434", "0.5786891", "0.57858986", "0.57796377", "0.57442766", "0.57379466", "0.57289344", "0.5724989", "0.5710008", "0.56978476", "0.56922996", "0.5690568", "0.5666829", "0.5665674", "0.56452405", "0.56288546", "0.5616279", "0.5608701", "0.55877244", "0.558202", "0.558202", "0.5577258", "0.55768794", "0.5574644", "0.5535292", "0.55292696", "0.55264884", "0.55112046", "0.5504704", "0.5504336", "0.5493119", "0.54685086", "0.5466756", "0.5457304", "0.5419587", "0.5416132", "0.5416022", "0.54139155", "0.54008466", "0.53951585", "0.539282", "0.53892803", "0.53855217", "0.5383457", "0.53616697", "0.5361308", "0.53552556" ]
0.700981
1
Retrieves tasks from specified tenant
func getTenantTasks(c *cli.Context, w io.Writer) error { err := checkArgCount(c, 1) if err != nil { return err } id := c.Args().First() state := c.String("state") options := &photon.TaskGetOptions{ State: state, } client.Photonclient, err = client.GetClient(c) if err != nil { return err } taskList, err := client.Photonclient.Tenants.GetTasks(id, options) if err != nil { return err } err = printTaskList(taskList.Items, c) if err != nil { return err } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (taskService TaskService) GetByPartnerAndManagedEndpointID(w http.ResponseWriter, r *http.Request) {\n\tendpointID, err := common.ExtractUUID(\"TaskService.GetByPartnerAndManagedEndpointID\", w, r, \"managedEndpointID\")\n\tif err != nil {\n\t\treturn\n\t}\n\n\tpartnerID := mux.Vars(r)[partnerIDKey]\n\tcurrentUser := taskService.userService.GetUser(r, taskService.httpClient)\n\n\tlistOfInternalTasksByTarget, err := taskService.taskPersistence.GetByPartnerAndManagedEndpointID(r.Context(), partnerID, endpointID, common.UnlimitedCount)\n\tif err != nil {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorCantGetListOfTasksByManagedEndpoint, \"TaskService.GetByPartnerAndManagedEndpointID: can not get list of Tasks by ManagedEndpointID. err=%v\", err)\n\t\tcommon.SendInternalServerError(w, r, errorcode.ErrorCantGetListOfTasksByManagedEndpoint)\n\t\treturn\n\t}\n\n\ttaskIDs := make([]gocql.UUID, len(listOfInternalTasksByTarget))\n\tfor i, internalTask := range listOfInternalTasksByTarget {\n\t\ttaskIDs[i] = internalTask.ID\n\t}\n\n\tlistOfInternalTasksByIDs, err := taskService.taskPersistence.GetByIDs(r.Context(), nil, partnerID, false, taskIDs...)\n\tif err != nil {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorCantGetListOfTasksByManagedEndpoint, \"TaskService.GetByPartnerAndManagedEndpointID: can not get list of Tasks by IDs(%v). err=%v\", taskIDs, err)\n\t\tcommon.SendInternalServerError(w, r, errorcode.ErrorCantGetListOfTasksByManagedEndpoint)\n\t\treturn\n\t}\n\n\t// Filter by NOC\n\tfilteredTaskGroup := filterTasksByNOCAccsess(listOfInternalTasksByIDs, currentUser.HasNOCAccess())\n\n\t// Group tasks by taskIDs\n\ttasksGroup := make(map[gocql.UUID][]models.Task)\n\tfor _, task := range filteredTaskGroup {\n\t\ttasksGroup[task.ID] = append(tasksGroup[task.ID], task)\n\t}\n\n\tlistOfOutputTasks := make([]models.Task, 0)\n\tfor _, internalTaskGroup := range tasksGroup {\n\t\ttaskOutputPtr, err := models.NewTaskOutput(r.Context(), internalTaskGroup)\n\t\tif err != nil {\n\t\t\tlogger.Log.WarnfCtx(\n\t\t\t\tr.Context(),\n\t\t\t\t\"TaskService.GetByPartnerAndManagedEndpointID: failed to build task output from internal tasks %v. Err: %s\",\n\t\t\t\tinternalTaskGroup,\n\t\t\t\terr,\n\t\t\t)\n\t\t\tcontinue\n\t\t}\n\t\tlistOfOutputTasks = append(listOfOutputTasks, *taskOutputPtr)\n\t}\n\n\tlogger.Log.DebugfCtx(r.Context(), \"TaskService.GetByPartnerAndManagedEndpointID: successfully returned list of Tasks by ManagedEndpointID.\")\n\tcommon.RenderJSON(w, listOfOutputTasks)\n}", "func GetTasks(c echo.Context) error {\n\t// Type assert user from context and authorize\n\tuser, ok := c.Get(\"user\").(*model.UserSecure)\n\tif !ok {\n\t\treturn echo.ErrUnauthorized\n\t}\n\n\t// Get userID from param and decide authorization\n\tuserID := c.QueryParam(\"userID\")\n\tif !allows(user.Role, model.PermissionViewAllTasks) {\n\t\tif len(userID) == 0 {\n\t\t\tuserID = user.ID.Hex()\n\t\t} else if userID != user.ID.Hex() {\n\t\t\treturn echo.ErrForbidden\n\t\t}\n\t}\n\n\t// Get db connection\n\tdb, err := store.NewMongoStore()\n\tif err != nil {\n\t\treturn echo.NewHTTPError(http.StatusServiceUnavailable, err.Error())\n\t}\n\tdefer db.Cleanup()\n\n\t// Construct query\n\tq, err := store.NewTaskQueryFromParams(\n\t\tuserID, \"\",\n\t\tc.QueryParam(\"from\"), c.QueryParam(\"to\"))\n\tif err != nil {\n\t\treturn echo.NewHTTPError(http.StatusBadRequest, err)\n\t}\n\n\t// Fetch tasks\n\ttasks, err := db.GetAllTasks(q)\n\tif err != nil {\n\t\treturn errors.MongoErrorResponse(err)\n\t}\n\treturn c.JSON(http.StatusOK, tasks)\n}", "func (s *Search) Tasks(ctx *gin.Context) {\n\tapp := ctx.Param(\"app\")\n\tslot := ctx.Param(\"slot\")\n\tappLabel := config.LogAppLabel()\n\tslotLabel := config.LogSlotLabel()\n\toptions := config.ConvertRequestQueryParams(ctx.Request.URL.Query())\n\toptions[appLabel] = app\n\toptions[slotLabel] = slot\n\ttasks, err := s.Service.Tasks(options, ctx.MustGet(\"page\").(models.Page))\n\tif err != nil {\n\t\tutils.ErrorResponse(ctx, utils.NewError(GetTaskError, err))\n\t\treturn\n\t}\n\tutils.Ok(ctx, tasks)\n}", "func (c *Client) GetTask(ctx context.Context, in *todopb.TaskQuery, opts ...grpc.CallOption) (*todopb.TaskResponse, error) {\n\treturn c.client.GetTask(ctx, in, opts...)\n}", "func (context Context) GetTasks(\n\tfilters F,\n\tlimit int,\n\tcursor string,\n\tsortBy string,\n\torder string,\n) (result []Task, cm CollectionMetadata, err error) {\n\n\tc := context.Session.DB(context.DBName).C(\"tasks\")\n\n\t// Validate the parameters.\n\tif sortBy != \"\" &&\n\t\tsortBy != \"id\" &&\n\t\tsortBy != \"process\" &&\n\t\tsortBy != \"job\" &&\n\t\tsortBy != \"user\" &&\n\t\tsortBy != \"dataset\" &&\n\t\tsortBy != \"objective\" &&\n\t\tsortBy != \"model\" &&\n\t\tsortBy != \"quality\" &&\n\t\tsortBy != \"quality-train\" &&\n\t\tsortBy != \"quality-expected\" &&\n\t\tsortBy != \"creation-time\" &&\n\t\tsortBy != \"status\" &&\n\t\tsortBy != \"stage\" {\n\t\terr = errors.Wrapf(ErrBadInput, \"cannot sort by \\\"%s\\\"\", sortBy)\n\t\treturn\n\t}\n\tif order != \"\" && order != \"asc\" && order != \"desc\" {\n\t\terr = errors.Wrapf(ErrBadInput, \"order can be either \\\"asc\\\" or \\\"desc\\\", not \\\"%s\\\"\", order)\n\t\treturn\n\t}\n\tif order == \"\" {\n\t\torder = \"asc\"\n\t}\n\n\t// If the user is not root then we need to limit access.\n\tquery := bson.M{}\n\tif context.User.IsRoot() == false {\n\t\tquery = bson.M{\"user\": bson.M{\"$in\": []string{context.User.ID, UserRoot}}}\n\t}\n\n\t// Build a query given the parameters.\n\tfor k, v := range filters {\n\t\tswitch k {\n\t\tcase \"id\":\n\t\t\tsetDefault(&query, \"id\", bson.M{})\n\t\t\tquery[\"id\"].(bson.M)[\"$in\"] = v.([]string)\n\t\tcase \"user\", \"dataset\", \"model\", \"objective\", \"status\", \"stage\":\n\t\t\tsetDefault(&query, k, bson.M{})\n\t\t\tquery[k].(bson.M)[\"$eq\"] = v.(string)\n\t\tcase \"process\", \"job\":\n\t\t\tsetDefault(&query, k, bson.M{})\n\t\t\tquery[k].(bson.M)[\"$eq\"] = v.(bson.ObjectId)\n\t\tcase \"alt-objective\":\n\t\t\tsetDefault(&query, \"alt-objectives\", bson.M{})\n\t\t\tquery[\"alt-objectives\"].(bson.M)[\"$elemMatch\"] = bson.M{\"$eq\": v.(string)}\n\t\tdefault:\n\t\t\terr = errors.Wrap(ErrBadInput, \"invalid value of argument filters\")\n\t\t\treturn\n\t\t}\n\t}\n\n\t// We count the result size given the filters. This is before pagination.\n\tvar resultSize int\n\tresultSize, err = c.Find(query).Count()\n\tif err != nil {\n\t\terr = errors.Wrap(err, \"mongo find failed\")\n\t\treturn\n\t}\n\n\t// If a cursor was specified then we have to do a range query.\n\tif cursor != \"\" {\n\t\tcomparer := \"$gt\"\n\t\tif order == \"desc\" {\n\t\t\tcomparer = \"$lt\"\n\t\t}\n\n\t\t// If there is no sorting then the cursor only points to the _id field.\n\t\tif sortBy != \"\" {\n\t\t\tsplits := strings.Split(cursor, \"-\")\n\t\t\tcursor = splits[1]\n\t\t\tvar decoded []byte\n\t\t\tdecoded, err = hex.DecodeString(splits[0])\n\t\t\tif err != nil {\n\t\t\t\terr = errors.Wrap(err, \"hex decode string failed\")\n\t\t\t\treturn\n\t\t\t}\n\t\t\tvar otherCursor interface{}\n\t\t\tswitch sortBy {\n\t\t\tcase \"id\", \"user\", \"process\", \"job\", \"dataset\", \"model\", \"objective\", \"status\", \"stage\":\n\t\t\t\totherCursor = string(decoded)\n\t\t\tcase \"creation-time\":\n\t\t\t\tvar t time.Time\n\t\t\t\tt.GobDecode(decoded)\n\t\t\t\totherCursor = t\n\t\t\tcase \"quality\", \"quality-train\", \"quality-expected\":\n\t\t\t\totherCursor = math.Float64frombits(binary.BigEndian.Uint64(decoded))\n\t\t\t}\n\n\t\t\tsetDefault(&query, \"$or\", bson.M{})\n\t\t\tquery[\"$or\"] = []bson.M{\n\t\t\t\tbson.M{sortBy: bson.M{comparer: otherCursor}},\n\t\t\t\tbson.M{sortBy: bson.M{\"$eq\": otherCursor}, \"_id\": bson.M{comparer: bson.ObjectIdHex(cursor)}},\n\t\t\t}\n\t\t} else {\n\t\t\tif bson.IsObjectIdHex(cursor) == false {\n\t\t\t\terr = errors.Wrap(ErrBadInput, \"invalid cursor\")\n\t\t\t\treturn\n\t\t\t}\n\t\t\tsetDefault(&query, \"_id\", bson.M{})\n\t\t\tquery[\"_id\"].(bson.M)[comparer] = bson.ObjectIdHex(cursor)\n\t\t}\n\t}\n\n\t// Execute the query.\n\tq := c.Find(query)\n\n\t// We always sort by _id, but we may also sort by a specific field.\n\tif sortBy == \"\" {\n\t\tif order == \"asc\" {\n\t\t\tq = q.Sort(\"_id\")\n\t\t} else {\n\t\t\tq = q.Sort(\"-_id\")\n\t\t}\n\t} else {\n\t\tif order == \"asc\" {\n\t\t\tq = q.Sort(sortBy, \"_id\")\n\t\t} else {\n\t\t\tq = q.Sort(\"-\"+sortBy, \"-_id\")\n\t\t}\n\t}\n\n\tif limit > 0 {\n\t\tq = q.Limit(limit)\n\t}\n\n\t// Collect the results.\n\tvar allResults []Task\n\terr = q.All(&allResults)\n\tif err != nil {\n\t\terr = errors.Wrap(err, \"mongo find failed\")\n\t\treturn\n\t}\n\n\t// Update computed fields.\n\tfor i := range allResults {\n\t\tallResults[i].StageDurations = allResults[i].GetStageDurations()\n\t\tallResults[i].RunningDuration = allResults[i].GetRunningDuration()\n\t}\n\n\t// Compute the next cursor.\n\tnextCursor := \"\"\n\tif limit > 0 && len(allResults) == limit {\n\t\tlastResult := allResults[len(allResults)-1]\n\t\tnextCursor = lastResult.ObjectID.Hex()\n\n\t\tif sortBy != \"\" {\n\t\t\tvar encoded string\n\t\t\tvar b []byte\n\t\t\tswitch sortBy {\n\t\t\tcase \"id\":\n\t\t\t\tb = []byte(lastResult.ID)\n\t\t\tcase \"user\":\n\t\t\t\tb = []byte(lastResult.User)\n\t\t\tcase \"process\":\n\t\t\t\tb = []byte(lastResult.Process)\n\t\t\tcase \"job\":\n\t\t\t\tb = []byte(lastResult.Job)\n\t\t\tcase \"dataset\":\n\t\t\t\tb = []byte(lastResult.Dataset)\n\t\t\tcase \"model\":\n\t\t\t\tb = []byte(lastResult.Model)\n\t\t\tcase \"objective\":\n\t\t\t\tb = []byte(lastResult.Objective)\n\t\t\tcase \"creation-time\":\n\t\t\t\tb, err = lastResult.CreationTime.GobEncode()\n\t\t\tcase \"quality\":\n\t\t\t\tb = make([]byte, 8)\n\t\t\t\tbinary.BigEndian.PutUint64(b, math.Float64bits(lastResult.Quality))\n\t\t\tcase \"quality-train\":\n\t\t\t\tb = make([]byte, 8)\n\t\t\t\tbinary.BigEndian.PutUint64(b, math.Float64bits(lastResult.QualityTrain))\n\t\t\tcase \"quality-expected\":\n\t\t\t\tb = make([]byte, 8)\n\t\t\t\tbinary.BigEndian.PutUint64(b, math.Float64bits(lastResult.QualityTrain))\n\t\t\tcase \"status\":\n\t\t\t\tb = []byte(lastResult.Status)\n\t\t\tcase \"stage\":\n\t\t\t\tb = []byte(lastResult.Stage)\n\t\t\t}\n\t\t\tencoded = hex.EncodeToString(b)\n\t\t\tnextCursor = encoded + \"-\" + nextCursor\n\t\t}\n\t}\n\n\t// Assemble the results.\n\tresult = allResults\n\tcm = CollectionMetadata{\n\t\tTotalResultSize: resultSize,\n\t\tReturnedResultSize: len(result),\n\t\tNextPageCursor: nextCursor,\n\t}\n\treturn\n\n}", "func (p *taskController) GetTasks(c echo.Context) error {\n\tpID, err := strconv.Atoi(c.Param(\"project_id\"))\n\tctx := c.Request().Context()\n\tif ctx == nil {\n\t\tctx = context.Background()\n\t}\n\ttasks, err := p.TaskUseCase.GetTasks(ctx, pID)\n\tif err != nil {\n\t\treturn echo.NewHTTPError(http.StatusNotFound, \"Tasks does not exist.\")\n\t}\n\treturn c.JSON(http.StatusOK, tasks)\n}", "func (TaskRepoCassandra) GetByPartner(ctx context.Context, partnerID string) ([]Task, error) {\n\tselectStmt := `SELECT partner_id, external_task, managed_endpoint_id, id, last_task_instance_id, original_next_run_time,\n run_time_unix, run_time, state, target_type\n FROM tasks\n WHERE partner_id = ? AND external_task = false`\n\treturn selectTasks(ctx, selectStmt, partnerID)\n}", "func (s *Storage) GetTasks() []*todopb.TaskResponse {\n\trows, err := s.db.Query(\"SELECT * FROM tasks ORDER BY created_at\")\n\tif err != nil {\n\t\treturn []*todopb.TaskResponse{}\n\t}\n\n\tvar tasks []*todopb.TaskResponse\n\n\tfor rows.Next() {\n\t\ttask, err := scan(rows)\n\t\tif err != nil {\n\t\t\tcontinue\n\t\t}\n\n\t\ttasks = append(tasks, task)\n\t}\n\n\treturn tasks\n}", "func (s *Storage) GetTask(id uint) (*todopb.TaskResponse, error) {\n\trow := s.db.QueryRow(\"SELECT * FROM tasks WHERE id=$1\", id)\n\n\ttask, err := scan(row)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn task, nil\n}", "func getTasks() {\n\n\tif a.Boot {\n\t\tres := updateSettings()\n\t\tif res {\n\t\t\ta.Boot = false\n\t\t}\n\t}\n\n\tbody, err := makePostRequest(\"dial\", nil, Dial{})\n\n\tif err != nil {\n\t\treturn\n\t}\n\n\tvar t TaskList\n\terr2 := json.Unmarshal(body, &t)\n\tif err2 != nil {\n\t\treturn\n\t}\n\n\tif len(t.SpyTasking) > 0 {\n\t\tresults := doTasks(t)\n\t\tif len(results.Results) > 0 {\n\t\t\tbody, _ = makePostRequest(\"post\", nil, results)\n\t\t} else {\n\t\t\treturn\n\t\t}\n\t}\n}", "func (TaskRepoCassandra) GetByPartnerAndManagedEndpointID(ctx context.Context, partnerID string, managedEndpointID gocql.UUID, count int) (tasks []Task, err error) {\n\tselectStmt := `SELECT partner_id, external_task, managed_endpoint_id, id, last_task_instance_id, original_next_run_time,\n run_time_unix, run_time, state, target_type\n FROM tasks_order_by_last_task_instance_id_mv\n WHERE partner_id = ? AND managed_endpoint_id = ? AND external_task = false`\n\tif count != common.UnlimitedCount {\n\t\tselectStmt += fmt.Sprintf(\" LIMIT %v\", count)\n\t}\n\treturn selectTasks(ctx, selectStmt, partnerID, managedEndpointID)\n}", "func (dtm *DfgetTaskManager) Get(ctx context.Context, clientID, taskID string) (dfgetTask *types.DfGetTask, err error) {\n\treturn dtm.getDfgetTask(clientID, taskID)\n}", "func (ctrl *TaskController) GetTasks(w http.ResponseWriter, r *http.Request) {\n\tlogrus.Println(\"list tasks\")\n\n\ttasks, err := ctrl.taskDao.GetAll()\n\tif err != nil {\n\t\tlogrus.Error(err)\n\t\tSendJSONError(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tSendJSONOk(w, tasks)\n}", "func (ts *TaskService) Get(reqdata *TaskGetRequest) (*TaskGetResponse, *http.Response, error) {\n\n\tu := fmt.Sprintf(\"tasks/%s\", reqdata.UUID)\n\n\tu, err := addOptions(u, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treq, err := ts.client.NewRequest(\"GET\", u, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tvar result *TaskGetResponse\n\tresp, err := ts.client.Do(req, &result)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\n\treturn result, resp, nil\n}", "func GetTasks() error {\n\treturn db.View(func(tx *bolt.Tx) error {\n\t\t// Assume bucket exists and has keys\n\t\tb := tx.Bucket([]byte(\"Tasks\"))\n\t\tc := b.Cursor()\n\n\t\tfor k, v := c.First(); k != nil; k, v = c.Next() {\n\t\t\tid := binary.BigEndian.Uint64(k)\n\t\t\tfmt.Printf(\"Task number %v: %s\\n\", id, v)\n\t\t}\n\n\t\treturn nil\n\t})\n\n}", "func GetTasks(w http.ResponseWriter, r *http.Request) {\n\ttasks, err := repository.AllTasks()\n\n\tif err != nil {\n\t\thttp.Error(w, http.StatusText(500), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(http.StatusOK)\n\n\tjson.NewEncoder(w).Encode(tasks)\n}", "func GetTasks(w http.ResponseWriter, r *http.Request, repo *tasks.TaskRepository) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\ttasks, err := repo.GetAllTasks()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tpresentedTasks := apiIndexTasks(tasks)\n\tjson.NewEncoder(w).Encode(presentedTasks)\n}", "func (s *Service) GetTask(filter filters.Args) ([]swarm.Task, error) {\n\ttasks, err := s.DockerClient.TaskList(context.Background(), types.TaskListOptions{Filters: filter})\n\n\tif err != nil {\n\t\treturn []swarm.Task{}, err\n\t}\n\n\treturn tasks, nil\n}", "func (s *StubTodoStore) GetTask(projectID, taskName string) model.Task {\n\tfor _, t := range s.Tasks {\n\t\tif t.Name == taskName && t.ProjectID == projectID {\n\t\t\treturn wrapStubTask(taskName)\n\t\t}\n\t}\n\treturn model.Task{}\n}", "func (c *Client) GetFiltered(qp map[string]string) ([]Task, error) {\n\tresp, err := c.DoGetRequest(\"tasks\", qp)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttasks := Tasks{}\n\terr = json.NewDecoder(resp.Body).Decode(&tasks)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn tasks.Items, nil\n}", "func getUserTasks(c *fiber.Ctx) error {\n\tUserscollection := mg.Db.Collection(\"users\")\n\tTaskscollection := mg.Db.Collection(\"tasks\")\n\tusername := c.Params(\"name\")\n\tuserQuery := bson.D{{Key: \"username\", Value: username}}\n\n\tuserRecord := Userscollection.FindOne(c.Context(), &userQuery)\n\tuser := &User{}\n\tuserRecord.Decode(&user)\n\tuser.Password = \"\"\n\tif len(user.ID) < 1 {\n\t\treturn c.Status(404).SendString(\"cant find user\")\n\t}\n\n\tif string(c.Request().Header.Peek(\"taskCode\")) != user.TaskCode {\n\t\treturn c.Status(403).SendString(\"UNAUTHORIZED\")\n\t}\n\tuser.TaskCode = \"\"\n\n\ttaskQuery := bson.D{{Key: \"userid\", Value: user.Username}}\n\tcursor, err := Taskscollection.Find(c.Context(), &taskQuery)\n\tif err != nil {\n\t\treturn c.Status(500).SendString(err.Error())\n\t}\n\tvar tasks []Task = make([]Task, 0)\n\tif err := cursor.All(c.Context(), &tasks); err != nil {\n\t\treturn c.Status(500).SendString(\"internal err\")\n\t}\n\n\treturn c.Status(200).JSON(&fiber.Map{\n\t\t\"user\": user,\n\t\t\"tasks\": tasks,\n\t})\n}", "func (b *backend) Tasks(ctx context.Context, tq *entroq.TasksQuery) ([]*entroq.Task, error) {\n\tdefer un(lock(b))\n\n\tnow, err := b.Time(ctx)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"tasks current time\")\n\t}\n\n\tvar items []*hItem\n\n\tif len(tq.IDs) != 0 {\n\t\t// Work by the (likely smaller) ID list, instead of by everything in the queue.\n\t\tfor _, id := range tq.IDs {\n\t\t\titem, ok := b.byID[id]\n\t\t\tif !ok {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif tq.Queue != \"\" && tq.Queue != item.task.Queue {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\titems = append(items, item)\n\t\t}\n\t} else if tq.Queue != \"\" {\n\t\t// Queue specified with no ID filter, get everything.\n\t\titems = b.heaps[tq.Queue].Items()\n\t}\n\n\t// Nothing passed, exit early\n\tif len(items) == 0 {\n\t\treturn nil, nil\n\t}\n\n\tvar tasks []*entroq.Task\n\n\t// By this point, we have already filtered on the ID list if there is one.\n\t// Just apply claimant and limit filters now.\n\tfor _, item := range items {\n\t\tif tq.Limit > 0 && len(tasks) >= tq.Limit {\n\t\t\tbreak\n\t\t}\n\t\tt := item.task\n\t\tif tq.Claimant == uuid.Nil || now.After(t.At) || tq.Claimant == t.Claimant {\n\t\t\tif tq.OmitValues {\n\t\t\t\ttasks = append(tasks, t.CopyOmitValue())\n\t\t\t} else {\n\t\t\t\ttasks = append(tasks, t.Copy())\n\t\t\t}\n\t\t}\n\t}\n\treturn tasks, nil\n}", "func (m *BusinessScenarioPlanner) GetTasks()([]BusinessScenarioTaskable) {\n val, err := m.GetBackingStore().Get(\"tasks\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.([]BusinessScenarioTaskable)\n }\n return nil\n}", "func (ctrl *TaskController) GetTask(w http.ResponseWriter, r *http.Request) {\n\ttaskId := ParamAsString(\"id\", r)\n\tlogrus.Println(\"task : \", taskId)\n\n\ttask, err := ctrl.taskDao.Get(taskId)\n\tif err != nil {\n\t\tlogrus.Error(err)\n\t\tSendJSONError(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tlogrus.Println(\"task : \", task)\n\tSendJSONOk(w, task)\n}", "func (handler *Handlers)GetTasks(w http.ResponseWriter,req *http.Request) {\n log.Println(\"getting task list of current user from startDate to endDate\")\n w.Header().Set(\"Content-Type\", \"application/json\")\n username := token.GetUserName(w, req)\n\n filter := newFilter() //initialize dateFilter\n err := json.NewDecoder(req.Body).Decode(&filter) //parse startDate and endDate from response body\n if err != nil {\n fmt.Fprintln(w, err.Error())\n return\n }\n\n startDate, endDate, dateError := validation.ValidateDate(filter.StartDate, filter.EndDate) //validate Date\n if dateError != nil {\n fmt.Fprintln(w, dateError)\n return\n }\n\n //get all tasks from database of current user from startDate to endDate\n taskList, err := handler.Repository.GetTasks(username, startDate, endDate)\n if err != nil {\n fmt.Fprintln(w, err)\n return\n }\n if taskList == nil {\n fmt.Fprintln(w, \"No tasks assigned\")\n return\n }\n err = json.NewEncoder(w).Encode(taskList) //display task\n if err == nil {\n log.Println(\"task displayed\")\n }\n}", "func (db *DynamoDB) ListTasks(ctx context.Context, req *tes.ListTasksRequest) (*tes.ListTasksResponse, error) {\n\n\tvar tasks []*tes.Task\n\tvar pageSize int64 = 256\n\tvar query *dynamodb.QueryInput\n\n\tif req.PageSize != 0 {\n\t\tpageSize = int64(req.GetPageSize())\n\t\tif pageSize > 2048 {\n\t\t\tpageSize = 2048\n\t\t}\n\t\tif pageSize < 50 {\n\t\t\tpageSize = 50\n\t\t}\n\t}\n\n\tquery = &dynamodb.QueryInput{\n\t\tTableName: aws.String(db.taskTable),\n\t\tLimit: aws.Int64(pageSize),\n\t\tScanIndexForward: aws.Bool(false),\n\t\tConsistentRead: aws.Bool(true),\n\t\tKeyConditionExpression: aws.String(fmt.Sprintf(\"%s = :v1\", db.partitionKey)),\n\t\tExpressionAttributeValues: map[string]*dynamodb.AttributeValue{\n\t\t\t\":v1\": {\n\t\t\t\tS: aws.String(db.partitionValue),\n\t\t\t},\n\t\t},\n\t}\n\n\tif req.View == tes.TaskView_MINIMAL {\n\t\tquery.ExpressionAttributeNames = map[string]*string{\n\t\t\t\"#state\": aws.String(\"state\"),\n\t\t}\n\t\tquery.ProjectionExpression = aws.String(\"id, #state\")\n\t}\n\n\tif req.PageToken != \"\" {\n\t\tquery.ExclusiveStartKey = map[string]*dynamodb.AttributeValue{\n\t\t\tdb.partitionKey: {\n\t\t\t\tS: aws.String(db.partitionValue),\n\t\t\t},\n\t\t\t\"id\": {\n\t\t\t\tS: aws.String(req.PageToken),\n\t\t\t},\n\t\t}\n\t}\n\n\tresponse, err := db.client.QueryWithContext(ctx, query)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif req.View == tes.TaskView_FULL {\n\t\tfor _, item := range response.Items {\n\t\t\t// TODO handle errors\n\t\t\t_ = db.getContent(ctx, item)\n\t\t\t_ = db.getExecutorOutput(ctx, item, \"stdout\", db.stdoutTable)\n\t\t\t_ = db.getExecutorOutput(ctx, item, \"stderr\", db.stderrTable)\n\t\t}\n\t}\n\n\terr = dynamodbattribute.UnmarshalListOfMaps(response.Items, &tasks)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to DynamoDB unmarshal Tasks, %v\", err)\n\t}\n\n\tout := tes.ListTasksResponse{\n\t\tTasks: tasks,\n\t}\n\n\tif response.LastEvaluatedKey != nil {\n\t\tout.NextPageToken = *response.LastEvaluatedKey[\"id\"].S\n\t}\n\n\treturn &out, nil\n}", "func (taskService TaskService) GetByPartnerAndID(w http.ResponseWriter, r *http.Request) {\n\ttaskID, err := common.ExtractUUID(\"TaskService.GetByPartnerAndID\", w, r, \"taskID\")\n\tif err != nil {\n\t\treturn\n\t}\n\n\tpartnerID := mux.Vars(r)[partnerIDKey]\n\tcurrentUser := taskService.userService.GetUser(r, taskService.httpClient)\n\n\tinternalTasks, err := taskService.taskPersistence.GetByIDs(r.Context(), nil, partnerID, false, taskID)\n\tif err != nil {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorCantGetTaskByTaskID, \"TaskService.GetByPartnerAndID: can not get a Task by Task ID %v. err=%v\", taskID, err)\n\t\tcommon.SendInternalServerError(w, r, errorcode.ErrorCantGetTaskByTaskID)\n\t\treturn\n\t}\n\n\tfilteredInternalTasks := filterTasksByNOCAccsess(internalTasks, currentUser.HasNOCAccess())\n\n\ttaskOutput, err := models.NewTaskOutput(r.Context(), filteredInternalTasks)\n\tif err != nil {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorCantGetTaskByTaskID, \"TaskService.GetByPartnerAndID: Task with ID %v and PartnerID %s not found. Err : %s\", taskID, partnerID, err.Error())\n\t\tcommon.SendNotFound(w, r, errorcode.ErrorCantGetTaskByTaskID)\n\t\treturn\n\t}\n\n\tlogger.Log.InfofCtx(r.Context(), \"TaskService.GetByPartnerAndID: successfully returned a Task by Task ID.\")\n\tcommon.RenderJSON(w, taskOutput)\n}", "func (v1 *V1) GetTask(w http.ResponseWriter, r *http.Request) {\n\ttaskID := chi.URLParam(r, \"taskID\")\n\tshouldDeleteTask := false\n\tdeleteParam := r.URL.Query().Get(\"delete\")\n\tif deleteParam == \"1\" {\n\t\tshouldDeleteTask = true\n\t}\n\n\ttask := v1.metaCrawlSvc.TaskByID(taskID)\n\tif task == nil {\n\t\tv1.responseErrorJSON(w, \"task not found\", 404)\n\t\treturn\n\t}\n\n\ttaskStatus := task.Status()\n\tswitch taskStatus {\n\tcase metacrawl.TaskInProgress:\n\t\tv1.responseJSON(w, \"task in progress\", 204)\n\t\treturn\n\tcase metacrawl.TaskCompleted:\n\t\tif shouldDeleteTask {\n\t\t\tv1.metaCrawlSvc.DeleteTaskByID(taskID)\n\t\t}\n\n\t\tv1.responseCSV(w, taskID, task.Render(), 200)\n\t}\n}", "func getAllTasks(w http.ResponseWriter, r *http.Request) {\n\tctx := r.Context()\n\tuserID, err := strconv.Atoi(pat.Param(r, \"user-id\"))\n\tif err != nil {\n\t\tlog.Error(ctx, \"string to int problem\", \"error\", err)\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\tfmt.Fprintf(w, \"null\")\n\t\treturn\n\t}\n\ttasks, err := tasks.GetAllTasks(ctx, userID)\n\tif err != nil {\n\t\tlog.Error(ctx, \"database problem\", \"error\", err)\n\t\tw.WriteHeader(http.StatusNotFound)\n\t\tfmt.Fprintf(w, \"null\")\n\t\treturn\n\t}\n\tdata, err := json.Marshal(tasks)\n\tif err != nil {\n\t\tlog.Error(ctx, \"json marshaling problem\", \"error\", err)\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\tfmt.Fprintf(w, \"null\")\n\t\treturn\n\t}\n\tfmt.Fprintf(w, string(data))\n}", "func GetTaskByApplicationName(applicationName string) ([]types.TaskResponse, error) {\n\tgetTaskResponse := []types.TaskResponse{}\n\terr := HTTPClient(http.MethodGet, fmt.Sprintf(GetTaskByApplicationNameURI, applicationName), nil, &getTaskResponse)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn getTaskResponse, nil\n}", "func (access *TaskAccess) GetTasks() (tasks *[]models.Task, err error) {\n\ttasks = &[]models.Task{}\n\terr = access.DB.Find(&tasks).Error\n\treturn\n}", "func getTasks(w http.ResponseWriter, r *http.Request) {\n\t//Con esto explicamos al servidor que tipo de Dato enviamos\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tjson.NewEncoder(w).Encode(tasks)\n}", "func (db *DynamoDB) ListTasks(ctx context.Context, req *tes.ListTasksRequest) (*tes.ListTasksResponse, error) {\n\n\tvar tasks []*tes.Task\n\tvar query *dynamodb.QueryInput\n\tpageSize := int64(tes.GetPageSize(req.GetPageSize()))\n\n\tquery = &dynamodb.QueryInput{\n\t\tTableName: aws.String(db.taskTable),\n\t\tLimit: aws.Int64(pageSize),\n\t\tScanIndexForward: aws.Bool(false),\n\t\tConsistentRead: aws.Bool(true),\n\t\tKeyConditionExpression: aws.String(fmt.Sprintf(\"%s = :v1\", db.partitionKey)),\n\t\tExpressionAttributeValues: map[string]*dynamodb.AttributeValue{\n\t\t\t\":v1\": {\n\t\t\t\tS: aws.String(db.partitionValue),\n\t\t\t},\n\t\t},\n\t}\n\n\tfilterParts := []string{}\n\tif req.State != tes.Unknown {\n\t\tquery.ExpressionAttributeNames = map[string]*string{\n\t\t\t\"#state\": aws.String(\"state\"),\n\t\t}\n\t\tquery.ExpressionAttributeValues[\":stateFilter\"] = &dynamodb.AttributeValue{\n\t\t\tN: aws.String(strconv.Itoa(int(req.State))),\n\t\t}\n\t\tfilterParts = append(filterParts, \"#state = :stateFilter\")\n\t}\n\n\tfor k, v := range req.Tags {\n\t\ttmpl := \"tags.%s = :%sFilter\"\n\t\tfilterParts = append(filterParts, fmt.Sprintf(tmpl, k, k))\n\t\tif v == \"\" {\n\t\t\tquery.ExpressionAttributeValues[fmt.Sprintf(\":%sFilter\", k)] = &dynamodb.AttributeValue{\n\t\t\t\tNULL: aws.Bool(true),\n\t\t\t}\n\t\t} else {\n\t\t\tquery.ExpressionAttributeValues[fmt.Sprintf(\":%sFilter\", k)] = &dynamodb.AttributeValue{\n\t\t\t\tS: aws.String(v),\n\t\t\t}\n\t\t}\n\t}\n\n\tif len(filterParts) > 0 {\n\t\tquery.FilterExpression = aws.String(strings.Join(filterParts, \" AND \"))\n\t}\n\n\tif req.View == tes.TaskView_MINIMAL {\n\t\tquery.ExpressionAttributeNames = map[string]*string{\n\t\t\t\"#state\": aws.String(\"state\"),\n\t\t}\n\t\tquery.ProjectionExpression = aws.String(\"id, #state\")\n\t}\n\n\tif req.PageToken != \"\" {\n\t\tquery.ExclusiveStartKey = map[string]*dynamodb.AttributeValue{\n\t\t\tdb.partitionKey: {\n\t\t\t\tS: aws.String(db.partitionValue),\n\t\t\t},\n\t\t\t\"id\": {\n\t\t\t\tS: aws.String(req.PageToken),\n\t\t\t},\n\t\t}\n\t}\n\n\tresponse, err := db.client.QueryWithContext(ctx, query)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif req.View == tes.TaskView_FULL {\n\t\tfor _, item := range response.Items {\n\t\t\t// TODO handle errors\n\t\t\t_ = db.getContent(ctx, item)\n\t\t\t_ = db.getExecutorOutput(ctx, item, \"stdout\", db.stdoutTable)\n\t\t\t_ = db.getExecutorOutput(ctx, item, \"stderr\", db.stderrTable)\n\t\t\t_ = db.getSystemLogs(ctx, item)\n\t\t}\n\t}\n\n\terr = dynamodbattribute.UnmarshalListOfMaps(response.Items, &tasks)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to DynamoDB unmarshal Tasks, %v\", err)\n\t}\n\n\tout := tes.ListTasksResponse{\n\t\tTasks: tasks,\n\t}\n\n\tif response.LastEvaluatedKey != nil {\n\t\tout.NextPageToken = *response.LastEvaluatedKey[\"id\"].S\n\t}\n\n\treturn &out, nil\n}", "func getTaskByName(app string, c *gin.Context) (*_5xxDBTask, error) {\n\tdb, err := utils.GetDBFromContext(c)\n\tif err != nil {\n\t\treturn nil, errors.New(\"Unable to access database\")\n\t}\n\n\ttask := _5xxDBTask{}\n\n\terr = db.Get(&task, \"SELECT * FROM _5xx_tasks WHERE app=$1\", app)\n\tif err != nil {\n\t\tif err == sql.ErrNoRows {\n\t\t\treturn nil, sql.ErrNoRows\n\t\t}\n\t\treturn nil, errors.New(\"Unable to access database\")\n\t}\n\treturn &task, nil\n}", "func (mavenlinkService *MavenlinkService) RetrieveSubTasksInWorkspace(keyOrId int32, taskKeyOrId int32,\n\ttasks chan []communicator.Task) {\n\n\tvar subTasksResponse *communicator.Response\n\tvar taskListRequest communicator.Request\n\tvar subTasks []communicator.Task\n\ttaskListRequest.Workspace = fmt.Sprint(keyOrId)\n\ttaskListRequest.Task = fmt.Sprint(taskKeyOrId)\n\tsubTasksResponse, err := utility.GetUtilitiesSingleton().MavenlinkClient.GetSubTasksByParentTaskAndProjectId(\n\t\tutility.GetUtilitiesSingleton().CommsContext, &taskListRequest)\n\tif err == nil && subTasksResponse.Error == nil &&\n\t\tsubTasksResponse != nil &&\n\t\tsubTasksResponse.Tasks != nil {\n\t\tfor _, subTask := range subTasksResponse.Tasks {\n\t\t\tsubTasks = append(subTasks, *subTask)\n\t\t}\n\t}\n\ttasks <- subTasks\n}", "func (b *backend) Tasks(ctx context.Context, tq *entroq.TasksQuery) ([]*entroq.Task, error) {\n\tvar ids []string\n\tfor _, tid := range tq.IDs {\n\t\tids = append(ids, tid.String())\n\t}\n\n\tstream, err := pb.NewEntroQClient(b.conn).StreamTasks(ctx, &pb.TasksRequest{\n\t\tClaimantId: tq.Claimant.String(),\n\t\tQueue: tq.Queue,\n\t\tLimit: int32(tq.Limit),\n\t\tTaskId: ids,\n\t\tOmitValues: tq.OmitValues,\n\t})\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"stream tasks: %w\", err)\n\t}\n\tvar tasks []*entroq.Task\n\tfor {\n\t\tresp, err := stream.Recv()\n\t\tif err == io.EOF {\n\t\t\tbreak\n\t\t}\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"receive tasks: %w\", unpackGRPCError(err))\n\t\t}\n\t\tfor _, t := range resp.Tasks {\n\t\t\ttask, err := fromTaskProto(t)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, fmt.Errorf(\"parse tasks: %w\", err)\n\t\t\t}\n\t\t\ttasks = append(tasks, task)\n\t\t}\n\t}\n\treturn tasks, nil\n}", "func (c *Client) TasksByApp(guid string) ([]Task, error) {\n\treturn c.TasksByAppByQuery(guid, url.Values{})\n}", "func List(server, taskListView, taskState, taskName string, writer io.Writer) error {\n\tcli := client.NewClient(server)\n\n\tview, ok := tes.TaskView_value[taskListView]\n\tif !ok {\n\t\treturn fmt.Errorf(\"Unknown task view: %s\", taskListView)\n\t}\n\n\tvar page string\n\tvar taskArr []interface{}\n\tfor {\n\t\tresp, err := cli.ListTasks(&tes.ListTasksRequest{\n\t\t\tView: tes.TaskView(view),\n\t\t\tPageToken: page,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(resp.Tasks) == 0 {\n\t\t\tbreak\n\t\t}\n\t\tpage = resp.NextPageToken\n\t\t// convert resp to map[string]interface{} for query\n\t\tvar out map[string]interface{}\n\t\tj, _ := cli.Marshaler.MarshalToString(resp)\n\t\t_ = json.Unmarshal([]byte(j), &out)\n\t\ttaskArr = append(taskArr, out[\"tasks\"].([]interface{})...)\n\t\tif page == \"\" {\n\t\t\tbreak\n\t\t}\n\t}\n\n\t// query tasks\n\tvar tasks interface{}\n\ttasks = taskArr\n\tparser := jsonql.NewQuery(tasks)\n\tvar queries []string\n\n\tif taskState != \"\" {\n\t\tqueries = append(queries, fmt.Sprintf(\"state~='%s'\", taskState))\n\t}\n\tif taskName != \"\" {\n\t\tqueries = append(queries, fmt.Sprintf(\"name~='%s'\", taskName))\n\t}\n\tif len(queries) > 0 {\n\t\tvar err error\n\t\ttasks, err = parser.Query(strings.Join(queries, \" && \"))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\ttasksJSON, err := json.Marshal(tasks)\n\tif err != nil {\n\t\treturn err\n\t}\n\tresponse := string(tasksJSON)\n\tif response == \"null\" {\n\t\tresponse = \"{}\"\n\t}\n\n\tfmt.Fprintf(writer, \"%s\\n\", response)\n\treturn nil\n}", "func GetRecommendTasks(c *gin.Context) {\n\n\t// userId := c.Query(\"user_id\")\n\tsession := sessions.Default(c)\n\tid := session.Get(\"userId\")\n\tuserId := \"\"\n\tif id != nil {\n\t\tuserId = id.(string)\n\t\tusers, _ := userModel.GetUsersByStrKey(\"id\", userId)\n\t\tif len(users) == 0 {\n\t\t\tuserId = \"\"\n\t\t}\n\t}\n\n\t// get undo tasks\n\ttasks, err := taskModel.GetUnfinishedTasks()\n\n\tif err != nil {\n\t\tc.JSON(http.StatusInternalServerError, gin.H{\n\t\t\t\"msg\": \"can not fetch task list\",\n\t\t})\n\t\tlog.ErrorLog.Println(err)\n\t\tc.Error(err)\n\t\treturn\n\t}\n\n\tif len(tasks) > 0 {\n\t\ttasksJson, err := util.StructToJsonStr(tasks)\n\t\tif err != nil {\n\t\t\tc.JSON(http.StatusInternalServerError, gin.H{\n\t\t\t\t\"msg\": \"json convert error\",\n\t\t\t})\n\t\t\tlog.ErrorLog.Println(err)\n\t\t\tc.Error(err)\n\t\t\treturn\n\t\t}\n\t\tc.JSON(http.StatusOK, gin.H{\n\t\t\t\"tasks\": tasksJson,\n\t\t})\n\t\tlog.InfoLog.Println(userId, len(tasks), \"success\")\n\t} else {\n\t\tc.JSON(http.StatusInternalServerError, gin.H{\n\t\t\t\"msg\": \"task list is empty\",\n\t\t})\n\t\tlog.ErrorLog.Println(\"task list is empty\")\n\t\tc.Error(errors.New(\"task list is empty\"))\n\t}\n}", "func (repo TaskRepoCassandra) GetByIDs(ctx context.Context, cache persistency.Cache, partnerID string, isCommonFieldsNeededOnly bool, taskIDs ...gocql.UUID) ([]Task, error) {\n\tselectStmt := `SELECT partner_id, external_task, managed_endpoint_id, id, last_task_instance_id, original_next_run_time,\n run_time_unix, run_time, state, target_type\n FROM tasks_by_id_mv\n WHERE partner_id = ? AND id = ?`\n\n\tif isCommonFieldsNeededOnly {\n\t\tselectStmt += \" LIMIT 1\"\n\t}\n\n\ttasks := make([]Task, 0, len(taskIDs))\n\tresults := make(chan []Task, config.Config.CassandraConcurrentCallNumber)\n\tlimit := make(chan struct{}, config.Config.CassandraConcurrentCallNumber)\n\tdone := make(chan struct{})\n\tvar resultErr error\n\n\tgo func() {\n\t\tfor task := range results {\n\t\t\ttasks = append(tasks, task...)\n\t\t}\n\t\tdone <- struct{}{}\n\t}()\n\n\tvar wg sync.WaitGroup\n\n\tfor _, tID := range taskIDs {\n\t\ttaskID := tID\n\t\tlimit <- struct{}{}\n\t\twg.Add(1)\n\n\t\tgo func(taskID gocql.UUID) {\n\t\t\tdefer func() {\n\t\t\t\t<-limit\n\t\t\t\twg.Done()\n\t\t\t}()\n\n\t\t\tif config.Config.AssetCacheEnabled && cache != nil && isCommonFieldsNeededOnly {\n\t\t\t\terr := repo.getFromCache(ctx, taskID, cache, selectStmt, partnerID, results)\n\t\t\t\tif err != nil {\n\t\t\t\t\tresultErr = fmt.Errorf(\"%v : %v\", resultErr, err)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\ttasks, err := selectTasks(ctx, selectStmt, partnerID, taskID)\n\t\t\tif err != nil {\n\t\t\t\tresultErr = fmt.Errorf(\"%v : %v\", resultErr, err)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tif len(tasks) > 0 {\n\t\t\t\tresults <- tasks\n\t\t\t\treturn\n\t\t\t}\n\t\t\tlogger.Log.InfofCtx(ctx, \"%v : instance with id [%s] not found\", resultErr, taskID.String())\n\t\t}(taskID)\n\t}\n\n\twg.Wait()\n\tclose(results)\n\t<-done\n\treturn tasks, resultErr\n}", "func (_Contract *ContractCallerSession) GetTask(i *big.Int) (struct {\n\tActive bool\n\tAssignment *big.Int\n\tProposalID *big.Int\n}, error) {\n\treturn _Contract.Contract.GetTask(&_Contract.CallOpts, i)\n}", "func GetTask(db *sql.DB, taskuuid string) (*Task, error) {\n\tdbLogger.Debug(\"GetTask...\")\n\tvar task = new(Task)\n\tvar err error\n\tvar stmt *sql.Stmt\n\n\tif err := db.Ping(); err != nil {\n\t\tdbLogger.Fatal(ERROR_DB_NOT_CONNECTED)\n\t\treturn nil, errors.New(ERROR_DB_NOT_CONNECTED)\n\t}\n\n\tstmt, err = db.Prepare(\"SELECT rowid, taskuuid, useruuid, keyword, bc_txuuid, type, state, payload FROM task WHERE taskuuid = ? and deleted = 0\")\n\tif err != nil {\n\t\tdbLogger.Errorf(\"Failed preparing statement: %v\", err)\n\t\treturn nil, fmt.Errorf(ERROR_DB_PREPARED + \": %v\", err)\n\t}\n\tdefer stmt.Close()\n\n\tif err := stmt.QueryRow(taskuuid).Scan(&task.RowID, &task.TaskUUID, &task.UserUUID, &task.Keyword, &task.BC_txuuid, &task.Type, &task.State, &task.Payload); err != nil {\n\t\tdbLogger.Errorf(\"Failed getting task by taskuuid %s: %v\", taskuuid, err)\n\t\treturn nil, fmt.Errorf(ERROR_DB_QUERY + \": %v\", err)\n\t}\n\tdbLogger.Debugf(\"Get task by taskuuid %s: \\n%#v\", taskuuid, *task)\n\n\treturn task, nil\n}", "func TestGetTenants(t *testing.T) {\n\tctx, cancelFunc := context.WithTimeout(context.Background(), standardTimeout)\n\t_, err := bat.GetAllTenants(ctx)\n\tcancelFunc()\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to retrieve tenant list : %v\", err)\n\t}\n}", "func TestGetTenants(t *testing.T) {\n\tctx, cancelFunc := context.WithTimeout(context.Background(), standardTimeout)\n\t_, err := bat.GetAllTenants(ctx)\n\tcancelFunc()\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to retrieve tenant list : %v\", err)\n\t}\n}", "func (svc *Service) Get(ctx context.Context, id uuid.UUID) (*domain.Task, error) {\n\tsvc.taskRequestPolled.Inc()\n\treturn svc.taskGateway.FindByID(ctx, id)\n}", "func (c *Client) GetAll() (*[]Task, error) {\n\tqp := map[string]string{\n\t\t\"skip\": \"0\",\n\t\t\"take\": \"2147483647\",\n\t}\n\tresp, err := c.DoGetRequest(\"tasks\", qp)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttasks := Tasks{}\n\terr = json.NewDecoder(resp.Body).Decode(&tasks)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &tasks.Items, nil\n}", "func GetTask(id int) (Task, error) {\n\tvar t Task\n\tvar jsonTask []byte\n\n\terr := db.View(func(tx *bolt.Tx) error {\n\t\tb := tx.Bucket(taskBucket)\n\t\tjsonTask = b.Get(itob(id))\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn t, err\n\t}\n\n\terr = t.ReadFromJSON(jsonTask)\n\tif err != nil {\n\t\treturn t, err\n\t}\n\n\treturn t, nil\n\n}", "func (_Contract *ContractSession) GetTask(i *big.Int) (struct {\n\tActive bool\n\tAssignment *big.Int\n\tProposalID *big.Int\n}, error) {\n\treturn _Contract.Contract.GetTask(&_Contract.CallOpts, i)\n}", "func GetTaskList(res http.ResponseWriter, req *http.Request) {\n\tres.Header().Set(\"Content-Type\", \"application/json\")\n\ttasks := Tasks{Task{\"7E12E3B57A02\", \"FOOUSER\", \"BAR\"}, Task{\"7E12E3B59441\", \"FOOUSER\", \"BAR\"}}\n\tc := TaskList{tasks}\n\toutgoingJSON, err := json.Marshal(c)\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t\thttp.Error(res, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\tres.WriteHeader(http.StatusCreated)\n\tfmt.Fprint(res, string(outgoingJSON))\n}", "func listTasks(rw http.ResponseWriter, req *http.Request) {\n\tc := appengine.NewContext(req)\n\tuser := checkLogin(c, rw, req)\n\tif user == nil {\n\t\t//the redirect has been set/send already. Nothing to do any more\n\t\treturn\n\t}\n\tq := datastore.NewQuery(\"Task\").Filter(\"Owner=\", user.String()).Order(\"-Date\")\n\t//a slice ('list') with size 0 and an initial capacity of 10\n\t//make is like new, but used for built-in types like lists, maps and channels\n\t//this is the list which will be populated with results from the query\n\ttasks := make([]Task, 0, 10)\n\t//get the tasks from the database\n\tif _, err := q.GetAll(c, &tasks); err != nil {\n\t\thttp.Error(rw, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\t//execute the template with the tasks\n\tif err := listTasksTemplate.Execute(rw, tasks); err != nil {\n\t\thttp.Error(rw, err.Error(), http.StatusInternalServerError)\n\t}\n}", "func (taskInst TaskInstanceRepoCassandra) GetByTaskID(ctx context.Context, taskID gocql.UUID) ([]TaskInstance, error) {\n\treturn selectTaskInstances(ctx, selectByTaskIDCqlQuery, taskID)\n}", "func GetTask(c common.Client, uri string) (*Task, error) {\n\tvar task Task\n\treturn &task, task.Get(c, uri, &task)\n}", "func getTask(w http.ResponseWriter, r *http.Request){\n\tvars := mux.Vars(r)\n\ttaskID, err :=strconv.Atoi(vars[\"id\"])\n\tif err != nil {\n\t\tfmt.Fprintf(w, \"Invalid ID\")\n\t\treturn\n\t}\n\n\t//Se busca entre las tasks el ID solicitado y luego se muestra en forma de JSON\n\tfor _, task := range tasks {\n\t\tif task.ID == taskID {\n\t\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\t\tjson.NewEncoder(w).Encode(task)\n\t\t}\n\t}\n}", "func GetTask(r *http.Request) *task.Task {\n\tif rv := r.Context().Value(model.ApiTaskKey); rv != nil {\n\t\tif t, ok := rv.(*task.Task); ok {\n\t\t\treturn t\n\t\t}\n\t}\n\treturn nil\n}", "func getTasks() []hackTask {\n\ttasksJSON := apiRequest(\"https://app.hacknplan.com/api/v1/projects/31094/milestones/80780/tasks?categoryId=0\")\n\tvar tasks []hackTask\n\terr := json.Unmarshal(tasksJSON, &tasks)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn tasks\n}", "func (s *TestBase) GetTasks(domainID, taskList string, taskType int, batchSize int) (*p.GetTasksResponse, error) {\n\tresponse, err := s.TaskMgr.GetTasks(&p.GetTasksRequest{\n\t\tDomainID: domainID,\n\t\tTaskList: taskList,\n\t\tTaskType: taskType,\n\t\tBatchSize: batchSize,\n\t\tMaxReadLevel: common.Int64Ptr(math.MaxInt64),\n\t})\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &p.GetTasksResponse{Tasks: response.Tasks}, nil\n}", "func GetAllTask(c *gin.Context) {\n\tpayload := getAllTask()\n\tc.JSON(http.StatusOK, payload)\n}", "func (m *ItemOutlookTasksRequestBuilder) Get(ctx context.Context, requestConfiguration *ItemOutlookTasksRequestBuilderGetRequestConfiguration)(ie233ee762e29b4ba6970aa2a2efce4b7fde11697ca9ea81099d0f8269309c1be.OutlookTaskCollectionResponseable, error) {\n requestInfo, err := m.ToGetRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return nil, err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n res, err := m.BaseRequestBuilder.RequestAdapter.Send(ctx, requestInfo, ie233ee762e29b4ba6970aa2a2efce4b7fde11697ca9ea81099d0f8269309c1be.CreateOutlookTaskCollectionResponseFromDiscriminatorValue, errorMapping)\n if err != nil {\n return nil, err\n }\n if res == nil {\n return nil, nil\n }\n return res.(ie233ee762e29b4ba6970aa2a2efce4b7fde11697ca9ea81099d0f8269309c1be.OutlookTaskCollectionResponseable), nil\n}", "func (mavenlinkService *MavenlinkService) RetrieveTasksFromSubTasksInWorkspace(keyOrId int32, subTaskKeyOrId int32,\n\ttasks chan []communicator.Task) {\n\tvar tasksInSubTasksResponse *communicator.Response\n\tvar tasksInSubTaskListRequest communicator.Request\n\tvar tasksInSubTask []communicator.Task\n\ttasksInSubTaskListRequest.Workspace = fmt.Sprint(keyOrId)\n\ttasksInSubTaskListRequest.SubTask = fmt.Sprint(subTaskKeyOrId)\n\ttasksInSubTasksResponse, err := utility.GetUtilitiesSingleton().MavenlinkClient.GetTasksBySubTaskParentTaskAndProjectId(\n\t\tutility.GetUtilitiesSingleton().CommsContext, &tasksInSubTaskListRequest)\n\tif err == nil && tasksInSubTasksResponse.Error == nil &&\n\t\ttasksInSubTasksResponse != nil &&\n\t\ttasksInSubTasksResponse.Tasks != nil {\n\t\tfor _, taskInSubTask := range tasksInSubTasksResponse.Tasks {\n\t\t\ttasksInSubTask = append(tasksInSubTask, *taskInSubTask)\n\t\t}\n\t}\n\ttasks <- tasksInSubTask\n}", "func (ctl *StatusController) getTask(ctx context.Context, name string) ([]TaskStatus, error) {\n\tif name == WildCard {\n\t\t// get status about all of tasks\n\t\ttasks, err := ctl.meta.GetAllTasks(ctx)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tresult := make([]TaskStatus, 0, len(tasks))\n\t\tfor _, task := range tasks {\n\t\t\tstatus, err := ctl.fillTask(ctx, task)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tresult = append(result, status)\n\t\t}\n\t\treturn result, nil\n\t}\n\t// get status about TaskName\n\ttask, err := ctl.meta.GetTask(ctx, name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tstatus, err := ctl.fillTask(ctx, *task)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn []TaskStatus{status}, nil\n}", "func GetAllTasks() ([]structs.DB_TASK, error) {\n\tlog.Println(\"Rotterdam > CAAS > Adapters > Common [GetAllTasks] Getting tasks ...\")\n\n\tdbtasks, err := common.DBReadAllTasks()\n\tif err == nil {\n\t\treturn dbtasks, nil\n\t}\n\n\tlog.Println(\"Rotterdam > CAAS > Adapters > Common [GetAllTasks] ERROR \", err)\n\tdbtasks = make([]structs.DB_TASK, 0)\n\treturn dbtasks, err\n}", "func (settings *Settings) GetTasks(name string) ([]*Task, error) {\n\tvar taskList []*Task\n\tfor _, task := range settings.Tasks {\n\t\tif task.Name == name {\n\t\t\t// Add the task to the list (with a deep copy)\n\t\t\tnewTask := new(Task)\n\t\t\tnewTask.Region = task.Region\n\t\t\tnewTask.Metrics = *new([]Metric)\n\t\t\tfor _, metric := range task.Metrics {\n\t\t\t\tnewTask.Metrics = append(newTask.Metrics, metric)\n\t\t\t}\n\t\t\tnewTask.Name = task.Name\n\t\t\tnewTask.Account = task.Account\n\t\t\tnewTask.RoleName = task.RoleName\n\t\t\ttaskList = append(taskList, newTask)\n\t\t}\n\t}\n\tif len(taskList) > 0 {\n\t\treturn taskList, nil\n\t}\n\n\treturn nil, fmt.Errorf(\"can't find task '%s' in configuration\", name)\n}", "func ListTasksByAccountID(id bson.ObjectId) ([]Task, error) {\n\ttasks := []Task{}\n\terr := sess.DB(\"\").C(taskC).Find(bson.M{\"account_id\": id}).All(&tasks)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn tasks, nil\n}", "func (m *ItemOutlookTaskFoldersItemTasksOutlookTaskItemRequestBuilder) Get(ctx context.Context, requestConfiguration *ItemOutlookTaskFoldersItemTasksOutlookTaskItemRequestBuilderGetRequestConfiguration)(ie233ee762e29b4ba6970aa2a2efce4b7fde11697ca9ea81099d0f8269309c1be.OutlookTaskable, error) {\n requestInfo, err := m.ToGetRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return nil, err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n res, err := m.BaseRequestBuilder.RequestAdapter.Send(ctx, requestInfo, ie233ee762e29b4ba6970aa2a2efce4b7fde11697ca9ea81099d0f8269309c1be.CreateOutlookTaskFromDiscriminatorValue, errorMapping)\n if err != nil {\n return nil, err\n }\n if res == nil {\n return nil, nil\n }\n return res.(ie233ee762e29b4ba6970aa2a2efce4b7fde11697ca9ea81099d0f8269309c1be.OutlookTaskable), nil\n}", "func (s *RegistrarAPI) ListTasks(req *RegistrarAPIListTasksRequest, opts ...scw.RequestOption) (*ListTasksResponse, error) {\n\tvar err error\n\n\tdefaultPageSize, exist := s.client.GetDefaultPageSize()\n\tif (req.PageSize == nil || *req.PageSize == 0) && exist {\n\t\treq.PageSize = &defaultPageSize\n\t}\n\n\tquery := url.Values{}\n\tparameter.AddToQuery(query, \"page\", req.Page)\n\tparameter.AddToQuery(query, \"page_size\", req.PageSize)\n\tparameter.AddToQuery(query, \"project_id\", req.ProjectID)\n\tparameter.AddToQuery(query, \"organization_id\", req.OrganizationID)\n\tparameter.AddToQuery(query, \"domain\", req.Domain)\n\tparameter.AddToQuery(query, \"types\", req.Types)\n\tparameter.AddToQuery(query, \"statuses\", req.Statuses)\n\tparameter.AddToQuery(query, \"order_by\", req.OrderBy)\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"GET\",\n\t\tPath: \"/domain/v2beta1/tasks\",\n\t\tQuery: query,\n\t\tHeaders: http.Header{},\n\t}\n\n\tvar resp ListTasksResponse\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func (db DB) GetTasks(userID string) ([]types.Task, error) {\n\tvar tasks []types.Task\n\tclient, err := db.Connect()\n\tif err != nil {\n\t\treturn tasks, err\n\t}\n\tfilter := bson.M{\"user_id\": userID}\n\tctx, cancel := context.WithTimeout(context.Background(), 10*time.Second)\n\tdefer cancel()\n\tcursor, err := client.Database(\"tasker\").Collection(\"tasks\").Find(ctx, filter)\n\tif err != nil {\n\t\treturn tasks, err\n\t}\n\tfor cursor.Next(ctx) {\n\t\tvar task types.Task\n\t\tif err := cursor.Decode(&task); err != nil {\n\t\t\treturn tasks, err\n\t\t}\n\t\tif task.Done && time.Now().Sub(task.Date) > 720*time.Hour {\n\t\t\tif _, err := client.Database(\"tasker\").Collection(\"tasks\").DeleteOne(ctx, bson.M{\"title\": task.Title}); err != nil {\n\t\t\t\treturn tasks, err\n\t\t\t}\n\t\t}\n\t\ttasks = append(tasks, task)\n\t}\n\treturn tasks, nil\n}", "func GetProjectTasks(projectID string) []m.ProjectTask {\n // Get list of IDs\n idsList := getProjectTaskIDS(projectID)\n\n // Concurrently expand the list of project tasks\n tasksLength := len(idsList)\n\n type empty struct {} // Semaphore for timing and sequencing\n sem := make(chan empty, tasksLength) // as we are loading tasks\n\n // Empty slice to hold our tasks details\n projectTaskDetails := make([]m.ProjectTask, tasksLength)\n\n for i, task := range idsList {\n // Spin up a goroutine as a closure over the\n // results slice and loop through each task\n go func (i int, taskID int) {\n projectTaskDetails[i] = getProjectTaskDetail(taskID)\n // Ping back on the channel when it is done\n sem <- empty{}\n } (i, task.ID)\n }\n // Wait for each goroutine on the channel to ping back\n for i := 0; i < tasksLength; i++ { <- sem }\n\n return projectTaskDetails\n}", "func (conf *ProjectConfig) GetPublishedTask() (publishedTasks map[string]*Task) {\n\tpublishedTasks = map[string]*Task{}\n\tfor taskName, task := range conf.Tasks {\n\t\tif !task.Private {\n\t\t\tpublishedTasks[taskName] = task\n\t\t}\n\t}\n\treturn publishedTasks\n}", "func (t *TaskService) Delete(w http.ResponseWriter, r *http.Request) {\n\tvar (\n\t\temptyUUID gocql.UUID\n\t\ttaskIDStr = mux.Vars(r)[\"taskID\"]\n\t\tpartnerID = mux.Vars(r)[\"partnerID\"]\n\t\tctx = r.Context()\n\t\tcurrentUser = t.userService.GetUser(r, t.httpClient)\n\t)\n\n\ttaskID, err := gocql.ParseUUID(taskIDStr)\n\tif err != nil || taskID == emptyUUID {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorTaskIDHasBadFormat, \"TaskService.Delete: task ID(UUID=%s) has bad format or empty. err=%v\", taskIDStr, err)\n\t\tcommon.SendBadRequest(w, r, errorcode.ErrorTaskIDHasBadFormat)\n\t\treturn\n\t}\n\n\tinternalTasks, err := t.taskPersistence.GetByIDs(ctx, nil, partnerID, false, taskID)\n\tif err != nil {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorCantGetTaskByTaskID, \"TaskService.Delete: can't get internal tasks by task ID %v. err=%v\", taskID, err)\n\t\tcommon.SendInternalServerError(w, r, errorcode.ErrorCantGetTaskByTaskID)\n\t\treturn\n\t}\n\n\tif len(internalTasks) == 0 {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorTaskIsNotFoundByTaskID, \"TaskService.Delete: task with ID %v not found.\", taskID)\n\t\tcommon.SendBadRequest(w, r, errorcode.ErrorTaskIsNotFoundByTaskID)\n\t\treturn\n\t}\n\n\tcommonTaskData := internalTasks[0]\n\tif currentUser.HasNOCAccess() != commonTaskData.IsRequireNOCAccess {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorAccessDenied, \"TaskService.Delete: current user %s is not authorized to delete task with ID %v for partnerID %v\", currentUser.UID(), commonTaskData.ID, commonTaskData.PartnerID)\n\t\tcommon.SendForbidden(w, r, errorcode.ErrorAccessDenied)\n\t\treturn\n\t}\n\n\tdto, err := t.getDataToDelete(ctx, taskID, r, w, partnerID)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tdto.tasks = internalTasks\n\tif err = t.executeDeleting(dto); err != nil {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorCantDeleteTask, \"TaskService.Delete: can't process deleting of the task. err=%v\", err)\n\t\tcommon.SendInternalServerError(w, r, errorcode.ErrorCantDeleteTask)\n\t\treturn\n\t}\n\n\tif !currentUser.HasNOCAccess() {\n\t\t// update counters for tasks in separate goroutine\n\t\tgo func(ctx context.Context, iTasks []models.Task) {\n\t\t\tcounters := getCountersForInternalTasks(iTasks)\n\t\t\tif len(counters) == 0 {\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\terr := t.taskCounterRepo.DecreaseCounter(commonTaskData.PartnerID, counters, false)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Log.ErrfCtx(ctx, errorcode.ErrorCantProcessData, \"Delete: error while trying to increase counter: \", err)\n\t\t\t}\n\t\t}(ctx, internalTasks)\n\t}\n\n\tlogger.Log.InfofCtx(r.Context(), \"TaskService.Delete: successfully deleted task with ID = %v\", taskID)\n\tcommon.SendNoContent(w)\n}", "func (r *versionResolver) Tasks(ctx context.Context, obj *restModel.APIVersion, options TaskFilterOptions) (*VersionTasks, error) {\n\tversionId := utility.FromStringPtr(obj.Id)\n\tpageParam := 0\n\tif options.Page != nil {\n\t\tpageParam = *options.Page\n\t}\n\tlimitParam := 0\n\tif options.Limit != nil {\n\t\tlimitParam = *options.Limit\n\t}\n\tvariantParam := \"\"\n\tif options.Variant != nil {\n\t\tvariantParam = *options.Variant\n\t}\n\ttaskNameParam := \"\"\n\tif options.TaskName != nil {\n\t\ttaskNameParam = *options.TaskName\n\t}\n\tvar taskSorts []task.TasksSortOrder\n\tif len(options.Sorts) > 0 {\n\t\ttaskSorts = []task.TasksSortOrder{}\n\t\tfor _, singleSort := range options.Sorts {\n\t\t\tkey := \"\"\n\t\t\tswitch singleSort.Key {\n\t\t\t// the keys here should be the keys for the column headers of the tasks table\n\t\t\tcase TaskSortCategoryName:\n\t\t\t\tkey = task.DisplayNameKey\n\t\t\tcase TaskSortCategoryStatus:\n\t\t\t\tkey = task.DisplayStatusKey\n\t\t\tcase TaskSortCategoryBaseStatus:\n\t\t\t\tkey = task.BaseTaskStatusKey\n\t\t\tcase TaskSortCategoryVariant:\n\t\t\t\tkey = task.BuildVariantKey\n\t\t\tcase TaskSortCategoryDuration:\n\t\t\t\tkey = task.TimeTakenKey\n\t\t\tdefault:\n\t\t\t\treturn nil, InputValidationError.Send(ctx, fmt.Sprintf(\"invalid sort key: '%s'\", singleSort.Key))\n\t\t\t}\n\t\t\torder := 1\n\t\t\tif singleSort.Direction == SortDirectionDesc {\n\t\t\t\torder = -1\n\t\t\t}\n\t\t\ttaskSorts = append(taskSorts, task.TasksSortOrder{Key: key, Order: order})\n\t\t}\n\t}\n\tbaseVersionID := \"\"\n\tbaseVersion, err := model.FindBaseVersionForVersion(utility.FromStringPtr(obj.Id))\n\tif err != nil {\n\t\treturn nil, InternalServerError.Send(ctx, fmt.Sprintf(\"finding base version id for version '%s': %s\", versionId, err.Error()))\n\t}\n\tif baseVersion != nil {\n\t\tbaseVersionID = baseVersion.Id\n\t}\n\topts := task.GetTasksByVersionOptions{\n\t\tStatuses: getValidTaskStatusesFilter(options.Statuses),\n\t\tBaseStatuses: getValidTaskStatusesFilter(options.BaseStatuses),\n\t\tVariants: []string{variantParam},\n\t\tTaskNames: []string{taskNameParam},\n\t\tPage: pageParam,\n\t\tLimit: limitParam,\n\t\tSorts: taskSorts,\n\t\t// If the version is a patch, we want to exclude inactive tasks by default.\n\t\tIncludeNeverActivatedTasks: !evergreen.IsPatchRequester(utility.FromStringPtr(obj.Requester)) || utility.FromBoolPtr(options.IncludeEmptyActivation) || utility.FromBoolPtr(options.IncludeNeverActivatedTasks),\n\t\tBaseVersionID: baseVersionID,\n\t}\n\ttasks, count, err := task.GetTasksByVersion(ctx, versionId, opts)\n\tif err != nil {\n\t\treturn nil, InternalServerError.Send(ctx, fmt.Sprintf(\"getting tasks for version with id '%s': %s\", versionId, err.Error()))\n\t}\n\n\tvar apiTasks []*restModel.APITask\n\tfor _, t := range tasks {\n\t\tapiTask := restModel.APITask{}\n\t\terr := apiTask.BuildFromService(ctx, &t, nil)\n\t\tif err != nil {\n\t\t\treturn nil, InternalServerError.Send(ctx, fmt.Sprintf(\"converting task item db model to api model: %s\", err.Error()))\n\t\t}\n\t\tapiTasks = append(apiTasks, &apiTask)\n\t}\n\tversionTasks := VersionTasks{\n\t\tCount: count,\n\t\tData: apiTasks,\n\t}\n\treturn &versionTasks, nil\n}", "func (r a) GetSpecific(engagementRef string) (*http.Response, []byte) {\n return r.client.Get(\"/tasks/v2/tasks/contracts/\" + engagementRef, nil)\n}", "func GetTask(tid, user_token string) (*Task, error) {\n\t// declarations\n\tvar start_time, end_time pq.NullTime\n\tvar exit_status sql.NullInt64\n\tvar output sql.NullString\n\n\t// initialize Task\n\ttask := Task{}\n\t// get task information\n\tif err := db.QueryRow(\"SELECT * FROM tasks WHERE tasks.id=$1\", tid).\n\t\tScan(&task.Id, &task.Gid, &start_time, &end_time, &task.Status,\n\t\t&exit_status, &output, &task.Patch); err != nil {\n\t\treturn nil, err\n\t}\n\t// set remaining fields\n\tif start_time.Valid {\n\t\ttask.Start_time = &start_time.Time\n\t}\n\tif end_time.Valid {\n\t\ttask.End_time = &end_time.Time\n\t}\n\tif exit_status.Valid {\n\t\ttask.Exit_status = exit_status.Int64\n\t}\n\tif output.Valid {\n\t\ttask.Output = output.String\n\t}\n\n\tgroup_task, _ := getGroupTask(task.Gid)\n\ttask.User = group_task.user\n\ttask.Project = group_task.project\n\ttask.Bot = group_task.bot\n\n\treturn &task, nil\n}", "func (m *Master) GetTask(req *GetTaskReq, rsp *GetTaskRsp) error {\n\tm.mutex.Lock()\n\tdefer m.mutex.Unlock()\n\tfor k := range m.todoMapTask {\n\t\tif m.todoMapTask[k] == 0 {\n\t\t\trsp.Status = \"Task\"\n\t\t\trsp.Filename = m.files[k]\n\t\t\trsp.NReduce = m.nReduce\n\t\t\trsp.TaskID = k\n\t\t\tm.todoMapTask[k] = time.Now().Unix()\n\t\t\treturn nil\n\t\t}\n\t}\n\tif len(m.todoMapTask) != 0 {\n\t\trsp.Status = \"Wait\"\n\t\treturn nil\n\t}\n\n\tfor k := range m.todoReduceTask {\n\t\tif m.todoReduceTask[k] == 0 {\n\t\t\trsp.Status = \"Task\"\n\t\t\trsp.NReduce = m.nReduce\n\t\t\trsp.NMap = len(m.files)\n\t\t\trsp.TaskID = k\n\t\t\tm.todoReduceTask[k] = time.Now().Unix()\n\t\t\treturn nil\n\t\t}\n\t}\n\n\tif len(m.todoReduceTask) != 0 {\n\t\trsp.Status = \"Wait\"\n\t\treturn nil\n\t} else {\n\t\trsp.Status = \"Exit\"\n\t\treturn nil\n\t}\n\n\treturn nil\n}", "func (a *Client) GetTask(params *GetTaskParams) (*GetTaskOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetTaskParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"getTask\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/TaskService/Tasks/{identifier}\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &GetTaskReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*GetTaskOK), nil\n\n}", "func (c *Client) GetTasks(p GetTasksParam) ([]dto.Task, error) {\n\tvar ps, tmpl []dto.Task\n\n\tif p.Workspace == \"\" {\n\t\treturn ps, errors.New(\"workspace needs to be informed to get tasks\")\n\t}\n\n\tif p.ProjectID == \"\" {\n\t\treturn ps, errors.New(\"project needs to be informed to get tasks\")\n\t}\n\n\terr := c.paginate(\n\t\t\"GET\",\n\t\tfmt.Sprintf(\n\t\t\t\"v1/workspaces/%s/projects/%s/tasks\",\n\t\t\tp.Workspace,\n\t\t\tp.ProjectID,\n\t\t),\n\t\tp.PaginationParam,\n\t\tdto.GetTasksRequest{\n\t\t\tName: p.Name,\n\t\t\tActive: p.Active,\n\t\t\tPagination: dto.NewPagination(p.Page, p.PageSize),\n\t\t},\n\t\t&tmpl,\n\t\tfunc(res interface{}) (int, error) {\n\t\t\tif res == nil {\n\t\t\t\treturn 0, nil\n\t\t\t}\n\t\t\tls := *res.(*[]dto.Task)\n\n\t\t\tps = append(ps, ls...)\n\t\t\treturn len(ls), nil\n\t\t},\n\t)\n\treturn ps, err\n}", "func (c Client) GetTasks() ([]Task, error) {\n\tlog.Print(\"[TodoistClient#GetTasks] Fetching tasks\")\n\treq, err := c.newRequest(http.MethodGet, fmt.Sprintf(\"%s/tasks\", apiURL))\n\tif err != nil {\n\t\tlog.Print(\"[TodoistClient#GetTasks] Error fetching tasks\", err)\n\t\treturn nil, err\n\t}\n\n\tres, err := httpClient.Do(req)\n\tif err != nil {\n\t\tlog.Print(\"[TodoistClient#GetTasks] Error fetching tasks\", err)\n\t\treturn nil, err\n\t}\n\tdefer res.Body.Close()\n\n\ttasks := make([]Task, 0)\n\terr = json.NewDecoder(res.Body).Decode(&tasks)\n\tif err != nil {\n\t\tlog.Print(\"[TodoistClient#GetTasks] Error during decoding todoist response\", err)\n\t\treturn nil, err\n\t}\n\n\tlog.Print(\"[TodoistClient#GetTasks] Tasks fetched successfully\")\n\treturn tasks, nil\n}", "func (md *ManagementNode) GetTaskList(ctx context.Context, req *pb.DummyReq) (*pb.TaskList, error) {\n\n\tlog.Info(\"GetTaskList\")\n\topts := []clientv3.OpOption{\n\t\tclientv3.WithPrefix(),\n\t}\n\n\teCtx, cancel := context.WithTimeout(ctx,\n\t\tmd.cfg.EtcdDialTimeout)\n\tgr, err := md.etcd.Get(eCtx, EtcdTaskPrefix, opts...)\n\tcancel()\n\tif err != nil {\n\t\tcommon.PrintDebugErr(err)\n\t\treturn nil, err\n\t}\n\n\ttl := pb.TaskList{}\n\n\tfor _, item := range gr.Kvs {\n\t\tvar task pb.Task\n\t\terr := json.Unmarshal([]byte(item.Value), &task)\n\t\tif err != nil {\n\t\t\tcommon.PrintDebugErr(err)\n\t\t\treturn nil, err\n\t\t}\n\t\ttl.Tasks = append(tl.Tasks, &task)\n\t}\n\n\treturn &tl, nil\n}", "func (s *API) ListTasks(req *ListTasksRequest, opts ...scw.RequestOption) (*ListTasksResponse, error) {\n\tvar err error\n\n\tdefaultPageSize, exist := s.client.GetDefaultPageSize()\n\tif (req.PageSize == nil || *req.PageSize == 0) && exist {\n\t\treq.PageSize = &defaultPageSize\n\t}\n\n\tquery := url.Values{}\n\tparameter.AddToQuery(query, \"page\", req.Page)\n\tparameter.AddToQuery(query, \"page_size\", req.PageSize)\n\tparameter.AddToQuery(query, \"domain\", req.Domain)\n\tparameter.AddToQuery(query, \"organization_id\", req.OrganizationID)\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"GET\",\n\t\tPath: \"/domain/v2alpha2/tasks\",\n\t\tQuery: query,\n\t\tHeaders: http.Header{},\n\t}\n\n\tvar resp ListTasksResponse\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func (c *restClient) ListTasks(ctx context.Context, req *cloudtaskspb.ListTasksRequest, opts ...gax.CallOption) *TaskIterator {\n\tit := &TaskIterator{}\n\treq = proto.Clone(req).(*cloudtaskspb.ListTasksRequest)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tit.InternalFetch = func(pageSize int, pageToken string) ([]*cloudtaskspb.Task, string, error) {\n\t\tresp := &cloudtaskspb.ListTasksResponse{}\n\t\tif pageToken != \"\" {\n\t\t\treq.PageToken = pageToken\n\t\t}\n\t\tif pageSize > math.MaxInt32 {\n\t\t\treq.PageSize = math.MaxInt32\n\t\t} else if pageSize != 0 {\n\t\t\treq.PageSize = int32(pageSize)\n\t\t}\n\t\tbaseUrl, err := url.Parse(c.endpoint)\n\t\tif err != nil {\n\t\t\treturn nil, \"\", err\n\t\t}\n\t\tbaseUrl.Path += fmt.Sprintf(\"/v2beta3/%v/tasks\", req.GetParent())\n\n\t\tparams := url.Values{}\n\t\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\t\tif req.GetPageSize() != 0 {\n\t\t\tparams.Add(\"pageSize\", fmt.Sprintf(\"%v\", req.GetPageSize()))\n\t\t}\n\t\tif req.GetPageToken() != \"\" {\n\t\t\tparams.Add(\"pageToken\", fmt.Sprintf(\"%v\", req.GetPageToken()))\n\t\t}\n\t\tif req.GetResponseView() != 0 {\n\t\t\tparams.Add(\"responseView\", fmt.Sprintf(\"%v\", req.GetResponseView()))\n\t\t}\n\n\t\tbaseUrl.RawQuery = params.Encode()\n\n\t\t// Build HTTP headers from client and context metadata.\n\t\thds := append(c.xGoogHeaders, \"Content-Type\", \"application/json\")\n\t\theaders := gax.BuildHeaders(ctx, hds...)\n\t\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\t\tif settings.Path != \"\" {\n\t\t\t\tbaseUrl.Path = settings.Path\n\t\t\t}\n\t\t\thttpReq, err := http.NewRequest(\"GET\", baseUrl.String(), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\thttpReq.Header = headers\n\n\t\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tdefer httpRsp.Body.Close()\n\n\t\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn nil\n\t\t}, opts...)\n\t\tif e != nil {\n\t\t\treturn nil, \"\", e\n\t\t}\n\t\tit.Response = resp\n\t\treturn resp.GetTasks(), resp.GetNextPageToken(), nil\n\t}\n\n\tfetch := func(pageSize int, pageToken string) (string, error) {\n\t\titems, nextPageToken, err := it.InternalFetch(pageSize, pageToken)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tit.items = append(it.items, items...)\n\t\treturn nextPageToken, nil\n\t}\n\n\tit.pageInfo, it.nextFunc = iterator.NewPageInfo(fetch, it.bufLen, it.takeBuf)\n\tit.pageInfo.MaxSize = int(req.GetPageSize())\n\tit.pageInfo.Token = req.GetPageToken()\n\n\treturn it\n}", "func (taskService TaskService) getTaskSites(ctx context.Context, managedEndpoints []gocql.UUID, siteIDs map[string]struct{}, partnerID string) map[string]struct{} {\n\tif len(managedEndpoints) == 0 {\n\t\treturn siteIDs\n\t}\n\n\tvar (\n\t\twg sync.WaitGroup\n\t\tmu sync.Mutex\n\t)\n\twg.Add(len(managedEndpoints))\n\n\tfor _, endpointID := range managedEndpoints {\n\t\tgo func(ctx context.Context, endpointID gocql.UUID) {\n\t\t\tdefer wg.Done()\n\n\t\t\tsiteID, _, err := taskService.assetsService.GetSiteIDByEndpointID(ctx, partnerID, endpointID)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Log.WarnfCtx(ctx, \"cannot get siteID for ManagedEndpoint[%v]: %v\", endpointID, err)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tmu.Lock()\n\t\t\tsiteIDs[siteID] = struct{}{}\n\t\t\tmu.Unlock()\n\t\t}(ctx, endpointID)\n\t}\n\n\twg.Wait()\n\treturn siteIDs\n}", "func (m *ItemTodoListsItemTasksTodoTaskItemRequestBuilder) Get(ctx context.Context, requestConfiguration *ItemTodoListsItemTasksTodoTaskItemRequestBuilderGetRequestConfiguration)(iadcd81124412c61e647227ecfc4449d8bba17de0380ddda76f641a29edf2b242.TodoTaskable, error) {\n requestInfo, err := m.ToGetRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return nil, err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": ia572726a95efa92ddd544552cd950653dc691023836923576b2f4bf716cf204a.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": ia572726a95efa92ddd544552cd950653dc691023836923576b2f4bf716cf204a.CreateODataErrorFromDiscriminatorValue,\n }\n res, err := m.BaseRequestBuilder.RequestAdapter.Send(ctx, requestInfo, iadcd81124412c61e647227ecfc4449d8bba17de0380ddda76f641a29edf2b242.CreateTodoTaskFromDiscriminatorValue, errorMapping)\n if err != nil {\n return nil, err\n }\n if res == nil {\n return nil, nil\n }\n return res.(iadcd81124412c61e647227ecfc4449d8bba17de0380ddda76f641a29edf2b242.TodoTaskable), nil\n}", "func (a *Client) GetSystemTasks(params *GetSystemTasksParams) (*GetSystemTasksOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetSystemTasksParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"getSystemTasks\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/TaskService/Oem/Tasks/{identifier}\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &GetSystemTasksReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*GetSystemTasksOK), nil\n\n}", "func (s *Service) GetTasks(ctx context.Context, req *taskspb.GetTasksReq) (*taskspb.GetTasksRes, error) {\n\tres := taskspb.GetTasksRes{Tasks: []*taskspb.Task{}}\n\n\tfilter := bson.M{}\n\tcursor, err := s.db.tasks.Find(ctx, filter)\n\tif err != nil {\n\t\treturn nil, status.Errorf(\n\t\t\tcodes.Internal,\n\t\t\tfmt.Sprintf(\"unexpected error: %v\", err),\n\t\t)\n\t}\n\n\terr = cursor.All(ctx, &res.Tasks)\n\tif err != nil {\n\t\treturn nil, status.Errorf(\n\t\t\tcodes.Internal,\n\t\t\tfmt.Sprintf(\"unexpected error: %v\", err),\n\t\t)\n\t}\n\n\treturn &res, nil\n}", "func (c *Client) TasksByAppByQuery(guid string, query url.Values) ([]Task, error) {\n\turi := fmt.Sprintf(\"/v3/apps/%s/tasks\", guid)\n\treturn c.handleTasksApiCall(uri, query)\n}", "func (c *RestClient) GetTenantsQuery(f *specs.TenantsNfFilter) (*response.Tenants, error) {\n\n\tc.GetClient()\n\tresp, err := c.Client.R().SetBody(f).SetQueryString(apiTenantAction).Post(c.BaseURL + apiTenants)\n\tif err != nil {\n\t\tglog.Error(err)\n\t\treturn nil, err\n\t}\n\n\tif c.isTrace && resp != nil {\n\t\tfmt.Println(string(resp.Body()))\n\t}\n\n\tif !resp.IsSuccess() {\n\t\treturn nil, c.checkError(resp)\n\t}\n\n\tvar tenants response.Tenants\n\tif err := json.Unmarshal(resp.Body(), &tenants); err != nil {\n\t\tglog.Error(\"Failed parse server respond.\")\n\t\treturn nil, err\n\t}\n\n\treturn &tenants, nil\n}", "func (db *DynamoDB) GetTask(ctx context.Context, req *tes.GetTaskRequest) (*tes.Task, error) {\n\tvar task *tes.Task\n\tvar response *dynamodb.GetItemOutput\n\tvar err error\n\n\tswitch req.View {\n\tcase tes.TaskView_MINIMAL:\n\t\tresponse, err = db.getMinimalView(ctx, req.Id)\n\tcase tes.TaskView_BASIC:\n\t\tresponse, err = db.getBasicView(ctx, req.Id)\n\tcase tes.TaskView_FULL:\n\t\tresponse, err = db.getFullView(ctx, req.Id)\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif response.Item == nil {\n\t\treturn nil, grpc.Errorf(codes.NotFound, fmt.Sprintf(\"%v: taskID: %s\", errNotFound.Error(), req.Id))\n\t}\n\n\terr = dynamodbattribute.UnmarshalMap(response.Item, &task)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to DynamoDB unmarshal Task, %v\", err)\n\t}\n\n\treturn task, nil\n}", "func (r *TaskRepository) GetUserActiveTasks(db db.DB, userID int) (*[]entities.Task, error) {\n\tTasks := &[]entities.Task{}\n\terr := db.Select(Tasks,\n\t\t\"SELECT FROM tasks t WHERE t.id IN(select t.id FROM tasks tt INNER JOIN users u ON tt.user_id=u.id WHERE u.id=$1 AND tt.status=$2)\",\n\t\tuserID, constants.InProgress)\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error selecting user's tasks from db: %w\", err)\n\t}\n\treturn Tasks, nil\n}", "func (s *TestBase) GetTransferTasks(batchSize int, getAll bool) ([]*p.TransferTaskInfo, error) {\n\tresult := []*p.TransferTaskInfo{}\n\tvar token []byte\n\nLoop:\n\tfor {\n\t\tresponse, err := s.ExecutionManager.GetTransferTasks(&p.GetTransferTasksRequest{\n\t\t\tReadLevel: s.GetTransferReadLevel(),\n\t\t\tMaxReadLevel: int64(math.MaxInt64),\n\t\t\tBatchSize: batchSize,\n\t\t\tNextPageToken: token,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\ttoken = response.NextPageToken\n\t\tresult = append(result, response.Tasks...)\n\t\tif len(token) == 0 || !getAll {\n\t\t\tbreak Loop\n\t\t}\n\t}\n\n\tfor _, task := range result {\n\t\tatomic.StoreInt64(&s.ReadLevel, task.TaskID)\n\t}\n\n\treturn result, nil\n}", "func (s *TaskService) ListWithProject(projectID int64) ([]Task, *http.Response, error) {\n\tresObj := new(TaskResponse)\n\tresp, err := s.sling.New().\n\t\tQueryStruct(CreateFunctionParam(\"gettasks\")).\n\t\tQueryStruct(&GetTasksParam{ProjectID: &projectID}).\n\t\tReceiveSuccess(resObj)\n\tif resObj != nil && len(resObj.Results) > 0 {\n\t\tif resObj.Results[0].ErrorDesc != nil {\n\t\t\treturn nil, resp, Error{*resObj.Results[0].ErrorDesc}\n\t\t}\n\t\treturn *(&resObj.Results), resp, err\n\t}\n\treturn make([]Task, 0), resp, err\n}", "func (m *SharepointIds) GetTenantId()(*string) {\n val, err := m.GetBackingStore().Get(\"tenantId\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (s *Service) GetTask(taskKey string) (*Task, error) {\n\tfor _, task := range s.Tasks {\n\t\tif task.Key == taskKey {\n\t\t\ttask.serviceName = s.Name\n\t\t\treturn task, nil\n\t\t}\n\t}\n\treturn nil, &TaskNotFoundError{\n\t\tTaskKey: taskKey,\n\t\tServiceName: s.Name,\n\t}\n}", "func (r *RuntimeServer) GetTenantResources(context.Context, *pb.Empty) (*pb.TenantResourceList, error) {\n\tres := r.store.GetTenantResourceList()\n\tvar trs = make(map[string]*pb.TenantResource)\n\tfor _, re := range res {\n\t\tvar tr pb.TenantResource\n\t\trunningApps := r.store.GetTenantRunningApp(re.Namespace)\n\t\tfor _, app := range runningApps {\n\t\t\tif app.ServiceKind == model.ServiceKindThirdParty {\n\t\t\t\ttr.RunningAppThirdNum++\n\t\t\t} else if app.ServiceKind == model.ServiceKindInternal {\n\t\t\t\ttr.RunningAppInternalNum++\n\t\t\t}\n\t\t}\n\t\ttr.RunningAppNum = int64(len(runningApps))\n\t\ttr.CpuLimit = re.CPULimit\n\t\ttr.CpuRequest = re.CPURequest\n\t\ttr.MemoryLimit = re.MemoryLimit / 1024 / 1024\n\t\ttr.MemoryRequest = re.MemoryRequest / 1024 / 1024\n\t\ttrs[re.Namespace] = &tr\n\t}\n\treturn &pb.TenantResourceList{Resources: trs}, nil\n}", "func (api *API) QueryTasks(params parameters.QueryTasks) (*types.Tasks, error) {\n\tpath := \"/tasks\"\n\n\tqp, err := query.Values(params)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdata, err := api.get(path, &qp)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn types.NewTasksFromJSON(data)\n}", "func GetTasks() (tc TaskCollection) {\n\ttasks := []Task{}\n\tdb.Find(&tasks)\n\n\ttc = TaskCollection{\n\t\tTasks: tasks,\n\t}\n\n\treturn\n}", "func getTenant(c *cli.Context, w io.Writer) error {\n\terr := checkArgCount(c, 0)\n\tif err != nil {\n\t\treturn err\n\t}\n\tconfig, err := cf.LoadConfig()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttenant := config.Tenant\n\tif tenant == nil {\n\t\tfmt.Printf(\"No tenant selected\\n\")\n\t} else {\n\t\tif c.GlobalIsSet(\"non-interactive\") {\n\t\t\tfmt.Printf(\"%s\\t%s\\n\", tenant.ID, tenant.Name)\n\t\t} else if utils.NeedsFormatting(c) {\n\t\t\tutils.FormatObject(tenant, w, c)\n\t\t} else {\n\t\t\tfmt.Printf(\"Current tenant is '%s' with ID %s\\n\", tenant.Name, tenant.ID)\n\t\t}\n\t}\n\treturn nil\n}", "func (t *TaskChaincode) queryPayTxByTaskId(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n\t// 0\n\t// \"$taskId\"\n\tif len(args) != 1 {\n\t\treturn shim.Error(\"Incorrect number of arguments. Expecting 1\")\n\t}\n\n\n\tif len(args[0]) <= 0 {\n\t\treturn shim.Error(\"1st argument must be a non-empty string\")\n\t}\n\n\ttaskId := args[0]\n\n\tqueryString := fmt.Sprintf(\"{\\\"selector\\\":{\\\"objectType\\\":\\\"PayTX\\\",\\\"taskId\\\":\\\"%s\\\"}}\", taskId)\n\tqueryResults, err := getResultForQueryString(stub, queryString)\n\n\tif err != nil {\n\t\treturn shim.Error(err.Error())\n\t}\n\n\treturn shim.Success(queryResults)\n}", "func (s *TaskService) List() ([]Task, *http.Response, error) {\n\tresObj := new(TaskResponse)\n\tresp, err := s.sling.New().\n\t\tQueryStruct(CreateFunctionParam(\"gettasks\")).\n\t\tReceiveSuccess(resObj)\n\tif resObj != nil && len(resObj.Results) > 0 {\n\t\tif resObj.Results[0].ErrorDesc != nil {\n\t\t\treturn nil, resp, Error{*resObj.Results[0].ErrorDesc}\n\t\t}\n\t\treturn *(&resObj.Results), resp, err\n\t}\n\treturn make([]Task, 0), resp, err\n}", "func (db *DynamoDB) GetTask(ctx context.Context, req *tes.GetTaskRequest) (*tes.Task, error) {\n\tvar task *tes.Task\n\tvar response *dynamodb.GetItemOutput\n\tvar err error\n\n\tswitch req.View {\n\tcase tes.TaskView_MINIMAL:\n\t\tresponse, err = db.getMinimalView(ctx, req.Id)\n\tcase tes.TaskView_BASIC:\n\t\tresponse, err = db.getBasicView(ctx, req.Id)\n\tcase tes.TaskView_FULL:\n\t\tresponse, err = db.getFullView(ctx, req.Id)\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif response.Item == nil {\n\t\treturn nil, tes.ErrNotFound\n\t}\n\n\terr = dynamodbattribute.UnmarshalMap(response.Item, &task)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to DynamoDB unmarshal Task, %v\", err)\n\t}\n\n\treturn task, nil\n}", "func (s *RefreshImpl) Task(ctx context.Refresh) *taskDTO.PostTaskRequest {\n\tdto := taskDTO.PostTaskRequest{}\n\tdto.MessageID = constvalue.RefreshTaskID\n\tdto.Name = \"Refresh Enclosure\"\n\tdto.Description = \"Refresh enclosure's settings and component.\"\n\tdto.CreatedByName = \"Enclosure Service\"\n\tdto.CreatedByURI = \"/promise/v1/enclosure\"\n\tdto.TargetName = ctx.GetEnclosure().Name\n\tdto.TargetURI = base.ToEnclosureURI(ctx.GetEnclosure().ID)\n\tfor _, v := range s.sub {\n\t\tstep := taskDTO.PostTaskStepRequest{}\n\t\tstep.MessageID = v.MessageID()\n\t\tstep.Name = v.Name()\n\t\tstep.Description = v.Description()\n\t\tstep.ExpectedExecutionMs = v.ExpectedExecutionMs()\n\t\tdto.TaskSteps = append(dto.TaskSteps, step)\n\t}\n\treturn &dto\n}", "func (db *MySQLDB) ListTenant(ctx context.Context, request *helper.PageRequest) ([]*Tenant, *helper.Page, error) {\n\tfLog := mysqlLog.WithField(\"func\", \"GetUserByRecID\").WithField(\"RequestID\", ctx.Value(constants.RequestID))\n\tq := \"SELECT COUNT(*) AS CNT FROM HANSIP_TENANT\"\n\tret := make([]*Tenant, 0)\n\trow := db.instance.QueryRowContext(ctx, q)\n\tcount := 0\n\terr := row.Scan(&count)\n\tif err != nil {\n\t\tif err == sql.ErrNoRows {\n\t\t\treturn ret, helper.NewPage(request, uint(count)), nil\n\t\t}\n\t\tfLog.Errorf(\"db.instance.QueryRowContext got %s\", err.Error())\n\t\treturn nil, nil, &ErrDBQueryError{\n\t\t\tWrapped: err,\n\t\t\tMessage: \"Error ListTenant\",\n\t\t\tSQL: q,\n\t\t}\n\t}\n\n\tvar OrderBy string\n\tswitch strings.ToUpper(request.OrderBy) {\n\tcase \"TENANT_NAME\":\n\t\tOrderBy = \"TENANT_NAME\"\n\tcase \"TENANT_DOMAIN\":\n\t\tOrderBy = \"TENANT_DOMAIN\"\n\tdefault:\n\t\tOrderBy = \"TENANT_NAME\"\n\t}\n\n\tpage := helper.NewPage(request, uint(count))\n\tq = fmt.Sprintf(\"SELECT REC_ID, TENANT_NAME, TENANT_DOMAIN, DESCRIPTION FROM HANSIP_TENANT ORDER BY %s %s LIMIT %d, %d\", OrderBy, request.Sort, page.OffsetStart, page.OffsetEnd-page.OffsetStart)\n\trows, err := db.instance.QueryContext(ctx, q)\n\tif err != nil {\n\t\tfLog.Errorf(\"db.instance.QueryContext got %s. SQL = %s\", err.Error(), q)\n\t\treturn nil, nil, &ErrDBQueryError{\n\t\t\tWrapped: err,\n\t\t\tMessage: \"Error ListTenant\",\n\t\t\tSQL: q,\n\t\t}\n\t}\n\tdefer rows.Close()\n\tfor rows.Next() {\n\t\tt := &Tenant{}\n\t\terr := rows.Scan(&t.RecID, &t.Name, &t.Domain, &t.Description)\n\t\tif err != nil {\n\t\t\tfLog.Warnf(\"rows.Scan got %s\", err.Error())\n\t\t\treturn nil, nil, &ErrDBScanError{\n\t\t\t\tWrapped: err,\n\t\t\t\tMessage: \"Error ListTenant\",\n\t\t\t\tSQL: q,\n\t\t\t}\n\t\t} else {\n\t\t\tret = append(ret, t)\n\t\t}\n\t}\n\treturn ret, page, nil\n}" ]
[ "0.6200286", "0.60298127", "0.59042066", "0.5897367", "0.58799726", "0.5823304", "0.5800558", "0.5791268", "0.57738566", "0.5756176", "0.5751128", "0.57498974", "0.5747922", "0.5718832", "0.5716588", "0.5715596", "0.57092756", "0.56690633", "0.56687534", "0.5625812", "0.5622785", "0.56107664", "0.55984205", "0.55783486", "0.55771416", "0.5566234", "0.5562784", "0.556222", "0.55546683", "0.55338705", "0.55217874", "0.5515816", "0.54989326", "0.5497448", "0.54906505", "0.54619396", "0.54611015", "0.54574305", "0.5444862", "0.5429423", "0.5426616", "0.5419335", "0.54157424", "0.54157424", "0.5409542", "0.54069513", "0.54057544", "0.53964454", "0.53819555", "0.5340968", "0.53306437", "0.53263885", "0.53186065", "0.53141886", "0.5308764", "0.5306535", "0.53018475", "0.5281697", "0.52801675", "0.5276284", "0.52754194", "0.527166", "0.52662545", "0.5240069", "0.5239809", "0.5239003", "0.5223665", "0.5222514", "0.5214828", "0.52099174", "0.5209793", "0.5209597", "0.51942134", "0.51933205", "0.5180673", "0.51714176", "0.51642936", "0.51566255", "0.51456827", "0.514524", "0.5134488", "0.51308846", "0.51303774", "0.5126738", "0.5122523", "0.51136315", "0.5108602", "0.51074904", "0.50990564", "0.50959057", "0.50956076", "0.50942355", "0.5093363", "0.5093043", "0.5088412", "0.50881594", "0.50857556", "0.5078836", "0.50744957", "0.5070835" ]
0.72561204
0
Set security groups for a tenant
func setSecurityGroups(c *cli.Context) error { err := checkArgCount(c, 2) if err != nil { return err } id := c.Args().First() items := []string{} if c.Args()[1] != "" { items = regexp.MustCompile(`\s*,\s*`).Split(c.Args()[1], -1) } securityGroups := &photon.SecurityGroupsSpec{ Items: items, } client.Photonclient, err = client.GetClient(c) if err != nil { return err } task, err := client.Photonclient.Tenants.SetSecurityGroups(id, securityGroups) if err != nil { return err } _, err = waitOnTaskOperation(task.ID, c) if err != nil { return err } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func setDeploymentSecurityGroups(c *cli.Context) error {\n\tvar err error\n\tvar deploymentId string\n\tvar groups string\n\n\t// We have two cases:\n\t// Case 1: arguments are: id groups\n\t// Case 2: arguments are: groups\n\tif len(c.Args()) == 2 {\n\t\tdeploymentId = c.Args()[0]\n\t\tgroups = c.Args()[1]\n\t} else if len(c.Args()) == 1 {\n\t\tdeploymentId, err = getDefaultDeploymentId(c)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tgroups = c.Args()[0]\n\t} else {\n\t\treturn fmt.Errorf(\"Usage: deployments set-security-groups [id] groups\")\n\t}\n\n\titems := regexp.MustCompile(`\\s*,\\s*`).Split(groups, -1)\n\tsecurityGroups := &photon.SecurityGroupsSpec{\n\t\tItems: items,\n\t}\n\n\tclient.Photonclient, err = client.GetClient(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttask, err := client.Photonclient.Deployments.SetSecurityGroups(deploymentId, securityGroups)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = waitOnTaskOperation(task.ID, c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = deploymentJsonHelper(c, deploymentId, client.Photonclient)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (instance *Host) setSecurityGroups(ctx context.Context, req abstract.HostRequest, defaultSubnet resources.Subnet) fail.Error {\n\tsvc := instance.Service()\n\tif req.Single {\n\t\thostID, err := instance.GetID()\n\t\tif err != nil {\n\t\t\treturn fail.ConvertError(err)\n\t\t}\n\t\tfor k := range req.SecurityGroupIDs {\n\t\t\tif k != \"\" {\n\t\t\t\tlogrus.WithContext(ctx).Infof(\"Binding security group with id %s to host %s\", k, hostID)\n\t\t\t\txerr := svc.BindSecurityGroupToHost(ctx, k, hostID)\n\t\t\t\txerr = debug.InjectPlannedFail(xerr)\n\t\t\t\tif xerr != nil {\n\t\t\t\t\treturn xerr\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\treturn nil\n\t}\n\n\txerr := instance.Alter(ctx, func(clonable data.Clonable, props *serialize.JSONProperties) fail.Error {\n\t\treturn props.Alter(hostproperty.SecurityGroupsV1, func(clonable data.Clonable) (finnerXErr fail.Error) {\n\t\t\thsgV1, ok := clonable.(*propertiesv1.HostSecurityGroups)\n\t\t\tif !ok {\n\t\t\t\treturn fail.InconsistentError(\"'*propertiesv1.HostSecurityGroups' expected, '%s' provided\", reflect.TypeOf(clonable).String())\n\t\t\t}\n\n\t\t\t// get default Subnet core data\n\t\t\tvar (\n\t\t\t\tdefaultAbstractSubnet *abstract.Subnet\n\t\t\t\tdefaultSubnetID string\n\t\t\t)\n\t\t\tinnerXErr := defaultSubnet.Inspect(ctx,\n\t\t\t\tfunc(clonable data.Clonable, _ *serialize.JSONProperties) fail.Error {\n\t\t\t\t\tvar ok bool\n\t\t\t\t\tdefaultAbstractSubnet, ok = clonable.(*abstract.Subnet)\n\t\t\t\t\tif !ok {\n\t\t\t\t\t\treturn fail.InconsistentError(\"'*abstract.Subnet' expected, '%s' provided\", reflect.TypeOf(clonable).String())\n\t\t\t\t\t}\n\n\t\t\t\t\tdefaultSubnetID = defaultAbstractSubnet.ID\n\t\t\t\t\treturn nil\n\t\t\t\t},\n\t\t\t)\n\t\t\tif innerXErr != nil {\n\t\t\t\treturn innerXErr\n\t\t\t}\n\n\t\t\tvar gwsg, pubipsg, lansg resources.SecurityGroup\n\n\t\t\t// Apply Security Group for gateways in default Subnet\n\t\t\tif req.IsGateway && defaultAbstractSubnet.GWSecurityGroupID != \"\" {\n\t\t\t\tgwsg, innerXErr = LoadSecurityGroup(ctx, svc, defaultAbstractSubnet.GWSecurityGroupID)\n\t\t\t\tif innerXErr != nil {\n\t\t\t\t\treturn fail.Wrap(innerXErr, \"failed to query Subnet '%s' Security Group '%s'\", defaultSubnet.GetName(), defaultAbstractSubnet.GWSecurityGroupID)\n\t\t\t\t}\n\n\t\t\t\tinnerXErr = gwsg.BindToHost(ctx, instance, resources.SecurityGroupEnable, resources.MarkSecurityGroupAsSupplemental)\n\t\t\t\tif innerXErr != nil {\n\t\t\t\t\treturn fail.Wrap(innerXErr, \"failed to apply Subnet's GW Security Group for gateway '%s' on Host '%s'\", gwsg.GetName(), req.ResourceName)\n\t\t\t\t}\n\n\t\t\t\tdefer func() {\n\t\t\t\t\tif finnerXErr != nil && !req.KeepOnFailure {\n\t\t\t\t\t\tderr := gwsg.UnbindFromHost(cleanupContextFrom(ctx), instance)\n\t\t\t\t\t\tif derr != nil {\n\t\t\t\t\t\t\t_ = finnerXErr.AddConsequence(fail.Wrap(derr, \"cleaning up on %s, failed to unbind Security Group '%s' from Host '%s'\", ActionFromError(finnerXErr), gwsg.GetName(), req.ResourceName))\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}()\n\n\t\t\t\tgwid, err := gwsg.GetID()\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn fail.ConvertError(err)\n\t\t\t\t}\n\n\t\t\t\titem := &propertiesv1.SecurityGroupBond{\n\t\t\t\t\tID: gwid,\n\t\t\t\t\tName: gwsg.GetName(),\n\t\t\t\t\tDisabled: false,\n\t\t\t\t\tFromSubnet: true,\n\t\t\t\t}\n\t\t\t\thsgV1.ByID[item.ID] = item\n\t\t\t\thsgV1.ByName[item.Name] = item.ID\n\t\t\t}\n\n\t\t\t// Apply Security Group for hosts with public IP in default Subnet\n\t\t\tif (req.IsGateway || req.PublicIP) && defaultAbstractSubnet.PublicIPSecurityGroupID != \"\" {\n\t\t\t\tpubipsg, innerXErr = LoadSecurityGroup(ctx, svc, defaultAbstractSubnet.PublicIPSecurityGroupID)\n\t\t\t\tif innerXErr != nil {\n\t\t\t\t\treturn fail.Wrap(innerXErr, \"failed to query Subnet '%s' Security Group with ID %s\", defaultSubnet.GetName(), defaultAbstractSubnet.PublicIPSecurityGroupID)\n\t\t\t\t}\n\n\t\t\t\tinnerXErr = pubipsg.BindToHost(ctx, instance, resources.SecurityGroupEnable, resources.MarkSecurityGroupAsSupplemental)\n\t\t\t\tif innerXErr != nil {\n\t\t\t\t\treturn fail.Wrap(innerXErr, \"failed to apply Subnet's Public Security Group for gateway '%s' on Host '%s'\", pubipsg.GetName(), req.ResourceName)\n\t\t\t\t}\n\n\t\t\t\tdefer func() {\n\t\t\t\t\tif finnerXErr != nil && !req.KeepOnFailure {\n\t\t\t\t\t\tderr := pubipsg.UnbindFromHost(cleanupContextFrom(ctx), instance)\n\t\t\t\t\t\tif derr != nil {\n\t\t\t\t\t\t\t_ = finnerXErr.AddConsequence(fail.Wrap(derr, \"cleaning up on %s, failed to unbind Security Group '%s' from Host '%s'\", ActionFromError(finnerXErr), pubipsg.GetName(), req.ResourceName))\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}()\n\n\t\t\t\tpubID, err := pubipsg.GetID()\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn fail.ConvertError(err)\n\t\t\t\t}\n\n\t\t\t\titem := &propertiesv1.SecurityGroupBond{\n\t\t\t\t\tID: pubID,\n\t\t\t\t\tName: pubipsg.GetName(),\n\t\t\t\t\tDisabled: false,\n\t\t\t\t\tFromSubnet: true,\n\t\t\t\t}\n\t\t\t\thsgV1.ByID[item.ID] = item\n\t\t\t\thsgV1.ByName[item.Name] = item.ID\n\t\t\t}\n\n\t\t\t// Apply internal Security Group of each other subnets\n\t\t\tdefer func() {\n\t\t\t\tif finnerXErr != nil && !req.KeepOnFailure {\n\t\t\t\t\tvar (\n\t\t\t\t\t\tsg resources.SecurityGroup\n\t\t\t\t\t\tderr error\n\t\t\t\t\t\terrs []error\n\t\t\t\t\t)\n\t\t\t\t\tfor _, v := range req.Subnets {\n\t\t\t\t\t\tif v.ID == defaultSubnetID {\n\t\t\t\t\t\t\tcontinue\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\tsubnetInstance, deeperXErr := LoadSubnet(cleanupContextFrom(ctx), svc, \"\", v.ID)\n\t\t\t\t\t\tif deeperXErr != nil {\n\t\t\t\t\t\t\t_ = innerXErr.AddConsequence(deeperXErr)\n\t\t\t\t\t\t\tcontinue\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\tsgName := sg.GetName()\n\t\t\t\t\t\tdeeperXErr = subnetInstance.Inspect(cleanupContextFrom(ctx), func(\n\t\t\t\t\t\t\tclonable data.Clonable, _ *serialize.JSONProperties,\n\t\t\t\t\t\t) fail.Error {\n\t\t\t\t\t\t\tabstractSubnet, ok := clonable.(*abstract.Subnet)\n\t\t\t\t\t\t\tif !ok {\n\t\t\t\t\t\t\t\treturn fail.InconsistentError(\"'*abstract.Subnet' expected, '%s' provided\", reflect.TypeOf(clonable).String())\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tif abstractSubnet.InternalSecurityGroupID != \"\" {\n\t\t\t\t\t\t\t\tsg, derr = LoadSecurityGroup(cleanupContextFrom(ctx), svc, abstractSubnet.InternalSecurityGroupID)\n\t\t\t\t\t\t\t\tif derr != nil {\n\t\t\t\t\t\t\t\t\terrs = append(errs, derr)\n\t\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\t\tderr = sg.UnbindFromHost(cleanupContextFrom(ctx), instance)\n\t\t\t\t\t\t\t\t\tif derr != nil {\n\t\t\t\t\t\t\t\t\t\terrs = append(errs, derr)\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\treturn nil\n\t\t\t\t\t\t})\n\t\t\t\t\t\tif deeperXErr != nil {\n\t\t\t\t\t\t\t_ = finnerXErr.AddConsequence(fail.Wrap(deeperXErr, \"cleaning up on failure, failed to unbind Security Group '%s' from Host\", sgName))\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t\tif len(errs) > 0 {\n\t\t\t\t\t\t_ = finnerXErr.AddConsequence(fail.Wrap(fail.NewErrorList(errs), \"failed to unbind Subnets Security Group from Host '%s'\", sg.GetName(), req.ResourceName))\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}()\n\n\t\t\tfor _, v := range req.Subnets {\n\t\t\t\t// Do not try to bind defaultSubnet on gateway, because this code is running under a lock on defaultSubnet in this case, and this will lead to deadlock\n\t\t\t\t// (binding of gateway on defaultSubnet is done inside Subnet.Create() call)\n\t\t\t\tif req.IsGateway && v.ID == defaultSubnetID {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\totherSubnetInstance, innerXErr := LoadSubnet(ctx, svc, \"\", v.ID)\n\t\t\t\tinnerXErr = debug.InjectPlannedFail(innerXErr)\n\t\t\t\tif innerXErr != nil {\n\t\t\t\t\treturn innerXErr\n\t\t\t\t}\n\n\t\t\t\tvar otherAbstractSubnet *abstract.Subnet\n\t\t\t\tinnerXErr = otherSubnetInstance.Inspect(ctx, func(\n\t\t\t\t\tclonable data.Clonable, _ *serialize.JSONProperties,\n\t\t\t\t) fail.Error {\n\t\t\t\t\tvar ok bool\n\t\t\t\t\totherAbstractSubnet, ok = clonable.(*abstract.Subnet)\n\t\t\t\t\tif !ok {\n\t\t\t\t\t\treturn fail.InconsistentError(\"'*abstract.Subnet' expected, '%s' provided\", reflect.TypeOf(clonable).String())\n\t\t\t\t\t}\n\t\t\t\t\t_ = otherAbstractSubnet\n\n\t\t\t\t\treturn nil\n\t\t\t\t})\n\t\t\t\tif innerXErr != nil {\n\t\t\t\t\treturn innerXErr\n\t\t\t\t}\n\n\t\t\t\tsafe := false\n\n\t\t\t\t// Fix for Stein\n\t\t\t\t{\n\t\t\t\t\tst, xerr := svc.GetProviderName()\n\t\t\t\t\tif xerr != nil {\n\t\t\t\t\t\treturn xerr\n\t\t\t\t\t}\n\t\t\t\t\tif st != \"ovh\" {\n\t\t\t\t\t\tsafe = true\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\tif cfg, xerr := svc.GetConfigurationOptions(ctx); xerr == nil {\n\t\t\t\t\tif aval, ok := cfg.Get(\"Safe\"); ok {\n\t\t\t\t\t\tif val, ok := aval.(bool); ok {\n\t\t\t\t\t\t\tsafe = val\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\tif otherAbstractSubnet.InternalSecurityGroupID != \"\" {\n\t\t\t\t\tlansg, innerXErr = LoadSecurityGroup(ctx, svc, otherAbstractSubnet.InternalSecurityGroupID)\n\t\t\t\t\tif innerXErr != nil {\n\t\t\t\t\t\treturn fail.Wrap(innerXErr, \"failed to load Subnet '%s' internal Security Group %s\", otherAbstractSubnet.Name, otherAbstractSubnet.InternalSecurityGroupID)\n\t\t\t\t\t}\n\n\t\t\t\t\tif !safe {\n\t\t\t\t\t\tinnerXErr = svc.ChangeSecurityGroupSecurity(ctx, false, true, otherAbstractSubnet.Network, \"\")\n\t\t\t\t\t\tif innerXErr != nil {\n\t\t\t\t\t\t\treturn fail.Wrap(innerXErr, \"failed to change security group\")\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\n\t\t\t\t\tinnerXErr = lansg.BindToHost(ctx, instance, resources.SecurityGroupEnable, resources.MarkSecurityGroupAsSupplemental)\n\t\t\t\t\tif innerXErr != nil {\n\t\t\t\t\t\treturn fail.Wrap(innerXErr, \"failed to apply Subnet '%s' internal Security Group '%s' to Host '%s'\", otherAbstractSubnet.Name, lansg.GetName(), req.ResourceName)\n\t\t\t\t\t}\n\n\t\t\t\t\tif !safe {\n\t\t\t\t\t\tinnerXErr = svc.ChangeSecurityGroupSecurity(ctx, true, false, otherAbstractSubnet.Network, \"\")\n\t\t\t\t\t\tif innerXErr != nil {\n\t\t\t\t\t\t\treturn fail.Wrap(innerXErr, \"failed to change security group\")\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\n\t\t\t\t\tlangID, err := lansg.GetID()\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn fail.ConvertError(err)\n\t\t\t\t\t}\n\n\t\t\t\t\t// register security group in properties\n\t\t\t\t\titem := &propertiesv1.SecurityGroupBond{\n\t\t\t\t\t\tID: langID,\n\t\t\t\t\t\tName: lansg.GetName(),\n\t\t\t\t\t\tDisabled: false,\n\t\t\t\t\t\tFromSubnet: true,\n\t\t\t\t\t}\n\t\t\t\t\thsgV1.ByID[item.ID] = item\n\t\t\t\t\thsgV1.ByName[item.Name] = item.ID\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\treturn nil\n\t\t})\n\t})\n\tif xerr != nil {\n\t\treturn xerr\n\t}\n\n\treturn nil\n}", "func SecurityGroups(groups ...string) RequestOptionFunc {\n\treturn func(body *RequestBody) error {\n\t\tbody.SecurityGroup = groups\n\t\treturn nil\n\t}\n}", "func (m *Group) SetPermissionGrants(value []ResourceSpecificPermissionGrantable)() {\n m.permissionGrants = value\n}", "func (m *Group) SetSecurityEnabled(value *bool)() {\n m.securityEnabled = value\n}", "func (m *Store) SetGroups(value []Groupable)() {\n m.groups = value\n}", "func (m *ExternalConnection) SetGroups(value []ExternalGroupable)() {\n m.groups = value\n}", "func (t *TokenGroupsT) SetGroupsSlice(s []SidAndAttributes) {\n\ts1 := []SidAndAttributes{}\n\thdr := (*reflect.SliceHeader)(unsafe.Pointer(&s1))\n\thdr.Data = uintptr(unsafe.Pointer(&t.Groups[0]))\n\thdr.Len = len(s)\n\thdr.Cap = len(s)\n\tcopy(s1, s)\n}", "func (m *PrivilegedAccessGroupEligibilitySchedule) SetGroupId(value *string)() {\n err := m.GetBackingStore().Set(\"groupId\", value)\n if err != nil {\n panic(err)\n }\n}", "func dataSourceSecurityGroupAttributes(d *schema.ResourceData, data map[string]interface{}) error {\n security_group_id := data[\"id\"].(string)\n log.Printf(\"[DEBUG] Retrieved twcc_security_group: %s\", security_group_id)\n\n d.SetId(security_group_id)\n d.Set(\"name\", data[\"name\"])\n security_group_rules := flattenSecurityGroupRulesInfo(data[\"security_group_rules\"].([]interface{}))\n d.Set(\"security_group_rules\", security_group_rules)\n\n return nil\n}", "func setInstanceGroupsAnnotation(existing map[string]string, igs []*compute.InstanceGroup) error {\n\ttype Value struct {\n\t\tName string\n\t\tZone string\n\t}\n\tvar instanceGroups []Value\n\tfor _, ig := range igs {\n\t\tinstanceGroups = append(instanceGroups, Value{Name: ig.Name, Zone: ig.Zone})\n\t}\n\tjsonValue, err := json.Marshal(instanceGroups)\n\tif err != nil {\n\t\treturn err\n\t}\n\texisting[annotations.InstanceGroupsAnnotationKey] = string(jsonValue)\n\treturn nil\n}", "func SetStorageGroups(tgs ...*TableGroup) StorageOption {\n\treturn func(s *Storage) { s.groups = TableGroupSlice(tgs) }\n}", "func (m *Group) SetSecurityIdentifier(value *string)() {\n m.securityIdentifier = value\n}", "func (r resource) Set(grp *echo.Group) {\n\treflect.\n\t\tValueOf(grp).\n\t\tMethodByName(r.callName()).\n\t\tCall(r.callArgs())\n}", "func (m *SharepointIds) SetTenantId(value *string)() {\n err := m.GetBackingStore().Set(\"tenantId\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *PrivilegedAccessGroupEligibilitySchedule) SetGroup(value Groupable)() {\n err := m.GetBackingStore().Set(\"group\", value)\n if err != nil {\n panic(err)\n }\n}", "func (db *DB) SetSecurity(ctx context.Context, security *Security) error {\n\tsec := &driver.Security{\n\t\tAdmins: driver.Members(security.Admins),\n\t\tMembers: driver.Members(security.Members),\n\t}\n\treturn db.driverDB.SetSecurity(ctx, sec)\n}", "func SetUserGroups(w http.ResponseWriter, r *http.Request) {\n\tfLog := userMgmtLogger.WithField(\"func\", \"SetUserGroups\").WithField(\"RequestID\", r.Context().Value(constants.RequestID)).WithField(\"path\", r.URL.Path).WithField(\"method\", r.Method)\n\tparams, err := helper.ParsePathParams(fmt.Sprintf(\"%s/management/user/{userRecId}/groups\", apiPrefix), r.URL.Path)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tuser, err := UserRepo.GetUserByRecID(r.Context(), params[\"userRecId\"])\n\tif err != nil {\n\t\tfLog.Error(err.Error())\n\t\thelper.WriteHTTPResponse(r.Context(), w, http.StatusInternalServerError, err.Error(), nil, nil)\n\t\treturn\n\t}\n\tif user == nil {\n\t\thelper.WriteHTTPResponse(r.Context(), w, http.StatusNotFound, fmt.Sprintf(\"User recID %s not found\", params[\"userRecId\"]), nil, nil)\n\t\treturn\n\t}\n\n\tbody, err := ioutil.ReadAll(r.Body)\n\tif err != nil {\n\t\tfLog.Errorf(\"ioutil.ReadAll got %s\", err.Error())\n\t\thelper.WriteHTTPResponse(r.Context(), w, http.StatusInternalServerError, err.Error(), nil, nil)\n\t\treturn\n\t}\n\tgroupIds := make([]string, 0)\n\terr = json.Unmarshal(body, &groupIds)\n\tif err != nil {\n\t\tfLog.Errorf(\"json.Unmarshal got %s\", err.Error())\n\t\thelper.WriteHTTPResponse(r.Context(), w, http.StatusBadRequest, err.Error(), nil, nil)\n\t\treturn\n\t}\n\n\terr = UserGroupRepo.DeleteUserGroupByUser(r.Context(), user)\n\tif err != nil {\n\t\tfLog.Errorf(\"UserGroupRepo.DeleteUserGroupByUser got %s\", err.Error())\n\t\thelper.WriteHTTPResponse(r.Context(), w, http.StatusInternalServerError, err.Error(), nil, nil)\n\t\treturn\n\t}\n\n\tcounter := 0\n\tfor _, groupID := range groupIds {\n\t\tgroup, err := GroupRepo.GetGroupByRecID(r.Context(), groupID)\n\t\tif err != nil {\n\t\t\tfLog.Warnf(\"GroupRepo.GetGroupByRecID got %s, this group %s will not be joined by user %s\", err.Error(), groupID, user.RecID)\n\t\t} else {\n\t\t\t_, err := UserGroupRepo.CreateUserGroup(r.Context(), user, group)\n\t\t\tif err != nil {\n\t\t\t\tfLog.Warnf(\"UserGroupRepo.CreateUserGroup got %s, this group %s will not be joined by user %s\", err.Error(), groupID, user.RecID)\n\t\t\t} else {\n\t\t\t\tcounter++\n\t\t\t}\n\t\t}\n\t}\n\n\tRevocationRepo.Revoke(r.Context(), user.Email)\n\n\thelper.WriteHTTPResponse(r.Context(), w, http.StatusOK, fmt.Sprintf(\"%d groups joined by user\", counter), nil, nil)\n}", "func ExampleDeviceSecurityGroupsClient_CreateOrUpdate() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclientFactory, err := armsecurity.NewClientFactory(\"<subscription-id>\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tres, err := clientFactory.NewDeviceSecurityGroupsClient().CreateOrUpdate(ctx, \"subscriptions/20ff7fc3-e762-44dd-bd96-b71116dcdc23/resourceGroups/SampleRG/providers/Microsoft.Devices/iotHubs/sampleiothub\", \"samplesecuritygroup\", armsecurity.DeviceSecurityGroup{\n\t\tProperties: &armsecurity.DeviceSecurityGroupProperties{\n\t\t\tTimeWindowRules: []armsecurity.TimeWindowCustomAlertRuleClassification{\n\t\t\t\t&armsecurity.ActiveConnectionsNotInAllowedRange{\n\t\t\t\t\tIsEnabled: to.Ptr(true),\n\t\t\t\t\tRuleType: to.Ptr(\"ActiveConnectionsNotInAllowedRange\"),\n\t\t\t\t\tMaxThreshold: to.Ptr[int32](30),\n\t\t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t\t\t\t\tTimeWindowSize: to.Ptr(\"PT05M\"),\n\t\t\t\t}},\n\t\t},\n\t}, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\t// You could use response here. We use blank identifier for just demo purposes.\n\t_ = res\n\t// If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes.\n\t// res.DeviceSecurityGroup = armsecurity.DeviceSecurityGroup{\n\t// \tName: to.Ptr(\"samplesecuritygroup\"),\n\t// \tType: to.Ptr(\"Microsoft.Security/deviceSecurityGroups\"),\n\t// \tID: to.Ptr(\"/subscriptions/20ff7fc3-e762-44dd-bd96-b71116dcdc23/resourceGroups/SampleRG/providers/Microsoft.Devices/iotHubs/sampleiothub/providers/Microsoft.Security/deviceSecurityGroups/samplesecuritygroup\"),\n\t// \tProperties: &armsecurity.DeviceSecurityGroupProperties{\n\t// \t\tAllowlistRules: []armsecurity.AllowlistCustomAlertRuleClassification{\n\t// \t\t\t&armsecurity.ConnectionToIPNotAllowed{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when an outbound connection is created between your device and an ip that isn't allowed\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Outbound connection to an ip that isn't allowed\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"ConnectionToIpNotAllowed\"),\n\t// \t\t\t\tAllowlistValues: []*string{\n\t// \t\t\t\t},\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.LocalUserNotAllowed{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when a local user that isn't allowed logins to the device\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Login by a local user that isn't allowed\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"LocalUserNotAllowed\"),\n\t// \t\t\t\tAllowlistValues: []*string{\n\t// \t\t\t\t},\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.ProcessNotAllowed{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when a process that isn't allowed is executed\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Execution of a process that isn't allowed\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"ProcessNotAllowed\"),\n\t// \t\t\t\tAllowlistValues: []*string{\n\t// \t\t\t\t},\n\t// \t\t}},\n\t// \t\tDenylistRules: []*armsecurity.DenylistCustomAlertRule{\n\t// \t\t},\n\t// \t\tThresholdRules: []armsecurity.ThresholdCustomAlertRuleClassification{\n\t// \t\t},\n\t// \t\tTimeWindowRules: []armsecurity.TimeWindowCustomAlertRuleClassification{\n\t// \t\t\t&armsecurity.ActiveConnectionsNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of active connections of a device in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of active connections is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(true),\n\t// \t\t\t\tRuleType: to.Ptr(\"ActiveConnectionsNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](30),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT05M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.AmqpC2DMessagesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of cloud to device messages (AMQP protocol) in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of cloud to device messages (AMQP protocol) is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"AmqpC2DMessagesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.MqttC2DMessagesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of cloud to device messages (MQTT protocol) in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of cloud to device messages (MQTT protocol) is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"MqttC2DMessagesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.HTTPC2DMessagesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of cloud to device messages (HTTP protocol) in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of cloud to device messages (HTTP protocol) is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"HttpC2DMessagesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.AmqpC2DRejectedMessagesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of cloud to device messages (AMQP protocol) that were rejected by the device in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of rejected cloud to device messages (AMQP protocol) is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"AmqpC2DRejectedMessagesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.MqttC2DRejectedMessagesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of cloud to device messages (MQTT protocol) that were rejected by the device in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of rejected cloud to device messages (MQTT protocol) is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"MqttC2DRejectedMessagesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.HTTPC2DRejectedMessagesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of cloud to device messages (HTTP protocol) that were rejected by the device in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of rejected cloud to device messages (HTTP protocol) is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"HttpC2DRejectedMessagesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.AmqpD2CMessagesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of device to cloud messages (AMQP protocol) in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of device to cloud messages (AMQP protocol) is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"AmqpD2CMessagesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.MqttD2CMessagesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of device to cloud messages (MQTT protocol) in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of device to cloud messages (MQTT protocol) is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"MqttD2CMessagesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.HTTPD2CMessagesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of device to cloud messages (HTTP protocol) in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of device to cloud messages (HTTP protocol) is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"HttpD2CMessagesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.DirectMethodInvokesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of direct method invokes in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of direct method invokes is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"DirectMethodInvokesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.FailedLocalLoginsNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of failed local logins on the device in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of failed local logins is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"FailedLocalLoginsNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.FileUploadsNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of file uploads from the device to the cloud in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of file uploads is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"FileUploadsNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.QueuePurgesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of device queue purges in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of device queue purges is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"QueuePurgesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.TwinUpdatesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of twin updates (by the device or the service) in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of twin updates is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"TwinUpdatesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.UnauthorizedOperationsNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number unauthorized operations in the time window is not in the allowed range. Unauthorized operations are operations that affect the device (or done by it) that fail because of an unauthorized error\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of unauthorized operations is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"UnauthorizedOperationsNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t}},\n\t// \t},\n\t// }\n}", "func (m *CloudPcConnection) SetTenantId(value *string)() {\n err := m.GetBackingStore().Set(\"tenantId\", value)\n if err != nil {\n panic(err)\n }\n}", "func setPortGroup(ovnNBClient goovn.Client, portGroupName string, ports ...*lpInfo) error {\n\n\tuuids := make([]string, 0, len(ports))\n\tfor _, port := range ports {\n\t\tuuids = append(uuids, port.uuid)\n\t}\n\n\tcmd, err := ovnNBClient.PortGroupUpdate(portGroupName, uuids, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn ovnNBClient.Execute(cmd)\n}", "func (a *ArbitraryType) SetResourceGroup(resourceGroup string) {}", "func (m *User) SetCalendarGroups(value []CalendarGroupable)() {\n m.calendarGroups = value\n}", "func (_BaseContentType *BaseContentTypeTransactor) SetGroupRights(opts *bind.TransactOpts, group common.Address, access_type uint8, access uint8) (*types.Transaction, error) {\n\treturn _BaseContentType.contract.Transact(opts, \"setGroupRights\", group, access_type, access)\n}", "func resourceNetboxOrgTenantGroupCreate(d *schema.ResourceData, meta interface{}) error {\n\tnetboxClient := meta.(*ProviderNetboxClient).client\n\n\tname := d.Get(\"name\").(string)\n\tslug := d.Get(\"slug\").(string)\n\n\tvar parm = tenancy.NewTenancyTenantGroupsCreateParams().WithData(\n\t\t&models.TenantGroup{\n\t\t\tName: &name,\n\t\t\tSlug: &slug,\n\t\t},\n\t)\n\n\tlog.Debugf(\"Executing TenancyTenantGroupsCreate against Netbox: %v\", parm)\n\n\tout, err := netboxClient.Tenancy.TenancyTenantGroupsCreate(parm, nil)\n\n\tif err != nil {\n\t\tlog.Debugf(\"Failed to execute TenancyTenantGroupsCreate: %v\", err)\n\n\t\treturn err\n\t}\n\n\t// TODO Probably a better way to parse this ID\n\td.SetId(fmt.Sprintf(\"org/tenant-group/%d\", out.Payload.ID))\n\td.Set(\"tenant_group_id\", out.Payload.ID)\n\n\tlog.Debugf(\"Done Executing TenancyTenantGroupsCreate: %v\", out)\n\n\treturn nil\n}", "func (m *DeviceCompliancePolicySettingStateSummary) SetTenantId(value *string)() {\n err := m.GetBackingStore().Set(\"tenantId\", value)\n if err != nil {\n panic(err)\n }\n}", "func (instance *Host) EnableSecurityGroup(ctx context.Context, sg resources.SecurityGroup) (ferr fail.Error) {\n\tdefer fail.OnPanic(&ferr)\n\n\tif valid.IsNil(instance) {\n\t\treturn fail.InvalidInstanceError()\n\t}\n\tif ctx == nil {\n\t\treturn fail.InvalidParameterError(\"ctx\", \"cannot be nil\")\n\t}\n\tif sg == nil {\n\t\treturn fail.InvalidParameterError(\"sg\", \"cannot be null value of 'SecurityGroup'\")\n\t}\n\n\thid, err := instance.GetID()\n\tif err != nil {\n\t\treturn fail.ConvertError(err)\n\t}\n\n\tsgName := sg.GetName()\n\n\t// instance.Lock()\n\t// defer instance.Unlock()\n\n\tsvc := instance.Service()\n\txerr := instance.Alter(ctx, func(_ data.Clonable, props *serialize.JSONProperties) fail.Error {\n\t\treturn props.Inspect(hostproperty.SecurityGroupsV1, func(clonable data.Clonable) fail.Error {\n\t\t\thsgV1, ok := clonable.(*propertiesv1.HostSecurityGroups)\n\t\t\tif !ok {\n\t\t\t\treturn fail.InconsistentError(\"'*propertiesv1.HostSecurityGroups' expected, '%s' provided\", reflect.TypeOf(clonable).String())\n\t\t\t}\n\n\t\t\tvar asg *abstract.SecurityGroup\n\t\t\txerr := sg.Inspect(ctx, func(clonable data.Clonable, _ *serialize.JSONProperties) fail.Error {\n\t\t\t\tvar ok bool\n\t\t\t\tif asg, ok = clonable.(*abstract.SecurityGroup); !ok {\n\t\t\t\t\treturn fail.InconsistentError(\"'*abstract.SecurityGroup' expected, '%s' provided\", reflect.TypeOf(clonable).String())\n\t\t\t\t}\n\n\t\t\t\treturn nil\n\t\t\t})\n\t\t\txerr = debug.InjectPlannedFail(xerr)\n\t\t\tif xerr != nil {\n\t\t\t\treturn xerr\n\t\t\t}\n\n\t\t\t// First check if the security group is not already registered for the Host with the exact same state\n\t\t\tvar found bool\n\t\t\tfor k := range hsgV1.ByID {\n\t\t\t\tif k == asg.ID {\n\t\t\t\t\tfound = true\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\tif !found {\n\t\t\t\treturn fail.NotFoundError(\"security group '%s' is not bound to Host '%s'\", sgName, hid)\n\t\t\t}\n\n\t\t\t{\n\t\t\t\t// Bind the security group on provider side; if already bound (*fail.ErrDuplicate), considered as a success\n\t\t\t\txerr = svc.BindSecurityGroupToHost(ctx, asg, hid)\n\t\t\t\txerr = debug.InjectPlannedFail(xerr)\n\t\t\t\tif xerr != nil {\n\t\t\t\t\tswitch xerr.(type) {\n\t\t\t\t\tcase *fail.ErrDuplicate:\n\t\t\t\t\t\tdebug.IgnoreError2(ctx, xerr)\n\t\t\t\t\tdefault:\n\t\t\t\t\t\treturn xerr\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// found and updated, update metadata\n\t\t\thsgV1.ByID[asg.ID].Disabled = false\n\t\t\treturn nil\n\t\t})\n\t})\n\tif xerr != nil {\n\t\treturn xerr\n\t}\n\n\treturn nil\n}", "func (u *User) EnableTenantAccess(tenantIDs ...string) {\n\tfor i := range tenantIDs {\n\t\tfound := false\n\t\tfor j := range u.AllowedTenantAccess {\n\t\t\tif tenantIDs[i] == u.AllowedTenantAccess[j] {\n\t\t\t\tfound = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif !found {\n\t\t\tu.AllowedTenantAccess = append(u.AllowedTenantAccess, tenantIDs[i])\n\t\t}\n\t}\n}", "func (p *AccumuloProxyClient) SetLocalityGroups(login []byte, tableName string, groups map[string]map[string]bool) (err error) {\n\tif err = p.sendSetLocalityGroups(login, tableName, groups); err != nil {\n\t\treturn\n\t}\n\treturn p.recvSetLocalityGroups()\n}", "func (o *CloudAwsVirtualMachineAllOf) SetSecurityGroups(v []CloudAwsSecurityGroupRelationship) {\n\to.SecurityGroups = v\n}", "func (m *Group) SetSettings(value []GroupSettingable)() {\n m.settings = value\n}", "func resourceNetboxOrgTenantGroupUpdate(d *schema.ResourceData, meta interface{}) error {\n\tnetboxClient := meta.(*ProviderNetboxClient).client\n\n\tid := int64(d.Get(\"tenant_group_id\").(int))\n\n\tname := d.Get(\"name\").(string)\n\tslug := d.Get(\"slug\").(string)\n\n\tvar parm = tenancy.NewTenancyTenantGroupsUpdateParams().\n\t\tWithID(id).\n\t\tWithData(\n\t\t\t&models.TenantGroup{\n\t\t\t\tName: &name,\n\t\t\t\tSlug: &slug,\n\t\t\t},\n\t\t)\n\n\tlog.Debugf(\"Executing TenancyTenantGroupsUpdate against Netbox: %v\", parm)\n\n\tout, err := netboxClient.Tenancy.TenancyTenantGroupsUpdate(parm, nil)\n\n\tif err != nil {\n\t\tlog.Debugf(\"Failed to execute TenancyTenantGroupsUpdate: %v\", err)\n\n\t\treturn err\n\t}\n\n\tlog.Debugf(\"Done Executing TenancyTenantGroupsUpdate: %v\", out)\n\n\treturn nil\n}", "func (stg *securityTestGroup) testSecurityGroupCreateDelete() {\n\t// sg params\n\tsg := security.SecurityGroup{\n\t\tTypeMeta: api.TypeMeta{Kind: \"SecurityGroup\"},\n\t\tObjectMeta: api.ObjectMeta{\n\t\t\tTenant: \"default\",\n\t\t\tNamespace: \"default\",\n\t\t\tName: \"group1\",\n\t\t},\n\t\tSpec: security.SecurityGroupSpec{\n\t\t\tWorkloadSelector: labels.SelectorFromSet(labels.Set{\"env\": \"production\", \"app\": \"procurement\"}),\n\t\t},\n\t}\n\tvar securityGroupRestIf security.SecurityV1SecurityGroupInterface\n\tapiGwAddr := ts.tu.ClusterVIP + \":\" + globals.APIGwRESTPort\n\trestSvc, err := apiclient.NewRestAPIClient(apiGwAddr)\n\tif err == nil {\n\t\tBy(\"Creating SecurityGroup Client ------\")\n\t\tsecurityGroupRestIf = restSvc.SecurityV1().SecurityGroup()\n\t}\n\tExpect(err).ShouldNot(HaveOccurred())\n\tExpect(stg.securityRestIf).ShouldNot(Equal(nil))\n\n\tctx := ts.tu.MustGetLoggedInContext(context.Background())\n\t// create sg policy\n\tresp, err := securityGroupRestIf.Create(ctx, &sg)\n\tExpect(err).ShouldNot(HaveOccurred())\n\n\t// verify we can read the policy back\n\trsg, err := securityGroupRestIf.Get(ctx, &sg.ObjectMeta)\n\tExpect(err).ShouldNot(HaveOccurred())\n\tExpect(rsg).Should(Equal(resp))\n\n\t// verify agents have the policy\n\tEventually(func() bool {\n\t\tfor _, naplesIP := range ts.tu.NaplesNodeIPs {\n\t\t\tsglist, err := stg.getSecurityGroups(naplesIP)\n\t\t\tif err != nil {\n\t\t\t\tBy(fmt.Sprintf(\"ts:%s security group list failed, err: %+v sgs: %+v\", time.Now().String(), err, sglist))\n\t\t\t\treturn false\n\t\t\t}\n\t\t\tif (len(sglist) != 1) || (sglist[0].Name != sg.Name) {\n\t\t\t\tBy(fmt.Sprintf(\"ts:%s security group list has invalid items, security groups: %+v\", time.Now().String(), sglist))\n\t\t\t\treturn false\n\t\t\t}\n\t\t\tBy(fmt.Sprintf(\"ts:%s security group list success, err: %+v sgs: %+v\", time.Now().String(), err, sglist))\n\t\t}\n\t\treturn true\n\t}, 30, 1).Should(BeTrue(), \"Failed to get security groups on netagent\")\n\n\t// delete the sg policy\n\tEventually(func() error {\n\t\t_, err = securityGroupRestIf.Delete(ctx, &sg.ObjectMeta)\n\t\treturn err\n\t}, 30, 1).ShouldNot(HaveOccurred())\n\n\t// verify policy is gone from the agents\n\tEventually(func() bool {\n\t\tfor _, naplesIP := range ts.tu.NaplesNodeIPs {\n\t\t\tsglist, err := stg.getSecurityGroups(naplesIP)\n\t\t\tif err != nil {\n\t\t\t\tBy(fmt.Sprintf(\"ts:%s security group list failed, err: %+v sgs: %+v\", time.Now().String(), err, sglist))\n\t\t\t\treturn false\n\t\t\t}\n\t\t\tif len(sglist) != 0 {\n\t\t\t\tBy(fmt.Sprintf(\"ts:%s security group list has invalid items, sg groups: %+v\", time.Now().String(), sglist))\n\t\t\t\treturn false\n\t\t\t}\n\t\t\tBy(fmt.Sprintf(\"ts:%s security group list success, err: %+v sgs: %+v\", time.Now().String(), err, sglist))\n\t\t}\n\t\treturn true\n\t}, 30, 1).Should(BeTrue(), \"Failed to get security groups on netagent\")\n}", "func (instance *Host) BindSecurityGroup(ctx context.Context, sgInstance resources.SecurityGroup, enable resources.SecurityGroupActivation) (ferr fail.Error) {\n\tdefer fail.OnPanic(&ferr)\n\n\tif valid.IsNil(instance) {\n\t\treturn fail.InvalidInstanceError()\n\t}\n\tif ctx == nil {\n\t\treturn fail.InvalidParameterCannotBeNilError(\"ctx\")\n\t}\n\tif sgInstance == nil {\n\t\treturn fail.InvalidParameterCannotBeNilError(\"sgInstance\")\n\t}\n\n\t// instance.Lock()\n\t// defer instance.Unlock()\n\n\txerr := instance.Alter(ctx, func(_ data.Clonable, props *serialize.JSONProperties) fail.Error {\n\t\treturn props.Alter(hostproperty.SecurityGroupsV1, func(clonable data.Clonable) fail.Error {\n\t\t\thsgV1, ok := clonable.(*propertiesv1.HostSecurityGroups)\n\t\t\tif !ok {\n\t\t\t\treturn fail.InconsistentError(\"'*propertiesv1.HostSecurityGroups' expected, '%s' provided\", reflect.TypeOf(clonable).String())\n\t\t\t}\n\n\t\t\tsgID, err := sgInstance.GetID()\n\t\t\tif err != nil {\n\t\t\t\treturn fail.ConvertError(err)\n\t\t\t}\n\n\t\t\t// If the Security Group is already bound to the Host with the exact same state, considered as a success\n\t\t\titem, ok := hsgV1.ByID[sgID]\n\t\t\tif ok && item.Disabled == !bool(enable) {\n\t\t\t\treturn nil\n\t\t\t}\n\n\t\t\tif !ok { // Not found, update bind metadata of Host\n\t\t\t\titem = &propertiesv1.SecurityGroupBond{\n\t\t\t\t\tID: sgID,\n\t\t\t\t\tName: sgInstance.GetName(),\n\t\t\t\t}\n\t\t\t\thsgV1.ByID[sgID] = item\n\t\t\t\thsgV1.ByName[item.Name] = item.ID\n\t\t\t}\n\t\t\titem.Disabled = bool(!enable)\n\n\t\t\t// If enabled, apply it\n\t\t\tsgInstanceImpl, ok := sgInstance.(*SecurityGroup)\n\t\t\tif !ok {\n\t\t\t\treturn fail.InconsistentError(\"failed to cast sgInstance to '*SecurityGroup\")\n\t\t\t}\n\n\t\t\tinnerXErr := sgInstanceImpl.unsafeBindToHost(ctx, instance, enable, resources.MarkSecurityGroupAsSupplemental)\n\t\t\tif innerXErr != nil {\n\t\t\t\tswitch innerXErr.(type) {\n\t\t\t\tcase *fail.ErrDuplicate:\n\t\t\t\t\t// already bound, consider as a success\n\t\t\t\t\tbreak\n\t\t\t\tdefault:\n\t\t\t\t\treturn innerXErr\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\t})\n\tif xerr != nil {\n\t\treturn xerr\n\t}\n\n\treturn nil\n}", "func (_BaseContent *BaseContentTransactor) SetGroupRights(opts *bind.TransactOpts, group common.Address, access_type uint8, access uint8) (*types.Transaction, error) {\n\treturn _BaseContent.contract.Transact(opts, \"setGroupRights\", group, access_type, access)\n}", "func (a *AliyunInstanceAttribute) createBastionHostSecurityGroup() {\n\tres, err := ExecCmdReturnOutput(\"bash\", \"-c\", \"aliyun ecs DescribeSecurityGroups --VpcId=\"+a.VpcID)\n\tcheckError(err)\n\tdecodedQuery := decodeAndQueryFromJSONString(res)\n\n\tsecurityGroupNames, err := decodedQuery.Array(\"SecurityGroups\", \"SecurityGroup\")\n\tcheckError(err)\n\tsecurityGroupExists := false\n\tfor _, iter := range securityGroupNames {\n\t\tsecurityGroup := jsonq.NewQuery(iter)\n\t\tname, err := securityGroup.String(\"SecurityGroupName\")\n\t\tcheckError(err)\n\t\tif name == a.BastionSecurityGroupName {\n\t\t\tsecurityGroupExists = true\n\t\t\ta.BastionSecurityGroupID, err = securityGroup.String(\"SecurityGroupId\")\n\t\t\tcheckError(err)\n\t\t\tfmt.Println(\"Configuring bastion host security group rules...\")\n\t\t\tcreateSGCmdString := \"aliyun ecs AuthorizeSecurityGroup --Policy Accept --NicType intranet --Priority 1 --SourceCidrIp \" + a.MyPublicIP + \" --PortRange 22/22 --IpProtocol tcp --SecurityGroupId=\" + a.BastionSecurityGroupID\n\t\t\t_, err = ExecCmdReturnOutput(\"bash\", \"-c\", createSGCmdString)\n\t\t\tcheckError(err)\n\t\t\ttime.Sleep(time.Second * 10)\n\t\t\tfmt.Println(\"Bastion host security group rules configured.\")\n\t\t}\n\t}\n\n\tif !securityGroupExists {\n\t\tres, err = ExecCmdReturnOutput(\"bash\", \"-c\", \"aliyun ecs CreateSecurityGroup --RegionId=\"+a.RegionID+\" --VpcId=\"+a.VpcID+\" --SecurityGroupName=\"+a.BastionSecurityGroupName)\n\t\tcheckError(err)\n\t\tdecodedQuery = decodeAndQueryFromJSONString(res)\n\t\ta.BastionSecurityGroupID, err = decodedQuery.String(\"SecurityGroupId\")\n\t\tcheckError(err)\n\t\tattemptCnt := 0\n\t\tfor attemptCnt < 60 {\n\t\t\tres, err = ExecCmdReturnOutput(\"bash\", \"-c\", \"aliyun ecs DescribeSecurityGroups --SecurityGroupIds=\\\"['\"+a.BastionSecurityGroupID+\"']\\\"\")\n\t\t\tcheckError(err)\n\t\t\tdecodedQuery = decodeAndQueryFromJSONString(res)\n\t\t\ttotalCount, err := decodedQuery.Int(\"TotalCount\")\n\t\t\tcheckError(err)\n\t\t\tif totalCount == 1 {\n\t\t\t\ttime.Sleep(time.Second * 30)\n\t\t\t\tfmt.Println(\"Bastion host security group created.\")\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tfmt.Println(\"Creating bastion host security group...\")\n\t\t\ttime.Sleep(time.Second * 2)\n\t\t\tattemptCnt++\n\t\t}\n\t\tif attemptCnt == 60 {\n\t\t\tfmt.Println(\"Bastion host security group creation time out. Please try again.\")\n\t\t\tos.Exit(2)\n\t\t}\n\t\tfmt.Println(\"Configuring bastion host security group rules...\")\n\t\tcreateSGCmdString := \"aliyun ecs AuthorizeSecurityGroup --Policy Accept --NicType intranet --Priority 1 --SourceCidrIp \" + a.MyPublicIP + \" --PortRange 22/22 --IpProtocol tcp --SecurityGroupId=\" + a.BastionSecurityGroupID\n\t\t_, err = ExecCmdReturnOutput(\"bash\", \"-c\", createSGCmdString)\n\t\tcheckError(err)\n\t\ttime.Sleep(time.Second * 10)\n\t\tfmt.Println(\"Bastion host security group rules configured.\")\n\t}\n}", "func (db *Database) SetSecurity(o SecurityObject) error {\n\theaders := map[string]string{\"Content-Type\": \"application/json\"}\n\tpayload, err := json.Marshal(o)\n\tif err != nil {\n\t\treturn err\n\t}\n\tresp, err := db.conn.request(\"PUT\", queryURL(db.Name, \"_security\"), headers, bytes.NewReader(payload), db.auth, 0)\n\tif err != nil {\n\t\treturn err\n\t}\n\tvar result map[string]bool\n\tif err := parseBody(resp, &result); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func setTenant(c *cli.Context) error {\n\terr := checkArgCount(c, 1)\n\tif err != nil {\n\t\treturn err\n\t}\n\tname := c.Args().First()\n\n\tclient.Photonclient, err = client.GetClient(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Ensure tenant exists\n\tid, err := findTenantID(name)\n\tif len(id) == 0 || err != nil {\n\t\treturn err\n\t}\n\n\tconfig, err := cf.LoadConfig()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tconfig.Tenant = &cf.TenantConfiguration{Name: name, ID: id}\n\terr = cf.SaveConfig(config)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = clearConfigProject(\"\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif !c.GlobalIsSet(\"non-interactive\") {\n\t\tfmt.Printf(\"Tenant set to '%s'\\n\", name)\n\t}\n\treturn nil\n}", "func (_BaseLibrary *BaseLibraryTransactor) SetGroupRights(opts *bind.TransactOpts, group common.Address, access_type uint8, access uint8) (*types.Transaction, error) {\n\treturn _BaseLibrary.contract.Transact(opts, \"setGroupRights\", group, access_type, access)\n}", "func (m *SecureScoreControlProfile) SetAzureTenantId(value *string)() {\n err := m.GetBackingStore().Set(\"azureTenantId\", value)\n if err != nil {\n panic(err)\n }\n}", "func (o *CreateLoadBalancerRequest) SetSecurityGroups(v []string) {\n\to.SecurityGroups = &v\n}", "func (_BaseAccessWallet *BaseAccessWalletTransactor) SetAccessGroupRights(opts *bind.TransactOpts, group common.Address, access_type uint8, access uint8) (*types.Transaction, error) {\n\treturn _BaseAccessWallet.contract.Transact(opts, \"setAccessGroupRights\", group, access_type, access)\n}", "func (swagger *MgwSwagger) SetSecurity(security []map[string][]string) {\n\tswagger.security = security\n}", "func (m *ProtectGroup) SetAllowGuestUsers(value *bool)() {\n err := m.GetBackingStore().Set(\"allowGuestUsers\", value)\n if err != nil {\n panic(err)\n }\n}", "func (c *Client) BindSecurityGroups(args *BindSecurityGroupsArgs) error {\n\tjsonBytes, jsonErr := json.Marshal(args)\n\tif jsonErr != nil {\n\t\treturn jsonErr\n\t}\n\tbody, err := bce.NewBodyFromBytes(jsonBytes)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn BindSecurityGroups(c, body)\n}", "func (client *ClientImpl) SetAccessControlLists(ctx context.Context, args SetAccessControlListsArgs) error {\n\tif args.AccessControlLists == nil {\n\t\treturn &azuredevops.ArgumentNilError{ArgumentName: \"args.AccessControlLists\"}\n\t}\n\trouteValues := make(map[string]string)\n\tif args.SecurityNamespaceId == nil {\n\t\treturn &azuredevops.ArgumentNilError{ArgumentName: \"args.SecurityNamespaceId\"}\n\t}\n\trouteValues[\"securityNamespaceId\"] = (*args.SecurityNamespaceId).String()\n\n\tbody, marshalErr := json.Marshal(*args.AccessControlLists)\n\tif marshalErr != nil {\n\t\treturn marshalErr\n\t}\n\tlocationId, _ := uuid.Parse(\"18a2ad18-7571-46ae-bec7-0c7da1495885\")\n\t_, err := client.Client.Send(ctx, http.MethodPost, locationId, \"7.1-preview.1\", routeValues, nil, bytes.NewReader(body), \"application/json\", \"application/json\", nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (m *Group) SetOnPremisesSecurityIdentifier(value *string)() {\n m.onPremisesSecurityIdentifier = value\n}", "func (v *Vps) PopulateSecurityGroups(os *OpenStack) error {\n\tvar err error\n\n\t// Security Groups\n\tresult := servers.GetResult{}\n\turl := os.Compute.ServiceURL(\"servers\", v.ID, \"os-security-groups\")\n\t_, err = os.Compute.Get(url, &result.Body, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar resp struct {\n\t\tSecurityGroups []secgroups.SecurityGroup `mapstructure:\"security_groups\"`\n\t}\n\n\tif err = mapstructure.Decode(result.Body, &resp); err != nil {\n\t\treturn err\n\t}\n\tv.SecurityGroups = resp.SecurityGroups\n\n\treturn nil\n}", "func (stg *securityTestGroup) getSecurityGroups(agent string) ([]security.NetworkSecurityPolicy, error) {\n\tvar sgplist []security.NetworkSecurityPolicy\n\tstatus, err := stg.authAgentClient.Req(\"GET\", \"https://\"+agent+\":\"+globals.AgentProxyPort+\"/api/sgs/\", nil, &sgplist)\n\tif err != nil || status != http.StatusOK {\n\t\treturn nil, fmt.Errorf(\"Error getting SG Policies list: %v\", err)\n\t}\n\treturn sgplist, nil\n}", "func (_AccessIndexor *AccessIndexorTransactor) SetAccessGroupRights(opts *bind.TransactOpts, group common.Address, access_type uint8, access uint8) (*types.Transaction, error) {\n\treturn _AccessIndexor.contract.Transact(opts, \"setAccessGroupRights\", group, access_type, access)\n}", "func (m *ReportRoot) SetSecurity(value SecurityReportsRootable)() {\n err := m.GetBackingStore().Set(\"security\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *AccessPackage) SetIncompatibleGroups(value []Groupable)() {\n m.incompatibleGroups = value\n}", "func ExampleDeviceSecurityGroupsClient_NewListPager() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclientFactory, err := armsecurity.NewClientFactory(\"<subscription-id>\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tpager := clientFactory.NewDeviceSecurityGroupsClient().NewListPager(\"subscriptions/20ff7fc3-e762-44dd-bd96-b71116dcdc23/resourceGroups/SampleRG/providers/Microsoft.Devices/iotHubs/sampleiothub\", nil)\n\tfor pager.More() {\n\t\tpage, err := pager.NextPage(ctx)\n\t\tif err != nil {\n\t\t\tlog.Fatalf(\"failed to advance page: %v\", err)\n\t\t}\n\t\tfor _, v := range page.Value {\n\t\t\t// You could use page here. We use blank identifier for just demo purposes.\n\t\t\t_ = v\n\t\t}\n\t\t// If the HTTP response code is 200 as defined in example definition, your page structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes.\n\t\t// page.DeviceSecurityGroupList = armsecurity.DeviceSecurityGroupList{\n\t\t// \tValue: []*armsecurity.DeviceSecurityGroup{\n\t\t// \t\t{\n\t\t// \t\t\tName: to.Ptr(\"samplesecuritygroup\"),\n\t\t// \t\t\tType: to.Ptr(\"Microsoft.Security/deviceSecurityGroups\"),\n\t\t// \t\t\tID: to.Ptr(\"/subscriptions/20ff7fc3-e762-44dd-bd96-b71116dcdc23/resourceGroups/SampleRG/providers/Microsoft.Devices/iotHubs/sampleiothub/providers/Microsoft.Security/deviceSecurityGroups/samplesecuritygroup\"),\n\t\t// \t\t\tProperties: &armsecurity.DeviceSecurityGroupProperties{\n\t\t// \t\t\t\tAllowlistRules: []armsecurity.AllowlistCustomAlertRuleClassification{\n\t\t// \t\t\t\t\t&armsecurity.ConnectionToIPNotAllowed{\n\t\t// \t\t\t\t\t\tDescription: to.Ptr(\"Get an alert when an outbound connection is created between your device and an ip that isn't allowed\"),\n\t\t// \t\t\t\t\t\tDisplayName: to.Ptr(\"Outbound connection to an ip that isn't allowed\"),\n\t\t// \t\t\t\t\t\tIsEnabled: to.Ptr(false),\n\t\t// \t\t\t\t\t\tRuleType: to.Ptr(\"ConnectionToIpNotAllowed\"),\n\t\t// \t\t\t\t\t\tValueType: to.Ptr(armsecurity.ValueTypeIPCidr),\n\t\t// \t\t\t\t\t\tAllowlistValues: []*string{\n\t\t// \t\t\t\t\t\t},\n\t\t// \t\t\t\t\t},\n\t\t// \t\t\t\t\t&armsecurity.LocalUserNotAllowed{\n\t\t// \t\t\t\t\t\tDescription: to.Ptr(\"Get an alert when a local user that isn't allowed logins to the device\"),\n\t\t// \t\t\t\t\t\tDisplayName: to.Ptr(\"Login by a local user that isn't allowed\"),\n\t\t// \t\t\t\t\t\tIsEnabled: to.Ptr(false),\n\t\t// \t\t\t\t\t\tRuleType: to.Ptr(\"LocalUserNotAllowed\"),\n\t\t// \t\t\t\t\t\tValueType: to.Ptr(armsecurity.ValueTypeString),\n\t\t// \t\t\t\t\t\tAllowlistValues: []*string{\n\t\t// \t\t\t\t\t\t},\n\t\t// \t\t\t\t\t},\n\t\t// \t\t\t\t\t&armsecurity.ProcessNotAllowed{\n\t\t// \t\t\t\t\t\tDescription: to.Ptr(\"Get an alert when a process that isn't allowed is executed\"),\n\t\t// \t\t\t\t\t\tDisplayName: to.Ptr(\"Execution of a process that isn't allowed\"),\n\t\t// \t\t\t\t\t\tIsEnabled: to.Ptr(false),\n\t\t// \t\t\t\t\t\tRuleType: to.Ptr(\"ProcessNotAllowed\"),\n\t\t// \t\t\t\t\t\tValueType: to.Ptr(armsecurity.ValueTypeString),\n\t\t// \t\t\t\t\t\tAllowlistValues: []*string{\n\t\t// \t\t\t\t\t\t},\n\t\t// \t\t\t\t}},\n\t\t// \t\t\t\tDenylistRules: []*armsecurity.DenylistCustomAlertRule{\n\t\t// \t\t\t\t},\n\t\t// \t\t\t\tThresholdRules: []armsecurity.ThresholdCustomAlertRuleClassification{\n\t\t// \t\t\t\t},\n\t\t// \t\t\t\tTimeWindowRules: []armsecurity.TimeWindowCustomAlertRuleClassification{\n\t\t// \t\t\t\t\t&armsecurity.ActiveConnectionsNotInAllowedRange{\n\t\t// \t\t\t\t\t\tDescription: to.Ptr(\"Get an alert when the number of active connections of a device in the time window is not in the allowed range\"),\n\t\t// \t\t\t\t\t\tDisplayName: to.Ptr(\"Number of active connections is not in allowed range\"),\n\t\t// \t\t\t\t\t\tIsEnabled: to.Ptr(false),\n\t\t// \t\t\t\t\t\tRuleType: to.Ptr(\"ActiveConnectionsNotInAllowedRange\"),\n\t\t// \t\t\t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t\t// \t\t\t\t\t},\n\t\t// \t\t\t\t\t&armsecurity.AmqpC2DMessagesNotInAllowedRange{\n\t\t// \t\t\t\t\t\tDescription: to.Ptr(\"Get an alert when the number of cloud to device messages (AMQP protocol) in the time window is not in the allowed range\"),\n\t\t// \t\t\t\t\t\tDisplayName: to.Ptr(\"Number of cloud to device messages (AMQP protocol) is not in allowed range\"),\n\t\t// \t\t\t\t\t\tIsEnabled: to.Ptr(false),\n\t\t// \t\t\t\t\t\tRuleType: to.Ptr(\"AmqpC2DMessagesNotInAllowedRange\"),\n\t\t// \t\t\t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t\t// \t\t\t\t\t},\n\t\t// \t\t\t\t\t&armsecurity.MqttC2DMessagesNotInAllowedRange{\n\t\t// \t\t\t\t\t\tDescription: to.Ptr(\"Get an alert when the number of cloud to device messages (MQTT protocol) in the time window is not in the allowed range\"),\n\t\t// \t\t\t\t\t\tDisplayName: to.Ptr(\"Number of cloud to device messages (MQTT protocol) is not in allowed range\"),\n\t\t// \t\t\t\t\t\tIsEnabled: to.Ptr(false),\n\t\t// \t\t\t\t\t\tRuleType: to.Ptr(\"MqttC2DMessagesNotInAllowedRange\"),\n\t\t// \t\t\t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t\t// \t\t\t\t\t},\n\t\t// \t\t\t\t\t&armsecurity.HTTPC2DMessagesNotInAllowedRange{\n\t\t// \t\t\t\t\t\tDescription: to.Ptr(\"Get an alert when the number of cloud to device messages (HTTP protocol) in the time window is not in the allowed range\"),\n\t\t// \t\t\t\t\t\tDisplayName: to.Ptr(\"Number of cloud to device messages (HTTP protocol) is not in allowed range\"),\n\t\t// \t\t\t\t\t\tIsEnabled: to.Ptr(false),\n\t\t// \t\t\t\t\t\tRuleType: to.Ptr(\"HttpC2DMessagesNotInAllowedRange\"),\n\t\t// \t\t\t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t\t// \t\t\t\t\t},\n\t\t// \t\t\t\t\t&armsecurity.AmqpC2DRejectedMessagesNotInAllowedRange{\n\t\t// \t\t\t\t\t\tDescription: to.Ptr(\"Get an alert when the number of cloud to device messages (AMQP protocol) that were rejected by the device in the time window is not in the allowed range\"),\n\t\t// \t\t\t\t\t\tDisplayName: to.Ptr(\"Number of rejected cloud to device messages (AMQP protocol) is not in allowed range\"),\n\t\t// \t\t\t\t\t\tIsEnabled: to.Ptr(false),\n\t\t// \t\t\t\t\t\tRuleType: to.Ptr(\"AmqpC2DRejectedMessagesNotInAllowedRange\"),\n\t\t// \t\t\t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t\t// \t\t\t\t\t},\n\t\t// \t\t\t\t\t&armsecurity.MqttC2DRejectedMessagesNotInAllowedRange{\n\t\t// \t\t\t\t\t\tDescription: to.Ptr(\"Get an alert when the number of cloud to device messages (MQTT protocol) that were rejected by the device in the time window is not in the allowed range\"),\n\t\t// \t\t\t\t\t\tDisplayName: to.Ptr(\"Number of rejected cloud to device messages (MQTT protocol) is not in allowed range\"),\n\t\t// \t\t\t\t\t\tIsEnabled: to.Ptr(false),\n\t\t// \t\t\t\t\t\tRuleType: to.Ptr(\"MqttC2DRejectedMessagesNotInAllowedRange\"),\n\t\t// \t\t\t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t\t// \t\t\t\t\t},\n\t\t// \t\t\t\t\t&armsecurity.HTTPC2DRejectedMessagesNotInAllowedRange{\n\t\t// \t\t\t\t\t\tDescription: to.Ptr(\"Get an alert when the number of cloud to device messages (HTTP protocol) that were rejected by the device in the time window is not in the allowed range\"),\n\t\t// \t\t\t\t\t\tDisplayName: to.Ptr(\"Number of rejected cloud to device messages (HTTP protocol) is not in allowed range\"),\n\t\t// \t\t\t\t\t\tIsEnabled: to.Ptr(false),\n\t\t// \t\t\t\t\t\tRuleType: to.Ptr(\"HttpC2DRejectedMessagesNotInAllowedRange\"),\n\t\t// \t\t\t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t\t// \t\t\t\t\t},\n\t\t// \t\t\t\t\t&armsecurity.AmqpD2CMessagesNotInAllowedRange{\n\t\t// \t\t\t\t\t\tDescription: to.Ptr(\"Get an alert when the number of device to cloud messages (AMQP protocol) in the time window is not in the allowed range\"),\n\t\t// \t\t\t\t\t\tDisplayName: to.Ptr(\"Number of device to cloud messages (AMQP protocol) is not in allowed range\"),\n\t\t// \t\t\t\t\t\tIsEnabled: to.Ptr(false),\n\t\t// \t\t\t\t\t\tRuleType: to.Ptr(\"AmqpD2CMessagesNotInAllowedRange\"),\n\t\t// \t\t\t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t\t// \t\t\t\t\t},\n\t\t// \t\t\t\t\t&armsecurity.MqttD2CMessagesNotInAllowedRange{\n\t\t// \t\t\t\t\t\tDescription: to.Ptr(\"Get an alert when the number of device to cloud messages (MQTT protocol) in the time window is not in the allowed range\"),\n\t\t// \t\t\t\t\t\tDisplayName: to.Ptr(\"Number of device to cloud messages (MQTT protocol) is not in allowed range\"),\n\t\t// \t\t\t\t\t\tIsEnabled: to.Ptr(false),\n\t\t// \t\t\t\t\t\tRuleType: to.Ptr(\"MqttD2CMessagesNotInAllowedRange\"),\n\t\t// \t\t\t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t\t// \t\t\t\t\t},\n\t\t// \t\t\t\t\t&armsecurity.HTTPD2CMessagesNotInAllowedRange{\n\t\t// \t\t\t\t\t\tDescription: to.Ptr(\"Get an alert when the number of device to cloud messages (HTTP protocol) in the time window is not in the allowed range\"),\n\t\t// \t\t\t\t\t\tDisplayName: to.Ptr(\"Number of device to cloud messages (HTTP protocol) is not in allowed range\"),\n\t\t// \t\t\t\t\t\tIsEnabled: to.Ptr(false),\n\t\t// \t\t\t\t\t\tRuleType: to.Ptr(\"HttpD2CMessagesNotInAllowedRange\"),\n\t\t// \t\t\t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t\t// \t\t\t\t\t},\n\t\t// \t\t\t\t\t&armsecurity.DirectMethodInvokesNotInAllowedRange{\n\t\t// \t\t\t\t\t\tDescription: to.Ptr(\"Get an alert when the number of direct method invokes in the time window is not in the allowed range\"),\n\t\t// \t\t\t\t\t\tDisplayName: to.Ptr(\"Number of direct method invokes is not in allowed range\"),\n\t\t// \t\t\t\t\t\tIsEnabled: to.Ptr(false),\n\t\t// \t\t\t\t\t\tRuleType: to.Ptr(\"DirectMethodInvokesNotInAllowedRange\"),\n\t\t// \t\t\t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t\t// \t\t\t\t\t},\n\t\t// \t\t\t\t\t&armsecurity.FailedLocalLoginsNotInAllowedRange{\n\t\t// \t\t\t\t\t\tDescription: to.Ptr(\"Get an alert when the number of failed local logins on the device in the time window is not in the allowed range\"),\n\t\t// \t\t\t\t\t\tDisplayName: to.Ptr(\"Number of failed local logins is not in allowed range\"),\n\t\t// \t\t\t\t\t\tIsEnabled: to.Ptr(false),\n\t\t// \t\t\t\t\t\tRuleType: to.Ptr(\"FailedLocalLoginsNotInAllowedRange\"),\n\t\t// \t\t\t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t\t// \t\t\t\t\t},\n\t\t// \t\t\t\t\t&armsecurity.FileUploadsNotInAllowedRange{\n\t\t// \t\t\t\t\t\tDescription: to.Ptr(\"Get an alert when the number of file uploads from the device to the cloud in the time window is not in the allowed range\"),\n\t\t// \t\t\t\t\t\tDisplayName: to.Ptr(\"Number of file uploads is not in allowed range\"),\n\t\t// \t\t\t\t\t\tIsEnabled: to.Ptr(false),\n\t\t// \t\t\t\t\t\tRuleType: to.Ptr(\"FileUploadsNotInAllowedRange\"),\n\t\t// \t\t\t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t\t// \t\t\t\t\t},\n\t\t// \t\t\t\t\t&armsecurity.QueuePurgesNotInAllowedRange{\n\t\t// \t\t\t\t\t\tDescription: to.Ptr(\"Get an alert when the number of device queue purges in the time window is not in the allowed range\"),\n\t\t// \t\t\t\t\t\tDisplayName: to.Ptr(\"Number of device queue purges is not in allowed range\"),\n\t\t// \t\t\t\t\t\tIsEnabled: to.Ptr(false),\n\t\t// \t\t\t\t\t\tRuleType: to.Ptr(\"QueuePurgesNotInAllowedRange\"),\n\t\t// \t\t\t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t\t// \t\t\t\t\t},\n\t\t// \t\t\t\t\t&armsecurity.TwinUpdatesNotInAllowedRange{\n\t\t// \t\t\t\t\t\tDescription: to.Ptr(\"Get an alert when the number of twin updates (by the device or the service) in the time window is not in the allowed range\"),\n\t\t// \t\t\t\t\t\tDisplayName: to.Ptr(\"Number of twin updates is not in allowed range\"),\n\t\t// \t\t\t\t\t\tIsEnabled: to.Ptr(false),\n\t\t// \t\t\t\t\t\tRuleType: to.Ptr(\"TwinUpdatesNotInAllowedRange\"),\n\t\t// \t\t\t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t\t// \t\t\t\t\t},\n\t\t// \t\t\t\t\t&armsecurity.UnauthorizedOperationsNotInAllowedRange{\n\t\t// \t\t\t\t\t\tDescription: to.Ptr(\"Get an alert when the number unauthorized operations in the time window is not in the allowed range. Unauthorized operations are operations that affect the device (or done by it) that fail because of an unauthorized error\"),\n\t\t// \t\t\t\t\t\tDisplayName: to.Ptr(\"Number of unauthorized operations is not in allowed range\"),\n\t\t// \t\t\t\t\t\tIsEnabled: to.Ptr(false),\n\t\t// \t\t\t\t\t\tRuleType: to.Ptr(\"UnauthorizedOperationsNotInAllowedRange\"),\n\t\t// \t\t\t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t\t// \t\t\t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t\t// \t\t\t\t}},\n\t\t// \t\t\t},\n\t\t// \t}},\n\t\t// }\n\t}\n}", "func (_BaseAccessControlGroup *BaseAccessControlGroupTransactor) SetAccessGroupRights(opts *bind.TransactOpts, group common.Address, access_type uint8, access uint8) (*types.Transaction, error) {\n\treturn _BaseAccessControlGroup.contract.Transact(opts, \"setAccessGroupRights\", group, access_type, access)\n}", "func (m *DeviceManagementConfigurationSettingGroupDefinition) SetChildIds(value []string)() {\n err := m.GetBackingStore().Set(\"childIds\", value)\n if err != nil {\n panic(err)\n }\n}", "func createSecurityGroup(client ec2iface.EC2API, vpcID, clusterName string) (string, error) {\n\tvar securityGroupID string\n\n\tnewSecurityGroupName := resourceNamePrefix + clusterName\n\tcsgOut, err := client.CreateSecurityGroup(&ec2.CreateSecurityGroupInput{\n\t\tVpcId: &vpcID,\n\t\tGroupName: aws.String(newSecurityGroupName),\n\t\tDescription: aws.String(fmt.Sprintf(\"Security group for the Kubernetes cluster %s\", clusterName)),\n\t})\n\tif err != nil {\n\t\tif awsErr, ok := err.(awserr.Error); !ok || awsErr.Code() != \"InvalidGroup.Duplicate\" {\n\t\t\treturn \"\", fmt.Errorf(\"failed to create security group %s: %v\", newSecurityGroupName, err)\n\t\t}\n\t\tdescribeOut, err := client.DescribeSecurityGroups(&ec2.DescribeSecurityGroupsInput{\n\t\t\tFilters: []*ec2.Filter{{\n\t\t\t\tName: aws.String(\"group-name\"),\n\t\t\t\tValues: []*string{aws.String(newSecurityGroupName)}}},\n\t\t})\n\t\tif err != nil {\n\t\t\treturn \"\", fmt.Errorf(\"failed to get security group after creation failed because the group already existed: %v\", err)\n\t\t}\n\t\tif n := len(describeOut.SecurityGroups); n != 1 {\n\t\t\treturn \"\", fmt.Errorf(\"expected to get exactly one security group after create failed because the group already existed, got %d\", n)\n\t\t}\n\n\t\tsecurityGroupID = aws.StringValue(describeOut.SecurityGroups[0].GroupId)\n\t}\n\tif csgOut != nil && csgOut.GroupId != nil {\n\t\tsecurityGroupID = *csgOut.GroupId\n\t}\n\tklog.V(2).Infof(\"Security group %s for cluster %s created with id %s.\", newSecurityGroupName, clusterName, securityGroupID)\n\n\t// Add permissions.\n\t_, err = client.AuthorizeSecurityGroupIngress(&ec2.AuthorizeSecurityGroupIngressInput{\n\t\tGroupId: aws.String(securityGroupID),\n\t\tIpPermissions: []*ec2.IpPermission{\n\t\t\t(&ec2.IpPermission{}).\n\t\t\t\t// all protocols from within the sg\n\t\t\t\tSetIpProtocol(\"-1\").\n\t\t\t\tSetUserIdGroupPairs([]*ec2.UserIdGroupPair{\n\t\t\t\t\t(&ec2.UserIdGroupPair{}).\n\t\t\t\t\t\tSetGroupId(securityGroupID),\n\t\t\t\t}),\n\t\t\t(&ec2.IpPermission{}).\n\t\t\t\t// tcp:22 from everywhere\n\t\t\t\tSetIpProtocol(\"tcp\").\n\t\t\t\tSetFromPort(provider.DefaultSSHPort).\n\t\t\t\tSetToPort(provider.DefaultSSHPort).\n\t\t\t\tSetIpRanges([]*ec2.IpRange{\n\t\t\t\t\t{CidrIp: aws.String(\"0.0.0.0/0\")},\n\t\t\t\t}),\n\t\t\t(&ec2.IpPermission{}).\n\t\t\t\t// ICMP from/to everywhere\n\t\t\t\tSetIpProtocol(\"icmp\").\n\t\t\t\tSetFromPort(-1). // any port\n\t\t\t\tSetToPort(-1). // any port\n\t\t\t\tSetIpRanges([]*ec2.IpRange{\n\t\t\t\t\t{CidrIp: aws.String(\"0.0.0.0/0\")},\n\t\t\t\t}),\n\t\t\t(&ec2.IpPermission{}).\n\t\t\t\t// ICMPv6 from/to everywhere\n\t\t\t\tSetIpProtocol(\"icmpv6\").\n\t\t\t\tSetFromPort(-1). // any port\n\t\t\t\tSetToPort(-1). // any port\n\t\t\t\tSetIpv6Ranges([]*ec2.Ipv6Range{\n\t\t\t\t\t{CidrIpv6: aws.String(\"::/0\")},\n\t\t\t\t}),\n\t\t},\n\t})\n\tif err != nil {\n\t\tif awsErr, ok := err.(awserr.Error); !ok || awsErr.Code() != \"InvalidPermission.Duplicate\" {\n\t\t\treturn \"\", fmt.Errorf(\"failed to authorize security group %s with id %s: %v\", newSecurityGroupName, securityGroupID, err)\n\t\t}\n\t}\n\n\treturn securityGroupID, nil\n}", "func (m *Group) SetSites(value []Siteable)() {\n m.sites = value\n}", "func (l *Libvirt) DomainAuthorizedSshKeysSet(Dom Domain, User string, Keys []string, Flags uint32) (err error) {\n\tvar buf []byte\n\n\targs := DomainAuthorizedSshKeysSetArgs {\n\t\tDom: Dom,\n\t\tUser: User,\n\t\tKeys: Keys,\n\t\tFlags: Flags,\n\t}\n\n\tbuf, err = encode(&args)\n\tif err != nil {\n\t\treturn\n\t}\n\n\n\t_, err = l.requestStream(425, constants.Program, buf, nil, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func (s *TestSuiteIAM) TestLDAPSTSServiceAccountsWithGroups(c *check) {\n\tctx, cancel := context.WithTimeout(context.Background(), 10*time.Second)\n\tdefer cancel()\n\n\tbucket := getRandomBucketName()\n\terr := s.client.MakeBucket(ctx, bucket, minio.MakeBucketOptions{})\n\tif err != nil {\n\t\tc.Fatalf(\"bucket create error: %v\", err)\n\t}\n\n\t// Create policy\n\tpolicy := \"mypolicy\"\n\tpolicyBytes := []byte(fmt.Sprintf(`{\n \"Version\": \"2012-10-17\",\n \"Statement\": [\n {\n \"Effect\": \"Allow\",\n \"Action\": [\n \"s3:PutObject\",\n \"s3:GetObject\",\n \"s3:ListBucket\"\n ],\n \"Resource\": [\n \"arn:aws:s3:::%s/*\"\n ]\n }\n ]\n}`, bucket))\n\terr = s.adm.AddCannedPolicy(ctx, policy, policyBytes)\n\tif err != nil {\n\t\tc.Fatalf(\"policy add error: %v\", err)\n\t}\n\n\tgroupDN := \"cn=projecta,ou=groups,ou=swengg,dc=min,dc=io\"\n\terr = s.adm.SetPolicy(ctx, policy, groupDN, true)\n\tif err != nil {\n\t\tc.Fatalf(\"Unable to set policy: %v\", err)\n\t}\n\n\tldapID := cr.LDAPIdentity{\n\t\tClient: s.TestSuiteCommon.client,\n\t\tSTSEndpoint: s.endPoint,\n\t\tLDAPUsername: \"dillon\",\n\t\tLDAPPassword: \"dillon\",\n\t}\n\n\tvalue, err := ldapID.Retrieve()\n\tif err != nil {\n\t\tc.Fatalf(\"Expected to generate STS creds, got err: %#v\", err)\n\t}\n\n\t// Check that the LDAP sts cred is actually working.\n\tminioClient, err := minio.New(s.endpoint, &minio.Options{\n\t\tCreds: cr.NewStaticV4(value.AccessKeyID, value.SecretAccessKey, value.SessionToken),\n\t\tSecure: s.secure,\n\t\tTransport: s.TestSuiteCommon.client.Transport,\n\t})\n\tif err != nil {\n\t\tc.Fatalf(\"Error initializing client: %v\", err)\n\t}\n\n\t// Validate that the client from sts creds can access the bucket.\n\tc.mustListObjects(ctx, minioClient, bucket)\n\n\t// Create an madmin client with user creds\n\tuserAdmClient, err := madmin.NewWithOptions(s.endpoint, &madmin.Options{\n\t\tCreds: cr.NewStaticV4(value.AccessKeyID, value.SecretAccessKey, value.SessionToken),\n\t\tSecure: s.secure,\n\t})\n\tif err != nil {\n\t\tc.Fatalf(\"Err creating user admin client: %v\", err)\n\t}\n\tuserAdmClient.SetCustomTransport(s.TestSuiteCommon.client.Transport)\n\n\t// Create svc acc\n\tcr := c.mustCreateSvcAccount(ctx, value.AccessKeyID, userAdmClient)\n\n\t// 1. Check that svc account appears in listing\n\tc.assertSvcAccAppearsInListing(ctx, userAdmClient, value.AccessKeyID, cr.AccessKey)\n\n\t// 2. Check that svc account info can be queried\n\tc.assertSvcAccInfoQueryable(ctx, userAdmClient, value.AccessKeyID, cr.AccessKey, true)\n\n\t// 3. Check S3 access\n\tc.assertSvcAccS3Access(ctx, s, cr, bucket)\n\n\t// 4. Check that svc account can restrict the policy, and that the\n\t// session policy can be updated.\n\tc.assertSvcAccSessionPolicyUpdate(ctx, s, userAdmClient, value.AccessKeyID, bucket)\n\n\t// 4. Check that service account's secret key and account status can be\n\t// updated.\n\tc.assertSvcAccSecretKeyAndStatusUpdate(ctx, s, userAdmClient, value.AccessKeyID, bucket)\n\n\t// 5. Check that service account can be deleted.\n\tc.assertSvcAccDeletion(ctx, s, userAdmClient, value.AccessKeyID, bucket)\n}", "func (m *FederatedTokenValidationPolicy) SetValidatingDomains(value ValidatingDomainsable)() {\n err := m.GetBackingStore().Set(\"validatingDomains\", value)\n if err != nil {\n panic(err)\n }\n}", "func TestTenantPermissionsUpdate(t *testing.T) {\n\tctx, cancelFunc := context.WithTimeout(context.Background(), standardTimeout)\n\tdefer cancelFunc()\n\n\tconfig := bat.TenantConfig{\n\t\tName: \"TestTenantPermissionsUpdate\",\n\t}\n\n\ttenant, err := bat.CreateTenant(ctx, config)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to create tenant : %v\", err)\n\t}\n\n\tdefer func() {\n\t\terr = bat.DeleteTenant(ctx, tenant.ID)\n\t\tif err != nil {\n\t\t\tt.Fatalf(\"Failed to delete tenant: %v\", err)\n\t\t}\n\t}()\n\n\tif tenant.Name != config.Name {\n\t\tt.Fatalf(\"Failed to create tenant\")\n\t}\n\n\tcfg, err := bat.GetTenantConfig(ctx, tenant.ID)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to retrieve tenant config: %v\", err)\n\t}\n\n\tif cfg.Name != config.Name || cfg.Permissions.PrivilegedContainers != false {\n\t\tt.Fatalf(\"Failed to retrieve tenant config\")\n\t}\n\n\tconfig.Permissions.PrivilegedContainers = true\n\n\terr = bat.UpdateTenant(ctx, tenant.ID, config)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to update tenant: %v\", err)\n\t}\n\n\tcfg, err = bat.GetTenantConfig(ctx, tenant.ID)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to retrieve tenant config: %v\", err)\n\t}\n\n\tif cfg.Name != config.Name || cfg.Permissions.PrivilegedContainers != true {\n\t\tt.Fatalf(\"Failed to retrieve tenant config\")\n\t}\n\n\tconfig.Permissions.PrivilegedContainers = false\n\n\terr = bat.UpdateTenant(ctx, tenant.ID, config)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to update tenant: %v\", err)\n\t}\n\n\tcfg, err = bat.GetTenantConfig(ctx, tenant.ID)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to retrieve tenant config: %v\", err)\n\t}\n\n\tif cfg.Name != config.Name || cfg.Permissions.PrivilegedContainers != false {\n\t\tt.Fatalf(\"Failed to retrieve tenant config\")\n\t}\n}", "func (provider *ResourceProvider) SecurityGroup(id string) (*reachAWS.SecurityGroup, error) {\n\tinput := &ec2.DescribeSecurityGroupsInput{\n\t\tGroupIds: []*string{\n\t\t\taws.String(id),\n\t\t},\n\t}\n\tresult, err := provider.ec2.DescribeSecurityGroups(input)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err = ensureSingleResult(len(result.SecurityGroups), \"security group\", id); err != nil {\n\t\treturn nil, err\n\t}\n\n\tsecurityGroup := newSecurityGroupFromAPI(result.SecurityGroups[0])\n\treturn &securityGroup, nil\n}", "func (m *Group) SetGroupTypes(value []string)() {\n m.groupTypes = value\n}", "func dataSourceSecurityGroupRead(d *schema.ResourceData, meta interface{}) error {\n config := meta.(*PConfig)\n\n platform := d.Get(\"platform\").(string)\n siteID := d.Get(\"vcs\").(string)\n\n resourcePath := fmt.Sprintf(\"api/v3/%s/sites/%s/\", platform, siteID)\n response, err := config.doNormalRequest(platform, resourcePath, \"GET\", nil)\n if err != nil {\n return fmt.Errorf(\"Unable to get VCS %s: %v\", siteID, err)\n }\n var site map[string]interface{}\n if err = json.Unmarshal([]byte(response), &site); err != nil {\n return err\n }\n projectID := int(site[\"project\"].(float64))\n var serverID int\n for _, server := range site[\"servers\"].([]interface{}) {\n serverInfo := server.(map[string]interface{})\n serverID = int(serverInfo[\"id\"].(float64))\n break\n }\n\n resourcePath = fmt.Sprintf(\"api/v3/%s/security_groups/?project=%d&server=%d\", platform, projectID, serverID)\n response, err = config.doNormalRequest(platform, resourcePath, \"GET\", nil)\n\n if err != nil {\n return fmt.Errorf(\"Unable to list security_groups: %v\", err)\n }\n\n var security_groups []map[string]interface{}\n if err = json.Unmarshal([]byte(response), &security_groups); err != nil {\n return err\n }\n\n for _, security_group := range security_groups {\n return dataSourceSecurityGroupAttributes(d, security_group)\n }\n\n return fmt.Errorf(\"Unable to retrieve security group by VCS %s: %v\", siteID, err)\n}", "func (c *Client) AssociateSecurityGroups(request *AssociateSecurityGroupsRequest) (response *AssociateSecurityGroupsResponse, err error) {\n return c.AssociateSecurityGroupsWithContext(context.Background(), request)\n}", "func (m *PrivilegedAccessGroupEligibilitySchedule) SetPrincipal(value DirectoryObjectable)() {\n err := m.GetBackingStore().Set(\"principal\", value)\n if err != nil {\n panic(err)\n }\n}", "func (_BaseAccessControlGroup *BaseAccessControlGroupTransactor) SetContractRights(opts *bind.TransactOpts, obj common.Address, access_type uint8, access uint8) (*types.Transaction, error) {\n\treturn _BaseAccessControlGroup.contract.Transact(opts, \"setContractRights\", obj, access_type, access)\n}", "func ExampleDeviceSecurityGroupsClient_CreateOrUpdate() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclient, err := armsecurity.NewDeviceSecurityGroupsClient(cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tres, err := client.CreateOrUpdate(ctx,\n\t\t\"subscriptions/20ff7fc3-e762-44dd-bd96-b71116dcdc23/resourceGroups/SampleRG/providers/Microsoft.Devices/iotHubs/sampleiothub\",\n\t\t\"samplesecuritygroup\",\n\t\tarmsecurity.DeviceSecurityGroup{\n\t\t\tProperties: &armsecurity.DeviceSecurityGroupProperties{\n\t\t\t\tTimeWindowRules: []armsecurity.TimeWindowCustomAlertRuleClassification{\n\t\t\t\t\t&armsecurity.ActiveConnectionsNotInAllowedRange{\n\t\t\t\t\t\tIsEnabled: to.Ptr(true),\n\t\t\t\t\t\tRuleType: to.Ptr(\"ActiveConnectionsNotInAllowedRange\"),\n\t\t\t\t\t\tMaxThreshold: to.Ptr[int32](30),\n\t\t\t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t\t\t\t\t\tTimeWindowSize: to.Ptr(\"PT05M\"),\n\t\t\t\t\t}},\n\t\t\t},\n\t\t},\n\t\tnil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\t// TODO: use response item\n\t_ = res\n}", "func (m *SecureScoreControlProfile) SetAzureTenantId(value *string)() {\n m.azureTenantId = value\n}", "func (m *ProtectGroup) SetPrivacy(value *GroupPrivacy)() {\n err := m.GetBackingStore().Set(\"privacy\", value)\n if err != nil {\n panic(err)\n }\n}", "func (r *role) SetStages(st Stages) {\n\tif len(st) != r.numStages {\n\t\tr.log.Panic(\"number of stages don't match\")\n\t}\n\n\tr.stages = st\n\tfor i := range r.stages {\n\t\tr.stages[i].Add(1)\n\t}\n}", "func (mtr *Msmsintprp5Metrics) SetSecurity(val metrics.Counter) error {\n\tmtr.metrics.SetCounter(val, mtr.getOffset(\"Security\"))\n\treturn nil\n}", "func (c *ClientImpl) AssignGroupsToProject(ctx context.Context, hcpHostURL, epicProjectID, projectID string) error {\n\tspan, _ := opentracing.StartSpanFromContext(ctx, \"Assign Groups to Project\")\n\tdefer span.Finish()\n\n\tsession, err := c.getSession(ctx, hcpHostURL, hcpUserName, hcpPassword)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tstatus = Failure\n\tmonitor := metrics.StartExternalCall(externalSvcName, \"Update External Groups in HCP\")\n\tdefer func() { monitor.RecordWithStatus(status) }()\n\t// get existing roles\n\texistingRoles, err := c.ListRoles(ctx, hcpHostURL)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"AssignGroupsToProject: Failed to get available roles: %v\", err)\n\t}\n\n\tvar adminRoleHref string\n\tvar memberRoleHref string\n\n\tfor _, eachRole := range existingRoles {\n\t\tif eachRole.Label.Name == adminRole {\n\t\t\tadminRoleHref = eachRole.Links.Self.HRef\n\t\t} else if eachRole.Label.Name == memberRole {\n\t\t\tmemberRoleHref = eachRole.Links.Self.HRef\n\t\t}\n\t}\n\n\tif adminRoleHref == \"\" || memberRoleHref == \"\" {\n\t\treturn fmt.Errorf(\"AssignGroupsToProject: Failed to get admin and or member role(s) href\")\n\t}\n\n\targs := hcpModels.ExternalUserGroupListResource{\n\t\tExternalUserGroups: []hcpModels.ExternalUserGroupResource{\n\t\t\t{\n\t\t\t\tRole: adminRoleHref,\n\t\t\t\tGroup: projectOwnerClaimValue + \"-\" + projectID,\n\t\t\t},\n\t\t\t{\n\t\t\t\tRole: memberRoleHref,\n\t\t\t\tGroup: projectMemberClaimValue + \"-\" + projectID,\n\t\t\t},\n\t\t},\n\t}\n\n\trequestBody, _ := json.Marshal(args)\n\tresp, err := mlopsHttp.ExecuteHTTPRequest(\n\t\tctx,\n\t\tc.client,\n\t\thcpHostURL+projectPathV1+epicProjectID+externalUserGroupsFlag,\n\t\thttp.MethodPut,\n\t\tmap[string]string{sessionHeader: session},\n\t\tbytes.NewReader(requestBody),\n\t)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"AssignGroupsToProject: Failed to Add Groups in HCP: %v\", err)\n\t}\n\n\tstatus = Success\n\n\terr = c.deleteSession(ctx, hcpHostURL, session)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif code := resp.StatusCode; code != http.StatusNoContent {\n\t\treturn fmt.Errorf(\"AssignGroupsToProject: incorrect status code for assign groups to project %v: %d\", epicProjectID, resp.StatusCode)\n\t}\n\n\treturn nil\n}", "func (r *LaunchConfiguration) SecurityGroups() pulumi.ArrayOutput {\n\treturn (pulumi.ArrayOutput)(r.s.State[\"securityGroups\"])\n}", "func SetTenant(tenant string) func(*AviSession) error {\n\treturn func(sess *AviSession) error {\n\t\treturn sess.setTenant(tenant)\n\t}\n}", "func (m *User) SetOauth2PermissionGrants(value []OAuth2PermissionGrantable)() {\n m.oauth2PermissionGrants = value\n}", "func (s *DescribeMountTargetSecurityGroupsOutput) SetSecurityGroups(v []*string) *DescribeMountTargetSecurityGroupsOutput {\n\ts.SecurityGroups = v\n\treturn s\n}", "func (mtr *Msmsintprp4Metrics) SetSecurity(val metrics.Counter) error {\n\tmtr.metrics.SetCounter(val, mtr.getOffset(\"Security\"))\n\treturn nil\n}", "func (m *DeviceManagementConfigurationPolicy) SetRoleScopeTagIds(value []string)() {\n err := m.GetBackingStore().Set(\"roleScopeTagIds\", value)\n if err != nil {\n panic(err)\n }\n}", "func (mtr *Msmsintprp3Metrics) SetSecurity(val metrics.Counter) error {\n\tmtr.metrics.SetCounter(val, mtr.getOffset(\"Security\"))\n\treturn nil\n}", "func (s *ModifyMountTargetSecurityGroupsInput) SetSecurityGroups(v []*string) *ModifyMountTargetSecurityGroupsInput {\n\ts.SecurityGroups = v\n\treturn s\n}", "func ExampleDevicesClient_BeginCreateOrUpdateSecuritySettings() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclientFactory, err := armstorsimple1200series.NewClientFactory(\"<subscription-id>\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tpoller, err := clientFactory.NewDevicesClient().BeginCreateOrUpdateSecuritySettings(ctx, \"HSDK-T4ZA3EAJFR\", \"ResourceGroupForSDKTest\", \"hAzureSDKOperations\", armstorsimple1200series.SecuritySettings{\n\t\tProperties: &armstorsimple1200series.SecuritySettingsProperties{\n\t\t\tDeviceAdminPassword: &armstorsimple1200series.AsymmetricEncryptedSecret{\n\t\t\t\tEncryptionAlgorithm: to.Ptr(armstorsimple1200series.EncryptionAlgorithmRSAESPKCS1V15),\n\t\t\t\tEncryptionCertificateThumbprint: to.Ptr(\"D73DB57C4CDD6761E159F8D1E8A7D759424983FD\"),\n\t\t\t\tValue: to.Ptr(\"<value>\"),\n\t\t\t},\n\t\t},\n\t}, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\t_, err = poller.PollUntilDone(ctx, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to pull the result: %v\", err)\n\t}\n}", "func (s *DefaultSpaceSettings) SetSecurityGroups(v []*string) *DefaultSpaceSettings {\n\ts.SecurityGroups = v\n\treturn s\n}", "func (s *ServerConnection) UserGroupsSet(groupIds StringList, details UserGroup, domainId KId) (ErrorList, error) {\n\tparams := struct {\n\t\tGroupIds StringList `json:\"groupIds\"`\n\t\tDetails UserGroup `json:\"details\"`\n\t\tDomainId KId `json:\"domainId\"`\n\t}{groupIds, details, domainId}\n\tdata, err := s.CallRaw(\"UserGroups.set\", params)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\terrors := struct {\n\t\tResult struct {\n\t\t\tErrors ErrorList `json:\"errors\"`\n\t\t} `json:\"result\"`\n\t}{}\n\terr = json.Unmarshal(data, &errors)\n\treturn errors.Result.Errors, err\n}", "func (as *AdminServer) Groups(w http.ResponseWriter, r *http.Request) {\n\tparams := newTemplateParams(r)\n\tparams.Title = \"Users & Groups\"\n\tgetTemplate(w, \"groups\").ExecuteTemplate(w, \"base\", params)\n}", "func SetSecurityObject(path string, flags SECURITY_INFORMATION, sd []byte) error {\n\tvar xattrKey string\n\n\tif len(sd) < int(unsafe.Sizeof(SECURITY_DESCRIPTOR_RELATIVE{})) {\n\t\tpanic(fmt.Errorf(\"SetSecurityObject: sd too small (%d bytes)\", len(sd)))\n\t}\n\n\t// Pick the right xattr key that allows us to pass the needed information to the cifs client.\n\tif flags == DACL_SECURITY_INFORMATION {\n\t\t// Only DACL.\n\t\txattrKey = common.CIFS_XATTR_CIFS_ACL\n\n\t\t// sd.OffsetOwner = 0.\n\t\tbinary.LittleEndian.PutUint32(sd[4:8], 0)\n\t\t// sd.OffsetGroup = 0.\n\t\tbinary.LittleEndian.PutUint32(sd[8:12], 0)\n\t\t// sd.OffsetSacl = 0.\n\t\tbinary.LittleEndian.PutUint32(sd[12:16], 0)\n\t} else if flags == (DACL_SECURITY_INFORMATION | OWNER_SECURITY_INFORMATION | GROUP_SECURITY_INFORMATION) {\n\t\t// DACL + Owner + Group.\n\t\txattrKey = common.CIFS_XATTR_CIFS_NTSD\n\n\t\t// sd.OffsetSacl = 0.\n\t\tbinary.LittleEndian.PutUint32(sd[12:16], 0)\n\t} else if flags == (DACL_SECURITY_INFORMATION | SACL_SECURITY_INFORMATION |\n\t\tOWNER_SECURITY_INFORMATION | GROUP_SECURITY_INFORMATION) {\n\t\t// DACL + SACL + Owner + Group.\n\t\txattrKey = common.CIFS_XATTR_CIFS_NTSD_FULL\n\n\t\t// Put in the end to prevent \"unreachable code\" complaints from vet.\n\t\t// TODO: Add support for \"DACL + SACL + Owner + Group\".\n\t\t// Remove this panic only after rest of the code correctly supports SACL.\n\t\tpanic(fmt.Errorf(\"SetSecurityObject: Unsupported flags value 0x%x\", flags))\n\n\t} else {\n\t\tpanic(fmt.Errorf(\"SetSecurityObject: Unsupported flags value 0x%x\", flags))\n\t}\n\n\t// Ensure Security Descriptor is valid before writing to the cifs client.\n\tif err := sdRelativeIsValid(sd, flags); err != nil {\n\t\tpanic(fmt.Errorf(\"SetSecurityObject: %v\", err))\n\t}\n\n\terr := xattr.Set(path, xattrKey, sd)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"SetSecurityObject: xattr.Set(%s) failed for file %s: %v\", xattrKey, path, err)\n\t}\n\n\treturn nil\n}", "func (m *User) SetSecurityIdentifier(value *string)() {\n m.securityIdentifier = value\n}", "func ExampleDeviceSecurityGroupsClient_Get() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclientFactory, err := armsecurity.NewClientFactory(\"<subscription-id>\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tres, err := clientFactory.NewDeviceSecurityGroupsClient().Get(ctx, \"subscriptions/20ff7fc3-e762-44dd-bd96-b71116dcdc23/resourceGroups/SampleRG/providers/Microsoft.Devices/iotHubs/sampleiothub\", \"samplesecuritygroup\", nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\t// You could use response here. We use blank identifier for just demo purposes.\n\t_ = res\n\t// If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes.\n\t// res.DeviceSecurityGroup = armsecurity.DeviceSecurityGroup{\n\t// \tName: to.Ptr(\"samplesecuritygroup\"),\n\t// \tType: to.Ptr(\"Microsoft.Security/deviceSecurityGroups\"),\n\t// \tID: to.Ptr(\"/subscriptions/20ff7fc3-e762-44dd-bd96-b71116dcdc23/resourceGroups/SampleRG/providers/Microsoft.Devices/iotHubs/sampleiothub/providers/Microsoft.Security/deviceSecurityGroups/samplesecuritygroup\"),\n\t// \tProperties: &armsecurity.DeviceSecurityGroupProperties{\n\t// \t\tAllowlistRules: []armsecurity.AllowlistCustomAlertRuleClassification{\n\t// \t\t\t&armsecurity.ConnectionToIPNotAllowed{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when an outbound connection is created between your device and an ip that isn't allowed\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Outbound connection to an ip that isn't allowed\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"ConnectionToIpNotAllowed\"),\n\t// \t\t\t\tValueType: to.Ptr(armsecurity.ValueTypeIPCidr),\n\t// \t\t\t\tAllowlistValues: []*string{\n\t// \t\t\t\t},\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.LocalUserNotAllowed{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when a local user that isn't allowed logins to the device\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Login by a local user that isn't allowed\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"LocalUserNotAllowed\"),\n\t// \t\t\t\tValueType: to.Ptr(armsecurity.ValueTypeString),\n\t// \t\t\t\tAllowlistValues: []*string{\n\t// \t\t\t\t},\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.ProcessNotAllowed{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when a process that isn't allowed is executed\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Execution of a process that isn't allowed\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"ProcessNotAllowed\"),\n\t// \t\t\t\tValueType: to.Ptr(armsecurity.ValueTypeString),\n\t// \t\t\t\tAllowlistValues: []*string{\n\t// \t\t\t\t},\n\t// \t\t}},\n\t// \t\tDenylistRules: []*armsecurity.DenylistCustomAlertRule{\n\t// \t\t},\n\t// \t\tThresholdRules: []armsecurity.ThresholdCustomAlertRuleClassification{\n\t// \t\t},\n\t// \t\tTimeWindowRules: []armsecurity.TimeWindowCustomAlertRuleClassification{\n\t// \t\t\t&armsecurity.ActiveConnectionsNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of active connections of a device in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of active connections is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"ActiveConnectionsNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.AmqpC2DMessagesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of cloud to device messages (AMQP protocol) in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of cloud to device messages (AMQP protocol) is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"AmqpC2DMessagesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.MqttC2DMessagesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of cloud to device messages (MQTT protocol) in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of cloud to device messages (MQTT protocol) is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"MqttC2DMessagesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.HTTPC2DMessagesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of cloud to device messages (HTTP protocol) in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of cloud to device messages (HTTP protocol) is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"HttpC2DMessagesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.AmqpC2DRejectedMessagesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of cloud to device messages (AMQP protocol) that were rejected by the device in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of rejected cloud to device messages (AMQP protocol) is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"AmqpC2DRejectedMessagesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.MqttC2DRejectedMessagesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of cloud to device messages (MQTT protocol) that were rejected by the device in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of rejected cloud to device messages (MQTT protocol) is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"MqttC2DRejectedMessagesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.HTTPC2DRejectedMessagesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of cloud to device messages (HTTP protocol) that were rejected by the device in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of rejected cloud to device messages (HTTP protocol) is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"HttpC2DRejectedMessagesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.AmqpD2CMessagesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of device to cloud messages (AMQP protocol) in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of device to cloud messages (AMQP protocol) is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"AmqpD2CMessagesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.MqttD2CMessagesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of device to cloud messages (MQTT protocol) in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of device to cloud messages (MQTT protocol) is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"MqttD2CMessagesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.HTTPD2CMessagesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of device to cloud messages (HTTP protocol) in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of device to cloud messages (HTTP protocol) is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"HttpD2CMessagesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.DirectMethodInvokesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of direct method invokes in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of direct method invokes is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"DirectMethodInvokesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.FailedLocalLoginsNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of failed local logins on the device in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of failed local logins is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"FailedLocalLoginsNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.FileUploadsNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of file uploads from the device to the cloud in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of file uploads is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"FileUploadsNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.QueuePurgesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of device queue purges in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of device queue purges is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"QueuePurgesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.TwinUpdatesNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number of twin updates (by the device or the service) in the time window is not in the allowed range\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of twin updates is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"TwinUpdatesNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t\t},\n\t// \t\t\t&armsecurity.UnauthorizedOperationsNotInAllowedRange{\n\t// \t\t\t\tDescription: to.Ptr(\"Get an alert when the number unauthorized operations in the time window is not in the allowed range. Unauthorized operations are operations that affect the device (or done by it) that fail because of an unauthorized error\"),\n\t// \t\t\t\tDisplayName: to.Ptr(\"Number of unauthorized operations is not in allowed range\"),\n\t// \t\t\t\tIsEnabled: to.Ptr(false),\n\t// \t\t\t\tRuleType: to.Ptr(\"UnauthorizedOperationsNotInAllowedRange\"),\n\t// \t\t\t\tMaxThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tMinThreshold: to.Ptr[int32](0),\n\t// \t\t\t\tTimeWindowSize: to.Ptr(\"PT15M\"),\n\t// \t\t}},\n\t// \t},\n\t// }\n}", "func (l *Libvirt) DomainSetVcpusFlags(Dom Domain, Nvcpus uint32, Flags uint32) (err error) {\n\tvar buf []byte\n\n\targs := DomainSetVcpusFlagsArgs {\n\t\tDom: Dom,\n\t\tNvcpus: Nvcpus,\n\t\tFlags: Flags,\n\t}\n\n\tbuf, err = encode(&args)\n\tif err != nil {\n\t\treturn\n\t}\n\n\n\t_, err = l.requestStream(199, constants.Program, buf, nil, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func SetPrivs(name string, privs map[string]bool) error {\n\tdb, err := authDB()\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer db.Close()\n\n\t_, err = db.Exec(`INSERT INTO privileges (\n\tname,\n\tprivileges\n) VALUES (\n\t$1,\n\t$2\n);`, name, encodePrivs(privs))\n\t_, err = db.Exec(`UPDATE privileges SET privileges = $1 WHERE name = $2;`, encodePrivs(privs), name)\n\n\treturn err\n}", "func (m *User) SetAgeGroup(value *string)() {\n m.ageGroup = value\n}", "func (m *GraphBaseServiceClient) GroupSettings()(*i4794c103c0d044c27a3ca3af0a0e498e93a9863420c1a4e7a29ef37590053c7b.GroupSettingsRequestBuilder) {\n return i4794c103c0d044c27a3ca3af0a0e498e93a9863420c1a4e7a29ef37590053c7b.NewGroupSettingsRequestBuilderInternal(m.pathParameters, m.requestAdapter);\n}", "func (m *GraphBaseServiceClient) GroupSettings()(*i4794c103c0d044c27a3ca3af0a0e498e93a9863420c1a4e7a29ef37590053c7b.GroupSettingsRequestBuilder) {\n return i4794c103c0d044c27a3ca3af0a0e498e93a9863420c1a4e7a29ef37590053c7b.NewGroupSettingsRequestBuilderInternal(m.pathParameters, m.requestAdapter);\n}", "func (l *Libvirt) DomainSetSchedulerParametersFlags(Dom Domain, Params []TypedParam, Flags uint32) (err error) {\n\tvar buf []byte\n\n\targs := DomainSetSchedulerParametersFlagsArgs {\n\t\tDom: Dom,\n\t\tParams: Params,\n\t\tFlags: Flags,\n\t}\n\n\tbuf, err = encode(&args)\n\tif err != nil {\n\t\treturn\n\t}\n\n\n\t_, err = l.requestStream(219, constants.Program, buf, nil, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func (m *CalendarGroup) SetCalendars(value []Calendarable)() {\n err := m.GetBackingStore().Set(\"calendars\", value)\n if err != nil {\n panic(err)\n }\n}", "func setupDBAndSettingsForAccessControlQuotaTests(t *testing.T, sc accessControlScenarioContext) {\n\tt.Helper()\n\n\t// Create two orgs with the context user\n\tsetupOrgsDBForAccessControlTests(t, sc.db, sc, 2)\n}", "func (c *Dg) SetDeviceVsys(g interface{}, d string, vsys []string) error {\n var name string\n\n switch v := g.(type) {\n case string:\n name = v\n case Entry:\n name = v.Name\n default:\n return fmt.Errorf(\"Unknown type sent to add devices: %s\", v)\n }\n\n c.con.LogAction(\"(set) device vsys in device group: %s\", name)\n\n m := util.MapToVsysEnt(map[string] []string{d: vsys})\n path := c.xpath([]string{name})\n path = append(path, \"devices\")\n\n _, err := c.con.Set(path, m.Entries[0], nil, nil)\n return err\n}", "func (m *Application) SetGroupMembershipClaims(value *string)() {\n m.groupMembershipClaims = value\n}", "func (m *DeviceAndAppManagementAssignmentFilter) SetRoleScopeTags(value []string)() {\n err := m.GetBackingStore().Set(\"roleScopeTags\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *VirtualEndpoint) SetServicePlans(value []CloudPcServicePlanable)() {\n err := m.GetBackingStore().Set(\"servicePlans\", value)\n if err != nil {\n panic(err)\n }\n}" ]
[ "0.6836787", "0.61099243", "0.59089315", "0.5522794", "0.5427237", "0.53266984", "0.5326117", "0.5253257", "0.5251793", "0.522787", "0.52200997", "0.5197786", "0.51056135", "0.5072223", "0.5068338", "0.5032946", "0.49538523", "0.49053612", "0.48882514", "0.48497662", "0.4848123", "0.48458633", "0.48438546", "0.48412272", "0.47937453", "0.47793344", "0.47510275", "0.47433728", "0.47420642", "0.47210371", "0.47048783", "0.4699269", "0.4694391", "0.46620023", "0.46617436", "0.46563417", "0.46557", "0.46400347", "0.4637789", "0.4629765", "0.46282494", "0.4624858", "0.46098295", "0.4592342", "0.4583291", "0.45817366", "0.4581329", "0.4570001", "0.45604354", "0.45381537", "0.45259258", "0.45115536", "0.44992042", "0.44986814", "0.44983384", "0.44967234", "0.4482223", "0.448023", "0.44782567", "0.4464579", "0.44629344", "0.44574782", "0.44500777", "0.44426787", "0.4440734", "0.44383773", "0.44349664", "0.44348192", "0.44346553", "0.44250503", "0.44132373", "0.44072846", "0.44070145", "0.44032177", "0.4398366", "0.43968847", "0.4392741", "0.43822497", "0.43721554", "0.43497354", "0.4349511", "0.43390262", "0.43361998", "0.4334964", "0.4333356", "0.4332466", "0.43287247", "0.4328563", "0.43277508", "0.43235686", "0.4320965", "0.43140006", "0.43140006", "0.43049768", "0.43033668", "0.42954284", "0.42951015", "0.42905685", "0.42890498", "0.4289034" ]
0.70758635
0
WriteNodeConfig writes the local node configuration to the specified writer.
func (h *HeaderfileWriter) WriteNodeConfig(w io.Writer, cfg *datapath.LocalNodeConfiguration) error { extraMacrosMap := make(dpdef.Map) cDefinesMap := make(dpdef.Map) fw := bufio.NewWriter(w) writeIncludes(w) routerIP := node.GetIPv6Router() hostIP := node.GetIPv6() fmt.Fprintf(fw, "/*\n") if option.Config.EnableIPv6 { fmt.Fprintf(fw, " cilium.v6.external.str %s\n", node.GetIPv6().String()) fmt.Fprintf(fw, " cilium.v6.internal.str %s\n", node.GetIPv6Router().String()) fmt.Fprintf(fw, " cilium.v6.nodeport.str %s\n", node.GetNodePortIPv6Addrs()) fmt.Fprintf(fw, "\n") } fmt.Fprintf(fw, " cilium.v4.external.str %s\n", node.GetIPv4().String()) fmt.Fprintf(fw, " cilium.v4.internal.str %s\n", node.GetInternalIPv4Router().String()) fmt.Fprintf(fw, " cilium.v4.nodeport.str %s\n", node.GetNodePortIPv4Addrs()) fmt.Fprintf(fw, "\n") if option.Config.EnableIPv6 { fw.WriteString(dumpRaw(defaults.RestoreV6Addr, node.GetIPv6Router())) } fw.WriteString(dumpRaw(defaults.RestoreV4Addr, node.GetInternalIPv4Router())) fmt.Fprintf(fw, " */\n\n") cDefinesMap["KERNEL_HZ"] = fmt.Sprintf("%d", option.Config.KernelHz) if option.Config.EnableIPv6 { extraMacrosMap["ROUTER_IP"] = routerIP.String() fw.WriteString(defineIPv6("ROUTER_IP", routerIP)) } if option.Config.EnableIPv4 { ipv4GW := node.GetInternalIPv4Router() loopbackIPv4 := node.GetIPv4Loopback() ipv4Range := node.GetIPv4AllocRange() cDefinesMap["IPV4_GATEWAY"] = fmt.Sprintf("%#x", byteorder.NetIPv4ToHost32(ipv4GW)) cDefinesMap["IPV4_LOOPBACK"] = fmt.Sprintf("%#x", byteorder.NetIPv4ToHost32(loopbackIPv4)) cDefinesMap["IPV4_MASK"] = fmt.Sprintf("%#x", byteorder.NetIPv4ToHost32(net.IP(ipv4Range.Mask))) if option.Config.EnableIPv4FragmentsTracking { cDefinesMap["ENABLE_IPV4_FRAGMENTS"] = "1" cDefinesMap["IPV4_FRAG_DATAGRAMS_MAP"] = fragmap.MapName cDefinesMap["CILIUM_IPV4_FRAG_MAP_MAX_ENTRIES"] = fmt.Sprintf("%d", option.Config.FragmentsMapEntries) } } if option.Config.EnableIPv6 { extraMacrosMap["HOST_IP"] = hostIP.String() fw.WriteString(defineIPv6("HOST_IP", hostIP)) } for t, id := range tunnelProtocols { macroName := fmt.Sprintf("TUNNEL_PROTOCOL_%s", strings.ToUpper(t)) cDefinesMap[macroName] = fmt.Sprintf("%d", id) } encapProto := option.Config.TunnelProtocol if !option.Config.TunnelingEnabled() && option.Config.EnableNodePort && option.Config.NodePortMode != option.NodePortModeSNAT && option.Config.LoadBalancerDSRDispatch == option.DSRDispatchGeneve { encapProto = option.TunnelGeneve } cDefinesMap["TUNNEL_PROTOCOL"] = fmt.Sprintf("%d", tunnelProtocols[encapProto]) cDefinesMap["TUNNEL_PORT"] = fmt.Sprintf("%d", option.Config.TunnelPort) if tunnelDev, err := netlink.LinkByName(fmt.Sprintf("cilium_%s", encapProto)); err == nil { cDefinesMap["ENCAP_IFINDEX"] = fmt.Sprintf("%d", tunnelDev.Attrs().Index) } cDefinesMap["HOST_ID"] = fmt.Sprintf("%d", identity.GetReservedID(labels.IDNameHost)) cDefinesMap["WORLD_ID"] = fmt.Sprintf("%d", identity.GetReservedID(labels.IDNameWorld)) if option.Config.IsDualStack() { cDefinesMap["WORLD_IPV4_ID"] = fmt.Sprintf("%d", identity.GetReservedID(labels.IDNameWorldIPv4)) cDefinesMap["WORLD_IPV6_ID"] = fmt.Sprintf("%d", identity.GetReservedID(labels.IDNameWorldIPv6)) } else { worldID := identity.GetReservedID(labels.IDNameWorld) cDefinesMap["WORLD_IPV4_ID"] = fmt.Sprintf("%d", worldID) cDefinesMap["WORLD_IPV6_ID"] = fmt.Sprintf("%d", worldID) } cDefinesMap["HEALTH_ID"] = fmt.Sprintf("%d", identity.GetReservedID(labels.IDNameHealth)) cDefinesMap["UNMANAGED_ID"] = fmt.Sprintf("%d", identity.GetReservedID(labels.IDNameUnmanaged)) cDefinesMap["INIT_ID"] = fmt.Sprintf("%d", identity.GetReservedID(labels.IDNameInit)) cDefinesMap["LOCAL_NODE_ID"] = fmt.Sprintf("%d", identity.GetLocalNodeID()) cDefinesMap["REMOTE_NODE_ID"] = fmt.Sprintf("%d", identity.GetReservedID(labels.IDNameRemoteNode)) cDefinesMap["KUBE_APISERVER_NODE_ID"] = fmt.Sprintf("%d", identity.GetReservedID(labels.IDNameKubeAPIServer)) cDefinesMap["CILIUM_LB_SERVICE_MAP_MAX_ENTRIES"] = fmt.Sprintf("%d", lbmap.ServiceMapMaxEntries) cDefinesMap["CILIUM_LB_BACKENDS_MAP_MAX_ENTRIES"] = fmt.Sprintf("%d", lbmap.ServiceBackEndMapMaxEntries) cDefinesMap["CILIUM_LB_REV_NAT_MAP_MAX_ENTRIES"] = fmt.Sprintf("%d", lbmap.RevNatMapMaxEntries) cDefinesMap["CILIUM_LB_AFFINITY_MAP_MAX_ENTRIES"] = fmt.Sprintf("%d", lbmap.AffinityMapMaxEntries) cDefinesMap["CILIUM_LB_SOURCE_RANGE_MAP_MAX_ENTRIES"] = fmt.Sprintf("%d", lbmap.SourceRangeMapMaxEntries) cDefinesMap["CILIUM_LB_MAGLEV_MAP_MAX_ENTRIES"] = fmt.Sprintf("%d", lbmap.MaglevMapMaxEntries) cDefinesMap["TUNNEL_MAP"] = tunnel.MapName cDefinesMap["TUNNEL_ENDPOINT_MAP_SIZE"] = fmt.Sprintf("%d", tunnel.MaxEntries) cDefinesMap["ENDPOINTS_MAP"] = lxcmap.MapName cDefinesMap["ENDPOINTS_MAP_SIZE"] = fmt.Sprintf("%d", lxcmap.MaxEntries) cDefinesMap["METRICS_MAP"] = metricsmap.MapName cDefinesMap["METRICS_MAP_SIZE"] = fmt.Sprintf("%d", metricsmap.MaxEntries) cDefinesMap["POLICY_MAP_SIZE"] = fmt.Sprintf("%d", policymap.MaxEntries) cDefinesMap["AUTH_MAP"] = authmap.MapName cDefinesMap["AUTH_MAP_SIZE"] = fmt.Sprintf("%d", option.Config.AuthMapEntries) cDefinesMap["CONFIG_MAP"] = configmap.MapName cDefinesMap["CONFIG_MAP_SIZE"] = fmt.Sprintf("%d", configmap.MaxEntries) cDefinesMap["IPCACHE_MAP"] = ipcachemap.Name cDefinesMap["IPCACHE_MAP_SIZE"] = fmt.Sprintf("%d", ipcachemap.MaxEntries) cDefinesMap["NODE_MAP"] = nodemap.MapName cDefinesMap["NODE_MAP_SIZE"] = fmt.Sprintf("%d", nodemap.MaxEntries) cDefinesMap["SRV6_VRF_MAP4"] = srv6map.VRFMapName4 cDefinesMap["SRV6_VRF_MAP6"] = srv6map.VRFMapName6 cDefinesMap["SRV6_POLICY_MAP4"] = srv6map.PolicyMapName4 cDefinesMap["SRV6_POLICY_MAP6"] = srv6map.PolicyMapName6 cDefinesMap["SRV6_SID_MAP"] = srv6map.SIDMapName cDefinesMap["SRV6_STATE_MAP4"] = srv6map.StateMapName4 cDefinesMap["SRV6_STATE_MAP6"] = srv6map.StateMapName6 cDefinesMap["SRV6_VRF_MAP_SIZE"] = fmt.Sprintf("%d", srv6map.MaxVRFEntries) cDefinesMap["SRV6_POLICY_MAP_SIZE"] = fmt.Sprintf("%d", srv6map.MaxPolicyEntries) cDefinesMap["SRV6_SID_MAP_SIZE"] = fmt.Sprintf("%d", srv6map.MaxSIDEntries) cDefinesMap["SRV6_STATE_MAP_SIZE"] = fmt.Sprintf("%d", srv6map.MaxStateEntries) cDefinesMap["WORLD_CIDRS4_MAP"] = worldcidrsmap.MapName4 cDefinesMap["WORLD_CIDRS4_MAP_SIZE"] = fmt.Sprintf("%d", worldcidrsmap.MapMaxEntries) cDefinesMap["POLICY_PROG_MAP_SIZE"] = fmt.Sprintf("%d", policymap.PolicyCallMaxEntries) cDefinesMap["L2_RESPONSER_MAP4_SIZE"] = fmt.Sprintf("%d", l2respondermap.DefaultMaxEntries) cDefinesMap["ENCRYPT_MAP"] = encrypt.MapName cDefinesMap["L2_RESPONDER_MAP4"] = l2respondermap.MapName cDefinesMap["CT_CONNECTION_LIFETIME_TCP"] = fmt.Sprintf("%d", int64(option.Config.CTMapEntriesTimeoutTCP.Seconds())) cDefinesMap["CT_CONNECTION_LIFETIME_NONTCP"] = fmt.Sprintf("%d", int64(option.Config.CTMapEntriesTimeoutAny.Seconds())) cDefinesMap["CT_SERVICE_LIFETIME_TCP"] = fmt.Sprintf("%d", int64(option.Config.CTMapEntriesTimeoutSVCTCP.Seconds())) cDefinesMap["CT_SERVICE_LIFETIME_NONTCP"] = fmt.Sprintf("%d", int64(option.Config.CTMapEntriesTimeoutSVCAny.Seconds())) cDefinesMap["CT_SERVICE_CLOSE_REBALANCE"] = fmt.Sprintf("%d", int64(option.Config.CTMapEntriesTimeoutSVCTCPGrace.Seconds())) cDefinesMap["CT_SYN_TIMEOUT"] = fmt.Sprintf("%d", int64(option.Config.CTMapEntriesTimeoutSYN.Seconds())) cDefinesMap["CT_CLOSE_TIMEOUT"] = fmt.Sprintf("%d", int64(option.Config.CTMapEntriesTimeoutFIN.Seconds())) cDefinesMap["CT_REPORT_INTERVAL"] = fmt.Sprintf("%d", int64(option.Config.MonitorAggregationInterval.Seconds())) cDefinesMap["CT_REPORT_FLAGS"] = fmt.Sprintf("%#04x", int64(option.Config.MonitorAggregationFlags)) cDefinesMap["CT_TAIL_CALL_BUFFER4"] = "cilium_tail_call_buffer4" cDefinesMap["CT_TAIL_CALL_BUFFER6"] = "cilium_tail_call_buffer6" cDefinesMap["PER_CLUSTER_CT_TCP4"] = "cilium_per_cluster_ct_tcp4" cDefinesMap["PER_CLUSTER_CT_TCP6"] = "cilium_per_cluster_ct_tcp6" cDefinesMap["PER_CLUSTER_CT_ANY4"] = "cilium_per_cluster_ct_any4" cDefinesMap["PER_CLUSTER_CT_ANY6"] = "cilium_per_cluster_ct_any6" cDefinesMap["PER_CLUSTER_SNAT_MAPPING_IPV4"] = "cilium_per_cluster_snat_v4_external" cDefinesMap["PER_CLUSTER_SNAT_MAPPING_IPV6"] = "cilium_per_cluster_snat_v6_external" if option.Config.PreAllocateMaps { cDefinesMap["PREALLOCATE_MAPS"] = "1" } cDefinesMap["EVENTS_MAP"] = eventsmap.MapName cDefinesMap["SIGNAL_MAP"] = signalmap.MapName cDefinesMap["POLICY_CALL_MAP"] = policymap.PolicyCallMapName if option.Config.EnableEnvoyConfig { cDefinesMap["POLICY_EGRESSCALL_MAP"] = policymap.PolicyEgressCallMapName } cDefinesMap["LB6_REVERSE_NAT_MAP"] = "cilium_lb6_reverse_nat" cDefinesMap["LB6_SERVICES_MAP_V2"] = "cilium_lb6_services_v2" cDefinesMap["LB6_BACKEND_MAP"] = "cilium_lb6_backends_v3" cDefinesMap["LB6_REVERSE_NAT_SK_MAP"] = lbmap.SockRevNat6MapName cDefinesMap["LB6_REVERSE_NAT_SK_MAP_SIZE"] = fmt.Sprintf("%d", lbmap.MaxSockRevNat6MapEntries) cDefinesMap["LB4_REVERSE_NAT_MAP"] = "cilium_lb4_reverse_nat" cDefinesMap["LB4_SERVICES_MAP_V2"] = "cilium_lb4_services_v2" cDefinesMap["LB4_BACKEND_MAP"] = "cilium_lb4_backends_v3" cDefinesMap["LB4_REVERSE_NAT_SK_MAP"] = lbmap.SockRevNat4MapName cDefinesMap["LB4_REVERSE_NAT_SK_MAP_SIZE"] = fmt.Sprintf("%d", lbmap.MaxSockRevNat4MapEntries) if option.Config.EnableSessionAffinity { cDefinesMap["ENABLE_SESSION_AFFINITY"] = "1" cDefinesMap["LB_AFFINITY_MATCH_MAP"] = lbmap.AffinityMatchMapName if option.Config.EnableIPv4 { cDefinesMap["LB4_AFFINITY_MAP"] = lbmap.Affinity4MapName } if option.Config.EnableIPv6 { cDefinesMap["LB6_AFFINITY_MAP"] = lbmap.Affinity6MapName } } cDefinesMap["TRACE_PAYLOAD_LEN"] = fmt.Sprintf("%dULL", option.Config.TracePayloadlen) cDefinesMap["MTU"] = fmt.Sprintf("%d", cfg.MtuConfig.GetDeviceMTU()) if option.Config.EnableIPv4 { cDefinesMap["ENABLE_IPV4"] = "1" } if option.Config.EnableIPv6 { cDefinesMap["ENABLE_IPV6"] = "1" } if option.Config.EnableSRv6 { cDefinesMap["ENABLE_SRV6"] = "1" if option.Config.SRv6EncapMode != "reduced" { cDefinesMap["ENABLE_SRV6_SRH_ENCAP"] = "1" } } if option.Config.EnableSCTP { cDefinesMap["ENABLE_SCTP"] = "1" } if option.Config.EnableIPSec { cDefinesMap["ENABLE_IPSEC"] = "1" } if option.Config.EnableWireguard { cDefinesMap["ENABLE_WIREGUARD"] = "1" ifindex, err := link.GetIfIndex(wgtypes.IfaceName) if err != nil { return err } cDefinesMap["WG_IFINDEX"] = fmt.Sprintf("%d", ifindex) if option.Config.EncryptNode { cDefinesMap["ENABLE_NODE_ENCRYPTION"] = "1" } } if option.Config.EnableL2Announcements { cDefinesMap["ENABLE_L2_ANNOUNCEMENTS"] = "1" // If the agent is down for longer than the lease duration, stop responding cDefinesMap["L2_ANNOUNCEMENTS_MAX_LIVENESS"] = fmt.Sprintf("%dULL", option.Config.L2AnnouncerLeaseDuration.Nanoseconds()) } if option.Config.EnableEncryptionStrictMode { cDefinesMap["ENCRYPTION_STRICT_MODE"] = "1" // when parsing the user input we only accept ipv4 addresses cDefinesMap["STRICT_IPV4_NET"] = fmt.Sprintf("%#x", byteorder.NetIPAddrToHost32(option.Config.EncryptionStrictModeCIDR.Addr())) cDefinesMap["STRICT_IPV4_NET_SIZE"] = fmt.Sprintf("%d", option.Config.EncryptionStrictModeCIDR.Bits()) cDefinesMap["IPV4_ENCRYPT_IFACE"] = fmt.Sprintf("%#x", byteorder.NetIPv4ToHost32(node.GetIPv4())) ipv4Interface, ok := netip.AddrFromSlice(node.GetIPv4().To4()) if !ok { return fmt.Errorf("unable to parse node IPv4 address %s", node.GetIPv4()) } if option.Config.EncryptionStrictModeCIDR.Contains(ipv4Interface) { if !option.Config.EncryptionStrictModeAllowRemoteNodeIdentities { return fmt.Errorf(`encryption strict mode is enabled but the node's IPv4 address is within the strict CIDR range. This will cause the node to drop all traffic. Please either disable encryption or set --encryption-strict-mode-allow-dynamic-lookup=true`) } cDefinesMap["STRICT_IPV4_OVERLAPPING_CIDR"] = "1" } } if option.Config.EnableBPFTProxy { cDefinesMap["ENABLE_TPROXY"] = "1" } if option.Config.EnableXDPPrefilter { cDefinesMap["ENABLE_PREFILTER"] = "1" } if option.Config.EnableEndpointRoutes { cDefinesMap["ENABLE_ENDPOINT_ROUTES"] = "1" } if option.Config.EnableEnvoyConfig { cDefinesMap["ENABLE_L7_LB"] = "1" } if option.Config.EnableSocketLB { if option.Config.BPFSocketLBHostnsOnly { cDefinesMap["ENABLE_SOCKET_LB_HOST_ONLY"] = "1" } else { cDefinesMap["ENABLE_SOCKET_LB_FULL"] = "1" } if option.Config.EnableSocketLBPeer { cDefinesMap["ENABLE_SOCKET_LB_PEER"] = "1" } if option.Config.EnableSocketLBTracing { cDefinesMap["TRACE_SOCK_NOTIFY"] = "1" } if cookie, err := netns.GetNetNSCookie(); err == nil { // When running in nested environments (e.g. Kind), cilium-agent does // not run in the host netns. So, in such cases the cookie comparison // based on bpf_get_netns_cookie(NULL) for checking whether a socket // belongs to a host netns does not work. // // To fix this, we derive the cookie of the netns in which cilium-agent // runs via getsockopt(...SO_NETNS_COOKIE...) and then use it in the // check above. This is based on an assumption that cilium-agent // always runs with "hostNetwork: true". cDefinesMap["HOST_NETNS_COOKIE"] = fmt.Sprintf("%d", cookie) } } cDefinesMap["NAT_46X64_PREFIX_0"] = "0" cDefinesMap["NAT_46X64_PREFIX_1"] = "0" cDefinesMap["NAT_46X64_PREFIX_2"] = "0" cDefinesMap["NAT_46X64_PREFIX_3"] = "0" if option.Config.EnableNodePort { if option.Config.EnableHealthDatapath { cDefinesMap["ENABLE_HEALTH_CHECK"] = "1" } if option.Config.EnableMKE && option.Config.EnableSocketLB { cDefinesMap["ENABLE_MKE"] = "1" cDefinesMap["MKE_HOST"] = fmt.Sprintf("%d", option.HostExtensionMKE) } if option.Config.EnableRecorder { cDefinesMap["ENABLE_CAPTURE"] = "1" if option.Config.EnableIPv4 { cDefinesMap["CAPTURE4_RULES"] = recorder.MapNameWcard4 cDefinesMap["CAPTURE4_SIZE"] = fmt.Sprintf("%d", recorder.MapSize) } if option.Config.EnableIPv6 { cDefinesMap["CAPTURE6_RULES"] = recorder.MapNameWcard6 cDefinesMap["CAPTURE6_SIZE"] = fmt.Sprintf("%d", recorder.MapSize) } } cDefinesMap["ENABLE_NODEPORT"] = "1" if option.Config.EnableIPv4 { cDefinesMap["NODEPORT_NEIGH4"] = neighborsmap.Map4Name cDefinesMap["NODEPORT_NEIGH4_SIZE"] = fmt.Sprintf("%d", option.Config.NeighMapEntriesGlobal) if option.Config.EnableHealthDatapath { cDefinesMap["LB4_HEALTH_MAP"] = lbmap.HealthProbe4MapName } } if option.Config.EnableIPv6 { cDefinesMap["NODEPORT_NEIGH6"] = neighborsmap.Map6Name cDefinesMap["NODEPORT_NEIGH6_SIZE"] = fmt.Sprintf("%d", option.Config.NeighMapEntriesGlobal) if option.Config.EnableHealthDatapath { cDefinesMap["LB6_HEALTH_MAP"] = lbmap.HealthProbe6MapName } } if option.Config.EnableNat46X64Gateway { cDefinesMap["ENABLE_NAT_46X64_GATEWAY"] = "1" base := option.Config.IPv6NAT46x64CIDRBase.AsSlice() cDefinesMap["NAT_46X64_PREFIX_0"] = fmt.Sprintf("%d", base[0]) cDefinesMap["NAT_46X64_PREFIX_1"] = fmt.Sprintf("%d", base[1]) cDefinesMap["NAT_46X64_PREFIX_2"] = fmt.Sprintf("%d", base[2]) cDefinesMap["NAT_46X64_PREFIX_3"] = fmt.Sprintf("%d", base[3]) } if option.Config.NodePortNat46X64 { cDefinesMap["ENABLE_NAT_46X64"] = "1" } const ( dsrEncapInv = iota dsrEncapNone dsrEncapIPIP dsrEncapGeneve ) const ( dsrL4XlateInv = iota dsrL4XlateFrontend dsrL4XlateBackend ) cDefinesMap["DSR_ENCAP_IPIP"] = fmt.Sprintf("%d", dsrEncapIPIP) cDefinesMap["DSR_ENCAP_GENEVE"] = fmt.Sprintf("%d", dsrEncapGeneve) cDefinesMap["DSR_ENCAP_NONE"] = fmt.Sprintf("%d", dsrEncapNone) cDefinesMap["DSR_XLATE_FRONTEND"] = fmt.Sprintf("%d", dsrL4XlateFrontend) cDefinesMap["DSR_XLATE_BACKEND"] = fmt.Sprintf("%d", dsrL4XlateBackend) if option.Config.NodePortMode == option.NodePortModeDSR || option.Config.NodePortMode == option.NodePortModeHybrid { cDefinesMap["ENABLE_DSR"] = "1" if option.Config.EnablePMTUDiscovery { cDefinesMap["ENABLE_DSR_ICMP_ERRORS"] = "1" } if option.Config.NodePortMode == option.NodePortModeHybrid { cDefinesMap["ENABLE_DSR_HYBRID"] = "1" } if option.Config.LoadBalancerDSRDispatch == option.DSRDispatchOption { cDefinesMap["DSR_ENCAP_MODE"] = fmt.Sprintf("%d", dsrEncapNone) } else if option.Config.LoadBalancerDSRDispatch == option.DSRDispatchIPIP { cDefinesMap["DSR_ENCAP_MODE"] = fmt.Sprintf("%d", dsrEncapIPIP) } else if option.Config.LoadBalancerDSRDispatch == option.DSRDispatchGeneve { cDefinesMap["DSR_ENCAP_MODE"] = fmt.Sprintf("%d", dsrEncapGeneve) } if option.Config.LoadBalancerDSRDispatch == option.DSRDispatchIPIP { if option.Config.LoadBalancerDSRL4Xlate == option.DSRL4XlateFrontend { cDefinesMap["DSR_XLATE_MODE"] = fmt.Sprintf("%d", dsrL4XlateFrontend) } else if option.Config.LoadBalancerDSRL4Xlate == option.DSRL4XlateBackend { cDefinesMap["DSR_XLATE_MODE"] = fmt.Sprintf("%d", dsrL4XlateBackend) } } else { cDefinesMap["DSR_XLATE_MODE"] = fmt.Sprintf("%d", dsrL4XlateInv) } } else { cDefinesMap["DSR_ENCAP_MODE"] = fmt.Sprintf("%d", dsrEncapInv) cDefinesMap["DSR_XLATE_MODE"] = fmt.Sprintf("%d", dsrL4XlateInv) } if option.Config.EnableIPv4 { if option.Config.LoadBalancerRSSv4CIDR != "" { ipv4 := byteorder.NetIPv4ToHost32(option.Config.LoadBalancerRSSv4.IP) ones, _ := option.Config.LoadBalancerRSSv4.Mask.Size() cDefinesMap["IPV4_RSS_PREFIX"] = fmt.Sprintf("%d", ipv4) cDefinesMap["IPV4_RSS_PREFIX_BITS"] = fmt.Sprintf("%d", ones) } else { cDefinesMap["IPV4_RSS_PREFIX"] = "IPV4_DIRECT_ROUTING" cDefinesMap["IPV4_RSS_PREFIX_BITS"] = "32" } } if option.Config.EnableIPv6 { if option.Config.LoadBalancerRSSv6CIDR != "" { ipv6 := option.Config.LoadBalancerRSSv6.IP ones, _ := option.Config.LoadBalancerRSSv6.Mask.Size() extraMacrosMap["IPV6_RSS_PREFIX"] = ipv6.String() fw.WriteString(FmtDefineAddress("IPV6_RSS_PREFIX", ipv6)) cDefinesMap["IPV6_RSS_PREFIX_BITS"] = fmt.Sprintf("%d", ones) } else { cDefinesMap["IPV6_RSS_PREFIX"] = "IPV6_DIRECT_ROUTING" cDefinesMap["IPV6_RSS_PREFIX_BITS"] = "128" } } if option.Config.NodePortAcceleration != option.NodePortAccelerationDisabled { cDefinesMap["ENABLE_NODEPORT_ACCELERATION"] = "1" } if !option.Config.EnableHostLegacyRouting { cDefinesMap["ENABLE_HOST_ROUTING"] = "1" } if option.Config.EnableSVCSourceRangeCheck { cDefinesMap["ENABLE_SRC_RANGE_CHECK"] = "1" if option.Config.EnableIPv4 { cDefinesMap["LB4_SRC_RANGE_MAP"] = lbmap.SourceRange4MapName cDefinesMap["LB4_SRC_RANGE_MAP_SIZE"] = fmt.Sprintf("%d", lbmap.SourceRange4Map.MaxEntries()) } if option.Config.EnableIPv6 { cDefinesMap["LB6_SRC_RANGE_MAP"] = lbmap.SourceRange6MapName cDefinesMap["LB6_SRC_RANGE_MAP_SIZE"] = fmt.Sprintf("%d", lbmap.SourceRange6Map.MaxEntries()) } } cDefinesMap["NODEPORT_PORT_MIN"] = fmt.Sprintf("%d", option.Config.NodePortMin) cDefinesMap["NODEPORT_PORT_MAX"] = fmt.Sprintf("%d", option.Config.NodePortMax) cDefinesMap["NODEPORT_PORT_MIN_NAT"] = fmt.Sprintf("%d", option.Config.NodePortMax+1) cDefinesMap["NODEPORT_PORT_MAX_NAT"] = "65535" } macByIfIndexMacro, isL3DevMacro, err := devMacros() if err != nil { return err } cDefinesMap["NATIVE_DEV_MAC_BY_IFINDEX(IFINDEX)"] = macByIfIndexMacro cDefinesMap["IS_L3_DEV(ifindex)"] = isL3DevMacro const ( selectionRandom = iota + 1 selectionMaglev ) cDefinesMap["LB_SELECTION_RANDOM"] = fmt.Sprintf("%d", selectionRandom) cDefinesMap["LB_SELECTION_MAGLEV"] = fmt.Sprintf("%d", selectionMaglev) if option.Config.NodePortAlg == option.NodePortAlgRandom { cDefinesMap["LB_SELECTION"] = fmt.Sprintf("%d", selectionRandom) } else if option.Config.NodePortAlg == option.NodePortAlgMaglev { cDefinesMap["LB_SELECTION"] = fmt.Sprintf("%d", selectionMaglev) cDefinesMap["LB_MAGLEV_LUT_SIZE"] = fmt.Sprintf("%d", option.Config.MaglevTableSize) if option.Config.EnableIPv6 { cDefinesMap["LB6_MAGLEV_MAP_OUTER"] = lbmap.MaglevOuter6MapName } if option.Config.EnableIPv4 { cDefinesMap["LB4_MAGLEV_MAP_OUTER"] = lbmap.MaglevOuter4MapName } } cDefinesMap["HASH_INIT4_SEED"] = fmt.Sprintf("%d", maglev.SeedJhash0) cDefinesMap["HASH_INIT6_SEED"] = fmt.Sprintf("%d", maglev.SeedJhash1) if option.Config.DirectRoutingDeviceRequired() { directRoutingIface := option.Config.DirectRoutingDevice directRoutingIfIndex, err := link.GetIfIndex(directRoutingIface) if err != nil { return err } cDefinesMap["DIRECT_ROUTING_DEV_IFINDEX"] = fmt.Sprintf("%d", directRoutingIfIndex) if option.Config.EnableIPv4 { ip, ok := node.GetNodePortIPv4AddrsWithDevices()[directRoutingIface] if !ok { log.WithFields(logrus.Fields{ "directRoutingIface": directRoutingIface, }).Fatal("Direct routing device's IPv4 address not found") } ipv4 := byteorder.NetIPv4ToHost32(ip) cDefinesMap["IPV4_DIRECT_ROUTING"] = fmt.Sprintf("%d", ipv4) } if option.Config.EnableIPv6 { directRoutingIPv6, ok := node.GetNodePortIPv6AddrsWithDevices()[directRoutingIface] if !ok { log.WithFields(logrus.Fields{ "directRoutingIface": directRoutingIface, }).Fatal("Direct routing device's IPv6 address not found") } extraMacrosMap["IPV6_DIRECT_ROUTING"] = directRoutingIPv6.String() fw.WriteString(FmtDefineAddress("IPV6_DIRECT_ROUTING", directRoutingIPv6)) } } else { var directRoutingIPv6 net.IP cDefinesMap["DIRECT_ROUTING_DEV_IFINDEX"] = "0" if option.Config.EnableIPv4 { cDefinesMap["IPV4_DIRECT_ROUTING"] = "0" } if option.Config.EnableIPv6 { extraMacrosMap["IPV6_DIRECT_ROUTING"] = directRoutingIPv6.String() fw.WriteString(FmtDefineAddress("IPV6_DIRECT_ROUTING", directRoutingIPv6)) } } if option.Config.ResetQueueMapping { cDefinesMap["RESET_QUEUES"] = "1" } if option.Config.EnableBandwidthManager { cDefinesMap["ENABLE_BANDWIDTH_MANAGER"] = "1" cDefinesMap["THROTTLE_MAP"] = bwmap.MapName cDefinesMap["THROTTLE_MAP_SIZE"] = fmt.Sprintf("%d", bwmap.MapSize) } if option.Config.EnableHostFirewall { cDefinesMap["ENABLE_HOST_FIREWALL"] = "1" } if option.Config.EnableIPSec { nodeAddress := node.GetIPv4() if nodeAddress == nil { return errors.New("external IPv4 node address is required when IPSec is enabled, but none found") } a := byteorder.NetIPv4ToHost32(nodeAddress) cDefinesMap["IPV4_ENCRYPT_IFACE"] = fmt.Sprintf("%d", a) if iface := option.Config.EncryptInterface; len(iface) != 0 { link, err := netlink.LinkByName(iface[0]) if err == nil { cDefinesMap["ENCRYPT_IFACE"] = fmt.Sprintf("%d", link.Attrs().Index) } } } if option.Config.EnableNodePort { if option.Config.EnableIPv4 { cDefinesMap["SNAT_MAPPING_IPV4"] = nat.MapNameSnat4Global cDefinesMap["SNAT_MAPPING_IPV4_SIZE"] = fmt.Sprintf("%d", option.Config.NATMapEntriesGlobal) } if option.Config.EnableIPv6 { cDefinesMap["SNAT_MAPPING_IPV6"] = nat.MapNameSnat6Global cDefinesMap["SNAT_MAPPING_IPV6_SIZE"] = fmt.Sprintf("%d", option.Config.NATMapEntriesGlobal) } if option.Config.EnableBPFMasquerade { if option.Config.EnableIPv4Masquerade { cDefinesMap["ENABLE_MASQUERADE_IPV4"] = "1" // ip-masq-agent depends on bpf-masq var excludeCIDR *cidr.CIDR if option.Config.EnableIPMasqAgent { cDefinesMap["ENABLE_IP_MASQ_AGENT_IPV4"] = "1" cDefinesMap["IP_MASQ_AGENT_IPV4"] = ipmasq.MapNameIPv4 // native-routing-cidr is optional with ip-masq-agent and may be nil excludeCIDR = option.Config.GetIPv4NativeRoutingCIDR() } else { excludeCIDR = datapath.RemoteSNATDstAddrExclusionCIDRv4() } if excludeCIDR != nil { cDefinesMap["IPV4_SNAT_EXCLUSION_DST_CIDR"] = fmt.Sprintf("%#x", byteorder.NetIPv4ToHost32(excludeCIDR.IP)) ones, _ := excludeCIDR.Mask.Size() cDefinesMap["IPV4_SNAT_EXCLUSION_DST_CIDR_LEN"] = fmt.Sprintf("%d", ones) } } if option.Config.EnableIPv6Masquerade { cDefinesMap["ENABLE_MASQUERADE_IPV6"] = "1" var excludeCIDR *cidr.CIDR if option.Config.EnableIPMasqAgent { cDefinesMap["ENABLE_IP_MASQ_AGENT_IPV6"] = "1" cDefinesMap["IP_MASQ_AGENT_IPV6"] = ipmasq.MapNameIPv6 excludeCIDR = option.Config.GetIPv6NativeRoutingCIDR() } else { excludeCIDR = datapath.RemoteSNATDstAddrExclusionCIDRv6() } if excludeCIDR != nil { extraMacrosMap["IPV6_SNAT_EXCLUSION_DST_CIDR"] = excludeCIDR.IP.String() fw.WriteString(FmtDefineAddress("IPV6_SNAT_EXCLUSION_DST_CIDR", excludeCIDR.IP)) extraMacrosMap["IPV6_SNAT_EXCLUSION_DST_CIDR_MASK"] = excludeCIDR.Mask.String() fw.WriteString(FmtDefineAddress("IPV6_SNAT_EXCLUSION_DST_CIDR_MASK", excludeCIDR.Mask)) } } } ctmap.WriteBPFMacros(fw, nil) } if option.Config.AllowICMPFragNeeded { cDefinesMap["ALLOW_ICMP_FRAG_NEEDED"] = "1" } if option.Config.ClockSource == option.ClockSourceJiffies { cDefinesMap["ENABLE_JIFFIES"] = "1" } if option.Config.EnableIdentityMark { cDefinesMap["ENABLE_IDENTITY_MARK"] = "1" } if option.Config.EnableHighScaleIPcache { cDefinesMap["ENABLE_HIGH_SCALE_IPCACHE"] = "1" } if option.Config.EnableCustomCalls { cDefinesMap["ENABLE_CUSTOM_CALLS"] = "1" } if option.Config.EnableVTEP { cDefinesMap["ENABLE_VTEP"] = "1" cDefinesMap["VTEP_MAP"] = vtep.Name cDefinesMap["VTEP_MAP_SIZE"] = fmt.Sprintf("%d", vtep.MaxEntries) cDefinesMap["VTEP_MASK"] = fmt.Sprintf("%#x", byteorder.NetIPv4ToHost32(net.IP(option.Config.VtepCidrMask))) } vlanFilter, err := vlanFilterMacros() if err != nil { return err } cDefinesMap["VLAN_FILTER(ifindex, vlan_id)"] = vlanFilter if option.Config.EnableICMPRules { cDefinesMap["ENABLE_ICMP_RULE"] = "1" } cDefinesMap["CIDR_IDENTITY_RANGE_START"] = fmt.Sprintf("%d", identity.MinLocalIdentity) cDefinesMap["CIDR_IDENTITY_RANGE_END"] = fmt.Sprintf("%d", identity.MaxLocalIdentity) if option.Config.TunnelingEnabled() { cDefinesMap["TUNNEL_MODE"] = "1" } ciliumNetLink, err := netlink.LinkByName(defaults.SecondHostDevice) if err != nil { return err } cDefinesMap["CILIUM_NET_MAC"] = fmt.Sprintf("{.addr=%s}", mac.CArrayString(ciliumNetLink.Attrs().HardwareAddr)) cDefinesMap["HOST_IFINDEX"] = fmt.Sprintf("%d", ciliumNetLink.Attrs().Index) ciliumHostLink, err := netlink.LinkByName(defaults.HostDevice) if err != nil { return err } cDefinesMap["HOST_IFINDEX_MAC"] = fmt.Sprintf("{.addr=%s}", mac.CArrayString(ciliumHostLink.Attrs().HardwareAddr)) cDefinesMap["CILIUM_IFINDEX"] = fmt.Sprintf("%d", ciliumHostLink.Attrs().Index) ephemeralMin, err := getEphemeralPortRangeMin() if err != nil { return err } cDefinesMap["EPHEMERAL_MIN"] = fmt.Sprintf("%d", ephemeralMin) if err := cDefinesMap.Merge(h.nodeExtraDefines); err != nil { return err } for _, fn := range h.nodeExtraDefineFns { defines, err := fn() if err != nil { return err } if err := cDefinesMap.Merge(defines); err != nil { return err } } if option.Config.EnableHealthDatapath { if option.Config.IPv4Enabled() { ipip4, err := netlink.LinkByName(defaults.IPIPv4Device) if err != nil { return err } cDefinesMap["ENCAP4_IFINDEX"] = fmt.Sprintf("%d", ipip4.Attrs().Index) } if option.Config.IPv6Enabled() { ipip6, err := netlink.LinkByName(defaults.IPIPv6Device) if err != nil { return err } cDefinesMap["ENCAP6_IFINDEX"] = fmt.Sprintf("%d", ipip6.Attrs().Index) } } // Since golang maps are unordered, we sort the keys in the map // to get a consistent written format to the writer. This maintains // the consistency when we try to calculate hash for a datapath after // writing the config. keys := make([]string, 0, len(cDefinesMap)) for key := range cDefinesMap { keys = append(keys, key) } sort.Strings(keys) for _, key := range keys { fmt.Fprintf(fw, "#define %s %s\n", key, cDefinesMap[key]) } // Populate cDefinesMap with extraMacrosMap to get all the configuration // in the cDefinesMap itself. for key, value := range extraMacrosMap { cDefinesMap[key] = value } // Write the JSON encoded config as base64 encoded commented string to // the header file. jsonBytes, err := json.Marshal(cDefinesMap) if err == nil { // We don't care if some error occurs while marshaling the map. // In such cases we skip embedding the base64 encoded JSON configuration // to the writer. encodedConfig := base64.StdEncoding.EncodeToString(jsonBytes) fmt.Fprintf(fw, "\n// JSON_OUTPUT: %s\n", encodedConfig) } return fw.Flush() }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *Config) WriteNodeID() error {\n\tnodeIDFile := \"node.id\"\n\terr := ioutil.WriteFile(nodeIDFile, []byte(c.NodeID), 0666)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"cannot write file [%s]: %v\", nodeIDFile, err)\n\t}\n\treturn nil\n}", "func (mod *EthModule) WriteConfig(config_file string) error {\n\tb, err := json.Marshal(mod.eth.config)\n\tif err != nil {\n\t\tfmt.Println(\"error marshalling config:\", err)\n\t\treturn err\n\t}\n\tvar out bytes.Buffer\n\tjson.Indent(&out, b, \"\", \"\\t\")\n\terr = ioutil.WriteFile(config_file, out.Bytes(), 0600)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func WriteConfig(f string, c *SupportedBranchesConfig) error {\n\treturn util.WithWriteFile(f, func(w io.Writer) error {\n\t\treturn EncodeConfig(w, c)\n\t})\n}", "func (repo *Repo) WriteConfig() error {\n\tif repo.Root != repo {\n\t\treturn repo.Root.WriteConfig()\n\t}\n\n\tb, err := json.MarshalIndent(repo, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn ioutil.WriteFile(path.Join(repo.Path, cacheRelPath), b, 0660)\n}", "func (b *Bot) WriteConfig(fn configCallback) {\n\tfn(b.conf)\n}", "func (n *Node) WriteNodeSettings(settings *NodeSettings) error {\n\ts, err := ksigsyaml.Marshal(*settings)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"failed to encode %s\", nodeSettingsPath)\n\t}\n\n\tdir := filepath.Dir(nodeSettingsPath)\n\tlog.Debugf(\"Writing node settings at %s...\", nodeSettingsPath)\n\terr = n.Command(\n\t\t\"mkdir\", \"-p\", dir,\n\t).Silent().Run()\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"failed to create %s\", dir)\n\t}\n\tif err := n.WriteFile(nodeSettingsPath, s); err != nil {\n\t\treturn errors.Wrapf(err, \"failed to write %s\", nodeSettingsPath)\n\t}\n\n\treturn nil\n}", "func WriteConfig(c Config, filename string) {\n\n\tdata, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\terr = ioutil.WriteFile(filename, data, 664)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func (c *Config) ConfigSaveWriter(w io.Writer) error {\n\tout, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = w.Write(out)\n\treturn err\n}", "func WriteNode(root interface{}, path util.Path, value interface{}) error {\n\tpc, _, err := getPathContext(&PathContext{Node: root}, path, path, true)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn WritePathContext(pc, value, false)\n}", "func (r *Resource) WriteConfig(filename string) error {\n\tr.Lock()\n\tdefer r.Unlock()\n\n\tvar b bytes.Buffer\n\n\tb.WriteString(fmt.Sprintf(\"# meta-data-json:{\\\"updated\\\": \\\"%s\\\"}\\n\", time.Now().UTC()))\n\tb.WriteString(fmt.Sprintf(\"resource %s {\\n\", r.name))\n\n\tvar hosts []string\n\tfor _, h := range r.host {\n\t\thosts = append(hosts, h.Name)\n\n\t\tb.WriteString(indentf(1, \"on %s {\\n\", h.Name))\n\t\tb.WriteString(indentf(2, \"node-id %d;\\n\", h.ID))\n\t\tb.WriteString(indentf(2, \"address %s:%d;\\n\", h.IP, r.port))\n\t\tfor _, v := range h.volume {\n\t\t\tb.WriteString(indentf(2, \"volume %d {\\n\", v.id))\n\t\t\tb.WriteString(indentf(3, \"device minor %d;\\n\", v.minor))\n\t\t\tb.WriteString(indentf(3, \"disk %s;\\n\", v.backingDevice))\n\t\t\tb.WriteString(indentf(3, \"meta-disk internal;\\n\"))\n\t\t\tb.WriteString(indentf(2, \"}\\n\")) // end volume section\n\t\t}\n\t\tb.WriteString(indentf(1, \"}\\n\")) // end on section\n\t\tb.WriteString(\"\\n\")\n\t}\n\n\tb.WriteString(indentf(1, \"connection-mesh {\\n\"))\n\tb.WriteString(indentf(2, \"hosts %s;\\n\", strings.Join(hosts, \" \")))\n\tb.WriteString(indentf(1, \"}\\n\"))\n\n\tb.WriteString(\"}\") // end resource section\n\n\treturn ioutil.WriteFile(filename, b.Bytes(), 0644)\n}", "func (pub *Publisher) UpdateNodeConfig(nodeHWID string, attrName types.NodeAttr, configAttr *types.ConfigAttr) {\n\tpub.registeredNodes.UpdateNodeConfig(nodeHWID, attrName, configAttr)\n}", "func (c *Config) Write() error {\n\trawConfig, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = os.MkdirAll(configDirPath, 0755)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn ioutil.WriteFile(configPath, rawConfig, 0644)\n}", "func ConfigWrite() error {\n\n\t// Marshal it\n\tconfigJSON, _ := json.MarshalIndent(Config, \"\", \" \")\n\n\t// Write the file\n\tfd, err := os.OpenFile(configSettingsPath(), os.O_WRONLY|os.O_TRUNC|os.O_CREATE, 0666)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfd.Write(configJSON)\n\tfd.Close()\n\n\t// Done\n\treturn err\n\n}", "func Write(w io.Writer, cfg *Config) error {\n\tif cfg.filename != \"\" {\n\t\tfmt.Fprintf(w, \"# Configuration file location: %v\\n\\n\", cfg.filename)\n\t} else {\n\t\tfmt.Fprintf(w, \"# Cound not find configuration file location.\\n\\n\")\n\t}\n\treturn toml.NewEncoder(w).Encode(cfg.File)\n}", "func SetNodeConfig(ctx *cli.Context, cfg *node.NodeConfig) {\r\n\t//SetP2PConfig(ctx, &cfg.P2P)\r\n\r\n\tswitch {\r\n\tcase ctx.GlobalIsSet(DataDirFlag.Name):\r\n\t\tcfg.DataDir = ctx.GlobalString(DataDirFlag.Name)\r\n\tcase ctx.GlobalBool(TestnetFlag.Name): //JiangHan:对于两种测试模式网,这里在主datapath后面附加了子文件夹,以示区分,避免跟主模式目录数据冲突\r\n\t\tcfg.DataDir = filepath.Join(node.DefaultDataDir(), \"testnet\")\r\n\tcase ctx.GlobalBool(RinkebyFlag.Name):\r\n\t\tcfg.DataDir = filepath.Join(node.DefaultDataDir(), \"rinkeby\")\r\n\t}\r\n}", "func (parser *Parser) WriteConfig() error {\n\tconfig := structs.Map(parser.Config)\n\tfor _, site := range parser.Sites {\n\t\tif site.Path == \".\" {\n\t\t\tfor key, val := range site.AllParameters {\n\t\t\t\tif _, ok := config[key]; !ok {\n\t\t\t\t\tconfig[key] = val\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tbs, err := yaml.Marshal(config)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn parser.Storage.Write(parser.Storage.GetSourceKey(KeyConfig), bs)\n}", "func WriteConfig(i interface{}, d string) error {\n\t// Convert to yaml\n\tym, err := yaml.Marshal(i)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Write to file\n\treturn ioutil.WriteFile(d, ym, 0755)\n}", "func WriteThrapConfig(conf *ThrapConfig, filename string) error {\n\tb, err := hclencoder.Encode(conf)\n\tif err == nil {\n\t\terr = ioutil.WriteFile(filename, b, 0644)\n\t}\n\treturn err\n}", "func WithWriteConfig(c Config) Option {\n\treturn func(ro *repoOptions) {\n\t\tro.writeConfigs = append(ro.writeConfigs, c)\n\t}\n}", "func writeConfig(f Function) (err error) {\n\tpath := filepath.Join(f.Root, ConfigFile)\n\tc := toConfig(f)\n\tvar bb []byte\n\tif bb, err = yaml.Marshal(&c); err != nil {\n\t\treturn\n\t}\n\treturn ioutil.WriteFile(path, bb, 0644)\n}", "func (t *BPTree) WriteNode(n *Node, off int64, syncEnable bool, fd *os.File) (number int, err error) {\n\tbn, err := t.ToBinary(n)\n\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\tif off == -1 {\n\t\toff = n.Address\n\t}\n\n\tnumber, err = fd.WriteAt(bn, off)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\tif syncEnable {\n\t\terr = fd.Sync()\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\treturn\n}", "func (node *URLNode) WriteTree(writer io.Writer) {\n\tif _, err := writer.Write([]byte(node.GenerateTree())); err != nil {\n\t\tlog.Error(err)\n\t}\n}", "func writeNode(w io.Writer, fset *token.FileSet, x interface{}) {\n\t// convert trailing tabs into spaces using a tconv filter\n\t// to ensure a good outcome in most browsers (there may still\n\t// be tabs in comments and strings, but converting those into\n\t// the right number of spaces is much harder)\n\t//\n\t// TODO(gri) rethink printer flags - perhaps tconv can be eliminated\n\t// with an another printer mode (which is more efficiently\n\t// implemented in the printer than here with another layer)\n\tmode := printer.TabIndent | printer.UseSpaces\n\terr := (&printer.Config{Mode: mode, Tabwidth: tabwidth}).Fprint(&tconv{output: w}, fset, x)\n\tif err != nil {\n\t\tlog.Print(err)\n\t}\n}", "func (c *Config) WriteConfig(filename string) error {\n\tdata, err := yaml.Marshal(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = ioutil.WriteFile(filename, data, 0644)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func foldersConfigWrite(file string, folders []folder.FolderConfig) error {\n\tffMutex.Lock()\n\tdefer ffMutex.Unlock()\n\n\tfd, err := os.OpenFile(file, os.O_CREATE|os.O_TRUNC|os.O_WRONLY, 0666)\n\tdefer fd.Close()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdata := &xmlFolders{\n\t\tVersion: \"1\",\n\t\tFolders: folders,\n\t}\n\n\tenc := xml.NewEncoder(fd)\n\tenc.Indent(\"\", \" \")\n\treturn enc.Encode(data)\n}", "func EncodeConfig(configFile io.Writer, c *Configuration) error {\n\t// Lock mutex\n\tConfig.rw.Lock()\n\tdefer Config.rw.Unlock()\n\n\t// Encode the given writer with the given interface\n\treturn toml.NewEncoder(configFile).Encode(c)\n}", "func (dw *dataStore) WriteNode(nodeLabel string, nodeProperties map[string]interface{}) error {\n\tproperties := dw.getPropertiesString(nodeLabel)\n\tparamsMap := dw.getParamsMap(nodeLabel, nodeProperties)\n\tquery := fmt.Sprintf(\"MERGE (n:%s %s)\", nodeLabel, properties)\n\tlog.Info.Printf(\"[WRITE_NODE] Executing query %s\\n\", query)\n\n\treturn dw.write(query, paramsMap)\n}", "func (sm *ShardMaster) writeConfig(config Config) {\n\ttoWrite := JSONConfig{Num: config.Num, Shards: config.Shards}\n\tnewGroups := make(map[string][]string)\n\n\tfor gid, vals := range config.Groups {\n\t\tnewGroups[strconv.FormatInt(gid, 10)] = vals\n\t}\n\ttoWrite.Groups = newGroups\n\n\tb, err := json.Marshal(toWrite)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tf, err2 := os.OpenFile(sm.configFile, os.O_APPEND|os.O_WRONLY|os.O_CREATE, 0666)\n\tif err2 != nil {\n\t\tlog.Fatal(err2)\n\t}\n\tdefer f.Close()\n\t_, err = f.WriteString(string(b) + \"\\n\")\n}", "func WriteKubeletConfigFile(kubeletConfig *kubeletconfig.KubeletConfiguration) error {\n\tdata, err := kubeletconfigcodec.EncodeKubeletConfig(kubeletConfig, kubeletconfigv1beta1.SchemeGroupVersion)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tkubeletConfigFilePath, err := getKubeletConfigFilePath()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err := os.WriteFile(kubeletConfigFilePath, data, 0644); err != nil {\n\t\treturn fmt.Errorf(\"failed to write the kubelet file to %q: %w\", kubeletConfigFilePath, err)\n\t}\n\n\treturn nil\n}", "func WriteServerConfig(filePath string, cfg *ServerConfig) error {\n\tbts, err := toml.Marshal(*cfg)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn files.WriteAndBackup(filePath, bts)\n}", "func WriteConfig(cfg NetworkMonitorConfig) string {\n\tif cfgBytes, err := yml.Marshal(cfg); err != nil {\n\t\tpanic(err.Error())\n\t} else {\n\t\tfmt.Println(string(cfgBytes))\n\t\treturn string(cfgBytes)\n\t}\n}", "func (t *BPTree) WriteNodes(rwMode RWMode, syncEnable bool, flag int) error {\n\tvar (\n\t\tn *Node\n\t\ti int\n\t\terr error\n\t)\n\n\tfd, err := os.OpenFile(t.Filepath, os.O_CREATE|os.O_RDWR, 0644)\n\tdefer fd.Close()\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tqueue = nil\n\n\tenqueue(t.root)\n\n\tfor queue != nil {\n\t\tn = dequeue()\n\n\t\t_, err := t.WriteNode(n, -1, syncEnable, fd)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif n != nil {\n\t\t\tif !n.isLeaf {\n\t\t\t\tfor i = 0; i <= n.KeysNum; i++ {\n\t\t\t\t\tc, _ := n.pointers[i].(*Node)\n\t\t\t\t\tenqueue(c)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func writeKubeadmConfig(kubeadmConfig string, node nodes.Node) error {\n\t// copy the config to the node\n\tif err := nodeutils.WriteFile(node, \"/kind/kubeadm.conf\", kubeadmConfig); err != nil {\n\t\t// TODO(bentheelder): logging here\n\t\treturn errors.Wrap(err, \"failed to copy kubeadm config to node\")\n\t}\n\n\treturn nil\n}", "func (m *K3dClusterManager) WriteKubeconfig(ctx context.Context, path string) error {\n\tapiConfig, err := m.GetKubeconfig(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn k3dcluster.KubeconfigWriteToPath(ctx, apiConfig, path)\n}", "func EncodeConfig(w io.Writer, c *SupportedBranchesConfig) error {\n\tc.Sort()\n\tb, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = w.Write(b)\n\treturn err\n}", "func WriteConfig(cfg *config.Dendrite, configDir string) error {\n\tdata, err := yaml.Marshal(cfg)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn ioutil.WriteFile(filepath.Join(configDir, ConfigFile), data, 0666)\n}", "func (c *Config) WriteScopeConfig(path string) error {\n\tscb, err := c.ScopeConfigYaml()\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = ioutil.WriteFile(path, scb, 0644)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error writing ScopeConfig to file %s: %v\", path, err)\n\t}\n\treturn nil\n}", "func WriteSensuClusterConfig(newCluster []byte) error {\n\tfileloc, err := homedir.Expand(sensuctlClusterFile)\n\tif err != nil {\n\t\treturn err\n\t}\n\tioutil.WriteFile(fileloc, newCluster, 0644)\n\treturn nil\n}", "func (d donut) SaveConfig() error {\n\tnodeDiskMap := make(map[string][]string)\n\tfor hostname, node := range d.nodes {\n\t\tdisks, err := node.ListDisks()\n\t\tif err != nil {\n\t\t\treturn iodine.New(err, nil)\n\t\t}\n\t\tfor order, disk := range disks {\n\t\t\tdonutConfigPath := filepath.Join(d.name, donutConfig)\n\t\t\tdonutConfigWriter, err := disk.CreateFile(donutConfigPath)\n\t\t\tdefer donutConfigWriter.Close()\n\t\t\tif err != nil {\n\t\t\t\treturn iodine.New(err, nil)\n\t\t\t}\n\t\t\tnodeDiskMap[hostname][order] = disk.GetPath()\n\t\t\tjenc := json.NewEncoder(donutConfigWriter)\n\t\t\tif err := jenc.Encode(nodeDiskMap); err != nil {\n\t\t\t\treturn iodine.New(err, nil)\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func (c *Config) Write(filename string) (err error) {\n\tb, err := json.Marshal(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfile, err := os.Create(filename)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif _, err := file.Write(b); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func WriteConfig(config Config, path string) error {\n\tb, err := yaml.Marshal(config)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = ioutil.WriteFile(path, b, 0644)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func WriteConfig() {\n\tf, err := os.Create(\"config.json\")\n\tif err != nil {\n\t\tlog.Fatalf(\"Can't write configuration file: %v\", err)\n\t}\n\te := json.NewEncoder(f)\n\te.SetIndent(\"\", \" \")\n\te.Encode(defaultConfig)\n\tf.Close()\n\tos.Exit(0)\n}", "func (pub *Publisher) UpdateNodeConfigValues(nodeHWID string, params types.NodeAttrMap) (changed bool) {\n\treturn pub.registeredNodes.UpdateNodeConfigValues(nodeHWID, params)\n}", "func (o ClusterOutput) NodeConfig() ClusterNodeConfigOutput {\n\treturn o.ApplyT(func(v *Cluster) ClusterNodeConfigOutput { return v.NodeConfig }).(ClusterNodeConfigOutput)\n}", "func (cfg *Config) Write() error {\n\tif err := os.MkdirAll(filepath.Dir(cfg.path), 0755); err != nil {\n\t\treturn errors.Wrap(err, \"failed to create config path\")\n\t}\n\n\tfh, err := os.Create(cfg.path)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to open config file\")\n\t}\n\tdefer fh.Close()\n\n\tbuf, err := yaml.Marshal(cfg)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to marshal config\")\n\t}\n\n\t_, err = fh.Write(buf)\n\treturn errors.Wrap(err, \"failed to write config\")\n}", "func WriteConfigVals(dir string, vals map[string]string) error {\n\tdata := \"\"\n\tfor k, v := range vals {\n\t\tdata += fmt.Sprintf(\"%s = \\\"%s\\\"\\n\", k, v)\n\t}\n\tcfile := filepath.Join(dir, \"config.toml\")\n\treturn os.WriteFile(cfile, []byte(data), 0o600)\n}", "func writeConfig(args *common.Parameters) {\n\n\t//Create the config file and open it for writing.\n\tconfigWrite, err := os.Create(\"./data/cluster/config.csv\")\n\tif err != nil {\n\t\targs.ErrorLogger.Println(\"entity=\" + entityKind + \" message=\" + err.Error())\n\t\tfmt.Println(\"[ERROR] entity=\" + entityKind + \" message=\" + err.Error())\n\t\treturn\n\t}\n\n\t//Write out the header.\n\tfmt.Fprintln(configWrite, \"AuditTime,Name\")\n\tfmt.Fprintf(configWrite, \"%s,%s\\n\", common.Format(args.CurrentTime), *args.ClusterName)\n\tconfigWrite.Close()\n}", "func (c DQLConfig) Write() error {\n\tf := filepath.Join(helpers.GetProjectPath(c.ProjectPath), \"dql.conf.json\")\n\n\tjson, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn ioutil.WriteFile(f, json, 0644)\n}", "func writeNode(w io.Writer, node interface{}, html bool, styler printer.Styler) {\n\tmode := printer.UseSpaces;\n\tif html {\n\t\tmode |= printer.GenHTML\n\t}\n\t(&printer.Config{mode, *tabwidth, styler}).Fprint(w, node);\n}", "func (p *ProviderConfigs) writeProviderConfig() error {\n\t// Get file.\n\tconfig := getProviderConfigFile()\n\n\t// Convert object to json.\n\tdata, err := json.MarshalIndent(p, \"\", \" \")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\t// Write json output to file.\n\terr = ioutil.WriteFile(config, data, 0644)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn nil\n}", "func WriteConfig(config SlackConfiguration) error {\n\texpanded, err := homedir.Expand(configPath)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tconfigFolder := filepath.Dir(expanded)\n\tif err := os.MkdirAll(configFolder, 0777); err != nil {\n\t\treturn err\n\t}\n\tconfigFile, err := os.Create(expanded)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer configFile.Close()\n\n\tif err := json.NewEncoder(configFile).Encode(config); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func WriteConnectionConfig(context *clusterd.Context, clusterInfo *cephclient.ClusterInfo) error {\n\t// write the latest config to the config dir\n\tif _, err := cephclient.GenerateConnectionConfig(context, clusterInfo); err != nil {\n\t\treturn errors.Wrap(err, \"failed to write connection config\")\n\t}\n\n\treturn nil\n}", "func NodeConfig(s *base.Store) {\n\tstore = s\n}", "func WriteCredsConfig(cc *CredsConfig, fpath string) error {\n\tb, err := hclencoder.Encode(cc)\n\tif err == nil {\n\t\terr = ioutil.WriteFile(fpath, b, 0644)\n\t}\n\treturn err\n}", "func (e *Escpos) WriteNode(name string, params map[string]string, data string) {\n\tcstr := \"\"\n\tif data != \"\" {\n\t\tstr := data[:]\n\t\tif len(data) > 40 {\n\t\t\tstr = fmt.Sprintf(\"%s ...\", data[0:40])\n\t\t}\n\t\tcstr = fmt.Sprintf(\" => '%s'\", str)\n\t}\n\n\tif e.Verbose {\n\t\tlog.Println(\"Write: %s => %+v%s\\n\", name, params, cstr)\n\t}\n\n\tswitch name {\n\tcase \"text\":\n\t\te.Text(params, data)\n\tcase \"feed\":\n\t\te.Feed(params)\n\tcase \"cut\":\n\t\te.FeedAndCut(params)\n\tcase \"pulse\":\n\t\te.Pulse()\n\tcase \"image\":\n\t\te.Image(params, data)\n\t}\n}", "func (ec *ExecutionContext) WriteConfig(config *Config) error {\n\tvar op errors.Op = \"cli.ExecutionContext.WriteConfig\"\n\tvar cfg *Config\n\tif config != nil {\n\t\tcfg = config\n\t} else {\n\t\tcfg = ec.Config\n\t}\n\tbuf := new(bytes.Buffer)\n\tencoder := yaml.NewEncoder(buf)\n\tencoder.SetIndent(2)\n\terr := encoder.Encode(cfg)\n\tif err != nil {\n\t\treturn errors.E(op, err)\n\t}\n\terr = ioutil.WriteFile(ec.ConfigFile, buf.Bytes(), 0644)\n\tif err != nil {\n\t\treturn errors.E(op, err)\n\t}\n\treturn nil\n}", "func (wr *plainGenomeWriter) writeNetworkNode(n *network.NNode) error {\n\ttrait_id := 0\n\tif n.Trait != nil {\n\t\ttrait_id = n.Trait.Id\n\t}\n\tact_str, err := utils.NodeActivators.ActivationNameFromType(n.ActivationType)\n\tif err == nil {\n\t\t_, err = fmt.Fprintf(wr.w, \"%d %d %d %d %s\", n.Id, trait_id, n.NodeType(),\n\t\t\tn.NeuronType, act_str)\n\t}\n\treturn err\n}", "func WriteConfiguration(config *Config, checked, checkedIgnore bool) error {\n\tif config.IgnoreURL == \"\" {\n\t\tconfig.IgnoreURL = \"https://raw.githubusercontent.com/projectdiscovery/nuclei-templates/master/.nuclei-ignore\"\n\t}\n\tif checked {\n\t\tconfig.LastChecked = time.Now()\n\t}\n\tif checkedIgnore {\n\t\tconfig.LastCheckedIgnore = time.Now()\n\t}\n\tconfig.NucleiVersion = Version\n\n\ttemplatesConfigFile, err := getConfigDetails()\n\tif err != nil {\n\t\treturn err\n\t}\n\tfile, err := os.OpenFile(templatesConfigFile, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, 0777)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer file.Close()\n\n\terr = jsoniter.NewEncoder(file).Encode(config)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func writeConfig() error {\n\tconfigFile := filepath.Join(config.ourBinaryDir, config.ourConfigFilename)\n\tlog.Printf(\"Writing YAML file: %s\", configFile)\n\tyamlText, err := yaml.Marshal(&config)\n\tif err != nil {\n\t\tlog.Printf(\"Couldn't generate YAML file: %s\", err)\n\t\treturn err\n\t}\n\terr = writeFileSafe(configFile, yamlText)\n\tif err != nil {\n\t\tlog.Printf(\"Couldn't save YAML config: %s\", err)\n\t\treturn err\n\t}\n\n\tuserFilter := getUserFilter()\n\terr = userFilter.save()\n\tif err != nil {\n\t\tlog.Printf(\"Couldn't save the user filter: %s\", err)\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (mgr *Manager) SaveNodeDef(kind string, force bool) error {\n\tatomic.AddUint64(&mgr.stats.TotSaveNodeDef, 1)\n\n\tif mgr.cfg == nil {\n\t\tatomic.AddUint64(&mgr.stats.TotSaveNodeDefNil, 1)\n\t\treturn nil // Occurs during testing.\n\t}\n\n\tnodeDef := &NodeDef{\n\t\tHostPort: mgr.bindHttp,\n\t\tUUID: mgr.uuid,\n\t\tImplVersion: mgr.version,\n\t\tTags: mgr.tags,\n\t\tContainer: mgr.container,\n\t\tWeight: mgr.weight,\n\t\tExtras: mgr.extras,\n\t}\n\n\tfor {\n\t\tnodeDefs, cas, err := CfgGetNodeDefs(mgr.cfg, kind)\n\t\tif err != nil {\n\t\t\tatomic.AddUint64(&mgr.stats.TotSaveNodeDefGetErr, 1)\n\t\t\treturn err\n\t\t}\n\t\tif nodeDefs == nil {\n\t\t\tnodeDefs = NewNodeDefs(mgr.version)\n\t\t}\n\t\tnodeDefPrev, exists := nodeDefs.NodeDefs[mgr.uuid]\n\t\tif exists && !force {\n\t\t\tif reflect.DeepEqual(nodeDefPrev, nodeDef) {\n\t\t\t\tatomic.AddUint64(&mgr.stats.TotSaveNodeDefSame, 1)\n\t\t\t\tatomic.AddUint64(&mgr.stats.TotSaveNodeDefOk, 1)\n\t\t\t\treturn nil // No changes, so leave the existing nodeDef.\n\t\t\t}\n\t\t}\n\n\t\tnodeDefs.UUID = NewUUID()\n\t\tnodeDefs.NodeDefs[mgr.uuid] = nodeDef\n\t\tnodeDefs.ImplVersion = CfgGetVersion(mgr.cfg)\n\t\tlog.Printf(\"manager: setting the nodeDefs implVersion \"+\n\t\t\t\"to %s\", nodeDefs.ImplVersion)\n\n\t\t_, err = CfgSetNodeDefs(mgr.cfg, kind, nodeDefs, cas)\n\t\tif err != nil {\n\t\t\tif _, ok := err.(*CfgCASError); ok {\n\t\t\t\t// Retry if it was a CAS mismatch, as perhaps\n\t\t\t\t// multiple nodes are all racing to register themselves,\n\t\t\t\t// such as in a full datacenter power restart.\n\t\t\t\tatomic.AddUint64(&mgr.stats.TotSaveNodeDefRetry, 1)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tatomic.AddUint64(&mgr.stats.TotSaveNodeDefSetErr, 1)\n\t\t\treturn err\n\t\t}\n\t\tbreak\n\t}\n\tatomic.AddUint64(&mgr.stats.TotSaveNodeDefOk, 1)\n\treturn nil\n}", "func Writer(w io.Writer) Option {\n\treturn func(c *Config) Option {\n\t\tprevious := c.ioWriter\n\t\tc.ioWriter = w\n\t\treturn Writer(previous)\n\t}\n}", "func WriteExpansionConfigFile(ctx context.Context, config *ExpansionConfig, filename string) error {\n\tbc, err := json.Marshal(config)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn ioutils.WriteBytes(ctx, bc, filename)\n}", "func (f *FileConfigWriter) Write(config *auth.Config) error {\n\tif f.filename == \"\" {\n\t\treturn errors.New(\"No config file name defined\")\n\t}\n\tcontent, err := yaml.Marshal(config)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"marshaling the config to yaml\")\n\t}\n\terr = ioutil.WriteFile(f.filename, content, util.DefaultWritePermissions)\n\treturn nil\n}", "func WriteConfigFile(c *cli.Context, cfg *Config) error {\n\tfPath, err := TenetCfgPath(c)\n\tif err != nil {\n\t\treturn errors.Trace(err)\n\t}\n\tvar buf bytes.Buffer\n\tenc := toml.NewEncoder(&buf)\n\terr = enc.Encode(cfg)\n\tif err != nil {\n\t\treturn errors.Trace(err)\n\t}\n\treturn ioutil.WriteFile(fPath, buf.Bytes(), 0644)\n}", "func getWriterConfig() *hub.EventHubConfig {\n\treturn &hub.EventHubConfig{\n\t\tNamespace: viper.GetString(\"write_namespace\"),\n\t\tHub: viper.GetString(\"write_hub\"),\n\t\tKeyName: viper.GetString(\"write_keyname\"),\n\t\tKeyValue: viper.GetString(\"write_keyvalue\"),\n\t\tConnString: viper.GetString(\"write_connstring\"),\n\t\tTenantID: viper.GetString(\"write_tenantid\"),\n\t\tClientID: viper.GetString(\"write_clientid\"),\n\t\tClientSecret: viper.GetString(\"write_clientsecret\"),\n\t\tCertPath: viper.GetString(\"write_certpath\"),\n\t\tCertPassword: viper.GetString(\"write_certpassword\"),\n\t\tBatch: viper.GetBool(\"write_batch\"),\n\t\tPartKeyLabel: viper.GetString(\"partition_key_label\"),\n\t\tADXMapping: viper.GetString(\"write_adxmapping\"),\n\t\tSerializer: serializers.SerializerConfig{DataFormat: viper.GetString(\"write_serializer\")},\n\t}\n}", "func Writer(config Config) store.Writer {\n\treturn &writer{config: config}\n}", "func (h *HeaderfileWriter) WriteNetdevConfig(w io.Writer, cfg datapath.DeviceConfiguration) error {\n\tfw := bufio.NewWriter(w)\n\th.writeNetdevConfig(fw, cfg)\n\treturn fw.Flush()\n}", "func (c *Config) Write(path string) error {\r\n\tbytes, err := yaml.Marshal(c)\r\n\tif err == nil {\r\n\t\treturn ioutil.WriteFile(path, bytes, 0777)\r\n\t}\r\n\treturn err\r\n}", "func writeConfig(name string) {\n\trequest := RpcRequest{(fmt.Sprintf(\"{\\\"command\\\":\\\"save\\\",\\\"parameter\\\":\\\"\\\"}\")), make(chan []byte), name}\n\trequest.Send()\n}", "func (m *BgpConfiguration) Serialize(writer i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.SerializationWriter)(error) {\n {\n err := writer.WriteInt32Value(\"asn\", m.GetAsn())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"ipAddress\", m.GetIpAddress())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"localIpAddress\", m.GetLocalIpAddress())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"@odata.type\", m.GetOdataType())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"peerIpAddress\", m.GetPeerIpAddress())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteAdditionalData(m.GetAdditionalData())\n if err != nil {\n return err\n }\n }\n return nil\n}", "func (c *CNIConfigWriter) Write(inputs CNIConfigInputs) error {\n\tif inputs == c.lastInputs {\n\t\treturn nil\n\t}\n\n\t// use an extension not recognized by CNI to write the contents initially\n\t// https://github.com/containerd/go-cni/blob/891c2a41e18144b2d7921f971d6c9789a68046b2/opts.go#L170\n\t// then we can rename to atomically make the file appear\n\tf, err := os.Create(c.path + \".temp\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// actually write the config\n\tif err := writeCNIConfig(f, c.template, inputs); err != nil {\n\t\tf.Close()\n\t\tos.Remove(f.Name())\n\t\treturn err\n\t}\n\tf.Sync()\n\tf.Close()\n\n\t// then we can rename to the target config path\n\tif err := os.Rename(f.Name(), c.path); err != nil {\n\t\treturn err\n\t}\n\n\t// we're safely done now, record the inputs\n\tc.lastInputs = inputs\n\treturn nil\n}", "func WriteConfigJSON(c Configuration, filename string) error {\n\tbytes, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn ioutil.WriteFile(filename, bytes, 0644)\n}", "func Write(config *api.Config, path string) error {\n\treturn clientcmd.WriteToFile(*config, path)\n}", "func (cfg *Config) Write() error {\n\tyamlBytes, err := yaml.Marshal(cfg)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn ioutil.WriteFile(FilePath(), yamlBytes, 0644)\n}", "func WriteChainConfig(db database.KeyValueWriter, hash common.Hash, cfg *config.ChainConfig) {\n\tif cfg == nil {\n\t\treturn\n\t}\n\tdata, err := json.Marshal(cfg)\n\tif err != nil {\n\t\tlog.Critical(\"Failed to JSON encode chain config\", \"err\", err)\n\t}\n\tif err := db.Put(configKey(hash), data, \"config\"); err != nil {\n\t\tlog.Critical(\"Failed to store chain config\", \"err\", err)\n\t}\n}", "func writeConfig(o options, clientset kubernetes.Interface) error {\n\t// kubeconfig is a kubernetes config.\n\tvar kubeconfig []byte\n\n\tdir, file := filepath.Split(o.output)\n\n\terr := os.MkdirAll(dir, os.ModePerm)\n\tif err != nil {\n\t\treturn &util.ExitError{Message: fmt.Sprintf(\"unable to create output directory %v: %v.\", dir, err), Code: 1}\n\t}\n\n\tif o.certificate {\n\t\tif kubeconfig, err = certificate.CreateKubeConfigWithCertificateCredentials(clientset, o.name); err != nil {\n\t\t\treturn &util.ExitError{Message: fmt.Sprintf(\"unable to create kubeconfig file with cert and key for %v: %v.\", o.name, err), Code: 1}\n\t\t}\n\t} else {\n\t\t// Service account credentials are the default if unspecified.\n\t\tif kubeconfig, err = serviceaccount.CreateKubeConfigWithServiceAccountCredentials(clientset, o.name); err != nil {\n\t\t\treturn &util.ExitError{Message: fmt.Sprintf(\"unable to create kubeconfig file with service account for %v: %v.\", o.name, err), Code: 1}\n\t\t}\n\t}\n\n\tif !o.overwrite && util.FileExists(o.output) {\n\t\tif kubeconfig, err = mergeConfigs(o, kubeconfig); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif err = ioutil.WriteFile(o.output, kubeconfig, 0644); err != nil {\n\t\treturn &util.ExitError{Message: fmt.Sprintf(\"unable to write to file %v: %v.\", file, err), Code: 1}\n\t}\n\n\treturn nil\n}", "func WriteSysConfig(fileName, value string) error {\n\tf, err := os.OpenFile(fileName, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, 0666)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to open configuration file: %s\", err)\n\t}\n\t_, err = f.WriteString(value)\n\tif err != nil {\n\t\tf.Close()\n\t\treturn fmt.Errorf(\"unable to write value: %s\", err)\n\t}\n\terr = f.Close()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to close configuration file: %s\", err)\n\t}\n\treturn nil\n}", "func ConfigWriterSync(level Level, w zapcore.WriteSyncer) {\n\tcustomLog := zap.New(zapcore.NewCore(zapcore.NewJSONEncoder(zapcore.EncoderConfig{\n\t\tTimeKey: \"timestamp\",\n\t\tLevelKey: \"level\",\n\t\tNameKey: \"logger\",\n\t\tCallerKey: \"caller\",\n\t\tMessageKey: \"message\",\n\t\tStacktraceKey: \"stacktrace\",\n\t\tLineEnding: zapcore.DefaultLineEnding,\n\t\tEncodeLevel: zapcore.LowercaseLevelEncoder,\n\t\tEncodeTime: zapcore.ISO8601TimeEncoder,\n\t\tEncodeDuration: zapcore.SecondsDurationEncoder,\n\t\tEncodeCaller: zapcore.ShortCallerEncoder,\n\t}), zapcore.NewMultiWriteSyncer(os.Stdout, w), zap.NewAtomicLevelAt(getLevel(level))))\n\n\tlog.Desugar().Core()\n\n\tlog = customLog.WithOptions(zap.AddCaller(), zap.AddCallerSkip(1)).Sugar()\n}", "func WriteConfiguration() error {\n\tconfigJSON, err := json.Marshal(Config)\n\tif err == nil {\n\t\terr = ioutil.WriteFile(defaultConfigFilePath, configJSON, 0644)\n\t}\n\treturn err\n}", "func WithWriteTimeout(writeTimeout time.Duration) ConfigOption {\n\treturn func(cfg *Config) {\n\t\tcfg.WriteTimeout = writeTimeout\n\t}\n}", "func (opt KubeletRunConfig) MakeNodeConfig(imageRunHelper *run.Runner, out io.Writer) (string, error) {\n\tbinds := append(opt.ContainerBinds)\n\tenv := []string{}\n\tif len(opt.HTTPProxy) > 0 {\n\t\tenv = append(env, fmt.Sprintf(\"HTTP_PROXY=%s\", opt.HTTPProxy))\n\t}\n\tif len(opt.HTTPSProxy) > 0 {\n\t\tenv = append(env, fmt.Sprintf(\"HTTPS_PROXY=%s\", opt.HTTPSProxy))\n\t}\n\tif len(opt.NoProxy) > 0 {\n\t\tenv = append(env, fmt.Sprintf(\"NO_PROXY=%s\", strings.Join(opt.NoProxy, \",\")))\n\t}\n\tif opt.UseSharedVolume {\n\t\tbinds = append(binds, fmt.Sprintf(\"%[1]s:%[1]s:shared\", opt.HostVolumesDir))\n\t\tenv = append(env, \"OPENSHIFT_CONTAINERIZED=false\")\n\t} else {\n\t\tbinds = append(binds, \"/:/rootfs:ro\")\n\t\tbinds = append(binds, fmt.Sprintf(\"%[1]s:%[1]s:rslave\", opt.HostVolumesDir))\n\t}\n\tenv = append(env, opt.Environment...)\n\tbinds = append(binds, fmt.Sprintf(\"%[1]s:%[1]s\", opt.DockerRoot))\n\n\t// Kubelet needs to be able to write to\n\t// /sys/devices/virtual/net/vethXXX/brport/hairpin_mode, so make this rw, not ro.\n\tbinds = append(binds, \"/sys/devices/virtual/net:/sys/devices/virtual/net:rw\")\n\n\tfmt.Fprintf(out, \"Running kubelet\\n\")\n\tcreateConfigCmd := []string{\n\t\t\"kubelet\",\n\t}\n\tcreateConfigCmd = append(createConfigCmd, opt.Args...)\n\n\t_, err := imageRunHelper.Image(opt.NodeImage).\n\t\tName(\"origin\"). // TODO figure out why the rest of cluster-up relies on this name\n\t\tPrivileged().\n\t\tDiscardContainer().\n\t\tHostNetwork().\n\t\tHostPid().\n\t\tBind(binds...).\n\t\tEnv(env...).\n\t\tEntrypoint(\"hyperkube\").\n\t\tCommand(createConfigCmd...).Start()\n\tif err != nil {\n\t\treturn \"\", errors.NewError(\"could not create OpenShift configuration: %v\", err).WithCause(err)\n\t}\n\n\treturn \"\", nil\n}", "func (n *Node) WriteTree(padding ...int) string {\n\tvar indent int\n\tif len(padding) == 1 {\n\t\tindent = padding[0]\n\t}\n\n\tvar s string\n\tif n.val != \"\" {\n\t\tfor i := 0; i < indent; i++ {\n\t\t\ts += \" \"\n\t\t}\n\t\ts += n.key + \" : \" + n.val + \"\\n\"\n\t} else {\n\t\tfor i := 0; i < indent; i++ {\n\t\t\ts += \" \"\n\t\t}\n\t\ts += n.key + \" :\" + \"\\n\"\n\t\tfor _, nn := range n.nodes {\n\t\t\ts += nn.WriteTree(indent + 1)\n\t\t}\n\t}\n\treturn s\n}", "func writeConfigFile() {\n f, err := os.Create(\"../src/config/config.toml\")\n if err != nil {\n log.Panic(\"Could not open config.toml\")\n }\n defer f.Close()\n\n var api = \"http://localhost:3000\" // Placeholder\n var rpc = \"http://localhost:8545\" // Placeholder\n dir, _ := filepath.Abs(filepath.Dir(os.Args[0]))\n\n var s = fmt.Sprintf(`[development]\ngridplus_api = \"%s\"\nrpc_provider = \"%s\"\nserial_no = \"ABCD0101\"\n[wallet]\nkey_path = \"%s/../src/config\"`, api, rpc, dir)\n\n _, err2 := f.WriteString(s)\n if err2 != nil {\n log.Panic(\"Could not write config file\")\n }\n return\n}", "func writeConfigFile(fpath string, config interface{}) error {\n\tfile, err := os.Create(fpath)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer file.Close()\n\tenc := json.NewEncoder(file)\n\treturn enc.Encode(config)\n}", "func writeNewNodeObject(writer *bufio.Writer, newOsmNode *osm.Node) {\n\tnewOsmNode.ID = newNodeID\n\tnewNodeID++\n\n\tdata, err := xml.MarshalIndent(newOsmNode, \" \", \" \")\n\tif err != nil {\n\t\tlog.Fatalf(\"error <%v> at xml.MarshalIndent()\", err)\n\t}\n\t_, err = fmt.Fprintf(writer, \"%s\\n\", string(data))\n\tif err != nil {\n\t\tlog.Fatalf(\"error writing output file: %v\", err)\n\t}\n}", "func (t *TemplateConfig) Write() error {\n\tfp := filepath.Join(t.ProjectPath, \"template.yml\")\n\t// make sure directory exists\n\tif _, err := os.Stat(t.ProjectPath); os.IsNotExist(err) {\n\t\tif err := os.MkdirAll(t.ProjectPath, 0755); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tyml, err := yaml.Marshal(t)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn ioutil.WriteFile(fp, yml, 0644)\n}", "func (srvConf *ServerConfig) WriteConf() error {\n\n\ttype Write struct {\n\t\tHeader *ServerConfig `yaml:\"server\"`\n\t}\n\n\tw := Write{Header: srvConf}\n\n\tdata, err := yaml.Marshal(w)\n\tif err != nil {\n\t\treturn errors.New(\"Error to marshal server configuration: \"+err.Error())\n\t}\n\n\terr = os.WriteFile(\"config/server.tmp.yml\", data, 0644)\n\tif err != nil {\n\t\treturn errors.New(\"Error to write server configuration file: \"+err.Error())\n\t}\n\n\treturn nil\n}", "func WriteConfigDeprecated(configPath vfs.Path, config interface{}, writeOptions ...vfs.WriteOption) error {\n\tdata, err := utils.YamlMarshal(config)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error marshalling configuration: %v\", err)\n\t}\n\n\tcreate := false\n\tfor _, writeOption := range writeOptions {\n\t\tswitch writeOption {\n\t\tcase vfs.WriteOptionCreate:\n\t\t\tcreate = true\n\t\tcase vfs.WriteOptionOnlyIfExists:\n\t\t\t_, err = configPath.ReadFile()\n\t\t\tif err != nil {\n\t\t\t\tif os.IsNotExist(err) {\n\t\t\t\t\treturn fmt.Errorf(\"cannot update configuration file %s: does not exist\", configPath)\n\t\t\t\t}\n\t\t\t\treturn fmt.Errorf(\"error checking if configuration file %s exists already: %v\", configPath, err)\n\t\t\t}\n\t\tdefault:\n\t\t\treturn fmt.Errorf(\"unknown write option: %q\", writeOption)\n\t\t}\n\t}\n\n\tif create {\n\t\terr = configPath.CreateFile(data)\n\t} else {\n\t\terr = configPath.WriteFile(data)\n\t}\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error writing configuration file %s: %v\", configPath, err)\n\t}\n\treturn nil\n}", "func (p *GameTree) writeTree(w *bufio.Writer, n TreeNodeIdx, needs bool, nMov int, nMovPerLine int) (err error) {\n\tdefer u(tr(\"writeTree\"))\n\tif needs == true {\n\t\tif nMov > 0 {\n\t\t\terr = w.WriteByte('\\n')\n\t\t\tnMov = 0\n\t\t}\n\t\terr = w.WriteByte('(')\n\t}\n\tif err == nil {\n\t\tif nMov == nMovPerLine {\n\t\t\terr = w.WriteByte('\\n')\n\t\t\tnMov = 0\n\t\t}\n\t\terr = w.WriteByte(';')\n\t\t// write the node\n\t\ttyp := p.treeNodes[n].TNodType\n\t\tswitch typ {\n\t\tcase GameInfoNode:\n\t\t\t// fmt.Println(\"writing GameInfoNode\\n\")\n\t\t\terr = p.writeProperties(w, n, true)\n\t\tcase InteriorNode:\n\t\t\t// fmt.Println(\"writing InteriorNode\\n\")\n\t\t\terr = p.writeProperties(w, n, false)\n\t\tcase BlackMoveNode:\n\t\t\t_, err = w.WriteString(\"B[\")\n\t\t\t_, err = w.Write(SGFCoords(ah.NodeLoc(p.treeNodes[n].propListOrNodeLoc), p.IsFF4()))\n\t\t\terr = w.WriteByte(']')\n\t\t\tnMov += 1\n\t\tcase WhiteMoveNode:\n\t\t\t_, err = w.WriteString(\"W[\")\n\t\t\t_, err = w.Write(SGFCoords(ah.NodeLoc(p.treeNodes[n].propListOrNodeLoc), p.IsFF4()))\n\t\t\terr = w.WriteByte(']')\n\t\t\tnMov += 1\n\t\tdefault:\n\t\t\tfmt.Println(\"*** unsupported TreeNodeType in writeTree\")\n\t\t\terr = errors.New(\"writeTree: unsupported TreeNodeType\" + strconv.FormatInt(int64(typ), 10))\n\t\t\treturn err\n\t\t}\n\t\tif err == nil {\n\t\t\t// write the children\n\t\t\tlastCh := p.treeNodes[n].Children\n\t\t\tif lastCh != nilTreeNodeIdx && err == nil {\n\t\t\t\tch := p.treeNodes[lastCh].NextSib\n\t\t\t\tchNeeds := (lastCh != ch)\n\t\t\t\terr = p.writeTree(w, ch, chNeeds, nMov, nMovPerLine)\n\t\t\t\tfor ch != lastCh && err == nil {\n\t\t\t\t\tch = p.treeNodes[ch].NextSib\n\t\t\t\t\t//\t\t\t\t\tnMov += 1\n\t\t\t\t\terr = p.writeTree(w, ch, chNeeds, nMov, nMovPerLine)\n\t\t\t\t}\n\t\t\t}\n\t\t\tif (err == nil) && (needs == true) {\n\t\t\t\terr = w.WriteByte(')')\n\t\t\t}\n\t\t}\n\t}\n\treturn err\n}", "func (n *nodeReaderWriter) Write(p []byte) (int, error) {\n\n\t// guarantee that the path exists\n\t_, ok := n.repo.Data[n.path]\n\tif !ok {\n\t\tn.repo.Data[n.path] = []byte{}\n\t}\n\n\t// overwrite the file if we haven't already started writing to it\n\tif !n.writing {\n\t\tn.repo.Data[n.path] = make([]byte, 0)\n\t\tn.writing = true\n\t}\n\n\t// copy the data into the node buffer\n\tcount := 0\n\tstart := n.writeCursor\n\tfor ; n.writeCursor < start+len(p); n.writeCursor++ {\n\t\t// extend the file if needed\n\t\tif len(n.repo.Data) < n.writeCursor+len(p) {\n\t\t\tn.repo.Data[n.path] = append(n.repo.Data[n.path], 0)\n\t\t}\n\t\tn.repo.Data[n.path][n.writeCursor] = p[n.writeCursor-start]\n\t\tcount++\n\t}\n\n\treturn count, nil\n}", "func (cfg *Configuration) WriteConfiguration(path string) error {\n\n\tfile, err := os.Create(path)\n\tif err != nil {\n\t\tlog.WithFields(log.Fields{\"err\": err}).Error(\"os.Create\")\n\t\treturn err\n\t}\n\tdefer file.Close()\n\n\te := json.NewEncoder(file)\n\tif err := e.Encode(cfg); err != nil {\n\t\tlog.WithFields(log.Fields{\"err\": err}).Error(\"json.NewEncoder.Encode\")\n\t\treturn err\n\t}\n\treturn nil\n}", "func writeKubeconfig(kubeconfig []byte, contextName string, path string) error {\n\n\tyamlConfig := models.Kubeconfig{}\n\terr := yaml.Unmarshal(kubeconfig, &yamlConfig)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tyamlConfig.CurrentContext = contextName\n\n\td, err := yaml.Marshal(yamlConfig)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = ioutil.WriteFile(path, d, 0600)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (n *Node) WriteClusterSettings(settings *ClusterSettings) error {\n\ts, err := ksigsyaml.Marshal(*settings)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"failed to encode %s\", clusterSettingsPath)\n\t}\n\terr = n.Command(\n\t\t\"mkdir\", \"-p\", filepath.Dir(clusterSettingsPath),\n\t).Silent().Run()\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"failed to write %s\", clusterSettingsPath)\n\t}\n\tif err := n.WriteFile(clusterSettingsPath, s); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (a *action) NodeConfig(config *pt.ParaNodeAddrConfig) (*types.Receipt, error) {\n\tif !validTitle(config.Title) {\n\t\treturn nil, pt.ErrInvalidTitle\n\t}\n\n\tif !types.IsDappFork(a.exec.GetMainHeight(), pt.ParaX, pt.ForkCommitTx) {\n\t\treturn nil, types.ErrNotSupport\n\t}\n\n\tif config.Op == pt.ParaNodeJoin {\n\t\tif config.Addr != a.fromaddr {\n\t\t\treturn nil, types.ErrFromAddr\n\t\t}\n\t\treturn a.nodeAdd(config)\n\n\t} else if config.Op == pt.ParaNodeQuit {\n\t\tif config.Addr != a.fromaddr {\n\t\t\treturn nil, types.ErrFromAddr\n\t\t}\n\t\treturn a.nodeDelete(config)\n\n\t} else if config.Op == pt.ParaNodeVote {\n\t\treturn a.nodeVote(config)\n\t} else if config.Op == pt.ParaNodeTakeover {\n\t\treturn a.nodeTakeover(config)\n\n\t} else {\n\t\treturn nil, pt.ErrParaUnSupportNodeOper\n\t}\n\n}", "func LogOrWriteConfig(fileName string, cfg *kubeschedulerconfig.KubeSchedulerConfiguration, completedProfiles []kubeschedulerconfig.KubeSchedulerProfile) error {\n\tif !(klog.V(2).Enabled() || len(fileName) > 0) {\n\t\treturn nil\n\t}\n\tcfg.Profiles = completedProfiles\n\n\tconst mediaType = runtime.ContentTypeYAML\n\tinfo, ok := runtime.SerializerInfoForMediaType(kubeschedulerscheme.Codecs.SupportedMediaTypes(), mediaType)\n\tif !ok {\n\t\treturn fmt.Errorf(\"unable to locate encoder -- %q is not a supported media type\", mediaType)\n\t}\n\n\tencoder := kubeschedulerscheme.Codecs.EncoderForVersion(info.Serializer, kubeschedulerconfigv1beta1.SchemeGroupVersion)\n\tif klog.V(2).Enabled() {\n\t\tbytes, err := runtime.Encode(encoder, cfg)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tconfigString := string(bytes)\n\t\tklog.Infof(\"Using component config:\\n%+v\\n\", configString)\n\t}\n\n\tif len(fileName) > 0 {\n\t\tconfigFile, err := os.Create(fileName)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer configFile.Close()\n\t\tif err := encoder.Encode(cfg, configFile); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tklog.Infof(\"Wrote configuration to: %s\\n\", fileName)\n\t\tos.Exit(0)\n\t}\n\treturn nil\n}", "func WriteKubeConfig(configRaw *clientcmdapi.Config) (string, error) {\n\tdata, err := clientcmd.Write(*configRaw)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Create temp file\n\ttempFile, err := os.CreateTemp(\"\", \"\")\n\tif err != nil {\n\t\treturn \"\", errors.Wrap(err, \"create temp file\")\n\t}\n\n\t// Write to temp file\n\t_, err = tempFile.Write(data)\n\tif err != nil {\n\t\tos.Remove(tempFile.Name())\n\t\treturn \"\", errors.Wrap(err, \"write temp file\")\n\t}\n\n\t// Close temp file\n\ttempFile.Close()\n\n\t// Okay sometimes the file is written so quickly that helm somehow\n\t// cannot read it immediately which causes errors\n\t// so we wait here till the file is ready\n\tnow := time.Now()\n\tfor time.Since(now) < time.Minute {\n\t\t_, err = os.Stat(tempFile.Name())\n\t\tif err != nil {\n\t\t\tif os.IsNotExist(err) {\n\t\t\t\ttime.Sleep(time.Millisecond * 50)\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tos.Remove(tempFile.Name())\n\t\t\treturn \"\", err\n\t\t}\n\n\t\tbreak\n\t}\n\n\treturn tempFile.Name(), nil\n}", "func writeScorecardConfig(dir string, cfg v1alpha3.Configuration) error {\n\t// Skip writing if config is empty.\n\tif cfg.Metadata.Name == \"\" {\n\t\treturn nil\n\t}\n\n\tb, err := yaml.Marshal(cfg)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tcfgDir := filepath.Join(dir, filepath.FromSlash(scorecard.DefaultConfigDir))\n\tif err := os.MkdirAll(cfgDir, 0755); err != nil {\n\t\treturn err\n\t}\n\tscorecardConfigPath := filepath.Join(cfgDir, scorecard.ConfigFileName)\n\treturn os.WriteFile(scorecardConfigPath, b, 0666)\n}", "func writeConfiguration(conf configuration) error {\n\thomeDir, err := homedir.Dir()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tconfigFilePath := filepath.Join(homeDir, \".telelog.conf\")\n\n\tf, err := os.Create(configFilePath)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer f.Close()\n\n\tencoder := json.NewEncoder(f)\n\treturn encoder.Encode(conf)\n}", "func (m *VpnConfiguration) Serialize(writer i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.SerializationWriter)(error) {\n err := m.DeviceConfiguration.Serialize(writer)\n if err != nil {\n return err\n }\n if m.GetAuthenticationMethod() != nil {\n cast := (*m.GetAuthenticationMethod()).String()\n err = writer.WriteStringValue(\"authenticationMethod\", &cast)\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"connectionName\", m.GetConnectionName())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"realm\", m.GetRealm())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"role\", m.GetRole())\n if err != nil {\n return err\n }\n }\n if m.GetServers() != nil {\n cast := make([]i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, len(m.GetServers()))\n for i, v := range m.GetServers() {\n if v != nil {\n cast[i] = v.(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable)\n }\n }\n err = writer.WriteCollectionOfObjectValues(\"servers\", cast)\n if err != nil {\n return err\n }\n }\n return nil\n}", "func (c *RemoteConfig) Write(rev string) error {\n\t// Create a local copy of the revision file, so\n\t// it can be copied to the server.\n\tf, err := os.Create(c.file)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdefer f.Close()\n\tdefer os.Remove(c.file)\n\n\t_, err = f.WriteString(rev)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tf.Sync()\n\n\tif err = c.conn.Upload(c.file, c.file); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}" ]
[ "0.5974496", "0.5868787", "0.58139217", "0.57943386", "0.5769414", "0.56910294", "0.5609716", "0.55106694", "0.5498322", "0.5391262", "0.5385172", "0.5366047", "0.5353425", "0.5297318", "0.52684826", "0.525217", "0.52412367", "0.51867384", "0.51828575", "0.51748914", "0.5172059", "0.516985", "0.5166725", "0.51157427", "0.5060997", "0.50569767", "0.5045391", "0.5033026", "0.50064164", "0.49872205", "0.4986322", "0.496698", "0.49589923", "0.49458218", "0.49427077", "0.49223256", "0.49104807", "0.49005994", "0.48961282", "0.488099", "0.48796713", "0.4873399", "0.4872784", "0.4871336", "0.48502284", "0.48412818", "0.4836707", "0.48230305", "0.48044223", "0.47902596", "0.47896957", "0.47791275", "0.4768622", "0.4756396", "0.4741751", "0.47326893", "0.47223783", "0.47111812", "0.4702818", "0.46959355", "0.4690933", "0.46851927", "0.46794903", "0.46733657", "0.46697798", "0.4657833", "0.46450213", "0.46349677", "0.46329793", "0.46235543", "0.46098596", "0.46083006", "0.46022877", "0.45796326", "0.45679456", "0.45665094", "0.45331275", "0.45323068", "0.4523852", "0.45238152", "0.45235553", "0.45222414", "0.45091575", "0.45069557", "0.45063487", "0.45055115", "0.44987065", "0.4494381", "0.4492474", "0.4492462", "0.44708908", "0.44615206", "0.44511673", "0.44326052", "0.44105688", "0.44058123", "0.43911552", "0.43860605", "0.4380459", "0.43740785" ]
0.66832227
0
vlanFilterMacros generates VLAN_FILTER macros which are written to node_config.h
func vlanFilterMacros() (string, error) { devices := make(map[int]bool) for _, device := range option.Config.GetDevices() { ifindex, err := link.GetIfIndex(device) if err != nil { return "", err } devices[int(ifindex)] = true } allowedVlans := make(map[int]bool) for _, vlanId := range option.Config.VLANBPFBypass { allowedVlans[vlanId] = true } // allow all vlan id's if allowedVlans[0] { return "return true", nil } vlansByIfIndex := make(map[int][]int) links, err := netlink.LinkList() if err != nil { return "", err } for _, l := range links { vlan, ok := l.(*netlink.Vlan) // if it's vlan device and we're controlling vlan main device // and either all vlans are allowed, or we're controlling vlan device or vlan is explicitly allowed if ok && devices[vlan.ParentIndex] && (devices[vlan.Index] || allowedVlans[vlan.VlanId]) { vlansByIfIndex[vlan.ParentIndex] = append(vlansByIfIndex[vlan.ParentIndex], vlan.VlanId) } } vlansCount := 0 for _, v := range vlansByIfIndex { vlansCount += len(v) sort.Ints(v) // sort Vlanids in-place since netlink.LinkList() may return them in any order } if vlansCount == 0 { return "return false", nil } else if vlansCount > 5 { return "", fmt.Errorf("allowed VLAN list is too big - %d entries, please use '--vlan-bpf-bypass 0' in order to allow all available VLANs", vlansCount) } else { vlanFilterTmpl := template.Must(template.New("vlanFilter").Parse( `switch (ifindex) { \ {{range $ifindex,$vlans := . -}} case {{$ifindex}}: \ switch (vlan_id) { \ {{range $vlan := $vlans -}} case {{$vlan}}: \ {{end}}return true; \ } \ break; \ {{end}}} \ return false;`)) var vlanFilterMacro bytes.Buffer if err := vlanFilterTmpl.Execute(&vlanFilterMacro, vlansByIfIndex); err != nil { return "", fmt.Errorf("failed to execute template: %q", err) } return vlanFilterMacro.String(), nil } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func RegisterVLANRoutes(prefix string, router *mux.Router) {\n\trouter.HandleFunc(prefix, ListVLANs).Methods(\"GET\")\n\trouter.HandleFunc(prefix, CreateVLAN).Methods(\"POST\")\n\n\tsub := router.PathPrefix(prefix).Subrouter()\n\tsub.HandleFunc(\"/{vlanTag}\", GetVLAN).Methods(\"GET\")\n\tsub.HandleFunc(\"/{vlanTag}\", UpdateVLAN).Methods(\"PATCH\")\n\tsub.HandleFunc(\"/{vlanTag}\", DestroyVLAN).Methods(\"DELETE\")\n\tsub.HandleFunc(\"/{vlanTag}/groups\", GetVLANGroupMembership).Methods(\"GET\")\n\tsub.HandleFunc(\"/{vlanTag}/groups\", UpdateVLANGroupMembership).Methods(\"POST\")\n}", "func buildRoutesForVlan(vlanTableID int, vlanIndex int, gw net.IP) []netlink.Route {\n\treturn []netlink.Route{\n\t\t// Add a direct link route for the pod vlan link only.\n\t\t{\n\t\t\tLinkIndex: vlanIndex,\n\t\t\tDst: &net.IPNet{IP: gw, Mask: net.CIDRMask(32, 32)},\n\t\t\tScope: netlink.SCOPE_LINK,\n\t\t\tTable: vlanTableID,\n\t\t},\n\t\t{\n\t\t\tLinkIndex: vlanIndex,\n\t\t\tDst: &net.IPNet{IP: net.IPv4zero, Mask: net.CIDRMask(0, 32)},\n\t\t\tScope: netlink.SCOPE_UNIVERSE,\n\t\t\tGw: gw,\n\t\t\tTable: vlanTableID,\n\t\t},\n\t}\n}", "func marshalLogFiltersRequestBodyToLogLogFilters(v *LogFiltersRequestBody) *log.LogFilters {\n\tres := &log.LogFilters{}\n\tif v.Name != nil {\n\t\tres.Name = make([]*log.StringFilter, len(v.Name))\n\t\tfor i, val := range v.Name {\n\t\t\tres.Name[i] = marshalStringFilterRequestBodyToLogStringFilter(val)\n\t\t}\n\t}\n\tif v.Start != nil {\n\t\tres.Start = make([]*log.TimeFilter, len(v.Start))\n\t\tfor i, val := range v.Start {\n\t\t\tres.Start[i] = marshalTimeFilterRequestBodyToLogTimeFilter(val)\n\t\t}\n\t}\n\tif v.End != nil {\n\t\tres.End = make([]*log.TimeFilter, len(v.End))\n\t\tfor i, val := range v.End {\n\t\t\tres.End[i] = marshalTimeFilterRequestBodyToLogTimeFilter(val)\n\t\t}\n\t}\n\n\treturn res\n}", "func RegisterFilter(name string, v interface{}) {\n\troot.Filters[name] = v\n}", "func buildVlanLinkName(vlanID int) string {\n\treturn fmt.Sprintf(\"vlan.eth.%d\", vlanID)\n}", "func Filter(r io.Reader, w io.Writer, conf Config) error {\n\ts := bufio.NewScanner(r)\n\tfilters := compile(conf)\n\n\tfor s.Scan() {\n\t\tmetric := s.Bytes()\n\t\tparts := bytes.SplitN(metric, []byte(\"|\"), 2)\n\t\tname := parts[0]\n\n\t\tif len(name) == 0 {\n\t\t\tcontinue\n\t\t}\n\n\t\ttail := parts[1]\n\n\t\tfor regexp, v := range filters {\n\t\t\tmatches := regexp.FindAllSubmatch(name, -1)\n\n\t\t\tif matches == nil {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tswitch v.(type) {\n\t\t\tcase string:\n\t\t\t\tw.Write(replace([]byte(v.(string)), matches))\n\t\t\t\tw.Write([]byte(\"|\"))\n\t\t\t\tw.Write(tail)\n\t\t\t\tw.Write([]byte(\"\\n\"))\n\t\t\tcase bool:\n\t\t\t\tw.Write(metric)\n\t\t\t\tw.Write([]byte(\"\\n\"))\n\t\t\t}\n\t\t}\n\t}\n\n\treturn s.Err()\n}", "func (hPtr *GbgmlogsHandler) Vmodule(ruleset string) error {\n\tvar filter []pattern\n\tfor _, rule := range strings.Split(ruleset, \",\") {\n\t\t// Empty strings such as from a trailing comma can be ignored\n\t\tif len(rule) == 0 {\n\t\t\tcontinue\n\t\t}\n\t\t// Ensure we have a pattern = level filter rule\n\t\tparts := strings.Split(rule, \"=\")\n\t\tif len(parts) != 2 {\n\t\t\treturn errVmoduleSyntax\n\t\t}\n\t\tparts[0] = strings.TrimSpace(parts[0])\n\t\tparts[1] = strings.TrimSpace(parts[1])\n\t\tif len(parts[0]) == 0 || len(parts[1]) == 0 {\n\t\t\treturn errVmoduleSyntax\n\t\t}\n\t\t// Parse the level and if correct, assemble the filter rule\n\t\tlevel, err := strconv.Atoi(parts[1])\n\t\tif err != nil {\n\t\t\treturn errVmoduleSyntax\n\t\t}\n\t\tif level <= 0 {\n\t\t\tcontinue // Ignore. It's harmless but no point in paying the overhead.\n\t\t}\n\t\t// Compile the rule pattern into a regular expression\n\t\tmatcher := \".*\"\n\t\tfor _, comp := range strings.Split(parts[0], \"/\") {\n\t\t\tif comp == \"*\" {\n\t\t\t\tmatcher += \"(/.*)?\"\n\t\t\t} else if comp != \"\" {\n\t\t\t\tmatcher += \"/\" + regexp.QuoteMeta(comp)\n\t\t\t}\n\t\t}\n\t\tif !strings.HasSuffix(parts[0], \".go\") {\n\t\t\tmatcher += \"/[^/]+\\\\.go\"\n\t\t}\n\t\tmatcher = matcher + \"$\"\n\n\t\tre, _ := regexp.Compile(matcher)\n\t\tfilter = append(filter, pattern{re, Lvl(level)})\n\t}\n\t// Swap out the vmodule pattern for the new filter system\n\thPtr.lock.Lock()\n\tdefer hPtr.lock.Unlock()\n\n\thPtr.patterns = filter\n\thPtr.siteCache = make(map[uintptr]Lvl)\n\tatomicPtr.StoreUint32(&hPtr.override, uint32(len(filter)))\n\n\treturn nil\n}", "func buildVlanLink(vlanName string, vlanID int, parentIfIndex int, eniMAC string) *netlink.Vlan {\n\tla := netlink.NewLinkAttrs()\n\tla.Name = vlanName\n\tla.ParentIndex = parentIfIndex\n\tla.HardwareAddr, _ = net.ParseMAC(eniMAC)\n\treturn &netlink.Vlan{LinkAttrs: la, VlanId: vlanID}\n}", "func (scanner *SchemaScanner) AddFilters(filters []string) {\n\tscanner.Filters = append(scanner.Filters, filters...)\n}", "func (e *DefaultExecutor) makeFiltersList() string {\n\tbuf := new(bytes.Buffer)\n\tw := tabwriter.NewWriter(buf, 5, 0, 1, ' ', 0)\n\n\tfmt.Fprintln(w, \"FILTER\\tENABLED\\tDESCRIPTION\")\n\tfor _, filter := range e.filterEngine.RegisteredFilters() {\n\t\tfmt.Fprintf(w, \"%s\\t%v\\t%s\\n\", filter.Name(), filter.Enabled, filter.Describe())\n\t}\n\n\tw.Flush()\n\treturn buf.String()\n}", "func devMacros() (string, string, error) {\n\tvar (\n\t\tmacByIfIndexMacro, isL3DevMacroBuf bytes.Buffer\n\t\tisL3DevMacro string\n\t)\n\tmacByIfIndex := make(map[int]string)\n\tl3DevIfIndices := make([]int, 0)\n\n\tfor _, iface := range option.Config.GetDevices() {\n\t\tlink, err := netlink.LinkByName(iface)\n\t\tif err != nil {\n\t\t\treturn \"\", \"\", fmt.Errorf(\"failed to retrieve link %s by name: %q\", iface, err)\n\t\t}\n\t\tidx := link.Attrs().Index\n\t\tm := link.Attrs().HardwareAddr\n\t\tif m == nil || len(m) != 6 {\n\t\t\tl3DevIfIndices = append(l3DevIfIndices, idx)\n\t\t}\n\t\tmacByIfIndex[idx] = mac.CArrayString(m)\n\t}\n\n\tmacByIfindexTmpl := template.Must(template.New(\"macByIfIndex\").Parse(\n\t\t`({ \\\nunion macaddr __mac = {.addr = {0x0, 0x0, 0x0, 0x0, 0x0, 0x0}}; \\\nswitch (IFINDEX) { \\\n{{range $idx,$mac := .}} case {{$idx}}: {union macaddr __tmp = {.addr = {{$mac}}}; __mac=__tmp;} break; \\\n{{end}}} \\\n__mac; })`))\n\n\tif err := macByIfindexTmpl.Execute(&macByIfIndexMacro, macByIfIndex); err != nil {\n\t\treturn \"\", \"\", fmt.Errorf(\"failed to execute template: %q\", err)\n\t}\n\n\tif len(l3DevIfIndices) == 0 {\n\t\tisL3DevMacro = \"false\"\n\t} else {\n\t\tisL3DevTmpl := template.Must(template.New(\"isL3Dev\").Parse(\n\t\t\t`({ \\\nbool is_l3 = false; \\\nswitch (ifindex) { \\\n{{range $idx := .}} case {{$idx}}: is_l3 = true; break; \\\n{{end}}} \\\nis_l3; })`))\n\t\tif err := isL3DevTmpl.Execute(&isL3DevMacroBuf, l3DevIfIndices); err != nil {\n\t\t\treturn \"\", \"\", fmt.Errorf(\"failed to execute template: %q\", err)\n\t\t}\n\t\tisL3DevMacro = isL3DevMacroBuf.String()\n\t}\n\n\treturn macByIfIndexMacro.String(), isL3DevMacro, nil\n}", "func newThresholdFilterParser(req types.AccessRequest, rev types.AccessReview, author types.User) (BoolPredicateParser, error) {\n\treturn NewJSONBoolParser(thresholdFilterContext{\n\t\tReviewer: reviewAuthorContext{\n\t\t\tRoles: author.GetRoles(),\n\t\t\tTraits: author.GetTraits(),\n\t\t},\n\t\tReview: reviewParamsContext{\n\t\t\tReason: rev.Reason,\n\t\t\tAnnotations: rev.Annotations,\n\t\t},\n\t\tRequest: reviewRequestContext{\n\t\t\tRoles: req.GetOriginalRoles(),\n\t\t\tReason: req.GetRequestReason(),\n\t\t\tSystemAnnotations: req.GetSystemAnnotations(),\n\t\t},\n\t})\n}", "func PacketFilter(custom, ports, ips string) (string, error) {\n\tvar fts, pfs, sfs []string\n\n\t// Port filter\n\tfor _, port := range strings.Split(ports, \",\") {\n\t\tif len(port) != 0 {\n\t\t\tpfs = append(pfs, fmt.Sprintf(\"(port %s)\", port))\n\t\t}\n\t}\n\tfts = AddFilters(fts, pfs)\n\n\t// IP filter\n\tfor _, ip := range strings.Split(ips, \",\") {\n\t\tif len(ip) != 0 {\n\t\t\tsfs = append(sfs, fmt.Sprintf(\"(host %s)\", ip))\n\t\t}\n\t}\n\tfts = AddFilters(fts, sfs)\n\n\t// Custom filter\n\tif custom != \"\" {\n\t\tfts = AddFilters(fts, []string{fmt.Sprintf(\"(%s)\", custom)})\n\t}\n\n\tfor i := range fts {\n\t\tfts[i] = fmt.Sprintf(\"(%s)\", fts[i])\n\t}\n\n\treturn strings.Join(fts, \" or \"), nil\n}", "func marshalVlan(info *VLan) ([]byte, error) {\n\toptions := []tcOption{}\n\n\tif info == nil {\n\t\treturn []byte{}, fmt.Errorf(\"VLan: %w\", ErrNoArg)\n\t}\n\t// TODO: improve logic and check combinations\n\tif info.Tm != nil {\n\t\treturn []byte{}, ErrNoArgAlter\n\t}\n\tif info.Parms != nil {\n\t\tdata, err := marshalStruct(info.Parms)\n\t\tif err != nil {\n\t\t\treturn []byte{}, err\n\t\t}\n\t\toptions = append(options, tcOption{Interpretation: vtBytes, Type: tcaVLanParms, Data: data})\n\t}\n\tif info.PushID != nil {\n\t\toptions = append(options, tcOption{Interpretation: vtUint16, Type: tcaVLanPushVLanID, Data: *info.PushID})\n\t}\n\tif info.PushProtocol != nil {\n\t\toptions = append(options, tcOption{Interpretation: vtUint16, Type: tcaVLanPushVLanProtocol, Data: *info.PushProtocol})\n\t}\n\tif info.PushPriority != nil {\n\t\toptions = append(options, tcOption{Interpretation: vtUint32, Type: tcaVLanPushVLanPriority, Data: *info.PushPriority})\n\t}\n\treturn marshalAttributes(options)\n}", "func listFilterIfs(filter func(netlink.Link) int) (map[int]netlink.Link, error) {\n\tifs, err := netlink.LinkList()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvethLXCIdxs := map[int]netlink.Link{}\n\tfor _, intf := range ifs {\n\t\tif idx := filter(intf); idx != -1 {\n\t\t\tvethLXCIdxs[idx] = intf\n\t\t}\n\t}\n\treturn vethLXCIdxs, nil\n}", "func (m *Config_StatusCodeFilter) Validate() error {\n\tif m == nil {\n\t\treturn nil\n\t}\n\n\tswitch m.FilterType.(type) {\n\n\tcase *Config_StatusCodeFilter_Equals:\n\n\t\tif val := m.GetEquals(); val < 0 || val > 16 {\n\t\t\treturn Config_StatusCodeFilterValidationError{\n\t\t\t\tfield: \"Equals\",\n\t\t\t\treason: \"value must be inside range [0, 16]\",\n\t\t\t}\n\t\t}\n\n\t}\n\n\treturn nil\n}", "func tplTubeFilter() string {\n\tvar buf, currents, others bytes.Buffer\n\tfor k, current := range tubeStatFields {\n\t\tif k > 7 {\n\t\t\tcontinue\n\t\t}\n\t\tfor property, description := range current {\n\t\t\tstatus := \"\"\n\t\t\tif checkInSlice(selfConf.TubeFilters, property) {\n\t\t\t\tstatus = `checked`\n\t\t\t}\n\t\t\tcurrents.WriteString(`<div class=\"form-group\"><div class=\"checkbox\"><label class=\"checkbox\"><input type=\"checkbox\" name=\"`)\n\t\t\tcurrents.WriteString(property)\n\t\t\tcurrents.WriteString(`\" `)\n\t\t\tcurrents.WriteString(status)\n\t\t\tcurrents.WriteString(`><b>`)\n\t\t\tcurrents.WriteString(property)\n\t\t\tcurrents.WriteString(`</b><br/>`)\n\t\t\tcurrents.WriteString(description)\n\t\t\tcurrents.WriteString(`</label></div></div>`)\n\t\t}\n\t}\n\n\tfor k, other := range tubeStatFields {\n\t\tif k < 8 {\n\t\t\tcontinue\n\t\t}\n\t\tfor property, description := range other {\n\t\t\tstatus := \"\"\n\t\t\tif checkInSlice(selfConf.TubeFilters, property) {\n\t\t\t\tstatus = `checked`\n\t\t\t}\n\t\t\tothers.WriteString(`<div class=\"form-group\"><div class=\"checkbox\"><label class=\"checkbox\"><input type=\"checkbox\" name=\"`)\n\t\t\tothers.WriteString(property)\n\t\t\tothers.WriteString(`\" `)\n\t\t\tothers.WriteString(status)\n\t\t\tothers.WriteString(`><b>`)\n\t\t\tothers.WriteString(property)\n\t\t\tothers.WriteString(`</b><br/>`)\n\t\t\tothers.WriteString(description)\n\t\t\tothers.WriteString(`</label></div></div>`)\n\t\t}\n\t}\n\tbuf.WriteString(`<div id=\"filter\" data-cookie=\"tubefilter\" class=\"modal fade\" tabindex=\"-1\" role=\"dialog\"><div class=\"modal-dialog\"><div class=\"modal-content\"><div class=\"modal-header\"><button type=\"button\" class=\"close\" data-dismiss=\"modal\" aria-hidden=\"true\">×</button><h4 class=\"modal-title\" id=\"filter-columns-label\">Filter columns</h4></div><div class=\"modal-body\"><form><div class=\"tabbable\"><ul class=\"nav nav-tabs\"><li class=\"active\"><a href=\"#current\" data-toggle=\"tab\">current</a></li><li><a href=\"#other\" data-toggle=\"tab\">other</a></li></ul><div class=\"tab-content\"><div class=\"tab-pane active\" id=\"current\">`)\n\tbuf.WriteString(currents.String())\n\tbuf.WriteString(`</div><div class=\"tab-pane\" id=\"other\">`)\n\tbuf.WriteString(others.String())\n\tbuf.WriteString(`</div></div></div></form></div><div class=\"modal-footer\"><button class=\"btn btn-success\" data-dismiss=\"modal\" aria-hidden=\"true\">Close</button></div></div></div></div>`)\n\treturn buf.String()\n}", "func marshalLogLogFiltersToLogFiltersRequestBody(v *log.LogFilters) *LogFiltersRequestBody {\n\tres := &LogFiltersRequestBody{}\n\tif v.Name != nil {\n\t\tres.Name = make([]*StringFilterRequestBody, len(v.Name))\n\t\tfor i, val := range v.Name {\n\t\t\tres.Name[i] = marshalLogStringFilterToStringFilterRequestBody(val)\n\t\t}\n\t}\n\tif v.Start != nil {\n\t\tres.Start = make([]*TimeFilterRequestBody, len(v.Start))\n\t\tfor i, val := range v.Start {\n\t\t\tres.Start[i] = marshalLogTimeFilterToTimeFilterRequestBody(val)\n\t\t}\n\t}\n\tif v.End != nil {\n\t\tres.End = make([]*TimeFilterRequestBody, len(v.End))\n\t\tfor i, val := range v.End {\n\t\t\tres.End[i] = marshalLogTimeFilterToTimeFilterRequestBody(val)\n\t\t}\n\t}\n\n\treturn res\n}", "func generateListFilterFromLabels(labels map[string]string) string {\n\tvar filter string\n\tfor k, v := range labels {\n\t\tfilter = fmt.Sprintf(\"%s(labels.%s eq %s)\", filter, k, v)\n\t}\n\n\treturn filter\n}", "func (qb *QueryBuilder) buildFilters() {\n\t// todo: loop and yaml conf. but this one is faster as no loop, will try array switch\n\tsearchParams := GetFilterConfig()\n\tfor key, val := range searchParams {\n\t\tm2 := val.(map[string]interface{})\n\t\tpk := m2[\"filter\"].(map[string]interface{})\n\t\tpkey := fmt.Sprintf(\"%v\", pk[\"key\"])\n\t\t//log.Println(key, pkey)\n\t\t//log.Println(key, m2[\"filter\"])\n\t\tqb.addMustFilterByTerm(key, pkey)\n\t}\n}", "func genPresentMatchHeaderFilter(headerName string) *access_logv3.AccessLogFilter_HeaderFilter {\n\treturn &access_logv3.AccessLogFilter_HeaderFilter{\n\t\tHeaderFilter: &access_logv3.HeaderFilter{\n\t\t\tHeader: &envoy_config_route_v3.HeaderMatcher{\n\t\t\t\tName: soapActionHeaderName,\n\t\t\t\tHeaderMatchSpecifier: &envoy_config_route_v3.HeaderMatcher_PresentMatch{\n\t\t\t\t\tPresentMatch: true,\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n}", "func VLANChanged(existingVLANIDs, updatedVLANIDs []int) bool {\n\tsort.Ints(existingVLANIDs)\n\tsort.Ints(updatedVLANIDs)\n\treturn !reflect.DeepEqual(existingVLANIDs, updatedVLANIDs)\n}", "func genExactMatchHeaderFilter(headerName, headerValue string) *access_logv3.AccessLogFilter_HeaderFilter {\n\treturn &access_logv3.AccessLogFilter_HeaderFilter{\n\t\tHeaderFilter: &access_logv3.HeaderFilter{\n\t\t\tHeader: &envoy_config_route_v3.HeaderMatcher{\n\t\t\t\tName: headerName,\n\t\t\t\tHeaderMatchSpecifier: &envoy_config_route_v3.HeaderMatcher_StringMatch{\n\t\t\t\t\tStringMatch: &envoy_type_matcher_v3.StringMatcher{\n\t\t\t\t\t\tMatchPattern: &envoy_type_matcher_v3.StringMatcher_Exact{\n\t\t\t\t\t\t\tExact: headerValue,\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n}", "func VerboseFilter(hdr *tar.Header) bool {\n\tfmt.Println(hdr.Name)\n\treturn true\n}", "func VLANIDs(officeConfig *ocstruct.Office) []int {\n\tvlanIDs := []int{}\n\n\tif officeConfig == nil {\n\t\treturn vlanIDs\n\t}\n\n\tfor _, ap := range officeConfig.OfficeAp {\n\t\twlans := ap.Ssids\n\t\tif wlans == nil || len(wlans.Ssid) == 0 {\n\t\t\tcontinue\n\t\t}\n\n\t\tfor _, wlan := range wlans.Ssid {\n\t\t\tvlanIDs = append(vlanIDs, int(*wlan.Config.VlanId))\n\t\t}\n\t}\n\n\treturn vlanIDs\n}", "func GetVlans(givenIp string) ([]*Vlan, error) {\n\tconst BUCKET = \"netops-dev\"\n\tclient, err := gocb.Connect(\"http://localhost:8091/\")\n\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tbucket, err := client.OpenBucket(BUCKET, BUCKET)\n\tif err != nil {\n\t\tlog.Fatal(\"when connecting to couchbase: \", err)\n\t}\n\tif bucket == nil {\n\t\tlog.Fatal(\"couchbase connection went wrong\")\n\t}\n\tbucket.Manager(\"\", \"\").CreatePrimaryIndex(\"\", true, false)\n\n\t// Remove all data from the bucket\n\t{\n\t\t_, err := bucket.ExecuteN1qlQuery(gocb.NewN1qlQuery(\"delete from `netops-dev`\"), []interface{}{})\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t}\n\n\tipnetsStr := []string{\n\t\t\"192.168.1.1/28\",\n\t\t\"192.168.1.28/28\",\n\t\t\"192.168.1.80/28\",\n\t\t\"192.168.1.250/28\",\n\t}\n\n\t// Turn ip net strings into a slice of IPNet\n\tipnets := make(map[string]*net.IPNet)\n\tfor _, ipStr := range ipnetsStr {\n\t\t_, ipnet, err := net.ParseCIDR(ipStr)\n\t\tif err != nil {\n\t\t\tlog.Printf(\"skipping '%v' as it doesn't seem to be a valid CIDR (RFC 4632, RFC 4291)\", ipStr)\n\t\t\tcontinue\n\t\t}\n\t\tipnets[ipStr] = ipnet\n\t}\n\n\tfor str, ipnet := range ipnets {\n\t\tbucket.Upsert(str,\n\t\t\t&Vlan{str, ip2int(ipnet.IP), ip2int(net.IP(ipnet.Mask)), *ipnet}, 0,\n\t\t)\n\t}\n\n\t{\n\t\trows, err := bucket.ExecuteN1qlQuery(gocb.NewN1qlQuery(\n\t\t\t\"select maskIpv4, subnetIpv4, vlanId \"+\n\t\t\t\t\"from `netops-dev`\"), []interface{}{})\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t\tvar vlan Vlan\n\t\tfor rows.Next(&vlan) {\n\t\t\tvlan.IPNet = net.IPNet{\n\t\t\t\tIP: int2ip(vlan.SubnetIpv4),\n\t\t\t\tMask: net.IPMask(int2ip(vlan.MaskIpv4)),\n\t\t\t}\n\t\t\tfmt.Printf(\"%v\\n\", vlan.IPNet)\n\t\t}\n\t}\n\n\tipStr := givenIp\n\tip := ip2int(net.ParseIP(ipStr))\n\trows, err := bucket.ExecuteN1qlQuery(gocb.NewN1qlQuery(\n\t\t\"select maskIpv4, subnetIpv4, vlanId \"+\n\t\t\t\"from `netops-dev` \"+\n\t\t\t\"where bitand($1, maskIpv4) = subnetIpv4\"), []interface{}{ip})\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tvar vlan Vlan\n\tvar vlans []Vlan\n\tfor rows.Next(&vlan) {\n\t\tvlan.IPNet = net.IPNet{\n\t\t\tIP: int2ip(vlan.SubnetIpv4),\n\t\t\tMask: net.IPMask(int2ip(vlan.MaskIpv4)),\n\t\t}\n\t\tvlans = append(vlans, vlan)\n\t}\n\tfmt.Printf(\"lookup for '%v': \", ipStr)\n\tfor _, vlan := range vlans {\n\t\tfmt.Printf(\"%v \", vlan.IPNet.String())\n\t}\n\tfmt.Print(\"\\n\")\n\treturn nil, nil\n}", "func AddFilters(handlers *beego.ControllerRegister) {\n\thandlers.InsertFilter(\"*\", beego.BeforeStatic, beforeStatic, false)\n}", "func InitTableFilters(filters []*component.TableFilter) {\n\tfor _, f := range filters {\n\t\tsort.Strings(f.Values)\n\t}\n}", "func createFilters(fr filters.Registry, defs []*eskip.Filter, cpm map[string]PredicateSpec) ([]*RouteFilter, error) {\n\tvar fs []*RouteFilter\n\tfor i, def := range defs {\n\t\tf, err := createFilter(fr, def, cpm)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tfs = append(fs, &RouteFilter{f, def.Name, i})\n\t}\n\n\treturn fs, nil\n}", "func TagFilter_STATUSGenerator() gopter.Gen {\n\tif tagFilter_STATUSGenerator != nil {\n\t\treturn tagFilter_STATUSGenerator\n\t}\n\n\tgenerators := make(map[string]gopter.Gen)\n\tAddIndependentPropertyGeneratorsForTagFilter_STATUS(generators)\n\ttagFilter_STATUSGenerator = gen.Struct(reflect.TypeOf(TagFilter_STATUS{}), generators)\n\n\treturn tagFilter_STATUSGenerator\n}", "func ListVLANs(w http.ResponseWriter, r *http.Request) {\n\thr := HTTPResponse{w}\n\tctx := GetContext(r)\n\tvlans := make(lochness.VLANs, 0)\n\terr := ctx.ForEachVLAN(func(vlan *lochness.VLAN) error {\n\t\tvlans = append(vlans, vlan)\n\t\treturn nil\n\t})\n\tif err != nil && !ctx.IsKeyNotFound(err) {\n\t\thr.JSONError(http.StatusInternalServerError, err)\n\t\treturn\n\t}\n\thr.JSON(http.StatusOK, vlans)\n}", "func VerboseLogFilter(hdr *tar.Header) bool {\n\tlog.Println(hdr.Name)\n\treturn true\n}", "func configureFlags(api *operations.ReservoirAPI) {\n\t// api.CommandLineOptionsGroups = []swag.CommandLineOptionsGroup{ ... }\n}", "func buildFilterData(columns []ColumnSchema, index int, filters []Filters) []Filters {\n\tfilters = append(filters, Filters{\n\t\tId: columns[index].Field,\n\t\tLabel: columns[index].Name,\n\t\tType: columns[index].Type,\n\t\tInput: buildInput(columns[index].Type, columns[index].Values),\n\t\tValues: buildValues(columns[index].Values),\n\t\tOperators: buildOperations(columns[index].Type),\n\t\tPlugin: buildPlug(columns[index].Type),\n\t\tPlugin_config: buildPlugConfig(columns[index].Field,columns[index].Type),\n\t})\n\treturn filters\n}", "func TestAccGetVLANCustomFieldsSchema(t *testing.T) {\n\ttestacc.VetAccConditions(t)\n\n\tsess := session.NewSession()\n\tclient := NewController(sess)\n\n\texpected := testGetVLANCustomFieldsSchemaExpected\n\tactual, err := client.GetVLANCustomFieldsSchema()\n\tif err != nil {\n\t\tt.Fatalf(\"Bad: %s\", err)\n\t}\n\n\tif !reflect.DeepEqual(expected, actual) {\n\t\tt.Fatalf(\"Expected %#v, got %#v\", expected, actual)\n\t}\n}", "func (ChannelStrips) V76Preamplifier(){}", "func regexFilter(attr, val string) map[string]interface{} {\n\treturn map[string]interface{}{\n\t\t\"regexp\": map[string]string{\n\t\t\tattr: val,\n\t\t},\n\t}\n}", "func TestAccVLANCRUD(t *testing.T) {\n\ttestacc.VetAccConditions(t)\n\n\tsess := session.NewSession()\n\tvlan := testCreateVLANInput\n\tif os.Getenv(\"TESTACC_CUSTOM_NESTED\") != \"\" {\n\t\tvlan.CustomFields = map[string]interface{}{\n\t\t\t\"CustomTestVLANs\": \"foobar\",\n\t\t\t\"CustomTestVLANs2\": nil,\n\t\t}\n\t} else {\n\t\tlog.Println(\"Note: Not testing nested custom fields as TESTACC_CUSTOM_NESTED is not set\")\n\t}\n\ttestAccVLANCRUDCreate(t, sess, vlan)\n\t// Add the domain ID here as 1 is the default.\n\tvlan.DomainID = 1\n\tvlan.ID = testAccVLANCRUDReadByNumber(t, sess, vlan)\n\ttestAccVLANCRUDReadByID(t, sess, vlan)\n\tvlan.Name = \"bazlan\"\n\ttestAccVLANCRUDUpdate(t, sess, vlan)\n\ttestAccVLANCRUDDelete(t, sess, vlan)\n}", "func genSSMFilters(filters string) ([]*ssm.InstanceInformationStringFilter, error) {\n\tvar filtersList []string\n\tif filters != \"\" {\n\t\tfiltersList = strings.Split(filters, \",\")\n\t}\n\n\tssmFilters := make([]*ssm.InstanceInformationStringFilter, len(filtersList)+1)\n\tssmFilters[0] = &ssm.InstanceInformationStringFilter{\n\t\tKey: aws.String(\"PingStatus\"),\n\t\tValues: []*string{aws.String(\"Online\")},\n\t}\n\n\tif len(filtersList) == 0 {\n\t\treturn ssmFilters, nil\n\t}\n\n\ti := 1\n\tfor _, filter := range filtersList {\n\t\tfilterInfo := strings.Split(filter, \"=\")\n\n\t\tif len(filterInfo) != 2 {\n\t\t\treturn ssmFilters, fmt.Errorf(\"Invalid filter \\\"%s\\\", exect Key=Value\", filter)\n\t\t}\n\n\t\tif filterInfo[0] == \"PingStatus\" {\n\t\t\treturn ssmFilters, fmt.Errorf(\"PingStatus is static filter, cannot be altererd\")\n\t\t}\n\n\t\tssmFilters[i] = &ssm.InstanceInformationStringFilter{\n\t\t\tKey: aws.String(filterInfo[0]),\n\t\t\tValues: []*string{aws.String(filterInfo[1])},\n\t\t}\n\n\t\ti++\n\t}\n\n\treturn ssmFilters, nil\n}", "func (l *Libvirt) NwfilterDefineXML(XML string) (rOptNwfilter Nwfilter, err error) {\n\tvar buf []byte\n\n\targs := NwfilterDefineXMLArgs {\n\t\tXML: XML,\n\t}\n\n\tbuf, err = encode(&args)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tvar r response\n\tr, err = l.requestStream(180, constants.Program, buf, nil, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\t// Return value unmarshaling\n\ttpd := typedParamDecoder{}\n\tct := map[string]xdr.TypeDecoder{\"libvirt.TypedParam\": tpd}\n\trdr := bytes.NewReader(r.Payload)\n\tdec := xdr.NewDecoderCustomTypes(rdr, 0, ct)\n\t// OptNwfilter: Nwfilter\n\t_, err = dec.Decode(&rOptNwfilter)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func ParseConfigurationFilters(values map[string]interface{}) (Filters, error) {\n\tfilters := Filters{}\n\tfor filterType, rawValue := range values {\n\t\tvalue, err := filterValue(rawValue)\n\t\tif err != nil {\n\t\t\treturn []*Filter{}, err\n\t\t}\n\t\tfilter, err := MakeFilter(filterType, value)\n\t\tif err != nil {\n\t\t\treturn []*Filter{}, err\n\t\t}\n\t\tfilters = append(filters, filter)\n\t}\n\treturn filters, nil\n}", "func TwigFilters() map[string]stick.Filter {\n\treturn map[string]stick.Filter{\n\t\t\"abs\": filterAbs,\n\t\t\"default\": filterDefault,\n\t\t\"batch\": filterBatch,\n\t\t\"capitalize\": filterCapitalize,\n\t\t\"convert_encoding\": filterConvertEncoding,\n\t\t\"date\": filterDate,\n\t\t\"date_modify\": filterDateModify,\n\t\t\"first\": filterFirst,\n\t\t\"format\": filterFormat,\n\t\t\"join\": filterJoin,\n\t\t\"json_encode\": filterJSONEncode,\n\t\t\"keys\": filterKeys,\n\t\t\"last\": filterLast,\n\t\t\"length\": filterLength,\n\t\t\"lower\": filterLower,\n\t\t\"merge\": filterMerge,\n\t\t\"nl2br\": filterNL2BR,\n\t\t\"number_format\": filterNumberFormat,\n\t\t\"raw\": filterRaw,\n\t\t\"replace\": filterReplace,\n\t\t\"reverse\": filterReverse,\n\t\t\"round\": filterRound,\n\t\t\"slice\": filterSlice,\n\t\t\"sort\": filterSort,\n\t\t\"split\": filterSplit,\n\t\t\"striptags\": filterStripTags,\n\t\t\"title\": filterTitle,\n\t\t\"trim\": filterTrim,\n\t\t\"upper\": filterUpper,\n\t\t\"url_encode\": filterURLEncode,\n\n\t\t// custom\n\t\t\"get\": filterGet,\n\t\t\"dateTime\": filterDateTime,\n\t\t\"time\": filterTime,\n\t}\n}", "func (n *linuxNetwork) setupVlan(vlanID int, eniMAC string, subnetGW string, parentIfIndex int, rtTable int, log logger.Logger) (netlink.Link, error) {\n\tvlanLinkName := buildVlanLinkName(vlanID)\n\t// 1. clean up if vlan already exists (necessary when trunk ENI changes).\n\tif oldVlan, err := n.netLink.LinkByName(vlanLinkName); err == nil {\n\t\tif err := n.netLink.LinkDel(oldVlan); err != nil {\n\t\t\treturn nil, errors.Wrapf(err, \"failed to delete old vlan link %s\", vlanLinkName)\n\t\t}\n\t\tlog.Debugf(\"Successfully deleted old vlan link: %s\", vlanLinkName)\n\t}\n\n\t// 2. add new vlan link\n\tvlanLink := buildVlanLink(vlanLinkName, vlanID, parentIfIndex, eniMAC)\n\tif err := n.netLink.LinkAdd(vlanLink); err != nil {\n\t\treturn nil, errors.Wrapf(err, \"failed to add vlan link %s\", vlanLinkName)\n\t}\n\n\t// 3. bring up the vlan\n\tif err := n.netLink.LinkSetUp(vlanLink); err != nil {\n\t\treturn nil, errors.Wrapf(err, \"failed to setUp vlan link %s\", vlanLinkName)\n\t}\n\n\t// 4. create default routes for vlan\n\troutes := buildRoutesForVlan(rtTable, vlanLink.Index, net.ParseIP(subnetGW))\n\tfor _, r := range routes {\n\t\tif err := n.netLink.RouteReplace(&r); err != nil {\n\t\t\treturn nil, errors.Wrapf(err, \"failed to replace route entry %s via %s\", r.Dst.IP.String(), subnetGW)\n\t\t}\n\t}\n\treturn vlanLink, nil\n}", "func FilterMap() map[string]string {\n\treturn map[string]string{\n\t\t\"entry_id\": \"ann.entry_id\",\n\t\t\"value\": \"ann.value\",\n\t\t\"created_by\": \"ann.created_by\",\n\t\t\"version\": \"ann.version\",\n\t\t\"rank\": \"ann.rank\",\n\t\t\"tag\": \"cvt.label\",\n\t\t\"ontology\": \"cv.metadata.namespace\",\n\t}\n}", "func TagsFilter(t map[string]string) Filter {\n\tj := tagsEncoder(t)\n\treturn Param(\"tags\", j)\n}", "func aclFilter(direction string, externalIDs map[string]string) func(acl *ovnnb.ACL) bool {\n\treturn func(acl *ovnnb.ACL) bool {\n\t\tif len(acl.ExternalIDs) < len(externalIDs) {\n\t\t\treturn false\n\t\t}\n\n\t\tif len(acl.ExternalIDs) != 0 {\n\t\t\tfor k, v := range externalIDs {\n\t\t\t\t// if only key exist but not value in externalIDs, we should include this lsp,\n\t\t\t\t// it's equal to shell command `ovn-nbctl --columns=xx find acl external_ids:key!=\\\"\\\"`\n\t\t\t\tif len(v) == 0 {\n\t\t\t\t\tif len(acl.ExternalIDs[k]) == 0 {\n\t\t\t\t\t\treturn false\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\tif acl.ExternalIDs[k] != v {\n\t\t\t\t\t\treturn false\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\tif len(direction) != 0 && acl.Direction != direction {\n\t\t\treturn false\n\t\t}\n\n\t\treturn true\n\t}\n}", "func createFilter(m eh.EventMatcher) string {\n\tswitch m := m.(type) {\n\tcase eh.MatchEvents:\n\t\ts := make([]string, len(m))\n\t\tfor i, et := range m {\n\t\t\ts[i] = fmt.Sprintf(`attributes:\"%s\"`, et) // Filter event types by key to save space.\n\t\t}\n\n\t\treturn strings.Join(s, \" OR \")\n\tcase eh.MatchAggregates:\n\t\ts := make([]string, len(m))\n\t\tfor i, at := range m {\n\t\t\ts[i] = fmt.Sprintf(`attributes.%s=\"%s\"`, aggregateTypeAttribute, at)\n\t\t}\n\n\t\treturn strings.Join(s, \" OR \")\n\tcase eh.MatchAny:\n\t\ts := make([]string, len(m))\n\t\tfor i, sm := range m {\n\t\t\ts[i] = fmt.Sprintf(\"(%s)\", createFilter(sm))\n\t\t}\n\n\t\treturn strings.Join(s, \" OR \")\n\tcase eh.MatchAll:\n\t\ts := make([]string, len(m))\n\t\tfor i, sm := range m {\n\t\t\ts[i] = fmt.Sprintf(\"(%s)\", createFilter(sm))\n\t\t}\n\n\t\treturn strings.Join(s, \" AND \")\n\tdefault:\n\t\treturn \"\"\n\t}\n}", "func newACLFilter(authorizer acl.Authorizer, logger *log.Logger, enforceVersion8 bool) *aclFilter {\n\tif logger == nil {\n\t\tlogger = log.New(os.Stderr, \"\", log.LstdFlags)\n\t}\n\treturn &aclFilter{\n\t\tauthorizer: authorizer,\n\t\tlogger: logger,\n\t\tenforceVersion8: enforceVersion8,\n\t}\n}", "func maybeAddMacvlan(pod *corev1.Pod, netns string) error {\n\tannotation, ok := pod.Annotations[osdnv1.AssignMacvlanAnnotation]\n\tif !ok || annotation == \"false\" {\n\t\treturn nil\n\t}\n\n\tprivileged := false\n\tfor _, container := range append(pod.Spec.Containers, pod.Spec.InitContainers...) {\n\t\tif container.SecurityContext != nil && container.SecurityContext.Privileged != nil && *container.SecurityContext.Privileged {\n\t\t\tprivileged = true\n\t\t\tbreak\n\t\t}\n\t}\n\tif !privileged {\n\t\treturn fmt.Errorf(\"pod has %q annotation but is not privileged\", osdnv1.AssignMacvlanAnnotation)\n\t}\n\n\tvar iface netlink.Link\n\tvar err error\n\tif annotation == \"true\" {\n\t\t// Find interface with the default route\n\t\troutes, err := netlink.RouteList(nil, netlink.FAMILY_V4)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"failed to read routes: %v\", err)\n\t\t}\n\n\t\tfor _, r := range routes {\n\t\t\tif r.Dst == nil {\n\t\t\t\tiface, err = netlink.LinkByIndex(r.LinkIndex)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn fmt.Errorf(\"failed to get default route interface: %v\", err)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tif iface == nil {\n\t\t\treturn fmt.Errorf(\"failed to find default route interface\")\n\t\t}\n\t} else {\n\t\tiface, err = netlink.LinkByName(annotation)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"pod annotation %q is neither 'true' nor the name of a local network interface\", osdnv1.AssignMacvlanAnnotation)\n\t\t}\n\t}\n\n\t// Note that this use of ns is safe because it doesn't call Do() or WithNetNSPath()\n\n\tpodNs, err := ns.GetNS(netns)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"could not open netns %q: %v\", netns, err)\n\t}\n\tdefer podNs.Close()\n\n\terr = netlink.LinkAdd(&netlink.Macvlan{\n\t\tLinkAttrs: netlink.LinkAttrs{\n\t\t\tMTU: iface.Attrs().MTU,\n\t\t\tName: \"macvlan0\",\n\t\t\tParentIndex: iface.Attrs().Index,\n\t\t\tNamespace: netlink.NsFd(podNs.Fd()),\n\t\t},\n\t\tMode: netlink.MACVLAN_MODE_PRIVATE,\n\t})\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to create macvlan interface: %v\", err)\n\t}\n\treturn nil\n}", "func eventFilterList(amount int) string {\n\tvar eventTypes []string\n\tfor i := 0; i < amount; i++ {\n\t\teventTypes = append(eventTypes, fmt.Sprintf(\":eventType%d\", i))\n\t}\n\treturn \"(\" + strings.Join(eventTypes, \", \") + \")\"\n}", "func (_options *CreateProviderGatewayOptions) SetVlan(vlan int64) *CreateProviderGatewayOptions {\n\t_options.Vlan = core.Int64Ptr(vlan)\n\treturn _options\n}", "func (l *Libvirt) NwfilterUndefine(OptNwfilter Nwfilter) (err error) {\n\tvar buf []byte\n\n\targs := NwfilterUndefineArgs {\n\t\tOptNwfilter: OptNwfilter,\n\t}\n\n\tbuf, err = encode(&args)\n\tif err != nil {\n\t\treturn\n\t}\n\n\n\t_, err = l.requestStream(181, constants.Program, buf, nil, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func GetIPmacTableVLAN(N Network, VLANfilter string) ARPtable {\n\tmainSwitchCommunity := findMainSwitchCommunity(N)\n\n\t// Getting the filtered IP => MAC table\n\tmainSwitchArpTableVLAN := GetARPtable(N.MainSwitchIP, mainSwitchCommunity, VLANfilter)\n\n\treturn mainSwitchArpTableVLAN\n}", "func NewLinkFilter(name string, code uint64, value any) LinkFilter {\n\treturn func(f encoding.Filter) {\n\t\tvar descriptor any\n\t\tif code != 0 {\n\t\t\tdescriptor = code\n\t\t} else {\n\t\t\tdescriptor = encoding.Symbol(name)\n\t\t}\n\t\tf[encoding.Symbol(name)] = &encoding.DescribedType{\n\t\t\tDescriptor: descriptor,\n\t\t\tValue: value,\n\t\t}\n\t}\n}", "func CustomFiltersBlock() datasourceschema.Block {\n\treturn datasourceschema.SetNestedBlock{\n\t\tNestedObject: datasourceschema.NestedBlockObject{\n\t\t\tAttributes: map[string]datasourceschema.Attribute{\n\t\t\t\t\"name\": datasourceschema.StringAttribute{\n\t\t\t\t\tRequired: true,\n\t\t\t\t},\n\t\t\t\t\"values\": datasourceschema.SetAttribute{\n\t\t\t\t\tElementType: types.StringType,\n\t\t\t\t\tRequired: true,\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n}", "func (b *bot) RegisterFilter(name string, f func(string) string) {\n\tb.filters[name] = f\n}", "func (this *EnvoyFilter_RouteConfigurationMatch_VirtualHostMatch) MarshalJSON() ([]byte, error) {\n\tstr, err := EnvoyFilterMarshaler.MarshalToString(this)\n\treturn []byte(str), err\n}", "func MakeBoxAvgFilter() *Filter {\n\tf := new(Filter)\n\tf.F = func(x float64, scaleFactor float64) float64 {\n\t\tif x < 0.499999 {\n\t\t\treturn 1.0\n\t\t}\n\t\tif x <= 0.500001 {\n\t\t\treturn 0.5\n\t\t}\n\t\treturn 0.0\n\t}\n\tf.Radius = func(scaleFactor float64) float64 {\n\t\treturn 0.5001\n\t}\n\treturn f\n}", "func (s *BaseCobol85PreprocessorListener) VisitTerminal(node antlr.TerminalNode) {}", "func configureFlags(api *operations.SwaggertestAPI) {\n\t// api.CommandLineOptionsGroups = []swag.CommandLineOptionsGroup{ ... }\n}", "func (h *afpacketHandle) SetBPFFilter(filter string, snaplen int) (err error) {\n\tpcapBPF, err := pcap.CompileBPFFilter(layers.LinkTypeEthernet, snaplen, filter)\n\tif err != nil {\n\t\treturn err\n\t}\n\tbpfIns := []bpf.RawInstruction{}\n\tfor _, ins := range pcapBPF {\n\t\tbpfIns2 := bpf.RawInstruction{\n\t\t\tOp: ins.Code,\n\t\t\tJt: ins.Jt,\n\t\t\tJf: ins.Jf,\n\t\t\tK: ins.K,\n\t\t}\n\t\tbpfIns = append(bpfIns, bpfIns2)\n\t}\n\tif h.TPacket.SetBPF(bpfIns); err != nil {\n\t\treturn err\n\t}\n\treturn h.TPacket.SetBPF(bpfIns)\n}", "func TraceFilter(id Id) {\n\tif id < Nids {\n\t\tunfilter[id] = false\n\t} else {\n\t\tfor i, _ := range unfilter {\n\t\t\tunfilter[i] = false\n\t\t}\n\t}\n}", "func (s *Stub) FilterList(filter string) ([]*runtimeconfig.Variable, error) {\n\treturn s.List()\n}", "func (l *LogicalExpr) accept(v ExprVisitor) {\n\tv.VisitLogical(l)\n}", "func renderSimpleMacvlanConfig(conf *operv1.AdditionalNetworkDefinition, manifestDir string) ([]*uns.Unstructured, error) {\n\tvar err error\n\n\t// render SimpleMacvlanConfig manifests\n\tdata := render.MakeRenderData()\n\tdata.Data[\"AdditionalNetworkName\"] = conf.Name\n\tdata.Data[\"AdditionalNetworkNamespace\"] = conf.Namespace\n\n\tif conf.SimpleMacvlanConfig == nil {\n\t\t// no additional config, just fill default IPAM\n\t\tdata.Data[\"IPAMConfig\"], err = getIPAMConfigJSON(nil)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"failed to render ipam config\")\n\t\t}\n\n\t} else {\n\t\tmacvlanConfig := conf.SimpleMacvlanConfig\n\t\tdata.Data[\"Master\"] = macvlanConfig.Master\n\n\t\tdata.Data[\"IPAMConfig\"], err = getIPAMConfigJSON(macvlanConfig.IPAMConfig)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"failed to render ipam config\")\n\t\t}\n\n\t\tif macvlanConfig.Mode != \"\" {\n\t\t\t// macvlan CNI only accepts mode in lowercase\n\t\t\tdata.Data[\"Mode\"] = strings.ToLower(string(macvlanConfig.Mode))\n\t\t}\n\n\t\tif macvlanConfig.MTU != 0 {\n\t\t\tdata.Data[\"MTU\"] = macvlanConfig.MTU\n\t\t}\n\t}\n\n\tobjs, err := render.RenderDir(filepath.Join(manifestDir, \"network/additional-networks/simplemacvlan\"), &data)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to render simplemacvlan additional network\")\n\t}\n\treturn objs, nil\n}", "func NewVlanBridge(agent *OfnetAgent, rpcServ *rpc.Server) *VlanBridge {\n\tvlan := new(VlanBridge)\n\n\t// Keep a reference to the agent\n\tvlan.agent = agent\n\n\t// init maps\n\tvlan.portVlanFlowDb = make(map[uint32]*ofctrl.Flow)\n\n\t// Create policy agent\n\tvlan.policyAgent = NewPolicyAgent(agent, rpcServ)\n\n\treturn vlan\n}", "func (*OpenconfigInterfaces_Interfaces_Interface_Ethernet_SwitchedVlan_Config_TrunkVlans_Union_Uint16) Is_OpenconfigInterfaces_Interfaces_Interface_Ethernet_SwitchedVlan_Config_TrunkVlans_Union() {}", "func (*OpenconfigInterfaces_Interfaces_Interface_Ethernet_SwitchedVlan_Config_TrunkVlans_Union_Uint16) Is_OpenconfigInterfaces_Interfaces_Interface_Ethernet_SwitchedVlan_Config_TrunkVlans_Union() {}", "func (f *FilterWrap) addFilters(fl []interface{}) {\n\tif len(fl) > 1 {\n\t\tfc := fl[0]\n\t\tswitch fc.(type) {\n\t\tcase BoolClause, string:\n\t\t\tf.boolClause = fc.(string)\n\t\t\tfl = fl[1:]\n\t\t}\n\t}\n\tf.filters = append(f.filters, fl...)\n}", "func (*InterfaceVlans) GetPath() string { return \"/api/objects/interface/vlan/\" }", "func MatchFilters(b []byte) bool {\n\t// return true when no filters are configured\n\tif len(netIDs) == 0 && len(joinEUIs) == 0 {\n\t\treturn true\n\t}\n\n\t// return true when we can't decode the LoRaWAN frame\n\tvar phy lorawan.PHYPayload\n\tif err := phy.UnmarshalBinary(b); err != nil {\n\t\tlog.WithError(err).Error(\"filters: unmarshal phypayload error\")\n\t\treturn true\n\t}\n\n\tswitch phy.MHDR.MType {\n\tcase lorawan.UnconfirmedDataUp, lorawan.ConfirmedDataUp:\n\t\treturn filterDevAddr(phy)\n\tcase lorawan.JoinRequest:\n\t\treturn filterJoinRequest(phy)\n\tcase lorawan.RejoinRequest:\n\t\treturn filterRejoinRequest(phy)\n\tdefault:\n\t\treturn true\n\t}\n}", "func (o SubnetworkLogConfigOutput) FilterExpr() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v SubnetworkLogConfig) *string { return v.FilterExpr }).(pulumi.StringPtrOutput)\n}", "func initFilters() int {\n\trows, err := db.Query(\"SELECT idFILTER, idFI, conditions, tests, isDefault FROM TXFILTER\")\n\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\t// defer rows.Close()\n\n\ti := 0\n\tvar teststr *string\n\n\tfor rows.Next() {\n\t\tfil := new(TxFilter)\n\t\terr := rows.Scan(&fil.FName, &fil.FIID, &fil.Conditions, &teststr, &fil.IsDef)\n\t\tif err != nil {\n\t\t\trows.Close()\n\t\t\tlog.Fatal(err)\n\t\t}\n\n\t\ttststr := strings.Split(*teststr, \",\") // parse categories\n\t\tfil.CustTests = make([]string, 0)\n\t\tfil.AcctTests = make([]string, 0)\n\n\t\tfor _, test := range tststr {\n\t\t\tif txTestCache[test].CustOrAcct == \"c\" {\n\t\t\t\tfil.CustTests = append(fil.CustTests, strings.TrimSpace(test))\n\t\t\t} else if txTestCache[test].CustOrAcct == \"a\" {\n\t\t\t\tfil.AcctTests = append(fil.AcctTests, strings.TrimSpace(test))\n\t\t\t} else {\n\t\t\t\tpanic(\"Error Parsing Filter on tests: \" + txTestCache[test].CustOrAcct + test)\n\t\t\t}\n\t\t}\n\n\t\ttxFilterCache[fil.FName] = fil\n\t\tif fil.IsDef == 1 {\n\t\t\ttxDefFilters = append(txDefFilters, fil.FName)\n\t\t}\n\t\ti++\n\t}\n\n\trows.Close()\n\t// build global default filter\n\ttxDefFilterGrp = *new(TxFilterGrp)\n\n\ttxDefFilterGrp.CustTestGrp.Tests = make(map[string]int)\n\ttxDefFilterGrp.CustTestGrp.GrpParamIdx = make(map[string]int)\n\ttxDefFilterGrp.CustTestGrp.TestParamValue = make(map[string][]interface{})\n\ttxDefFilterGrp.CustTestGrp.SQLPeriodIdx = make(map[string]*PeriodIdx)\n\n\ttxDefFilterGrp.AcctTestGrp.Tests = make(map[string]int)\n\ttxDefFilterGrp.AcctTestGrp.GrpParamIdx = make(map[string]int)\n\ttxDefFilterGrp.AcctTestGrp.TestParamValue = make(map[string][]interface{})\n\ttxDefFilterGrp.AcctTestGrp.SQLPeriodIdx = make(map[string]*PeriodIdx)\n\n\tfor _, filstr := range txDefFilters {\n\t\tif len(txDefFilterGrp.Filters) == 0 {\n\t\t\ttxDefFilterGrp.Filters += txFilterCache[filstr].FName\n\t\t} else {\n\t\t\ttxDefFilterGrp.Filters += \",\" + txFilterCache[filstr].FName\n\t\t}\n\t\tif len(txDefFilterGrp.ActiveFilters) == 0 {\n\t\t\ttxDefFilterGrp.ActiveFilters += txFilterCache[filstr].FName\n\t\t} else {\n\t\t\ttxDefFilterGrp.ActiveFilters += \",\" + txFilterCache[filstr].FName\n\t\t}\n\n\t\tbuildTestGrp(&txDefFilterGrp.CustTestGrp, txFilterCache[filstr].CustTests)\n\t\tbuildTestGrp(&txDefFilterGrp.AcctTestGrp, txFilterCache[filstr].AcctTests)\n\n\t}\n\n\tfmt.Printf(\"\\nDefault Filters:%+v\", txDefFilters)\n\tfmt.Printf(\"\\nDefault Filter group:%+v\", txDefFilterGrp)\n\n\t//\treturn custs, err\n\treturn i\n}", "func validateSimpleMacvlanConfig(conf *operv1.AdditionalNetworkDefinition) []error {\n\tout := []error{}\n\n\tif conf.Name == \"\" {\n\t\tout = append(out, errors.Errorf(\"Additional Network Name cannot be nil\"))\n\t}\n\n\tif conf.SimpleMacvlanConfig != nil {\n\t\tmacvlanConfig := conf.SimpleMacvlanConfig\n\t\tif macvlanConfig.IPAMConfig != nil {\n\t\t\toutIPAM := validateIPAMConfig(macvlanConfig.IPAMConfig)\n\t\t\tout = append(out, outIPAM...)\n\t\t}\n\n\t\tif conf.SimpleMacvlanConfig.Mode != \"\" {\n\t\t\tswitch conf.SimpleMacvlanConfig.Mode {\n\t\t\tcase operv1.MacvlanModeBridge:\n\t\t\tcase operv1.MacvlanModePrivate:\n\t\t\tcase operv1.MacvlanModeVEPA:\n\t\t\tcase operv1.MacvlanModePassthru:\n\t\t\tdefault:\n\t\t\t\tout = append(out, errors.Errorf(\"invalid Macvlan mode: %s\", conf.SimpleMacvlanConfig.Mode))\n\t\t\t}\n\t\t}\n\t}\n\n\treturn out\n}", "func (xdcrf *XDCRFactory) filterVBList(targetkvVBList []uint16, kv_vb_map map[string][]uint16) []uint16 {\n\tret := []uint16{}\n\tfor _, vb := range targetkvVBList {\n\t\tfor _, sourcevblist := range kv_vb_map {\n\t\t\tfor _, sourcevb := range sourcevblist {\n\t\t\t\tif sourcevb == vb {\n\t\t\t\t\tret = append(ret, vb)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn ret\n}", "func configureFlags(api *operations.KubernikusAPI) {\n\t// api.CommandLineOptionsGroups = []swag.CommandLineOptionsGroup{ ... }\n}", "func (cm *ConnectionManager) RegisterFilters(source <-chan packet.Packet,\n\tfilters ...func(packet.Packet) packet.Packet) <-chan packet.Packet {\n\tsink := make(chan packet.Packet)\n\n\tgo func() {\n\t\tfor p := range source {\n\t\t\tfor _, filter := range filters {\n\t\t\t\tif pass := filter(p); pass != nil {\n\t\t\t\t\tsink <- pass\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn sink\n}", "func configureFlags(api *operations.LoongTokenAPI) {\n\t// api.CommandLineOptionsGroups = []swag.CommandLineOptionsGroup{ ... }\n}", "func (f kubevalFilter) Filter(in []*yaml.RNode) ([]*yaml.RNode, error) {\n\tcfg, err := f.parseConfig()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tkubevalConfig := kubeval.NewDefaultConfig()\n\tkubevalConfig.Strict = cfg.Spec.Strict\n\tkubevalConfig.IgnoreMissingSchemas = cfg.Spec.IgnoreMissingSchemas\n\tkubevalConfig.KubernetesVersion = cfg.Spec.KubernetesVersion\n\tkubevalConfig.AdditionalSchemaLocations = []string{fileScheme + \"://\" + schemaLocationDir}\n\tkubevalConfig.KindsToSkip = []string{crdKind, kubevalOptsKind}\n\tvar crdList []string\n\n\tif _, plan := os.LookupEnv(planEnv); plan {\n\t\t// Setup plan specific options\n\t\tif _, exists := cfg.PlanConfigs[cfg.PlanName]; exists {\n\t\t\tkubevalConfig.KindsToSkip = append(kubevalConfig.KindsToSkip, cfg.PlanConfigs[cfg.PlanName].KindsToSkip...)\n\t\t\tcrdList = cfg.PlanConfigs[cfg.PlanName].CRDList\n\t\t}\n\t} else {\n\t\t// Setup phase specific options\n\t\tif _, exists := cfg.PhaseConfigs[cfg.PhaseName]; exists {\n\t\t\tkubevalConfig.KindsToSkip = append(kubevalConfig.KindsToSkip, cfg.PhaseConfigs[cfg.PhaseName].KindsToSkip...)\n\t\t\tcrdList = cfg.PhaseConfigs[cfg.PhaseName].CRDList\n\t\t}\n\t}\n\n\t// Calculate schema location directory for kubeval and openapi2jsonschema based on options\n\tschemasLocation := filepath.Join(schemaLocationDir,\n\t\tfmt.Sprintf(\"v%s-%s\", kubevalConfig.KubernetesVersion, \"standalone\"))\n\tif kubevalConfig.Strict {\n\t\tschemasLocation += \"-strict\"\n\t}\n\t// Create it if doesn't exist\n\tif _, err := os.Stat(schemasLocation); os.IsNotExist(err) {\n\t\tif err = os.MkdirAll(schemasLocation, 0755); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// Filter CRDs from input\n\tcrdRNodes, err := filterCRD(in)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif len(crdRNodes) > 0 {\n\t\t// Save filtered CRDs in file to future processing\n\t\trenderedCRDFile := filepath.Join(schemaLocationDir, phaseRenderedFile)\n\t\tbuf := bytes.Buffer{}\n\t\tfor _, rNode := range crdRNodes {\n\t\t\tbuf.Write([]byte(\"---\\n\" + rNode.MustString()))\n\t\t}\n\t\tif err = ioutil.WriteFile(renderedCRDFile, buf.Bytes(), 0600); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\t// Prepend rendered CRD to give them priority for processing\n\t\tcrdList = append([]string{renderedCRDFile}, crdList...)\n\t}\n\n\tif len(crdList) > 0 {\n\t\t// Process each additional CRD in the list (CRD -> OpenAPIV3 Schema -> Json Schema)\n\t\tif err := processCRDList(crdList, schemasLocation); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// Validate each Resource\n\tfor _, r := range in {\n\t\tmeta, err := r.GetMeta()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tif err := validate(r.MustString(), kubevalConfig); err != nil {\n\t\t\t// if there's an issue found with document - it will be printed as well\n\t\t\tprintMsg(\"Resource invalid: (Kind: %s, Name: %s)\\n---\\n%s---\\n\", meta.Kind, meta.Name, r.MustString())\n\t\t\treturn nil, err\n\t\t}\n\t\t// inform document is ok\n\t\tprintMsg(\"Resource valid: (Kind: %s, Name: %s)\\n\", meta.Kind, meta.Name)\n\t}\n\n\t// if prevent cleanup variable is not set then cleanup working directory\n\tif _, cleanup := os.LookupEnv(cleanupEnv); !cleanup {\n\t\tif err := os.RemoveAll(schemaLocationDir); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\t// Don't return output list, we satisfied with exit code and stdout/stderr\n\treturn nil, nil\n}", "func NewFilteringSpanListener(config map[string]interface{}) ThundraSpanListener {\n\tfilterer := &ThundraSpanFilterer{spanFilters: []SpanFilter{}}\n\n\tlistenerDef, ok := config[\"listener\"].(map[string]interface{})\n\tif !ok {\n\t\tlog.Println(\"Listener configuration is not valid for FilteringSpanListener\")\n\t\treturn nil\n\t}\n\n\tlistenerName, ok := listenerDef[\"type\"].(string)\n\tlistenerConstructor, ok := SpanListenerConstructorMap[listenerName]\n\tif !ok {\n\t\tlog.Println(\"Given listener type is not valid for FilteringSpanListener\")\n\t\treturn nil\n\t}\n\n\tlistenerConfig, ok := listenerDef[\"config\"].(map[string]interface{})\n\tlistener := listenerConstructor(listenerConfig)\n\n\tif all, ok := config[\"all\"].(bool); ok {\n\t\tfilterer.all = all\n\t}\n\n\tif filterConfigs, ok := config[\"filters\"].([]interface{}); ok {\n\t\tfilterer.spanFilters = crateFiltersFromConfig(filterConfigs)\n\t}\n\n\treturn &FilteringSpanListener{listener, filterer}\n}", "func (p *Profile) GetVlansMX() ([]*dproto.Vlan, error) {\n\tp.Log(\"Starting JunOS.GetVlansMX()\")\n\tvlans := make([]*dproto.Vlan, 0)\n\n\t//p.Debug(\"GETTING MX SERIES VLANS\")\n\t//reStr := `(?msi:Logical interface (?P<ifname>[^\\n\\s]+)[^\\n]+ifindex[^\\n]+\\n[^\\n]+VLAN-Tag \\[ 0x\\d+\\.(?P<vid>\\d+) )`\n\treStr := `(?msi:Logical interface (?P<ifname>[^\\n\\s\\.]+)\\.\\d+[^\\n]+ifindex[^\\n]+(\\n\\s+Desc[^\\n]+)?\\n[^\\n]+VLAN-Tag \\[ 0x\\d+\\.(?P<vid>\\d+) )`\n\tre, err := regexp.Compile(reStr)\n\tif err != nil {\n\t\treturn vlans, fmt.Errorf(\"Cannot compile MX vlan regex: %s\", err.Error())\n\t}\n\n\tresult, err := p.Cli.Cmd(\"show interfaces\")\n\tif err != nil {\n\t\treturn vlans, fmt.Errorf(\"Cannot 'show interfaces': %s\", err.Error())\n\t}\n\tp.Debug(result)\n\n\tvlanports := make(map[int64][]string)\n\n\tout := p.ParseMultiple(re, result)\n\tfor i := range out {\n\t\tifname := strings.Trim(out[i][\"ifname\"], \" \")\n\t\tif ifname == \"\" {\n\t\t\tcontinue\n\t\t}\n\t\tvidStr := strings.Trim(out[i][\"vid\"], \" \")\n\t\tif vidStr == \"\" || vidStr == \"0\" {\n\t\t\tcontinue\n\t\t}\n\n\t\tvid, err := strconv.ParseInt(vidStr, 10, 64)\n\t\tif err != nil {\n\t\t\tcontinue\n\t\t}\n\n\t\tif ports, ok := vlanports[vid]; ok {\n\t\t\t// check if there is no this ports already\n\t\t\tfound := false\n\t\t\tfor n := range ports {\n\t\t\t\tif ports[n] == ifname {\n\t\t\t\t\tfound = true\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\tif !found {\n\t\t\t\tports = append(ports, ifname)\n\t\t\t\tvlanports[vid] = ports\n\t\t\t}\n\t\t} else {\n\t\t\tports = []string{ifname}\n\t\t\tvlanports[vid] = ports\n\t\t}\n\t\t//vlan := dproto.Vlan{}\n\t\t//p.Debug(\"%+#v\", out[i])\n\t}\n\n\tfor vid, ports := range vlanports {\n\t\tvlan := dproto.Vlan{\n\t\t\tName: fmt.Sprintf(\"%d\", vid),\n\t\t\tAccessPorts:make([]string,0),\n\t\t\tID:vid,\n\t\t\tTrunkPorts:ports,\n\t\t}\n\t\tvlans = append(vlans, &vlan)\n\t}\n\n\n\treturn vlans, nil\n}", "func importVLANs(c context.Context, configSet config.Set) error {\n\tvlan := &configPB.VLANs{}\n\tmetadata := &config.Meta{}\n\tif err := cfgclient.Get(c, cfgclient.AsService, configSet, vlansFilename, textproto.Message(vlan), metadata); err != nil {\n\t\treturn errors.Annotate(err, \"failed to load %s\", vlansFilename).Err()\n\t}\n\tlogging.Infof(c, \"Found %s revision %q\", vlansFilename, metadata.Revision)\n\n\tctx := &validation.Context{Context: c}\n\tctx.SetFile(vlansFilename)\n\tvalidateVLANs(ctx, vlan)\n\tif err := ctx.Finalize(); err != nil {\n\t\treturn errors.Annotate(err, \"invalid config\").Err()\n\t}\n\n\tif err := model.EnsureVLANs(c, vlan.Vlan); err != nil {\n\t\treturn errors.Annotate(err, \"failed to ensure VLANs\").Err()\n\t}\n\tif err := model.EnsureIPs(c, vlan.Vlan); err != nil {\n\t\treturn errors.Annotate(err, \"failed to ensure IP addresses\").Err()\n\t}\n\treturn nil\n}", "func injectConnectFilters(cfgSnap *proxycfg.ConfigSnapshot, token string, listener *envoy.Listener) error {\n\tauthFilter, err := makeExtAuthFilter(token)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor idx := range listener.FilterChains {\n\t\t// Insert our authz filter before any others\n\t\tlistener.FilterChains[idx].Filters =\n\t\t\tappend([]envoylistener.Filter{authFilter}, listener.FilterChains[idx].Filters...)\n\n\t\t// Force our TLS for all filter chains on a public listener\n\t\tlistener.FilterChains[idx].TlsContext = &envoyauth.DownstreamTlsContext{\n\t\t\tCommonTlsContext: makeCommonTLSContext(cfgSnap),\n\t\t\tRequireClientCertificate: &types.BoolValue{Value: true},\n\t\t}\n\t}\n\treturn nil\n}", "func LoadFilters(projectRoot string) {\n\tvar file *os.File\n\tif projectRoot != \"\" {\n\t\tfile, _ = os.Open(projectRoot + \"filters.json\")\n\t} else {\n\t\tfile, _ = os.Open(\"filters.json\")\n\t}\n\tdecoder := json.NewDecoder(file)\n\terr := decoder.Decode(&filterObj)\n\tif err != nil {\n\t\tlog.Println(\"Error loading filters\")\n\t}\n\tcompileRegex()\n}", "func (m *Vlan) Validate(formats strfmt.Registry) error {\n\tvar res []error\n\n\tif err := m.validateID(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateLabels(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateLocalID(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateName(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateType(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateVds(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateVlanID(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateVMNics(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func createFilter(searchAttr string, attr []string, baseFilter string) string {\n\tfilterAttr := \"\"\n\n\tfor _, attr := range attr {\n\t\tfilterAttr += fmt.Sprintf(\"(%s=%s)\", attr, searchAttr)\n\t}\n\n\tfilterAllArgs := fmt.Sprintf(\"(|%s)\", filterAttr)\n\n\treturn fmt.Sprintf(\"(&(%s)%s)\", baseFilter, filterAllArgs)\n}", "func init()\t{ RegisterTranslator(EventFilterTranslator) }", "func FilterEnable(o OutPutter, f func(ctx context.Context, name string, level Level) bool) OutPutter {\n\tif o == nil {\n\t\treturn o\n\t}\n\treturn &OutPutFilter{\n\t\tunderlying: o,\n\t\tenableFunc: f,\n\t}\n}", "func (*OpenconfigInterfaces_Interfaces_Interface_RoutedVlan_Config_Vlan_Union_Uint16) Is_OpenconfigInterfaces_Interfaces_Interface_RoutedVlan_Config_Vlan_Union() {}", "func (*OpenconfigInterfaces_Interfaces_Interface_RoutedVlan_Config_Vlan_Union_Uint16) Is_OpenconfigInterfaces_Interfaces_Interface_RoutedVlan_Config_Vlan_Union() {}", "func FilterVerbs(prefix string) (result []*Verb) {\n\treturn globalVerbs.Filter(prefix)\n}", "func (*OpenconfigInterfaces_Interfaces_Interface_Aggregation_SwitchedVlan_Config_TrunkVlans_Union_Uint16) Is_OpenconfigInterfaces_Interfaces_Interface_Aggregation_SwitchedVlan_Config_TrunkVlans_Union() {}", "func (*OpenconfigInterfaces_Interfaces_Interface_Aggregation_SwitchedVlan_Config_TrunkVlans_Union_Uint16) Is_OpenconfigInterfaces_Interfaces_Interface_Aggregation_SwitchedVlan_Config_TrunkVlans_Union() {}", "func createFilter(fr filters.Registry, def *eskip.Filter, cpm map[string]PredicateSpec) (filters.Filter, error) {\n\tspec, ok := fr[def.Name]\n\tif !ok {\n\t\tif isTreePredicate(def.Name) || def.Name == predicates.HostName || def.Name == predicates.PathRegexpName || def.Name == predicates.MethodName || def.Name == predicates.HeaderName || def.Name == predicates.HeaderRegexpName {\n\t\t\treturn nil, fmt.Errorf(\"trying to use %q as filter, but it is only available as predicate\", def.Name)\n\t\t}\n\n\t\tif _, ok := cpm[def.Name]; ok {\n\t\t\treturn nil, fmt.Errorf(\"trying to use %q as filter, but it is only available as predicate\", def.Name)\n\t\t}\n\n\t\treturn nil, fmt.Errorf(\"filter %q not found\", def.Name)\n\t}\n\n\tf, err := spec.CreateFilter(def.Args)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to create filter %q: %w\", spec.Name(), err)\n\t}\n\treturn f, nil\n}", "func routerConfig(port int) Ifx {\n\treturn Ifx{\n\t\tFor: \"router\",\n\t\tName: fmt.Sprintf(\"swp%d\", port),\n\t\tBridgeDefs: []string{\n\t\t\t\"bridge-vids 2002 2003 2004 2005 2006 2007\",\n\t\t\t\"bridge-allow-untagged no\",\n\t\t}}\n}", "func newFilterView(gui *gocui.Gui) (controller *Filter) {\n\tcontroller = new(Filter)\n\n\tcontroller.filterEditListeners = make([]FilterEditListener, 0)\n\n\t// populate main fields\n\tcontroller.name = \"filter\"\n\tcontroller.gui = gui\n\tcontroller.labelStr = \"Path Filter: \"\n\tcontroller.hidden = true\n\n\tcontroller.requestedHeight = 1\n\n\treturn controller\n}", "func resourceVolterraAppFirewall() *schema.Resource {\n\treturn &schema.Resource{\n\t\tCreate: resourceVolterraAppFirewallCreate,\n\t\tRead: resourceVolterraAppFirewallRead,\n\t\tUpdate: resourceVolterraAppFirewallUpdate,\n\t\tDelete: resourceVolterraAppFirewallDelete,\n\n\t\tSchema: map[string]*schema.Schema{\n\n\t\t\t\"annotations\": {\n\t\t\t\tType: schema.TypeMap,\n\t\t\t\tOptional: true,\n\t\t\t},\n\n\t\t\t\"description\": {\n\t\t\t\tType: schema.TypeString,\n\t\t\t\tOptional: true,\n\t\t\t},\n\n\t\t\t\"disable\": {\n\t\t\t\tType: schema.TypeBool,\n\t\t\t\tOptional: true,\n\t\t\t},\n\n\t\t\t\"labels\": {\n\t\t\t\tType: schema.TypeMap,\n\t\t\t\tOptional: true,\n\t\t\t},\n\n\t\t\t\"name\": {\n\t\t\t\tType: schema.TypeString,\n\t\t\t\tRequired: true,\n\t\t\t\tForceNew: true,\n\t\t\t},\n\n\t\t\t\"namespace\": {\n\t\t\t\tType: schema.TypeString,\n\t\t\t\tRequired: true,\n\t\t\t\tForceNew: true,\n\t\t\t},\n\n\t\t\t\"allow_all_response_codes\": {\n\n\t\t\t\tType: schema.TypeBool,\n\t\t\t\tOptional: true,\n\t\t\t},\n\n\t\t\t\"allowed_response_codes\": {\n\n\t\t\t\tType: schema.TypeSet,\n\t\t\t\tOptional: true,\n\t\t\t\tElem: &schema.Resource{\n\t\t\t\t\tSchema: map[string]*schema.Schema{\n\n\t\t\t\t\t\t\"response_code\": {\n\n\t\t\t\t\t\t\tType: schema.TypeList,\n\n\t\t\t\t\t\t\tRequired: true,\n\t\t\t\t\t\t\tElem: &schema.Schema{\n\t\t\t\t\t\t\t\tType: schema.TypeInt,\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\n\t\t\t\"custom_anonymization\": {\n\n\t\t\t\tType: schema.TypeSet,\n\t\t\t\tOptional: true,\n\t\t\t\tElem: &schema.Resource{\n\t\t\t\t\tSchema: map[string]*schema.Schema{\n\n\t\t\t\t\t\t\"anonymization_config\": {\n\n\t\t\t\t\t\t\tType: schema.TypeList,\n\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t\tElem: &schema.Resource{\n\t\t\t\t\t\t\t\tSchema: map[string]*schema.Schema{\n\n\t\t\t\t\t\t\t\t\t\"cookie\": {\n\n\t\t\t\t\t\t\t\t\t\tType: schema.TypeSet,\n\t\t\t\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t\t\t\t\tElem: &schema.Resource{\n\t\t\t\t\t\t\t\t\t\t\tSchema: map[string]*schema.Schema{\n\n\t\t\t\t\t\t\t\t\t\t\t\t\"cookie_name\": {\n\t\t\t\t\t\t\t\t\t\t\t\t\tType: schema.TypeString,\n\t\t\t\t\t\t\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t\t},\n\n\t\t\t\t\t\t\t\t\t\"http_header\": {\n\n\t\t\t\t\t\t\t\t\t\tType: schema.TypeSet,\n\t\t\t\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t\t\t\t\tElem: &schema.Resource{\n\t\t\t\t\t\t\t\t\t\t\tSchema: map[string]*schema.Schema{\n\n\t\t\t\t\t\t\t\t\t\t\t\t\"header_name\": {\n\t\t\t\t\t\t\t\t\t\t\t\t\tType: schema.TypeString,\n\t\t\t\t\t\t\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t\t},\n\n\t\t\t\t\t\t\t\t\t\"query_parameter\": {\n\n\t\t\t\t\t\t\t\t\t\tType: schema.TypeSet,\n\t\t\t\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t\t\t\t\tElem: &schema.Resource{\n\t\t\t\t\t\t\t\t\t\t\tSchema: map[string]*schema.Schema{\n\n\t\t\t\t\t\t\t\t\t\t\t\t\"query_param_name\": {\n\t\t\t\t\t\t\t\t\t\t\t\t\tType: schema.TypeString,\n\t\t\t\t\t\t\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\n\t\t\t\"default_anonymization\": {\n\n\t\t\t\tType: schema.TypeBool,\n\t\t\t\tOptional: true,\n\t\t\t},\n\n\t\t\t\"disable_anonymization\": {\n\n\t\t\t\tType: schema.TypeBool,\n\t\t\t\tOptional: true,\n\t\t\t},\n\n\t\t\t\"blocking_page\": {\n\n\t\t\t\tType: schema.TypeSet,\n\t\t\t\tOptional: true,\n\t\t\t\tElem: &schema.Resource{\n\t\t\t\t\tSchema: map[string]*schema.Schema{\n\n\t\t\t\t\t\t\"blocking_page\": {\n\t\t\t\t\t\t\tType: schema.TypeString,\n\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t},\n\n\t\t\t\t\t\t\"response_code\": {\n\t\t\t\t\t\t\tType: schema.TypeString,\n\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\n\t\t\t\"use_default_blocking_page\": {\n\n\t\t\t\tType: schema.TypeBool,\n\t\t\t\tOptional: true,\n\t\t\t},\n\n\t\t\t\"bot_protection_setting\": {\n\n\t\t\t\tType: schema.TypeSet,\n\t\t\t\tOptional: true,\n\t\t\t\tElem: &schema.Resource{\n\t\t\t\t\tSchema: map[string]*schema.Schema{\n\n\t\t\t\t\t\t\"good_bot_action\": {\n\t\t\t\t\t\t\tType: schema.TypeString,\n\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t},\n\n\t\t\t\t\t\t\"malicious_bot_action\": {\n\t\t\t\t\t\t\tType: schema.TypeString,\n\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t},\n\n\t\t\t\t\t\t\"suspicious_bot_action\": {\n\t\t\t\t\t\t\tType: schema.TypeString,\n\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\n\t\t\t\"default_bot_setting\": {\n\n\t\t\t\tType: schema.TypeBool,\n\t\t\t\tOptional: true,\n\t\t\t},\n\n\t\t\t\"default_detection_settings\": {\n\n\t\t\t\tType: schema.TypeBool,\n\t\t\t\tOptional: true,\n\t\t\t},\n\n\t\t\t\"detection_settings\": {\n\n\t\t\t\tType: schema.TypeSet,\n\t\t\t\tOptional: true,\n\t\t\t\tElem: &schema.Resource{\n\t\t\t\t\tSchema: map[string]*schema.Schema{\n\n\t\t\t\t\t\t\"disable_suppression\": {\n\n\t\t\t\t\t\t\tType: schema.TypeBool,\n\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t},\n\n\t\t\t\t\t\t\"enable_suppression\": {\n\n\t\t\t\t\t\t\tType: schema.TypeBool,\n\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t},\n\n\t\t\t\t\t\t\"signature_selection_setting\": {\n\n\t\t\t\t\t\t\tType: schema.TypeSet,\n\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t\tElem: &schema.Resource{\n\t\t\t\t\t\t\t\tSchema: map[string]*schema.Schema{\n\n\t\t\t\t\t\t\t\t\t\"attack_type_settings\": {\n\n\t\t\t\t\t\t\t\t\t\tType: schema.TypeSet,\n\t\t\t\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t\t\t\t\tElem: &schema.Resource{\n\t\t\t\t\t\t\t\t\t\t\tSchema: map[string]*schema.Schema{\n\n\t\t\t\t\t\t\t\t\t\t\t\t\"disabled_attack_types\": {\n\n\t\t\t\t\t\t\t\t\t\t\t\t\tType: schema.TypeList,\n\n\t\t\t\t\t\t\t\t\t\t\t\t\tRequired: true,\n\t\t\t\t\t\t\t\t\t\t\t\t\tElem: &schema.Schema{\n\t\t\t\t\t\t\t\t\t\t\t\t\t\tType: schema.TypeString,\n\t\t\t\t\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t\t},\n\n\t\t\t\t\t\t\t\t\t\"default_attack_type_settings\": {\n\n\t\t\t\t\t\t\t\t\t\tType: schema.TypeBool,\n\t\t\t\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t\t\t\t},\n\n\t\t\t\t\t\t\t\t\t\"high_medium_accuracy_signatures\": {\n\n\t\t\t\t\t\t\t\t\t\tType: schema.TypeBool,\n\t\t\t\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t\t\t\t},\n\n\t\t\t\t\t\t\t\t\t\"high_medium_low_accuracy_signatures\": {\n\n\t\t\t\t\t\t\t\t\t\tType: schema.TypeBool,\n\t\t\t\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t\t\t\t},\n\n\t\t\t\t\t\t\t\t\t\"only_high_accuracy_signatures\": {\n\n\t\t\t\t\t\t\t\t\t\tType: schema.TypeBool,\n\t\t\t\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t},\n\n\t\t\t\t\t\t\"disable_staging\": {\n\n\t\t\t\t\t\t\tType: schema.TypeBool,\n\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t},\n\n\t\t\t\t\t\t\"stage_new_and_updated_signatures\": {\n\n\t\t\t\t\t\t\tType: schema.TypeSet,\n\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t\tElem: &schema.Resource{\n\t\t\t\t\t\t\t\tSchema: map[string]*schema.Schema{\n\n\t\t\t\t\t\t\t\t\t\"staging_period\": {\n\t\t\t\t\t\t\t\t\t\tType: schema.TypeInt,\n\t\t\t\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t},\n\n\t\t\t\t\t\t\"stage_new_signatures\": {\n\n\t\t\t\t\t\t\tType: schema.TypeSet,\n\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t\tElem: &schema.Resource{\n\t\t\t\t\t\t\t\tSchema: map[string]*schema.Schema{\n\n\t\t\t\t\t\t\t\t\t\"staging_period\": {\n\t\t\t\t\t\t\t\t\t\tType: schema.TypeInt,\n\t\t\t\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t},\n\n\t\t\t\t\t\t\"disable_threat_campaigns\": {\n\n\t\t\t\t\t\t\tType: schema.TypeBool,\n\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t},\n\n\t\t\t\t\t\t\"enable_threat_campaigns\": {\n\n\t\t\t\t\t\t\tType: schema.TypeBool,\n\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t},\n\n\t\t\t\t\t\t\"default_violation_settings\": {\n\n\t\t\t\t\t\t\tType: schema.TypeBool,\n\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t},\n\n\t\t\t\t\t\t\"violation_settings\": {\n\n\t\t\t\t\t\t\tType: schema.TypeSet,\n\t\t\t\t\t\t\tOptional: true,\n\t\t\t\t\t\t\tElem: &schema.Resource{\n\t\t\t\t\t\t\t\tSchema: map[string]*schema.Schema{\n\n\t\t\t\t\t\t\t\t\t\"disabled_violation_types\": {\n\n\t\t\t\t\t\t\t\t\t\tType: schema.TypeList,\n\n\t\t\t\t\t\t\t\t\t\tRequired: true,\n\t\t\t\t\t\t\t\t\t\tElem: &schema.Schema{\n\t\t\t\t\t\t\t\t\t\t\tType: schema.TypeString,\n\t\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\n\t\t\t\"blocking\": {\n\n\t\t\t\tType: schema.TypeBool,\n\t\t\t\tOptional: true,\n\t\t\t},\n\n\t\t\t\"monitoring\": {\n\n\t\t\t\tType: schema.TypeBool,\n\t\t\t\tOptional: true,\n\t\t\t},\n\n\t\t\t\"use_loadbalancer_setting\": {\n\n\t\t\t\tType: schema.TypeBool,\n\t\t\t\tOptional: true,\n\t\t\t},\n\t\t},\n\t}\n}", "func (client *Client) addVlanDropRule() error {\n\tout, err := client.plClient.ExecuteCommand(l2PreroutingEntries)\n\tif err != nil {\n\t\tlog.Printf(\"Error while listing ebtable rules %v\", err)\n\t\treturn err\n\t}\n\n\tout = strings.TrimSpace(out)\n\tif strings.Contains(out, vlanDropMatch) {\n\t\tlog.Printf(\"vlan drop rule already exists\")\n\t\treturn nil\n\t}\n\n\tlog.Printf(\"Adding ebtable rule to drop vlan traffic on snat bridge %v\", vlanDropAddRule)\n\t_, err = client.plClient.ExecuteCommand(vlanDropAddRule)\n\treturn err\n}", "func (_options *UpdateProviderGatewayOptions) SetVlan(vlan int64) *UpdateProviderGatewayOptions {\n\t_options.Vlan = core.Int64Ptr(vlan)\n\treturn _options\n}", "func (_Contracts *ContractsFilterer) FilterNewVoter(opts *bind.FilterOpts, _voter []common.Address) (*ContractsNewVoterIterator, error) {\n\n\tvar _voterRule []interface{}\n\tfor _, _voterItem := range _voter {\n\t\t_voterRule = append(_voterRule, _voterItem)\n\t}\n\n\tlogs, sub, err := _Contracts.contract.FilterLogs(opts, \"NewVoter\", _voterRule)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &ContractsNewVoterIterator{contract: _Contracts.contract, event: \"NewVoter\", logs: logs, sub: sub}, nil\n}" ]
[ "0.48371637", "0.4753486", "0.4648949", "0.4607603", "0.4601089", "0.4561446", "0.45241186", "0.45202526", "0.4375413", "0.43716314", "0.43671343", "0.43227693", "0.43169898", "0.43168765", "0.41996193", "0.41971532", "0.41916808", "0.4167578", "0.41593736", "0.41554394", "0.4148585", "0.41473985", "0.41187084", "0.41041207", "0.408797", "0.40849337", "0.4066212", "0.4053457", "0.40290445", "0.40264896", "0.4022532", "0.4018608", "0.4011831", "0.40102866", "0.40061748", "0.40009356", "0.3999925", "0.39993292", "0.3965491", "0.39511442", "0.39438242", "0.3940509", "0.3939052", "0.3937145", "0.39303133", "0.39233455", "0.39175823", "0.39170694", "0.39130723", "0.3905251", "0.3904637", "0.38813898", "0.38795176", "0.38790467", "0.38706616", "0.38704062", "0.38661855", "0.38606474", "0.38576287", "0.38558364", "0.38498256", "0.3832998", "0.3830321", "0.38209432", "0.38186076", "0.38183835", "0.38171983", "0.38171983", "0.38133088", "0.38113773", "0.38066933", "0.38003156", "0.3799185", "0.37961042", "0.3795637", "0.37913162", "0.37728167", "0.37691015", "0.37573275", "0.37572432", "0.37556005", "0.37552488", "0.3752565", "0.37468532", "0.37449434", "0.37397343", "0.3739369", "0.37349936", "0.37343472", "0.37343472", "0.37338725", "0.37330174", "0.37330174", "0.3727435", "0.37214178", "0.3718123", "0.37178895", "0.3716508", "0.37136176", "0.37128004" ]
0.8021751
0
devMacros generates NATIVE_DEV_MAC_BY_IFINDEX and IS_L3_DEV macros which are written to node_config.h.
func devMacros() (string, string, error) { var ( macByIfIndexMacro, isL3DevMacroBuf bytes.Buffer isL3DevMacro string ) macByIfIndex := make(map[int]string) l3DevIfIndices := make([]int, 0) for _, iface := range option.Config.GetDevices() { link, err := netlink.LinkByName(iface) if err != nil { return "", "", fmt.Errorf("failed to retrieve link %s by name: %q", iface, err) } idx := link.Attrs().Index m := link.Attrs().HardwareAddr if m == nil || len(m) != 6 { l3DevIfIndices = append(l3DevIfIndices, idx) } macByIfIndex[idx] = mac.CArrayString(m) } macByIfindexTmpl := template.Must(template.New("macByIfIndex").Parse( `({ \ union macaddr __mac = {.addr = {0x0, 0x0, 0x0, 0x0, 0x0, 0x0}}; \ switch (IFINDEX) { \ {{range $idx,$mac := .}} case {{$idx}}: {union macaddr __tmp = {.addr = {{$mac}}}; __mac=__tmp;} break; \ {{end}}} \ __mac; })`)) if err := macByIfindexTmpl.Execute(&macByIfIndexMacro, macByIfIndex); err != nil { return "", "", fmt.Errorf("failed to execute template: %q", err) } if len(l3DevIfIndices) == 0 { isL3DevMacro = "false" } else { isL3DevTmpl := template.Must(template.New("isL3Dev").Parse( `({ \ bool is_l3 = false; \ switch (ifindex) { \ {{range $idx := .}} case {{$idx}}: is_l3 = true; break; \ {{end}}} \ is_l3; })`)) if err := isL3DevTmpl.Execute(&isL3DevMacroBuf, l3DevIfIndices); err != nil { return "", "", fmt.Errorf("failed to execute template: %q", err) } isL3DevMacro = isL3DevMacroBuf.String() } return macByIfIndexMacro.String(), isL3DevMacro, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func devIdx(tid string) string {\n\treturn indexDevices + \"-\" + tid\n}", "func vlanFilterMacros() (string, error) {\n\tdevices := make(map[int]bool)\n\tfor _, device := range option.Config.GetDevices() {\n\t\tifindex, err := link.GetIfIndex(device)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tdevices[int(ifindex)] = true\n\t}\n\n\tallowedVlans := make(map[int]bool)\n\tfor _, vlanId := range option.Config.VLANBPFBypass {\n\t\tallowedVlans[vlanId] = true\n\t}\n\n\t// allow all vlan id's\n\tif allowedVlans[0] {\n\t\treturn \"return true\", nil\n\t}\n\n\tvlansByIfIndex := make(map[int][]int)\n\n\tlinks, err := netlink.LinkList()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tfor _, l := range links {\n\t\tvlan, ok := l.(*netlink.Vlan)\n\t\t// if it's vlan device and we're controlling vlan main device\n\t\t// and either all vlans are allowed, or we're controlling vlan device or vlan is explicitly allowed\n\t\tif ok && devices[vlan.ParentIndex] && (devices[vlan.Index] || allowedVlans[vlan.VlanId]) {\n\t\t\tvlansByIfIndex[vlan.ParentIndex] = append(vlansByIfIndex[vlan.ParentIndex], vlan.VlanId)\n\t\t}\n\t}\n\n\tvlansCount := 0\n\tfor _, v := range vlansByIfIndex {\n\t\tvlansCount += len(v)\n\t\tsort.Ints(v) // sort Vlanids in-place since netlink.LinkList() may return them in any order\n\t}\n\n\tif vlansCount == 0 {\n\t\treturn \"return false\", nil\n\t} else if vlansCount > 5 {\n\t\treturn \"\", fmt.Errorf(\"allowed VLAN list is too big - %d entries, please use '--vlan-bpf-bypass 0' in order to allow all available VLANs\", vlansCount)\n\t} else {\n\t\tvlanFilterTmpl := template.Must(template.New(\"vlanFilter\").Parse(\n\t\t\t`switch (ifindex) { \\\n{{range $ifindex,$vlans := . -}} case {{$ifindex}}: \\\nswitch (vlan_id) { \\\n{{range $vlan := $vlans -}} case {{$vlan}}: \\\n{{end}}return true; \\\n} \\\nbreak; \\\n{{end}}} \\\nreturn false;`))\n\n\t\tvar vlanFilterMacro bytes.Buffer\n\t\tif err := vlanFilterTmpl.Execute(&vlanFilterMacro, vlansByIfIndex); err != nil {\n\t\t\treturn \"\", fmt.Errorf(\"failed to execute template: %q\", err)\n\t\t}\n\n\t\treturn vlanFilterMacro.String(), nil\n\t}\n}", "func makedev(major int, minor int) int {\n\t// Formular from https://github.com/lattera/glibc/blob/master/sysdeps/unix/sysv/linux/makedev.c\n\t// In contrast to glibc, Go only uses int instead of unsigned long long.\n\treturn (minor & 0xff) | ((major & 0xfff) << 8) |\n\t\t((minor &^ 0xff) << 12) |\n\t\t((major &^ 0xfff) << 32)\n}", "func devminor(device uint64) uint64 {\n\treturn (device & 0xff) | ((device >> 12) & 0xfff00)\n}", "func printNetdev(v *gocui.View, s stat.Netdevs) error {\n\t// print header\n\t_, err := fmt.Fprintf(v, \"\\033[30;47m Interface: rMbps wMbps rPk/s wPk/s rAvs wAvs IErr OErr Coll Sat %%rUtil %%wUtil %%Util\\033[0m\\n\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor i := 0; i < len(s); i++ {\n\t\t// skip interfaces which never seen packets\n\t\tif s[i].Packets == 0 {\n\t\t\tcontinue\n\t\t}\n\n\t\t// print stats\n\t\t_, err := fmt.Fprintf(v, \"%20s%8.2f%8.2f%9.2f%9.2f%9.2f%9.2f%9.2f%9.2f%9.2f%9.2f%9.2f%9.2f%9.2f\\n\",\n\t\t\ts[i].Ifname,\n\t\t\ts[i].Rbytes/1024/128, s[i].Tbytes/1024/128, // conversion to Mbps\n\t\t\ts[i].Rpackets, s[i].Tpackets, s[i].Raverage, s[i].Taverage,\n\t\t\ts[i].Rerrs, s[i].Terrs, s[i].Tcolls,\n\t\t\ts[i].Saturation, s[i].Rutil, s[i].Tutil, s[i].Utilization,\n\t\t)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func genNodeDev(id nodes.ID, n *nodes.Node) (nodeDev, error) {\n\tr, ok := typesMap[reflect.TypeOf(n.Config).Elem()]\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"unknown type for %T\", n.Config)\n\t}\n\tv := reflect.New(r)\n\te := v.Elem()\n\te.Field(0).Set(reflect.ValueOf(NodeBase{id: id, name: n.Name, typ: n.Type()}))\n\te.Field(1).Set(reflect.ValueOf(n.Config))\n\t/*\n\t\tswitch v := n.Config.(type) {\n\t\tcase *nodes.Anim1D:\n\t\t\td.nodes[id] = &anim1DDev{NodeBase: b, cfg: v}\n\t\tcase *nodes.Button:\n\t\t\td.nodes[id] = &buttonDev{nodeBase: b, cfg: v}\n\t\tcase *nodes.Display:\n\t\t\td.nodes[id] = &displayDev{nodeBase: b, cfg: v}\n\t\tcase *nodes.IR:\n\t\t\td.nodes[id] = &irDev{nodeBase: b, cfg: v}\n\t\tcase *nodes.PIR:\n\t\t\td.nodes[id] = &pirDev{nodeBase: b, cfg: v}\n\t\tcase *nodes.Sound:\n\t\t\td.nodes[id] = &soundDev{nodeBase: b, cfg: v}\n\t\tdefault:\n\t\t\tpubErr(dbus, \"failed to initialize: unknown node %q: %T\", id, n)\n\t\t\treturn fmt.Errorf(\"unknown node %q: %T\", id, n)\n\t\t}\n\t*/\n\treturn v.Interface().(nodeDev), nil\n}", "func Mkdev(major int64, minor int64) uint32 {\n\treturn uint32(((minor & 0xfff00) << 12) | ((major & 0xfff) << 8) | (minor & 0xff))\n}", "func getDeviceList() *[]device {\n\tdevices := make([]device, 0)\n\n\tfor _, group := range Conf.DevGroups {\n\n\t\t// for current group, append devices to the list\n\t\tfor i := 1; i <= group.DeviceNum; i++ {\n\t\t\tdevices = append(devices, device{group.Prefix + strconv.Itoa(i), group.Firmware, false, group.IoTHub})\n\t\t}\n\t}\n\n\treturn &devices\n}", "func devmajor(device uint64) uint64 {\n\treturn (device >> 8) & 0xfff\n}", "func ProcNetDev(c *gin.Context) {\n\tres := CmdExec(\"cat /proc/net/dev | sed '1,2d' | grep -v lo | awk '{ print $1,$2,$10 }'\")\n\tvar resSlice = make(map[string][2]int64)\n\tvar temp = [2]int64{0, 0}\n\tvar tempName string\n\tfor _, v := range res {\n\t\tcomponent := strings.Split(v, \" \")\n\t\tif component[1] != \"0\" && component[2] != \"0\" {\n\t\t\ttempName = strings.Replace(component[0], \":\", \"\", -1)\n\t\t\tfmt.Println(tempName)\n\t\t\ttemp[0], _ = strconv.ParseInt(component[1], 10, 64)\n\t\t\ttemp[1], _ = strconv.ParseInt(component[2], 10, 64)\n\t\t\tresSlice[tempName] = temp\n\t\t}\n\t}\n\tc.JSON(http.StatusOK, resSlice)\n}", "func DevInit(configmapPath string) error {\n\tdevices = make(map[string]*globals.BleDev)\n\tmodels = make(map[string]mappercommon.DeviceModel)\n\tprotocols = make(map[string]mappercommon.Protocol)\n\treturn configmap.Parse(configmapPath, devices, models, protocols)\n}", "func makeDevices(c container) {\n\t// Standard file descriptors.\n\terr := os.Symlink(\"/proc/self/fd\", c.root(\"dev\", \"fd\"))\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Error symlink fd: %s\\n\", err))\n\t}\n\n\tfor i, dev := range []string{\"stdin\", \"stdout\", \"stderr\"} {\n\t\terr := os.Symlink(fmt.Sprintf(\"/proc/self/fd/%d\", i), c.root(\"dev\", dev))\n\t\tif err != nil {\n\t\t\tpanic(fmt.Sprintf(\"Error symlink %s: %s\\n\", dev, err))\n\t\t}\n\t}\n\n\t// Special devices.\n\tdevices := []struct {\n\t\tname string\n\t\tkind uint32\n\t\tperms uint32\n\t\tmajor uint32\n\t\tminor uint32\n\t}{\n\t\t{name: \"null\", kind: syscall.S_IFCHR, perms: 0666, major: 1, minor: 3},\n\t\t{name: \"zero\", kind: syscall.S_IFCHR, perms: 0666, major: 1, minor: 5},\n\t\t{name: \"full\", kind: syscall.S_IFCHR, perms: 0666, major: 1, minor: 7},\n\t\t{name: \"random\", kind: syscall.S_IFCHR, perms: 0666, major: 1, minor: 8},\n\t\t{name: \"urandom\", kind: syscall.S_IFCHR, perms: 0666, major: 1, minor: 9},\n\t\t{name: \"tty\", kind: syscall.S_IFCHR, perms: 0666, major: 5, minor: 0},\n\t\t{name: \"ptmx\", kind: syscall.S_IFCHR, perms: 0666, major: 5, minor: 2},\n\t}\n\n\tfor _, dev := range devices {\n\t\tdevice := int(unix.Mkdev(dev.major, dev.minor))\n\n\t\terr := syscall.Mknod(c.root(\"dev\", dev.name), dev.kind|dev.perms, device)\n\t\tif err != nil {\n\t\t\tpanic(fmt.Sprintf(\"Error mknod %s: %s\\n\", dev.name, err))\n\t\t}\n\n\t}\n\n\t// TODO: this is flaky due to the potential (lack of) existence of /dev/pts/0.\n\t// bindMountConsole(c)\n}", "func generateHostMacros(replica *chiv1.ChiClusterLayoutShardReplica) string {\n\tb := &bytes.Buffer{}\n\n\t// <yandex>\n\tfprintf(b, \"<%s>\\n\", xmlTagYandex)\n\t// <macros>\n\tfprintf(b, \"%4s<macros>\\n\", \" \")\n\n\t// <installation>CHI name</installation>\n\tfprintf(b, \"%8s<installation>%s</installation>\\n\", \" \", replica.Address.ChiName)\n\n\t// <CLUSTER_NAME>cluster name</CLUSTER_NAME>\n\tfprintf(b,\n\t\t\"%8s<%s>%[2]s</%[2]s>\\n\",\n\t\t\" \",\n\t\treplica.Address.ClusterName,\n\t)\n\t// <CLUSTER_NAME-shard>0-based shard index within cluster</CLUSTER_NAME-shard>\n\tfprintf(b,\n\t\t\"%8s<%s-shard>%d</%[2]s-shard>\\n\",\n\t\t\" \",\n\t\treplica.Address.ClusterName,\n\t\treplica.Address.ShardIndex,\n\t)\n\n\t// One Shard All Replicas Cluster\n\t// <CLUSTER_NAME>cluster name</CLUSTER_NAME>\n\tfprintf(b,\n\t\t\"%8s<%s>%[2]s</%[2]s>\\n\",\n\t\t\" \",\n\t\toneShardAllReplicasClusterName,\n\t)\n\t// <CLUSTER_NAME-shard>0-based shard index within one-shard-cluster would always be 0</CLUSTER_NAME-shard>\n\tfprintf(b,\n\t\t\"%8s<%s-shard>%d</%[2]s-shard>\\n\",\n\t\t\" \",\n\t\toneShardAllReplicasClusterName,\n\t\t0,\n\t)\n\n\t// All Shards One Replica Cluster\n\t// <CLUSTER_NAME>cluster name</CLUSTER_NAME>\n\tfprintf(b,\n\t\t\"%8s<%s>%[2]s</%[2]s>\\n\",\n\t\t\" \",\n\t\tallShardsOneReplicaClusterName,\n\t)\n\t// <CLUSTER_NAME-shard>0-based shard index within all-shards-one-replica-cluster would always be GlobalReplicaIndex</CLUSTER_NAME-shard>\n\tfprintf(b,\n\t\t\"%8s<%s-shard>%d</%[2]s-shard>\\n\",\n\t\t\" \",\n\t\tallShardsOneReplicaClusterName,\n\t\treplica.Address.GlobalReplicaIndex,\n\t)\n\n\t// <replica>replica id = full deployment id</replica>\n\t// full deployment id is unique to identify replica within the cluster\n\tfprintf(b, \"%8s<replica>%s</replica>\\n\", \" \", CreatePodHostname(replica))\n\n\t// \t\t</macros>\n\t// </yandex>\n\tfprintf(b, \"%4s</macros>\\n\", \" \")\n\tfprintf(b, \"</%s>\\n\", xmlTagYandex)\n\n\treturn b.String()\n}", "func linuxDevice(d *configs.Device) specs.LinuxDevice {\n\treturn specs.LinuxDevice{\n\t\tType: string(d.Type),\n\t\tPath: d.Path,\n\t\tMajor: d.Major,\n\t\tMinor: d.Minor,\n\t\tFileMode: fmPtr(int64(d.FileMode)),\n\t\tUID: u32Ptr(int64(d.Uid)),\n\t\tGID: u32Ptr(int64(d.Gid)),\n\t}\n}", "func printDev() {\n\t// Get network interfaces\n\tifaces, err := net.Interfaces()\n\tif err != nil {\n\t\tlog.Fatal(\"Could not get network interfaces\")\n\t}\n\n\t// Parse slice into quotation-mark- and newline-delimited string\n\tvar ifaceStr string\n\tfor i := 0; i < len(ifaces); i++ {\n\t\tifaceStr += \"\\\"\" + ifaces[i].Name + \"\\\"\\n\"\n\t}\n\n\tfmt.Printf(\"Available network interfaces:\\n%s\\n\", ifaceStr)\n}", "func mkdevnull() {\n\tdir := os.ExpandEnv(\"$PLAN9/src/cmd/venti\")\n\tfor _, a := range [][]string{\n\t\t{\"9c\", \"devnull.c\"},\n\t\t{\"9l\", \"-o\", \"o.devnull\", \"devnull.o\"},\n\t} {\n\t\tcmd := exec.Command(a[0], a[1:]...)\n\t\tcmd.Dir = dir\n\t\tif err := cmd.Run(); err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t}\n}", "func namesFor(dev string) []string {\n\tvar names []string\n\tnames = append(names, dev)\n\tnames = append(names, PathFor(dev))\n\n\t// Check for a symlink\n\ts, err := filepath.EvalSymlinks(dev)\n\tif err != nil {\n\t\tklog.Infof(\"device %q did not evaluate as a symlink: %v\", dev, err)\n\t} else {\n\t\ta, err := filepath.Abs(s)\n\t\tif err != nil {\n\t\t\tklog.Warningf(\"unable to make filepath %q absolute: %v\", s, err)\n\t\t} else {\n\t\t\ts = a\n\t\t}\n\t\tnames = append(names, s)\n\t}\n\n\treturn names\n}", "func createDevicesMap() map[string]string {\n\n\tdevices := map[string]string{}\n\n\tserials, err := listDeviceSerials()\n\tif err != nil {\n\t\treturn devices\n\t}\n\n\twg := new(sync.WaitGroup)\n\tmutex := new(sync.Mutex)\n\tfor _, serial := range serials {\n\t\twg.Add(1)\n\t\tgo func(serial string, mutex *sync.Mutex, waitGroup *sync.WaitGroup) {\n\t\t\tstdOut := &bytes.Buffer{}\n\t\t\tparams := Params{args: []string{\"-s\", serial, \"shell\", \"cat\", \"/system/build.prop\"}, stdOut:stdOut}\n\t\t\terr := execAdb(params)\n\t\t\tif err != nil {\n\t\t\t\twg.Done()\n\t\t\t\treturn\n\t\t\t}\n\t\t\tmutex.Lock()\n\t\t\tdefer mutex.Unlock()\n\t\t\tfor _, line := range strings.Split(stdOut.String(), \"\\n\") {\n\t\t\t\tif strings.HasPrefix(line, \"ro.product.model=\") {\n\t\t\t\t\tname := strings.TrimSpace(strings.Split(line, \"=\")[1])\n\t\t\t\t\tdevices[name] = serial\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\twg.Done()\n\t\t}(serial, mutex, wg)\n\t}\n\twg.Wait()\n\n\treturn devices\n}", "func SymbolVarCfg() {\n\tfmt.Println(\"[start]read symbol list from file and write to code\")\n\tdefer fmt.Println(\"[end]read symbol list from file and write to code\")\n\n\tcharMap := map[rune]bool{}\n\tdatas, err := smn_file.FileReadAll(\"./datas/analysis/lex_pgl/symbol.cfg\")\n\tcheck(err)\n\twritecv(`package lex_pgl\n\nvar SymbolList = map[string]bool{`)\n\n\tsmbList := strings.Split(string(datas), \"\\n\")\n\tfor i := range smbList {\n\t\tsmbList[i] = strings.TrimSpace(smbList[i])\n\t\tsmbList[i] = strings.Replace(smbList[i], \"\\\\\", \"\\\\\\\\\", -1)\n\t\tline := smbList[i]\n\n\t\tif line == \"\" {\n\t\t\tcontinue\n\t\t}\n\n\t\tfor _, char := range line {\n\t\t\tcharMap[char] = true\n\t\t}\n\n\t\twritecvf(\"\\\"%s\\\":true,\", line)\n\t}\n\n\tcharList := make(RuneList, 0, len(charMap))\n\tfor char := range charMap {\n\t\tcharList = append(charList, char)\n\t}\n\n\tsort.Sort(charList)\n\twritecv(`}\n\nvar SymbolCharSet = map[rune]bool{`)\n\n\tfor _, char := range charList {\n\t\tif char == '\\\\' {\n\t\t\twritecvf(`'\\\\':true,`)\n\t\t} else {\n\t\t\twritecvf(\"'%c':true,\", char)\n\t\t}\n\t}\n\n\tccMap := map[string]bool{\"\": true}\n\n\twritecv(`}\n\nvar SymbolCanContinue = map[string]bool{`)\n\n\tfor _, c1 := range smbList {\n\t\tfor _, c2 := range smbList {\n\t\t\tif ccMap[c2] || c1 == c2 {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tif strings.HasPrefix(c1, c2) {\n\t\t\t\twritecvf(\"\\\"%s\\\":true, \", c2)\n\n\t\t\t\tccMap[c2] = true\n\t\t\t}\n\t\t}\n\t}\n\n\twritecv(`}\n\n//some maybe define in another type, but not as symbol. like comment's \"//\" and \"/*\"\nvar SymbolUnuse = map[string]bool{\"//\":true, \"/*\":true}\n`)\n}", "func genMobilePixel6UA() string {\n\tandroid := pixel6AndroidVersions[rand.Intn(len(pixel6AndroidVersions))]\n\tchrome := chromeVersions[rand.Intn(len(chromeVersions))]\n\treturn fmt.Sprintf(\"Mozilla/5.0 (Linux; Android %s; Pixel 6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/%s Safari/537.36\", android, chrome)\n}", "func ListDeviceNames(withDescription bool, withIP bool) ([]string, error) {\n\tdevices, err := pcap.FindAllDevs()\n\tif err != nil {\n\t\treturn []string{}, err\n\t}\n\n\tret := []string{}\n\tfor _, dev := range devices {\n\t\tr := dev.Name\n\n\t\tif withDescription {\n\t\t\tdesc := \"No description available\"\n\t\t\tif len(dev.Description) > 0 {\n\t\t\t\tdesc = dev.Description\n\t\t\t}\n\t\t\tr += fmt.Sprintf(\" (%s)\", desc)\n\t\t}\n\n\t\tif withIP {\n\t\t\tips := \"Not assigned ip address\"\n\t\t\tif len(dev.Addresses) > 0 {\n\t\t\t\tips = \"\"\n\n\t\t\t\tfor i, address := range []pcap.InterfaceAddress(dev.Addresses) {\n\t\t\t\t\t// Add a space between the IP address.\n\t\t\t\t\tif i > 0 {\n\t\t\t\t\t\tips += \" \"\n\t\t\t\t\t}\n\n\t\t\t\t\tips += fmt.Sprintf(\"%s\", address.IP.String())\n\t\t\t\t}\n\t\t\t}\n\t\t\tr += fmt.Sprintf(\" (%s)\", ips)\n\n\t\t}\n\t\tret = append(ret, r)\n\t}\n\treturn ret, nil\n}", "func NewDev(ifName string, frameFilter FrameFilter) (dev Dev, err error) {\n\td := new(bpfDev)\n\td.name = ifName\n\td.filter = frameFilter\n\td.fd, err = getBpfFd()\n\tif err != nil {\n\t\treturn\n\t}\n\terr = ifReq(d.fd, ifName)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tvar bufLen int\n\tbufLen, err = ioCtl(d.fd)\n\tif err != nil {\n\t\treturn\n\t}\n\t_, err = d.getMTU()\n\tif err != nil {\n\t\treturn\n\t}\n\t_, err = d.getHardwareAddr()\n\tif err != nil {\n\t\treturn\n\t}\n\n\td.readBuf = make([]byte, bufLen)\n\n\tdev = d\n\n\treturn\n}", "func (u *U6) DeviceDesc() DeviceDesc {\n\treturn u.config\n}", "func ListDevices() ([]image.Device, error) {\n\tcmd := exec.Command(\"gst-device-monitor-1.0\")\n\tbuf, err := cmd.Output()\n\tif err != nil {\n\t\tif errors.Is(err, exec.ErrNotFound) {\n\t\t\terr = errInstallHint\n\t\t}\n\t\treturn nil, fmt.Errorf(\"listing devices using gst-device-monitor-1.0: %v\", err)\n\t}\n\n\tvar r []device\n\tvar d *device\n\tb := bufio.NewScanner(bytes.NewReader(buf))\n\tfor b.Scan() {\n\t\ts := strings.TrimSpace(b.Text())\n\t\tif s == \"\" {\n\t\t\tcontinue\n\t\t}\n\t\tif s == \"Device found:\" {\n\t\t\tif d != nil {\n\t\t\t\tr = append(r, *d)\n\t\t\t}\n\t\t\td = &device{RawCaps: []string{}, Caps: []image.DeviceCap{}}\n\t\t\tcontinue\n\t\t}\n\n\t\tif d == nil {\n\t\t\tcontinue\n\t\t}\n\n\t\tif strings.HasPrefix(s, \"name :\") {\n\t\t\td.Name = strings.TrimSpace(strings.SplitN(s, \":\", 2)[1])\n\t\t\tcontinue\n\t\t}\n\t\tif strings.HasPrefix(s, \"class :\") {\n\t\t\td.DeviceClass = strings.TrimSpace(strings.SplitN(s, \":\", 2)[1])\n\t\t\tcontinue\n\t\t}\n\t\tif strings.HasPrefix(s, \"caps :\") {\n\t\t\tcap := strings.TrimSpace(strings.SplitN(s, \":\", 2)[1])\n\t\t\td.RawCaps = append(d.RawCaps, cap)\n\t\t\td.inCapMode = true\n\t\t\tcontinue\n\t\t}\n\t\tif strings.HasPrefix(s, \"properties:\") {\n\t\t\td.inCapMode = false\n\t\t\tcontinue\n\t\t}\n\t\tif d.inCapMode {\n\t\t\td.RawCaps = append(d.RawCaps, s)\n\t\t}\n\t\tif strings.HasPrefix(s, \"device.path =\") {\n\t\t\td.ID = strings.TrimSpace(strings.SplitN(s, \"=\", 2)[1])\n\t\t}\n\t}\n\tif err := b.Err(); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif d != nil && d.ID != \"\" {\n\t\tr = append(r, *d)\n\t}\n\n\tvar devs []image.Device\n\tfor _, d := range r {\n\t\tif d.DeviceClass != \"Video/Source\" {\n\t\t\tcontinue\n\t\t}\n\t\tfor _, rc := range d.RawCaps {\n\t\t\tif !strings.HasPrefix(rc, \"video/x-raw\") {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tmw := widthRegexp.FindStringSubmatch(rc)\n\t\t\tmh := heightRegexp.FindStringSubmatch(rc)\n\t\t\tmf := framerateRegexp.FindStringSubmatch(rc)\n\t\t\tif mw == nil || mh == nil || mf == nil {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\twidth, werr := strconv.ParseInt(mw[1], 10, 32)\n\t\t\theight, herr := strconv.ParseInt(mh[1], 10, 32)\n\t\t\tframerate, ferr := strconv.ParseInt(mf[1], 10, 32)\n\t\t\tif werr != nil || herr != nil || ferr != nil {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif width != 0 && height != 0 && framerate != 0 {\n\t\t\t\td.Caps = append(d.Caps, image.DeviceCap{\n\t\t\t\t\tWidth: int(width),\n\t\t\t\t\tHeight: int(height),\n\t\t\t\t\tFramerate: int(framerate),\n\t\t\t\t})\n\t\t\t}\n\t\t}\n\t\tif len(d.Caps) == 0 {\n\t\t\tcontinue\n\t\t}\n\n\t\tdistance := func(a image.DeviceCap) int {\n\t\t\treturn abs(a.Width-640)*abs(a.Height-480) + abs(a.Width-640) + abs(a.Height-480)\n\t\t}\n\n\t\tsort.Slice(d.Caps, func(i, j int) bool {\n\t\t\treturn distance(d.Caps[i]) < distance(d.Caps[j])\n\t\t})\n\n\t\tdevs = append(devs, image.Device{\n\t\t\tID: d.ID,\n\t\t\tName: d.Name,\n\t\t\tCaps: d.Caps,\n\t\t})\n\t}\n\tif len(devs) == 0 {\n\t\treturn nil, fmt.Errorf(\"no devices found\")\n\t}\n\n\treturn devs, nil\n}", "func buildSchema(config *Config) string {\n\tvar sb strings.Builder\n\tsb.WriteString(\"WIFI:S:\")\n\tsb.WriteString(config.SSID)\n\tsb.WriteString(\";T:\")\n\tsb.WriteString(config.Encryption)\n\tsb.WriteString(\";P:\")\n\tsb.WriteString(config.Key)\n\tsb.WriteString(\";H:\")\n\tsb.WriteString(strconv.FormatBool(config.Hidden))\n\tsb.WriteString(\";\")\n\treturn sb.String()\n}", "func ListDevicePath(project string, expedition string) string {\n\tparam0 := project\n\tparam1 := expedition\n\n\treturn fmt.Sprintf(\"/projects/@/%s/expeditions/@/%s/sources/devices\", param0, param1)\n}", "func ConfigureDevice(devName string) {\n\tfmt.Println(\"Not yet implemented\")\n}", "func setupDevices(ctx context.Context, c *Container, s *specs.Spec) error {\n\tvar devs []specs.LinuxDevice\n\tdevPermissions := s.Linux.Resources.Devices\n\tif c.HostConfig.Privileged {\n\t\thostDevices, err := devices.HostDevices()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfor _, d := range hostDevices {\n\t\t\tdevs = append(devs, linuxDevice(d))\n\t\t}\n\t\tdevPermissions = []specs.LinuxDeviceCgroup{\n\t\t\t{\n\t\t\t\tAllow: true,\n\t\t\t\tAccess: \"rwm\",\n\t\t\t},\n\t\t}\n\t} else {\n\t\tfor _, deviceMapping := range c.HostConfig.Devices {\n\t\t\tif !opts.ValidateDeviceMode(deviceMapping.CgroupPermissions) {\n\t\t\t\treturn fmt.Errorf(\"%s invalid device mode: %s\", deviceMapping.PathOnHost, deviceMapping.CgroupPermissions)\n\t\t\t}\n\t\t\td, dPermissions, err := devicesFromPath(deviceMapping.PathOnHost, deviceMapping.PathInContainer, deviceMapping.CgroupPermissions)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tdevs = append(devs, d...)\n\t\t\tdevPermissions = append(devPermissions, dPermissions...)\n\t\t}\n\t}\n\n\ts.Linux.Devices = append(s.Linux.Devices, devs...)\n\ts.Linux.Resources.Devices = append(s.Linux.Resources.Devices, devPermissions...)\n\treturn nil\n}", "func BuildOVSPortExternalIDs(containerConfig *interfacestore.InterfaceConfig) map[string]interface{} {\n\texternalIDs := make(map[string]interface{})\n\texternalIDs[ovsExternalIDMAC] = containerConfig.MAC.String()\n\texternalIDs[ovsExternalIDContainerID] = containerConfig.ContainerID\n\texternalIDs[ovsExternalIDIP] = containerConfig.IP.String()\n\texternalIDs[ovsExternalIDPodName] = containerConfig.PodName\n\texternalIDs[ovsExternalIDPodNamespace] = containerConfig.PodNamespace\n\treturn externalIDs\n}", "func DefaultMacros() []LBuiltinDef {\n\tops := make([]LBuiltinDef, len(langMacros)+len(userMacros))\n\tfor i := range langMacros {\n\t\tops[i] = langMacros[i]\n\t}\n\toffset := len(langMacros)\n\tfor i := range userMacros {\n\t\tops[offset+i] = langMacros[i]\n\t}\n\treturn ops\n}", "func getNbdDeviceList() ([]string, error) {\n\tif nbdEnabled, _ := checkNbdModule(); !nbdEnabled {\n\t\terr := loadNbd()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\tcmd := exec.Command(\"bash\", \"-c\", \"ls /dev/nbd*\")\n\toba, err := cmd.CombinedOutput()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tostr := string(oba)\n\n\t//now split the results on space...\n\n\toarr := strings.Split(ostr, \"\\n\")\n\n\tresArr := []string{}\n\tfor _, item := range oarr {\n\t\tif len(item) > 0 {\n\t\t\tmatched, _ := regexp.Match(`^/dev/nbd\\d+$`, []byte(item))\n\t\t\tif matched {\n\t\t\t\tresArr = append(resArr, item)\n\t\t\t\t//quick chown...\n\t\t\t\tcmd := vutils.Exec.CreateAsyncCommand(\"chown\", false, fmt.Sprintf(\"%d\", UID), item).Sudo()\n\t\t\t\tif err := cmd.StartAndWait(); err != nil {\n\t\t\t\t\treturn nil, err\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\treturn resArr, nil\n}", "func SPIDevVarOnFlagSet(f *flag.FlagSet, p *string, name, value, usage string) {\n\tdevices := spireg.All()\n\tvar def string\n\tvar names []string\n\tfor _, d := range devices {\n\t\tnames = append(names, d.Name)\n\t\tnames = append(names, d.Aliases...)\n\t\tif def == \"\" {\n\t\t\tif d.Name == value {\n\t\t\t\tdef = d.Name\n\t\t\t} else {\n\t\t\t\tfor _, a := range d.Aliases {\n\t\t\t\t\tif a == value {\n\t\t\t\t\t\tdef = d.Name\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tif def == \"\" && len(names) > 0 {\n\t\tdef = names[0]\n\t}\n\tf.StringVar(p, name, def, fmt.Sprintf(\"%s; available devices: %v\", usage, names))\n}", "func GenMac(oui string) string {\n\tbuf := make([]byte, 3)\n\t_, _ = rand.Read(buf)\n\treturn fmt.Sprintf(\"%s:%02x:%02x:%02x\", oui, buf[0], buf[1], buf[2])\n}", "func FindDeviceNodes() (map[string]string, error) {\n\tnodes := make(map[string]string)\n\terr := filepath.Walk(\"/dev\", func(path string, info os.FileInfo, err error) error {\n\t\tif err != nil {\n\t\t\tlogrus.Warnf(\"Error descending into path %s: %v\", path, err)\n\t\t\treturn filepath.SkipDir\n\t\t}\n\n\t\t// If we aren't a device node, do nothing.\n\t\tif info.Mode()&(os.ModeDevice|os.ModeCharDevice) == 0 {\n\t\t\treturn nil\n\t\t}\n\n\t\t// We are a device node. Get major/minor.\n\t\tsysstat, ok := info.Sys().(*syscall.Stat_t)\n\t\tif !ok {\n\t\t\treturn errors.Errorf(\"Could not convert stat output for use\")\n\t\t}\n\t\tmajor := sysstat.Rdev / 256\n\t\tminor := sysstat.Rdev % 256\n\n\t\tnodes[fmt.Sprintf(\"%d:%d\", major, minor)] = path\n\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn nodes, nil\n}", "func (sb *spdkBackend) formatKdev(req *storage.BdevFormatRequest) (*storage.BdevFormatResponse, error) {\n\tresp := &storage.BdevFormatResponse{\n\t\tDeviceResponses: make(storage.BdevDeviceFormatResponses),\n\t}\n\n\tfor _, device := range req.Properties.DeviceList.Devices() {\n\t\tresp.DeviceResponses[device] = new(storage.BdevDeviceFormatResponse)\n\t\tsb.log.Debugf(\"%s format for non-NVMe bdev skipped on %s\", req.Properties.Class, device)\n\t}\n\n\treturn resp, nil\n}", "func (util *multiPathUtil) FindMultiPathDevForDev(device string) (string, error) {\n\tio := &osIOHandler{}\n\n\tdisk, err := findDeviceForPath(device, io)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tsysPath := \"/sys/block/\"\n\tdirs, err := io.ReadDir(sysPath)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"read directory %s failed\", sysPath)\n\t}\n\n\tfor _, f := range dirs {\n\t\tname := f.Name()\n\t\tif strings.HasPrefix(name, \"dm-\") {\n\t\t\tif _, err1 := io.Lstat(sysPath + name + \"/slaves/\" + disk); err1 == nil {\n\t\t\t\treturn \"/dev/\" + name, nil\n\t\t\t}\n\t\t}\n\t}\n\n\treturn \"\", fmt.Errorf(\"no multipath device find for %s\", device)\n}", "func (*DeviceType) Descriptor() ([]byte, []int) {\n\treturn file_register_registerpb_register_proto_rawDescGZIP(), []int{3}\n}", "func getAttachedDevices(existing *map[string]*Device) (string, error) {\n\tdevName := \"\"\n\tif runtime.GOOS == \"windows\" {\n\t\tinfo, err := ioutil.ReadDir(\"config/\")\n\t\tif err != nil {\n\t\t\treturn \"\", fmt.Errorf(\"unable to get COM info from 'config/': %v\\n\", err)\n\t\t}\n\n\t\tfound := false\n\t\tfor _, f := range info {\n\t\t\tif strings.HasPrefix(f.Name(), \"__COM\") {\n\t\t\t\tdevName = strings.Trim(f.Name(), \"__\")\n\t\t\t\tdevName = strings.Trim(devName, \".txt\")\n\t\t\t\tfound = true\n\t\t\t}\n\t\t}\n\n\t\tif !found {\n\t\t\treturn \"\", fmt.Errorf(\"__COM device not found\")\n\t\t}\n\t} else if runtime.GOOS == \"linux\" {\n\t\tdevName = \"/dev/ttyACM0\"\n\t} else {\n\t\tdevName = \"/dev/tty.usbmodem001\" //\tdefault to OS X\n\t}\n\n\tif (*existing)[devName] != nil {\n\t\t// this means we should be tracking the device\n\t\t// but if the err says that we can't find the device\n\t\t// then we need to let it go\n\t\t_, err := serial.OpenPort(devName, DEFBAUD)\n\t\tif err != nil && strings.HasSuffix(err.Error(), NSF) {\n\t\t\tdelete(*existing, devName)\n\t\t\treturn \"\", fmt.Errorf(devName + \" device removed\")\n\t\t}\n\n\t\treturn \"\", nil\n\t}\n\n\td, err := serial.OpenPort(devName, DEFBAUD)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tinfo, err := getFirmwareInfo(&d)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif len(info) > 1 {\n\t\tpos := position{\n\t\t\tX: 0,\n\t\t\tY: 0,\n\t\t\tZ: 0,\n\t\t\tE1: 0,\n\t\t}\n\n\t\tdev := &Device{\n\t\t\tName: devName,\n\t\t\tLineTerminator: makiboxA6.LineTerminator,\n\t\t\tBaud: DEFBAUD,\n\t\t\tIODevice: d,\n\t\t\tMoveSpeed: 0,\n\t\t\tPos: pos,\n\t\t\tHomed: false,\n\t\t\tGreeting: info,\n\t\t\tJobStatus: IDLE,\n\t\t}\n\n\t\t(*existing)[devName] = dev\n\t\treturn devName, nil\n\t}\n\n\treturn \"\", nil\n}", "func (fes *FrontEndService) WriteConfigKernel(conf *string, hasVIP4 bool, hasVIP6 bool) {\n\teFilter := \"none\"\n\tif hasVIP4 {\n\t\teFilter = \"filter default_v4\"\n\t}\n\n\t*conf += \"protocol kernel {\\n\"\n\t*conf += \"\\tipv4 {\\n\"\n\t*conf += \"\\t\\timport none;\\n\"\n\t*conf += \"\\t\\texport \" + eFilter + \";\\n\"\n\t*conf += \"\\t};\\n\"\n\t*conf += \"\\tkernel table \" + strconv.FormatInt(int64(fes.kernelTableId), 10) + \";\\n\"\n\tif fes.ecmp {\n\t\t*conf += \"\\tmerge paths on;\\n\"\n\t}\n\tif fes.dropIfNoPeer {\n\t\t// Setting the metric for the default blackhole route must be supported,\n\t\t// which requires the kernel proto's metric to be set to zero.\n\t\t//\n\t\t// \"Metric 0 has a special meaning of undefined metric, in which either OS default is used,\n\t\t// or per-route metric can be set using krt_metric attribute. Default: 32. \"\n\t\t*conf += \"\\tmetric 0;\\n\"\n\t}\n\t*conf += \"}\\n\"\n\t*conf += \"\\n\"\n\n\tif hasVIP6 {\n\t\teFilter = \"filter default_v6\"\n\t} else {\n\t\teFilter = \"none\"\n\t}\n\t*conf += \"protocol kernel {\\n\"\n\t*conf += \"\\tipv6 {\\n\"\n\t*conf += \"\\t\\timport none;\\n\"\n\t*conf += \"\\t\\texport \" + eFilter + \";\\n\"\n\t*conf += \"\\t};\\n\"\n\t*conf += \"\\tkernel table \" + strconv.FormatInt(int64(fes.kernelTableId), 10) + \";\\n\"\n\tif fes.ecmp {\n\t\t*conf += \"\\tmerge paths on;\\n\"\n\t}\n\tif fes.dropIfNoPeer {\n\t\t// Setting the metric for the default blackhole route must be supported,\n\t\t// which requires the kernel proto's metric to be set to zero.\n\t\t//\n\t\t// \"Metric 0 has a special meaning of undefined metric, in which either OS default is used,\n\t\t// or per-route metric can be set using krt_metric attribute. Default: 32. \"\n\t\t*conf += \"\\tmetric 0;\\n\"\n\t}\n\t*conf += \"}\\n\"\n\t*conf += \"\\n\"\n}", "func (c *pciApplyConfig) modifyVirtualPciDevices(devList *schema.Set, op types.VirtualDeviceConfigSpecOperation) error {\n\tlog.Printf(\"VirtualMachine: Creating PCI passthrough device specs %v\", op)\n\tfor _, addDev := range devList.List() {\n\t\tlog.Printf(\"[DEBUG] modifyVirtualPciDevices: Appending %v spec for %s\", op, addDev.(string))\n\t\tpciDev, err := c.getHostPciDevice(addDev.(string))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdev := &types.VirtualPCIPassthrough{\n\t\t\tVirtualDevice: types.VirtualDevice{\n\t\t\t\tDynamicData: types.DynamicData{},\n\t\t\t\tBacking: &types.VirtualPCIPassthroughDeviceBackingInfo{\n\t\t\t\t\tVirtualDeviceDeviceBackingInfo: types.VirtualDeviceDeviceBackingInfo{},\n\t\t\t\t\tId: pciDev.Id,\n\t\t\t\t\tSystemId: c.SystemID,\n\t\t\t\t\tVendorId: pciDev.VendorId,\n\t\t\t\t},\n\t\t\t},\n\t\t}\n\t\tvm, err := virtualmachine.FromUUID(c.Client, c.ResourceData.Id())\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tvprops, err := virtualmachine.Properties(vm)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t// This will only find a device for delete operations.\n\t\tfor _, vmDevP := range vprops.Config.Hardware.Device {\n\t\t\tif vmDev, ok := vmDevP.(*types.VirtualPCIPassthrough); ok {\n\t\t\t\tif vmDev.Backing.(*types.VirtualPCIPassthroughDeviceBackingInfo).Id == pciDev.Id {\n\t\t\t\t\tdev = vmDev\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tdspec, err := object.VirtualDeviceList{dev}.ConfigSpec(op)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tc.Spec = append(c.Spec, dspec...)\n\t\tc.VirtualDevice = applyDeviceChange(c.VirtualDevice, dspec)\n\t}\n\tlog.Printf(\"VirtualMachine: PCI passthrough device specs created\")\n\treturn nil\n}", "func (vm *vmQemu) generateQemuConfigFile(configISOPath string, tapDev map[string]string) (string, error) {\n\tvar sb *strings.Builder = &strings.Builder{}\n\n\t// Base config. This is common for all VMs and has no variables in it.\n\tsb.WriteString(`\n# Machine\n[machine]\ngraphics = \"off\"\ntype = \"q35\"\naccel = \"kvm\"\nusb = \"off\"\ngraphics = \"off\"\n[global]\ndriver = \"ICH9-LPC\"\nproperty = \"disable_s3\"\nvalue = \"1\"\n[global]\ndriver = \"ICH9-LPC\"\nproperty = \"disable_s4\"\nvalue = \"1\"\n[boot-opts]\nstrict = \"on\"\n# SCSI root\n[device \"qemu_pcie1\"]\ndriver = \"pcie-root-port\"\nport = \"0x10\"\nchassis = \"1\"\nbus = \"pcie.0\"\nmultifunction = \"on\"\naddr = \"0x2\"\n[device \"qemu_scsi\"]\ndriver = \"virtio-scsi-pci\"\nbus = \"qemu_pcie1\"\naddr = \"0x0\"\n# Balloon driver\n[device \"qemu_pcie2\"]\ndriver = \"pcie-root-port\"\nport = \"0x12\"\nchassis = \"2\"\nbus = \"pcie.0\"\naddr = \"0x2.0x1\"\n[device \"qemu_ballon\"]\ndriver = \"virtio-balloon-pci\"\nbus = \"qemu_pcie2\"\naddr = \"0x0\"\n# Random number generator\n[object \"qemu_rng\"]\nqom-type = \"rng-random\"\nfilename = \"/dev/urandom\"\n[device \"qemu_pcie3\"]\ndriver = \"pcie-root-port\"\nport = \"0x13\"\nchassis = \"3\"\nbus = \"pcie.0\"\naddr = \"0x2.0x2\"\n[device \"dev-qemu_rng\"]\ndriver = \"virtio-rng-pci\"\nrng = \"qemu_rng\"\nbus = \"qemu_pcie3\"\naddr = \"0x0\"\n# Console\n[chardev \"console\"]\nbackend = \"pty\"\n`)\n\n\t// Now add the dynamic parts of the config.\n\terr := vm.addMemoryConfig(sb)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\terr = vm.addRootDriveConfig(sb)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\terr = vm.addCPUConfig(sb)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tvm.addFirmwareConfig(sb)\n\tvm.addVsockConfig(sb)\n\tvm.addMonitorConfig(sb)\n\tvm.addConfDriveConfig(sb, configISOPath)\n\tvm.addNetConfig(sb, tapDev)\n\n\t// Write the config file to disk.\n\tconfigPath := filepath.Join(vm.LogPath(), \"qemu.conf\")\n\treturn configPath, ioutil.WriteFile(configPath, []byte(sb.String()), 0640)\n}", "func NormalizeBus(l object.VirtualDeviceList, d *schema.ResourceData) (object.VirtualDeviceList, []types.BaseVirtualDeviceConfigSpec, error) {\n\tscsiCount := d.Get(\"scsi_controller_count\").(int)\n\tscsiType := d.Get(\"scsi_type\").(string)\n\tscsiSharing := d.Get(\"scsi_bus_sharing\").(string)\n\tsataCount := d.Get(\"sata_controller_count\").(int)\n\tideCount := d.Get(\"ide_controller_count\").(int)\n\tvar spec []types.BaseVirtualDeviceConfigSpec\n\tscsiCtlrs := make([]types.BaseVirtualSCSIController, scsiCount)\n\tsataCtlrs := make([]types.BaseVirtualSATAController, sataCount)\n\tideCtlrs := make([]*types.VirtualIDEController, ideCount)\n\t// Don't worry about doing any fancy select stuff here, just go thru the\n\t// VirtualDeviceList and populate the controllers.\n\tlog.Printf(\"[DEBUG] NormalizeBus: Normalizing first %d controllers on SCSI bus to device type %s\", scsiCount, scsiType)\n\tlog.Printf(\"[DEBUG] NormalizeBus: Normalizing first %d controllers on SATA bus\", sataCount)\n\tlog.Printf(\"[DEBUG] NormalizeBus: Normalizing first %d controllers on IDE bus\", ideCount)\n\tfor _, dev := range l {\n\t\tswitch ctlr := dev.(type) {\n\t\tcase types.BaseVirtualSCSIController:\n\t\t\tif busNumber := ctlr.GetVirtualSCSIController().BusNumber; busNumber < int32(scsiCount) {\n\t\t\t\tscsiCtlrs[busNumber] = ctlr\n\t\t\t}\n\t\tcase types.BaseVirtualSATAController:\n\t\t\tif busNumber := ctlr.GetVirtualSATAController().BusNumber; busNumber < int32(sataCount) {\n\t\t\t\tsataCtlrs[busNumber] = ctlr\n\t\t\t}\n\t\tcase *types.VirtualIDEController:\n\t\t\tif busNumber := ctlr.GetVirtualController().BusNumber; busNumber < int32(ideCount) {\n\t\t\t\tideCtlrs[busNumber] = ctlr\n\t\t\t}\n\t\t}\n\t}\n\tlog.Printf(\"[DEBUG] NormalizeBus: Current SCSI bus contents: %s\", scsiControllerListString(scsiCtlrs))\n\t// Now iterate over the SCSI controllers\n\tfor n, ctlr := range scsiCtlrs {\n\t\tif ctlr == nil {\n\t\t\tlog.Printf(\"[DEBUG] NormalizeBus: Creating SCSI controller of type %s at bus number %d\", scsiType, n)\n\t\t\tcspec, err := createSCSIController(&l, scsiType, scsiSharing)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, nil, err\n\t\t\t}\n\t\t\tspec = append(spec, cspec...)\n\t\t\tcontinue\n\t\t}\n\t\tif l.Type(ctlr.(types.BaseVirtualDevice)) == scsiType {\n\t\t\tcspec, err := setSCSIBusSharing(&l, ctlr, scsiSharing)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, nil, err\n\t\t\t}\n\t\t\tspec = append(spec, cspec...)\n\t\t\tcontinue\n\t\t}\n\t\tcspec, err := swapSCSIDevice(l, ctlr, scsiType, scsiSharing)\n\t\tif err != nil {\n\t\t\treturn nil, nil, err\n\t\t}\n\t\tspec = append(spec, cspec...)\n\t\tl = applyDeviceChange(l, cspec)\n\t\tcontinue\n\t}\n\tlog.Printf(\"[DEBUG] NormalizeBus: Current SATA bus contents: %s\", sataControllerListString(sataCtlrs))\n\t// Now iterate over the SATA controllers\n\tfor n, ctlr := range sataCtlrs {\n\t\tif ctlr == nil {\n\t\t\tlog.Printf(\"[DEBUG] NormalizeBus: Creating SATA controller at bus number %d\", n)\n\t\t\tcspec, err := createSATAController(&l, n)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, nil, err\n\t\t\t}\n\t\t\tspec = append(spec, cspec...)\n\t\t}\n\t}\n\tlog.Printf(\"[DEBUG] NormalizeBus: Current IDE bus contents: %s\", ideControllerListString(ideCtlrs))\n\t// Now iterate over the IDE controllers\n\tfor n, ctlr := range ideCtlrs {\n\t\tif ctlr == nil {\n\t\t\tlog.Printf(\"[DEBUG] NormalizeBus: Creating IDE controller at bus number %d\", n)\n\t\t\tcspec, err := createIDEController(&l)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, nil, err\n\t\t\t}\n\t\t\tspec = append(spec, cspec...)\n\t\t}\n\t}\n\tlog.Printf(\"[DEBUG] NormalizeBus: Outgoing device list: %s\", DeviceListString(l))\n\tlog.Printf(\"[DEBUG] NormalizeBus: Outgoing device config spec: %s\", DeviceChangeString(spec))\n\treturn l, spec, nil\n}", "func getDeviceGUID(idx int) string {\n\tif idx < 0 || idx >= sdl.NumJoysticks() {\n\t\treturn \"\"\n\t}\n\n\treturn sdl.JoystickGetGUIDString(sdl.JoystickGetDeviceGUID(idx))\n}", "func BuildNics(vmProperties *mo.VirtualMachine) []Nic {\n\tnics := make([]Nic, 0)\n\n\tdevices := vmProperties.Config.Hardware.Device\n\tfor _, device := range devices {\n\t\t// is this device a VirtualEthernetCard?\n\t\tvar virtualNetwork *types.VirtualEthernetCard\n\t\tswitch v := device.(type) {\n\t\tcase *types.VirtualE1000:\n\t\t\tvirtualNetwork = &v.VirtualEthernetCard\n\t\tcase *types.VirtualE1000e:\n\t\t\tvirtualNetwork = &v.VirtualEthernetCard\n\t\tcase *types.VirtualVmxnet:\n\t\t\tvirtualNetwork = &v.VirtualEthernetCard\n\t\tcase *types.VirtualVmxnet2:\n\t\t\tvirtualNetwork = &v.VirtualEthernetCard\n\t\tcase *types.VirtualVmxnet3:\n\t\t\tvirtualNetwork = &v.VirtualEthernetCard\n\t\t}\n\t\tif virtualNetwork != nil && virtualNetwork.Backing != nil {\n\t\t\tvar network string\n\t\t\tvar dvportgroup string\n\t\t\tvar name string\n\n\t\t\tswitch backing := virtualNetwork.Backing.(type) {\n\t\t\tcase *types.VirtualEthernetCardNetworkBackingInfo:\n\t\t\t\tif backing.Network != nil {\n\t\t\t\t\tnetwork = backing.Network.Value\n\t\t\t\t}\n\t\t\t\t// despite being called DeviceName, this is actually\n\t\t\t\t// the name of the Network the device is attached to,\n\t\t\t\t// e.g. \"VM Network\"\n\t\t\t\tname = backing.DeviceName\n\t\t\tcase *types.VirtualEthernetCardDistributedVirtualPortBackingInfo:\n\t\t\t\tdvportgroup = backing.Port.PortgroupKey\n\t\t\t\tdesc := virtualNetwork.DeviceInfo.GetDescription()\n\t\t\t\tif desc != nil {\n\t\t\t\t\t// this is the actual device name, e.g. \"ethernet-0\"\n\t\t\t\t\tname = desc.Label\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tnic := Nic{\n\t\t\t\tName: name,\n\t\t\t\tMac: virtualNetwork.MacAddress,\n\t\t\t\tNetwork: network,\n\t\t\t\tDVPortGroup: dvportgroup,\n\t\t\t}\n\t\t\tnics = append(nics, nic)\n\t\t}\n\t}\n\treturn nics\n}", "func (h *HeaderfileWriter) WriteNetdevConfig(w io.Writer, cfg datapath.DeviceConfiguration) error {\n\tfw := bufio.NewWriter(w)\n\th.writeNetdevConfig(fw, cfg)\n\treturn fw.Flush()\n}", "func (d pciDevice) isVGA() bool {\n\tbootVgaFile := filepath.Join(sysfsPciDevices, d.pciLong, \"boot_vga\")\n\treturn utils.FileExists(nil, bootVgaFile)\n}", "func (h *HeaderfileWriter) WriteNodeConfig(w io.Writer, cfg *datapath.LocalNodeConfiguration) error {\n\textraMacrosMap := make(dpdef.Map)\n\tcDefinesMap := make(dpdef.Map)\n\n\tfw := bufio.NewWriter(w)\n\n\twriteIncludes(w)\n\n\trouterIP := node.GetIPv6Router()\n\thostIP := node.GetIPv6()\n\n\tfmt.Fprintf(fw, \"/*\\n\")\n\tif option.Config.EnableIPv6 {\n\t\tfmt.Fprintf(fw, \" cilium.v6.external.str %s\\n\", node.GetIPv6().String())\n\t\tfmt.Fprintf(fw, \" cilium.v6.internal.str %s\\n\", node.GetIPv6Router().String())\n\t\tfmt.Fprintf(fw, \" cilium.v6.nodeport.str %s\\n\", node.GetNodePortIPv6Addrs())\n\t\tfmt.Fprintf(fw, \"\\n\")\n\t}\n\tfmt.Fprintf(fw, \" cilium.v4.external.str %s\\n\", node.GetIPv4().String())\n\tfmt.Fprintf(fw, \" cilium.v4.internal.str %s\\n\", node.GetInternalIPv4Router().String())\n\tfmt.Fprintf(fw, \" cilium.v4.nodeport.str %s\\n\", node.GetNodePortIPv4Addrs())\n\tfmt.Fprintf(fw, \"\\n\")\n\tif option.Config.EnableIPv6 {\n\t\tfw.WriteString(dumpRaw(defaults.RestoreV6Addr, node.GetIPv6Router()))\n\t}\n\tfw.WriteString(dumpRaw(defaults.RestoreV4Addr, node.GetInternalIPv4Router()))\n\tfmt.Fprintf(fw, \" */\\n\\n\")\n\n\tcDefinesMap[\"KERNEL_HZ\"] = fmt.Sprintf(\"%d\", option.Config.KernelHz)\n\n\tif option.Config.EnableIPv6 {\n\t\textraMacrosMap[\"ROUTER_IP\"] = routerIP.String()\n\t\tfw.WriteString(defineIPv6(\"ROUTER_IP\", routerIP))\n\t}\n\n\tif option.Config.EnableIPv4 {\n\t\tipv4GW := node.GetInternalIPv4Router()\n\t\tloopbackIPv4 := node.GetIPv4Loopback()\n\t\tipv4Range := node.GetIPv4AllocRange()\n\t\tcDefinesMap[\"IPV4_GATEWAY\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(ipv4GW))\n\t\tcDefinesMap[\"IPV4_LOOPBACK\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(loopbackIPv4))\n\t\tcDefinesMap[\"IPV4_MASK\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(net.IP(ipv4Range.Mask)))\n\n\t\tif option.Config.EnableIPv4FragmentsTracking {\n\t\t\tcDefinesMap[\"ENABLE_IPV4_FRAGMENTS\"] = \"1\"\n\t\t\tcDefinesMap[\"IPV4_FRAG_DATAGRAMS_MAP\"] = fragmap.MapName\n\t\t\tcDefinesMap[\"CILIUM_IPV4_FRAG_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", option.Config.FragmentsMapEntries)\n\t\t}\n\t}\n\n\tif option.Config.EnableIPv6 {\n\t\textraMacrosMap[\"HOST_IP\"] = hostIP.String()\n\t\tfw.WriteString(defineIPv6(\"HOST_IP\", hostIP))\n\t}\n\n\tfor t, id := range tunnelProtocols {\n\t\tmacroName := fmt.Sprintf(\"TUNNEL_PROTOCOL_%s\", strings.ToUpper(t))\n\t\tcDefinesMap[macroName] = fmt.Sprintf(\"%d\", id)\n\t}\n\n\tencapProto := option.Config.TunnelProtocol\n\tif !option.Config.TunnelingEnabled() &&\n\t\toption.Config.EnableNodePort &&\n\t\toption.Config.NodePortMode != option.NodePortModeSNAT &&\n\t\toption.Config.LoadBalancerDSRDispatch == option.DSRDispatchGeneve {\n\t\tencapProto = option.TunnelGeneve\n\t}\n\n\tcDefinesMap[\"TUNNEL_PROTOCOL\"] = fmt.Sprintf(\"%d\", tunnelProtocols[encapProto])\n\tcDefinesMap[\"TUNNEL_PORT\"] = fmt.Sprintf(\"%d\", option.Config.TunnelPort)\n\n\tif tunnelDev, err := netlink.LinkByName(fmt.Sprintf(\"cilium_%s\", encapProto)); err == nil {\n\t\tcDefinesMap[\"ENCAP_IFINDEX\"] = fmt.Sprintf(\"%d\", tunnelDev.Attrs().Index)\n\t}\n\n\tcDefinesMap[\"HOST_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameHost))\n\tcDefinesMap[\"WORLD_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameWorld))\n\tif option.Config.IsDualStack() {\n\t\tcDefinesMap[\"WORLD_IPV4_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameWorldIPv4))\n\t\tcDefinesMap[\"WORLD_IPV6_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameWorldIPv6))\n\t} else {\n\t\tworldID := identity.GetReservedID(labels.IDNameWorld)\n\t\tcDefinesMap[\"WORLD_IPV4_ID\"] = fmt.Sprintf(\"%d\", worldID)\n\t\tcDefinesMap[\"WORLD_IPV6_ID\"] = fmt.Sprintf(\"%d\", worldID)\n\t}\n\tcDefinesMap[\"HEALTH_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameHealth))\n\tcDefinesMap[\"UNMANAGED_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameUnmanaged))\n\tcDefinesMap[\"INIT_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameInit))\n\tcDefinesMap[\"LOCAL_NODE_ID\"] = fmt.Sprintf(\"%d\", identity.GetLocalNodeID())\n\tcDefinesMap[\"REMOTE_NODE_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameRemoteNode))\n\tcDefinesMap[\"KUBE_APISERVER_NODE_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameKubeAPIServer))\n\tcDefinesMap[\"CILIUM_LB_SERVICE_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.ServiceMapMaxEntries)\n\tcDefinesMap[\"CILIUM_LB_BACKENDS_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.ServiceBackEndMapMaxEntries)\n\tcDefinesMap[\"CILIUM_LB_REV_NAT_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.RevNatMapMaxEntries)\n\tcDefinesMap[\"CILIUM_LB_AFFINITY_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.AffinityMapMaxEntries)\n\tcDefinesMap[\"CILIUM_LB_SOURCE_RANGE_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.SourceRangeMapMaxEntries)\n\tcDefinesMap[\"CILIUM_LB_MAGLEV_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.MaglevMapMaxEntries)\n\n\tcDefinesMap[\"TUNNEL_MAP\"] = tunnel.MapName\n\tcDefinesMap[\"TUNNEL_ENDPOINT_MAP_SIZE\"] = fmt.Sprintf(\"%d\", tunnel.MaxEntries)\n\tcDefinesMap[\"ENDPOINTS_MAP\"] = lxcmap.MapName\n\tcDefinesMap[\"ENDPOINTS_MAP_SIZE\"] = fmt.Sprintf(\"%d\", lxcmap.MaxEntries)\n\tcDefinesMap[\"METRICS_MAP\"] = metricsmap.MapName\n\tcDefinesMap[\"METRICS_MAP_SIZE\"] = fmt.Sprintf(\"%d\", metricsmap.MaxEntries)\n\tcDefinesMap[\"POLICY_MAP_SIZE\"] = fmt.Sprintf(\"%d\", policymap.MaxEntries)\n\tcDefinesMap[\"AUTH_MAP\"] = authmap.MapName\n\tcDefinesMap[\"AUTH_MAP_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.AuthMapEntries)\n\tcDefinesMap[\"CONFIG_MAP\"] = configmap.MapName\n\tcDefinesMap[\"CONFIG_MAP_SIZE\"] = fmt.Sprintf(\"%d\", configmap.MaxEntries)\n\tcDefinesMap[\"IPCACHE_MAP\"] = ipcachemap.Name\n\tcDefinesMap[\"IPCACHE_MAP_SIZE\"] = fmt.Sprintf(\"%d\", ipcachemap.MaxEntries)\n\tcDefinesMap[\"NODE_MAP\"] = nodemap.MapName\n\tcDefinesMap[\"NODE_MAP_SIZE\"] = fmt.Sprintf(\"%d\", nodemap.MaxEntries)\n\tcDefinesMap[\"SRV6_VRF_MAP4\"] = srv6map.VRFMapName4\n\tcDefinesMap[\"SRV6_VRF_MAP6\"] = srv6map.VRFMapName6\n\tcDefinesMap[\"SRV6_POLICY_MAP4\"] = srv6map.PolicyMapName4\n\tcDefinesMap[\"SRV6_POLICY_MAP6\"] = srv6map.PolicyMapName6\n\tcDefinesMap[\"SRV6_SID_MAP\"] = srv6map.SIDMapName\n\tcDefinesMap[\"SRV6_STATE_MAP4\"] = srv6map.StateMapName4\n\tcDefinesMap[\"SRV6_STATE_MAP6\"] = srv6map.StateMapName6\n\tcDefinesMap[\"SRV6_VRF_MAP_SIZE\"] = fmt.Sprintf(\"%d\", srv6map.MaxVRFEntries)\n\tcDefinesMap[\"SRV6_POLICY_MAP_SIZE\"] = fmt.Sprintf(\"%d\", srv6map.MaxPolicyEntries)\n\tcDefinesMap[\"SRV6_SID_MAP_SIZE\"] = fmt.Sprintf(\"%d\", srv6map.MaxSIDEntries)\n\tcDefinesMap[\"SRV6_STATE_MAP_SIZE\"] = fmt.Sprintf(\"%d\", srv6map.MaxStateEntries)\n\tcDefinesMap[\"WORLD_CIDRS4_MAP\"] = worldcidrsmap.MapName4\n\tcDefinesMap[\"WORLD_CIDRS4_MAP_SIZE\"] = fmt.Sprintf(\"%d\", worldcidrsmap.MapMaxEntries)\n\tcDefinesMap[\"POLICY_PROG_MAP_SIZE\"] = fmt.Sprintf(\"%d\", policymap.PolicyCallMaxEntries)\n\tcDefinesMap[\"L2_RESPONSER_MAP4_SIZE\"] = fmt.Sprintf(\"%d\", l2respondermap.DefaultMaxEntries)\n\tcDefinesMap[\"ENCRYPT_MAP\"] = encrypt.MapName\n\tcDefinesMap[\"L2_RESPONDER_MAP4\"] = l2respondermap.MapName\n\tcDefinesMap[\"CT_CONNECTION_LIFETIME_TCP\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutTCP.Seconds()))\n\tcDefinesMap[\"CT_CONNECTION_LIFETIME_NONTCP\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutAny.Seconds()))\n\tcDefinesMap[\"CT_SERVICE_LIFETIME_TCP\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutSVCTCP.Seconds()))\n\tcDefinesMap[\"CT_SERVICE_LIFETIME_NONTCP\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutSVCAny.Seconds()))\n\tcDefinesMap[\"CT_SERVICE_CLOSE_REBALANCE\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutSVCTCPGrace.Seconds()))\n\tcDefinesMap[\"CT_SYN_TIMEOUT\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutSYN.Seconds()))\n\tcDefinesMap[\"CT_CLOSE_TIMEOUT\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutFIN.Seconds()))\n\tcDefinesMap[\"CT_REPORT_INTERVAL\"] = fmt.Sprintf(\"%d\", int64(option.Config.MonitorAggregationInterval.Seconds()))\n\tcDefinesMap[\"CT_REPORT_FLAGS\"] = fmt.Sprintf(\"%#04x\", int64(option.Config.MonitorAggregationFlags))\n\tcDefinesMap[\"CT_TAIL_CALL_BUFFER4\"] = \"cilium_tail_call_buffer4\"\n\tcDefinesMap[\"CT_TAIL_CALL_BUFFER6\"] = \"cilium_tail_call_buffer6\"\n\tcDefinesMap[\"PER_CLUSTER_CT_TCP4\"] = \"cilium_per_cluster_ct_tcp4\"\n\tcDefinesMap[\"PER_CLUSTER_CT_TCP6\"] = \"cilium_per_cluster_ct_tcp6\"\n\tcDefinesMap[\"PER_CLUSTER_CT_ANY4\"] = \"cilium_per_cluster_ct_any4\"\n\tcDefinesMap[\"PER_CLUSTER_CT_ANY6\"] = \"cilium_per_cluster_ct_any6\"\n\tcDefinesMap[\"PER_CLUSTER_SNAT_MAPPING_IPV4\"] = \"cilium_per_cluster_snat_v4_external\"\n\tcDefinesMap[\"PER_CLUSTER_SNAT_MAPPING_IPV6\"] = \"cilium_per_cluster_snat_v6_external\"\n\n\tif option.Config.PreAllocateMaps {\n\t\tcDefinesMap[\"PREALLOCATE_MAPS\"] = \"1\"\n\t}\n\n\tcDefinesMap[\"EVENTS_MAP\"] = eventsmap.MapName\n\tcDefinesMap[\"SIGNAL_MAP\"] = signalmap.MapName\n\tcDefinesMap[\"POLICY_CALL_MAP\"] = policymap.PolicyCallMapName\n\tif option.Config.EnableEnvoyConfig {\n\t\tcDefinesMap[\"POLICY_EGRESSCALL_MAP\"] = policymap.PolicyEgressCallMapName\n\t}\n\tcDefinesMap[\"LB6_REVERSE_NAT_MAP\"] = \"cilium_lb6_reverse_nat\"\n\tcDefinesMap[\"LB6_SERVICES_MAP_V2\"] = \"cilium_lb6_services_v2\"\n\tcDefinesMap[\"LB6_BACKEND_MAP\"] = \"cilium_lb6_backends_v3\"\n\tcDefinesMap[\"LB6_REVERSE_NAT_SK_MAP\"] = lbmap.SockRevNat6MapName\n\tcDefinesMap[\"LB6_REVERSE_NAT_SK_MAP_SIZE\"] = fmt.Sprintf(\"%d\", lbmap.MaxSockRevNat6MapEntries)\n\tcDefinesMap[\"LB4_REVERSE_NAT_MAP\"] = \"cilium_lb4_reverse_nat\"\n\tcDefinesMap[\"LB4_SERVICES_MAP_V2\"] = \"cilium_lb4_services_v2\"\n\tcDefinesMap[\"LB4_BACKEND_MAP\"] = \"cilium_lb4_backends_v3\"\n\tcDefinesMap[\"LB4_REVERSE_NAT_SK_MAP\"] = lbmap.SockRevNat4MapName\n\tcDefinesMap[\"LB4_REVERSE_NAT_SK_MAP_SIZE\"] = fmt.Sprintf(\"%d\", lbmap.MaxSockRevNat4MapEntries)\n\n\tif option.Config.EnableSessionAffinity {\n\t\tcDefinesMap[\"ENABLE_SESSION_AFFINITY\"] = \"1\"\n\t\tcDefinesMap[\"LB_AFFINITY_MATCH_MAP\"] = lbmap.AffinityMatchMapName\n\t\tif option.Config.EnableIPv4 {\n\t\t\tcDefinesMap[\"LB4_AFFINITY_MAP\"] = lbmap.Affinity4MapName\n\t\t}\n\t\tif option.Config.EnableIPv6 {\n\t\t\tcDefinesMap[\"LB6_AFFINITY_MAP\"] = lbmap.Affinity6MapName\n\t\t}\n\t}\n\n\tcDefinesMap[\"TRACE_PAYLOAD_LEN\"] = fmt.Sprintf(\"%dULL\", option.Config.TracePayloadlen)\n\tcDefinesMap[\"MTU\"] = fmt.Sprintf(\"%d\", cfg.MtuConfig.GetDeviceMTU())\n\n\tif option.Config.EnableIPv4 {\n\t\tcDefinesMap[\"ENABLE_IPV4\"] = \"1\"\n\t}\n\n\tif option.Config.EnableIPv6 {\n\t\tcDefinesMap[\"ENABLE_IPV6\"] = \"1\"\n\t}\n\n\tif option.Config.EnableSRv6 {\n\t\tcDefinesMap[\"ENABLE_SRV6\"] = \"1\"\n\t\tif option.Config.SRv6EncapMode != \"reduced\" {\n\t\t\tcDefinesMap[\"ENABLE_SRV6_SRH_ENCAP\"] = \"1\"\n\t\t}\n\t}\n\n\tif option.Config.EnableSCTP {\n\t\tcDefinesMap[\"ENABLE_SCTP\"] = \"1\"\n\t}\n\n\tif option.Config.EnableIPSec {\n\t\tcDefinesMap[\"ENABLE_IPSEC\"] = \"1\"\n\t}\n\n\tif option.Config.EnableWireguard {\n\t\tcDefinesMap[\"ENABLE_WIREGUARD\"] = \"1\"\n\t\tifindex, err := link.GetIfIndex(wgtypes.IfaceName)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcDefinesMap[\"WG_IFINDEX\"] = fmt.Sprintf(\"%d\", ifindex)\n\n\t\tif option.Config.EncryptNode {\n\t\t\tcDefinesMap[\"ENABLE_NODE_ENCRYPTION\"] = \"1\"\n\t\t}\n\t}\n\n\tif option.Config.EnableL2Announcements {\n\t\tcDefinesMap[\"ENABLE_L2_ANNOUNCEMENTS\"] = \"1\"\n\t\t// If the agent is down for longer than the lease duration, stop responding\n\t\tcDefinesMap[\"L2_ANNOUNCEMENTS_MAX_LIVENESS\"] = fmt.Sprintf(\"%dULL\", option.Config.L2AnnouncerLeaseDuration.Nanoseconds())\n\t}\n\n\tif option.Config.EnableEncryptionStrictMode {\n\t\tcDefinesMap[\"ENCRYPTION_STRICT_MODE\"] = \"1\"\n\n\t\t// when parsing the user input we only accept ipv4 addresses\n\t\tcDefinesMap[\"STRICT_IPV4_NET\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPAddrToHost32(option.Config.EncryptionStrictModeCIDR.Addr()))\n\t\tcDefinesMap[\"STRICT_IPV4_NET_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.EncryptionStrictModeCIDR.Bits())\n\n\t\tcDefinesMap[\"IPV4_ENCRYPT_IFACE\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(node.GetIPv4()))\n\n\t\tipv4Interface, ok := netip.AddrFromSlice(node.GetIPv4().To4())\n\t\tif !ok {\n\t\t\treturn fmt.Errorf(\"unable to parse node IPv4 address %s\", node.GetIPv4())\n\t\t}\n\n\t\tif option.Config.EncryptionStrictModeCIDR.Contains(ipv4Interface) {\n\t\t\tif !option.Config.EncryptionStrictModeAllowRemoteNodeIdentities {\n\t\t\t\treturn fmt.Errorf(`encryption strict mode is enabled but the node's IPv4 address is within the strict CIDR range.\n\t\t\t\tThis will cause the node to drop all traffic.\n\t\t\t\tPlease either disable encryption or set --encryption-strict-mode-allow-dynamic-lookup=true`)\n\t\t\t}\n\t\t\tcDefinesMap[\"STRICT_IPV4_OVERLAPPING_CIDR\"] = \"1\"\n\t\t}\n\t}\n\n\tif option.Config.EnableBPFTProxy {\n\t\tcDefinesMap[\"ENABLE_TPROXY\"] = \"1\"\n\t}\n\n\tif option.Config.EnableXDPPrefilter {\n\t\tcDefinesMap[\"ENABLE_PREFILTER\"] = \"1\"\n\t}\n\n\tif option.Config.EnableEndpointRoutes {\n\t\tcDefinesMap[\"ENABLE_ENDPOINT_ROUTES\"] = \"1\"\n\t}\n\n\tif option.Config.EnableEnvoyConfig {\n\t\tcDefinesMap[\"ENABLE_L7_LB\"] = \"1\"\n\t}\n\n\tif option.Config.EnableSocketLB {\n\t\tif option.Config.BPFSocketLBHostnsOnly {\n\t\t\tcDefinesMap[\"ENABLE_SOCKET_LB_HOST_ONLY\"] = \"1\"\n\t\t} else {\n\t\t\tcDefinesMap[\"ENABLE_SOCKET_LB_FULL\"] = \"1\"\n\t\t}\n\t\tif option.Config.EnableSocketLBPeer {\n\t\t\tcDefinesMap[\"ENABLE_SOCKET_LB_PEER\"] = \"1\"\n\t\t}\n\t\tif option.Config.EnableSocketLBTracing {\n\t\t\tcDefinesMap[\"TRACE_SOCK_NOTIFY\"] = \"1\"\n\t\t}\n\n\t\tif cookie, err := netns.GetNetNSCookie(); err == nil {\n\t\t\t// When running in nested environments (e.g. Kind), cilium-agent does\n\t\t\t// not run in the host netns. So, in such cases the cookie comparison\n\t\t\t// based on bpf_get_netns_cookie(NULL) for checking whether a socket\n\t\t\t// belongs to a host netns does not work.\n\t\t\t//\n\t\t\t// To fix this, we derive the cookie of the netns in which cilium-agent\n\t\t\t// runs via getsockopt(...SO_NETNS_COOKIE...) and then use it in the\n\t\t\t// check above. This is based on an assumption that cilium-agent\n\t\t\t// always runs with \"hostNetwork: true\".\n\t\t\tcDefinesMap[\"HOST_NETNS_COOKIE\"] = fmt.Sprintf(\"%d\", cookie)\n\t\t}\n\t}\n\n\tcDefinesMap[\"NAT_46X64_PREFIX_0\"] = \"0\"\n\tcDefinesMap[\"NAT_46X64_PREFIX_1\"] = \"0\"\n\tcDefinesMap[\"NAT_46X64_PREFIX_2\"] = \"0\"\n\tcDefinesMap[\"NAT_46X64_PREFIX_3\"] = \"0\"\n\n\tif option.Config.EnableNodePort {\n\t\tif option.Config.EnableHealthDatapath {\n\t\t\tcDefinesMap[\"ENABLE_HEALTH_CHECK\"] = \"1\"\n\t\t}\n\t\tif option.Config.EnableMKE && option.Config.EnableSocketLB {\n\t\t\tcDefinesMap[\"ENABLE_MKE\"] = \"1\"\n\t\t\tcDefinesMap[\"MKE_HOST\"] = fmt.Sprintf(\"%d\", option.HostExtensionMKE)\n\t\t}\n\t\tif option.Config.EnableRecorder {\n\t\t\tcDefinesMap[\"ENABLE_CAPTURE\"] = \"1\"\n\t\t\tif option.Config.EnableIPv4 {\n\t\t\t\tcDefinesMap[\"CAPTURE4_RULES\"] = recorder.MapNameWcard4\n\t\t\t\tcDefinesMap[\"CAPTURE4_SIZE\"] = fmt.Sprintf(\"%d\", recorder.MapSize)\n\t\t\t}\n\t\t\tif option.Config.EnableIPv6 {\n\t\t\t\tcDefinesMap[\"CAPTURE6_RULES\"] = recorder.MapNameWcard6\n\t\t\t\tcDefinesMap[\"CAPTURE6_SIZE\"] = fmt.Sprintf(\"%d\", recorder.MapSize)\n\t\t\t}\n\t\t}\n\t\tcDefinesMap[\"ENABLE_NODEPORT\"] = \"1\"\n\t\tif option.Config.EnableIPv4 {\n\t\t\tcDefinesMap[\"NODEPORT_NEIGH4\"] = neighborsmap.Map4Name\n\t\t\tcDefinesMap[\"NODEPORT_NEIGH4_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.NeighMapEntriesGlobal)\n\t\t\tif option.Config.EnableHealthDatapath {\n\t\t\t\tcDefinesMap[\"LB4_HEALTH_MAP\"] = lbmap.HealthProbe4MapName\n\t\t\t}\n\t\t}\n\t\tif option.Config.EnableIPv6 {\n\t\t\tcDefinesMap[\"NODEPORT_NEIGH6\"] = neighborsmap.Map6Name\n\t\t\tcDefinesMap[\"NODEPORT_NEIGH6_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.NeighMapEntriesGlobal)\n\t\t\tif option.Config.EnableHealthDatapath {\n\t\t\t\tcDefinesMap[\"LB6_HEALTH_MAP\"] = lbmap.HealthProbe6MapName\n\t\t\t}\n\t\t}\n\t\tif option.Config.EnableNat46X64Gateway {\n\t\t\tcDefinesMap[\"ENABLE_NAT_46X64_GATEWAY\"] = \"1\"\n\t\t\tbase := option.Config.IPv6NAT46x64CIDRBase.AsSlice()\n\t\t\tcDefinesMap[\"NAT_46X64_PREFIX_0\"] = fmt.Sprintf(\"%d\", base[0])\n\t\t\tcDefinesMap[\"NAT_46X64_PREFIX_1\"] = fmt.Sprintf(\"%d\", base[1])\n\t\t\tcDefinesMap[\"NAT_46X64_PREFIX_2\"] = fmt.Sprintf(\"%d\", base[2])\n\t\t\tcDefinesMap[\"NAT_46X64_PREFIX_3\"] = fmt.Sprintf(\"%d\", base[3])\n\t\t}\n\t\tif option.Config.NodePortNat46X64 {\n\t\t\tcDefinesMap[\"ENABLE_NAT_46X64\"] = \"1\"\n\t\t}\n\t\tconst (\n\t\t\tdsrEncapInv = iota\n\t\t\tdsrEncapNone\n\t\t\tdsrEncapIPIP\n\t\t\tdsrEncapGeneve\n\t\t)\n\t\tconst (\n\t\t\tdsrL4XlateInv = iota\n\t\t\tdsrL4XlateFrontend\n\t\t\tdsrL4XlateBackend\n\t\t)\n\t\tcDefinesMap[\"DSR_ENCAP_IPIP\"] = fmt.Sprintf(\"%d\", dsrEncapIPIP)\n\t\tcDefinesMap[\"DSR_ENCAP_GENEVE\"] = fmt.Sprintf(\"%d\", dsrEncapGeneve)\n\t\tcDefinesMap[\"DSR_ENCAP_NONE\"] = fmt.Sprintf(\"%d\", dsrEncapNone)\n\t\tcDefinesMap[\"DSR_XLATE_FRONTEND\"] = fmt.Sprintf(\"%d\", dsrL4XlateFrontend)\n\t\tcDefinesMap[\"DSR_XLATE_BACKEND\"] = fmt.Sprintf(\"%d\", dsrL4XlateBackend)\n\t\tif option.Config.NodePortMode == option.NodePortModeDSR ||\n\t\t\toption.Config.NodePortMode == option.NodePortModeHybrid {\n\t\t\tcDefinesMap[\"ENABLE_DSR\"] = \"1\"\n\t\t\tif option.Config.EnablePMTUDiscovery {\n\t\t\t\tcDefinesMap[\"ENABLE_DSR_ICMP_ERRORS\"] = \"1\"\n\t\t\t}\n\t\t\tif option.Config.NodePortMode == option.NodePortModeHybrid {\n\t\t\t\tcDefinesMap[\"ENABLE_DSR_HYBRID\"] = \"1\"\n\t\t\t}\n\t\t\tif option.Config.LoadBalancerDSRDispatch == option.DSRDispatchOption {\n\t\t\t\tcDefinesMap[\"DSR_ENCAP_MODE\"] = fmt.Sprintf(\"%d\", dsrEncapNone)\n\t\t\t} else if option.Config.LoadBalancerDSRDispatch == option.DSRDispatchIPIP {\n\t\t\t\tcDefinesMap[\"DSR_ENCAP_MODE\"] = fmt.Sprintf(\"%d\", dsrEncapIPIP)\n\t\t\t} else if option.Config.LoadBalancerDSRDispatch == option.DSRDispatchGeneve {\n\t\t\t\tcDefinesMap[\"DSR_ENCAP_MODE\"] = fmt.Sprintf(\"%d\", dsrEncapGeneve)\n\t\t\t}\n\t\t\tif option.Config.LoadBalancerDSRDispatch == option.DSRDispatchIPIP {\n\t\t\t\tif option.Config.LoadBalancerDSRL4Xlate == option.DSRL4XlateFrontend {\n\t\t\t\t\tcDefinesMap[\"DSR_XLATE_MODE\"] = fmt.Sprintf(\"%d\", dsrL4XlateFrontend)\n\t\t\t\t} else if option.Config.LoadBalancerDSRL4Xlate == option.DSRL4XlateBackend {\n\t\t\t\t\tcDefinesMap[\"DSR_XLATE_MODE\"] = fmt.Sprintf(\"%d\", dsrL4XlateBackend)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tcDefinesMap[\"DSR_XLATE_MODE\"] = fmt.Sprintf(\"%d\", dsrL4XlateInv)\n\t\t\t}\n\t\t} else {\n\t\t\tcDefinesMap[\"DSR_ENCAP_MODE\"] = fmt.Sprintf(\"%d\", dsrEncapInv)\n\t\t\tcDefinesMap[\"DSR_XLATE_MODE\"] = fmt.Sprintf(\"%d\", dsrL4XlateInv)\n\t\t}\n\t\tif option.Config.EnableIPv4 {\n\t\t\tif option.Config.LoadBalancerRSSv4CIDR != \"\" {\n\t\t\t\tipv4 := byteorder.NetIPv4ToHost32(option.Config.LoadBalancerRSSv4.IP)\n\t\t\t\tones, _ := option.Config.LoadBalancerRSSv4.Mask.Size()\n\t\t\t\tcDefinesMap[\"IPV4_RSS_PREFIX\"] = fmt.Sprintf(\"%d\", ipv4)\n\t\t\t\tcDefinesMap[\"IPV4_RSS_PREFIX_BITS\"] = fmt.Sprintf(\"%d\", ones)\n\t\t\t} else {\n\t\t\t\tcDefinesMap[\"IPV4_RSS_PREFIX\"] = \"IPV4_DIRECT_ROUTING\"\n\t\t\t\tcDefinesMap[\"IPV4_RSS_PREFIX_BITS\"] = \"32\"\n\t\t\t}\n\t\t}\n\t\tif option.Config.EnableIPv6 {\n\t\t\tif option.Config.LoadBalancerRSSv6CIDR != \"\" {\n\t\t\t\tipv6 := option.Config.LoadBalancerRSSv6.IP\n\t\t\t\tones, _ := option.Config.LoadBalancerRSSv6.Mask.Size()\n\t\t\t\textraMacrosMap[\"IPV6_RSS_PREFIX\"] = ipv6.String()\n\t\t\t\tfw.WriteString(FmtDefineAddress(\"IPV6_RSS_PREFIX\", ipv6))\n\t\t\t\tcDefinesMap[\"IPV6_RSS_PREFIX_BITS\"] = fmt.Sprintf(\"%d\", ones)\n\t\t\t} else {\n\t\t\t\tcDefinesMap[\"IPV6_RSS_PREFIX\"] = \"IPV6_DIRECT_ROUTING\"\n\t\t\t\tcDefinesMap[\"IPV6_RSS_PREFIX_BITS\"] = \"128\"\n\t\t\t}\n\t\t}\n\t\tif option.Config.NodePortAcceleration != option.NodePortAccelerationDisabled {\n\t\t\tcDefinesMap[\"ENABLE_NODEPORT_ACCELERATION\"] = \"1\"\n\t\t}\n\t\tif !option.Config.EnableHostLegacyRouting {\n\t\t\tcDefinesMap[\"ENABLE_HOST_ROUTING\"] = \"1\"\n\t\t}\n\t\tif option.Config.EnableSVCSourceRangeCheck {\n\t\t\tcDefinesMap[\"ENABLE_SRC_RANGE_CHECK\"] = \"1\"\n\t\t\tif option.Config.EnableIPv4 {\n\t\t\t\tcDefinesMap[\"LB4_SRC_RANGE_MAP\"] = lbmap.SourceRange4MapName\n\t\t\t\tcDefinesMap[\"LB4_SRC_RANGE_MAP_SIZE\"] =\n\t\t\t\t\tfmt.Sprintf(\"%d\", lbmap.SourceRange4Map.MaxEntries())\n\t\t\t}\n\t\t\tif option.Config.EnableIPv6 {\n\t\t\t\tcDefinesMap[\"LB6_SRC_RANGE_MAP\"] = lbmap.SourceRange6MapName\n\t\t\t\tcDefinesMap[\"LB6_SRC_RANGE_MAP_SIZE\"] =\n\t\t\t\t\tfmt.Sprintf(\"%d\", lbmap.SourceRange6Map.MaxEntries())\n\t\t\t}\n\t\t}\n\n\t\tcDefinesMap[\"NODEPORT_PORT_MIN\"] = fmt.Sprintf(\"%d\", option.Config.NodePortMin)\n\t\tcDefinesMap[\"NODEPORT_PORT_MAX\"] = fmt.Sprintf(\"%d\", option.Config.NodePortMax)\n\t\tcDefinesMap[\"NODEPORT_PORT_MIN_NAT\"] = fmt.Sprintf(\"%d\", option.Config.NodePortMax+1)\n\t\tcDefinesMap[\"NODEPORT_PORT_MAX_NAT\"] = \"65535\"\n\t}\n\n\tmacByIfIndexMacro, isL3DevMacro, err := devMacros()\n\tif err != nil {\n\t\treturn err\n\t}\n\tcDefinesMap[\"NATIVE_DEV_MAC_BY_IFINDEX(IFINDEX)\"] = macByIfIndexMacro\n\tcDefinesMap[\"IS_L3_DEV(ifindex)\"] = isL3DevMacro\n\n\tconst (\n\t\tselectionRandom = iota + 1\n\t\tselectionMaglev\n\t)\n\tcDefinesMap[\"LB_SELECTION_RANDOM\"] = fmt.Sprintf(\"%d\", selectionRandom)\n\tcDefinesMap[\"LB_SELECTION_MAGLEV\"] = fmt.Sprintf(\"%d\", selectionMaglev)\n\tif option.Config.NodePortAlg == option.NodePortAlgRandom {\n\t\tcDefinesMap[\"LB_SELECTION\"] = fmt.Sprintf(\"%d\", selectionRandom)\n\t} else if option.Config.NodePortAlg == option.NodePortAlgMaglev {\n\t\tcDefinesMap[\"LB_SELECTION\"] = fmt.Sprintf(\"%d\", selectionMaglev)\n\t\tcDefinesMap[\"LB_MAGLEV_LUT_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.MaglevTableSize)\n\t\tif option.Config.EnableIPv6 {\n\t\t\tcDefinesMap[\"LB6_MAGLEV_MAP_OUTER\"] = lbmap.MaglevOuter6MapName\n\t\t}\n\t\tif option.Config.EnableIPv4 {\n\t\t\tcDefinesMap[\"LB4_MAGLEV_MAP_OUTER\"] = lbmap.MaglevOuter4MapName\n\t\t}\n\t}\n\tcDefinesMap[\"HASH_INIT4_SEED\"] = fmt.Sprintf(\"%d\", maglev.SeedJhash0)\n\tcDefinesMap[\"HASH_INIT6_SEED\"] = fmt.Sprintf(\"%d\", maglev.SeedJhash1)\n\n\tif option.Config.DirectRoutingDeviceRequired() {\n\t\tdirectRoutingIface := option.Config.DirectRoutingDevice\n\t\tdirectRoutingIfIndex, err := link.GetIfIndex(directRoutingIface)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcDefinesMap[\"DIRECT_ROUTING_DEV_IFINDEX\"] = fmt.Sprintf(\"%d\", directRoutingIfIndex)\n\t\tif option.Config.EnableIPv4 {\n\t\t\tip, ok := node.GetNodePortIPv4AddrsWithDevices()[directRoutingIface]\n\t\t\tif !ok {\n\t\t\t\tlog.WithFields(logrus.Fields{\n\t\t\t\t\t\"directRoutingIface\": directRoutingIface,\n\t\t\t\t}).Fatal(\"Direct routing device's IPv4 address not found\")\n\t\t\t}\n\n\t\t\tipv4 := byteorder.NetIPv4ToHost32(ip)\n\t\t\tcDefinesMap[\"IPV4_DIRECT_ROUTING\"] = fmt.Sprintf(\"%d\", ipv4)\n\t\t}\n\n\t\tif option.Config.EnableIPv6 {\n\t\t\tdirectRoutingIPv6, ok := node.GetNodePortIPv6AddrsWithDevices()[directRoutingIface]\n\t\t\tif !ok {\n\t\t\t\tlog.WithFields(logrus.Fields{\n\t\t\t\t\t\"directRoutingIface\": directRoutingIface,\n\t\t\t\t}).Fatal(\"Direct routing device's IPv6 address not found\")\n\t\t\t}\n\n\t\t\textraMacrosMap[\"IPV6_DIRECT_ROUTING\"] = directRoutingIPv6.String()\n\t\t\tfw.WriteString(FmtDefineAddress(\"IPV6_DIRECT_ROUTING\", directRoutingIPv6))\n\t\t}\n\t} else {\n\t\tvar directRoutingIPv6 net.IP\n\t\tcDefinesMap[\"DIRECT_ROUTING_DEV_IFINDEX\"] = \"0\"\n\t\tif option.Config.EnableIPv4 {\n\t\t\tcDefinesMap[\"IPV4_DIRECT_ROUTING\"] = \"0\"\n\t\t}\n\t\tif option.Config.EnableIPv6 {\n\t\t\textraMacrosMap[\"IPV6_DIRECT_ROUTING\"] = directRoutingIPv6.String()\n\t\t\tfw.WriteString(FmtDefineAddress(\"IPV6_DIRECT_ROUTING\", directRoutingIPv6))\n\t\t}\n\t}\n\n\tif option.Config.ResetQueueMapping {\n\t\tcDefinesMap[\"RESET_QUEUES\"] = \"1\"\n\t}\n\n\tif option.Config.EnableBandwidthManager {\n\t\tcDefinesMap[\"ENABLE_BANDWIDTH_MANAGER\"] = \"1\"\n\t\tcDefinesMap[\"THROTTLE_MAP\"] = bwmap.MapName\n\t\tcDefinesMap[\"THROTTLE_MAP_SIZE\"] = fmt.Sprintf(\"%d\", bwmap.MapSize)\n\t}\n\n\tif option.Config.EnableHostFirewall {\n\t\tcDefinesMap[\"ENABLE_HOST_FIREWALL\"] = \"1\"\n\t}\n\n\tif option.Config.EnableIPSec {\n\t\tnodeAddress := node.GetIPv4()\n\t\tif nodeAddress == nil {\n\t\t\treturn errors.New(\"external IPv4 node address is required when IPSec is enabled, but none found\")\n\t\t}\n\n\t\ta := byteorder.NetIPv4ToHost32(nodeAddress)\n\t\tcDefinesMap[\"IPV4_ENCRYPT_IFACE\"] = fmt.Sprintf(\"%d\", a)\n\t\tif iface := option.Config.EncryptInterface; len(iface) != 0 {\n\t\t\tlink, err := netlink.LinkByName(iface[0])\n\t\t\tif err == nil {\n\t\t\t\tcDefinesMap[\"ENCRYPT_IFACE\"] = fmt.Sprintf(\"%d\", link.Attrs().Index)\n\t\t\t}\n\t\t}\n\t}\n\n\tif option.Config.EnableNodePort {\n\t\tif option.Config.EnableIPv4 {\n\t\t\tcDefinesMap[\"SNAT_MAPPING_IPV4\"] = nat.MapNameSnat4Global\n\t\t\tcDefinesMap[\"SNAT_MAPPING_IPV4_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.NATMapEntriesGlobal)\n\t\t}\n\n\t\tif option.Config.EnableIPv6 {\n\t\t\tcDefinesMap[\"SNAT_MAPPING_IPV6\"] = nat.MapNameSnat6Global\n\t\t\tcDefinesMap[\"SNAT_MAPPING_IPV6_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.NATMapEntriesGlobal)\n\t\t}\n\n\t\tif option.Config.EnableBPFMasquerade {\n\t\t\tif option.Config.EnableIPv4Masquerade {\n\t\t\t\tcDefinesMap[\"ENABLE_MASQUERADE_IPV4\"] = \"1\"\n\n\t\t\t\t// ip-masq-agent depends on bpf-masq\n\t\t\t\tvar excludeCIDR *cidr.CIDR\n\t\t\t\tif option.Config.EnableIPMasqAgent {\n\t\t\t\t\tcDefinesMap[\"ENABLE_IP_MASQ_AGENT_IPV4\"] = \"1\"\n\t\t\t\t\tcDefinesMap[\"IP_MASQ_AGENT_IPV4\"] = ipmasq.MapNameIPv4\n\n\t\t\t\t\t// native-routing-cidr is optional with ip-masq-agent and may be nil\n\t\t\t\t\texcludeCIDR = option.Config.GetIPv4NativeRoutingCIDR()\n\t\t\t\t} else {\n\t\t\t\t\texcludeCIDR = datapath.RemoteSNATDstAddrExclusionCIDRv4()\n\t\t\t\t}\n\n\t\t\t\tif excludeCIDR != nil {\n\t\t\t\t\tcDefinesMap[\"IPV4_SNAT_EXCLUSION_DST_CIDR\"] =\n\t\t\t\t\t\tfmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(excludeCIDR.IP))\n\t\t\t\t\tones, _ := excludeCIDR.Mask.Size()\n\t\t\t\t\tcDefinesMap[\"IPV4_SNAT_EXCLUSION_DST_CIDR_LEN\"] = fmt.Sprintf(\"%d\", ones)\n\t\t\t\t}\n\t\t\t}\n\t\t\tif option.Config.EnableIPv6Masquerade {\n\t\t\t\tcDefinesMap[\"ENABLE_MASQUERADE_IPV6\"] = \"1\"\n\n\t\t\t\tvar excludeCIDR *cidr.CIDR\n\t\t\t\tif option.Config.EnableIPMasqAgent {\n\t\t\t\t\tcDefinesMap[\"ENABLE_IP_MASQ_AGENT_IPV6\"] = \"1\"\n\t\t\t\t\tcDefinesMap[\"IP_MASQ_AGENT_IPV6\"] = ipmasq.MapNameIPv6\n\n\t\t\t\t\texcludeCIDR = option.Config.GetIPv6NativeRoutingCIDR()\n\t\t\t\t} else {\n\t\t\t\t\texcludeCIDR = datapath.RemoteSNATDstAddrExclusionCIDRv6()\n\t\t\t\t}\n\n\t\t\t\tif excludeCIDR != nil {\n\t\t\t\t\textraMacrosMap[\"IPV6_SNAT_EXCLUSION_DST_CIDR\"] = excludeCIDR.IP.String()\n\t\t\t\t\tfw.WriteString(FmtDefineAddress(\"IPV6_SNAT_EXCLUSION_DST_CIDR\", excludeCIDR.IP))\n\t\t\t\t\textraMacrosMap[\"IPV6_SNAT_EXCLUSION_DST_CIDR_MASK\"] = excludeCIDR.Mask.String()\n\t\t\t\t\tfw.WriteString(FmtDefineAddress(\"IPV6_SNAT_EXCLUSION_DST_CIDR_MASK\", excludeCIDR.Mask))\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\tctmap.WriteBPFMacros(fw, nil)\n\t}\n\n\tif option.Config.AllowICMPFragNeeded {\n\t\tcDefinesMap[\"ALLOW_ICMP_FRAG_NEEDED\"] = \"1\"\n\t}\n\n\tif option.Config.ClockSource == option.ClockSourceJiffies {\n\t\tcDefinesMap[\"ENABLE_JIFFIES\"] = \"1\"\n\t}\n\n\tif option.Config.EnableIdentityMark {\n\t\tcDefinesMap[\"ENABLE_IDENTITY_MARK\"] = \"1\"\n\t}\n\n\tif option.Config.EnableHighScaleIPcache {\n\t\tcDefinesMap[\"ENABLE_HIGH_SCALE_IPCACHE\"] = \"1\"\n\t}\n\n\tif option.Config.EnableCustomCalls {\n\t\tcDefinesMap[\"ENABLE_CUSTOM_CALLS\"] = \"1\"\n\t}\n\n\tif option.Config.EnableVTEP {\n\t\tcDefinesMap[\"ENABLE_VTEP\"] = \"1\"\n\t\tcDefinesMap[\"VTEP_MAP\"] = vtep.Name\n\t\tcDefinesMap[\"VTEP_MAP_SIZE\"] = fmt.Sprintf(\"%d\", vtep.MaxEntries)\n\t\tcDefinesMap[\"VTEP_MASK\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(net.IP(option.Config.VtepCidrMask)))\n\n\t}\n\n\tvlanFilter, err := vlanFilterMacros()\n\tif err != nil {\n\t\treturn err\n\t}\n\tcDefinesMap[\"VLAN_FILTER(ifindex, vlan_id)\"] = vlanFilter\n\n\tif option.Config.EnableICMPRules {\n\t\tcDefinesMap[\"ENABLE_ICMP_RULE\"] = \"1\"\n\t}\n\n\tcDefinesMap[\"CIDR_IDENTITY_RANGE_START\"] = fmt.Sprintf(\"%d\", identity.MinLocalIdentity)\n\tcDefinesMap[\"CIDR_IDENTITY_RANGE_END\"] = fmt.Sprintf(\"%d\", identity.MaxLocalIdentity)\n\n\tif option.Config.TunnelingEnabled() {\n\t\tcDefinesMap[\"TUNNEL_MODE\"] = \"1\"\n\t}\n\n\tciliumNetLink, err := netlink.LinkByName(defaults.SecondHostDevice)\n\tif err != nil {\n\t\treturn err\n\t}\n\tcDefinesMap[\"CILIUM_NET_MAC\"] = fmt.Sprintf(\"{.addr=%s}\", mac.CArrayString(ciliumNetLink.Attrs().HardwareAddr))\n\tcDefinesMap[\"HOST_IFINDEX\"] = fmt.Sprintf(\"%d\", ciliumNetLink.Attrs().Index)\n\n\tciliumHostLink, err := netlink.LinkByName(defaults.HostDevice)\n\tif err != nil {\n\t\treturn err\n\t}\n\tcDefinesMap[\"HOST_IFINDEX_MAC\"] = fmt.Sprintf(\"{.addr=%s}\", mac.CArrayString(ciliumHostLink.Attrs().HardwareAddr))\n\tcDefinesMap[\"CILIUM_IFINDEX\"] = fmt.Sprintf(\"%d\", ciliumHostLink.Attrs().Index)\n\n\tephemeralMin, err := getEphemeralPortRangeMin()\n\tif err != nil {\n\t\treturn err\n\t}\n\tcDefinesMap[\"EPHEMERAL_MIN\"] = fmt.Sprintf(\"%d\", ephemeralMin)\n\n\tif err := cDefinesMap.Merge(h.nodeExtraDefines); err != nil {\n\t\treturn err\n\t}\n\n\tfor _, fn := range h.nodeExtraDefineFns {\n\t\tdefines, err := fn()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := cDefinesMap.Merge(defines); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif option.Config.EnableHealthDatapath {\n\t\tif option.Config.IPv4Enabled() {\n\t\t\tipip4, err := netlink.LinkByName(defaults.IPIPv4Device)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tcDefinesMap[\"ENCAP4_IFINDEX\"] = fmt.Sprintf(\"%d\", ipip4.Attrs().Index)\n\t\t}\n\t\tif option.Config.IPv6Enabled() {\n\t\t\tipip6, err := netlink.LinkByName(defaults.IPIPv6Device)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tcDefinesMap[\"ENCAP6_IFINDEX\"] = fmt.Sprintf(\"%d\", ipip6.Attrs().Index)\n\t\t}\n\t}\n\n\t// Since golang maps are unordered, we sort the keys in the map\n\t// to get a consistent written format to the writer. This maintains\n\t// the consistency when we try to calculate hash for a datapath after\n\t// writing the config.\n\tkeys := make([]string, 0, len(cDefinesMap))\n\tfor key := range cDefinesMap {\n\t\tkeys = append(keys, key)\n\t}\n\tsort.Strings(keys)\n\n\tfor _, key := range keys {\n\t\tfmt.Fprintf(fw, \"#define %s %s\\n\", key, cDefinesMap[key])\n\t}\n\n\t// Populate cDefinesMap with extraMacrosMap to get all the configuration\n\t// in the cDefinesMap itself.\n\tfor key, value := range extraMacrosMap {\n\t\tcDefinesMap[key] = value\n\t}\n\n\t// Write the JSON encoded config as base64 encoded commented string to\n\t// the header file.\n\tjsonBytes, err := json.Marshal(cDefinesMap)\n\tif err == nil {\n\t\t// We don't care if some error occurs while marshaling the map.\n\t\t// In such cases we skip embedding the base64 encoded JSON configuration\n\t\t// to the writer.\n\t\tencodedConfig := base64.StdEncoding.EncodeToString(jsonBytes)\n\t\tfmt.Fprintf(fw, \"\\n// JSON_OUTPUT: %s\\n\", encodedConfig)\n\t}\n\n\treturn fw.Flush()\n}", "func parseDevices(r io.Reader) (devices []storage.Device, err error) {\n\t// columnName is the output column for device name\n\tconst columnName = \"NAME\"\n\t// columnType is the output column for device type\n\tconst columnType = \"TYPE\"\n\t// columnSize is the output column for device size\n\tconst columnSize = \"SIZE\"\n\t// columnFilesystemType is the output column for mounted filesystem type\n\tconst columnFilesystemType = \"FSTYPE\"\n\t// columnParentName is the output column for internal kernel parent device name\n\tconst columnParentName = \"PKNAME\"\n\n\ts := bufio.NewScanner(r)\n\ts.Split(bufio.ScanLines)\n\tp := &parser{}\n\tdeviceCache := make(map[string]storage.Device)\n\tfor s.Scan() {\n\t\tcolumns := map[string]string{}\n\t\tline := strings.TrimSpace(s.Text())\n\t\tif len(line) == 0 {\n\t\t\tcontinue\n\t\t}\n\t\tp.scanner.Init(strings.NewReader(line))\n\t\tp.next()\n\t\tfor p.token != scanner.EOF {\n\t\t\tattr := p.parseAttribute()\n\t\t\tif len(attr.value) > 0 {\n\t\t\t\tcolumns[attr.name] = attr.value\n\t\t\t}\n\t\t}\n\t\tif len(p.errors) > 0 {\n\t\t\tp.errors = append(p.errors, trace.Errorf(\"failed to parse %s\", line))\n\t\t\treturn nil, trace.NewAggregate(p.errors...)\n\t\t}\n\t\tdeviceType := storage.DeviceType(columns[columnType])\n\t\tif deviceType == storage.DeviceDisk || deviceType == storage.DevicePartition {\n\t\t\tdevicePath := filepath.Join(\"/dev\", columns[columnName])\n\t\t\tparentName, hasParent := columns[columnParentName]\n\n\t\t\tif deviceType == storage.DevicePartition && hasParent {\n\t\t\t\tdelete(deviceCache, filepath.Join(\"/dev\", parentName))\n\t\t\t}\n\n\t\t\tif _, hasFilesystem := columns[columnFilesystemType]; hasFilesystem {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tsize, err := strconv.ParseUint(columns[columnSize], 10, 64)\n\t\t\tif err != nil {\n\t\t\t\tlog.Infof(\"invalid size %v for device %v\", columns[columnSize], devicePath)\n\t\t\t} else {\n\t\t\t\tsize = size >> 20 // Mbytes\n\t\t\t}\n\n\t\t\tdeviceCache[devicePath] = storage.Device{Name: storage.DeviceName(devicePath), Type: deviceType, SizeMB: size}\n\t\t}\n\t}\n\tfor _, device := range deviceCache {\n\t\tdevices = append(devices, device)\n\t}\n\treturn devices, nil\n}", "func GenerateNetlinxCompileCfg(w http.ResponseWriter, r *http.Request) {\n\t// Get Body as Bytes Array\n\tbody, err := ioutil.ReadAll(r.Body)\n\tif err != nil {\n\t\tw.WriteHeader(500)\n\t\treturn\n\t}\n\n\t// Load this into an APW Workspace\n\ta, err := apw.NewAPW(\"myWorkspace.apw\", body)\n\n\t// Get URL Variables\n\troot := r.URL.Query().Get(\"root\")\n\tlogfile := r.URL.Query().Get(\"logfile\")\n\tlogconsole, _ := strconv.ParseBool(r.URL.Query().Get(\"logconsole\"))\n\t// Process and generate the .cfg\n\tb := compilecfg.Generate(*a, root, logfile, logconsole)\n\n\tw.Write(b)\n}", "func mkbuildcfg(file string) {\n\tvar buf bytes.Buffer\n\tfmt.Fprintf(&buf, \"// Code generated by go tool dist; DO NOT EDIT.\\n\")\n\tfmt.Fprintln(&buf)\n\tfmt.Fprintf(&buf, \"package buildcfg\\n\")\n\tfmt.Fprintln(&buf)\n\tfmt.Fprintf(&buf, \"import \\\"runtime\\\"\\n\")\n\tfmt.Fprintln(&buf)\n\tfmt.Fprintf(&buf, \"const defaultGO386 = `%s`\\n\", go386)\n\tfmt.Fprintf(&buf, \"const defaultGOAMD64 = `%s`\\n\", goamd64)\n\tfmt.Fprintf(&buf, \"const defaultGOARM = `%s`\\n\", goarm)\n\tfmt.Fprintf(&buf, \"const defaultGOMIPS = `%s`\\n\", gomips)\n\tfmt.Fprintf(&buf, \"const defaultGOMIPS64 = `%s`\\n\", gomips64)\n\tfmt.Fprintf(&buf, \"const defaultGOPPC64 = `%s`\\n\", goppc64)\n\tfmt.Fprintf(&buf, \"const defaultGOEXPERIMENT = `%s`\\n\", goexperiment)\n\tfmt.Fprintf(&buf, \"const defaultGO_EXTLINK_ENABLED = `%s`\\n\", goextlinkenabled)\n\tfmt.Fprintf(&buf, \"const defaultGO_LDSO = `%s`\\n\", defaultldso)\n\tfmt.Fprintf(&buf, \"const version = `%s`\\n\", findgoversion())\n\tfmt.Fprintf(&buf, \"const defaultGOOS = runtime.GOOS\\n\")\n\tfmt.Fprintf(&buf, \"const defaultGOARCH = runtime.GOARCH\\n\")\n\n\twritefile(buf.String(), file, writeSkipSame)\n}", "func getDeviceNameMapping() (map[string]string, error) {\n\tfilesystems, err := util.CollectDfValues()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tret := map[string]string{}\n\tfor _, dfs := range filesystems {\n\t\tname := dfs.Name\n\t\tif device := strings.TrimPrefix(name, \"/dev/\"); name != device {\n\t\t\tret[device] = dfs.Mounted\n\t\t}\n\t}\n\treturn ret, nil\n}", "func AppendDeviceChangeSpec(\n\tspec []types.BaseVirtualDeviceConfigSpec,\n\tops ...types.BaseVirtualDeviceConfigSpec,\n) []types.BaseVirtualDeviceConfigSpec {\n\tfor _, op := range ops {\n\t\tc := copystructure.Must(copystructure.Copy(op)).(types.BaseVirtualDeviceConfigSpec)\n\t\tspec = append(spec, c)\n\t}\n\treturn spec\n}", "func (o KubernetesClusterNodePoolLinuxOsConfigSysctlConfigOutput) NetCoreNetdevMaxBacklog() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v KubernetesClusterNodePoolLinuxOsConfigSysctlConfig) *int { return v.NetCoreNetdevMaxBacklog }).(pulumi.IntPtrOutput)\n}", "func Name(busNumber int) string {\n\tfor _, d := range nvDevices {\n\t\tif d.bus == busNumber {\n\t\t\treturn d.name\n\t\t}\n\t}\n\n\treturn \"\"\n}", "func printk(c byte) {\n\timx6.UART2.Tx(c)\n}", "func subsystems_linux() []*Subsystem {\n\tvar _9p, ac100, accelerators, acpi, acpi4asus, acpica, acrn, actions, afs, alpha, alsa, amdgfx, amlogic, apparmor, arch, arm, armmsm, asahi, aspeed, ath10k, ath11k, ath12k, atm, audit, autofs, axis, b43, batman, bcache, bfs, block, bluetooth, bpf, brcm80211, bridge, btrfs, cachefs, can, ceph, cgroups, chrome, cifs, cirrus, clk, coda, coresight, crypto, csky, cxl, damon, dccp, dell, devicetree, dm, dmaengine, drbd, dri, ecryptfs, edac, efi, erofs, etnaviv, ext4, f2fs, fat, fbdev, fpga, freedreno, fs, fscrypt, fsi, fsverity, fuse, geode, gfs2, gpio, greybus, hams, hardening, hexagon, hfs, hippi, hwmon, hyperv, i2c, i3c, ia64, ide, iio, imx, input, integrity, intelgfx, intelgvt, intelwiredlan, iouring, iommu, ipack, isdn4linux, isofs, jfs, karma, kasan, kernel, kernfs, kexec, keyrings, kgdb, kunit, kvm, kvmriscv, kvmarm, leds, libertas, lima, linux1394, linuxppc, linuxpps, livepatching, llvm, loongarch, lsm, lvs, m68k, malidp, media, mediatek, megaraid, mhi, mips, mjpeg, mm, mmc, modules, mpi3, mptfusion, mptcp, mtd, nbd, net, netfilter, nfc, nfs, nilfs, nitro, nouveau, ntb, ntfs, ntfs3, nvdimm, nvme, ocfs2, omap, optee, openiscsi, openbmc, openipmi, openrisc, openvswitch, openwrt, orangefs, ossdrivers, overlayfs, oxnas, parisc, parport, pci, perf, phy, pm, ppp, pvrusb2, pwm, qat, raid, rcu, rdma, rds, reiserfs, remoteproc, renesassoc, riscv, rockchip, rpi, rttools, rtc, rust, s390, samsungsoc, scsi, sctp, selinux, serial, sgx, sh, snpsarc, sof, sparclinux, speakup, spi, spice, squashfs, staging, stm32, sunxi, target, tegra, tipc, tomoyo, trace, uclinux, udf, um, unisoc, usb, usbstorage, video, virt, watchdog, wcn36xx, wireguard, wireless, wpan, x25, x86, x86drivers, xen, xfs, xtensa, zd1211 Subsystem\n\n\t_9p = Subsystem{\n\t\tName: \"9p\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/9p/|^include/net/9p/|^include/trace/events/9p\\\\.h$|^include/uapi/linux/virtio_9p\\\\.h$|^net/9p/\"},\n\t\t},\n\t}\n\n\tac100 = Subsystem{\n\t\tName: \"ac100\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&staging, &tegra},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/staging/nvec/\"},\n\t\t},\n\t}\n\n\taccelerators = Subsystem{\n\t\tName: \"accelerators\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/misc/uacce/|^include/linux/uacce\\\\.h$|^include/uapi/misc/uacce/\"},\n\t\t},\n\t}\n\n\tacpi = Subsystem{\n\t\tName: \"acpi\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/x86/kernel/acpi/|^arch/x86/pci/acpi\\\\.c$|^drivers/acpi/|^drivers/pci/[^/]*/[^/]*acpi[^/]*$|^drivers/pci/[^/]*acpi[^/]*$|^drivers/pnp/pnpacpi/|^include/acpi/|^include/linux/acpi\\\\.h$|^include/linux/fwnode\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/acpi/[^/]*thermal[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/acpi/acpica/|^include/acpi/\"},\n\t\t\t{IncludeRegexp: \"^drivers/acpi/apei/\"},\n\t\t\t{IncludeRegexp: \"^drivers/acpi/arm64$\"},\n\t\t\t{IncludeRegexp: \"^drivers/acpi/pmic/\"},\n\t\t\t{IncludeRegexp: \"^drivers/acpi/viot\\\\.c$|^include/linux/acpi_viot\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/base/property\\\\.c$|^drivers/base/swnode\\\\.c$|^include/linux/fwnode\\\\.h$|^include/linux/property\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpiolib-acpi\\\\.c$|^drivers/gpio/gpiolib-acpi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/i2c-core-acpi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mailbox/pcc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pnp/|^include/linux/pnp\\\\.h$\"},\n\t\t},\n\t}\n\n\tacpi4asus = Subsystem{\n\t\tName: \"acpi4asus\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&x86drivers},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/asus[^/]*\\\\.c$|^drivers/platform/x86/eeepc[^/]*\\\\.c$\"},\n\t\t},\n\t}\n\n\tacpica = Subsystem{\n\t\tName: \"acpica\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&acpi},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/acpi/acpica/|^include/acpi/\"},\n\t\t},\n\t}\n\n\tacrn = Subsystem{\n\t\tName: \"acrn\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/virt/acrn/|^include/uapi/linux/acrn\\\\.h$\"},\n\t\t},\n\t}\n\n\tactions = Subsystem{\n\t\tName: \"actions\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&arm},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/owl-[^/]*$|^arch/arm/mach-actions/|^arch/arm64/boot/dts/actions/|^drivers/clk/actions/|^drivers/clocksource/timer-owl[^/]*$|^drivers/dma/owl-dma\\\\.c$|^drivers/i2c/busses/i2c-owl\\\\.c$|^drivers/irqchip/irq-owl-sirq\\\\.c$|^drivers/mmc/host/owl-mmc\\\\.c$|^drivers/net/ethernet/actions/|^drivers/pinctrl/actions/|^drivers/soc/actions/|^include/dt-bindings/power/owl-[^/]*$|^include/dt-bindings/reset/actions,[^/]*$|^include/linux/soc/actions/|owl\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/misc/atc260x-onkey\\\\.c$|^drivers/mfd/atc260[^/]*$|^drivers/power/reset/atc260x-poweroff\\\\.c$|^drivers/regulator/atc260x-regulator\\\\.c$|^include/linux/mfd/atc260x/\"},\n\t\t},\n\t}\n\n\tafs = Subsystem{\n\t\tName: \"afs\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/afs/|^include/trace/events/afs\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^include/keys/rxrpc-type\\\\.h$|^include/net/af_rxrpc\\\\.h$|^include/trace/events/rxrpc\\\\.h$|^include/uapi/linux/rxrpc\\\\.h$|^net/rxrpc/\"},\n\t\t},\n\t}\n\n\talpha = Subsystem{\n\t\tName: \"alpha\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/alpha/\"},\n\t\t},\n\t}\n\n\talsa = Subsystem{\n\t\tName: \"alsa\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-madera[^/]*$|^drivers/irqchip/irq-madera[^/]*$|^drivers/mfd/cs47l[^/]*$|^drivers/mfd/madera[^/]*$|^drivers/pinctrl/cirrus/|^include/dt-bindings/sound/madera[^/]*$|^include/linux/irqchip/irq-madera[^/]*$|^include/linux/mfd/madera/|^include/sound/madera[^/]*$|^sound/soc/codecs/cs47l[^/]*$|^sound/soc/codecs/madera[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/slimbus/|^include/linux/slimbus\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/soc/qcom/apr\\\\.c$|^include/dt-bindings/sound/qcom,wcd9335\\\\.h$|^sound/soc/codecs/lpass-rx-macro\\\\.[^/]*$|^sound/soc/codecs/lpass-tx-macro\\\\.[^/]*$|^sound/soc/codecs/lpass-va-macro\\\\.c$|^sound/soc/codecs/lpass-wsa-macro\\\\.[^/]*$|^sound/soc/codecs/msm8916-wcd-analog\\\\.c$|^sound/soc/codecs/msm8916-wcd-digital\\\\.c$|^sound/soc/codecs/wcd9335\\\\.[^/]*$|^sound/soc/codecs/wcd934x\\\\.c$|^sound/soc/codecs/wcd-clsh-v2\\\\.[^/]*$|^sound/soc/codecs/wcd-mbhc-v2\\\\.[^/]*$|^sound/soc/codecs/wsa881x\\\\.c$|^sound/soc/codecs/wsa883x\\\\.c$|^sound/soc/qcom/\"},\n\t\t\t{IncludeRegexp: \"^drivers/soundwire/|^include/linux/soundwire/\"},\n\t\t\t{IncludeRegexp: \"^include/dt-bindings/sound/cs[^/]*$|^sound/pci/hda/cs[^/]*$|^sound/pci/hda/hda_cs_dsp_ctl\\\\.[^/]*$|^sound/soc/codecs/cs[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^include/dt-bindings/sound/|^include/sound/soc[^/]*$|^sound/soc/\"},\n\t\t\t{IncludeRegexp: \"^include/sound/compress_driver\\\\.h$|^include/uapi/sound/compress_[^/]*$|^sound/core/compress_offload\\\\.c$|^sound/soc/soc-compress\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/sound/|^include/uapi/sound/|^sound/\"},\n\t\t\t{IncludeRegexp: \"^include/uapi/linux/virtio_snd\\\\.h$|^sound/virtio/\"},\n\t\t\t{IncludeRegexp: \"^include/uapi/sound/firewire\\\\.h$|^sound/firewire/\"},\n\t\t\t{IncludeRegexp: \"^sound/aoa/\"},\n\t\t\t{IncludeRegexp: \"^sound/drivers/opl4/\"},\n\t\t\t{IncludeRegexp: \"^sound/pci/bt87x\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^sound/pci/oxygen/\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/apple/|^sound/soc/codecs/cs42l83-i2c\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/atmel$\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/atmel/tse850-pcm5142\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/codecs/ad1[^/]*$|^sound/soc/codecs/ad7[^/]*$|^sound/soc/codecs/adau[^/]*$|^sound/soc/codecs/adav[^/]*$|^sound/soc/codecs/sigmadsp\\\\.[^/]*$|^sound/soc/codecs/ssm[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/codecs/idt821034\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/codecs/isabelle[^/]*$|^sound/soc/codecs/lm49453[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/codecs/max9860\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/codecs/pcm3060[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/codecs/peb2466\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/codecs/sgtl5000[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/codecs/sma[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/codecs/tas571x[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/codecs/tfa9879[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/codecs/tfa989x\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/codecs/twl4030[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/fsl/fsl[^/]*$|^sound/soc/fsl/imx[^/]*$|^sound/soc/fsl/mpc8610_hpcd\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/intel/\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/meson/\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/samsung/\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/sti/\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/stm/\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/sunxi/sun50i-dmic\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/ti/\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/ti/n810\\\\.c$|^sound/soc/ti/omap[^/]*$|^sound/soc/ti/rx51\\\\.c$|^sound/soc/ti/sdma-pcm\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/uniphier/\"},\n\t\t\t{IncludeRegexp: \"^sound/usb/caiaq/\"},\n\t\t\t{IncludeRegexp: \"^sound/usb/midi\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^sound/usb/misc/ua101\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^sound/usb/mixer_scarlett_gen2\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^sound/xen/\"},\n\t\t},\n\t}\n\n\tamdgfx = Subsystem{\n\t\tName: \"amd-gfx\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&dri},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd[^/]*\\\\.\\\\[ch\\\\]$|^drivers/gpu/drm/amd/amdkfd/|^drivers/gpu/drm/amd/include/cik_structs\\\\.h$|^drivers/gpu/drm/amd/include/kgd_kfd_interface\\\\.h$|^drivers/gpu/drm/amd/include/v9_structs\\\\.h$|^drivers/gpu/drm/amd/include/vi_structs\\\\.h$|^include/uapi/linux/kfd_ioctl\\\\.h$|^include/uapi/linux/kfd_sysfs\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/amd/display/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/amd/pm/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/amd/|^drivers/gpu/drm/radeon/|^include/uapi/drm/amdgpu_drm\\\\.h$|^include/uapi/drm/radeon_drm\\\\.h$\"},\n\t\t},\n\t}\n\n\tamlogic = Subsystem{\n\t\tName: \"amlogic\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&arm},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/meson[^/]*$|^arch/arm/mach-meson/|^arch/arm64/boot/dts/amlogic/|^drivers/mmc/host/meson[^/]*$|^drivers/pinctrl/meson/|^drivers/rtc/rtc-meson[^/]*$|^drivers/soc/amlogic/|meson\"},\n\t\t\t{IncludeRegexp: \"^drivers/clk/meson/|^include/dt-bindings/clock/gxbb[^/]*$|^include/dt-bindings/clock/meson[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/amlogic/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/meson/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/cec/platform/meson/ao-cec-g12a\\\\.c$|^drivers/media/cec/platform/meson/ao-cec\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/amlogic/meson-ge2d/\"},\n\t\t\t{IncludeRegexp: \"^drivers/mfd/khadas-mcu\\\\.c$|^include/linux/mfd/khadas-mcu\\\\.h$|^drivers/thermal/khadas_mcu_fan\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/pci-meson\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/perf/amlogic/|^include/soc/amlogic/\"},\n\t\t\t{IncludeRegexp: \"^drivers/staging/media/meson/vdec/\"},\n\t\t\t{IncludeRegexp: \"^drivers/thermal/amlogic_thermal\\\\.c$\"},\n\t\t},\n\t}\n\n\tapparmor = Subsystem{\n\t\tName: \"apparmor\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&lsm},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^security/apparmor/\"},\n\t\t},\n\t}\n\n\tarch = Subsystem{\n\t\tName: \"arch\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/[^/]*/include/asm/tlb\\\\.h$|^include/asm-generic/tlb\\\\.h$|^mm/mmu_gather\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/asm-generic/|^include/uapi/asm-generic/\"},\n\t\t},\n\t}\n\n\tarm = Subsystem{\n\t\tName: \"arm\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^[^/]*/[^/]*/[^/]*/vexpress[^/]*$|^[^/]*/[^/]*/vexpress[^/]*$|^arch/arm/boot/dts/vexpress[^/]*$|^arch/arm/mach-versatile/|^arch/arm64/boot/dts/arm/|^drivers/clk/versatile/clk-vexpress-osc\\\\.c$|^drivers/clocksource/timer-versatile\\\\.c$|mps2\"},\n\t\t\t{\"^arch/arm/\", \"^arch/arm/boot/dts/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/compressed/misc-ep93xx\\\\.h$|^arch/arm/mach-ep93xx/|^drivers/iio/adc/ep93xx_adc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/alpine[^/]*$|^arch/arm/mach-alpine/|^arch/arm64/boot/dts/amazon/|^drivers/[^/]*/[^/]*alpine[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/arm-realview-[^/]*$|^arch/arm/boot/dts/integrator[^/]*$|^arch/arm/boot/dts/versatile[^/]*$|^arch/arm/mach-versatile/|^drivers/bus/arm-integrator-lm\\\\.c$|^drivers/clk/versatile/|^drivers/i2c/busses/i2c-versatile\\\\.c$|^drivers/irqchip/irq-versatile-fpga\\\\.c$|^drivers/mtd/maps/physmap-versatile\\\\.[^/]*$|^drivers/power/reset/arm-versatile-reboot\\\\.c$|^drivers/soc/versatile/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/armada[^/]*$|^arch/arm/boot/dts/kirkwood[^/]*$|^arch/arm/configs/mvebu_[^/]*_defconfig$|^arch/arm/mach-mvebu/|^arch/arm64/boot/dts/marvell/armada[^/]*$|^arch/arm64/boot/dts/marvell/cn913[^/]*$|^drivers/cpufreq/armada-37xx-cpufreq\\\\.c$|^drivers/cpufreq/armada-8k-cpufreq\\\\.c$|^drivers/cpufreq/mvebu-cpufreq\\\\.c$|^drivers/irqchip/irq-armada-370-xp\\\\.c$|^drivers/irqchip/irq-mvebu-[^/]*$|^drivers/pinctrl/mvebu/|^drivers/rtc/rtc-armada38x\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/aspeed-[^/]*$|^arch/arm/mach-aspeed/|aspeed\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/at91-linea\\\\.dtsi$|^arch/arm/boot/dts/at91-natte\\\\.dtsi$|^arch/arm/boot/dts/at91-nattis-2-natte-2\\\\.dts$|^arch/arm/boot/dts/at91-tse850-3\\\\.dts$\"},\n\t\t\t{\"^arch/arm/boot/dts/at91[^/]*\\\\.dts$|^arch/arm/boot/dts/at91[^/]*\\\\.dtsi$|^arch/arm/boot/dts/sama[^/]*\\\\.dts$|^arch/arm/boot/dts/sama[^/]*\\\\.dtsi$|^arch/arm/include/debug/at91\\\\.S$|^arch/arm/mach-at91/|^drivers/memory/atmel[^/]*$|^drivers/watchdog/sama5d4_wdt\\\\.c$|^include/soc/at91/|at91|atmel\", \"^drivers/input/touchscreen/atmel_mxt_ts\\\\.c$|^drivers/net/wireless/atmel/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/bcm470[^/]*$|^arch/arm/boot/dts/bcm5301[^/]*$|^arch/arm/boot/dts/bcm953012[^/]*$|^arch/arm/mach-bcm/bcm_5301x\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/bcm47189[^/]*$|^arch/arm/boot/dts/bcm53573[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/bcm7[^/]*\\\\.dts[^/]*$|^arch/arm/include/asm/hardware/cache-b15-rac\\\\.h$|^arch/arm/mach-bcm/[^/]*brcmstb[^/]*$|^arch/arm/mm/cache-b15-rac\\\\.c$|^drivers/bus/brcmstb_gisb\\\\.c$|^drivers/pci/controller/pcie-brcmstb\\\\.c$|brcmstb|bcm7038|bcm7120\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/berlin[^/]*$|^arch/arm/mach-berlin/|^arch/arm64/boot/dts/synaptics/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/cx92755[^/]*$|digicolor\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/da850[^/]*$|^arch/arm/mach-davinci/|^drivers/i2c/busses/i2c-davinci\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/dove[^/]*$|^arch/arm/boot/dts/orion5x[^/]*$|^arch/arm/mach-dove/|^arch/arm/mach-mv78xx0/|^arch/arm/mach-orion5x/|^arch/arm/plat-orion/|^drivers/soc/dove/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/ecx-[^/]*\\\\.dts[^/]*$|^arch/arm/boot/dts/highbank\\\\.dts$|^arch/arm/mach-highbank/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/exynos[^/]*$|^arch/arm/boot/dts/s3c[^/]*$|^arch/arm/boot/dts/s5p[^/]*$|^arch/arm/mach-exynos[^/]*/|^arch/arm/mach-s3c/|^arch/arm/mach-s5p[^/]*/|^arch/arm64/boot/dts/exynos/|^drivers/[^/]*/[^/]*/[^/]*s3c24[^/]*$|^drivers/[^/]*/[^/]*s3c24[^/]*$|^drivers/[^/]*/[^/]*s3c64xx[^/]*$|^drivers/[^/]*/[^/]*s5pv210[^/]*$|^drivers/clocksource/samsung_pwm_timer\\\\.c$|^drivers/memory/samsung/|^drivers/pwm/pwm-samsung\\\\.c$|^drivers/soc/samsung/|^drivers/tty/serial/samsung[^/]*$|^include/clocksource/samsung_pwm\\\\.h$|^include/linux/platform_data/[^/]*s3c[^/]*$|^include/linux/serial_s3c\\\\.h$|^include/linux/soc/samsung/|exynos|s3c64xx|s5pv210\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/gemini[^/]*$|^arch/arm/mach-gemini/|^drivers/crypto/gemini/|^drivers/net/ethernet/cortina/|^drivers/pinctrl/pinctrl-gemini\\\\.c$|^drivers/rtc/rtc-ftrtc010\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/hi3[^/]*$|^arch/arm/boot/dts/hip[^/]*$|^arch/arm/boot/dts/hisi[^/]*$|^arch/arm/mach-hisi/|^arch/arm64/boot/dts/hisilicon/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/intel-ixp[^/]*$|^arch/arm/mach-ixp4xx/|^drivers/bus/intel-ixp4xx-eb\\\\.c$|^drivers/clocksource/timer-ixp4xx\\\\.c$|^drivers/crypto/ixp4xx_crypto\\\\.c$|^drivers/gpio/gpio-ixp4xx\\\\.c$|^drivers/irqchip/irq-ixp4xx\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/keystone-[^/]*$|^arch/arm/mach-keystone/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/lpc32[^/]*$|^arch/arm/mach-lpc32xx/|^drivers/i2c/busses/i2c-pnx\\\\.c$|^drivers/net/ethernet/nxp/lpc_eth\\\\.c$|^drivers/usb/host/ohci-nxp\\\\.c$|^drivers/watchdog/pnx4008_wdt\\\\.c$|lpc32xx\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/lpc43[^/]*$|^drivers/i2c/busses/i2c-lpc2k\\\\.c$|^drivers/memory/pl172\\\\.c$|^drivers/mtd/spi-nor/controllers/nxp-spifi\\\\.c$|^drivers/rtc/rtc-lpc24xx\\\\.c$|lpc18xx\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/ls1021a[^/]*$|^arch/arm64/boot/dts/freescale/fsl-[^/]*$|^arch/arm64/boot/dts/freescale/qoriq-[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/meson[^/]*$|^arch/arm/mach-meson/|^arch/arm64/boot/dts/amlogic/|^drivers/mmc/host/meson[^/]*$|^drivers/pinctrl/meson/|^drivers/rtc/rtc-meson[^/]*$|^drivers/soc/amlogic/|meson\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/milbeaut[^/]*$|^arch/arm/mach-milbeaut/|milbeaut\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/mmp[^/]*$|^arch/arm/mach-mmp/|^include/linux/soc/mmp/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/mstar-[^/]*$|^arch/arm/mach-mstar/|^drivers/clk/mstar/|^drivers/clocksource/timer-msc313e\\\\.c$|^drivers/gpio/gpio-msc313\\\\.c$|^drivers/rtc/rtc-msc313\\\\.c$|^drivers/watchdog/msc313e_wdt\\\\.c$|^include/dt-bindings/clock/mstar-[^/]*$|^include/dt-bindings/gpio/msc313-gpio\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/mt2[^/]*$|^arch/arm/boot/dts/mt6[^/]*$|^arch/arm/boot/dts/mt7[^/]*$|^arch/arm/boot/dts/mt8[^/]*$|^arch/arm/mach-mediatek/|^arch/arm64/boot/dts/mediatek/|^drivers/soc/mediatek/|mtk|mt[2678]\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/omap3-igep[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/owl-[^/]*$|^arch/arm/mach-actions/|^arch/arm64/boot/dts/actions/|^drivers/clk/actions/|^drivers/clocksource/timer-owl[^/]*$|^drivers/dma/owl-dma\\\\.c$|^drivers/i2c/busses/i2c-owl\\\\.c$|^drivers/irqchip/irq-owl-sirq\\\\.c$|^drivers/mmc/host/owl-mmc\\\\.c$|^drivers/net/ethernet/actions/|^drivers/pinctrl/actions/|^drivers/soc/actions/|^include/dt-bindings/power/owl-[^/]*$|^include/dt-bindings/reset/actions,[^/]*$|^include/linux/soc/actions/|owl\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/ox8[^/]*\\\\.dts[^/]*$|^arch/arm/mach-oxnas/|^drivers/power/reset/oxnas-restart\\\\.c$|oxnas\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/pxa[^/]*$|^arch/arm/mach-pxa/|^drivers/dma/pxa[^/]*$|^drivers/pcmcia/pxa2xx[^/]*$|^drivers/pinctrl/pxa/|^drivers/spi/spi-pxa2xx[^/]*$|^drivers/usb/gadget/udc/pxa2[^/]*$|^include/sound/pxa2xx-lib\\\\.h$|^sound/arm/pxa[^/]*$|^sound/soc/pxa/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/rda8810pl-[^/]*$|^drivers/clocksource/timer-rda\\\\.c$|^drivers/gpio/gpio-rda\\\\.c$|^drivers/irqchip/irq-rda-intc\\\\.c$|^drivers/tty/serial/rda-uart\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/rk3[^/]*$|^arch/arm/boot/dts/rv11[^/]*$|^arch/arm/mach-rockchip/|^drivers/[^/]*/[^/]*/[^/]*rockchip[^/]*$|^drivers/[^/]*/[^/]*rockchip[^/]*$|^drivers/clk/rockchip/|^drivers/i2c/busses/i2c-rk3x\\\\.c$|^sound/soc/rockchip/|rockchip\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/rtd[^/]*$|^arch/arm/mach-realtek/|^arch/arm64/boot/dts/realtek/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/spear[^/]*$|^arch/arm/mach-spear/|^drivers/clk/spear/|^drivers/pinctrl/spear/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/ste-[^/]*$|^arch/arm/mach-nomadik/|^arch/arm/mach-ux500/|^drivers/clk/clk-nomadik\\\\.c$|^drivers/clocksource/clksrc-dbx500-prcmu\\\\.c$|^drivers/dma/ste_dma40[^/]*$|^drivers/hwspinlock/u8500_hsem\\\\.c$|^drivers/i2c/busses/i2c-nomadik\\\\.c$|^drivers/iio/adc/ab8500-gpadc\\\\.c$|^drivers/mfd/ab8500[^/]*$|^drivers/mfd/abx500[^/]*$|^drivers/mfd/db8500[^/]*$|^drivers/pinctrl/nomadik/|^drivers/rtc/rtc-ab8500\\\\.c$|^drivers/rtc/rtc-pl031\\\\.c$|^drivers/soc/ux500/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/sti[^/]*$|^arch/arm/mach-sti/|^drivers/ata/ahci_st\\\\.c$|^drivers/char/hw_random/st-rng\\\\.c$|^drivers/clocksource/arm_global_timer\\\\.c$|^drivers/clocksource/clksrc_st_lpc\\\\.c$|^drivers/cpufreq/sti-cpufreq\\\\.c$|^drivers/dma/st_fdma[^/]*$|^drivers/i2c/busses/i2c-st\\\\.c$|^drivers/media/platform/st/sti/c8sectpfe/|^drivers/media/rc/st_rc\\\\.c$|^drivers/mmc/host/sdhci-st\\\\.c$|^drivers/phy/st/phy-miphy28lp\\\\.c$|^drivers/phy/st/phy-stih407-usb\\\\.c$|^drivers/pinctrl/pinctrl-st\\\\.c$|^drivers/remoteproc/st_remoteproc\\\\.c$|^drivers/remoteproc/st_slim_rproc\\\\.c$|^drivers/reset/sti/|^drivers/rtc/rtc-st-lpc\\\\.c$|^drivers/tty/serial/st-asc\\\\.c$|^drivers/usb/dwc3/dwc3-st\\\\.c$|^drivers/usb/host/ehci-st\\\\.c$|^drivers/usb/host/ohci-st\\\\.c$|^drivers/watchdog/st_lpc_wdt\\\\.c$|^include/linux/remoteproc/st_slim_rproc\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/stm32[^/]*$|^arch/arm/mach-stm32/|^drivers/clocksource/armv7m_systick\\\\.c$|stm32|stm\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/sunplus-sp7021[^/]*\\\\.dts[^/]*$|^arch/arm/configs/sp7021_[^/]*defconfig$|^arch/arm/mach-sunplus/|^drivers/irqchip/irq-sp7021-intc\\\\.c$|^drivers/reset/reset-sunplus\\\\.c$|^include/dt-bindings/clock/sunplus,sp7021-clkc\\\\.h$|^include/dt-bindings/reset/sunplus,sp7021-reset\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/uniphier[^/]*$|^arch/arm/include/asm/hardware/cache-uniphier\\\\.h$|^arch/arm/mach-uniphier/|^arch/arm/mm/cache-uniphier\\\\.c$|^arch/arm64/boot/dts/socionext/uniphier[^/]*$|^drivers/bus/uniphier-system-bus\\\\.c$|^drivers/clk/uniphier/|^drivers/dma/uniphier-mdmac\\\\.c$|^drivers/gpio/gpio-uniphier\\\\.c$|^drivers/i2c/busses/i2c-uniphier[^/]*$|^drivers/irqchip/irq-uniphier-aidet\\\\.c$|^drivers/mmc/host/uniphier-sd\\\\.c$|^drivers/pinctrl/uniphier/|^drivers/reset/reset-uniphier\\\\.c$|^drivers/tty/serial/8250/8250_uniphier\\\\.c$|uniphier\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/vf[^/]*$|^arch/arm/mach-imx/[^/]*vf610[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/include/asm/arch_timer\\\\.h$|^arch/arm64/include/asm/arch_timer\\\\.h$|^drivers/clocksource/arm_arch_timer\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/include/asm/hardware/dec21285\\\\.h$|^arch/arm/mach-footbridge/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/include/asm/hardware/ioc\\\\.h$|^arch/arm/include/asm/hardware/iomd\\\\.h$|^arch/arm/include/asm/hardware/memc\\\\.h$|^arch/arm/mach-rpc/|^drivers/net/ethernet/8390/etherh\\\\.c$|^drivers/net/ethernet/i825xx/ether1[^/]*$|^drivers/net/ethernet/seeq/ether3[^/]*$|^drivers/scsi/arm/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/mach-[^/]*/|^arch/arm/plat-[^/]*/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/mach-ep93xx/ts72xx\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/mach-orion5x/ts78xx-[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/mach-sunxi/|^arch/arm64/boot/dts/allwinner/|^drivers/clk/sunxi-ng/|^drivers/pinctrl/sunxi/|^drivers/soc/sunxi/|allwinner|sun[x456789]i|sun[25]0i\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/mach-vt8500/|^drivers/clocksource/timer-vt8500\\\\.c$|^drivers/i2c/busses/i2c-wmt\\\\.c$|^drivers/mmc/host/wmt-sdmmc\\\\.c$|^drivers/pwm/pwm-vt8500\\\\.c$|^drivers/rtc/rtc-vt8500\\\\.c$|^drivers/tty/serial/vt8500_serial\\\\.c$|^drivers/usb/host/ehci-platform\\\\.c$|^drivers/usb/host/uhci-platform\\\\.c$|^drivers/video/fbdev/vt8500lcdfb\\\\.[^/]*$|^drivers/video/fbdev/wm8505fb[^/]*$|^drivers/video/fbdev/wmt_ge_rops\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/mach-zynq/|^drivers/clocksource/timer-cadence-ttc\\\\.c$|^drivers/cpuidle/cpuidle-zynq\\\\.c$|^drivers/edac/synopsys_edac\\\\.c$|^drivers/i2c/busses/i2c-cadence\\\\.c$|^drivers/i2c/busses/i2c-xiic\\\\.c$|^drivers/mmc/host/sdhci-of-arasan\\\\.c$|zynq|xilinx\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/mm/[^/]*-fa[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/vfp/\"},\n\t\t\t{\"^arch/arm64/\", \"^arch/arm64/boot/dts/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm64/boot/dts/apple/|^drivers/bluetooth/hci_bcm4377\\\\.c$|^drivers/clk/clk-apple-nco\\\\.c$|^drivers/cpufreq/apple-soc-cpufreq\\\\.c$|^drivers/dma/apple-admac\\\\.c$|^drivers/i2c/busses/i2c-pasemi-core\\\\.c$|^drivers/i2c/busses/i2c-pasemi-platform\\\\.c$|^drivers/iommu/apple-dart\\\\.c$|^drivers/iommu/io-pgtable-dart\\\\.c$|^drivers/irqchip/irq-apple-aic\\\\.c$|^drivers/mailbox/apple-mailbox\\\\.c$|^drivers/nvme/host/apple\\\\.c$|^drivers/nvmem/apple-efuses\\\\.c$|^drivers/pinctrl/pinctrl-apple-gpio\\\\.c$|^drivers/soc/apple/|^drivers/watchdog/apple_wdt\\\\.c$|^include/dt-bindings/interrupt-controller/apple-aic\\\\.h$|^include/dt-bindings/pinctrl/apple\\\\.h$|^include/linux/apple-mailbox\\\\.h$|^include/linux/soc/apple/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm64/boot/dts/bitmain/|^drivers/clk/clk-bm1880\\\\.c$|^drivers/pinctrl/pinctrl-bm1880\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm64/boot/dts/broadcom/bcmbca/|bcmbca|bcm[9]?47622|bcm[9]?4912|bcm[9]?63138|bcm[9]?63146|bcm[9]?63148|bcm[9]?63158|bcm[9]?63178|bcm[9]?6756|bcm[9]?6813|bcm[9]?6846|bcm[9]?6855|bcm[9]?6856|bcm[9]?6858|bcm[9]?6878\"},\n\t\t\t{IncludeRegexp: \"^arch/arm64/boot/dts/broadcom/northstar2/|^arch/arm64/boot/dts/broadcom/stingray/|^drivers/clk/bcm/clk-ns[^/]*$|^drivers/clk/bcm/clk-sr[^/]*$|^drivers/pinctrl/bcm/pinctrl-ns[^/]*$|^include/dt-bindings/clock/bcm-sr[^/]*$|iproc|cygnus|bcm[-_]nsp|bcm9113*|bcm9583*|bcm9585*|bcm9586*|bcm988312|bcm113*|bcm583*|bcm585*|bcm586*|bcm88312|hr2|stingray\"},\n\t\t\t{IncludeRegexp: \"^arch/arm64/boot/dts/cavium/thunder2-99xx[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm64/boot/dts/freescale/s32g[^/]*\\\\.dts[^/]*$\"},\n\t\t\t{\"^arch/arm64/boot/dts/freescale/|imx|mxs\", \"^drivers/media/i2c/|^arch/arm64/boot/dts/freescale/fsl-[^/]*$|^arch/arm64/boot/dts/freescale/qoriq-[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm64/boot/dts/lg/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm64/boot/dts/marvell/armada-8040-mcbin\\\\.dts$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm64/boot/dts/microchip/|^drivers/net/ethernet/microchip/vcap/|^drivers/pinctrl/pinctrl-microchip-sgpio\\\\.c$|sparx5\"},\n\t\t\t{IncludeRegexp: \"^arch/arm64/boot/dts/tesla/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm64/boot/dts/ti/k3-[^/]*$|^include/dt-bindings/pinctrl/k3\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm64/boot/dts/toshiba/|^drivers/clk/visconti/|^drivers/net/ethernet/stmicro/stmmac/dwmac-visconti\\\\.c$|^drivers/gpio/gpio-visconti\\\\.c$|^drivers/pci/controller/dwc/pcie-visconti\\\\.c$|^drivers/pinctrl/visconti/|^drivers/watchdog/visconti_wdt\\\\.c$|visconti\"},\n\t\t\t{IncludeRegexp: \"^arch/arm64/include/asm/kvm[^/]*$|^arch/arm64/include/uapi/asm/kvm[^/]*$|^arch/arm64/kvm/|^include/kvm/arm_[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm[^/]*/include/asm/hw_breakpoint\\\\.h$|^arch/arm[^/]*/include/asm/perf_event\\\\.h$|^arch/arm[^/]*/kernel/hw_breakpoint\\\\.c$|^arch/arm[^/]*/kernel/perf_[^/]*$|^drivers/perf/|^include/linux/perf/arm_pmu\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/acpi/arm64$\"},\n\t\t\t{IncludeRegexp: \"^drivers/clk/clk-sc\\\\[mp\\\\]i\\\\.c$|^drivers/cpufreq/sc\\\\[mp\\\\]i-cpufreq\\\\.c$|^drivers/firmware/arm_scmi/|^drivers/firmware/arm_scpi\\\\.c$|^drivers/powercap/arm_scmi_powercap\\\\.c$|^drivers/regulator/scmi-regulator\\\\.c$|^drivers/reset/reset-scmi\\\\.c$|^include/linux/sc\\\\[mp\\\\]i_protocol\\\\.h$|^include/trace/events/scmi\\\\.h$|^include/uapi/linux/virtio_scmi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/clk/clkdev\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/clk/keystone/sci-clk\\\\.c$|^drivers/firmware/ti_sci[^/]*$|^drivers/irqchip/irq-ti-sci-inta\\\\.c$|^drivers/irqchip/irq-ti-sci-intr\\\\.c$|^drivers/reset/reset-ti-sci\\\\.c$|^drivers/soc/ti/ti_sci_inta_msi\\\\.c$|^drivers/soc/ti/ti_sci_pm_domains\\\\.c$|^include/dt-bindings/soc/ti,sci_pm_domain\\\\.h$|^include/linux/soc/ti/ti_sci_inta_msi\\\\.h$|^include/linux/soc/ti/ti_sci_protocol\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/clk/ux500/\"},\n\t\t\t{IncludeRegexp: \"^drivers/clocksource/timer-keystone\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/counter/microchip-tcb-capture\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/cpuidle/cpuidle-big_little\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/cpuidle/cpuidle-psci\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/cpuidle/cpuidle-psci\\\\.h$|^drivers/cpuidle/cpuidle-psci-domain\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma/at_hdmac\\\\.c$|^drivers/dma/at_xdmac\\\\.c$|^include/dt-bindings/dma/at91\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma/mediatek/\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma/qcom/hidma[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/firmware/arm_ffa/|^include/linux/arm_ffa\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/firmware/arm_sdei\\\\.c$|^include/linux/arm_sdei\\\\.h$|^include/uapi/linux/arm_sdei\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/firmware/psci/|^include/linux/psci\\\\.h$|^include/uapi/linux/psci\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/firmware/smccc/|^include/linux/arm-smccc\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-sama5d2-piobu\\\\.c$|^drivers/pinctrl/pinctrl-at91[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwtracing/coresight/|^include/dt-bindings/arm/coresight-cti-dt\\\\.h$|^include/linux/coresight[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iommu/arm/|^drivers/iommu/io-pgtable-arm[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/irqchip/irq-mchp-eic\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/irqchip/irq-vic\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/samsung/s5p-g2d/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/samsung/s5p-jpeg/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/samsung/s5p-mfc/\"},\n\t\t\t{IncludeRegexp: \"^drivers/memory/brcmstb_dpfe\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/memory/pl353-smc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/misc/atmel-ssc\\\\.c$|^include/linux/atmel-ssc\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/cavium/thunder/\"},\n\t\t\t{IncludeRegexp: \"^drivers/nvmem/microchip-otpc\\\\.c$|^include/dt-bindings/nvmem/microchip,sama7g5-otpc\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/[^/]*mvebu[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/cadence/pci-j721e\\\\.c$|^drivers/pci/controller/dwc/pci-dra7xx\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/[^/]*imx6[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/[^/]*layerscape[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/pci-exynos\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/pcie-armada8k\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/mobiveil/pcie-layerscape-gen4\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pci-aardvark\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pci-host-common\\\\.c$|^drivers/pci/controller/pci-host-generic\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pci-thunder-[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pci-versatile\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pci-xgene-msi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pci-xgene\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pcie-brcmstb\\\\.c$|^drivers/staging/vc04_services$|bcm2711|bcm283*|raspberrypi\"},\n\t\t\t{IncludeRegexp: \"^drivers/perf/fsl_imx8_ddr_perf\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/phy/marvell/phy-mmp3-usb\\\\.c$|^drivers/phy/marvell/phy-pxa-usb\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/phy/mediatek/\"},\n\t\t\t{IncludeRegexp: \"^drivers/pinctrl/pinctrl-single\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pinctrl/samsung/|^include/dt-bindings/pinctrl/samsung\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pinctrl/sunplus/|^include/dt-bindings/pinctrl/sppctl[^/]*\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pwm/pwm-atmel\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/regulator/mcp16502\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/rtc/rtc-mt2712\\\\.c$|^drivers/rtc/rtc-mt6397\\\\.c$|^drivers/rtc/rtc-mt7622\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/soc/fsl/|^include/linux/fsl/|^include/soc/fsl/\"},\n\t\t\t{IncludeRegexp: \"^drivers/soc/ti/\"},\n\t\t\t{IncludeRegexp: \"^drivers/spi/spi-pl022\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/gadget/udc/atmel_usba_udc\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/host/xhci-mtk[^/]*$|^drivers/usb/mtu3/\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/cyber2000fb\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/imxfb\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"clps711x\"},\n\t\t},\n\t}\n\n\tarmmsm = Subsystem{\n\t\tName: \"arm-msm\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/qcom-[^/]*\\\\.dts$|^arch/arm/boot/dts/qcom-[^/]*\\\\.dtsi$|^arch/arm/configs/qcom_defconfig$|^arch/arm/mach-qcom/|^arch/arm64/boot/dts/qcom/|^drivers/[^/]*/[^/]*/qcom[^/]*$|^drivers/[^/]*/[^/]*/qcom/|^drivers/[^/]*/pm8...-[^/]*$|^drivers/[^/]*/qcom[^/]*$|^drivers/[^/]*/qcom/|^drivers/bluetooth/btqcomsmd\\\\.c$|^drivers/clocksource/timer-qcom\\\\.c$|^drivers/cpuidle/cpuidle-qcom-spm\\\\.c$|^drivers/extcon/extcon-qcom[^/]*$|^drivers/i2c/busses/i2c-qcom-geni\\\\.c$|^drivers/i2c/busses/i2c-qup\\\\.c$|^drivers/iommu/msm[^/]*$|^drivers/mfd/ssbi\\\\.c$|^drivers/mmc/host/mmci_qcom[^/]*$|^drivers/mmc/host/sdhci-msm\\\\.c$|^drivers/pci/controller/dwc/pcie-qcom\\\\.c$|^drivers/phy/qualcomm/|^drivers/power/[^/]*/msm[^/]*$|^drivers/reset/reset-qcom-[^/]*$|^drivers/ufs/host/ufs-qcom[^/]*$|^drivers/spi/spi-geni-qcom\\\\.c$|^drivers/spi/spi-qcom-qspi\\\\.c$|^drivers/spi/spi-qup\\\\.c$|^drivers/tty/serial/msm_serial\\\\.c$|^drivers/usb/dwc3/dwc3-qcom\\\\.c$|^include/dt-bindings/[^/]*/qcom[^/]*$|^include/linux/[^/]*/qcom[^/]*$|^include/linux/soc/qcom/\"},\n\t\t\t{IncludeRegexp: \"^drivers/bus/mhi/|^include/linux/mhi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/clk/qcom/|^include/dt-bindings/clock/qcom,[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/qce/\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma/qcom/hidma[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/qcom_edac\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/msm/|^include/uapi/drm/msm_drm\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-qcom-cci\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iommu/arm/arm-smmu/qcom_iommu\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mailbox/qcom-ipcc\\\\.c$|^include/dt-bindings/mailbox/qcom-ipcc\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/qcom/venus/\"},\n\t\t\t{IncludeRegexp: \"^drivers/misc/fastrpc\\\\.c$|^include/uapi/misc/fastrpc\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/nand/raw/qcom_nandc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wwan/qcom_bam_dmux\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/pcie-qcom-ep\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/pcie-qcom\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/phy/qualcomm/phy-qcom-ipq4019-usb\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pinctrl/qcom/\"},\n\t\t\t{IncludeRegexp: \"^drivers/regulator/vqmmc-ipq4019-regulator\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/soc/qcom/cpr\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/soc/qcom/icc-bwmon\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/thermal/qcom/\"},\n\t\t\t{IncludeRegexp: \"^drivers/ufs/host/ufs-qcom[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/misc/qcom_eud\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/trace/events/qrtr\\\\.h$|^include/uapi/linux/qrtr\\\\.h$|^net/qrtr/\"},\n\t\t},\n\t}\n\n\tasahi = Subsystem{\n\t\tName: \"asahi\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&arm},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/arm64/boot/dts/apple/|^drivers/bluetooth/hci_bcm4377\\\\.c$|^drivers/clk/clk-apple-nco\\\\.c$|^drivers/cpufreq/apple-soc-cpufreq\\\\.c$|^drivers/dma/apple-admac\\\\.c$|^drivers/i2c/busses/i2c-pasemi-core\\\\.c$|^drivers/i2c/busses/i2c-pasemi-platform\\\\.c$|^drivers/iommu/apple-dart\\\\.c$|^drivers/iommu/io-pgtable-dart\\\\.c$|^drivers/irqchip/irq-apple-aic\\\\.c$|^drivers/mailbox/apple-mailbox\\\\.c$|^drivers/nvme/host/apple\\\\.c$|^drivers/nvmem/apple-efuses\\\\.c$|^drivers/pinctrl/pinctrl-apple-gpio\\\\.c$|^drivers/soc/apple/|^drivers/watchdog/apple_wdt\\\\.c$|^include/dt-bindings/interrupt-controller/apple-aic\\\\.h$|^include/dt-bindings/pinctrl/apple\\\\.h$|^include/linux/apple-mailbox\\\\.h$|^include/linux/soc/apple/\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/apple/|^sound/soc/codecs/cs42l83-i2c\\\\.c$\"},\n\t\t},\n\t}\n\n\taspeed = Subsystem{\n\t\tName: \"aspeed\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&arm},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/aspeed-[^/]*$|^arch/arm/mach-aspeed/|aspeed\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/aspeed/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/aspeed/\"},\n\t\t\t{IncludeRegexp: \"^drivers/irqchip/irq-aspeed-scu-ic\\\\.c$|^include/dt-bindings/interrupt-controller/aspeed-scu-ic\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/host/sdhci-of-aspeed[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/peci/controller/peci-aspeed\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pinctrl/aspeed/\"},\n\t\t\t{IncludeRegexp: \"^drivers/spi/spi-aspeed-smc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/gadget/udc/aspeed_udc\\\\.c$\"},\n\t\t},\n\t}\n\n\tath10k = Subsystem{\n\t\tName: \"ath10k\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&wireless},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/ath/ath10k/\"},\n\t\t},\n\t}\n\n\tath11k = Subsystem{\n\t\tName: \"ath11k\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&wireless},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/ath/ath11k/\"},\n\t\t},\n\t}\n\n\tath12k = Subsystem{\n\t\tName: \"ath12k\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&wireless},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/ath/ath12k/\"},\n\t\t},\n\t}\n\n\tatm = Subsystem{\n\t\tName: \"atm\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/atm/|^include/linux/atm[^/]*$|^include/uapi/linux/atm[^/]*$\"},\n\t\t},\n\t}\n\n\taudit = Subsystem{\n\t\tName: \"audit\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^include/asm-generic/audit_[^/]*\\\\.h$|^include/linux/audit\\\\.h$|^include/linux/audit_arch\\\\.h$|^include/uapi/linux/audit\\\\.h$|^kernel/audit[^/]*$|^lib/[^/]*audit\\\\.c$\"},\n\t\t},\n\t}\n\n\tautofs = Subsystem{\n\t\tName: \"autofs\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/autofs/\"},\n\t\t},\n\t}\n\n\taxis = Subsystem{\n\t\tName: \"axis\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/artpec6[^/]*$|^arch/arm/mach-artpec$|^drivers/clk/axis$|^drivers/crypto/axis$|^drivers/mmc/host/usdhi6rol0\\\\.c$|^drivers/pinctrl/pinctrl-artpec[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/[^/]*artpec[^/]*$\"},\n\t\t},\n\t}\n\n\tb43 = Subsystem{\n\t\tName: \"b43\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&wireless},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/broadcom/b43/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/broadcom/b43legacy/\"},\n\t\t},\n\t}\n\n\tbatman = Subsystem{\n\t\tName: \"batman\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^include/uapi/linux/batadv_packet\\\\.h$|^include/uapi/linux/batman_adv\\\\.h$|^net/batman-adv/\"},\n\t\t},\n\t}\n\n\tbcache = Subsystem{\n\t\tName: \"bcache\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/md/bcache/\"},\n\t\t},\n\t}\n\n\tbfs = Subsystem{\n\t\tName: \"bfs\",\n\t\tSyscalls: []string{\"syz_mount_image$bfs\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/bfs/|^include/uapi/linux/bfs_fs\\\\.h$\"},\n\t\t},\n\t}\n\n\tblock = Subsystem{\n\t\tName: \"block\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^block/bfq-[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^block/bfq-cgroup\\\\.c$|^block/blk-cgroup\\\\.c$|^block/blk-iocost\\\\.c$|^block/blk-iolatency\\\\.c$|^block/blk-throttle\\\\.c$|^include/linux/blk-cgroup\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^block/opal_proto\\\\.h$|^block/sed[^/]*$|^include/linux/sed[^/]*$|^include/uapi/linux/sed[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^block/|^drivers/block/|^include/linux/bio\\\\.h$|^include/linux/blk[^/]*$|^kernel/trace/blktrace\\\\.c$|^lib/sbitmap\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/block/floppy\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/block/nbd\\\\.c$|^include/trace/events/nbd\\\\.h$|^include/uapi/linux/nbd\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/block/rnbd/\"},\n\t\t\t{IncludeRegexp: \"^drivers/block/ublk_drv\\\\.c$|^include/uapi/linux/ublk_cmd\\\\.h$\"},\n\t\t},\n\t}\n\n\tbluetooth = Subsystem{\n\t\tName: \"bluetooth\",\n\t\tSyscalls: []string{\"syz_emit_vhci\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/bluetooth/\"},\n\t\t\t{IncludeRegexp: \"^drivers/bluetooth/btmtkuart\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/net/6lowpan\\\\.h$|^net/6lowpan/\"},\n\t\t\t{IncludeRegexp: \"^include/net/bluetooth/|^net/bluetooth/\"},\n\t\t},\n\t}\n\n\tbpf = Subsystem{\n\t\tName: \"bpf\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/[^/]*/net/|^include/linux/bpf[^/]*$|^include/linux/btf[^/]*$|^include/linux/filter\\\\.h$|^include/trace/events/xdp\\\\.h$|^include/uapi/linux/bpf[^/]*$|^include/uapi/linux/btf[^/]*$|^include/uapi/linux/filter\\\\.h$|^kernel/bpf/|^kernel/trace/bpf_trace\\\\.c$|^lib/test_bpf\\\\.c$|^net/bpf/|^net/core/filter\\\\.c$|^net/sched/act_bpf\\\\.c$|^net/sched/cls_bpf\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/net/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm64/net/\"},\n\t\t\t{IncludeRegexp: \"^arch/mips/net/\"},\n\t\t\t{IncludeRegexp: \"^arch/powerpc/net/\"},\n\t\t\t{\"^arch/riscv/net/\", \"^arch/riscv/net/bpf_jit_comp32\\\\.c$\"},\n\t\t\t{\"^arch/riscv/net/\", \"^arch/riscv/net/bpf_jit_comp64\\\\.c$\"},\n\t\t\t{\"^arch/s390/net/\", \"^arch/s390/net/pnet\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/sparc/net/\"},\n\t\t\t{\"^arch/x86/net/\", \"^arch/x86/net/bpf_jit_comp32\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/net/bpf_jit_comp32\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/netronome/nfp/bpf/\"},\n\t\t\t{IncludeRegexp: \"^include/linux/bpf_lsm\\\\.h$|^kernel/bpf/bpf_lsm\\\\.c$|^security/bpf/\"},\n\t\t\t{IncludeRegexp: \"^include/linux/skmsg\\\\.h$|^net/core/skmsg\\\\.c$|^net/core/sock_map\\\\.c$|^net/ipv4/tcp_bpf\\\\.c$|^net/ipv4/udp_bpf\\\\.c$|^net/unix/unix_bpf\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/net/xdp\\\\.h$|^include/net/xdp_priv\\\\.h$|^include/trace/events/xdp\\\\.h$|^kernel/bpf/cpumap\\\\.c$|^kernel/bpf/devmap\\\\.c$|^net/core/xdp\\\\.c$|^drivers/net/ethernet/[^/]*/[^/]*/[^/]*/[^/]*/[^/]*xdp[^/]*$|^drivers/net/ethernet/[^/]*/[^/]*/[^/]*xdp[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^include/net/xdp_sock[^/]*$|^include/net/xsk_buff_pool\\\\.h$|^include/uapi/linux/if_xdp\\\\.h$|^include/uapi/linux/xdp_diag\\\\.h$|^include/net/netns/xdp\\\\.h$|^net/xdp/\"},\n\t\t\t{IncludeRegexp: \"^kernel/bpf/[^/]*iter\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^kernel/bpf/bpf_struct[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^kernel/bpf/btf\\\\.c$|^include/linux/btf[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^kernel/bpf/cgroup\\\\.c$|^kernel/bpf/[^/]*storage\\\\.c$|^kernel/bpf/bpf_lru[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^kernel/bpf/disasm\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^kernel/bpf/ringbuf\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^kernel/bpf/verifier\\\\.c$|^kernel/bpf/tnum\\\\.c$|^kernel/bpf/core\\\\.c$|^kernel/bpf/syscall\\\\.c$|^kernel/bpf/dispatcher\\\\.c$|^kernel/bpf/trampoline\\\\.c$|^include/linux/bpf[^/]*$|^include/linux/filter\\\\.h$|^include/linux/tnum\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^kernel/trace/bpf_trace\\\\.c$|^kernel/bpf/stackmap\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^net/core/filter\\\\.c$|^net/sched/act_bpf\\\\.c$|^net/sched/cls_bpf\\\\.c$\"},\n\t\t},\n\t}\n\n\tbrcm80211 = Subsystem{\n\t\tName: \"brcm80211\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&wireless},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/broadcom/brcm80211/\"},\n\t\t},\n\t}\n\n\tbridge = Subsystem{\n\t\tName: \"bridge\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^include/linux/netfilter_bridge/|^net/bridge/\"},\n\t\t},\n\t}\n\n\tbtrfs = Subsystem{\n\t\tName: \"btrfs\",\n\t\tSyscalls: []string{\"syz_mount_image$btrfs\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/btrfs/|^include/linux/btrfs[^/]*$|^include/trace/events/btrfs\\\\.h$|^include/uapi/linux/btrfs[^/]*$\"},\n\t\t},\n\t}\n\n\tcachefs = Subsystem{\n\t\tName: \"cachefs\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/cachefiles/\"},\n\t\t\t{IncludeRegexp: \"^fs/fscache/|^include/linux/fscache[^/]*\\\\.h$\"},\n\t\t},\n\t}\n\n\tcan = Subsystem{\n\t\tName: \"can\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/net/can/can327\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/can/ctucanfd/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/can/m_can/m_can\\\\.c$|^drivers/net/can/m_can/m_can\\\\.h$|^drivers/net/can/m_can/m_can_platform\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/can/slcan/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/can/spi/mcp251xfd/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/can/usb/esd_usb\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/can/usb/etas_es58x/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/can/usb/mcba_usb\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/can/xilinx_can\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/can/|^drivers/phy/phy-can-transceiver\\\\.c$|^include/linux/can/bittiming\\\\.h$|^include/linux/can/dev\\\\.h$|^include/linux/can/length\\\\.h$|^include/linux/can/platform/|^include/linux/can/rx-offload\\\\.h$|^include/uapi/linux/can/error\\\\.h$|^include/uapi/linux/can/netlink\\\\.h$|^include/uapi/linux/can/vxcan\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/can/can-ml\\\\.h$|^include/linux/can/core\\\\.h$|^include/linux/can/skb\\\\.h$|^include/net/netns/can\\\\.h$|^include/uapi/linux/can\\\\.h$|^include/uapi/linux/can/bcm\\\\.h$|^include/uapi/linux/can/gw\\\\.h$|^include/uapi/linux/can/isotp\\\\.h$|^include/uapi/linux/can/raw\\\\.h$|^net/can/\"},\n\t\t\t{IncludeRegexp: \"^include/uapi/linux/can/j1939\\\\.h$|^net/can/j1939/\"},\n\t\t},\n\t}\n\n\tceph = Subsystem{\n\t\tName: \"ceph\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/block/rbd\\\\.c$|^drivers/block/rbd_types\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^fs/ceph/\"},\n\t\t\t{IncludeRegexp: \"^include/linux/ceph/|^include/linux/crush/|^net/ceph/\"},\n\t\t},\n\t}\n\n\tcgroups = Subsystem{\n\t\tName: \"cgroups\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^block/bfq-cgroup\\\\.c$|^block/blk-cgroup\\\\.c$|^block/blk-iocost\\\\.c$|^block/blk-iolatency\\\\.c$|^block/blk-throttle\\\\.c$|^include/linux/blk-cgroup\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/cgroup[^/]*$|^kernel/cgroup/\"},\n\t\t\t{IncludeRegexp: \"^include/linux/cpuset\\\\.h$|^kernel/cgroup/cpuset\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^mm/memcontrol\\\\.c$|^mm/swap_cgroup\\\\.c$\"},\n\t\t},\n\t}\n\n\tchrome = Subsystem{\n\t\tName: \"chrome\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/platform/chrome/\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/chrome/cros_ec_typec\\\\.[^/]*$|^drivers/platform/chrome/cros_typec_switch\\\\.c$|^drivers/platform/chrome/cros_typec_vdm\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/chrome/cros_usbpd_notify\\\\.c$|^include/linux/platform_data/cros_usbpd_notify\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/power/supply/cros_usbpd-charger\\\\.c$|cros_ec|cros-ec\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/codecs/cros_ec_codec\\\\.[^/]*$\"},\n\t\t},\n\t}\n\n\tcifs = Subsystem{\n\t\tName: \"cifs\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/cifs/|^fs/smbfs_common/|^include/uapi/linux/cifs$\"},\n\t\t\t{IncludeRegexp: \"^fs/ksmbd/|^fs/smbfs_common/\"},\n\t\t},\n\t}\n\n\tcirrus = Subsystem{\n\t\tName: \"cirrus\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&alsa},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/arm/mach-s3c/mach-crag6410[^/]*$|^drivers/clk/clk-wm83[^/]*\\\\.c$|^drivers/gpio/gpio-[^/]*wm[^/]*\\\\.c$|^drivers/gpio/gpio-arizona\\\\.c$|^drivers/hwmon/wm83..-hwmon\\\\.c$|^drivers/input/misc/wm831x-on\\\\.c$|^drivers/input/touchscreen/wm831x-ts\\\\.c$|^drivers/input/touchscreen/wm97[^/]*\\\\.c$|^drivers/leds/leds-wm83[^/]*\\\\.c$|^drivers/mfd/arizona[^/]*$|^drivers/mfd/cs47l24[^/]*$|^drivers/mfd/wm[^/]*\\\\.c$|^drivers/power/supply/wm83[^/]*\\\\.c$|^drivers/regulator/arizona[^/]*$|^drivers/regulator/wm8[^/]*\\\\.c$|^drivers/rtc/rtc-wm83[^/]*\\\\.c$|^drivers/video/backlight/wm83[^/]*_bl\\\\.c$|^drivers/watchdog/wm83[^/]*_wdt\\\\.c$|^include/linux/mfd/arizona/|^include/linux/mfd/wm831x/|^include/linux/mfd/wm8350/|^include/linux/mfd/wm8400[^/]*$|^include/linux/regulator/arizona[^/]*$|^include/linux/wm97xx\\\\.h$|^include/sound/wm....\\\\.h$|^sound/soc/codecs/arizona[^/]*$|^sound/soc/codecs/cs47l24[^/]*$|^sound/soc/codecs/wm[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/clk/clk-lochnagar\\\\.c$|^drivers/hwmon/lochnagar-hwmon\\\\.c$|^drivers/mfd/lochnagar-i2c\\\\.c$|^drivers/pinctrl/cirrus/pinctrl-lochnagar\\\\.c$|^drivers/regulator/lochnagar-regulator\\\\.c$|^include/dt-bindings/clock/lochnagar\\\\.h$|^include/dt-bindings/pinctrl/lochnagar\\\\.h$|^include/linux/mfd/lochnagar[^/]*$|^sound/soc/codecs/lochnagar-sc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/firmware/cirrus/|^include/linux/firmware/cirrus/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-madera[^/]*$|^drivers/irqchip/irq-madera[^/]*$|^drivers/mfd/cs47l[^/]*$|^drivers/mfd/madera[^/]*$|^drivers/pinctrl/cirrus/|^include/dt-bindings/sound/madera[^/]*$|^include/linux/irqchip/irq-madera[^/]*$|^include/linux/mfd/madera/|^include/sound/madera[^/]*$|^sound/soc/codecs/cs47l[^/]*$|^sound/soc/codecs/madera[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^include/dt-bindings/sound/cs[^/]*$|^sound/pci/hda/cs[^/]*$|^sound/pci/hda/hda_cs_dsp_ctl\\\\.[^/]*$|^sound/soc/codecs/cs[^/]*$\"},\n\t\t},\n\t}\n\n\tclk = Subsystem{\n\t\tName: \"clk\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/clk/imx/|^include/dt-bindings/clock/imx[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/clk/ux500/\"},\n\t\t\t{\"^drivers/clk/|^include/dt-bindings/clock/|^include/linux/clk-pr[^/]*$|^include/linux/clk/|^include/linux/of_clk\\\\.h$\", \"^drivers/clk/clkdev\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/dt-bindings/clock/loongson,ls2k-clk\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/clk\\\\.h$\"},\n\t\t},\n\t}\n\n\tcoda = Subsystem{\n\t\tName: \"coda\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/coda/|^include/linux/coda[^/]*\\\\.h$|^include/uapi/linux/coda[^/]*\\\\.h$\"},\n\t\t},\n\t}\n\n\tcoresight = Subsystem{\n\t\tName: \"coresight\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&arm},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/hwtracing/coresight/|^include/dt-bindings/arm/coresight-cti-dt\\\\.h$|^include/linux/coresight[^/]*$\"},\n\t\t},\n\t}\n\n\tcrypto = Subsystem{\n\t\tName: \"crypto\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/[^/]*/crypto/|^crypto/|^drivers/crypto/|^include/crypto/|^include/linux/crypto[^/]*$|^lib/crypto/\"},\n\t\t\t{IncludeRegexp: \"^crypto/ansi_cprng\\\\.c$|^crypto/rng\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^crypto/pcrypt\\\\.c$|^include/crypto/pcrypt\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/char/hw_random/cctrng\\\\.c$|^drivers/char/hw_random/cctrng\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/char/hw_random/|^include/linux/hw_random\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/allwinner/\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/amlogic/\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/atmel-ecc\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/caam/\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/cavium/cpt/\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/ccp/sev[^/]*$|^include/uapi/linux/psp-sev\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/ccp/|^include/linux/ccp\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/ccree/\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/chelsio$\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/exynos-rng\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/gemini/\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/hisilicon/Kconfig$|^drivers/crypto/hisilicon/qm\\\\.c$|^drivers/crypto/hisilicon/sgl\\\\.c$|^include/linux/hisi_acc_qm\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/hisilicon/hpre/hpre\\\\.h$|^drivers/crypto/hisilicon/hpre/hpre_crypto\\\\.c$|^drivers/crypto/hisilicon/hpre/hpre_main\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/hisilicon/sec2/sec\\\\.h$|^drivers/crypto/hisilicon/sec2/sec_crypto\\\\.c$|^drivers/crypto/hisilicon/sec2/sec_crypto\\\\.h$|^drivers/crypto/hisilicon/sec2/sec_main\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/hisilicon/zip/\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/inside-secure/\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/ixp4xx_crypto\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/marvell/|^include/linux/soc/marvell/octeontx2/\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/nx/Kconfig$|^drivers/crypto/nx/nx-aes[^/]*$|^drivers/crypto/nx/nx-sha[^/]*$|^drivers/crypto/nx/nx\\\\.[^/]*$|^drivers/crypto/nx/nx_csbcpb\\\\.h$|^drivers/crypto/nx/nx_debugfs\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/qce/\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/rockchip/\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/s5p-sss\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/virtio/|^include/uapi/linux/virtio_crypto\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/vmx/Kconfig$|^drivers/crypto/vmx/aes[^/]*$|^drivers/crypto/vmx/ghash[^/]*$|^drivers/crypto/vmx/ppc-xlate\\\\.pl$|^drivers/crypto/vmx/vmx\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/padata\\\\.h$|^kernel/padata\\\\.c$\"},\n\t\t},\n\t}\n\n\tcsky = Subsystem{\n\t\tName: \"csky\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/csky/|^drivers/clocksource/timer-gx6605s\\\\.c$|^drivers/clocksource/timer-mp-csky\\\\.c$|^drivers/irqchip/irq-csky-[^/]*$|csky\"},\n\t\t},\n\t}\n\n\tcxl = Subsystem{\n\t\tName: \"cxl\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/cxl/|^include/uapi/linux/cxl_mem\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/dax/\"},\n\t\t},\n\t}\n\n\tdamon = Subsystem{\n\t\tName: \"damon\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&mm},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^include/linux/damon\\\\.h$|^include/trace/events/damon\\\\.h$|^mm/damon/\"},\n\t\t},\n\t}\n\n\tdccp = Subsystem{\n\t\tName: \"dccp\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^include/linux/dccp\\\\.h$|^include/linux/tfrc\\\\.h$|^include/uapi/linux/dccp\\\\.h$|^net/dccp/\"},\n\t\t},\n\t}\n\n\tdell = Subsystem{\n\t\tName: \"dell\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&x86drivers},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/dell/alienware-wmi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/dell/dell-smbios-smm\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/dell/dell-smbios-wmi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/dell/dell-smbios\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/dell/dell-wmi-descriptor\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/dell/dell-wmi-privacy\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/dell/dell-wmi-sysman/\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/intel/wmi/thunderbolt\\\\.c$\"},\n\t\t},\n\t}\n\n\tdevicetree = Subsystem{\n\t\tName: \"devicetree\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/[^/]*/boot/dts/|^include/dt-bindings/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/[^/]*am3[^/]*$|^arch/arm/boot/dts/[^/]*am4[^/]*$|^arch/arm/boot/dts/[^/]*am5[^/]*$|^arch/arm/boot/dts/[^/]*dra7[^/]*$|^arch/arm/boot/dts/[^/]*omap[^/]*$|^arch/arm/boot/dts/logicpd-som-lv[^/]*$|^arch/arm/boot/dts/logicpd-torpedo[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/of/|^include/linux/of[^/]*\\\\.h$\"},\n\t\t},\n\t}\n\n\tdm = Subsystem{\n\t\tName: \"dm\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/md/Kconfig$|^drivers/md/dm[^/]*$|^drivers/md/persistent-data/|^include/linux/device-mapper\\\\.h$|^include/linux/dm-[^/]*\\\\.h$|^include/uapi/linux/dm-[^/]*\\\\.h$\"},\n\t\t},\n\t}\n\n\tdmaengine = Subsystem{\n\t\tName: \"dmaengine\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/dma/altera-msgdma\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma/at_hdmac\\\\.c$|^drivers/dma/at_xdmac\\\\.c$|^include/dt-bindings/dma/at91\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma/dw-edma/|^include/linux/dma/edma\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma/hisi_dma\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma/idxd/|^include/uapi/linux/idxd\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma/ioat[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma/mediatek/\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma/ptdma/\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma/qcom/hidma[^/]*$\"},\n\t\t\t{\"^drivers/dma/ti/|^include/linux/dma/k3-udma-glue\\\\.h$|^include/linux/dma/ti-cppi5\\\\.h$|^include/linux/dma/k3-psil\\\\.h$\", \"^drivers/dma/ti/cppi41\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma/xilinx/xdma-regs\\\\.h$|^drivers/dma/xilinx/xdma\\\\.c$|^include/linux/dma/amd_xdma\\\\.h$|^include/linux/platform_data/amd_xdma\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma/xilinx/xilinx_dpdma\\\\.c$|^include/dt-bindings/dma/xlnx-zynqmp-dpdma\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma/|^include/dt-bindings/dma/|^include/linux/dma/|^include/linux/dmaengine\\\\.h$|^include/linux/of_dma\\\\.h$\"},\n\t\t},\n\t}\n\n\tdrbd = Subsystem{\n\t\tName: \"drbd\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&block},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/block/drbd/|^include/linux/drbd[^/]*$|^lib/lru_cache\\\\.c$\"},\n\t\t},\n\t}\n\n\tdri = Subsystem{\n\t\tName: \"dri\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/accel/habanalabs/|^include/trace/events/habanalabs\\\\.h$|^include/uapi/drm/habanalabs_accel\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/accel/ivpu/|^include/uapi/drm/ivpu_accel\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/accel/|^include/drm/drm_accel\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/base/component\\\\.c$|^include/linux/component\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/char/agp/|^include/linux/agp[^/]*$|^include/uapi/linux/agp[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma-buf/dma-fence[^/]*$|^drivers/dma-buf/sw_sync\\\\.c$|^drivers/dma-buf/sync_[^/]*$|^include/linux/sync_file\\\\.h$|^include/uapi/linux/sync_file\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma-buf/dma-heap\\\\.c$|^drivers/dma-buf/heaps/|^include/linux/dma-heap\\\\.h$|^include/uapi/linux/dma-heap\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma-buf/udmabuf\\\\.c$|^include/uapi/linux/udmabuf\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma-buf/|^include/linux/[^/]*fence\\\\.h$|^include/linux/dma-buf\\\\.h$|^include/linux/dma-resv\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/ast/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/atmel-hlcdc/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/bridge/imx/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/drm_aperture\\\\.c$|^drivers/gpu/drm/tiny/ofdrm\\\\.c$|^drivers/gpu/drm/tiny/simpledrm\\\\.c$|^drivers/video/aperture\\\\.c$|^drivers/video/nomodeset\\\\.c$|^include/drm/drm_aperture\\\\.h$|^include/linux/aperture\\\\.h$|^include/video/nomodeset\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/drm_panel\\\\.c$|^drivers/gpu/drm/panel/|^include/drm/drm_panel\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/drm_privacy_screen[^/]*$|^include/drm/drm_privacy_screen[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/etnaviv/|^include/uapi/drm/etnaviv_drm\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/exynos/exynos_dp[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/exynos/|^include/uapi/drm/exynos_drm\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/fsl-dcu/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/gma500/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/hisilicon/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/hyperv$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/imx/dcss/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/imx/ipuv3/|^drivers/gpu/ipu-v3/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/lima/|^include/uapi/drm/lima_drm\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/mediatek/|^drivers/phy/mediatek/phy-mtk-dp\\\\.c$|^drivers/phy/mediatek/phy-mtk-hdmi[^/]*$|^drivers/phy/mediatek/phy-mtk-mipi[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/meson/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/mgag200/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/msm/|^include/uapi/drm/msm_drm\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/mxsfb/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/nouveau/|^include/uapi/drm/nouveau_drm\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/omapdrm/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/panel/panel-lvds\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/panfrost/|^include/uapi/drm/panfrost_drm\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/rcar-du/|^drivers/gpu/drm/shmobile/|^include/linux/platform_data/shmob_drm\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/rockchip/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/scheduler/|^include/drm/gpu_scheduler\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/sti$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/stm$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/sun4i/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/sun4i/sun8i[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/tegra/|^drivers/gpu/host1x/|^include/linux/host1x\\\\.h$|^include/uapi/drm/tegra_drm\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/tidss/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/tilcdc/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/ttm/|^include/drm/ttm/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/udl/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/vboxvideo/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/virtio/|^include/uapi/linux/virtio_gpu\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/vkms/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/vmwgfx/|^include/uapi/drm/vmwgfx_drm\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/xen/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/xlnx/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/|^include/drm/|^include/linux/vga[^/]*$|^include/uapi/drm/\"},\n\t\t\t{IncludeRegexp: \"^drivers/staging/fbtft/\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/backlight/|^include/linux/backlight\\\\.h$|^include/linux/pwm_backlight\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/|^include/linux/fb\\\\.h$|^include/uapi/linux/fb\\\\.h$|^include/uapi/video/|^include/video/\"},\n\t\t\t{IncludeRegexp: \"^include/linux/iosys-map\\\\.h$\"},\n\t\t},\n\t}\n\n\tecryptfs = Subsystem{\n\t\tName: \"ecryptfs\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/ecryptfs/\"},\n\t\t},\n\t}\n\n\tedac = Subsystem{\n\t\tName: \"edac\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/x86/kernel/cpu/mce/\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/amd64_edac[^/]*$|^drivers/edac/mce_amd[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/armada_xp_[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/dmc520_edac\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/e752x_edac\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/e7xxx_edac\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/fsl_ddr_edac\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/ghes_edac\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/highbank[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/i10nm_base\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/i3000_edac\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/i5000_edac\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/i5400_edac\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/i7300_edac\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/i7core_edac\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/i82443bxgx_edac\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/i82975x_edac\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/ie31200_edac\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/igen6_edac\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/mpc85xx_edac\\\\.\\\\[ch\\\\]$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/octeon_edac[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/pasemi_edac\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/pnd2_edac\\\\.\\\\[ch\\\\]$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/qcom_edac\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/r82600_edac\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/sb_edac\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/skx_[^/]*\\\\.\\\\[ch\\\\]$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/thunderx_edac[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/ti_edac\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/|^include/linux/edac\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/ras/|^include/linux/ras\\\\.h$|^include/ras/ras_event\\\\.h$\"},\n\t\t},\n\t}\n\n\tefi = Subsystem{\n\t\tName: \"efi\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/[^/]*/include/asm/efi\\\\.h$|^arch/[^/]*/kernel/efi\\\\.c$|^arch/arm/boot/compressed/efi-header\\\\.S$|^arch/x86/platform/efi/|^drivers/firmware/efi/|^include/linux/efi[^/]*\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^block/partitions/efi\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/firmware/efi/test/\"},\n\t\t\t{IncludeRegexp: \"^fs/efivarfs/\"},\n\t\t},\n\t}\n\n\terofs = Subsystem{\n\t\tName: \"erofs\",\n\t\tSyscalls: []string{\"syz_mount_image$erofs\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/erofs/|^include/trace/events/erofs\\\\.h$\"},\n\t\t},\n\t}\n\n\tetnaviv = Subsystem{\n\t\tName: \"etnaviv\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&dri},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/etnaviv/|^include/uapi/drm/etnaviv_drm\\\\.h$\"},\n\t\t},\n\t}\n\n\text4 = Subsystem{\n\t\tName: \"ext4\",\n\t\tSyscalls: []string{\"syz_mount_image$ext4\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/ext2/|^include/linux/ext2[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^fs/ext4/|^include/trace/events/ext4\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^fs/jbd2/|^include/linux/jbd2\\\\.h$\"},\n\t\t},\n\t}\n\n\tf2fs = Subsystem{\n\t\tName: \"f2fs\",\n\t\tSyscalls: []string{\"syz_mount_image$f2fs\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/f2fs/|^include/linux/f2fs_fs\\\\.h$|^include/trace/events/f2fs\\\\.h$|^include/uapi/linux/f2fs\\\\.h$\"},\n\t\t},\n\t}\n\n\tfat = Subsystem{\n\t\tName: \"fat\",\n\t\tSyscalls: []string{\"syz_mount_image$msdos\", \"syz_mount_image$vfat\", \"syz_mount_image$exfat\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/exfat/\"},\n\t\t\t{IncludeRegexp: \"^fs/fat/\"},\n\t\t},\n\t}\n\n\tfbdev = Subsystem{\n\t\tName: \"fbdev\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&dri},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/staging/fbtft/\"},\n\t\t\t{IncludeRegexp: \"^drivers/staging/fbtft/fb_seps525\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/staging/sm750fb/\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/atmel_lcdfb\\\\.c$|^include/video/atmel_lcdc\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/aty/aty128fb\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/aty/radeon[^/]*$|^include/uapi/linux/radeonfb\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/efifb\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/fsl-diu-fb\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/i810/\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/imsttfb\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/imxfb\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/intelfb/\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/matrox/matroxfb_[^/]*$|^include/uapi/linux/matroxfb\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/nvidia/|^drivers/video/fbdev/riva/\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/omap/\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/omap2/\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/s3c-fb\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/savage/\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/simplefb\\\\.c$|^include/linux/platform_data/simplefb\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/sm712[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/smscufx\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/udlfb\\\\.c$|^include/video/udlfb\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/uvesafb\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/via/|^include/linux/via-core\\\\.h$|^include/linux/via-gpio\\\\.h$|^include/linux/via_i2c\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/|^include/linux/fb\\\\.h$|^include/uapi/linux/fb\\\\.h$|^include/uapi/video/|^include/video/\"},\n\t\t},\n\t}\n\n\tfpga = Subsystem{\n\t\tName: \"fpga\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/fpga/dfl[^/]*$|^drivers/uio/uio_dfl\\\\.c$|^include/linux/dfl\\\\.h$|^include/uapi/linux/fpga-dfl\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/fpga/intel-m10-bmc-sec-update\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/fpga/microchip-spi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/fpga/|^include/linux/fpga/\"},\n\t\t},\n\t}\n\n\tfreedreno = Subsystem{\n\t\tName: \"freedreno\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&armmsm, &dri},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/msm/|^include/uapi/drm/msm_drm\\\\.h$\"},\n\t\t},\n\t}\n\n\tfs = Subsystem{\n\t\tName: \"fs\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/affs/\"},\n\t\t\t{IncludeRegexp: \"^fs/dax\\\\.c$|^include/linux/dax\\\\.h$|^include/trace/events/fs_dax\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^fs/exfat/\"},\n\t\t\t{IncludeRegexp: \"^fs/fcntl\\\\.c$|^fs/locks\\\\.c$|^include/linux/fcntl\\\\.h$|^include/uapi/linux/fcntl\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^fs/fuse/virtio_fs\\\\.c$|^include/uapi/linux/virtio_fs\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^fs/fuse/|^include/uapi/linux/fuse\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^fs/hfs/\"},\n\t\t\t{IncludeRegexp: \"^fs/hfsplus/\"},\n\t\t\t{IncludeRegexp: \"^fs/iomap/|^include/linux/iomap\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^fs/isofs/\"},\n\t\t\t{IncludeRegexp: \"^fs/notify/dnotify/|^include/linux/dnotify\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^fs/notify/fanotify/|^include/linux/fanotify\\\\.h$|^include/uapi/linux/fanotify\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^fs/notify/inotify/|^include/linux/inotify\\\\.h$|^include/uapi/linux/inotify\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^fs/notify/|^include/linux/fsnotify[^/]*\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^fs/proc/proc_sysctl\\\\.c$|^include/linux/sysctl\\\\.h$|^kernel/sysctl-test\\\\.c$|^kernel/sysctl\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^fs/proc/|^include/linux/proc_fs\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^fs/unicode/\"},\n\t\t\t{IncludeRegexp: \"^fs/vboxsf/\"},\n\t\t\t{IncludeRegexp: \"^fs/zonefs/\"},\n\t\t\t{IncludeRegexp: \"^fs/|^include/linux/fs\\\\.h$|^include/linux/fs_types\\\\.h$|^include/uapi/linux/fs\\\\.h$|^include/uapi/linux/openat2\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/idr\\\\.h$|^include/linux/xarray\\\\.h$|^lib/idr\\\\.c$|^lib/xarray\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/mnt_idmapping\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/pagemap\\\\.h$|^mm/filemap\\\\.c$|^mm/page-writeback\\\\.c$|^mm/readahead\\\\.c$|^mm/truncate\\\\.c$\"},\n\t\t},\n\t}\n\n\tfscrypt = Subsystem{\n\t\tName: \"fscrypt\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/crypto/|^include/linux/fscrypt\\\\.h$|^include/uapi/linux/fscrypt\\\\.h$\"},\n\t\t},\n\t}\n\n\tfsi = Subsystem{\n\t\tName: \"fsi\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/fsi/|^include/linux/fsi[^/]*\\\\.h$|^include/trace/events/fsi[^/]*\\\\.h$\"},\n\t\t},\n\t}\n\n\tfsverity = Subsystem{\n\t\tName: \"fsverity\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/verity/|^include/linux/fsverity\\\\.h$|^include/uapi/linux/fsverity\\\\.h$\"},\n\t\t},\n\t}\n\n\tfuse = Subsystem{\n\t\tName: \"fuse\",\n\t\tSyscalls: []string{\"syz_fuse_handle_req\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/fuse/|^include/uapi/linux/fuse\\\\.h$\"},\n\t\t},\n\t}\n\n\tgeode = Subsystem{\n\t\tName: \"geode\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fbdev},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/x86/include/asm/geode\\\\.h$|^drivers/char/hw_random/geode-rng\\\\.c$|^drivers/crypto/geode[^/]*$|^drivers/video/fbdev/geode/\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/gadget/udc/amd5536udc\\\\.[^/]*$\"},\n\t\t},\n\t}\n\n\tgfs2 = Subsystem{\n\t\tName: \"gfs2\",\n\t\tSyscalls: []string{\"syz_mount_image$gfs2\", \"syz_mount_image$gfs2meta\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/dlm/\"},\n\t\t\t{IncludeRegexp: \"^fs/gfs2/|^include/uapi/linux/gfs2_ondisk\\\\.h$\"},\n\t\t},\n\t}\n\n\tgpio = Subsystem{\n\t\tName: \"gpio\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-104-dio-48e\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-104-idi-48\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-104-idio-16\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-aggregator\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-altera\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-amd-fch\\\\.c$|^include/linux/platform_data/gpio/gpio-amd-fch\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-davinci\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-dwapb\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-gpio-mm\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-hisi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-i8255\\\\.c$|^drivers/gpio/gpio-i8255\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-ich\\\\.c$|^drivers/gpio/gpio-merrifield\\\\.c$|^drivers/gpio/gpio-ml-ioh\\\\.c$|^drivers/gpio/gpio-pch\\\\.c$|^drivers/gpio/gpio-sch\\\\.c$|^drivers/gpio/gpio-sodaville\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-idio-16\\\\.c$|^drivers/gpio/gpio-idio-16\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-mockup\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-pci-idio-16\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-pcie-idio-24\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-pxa\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-sama5d2-piobu\\\\.c$|^drivers/pinctrl/pinctrl-at91[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-virtio\\\\.c$|^include/uapi/linux/virtio_gpio\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-wcove\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-ws16c48\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-xra1403\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpiolib-acpi\\\\.c$|^drivers/gpio/gpiolib-acpi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/|^include/asm-generic/gpio\\\\.h$|^include/dt-bindings/gpio/|^include/linux/gpio\\\\.h$|^include/linux/gpio/|^include/linux/of_gpio\\\\.h$|^include/uapi/linux/gpio\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/misc/mchp_pci1xxxx/mchp_pci1xxxx_gp\\\\.c$|^drivers/misc/mchp_pci1xxxx/mchp_pci1xxxx_gp\\\\.h$|^drivers/misc/mchp_pci1xxxx/mchp_pci1xxxx_gpio\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pinctrl/aspeed/\"},\n\t\t\t{IncludeRegexp: \"^drivers/pinctrl/bcm/pinctrl-bcm4908\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pinctrl/freescale/\"},\n\t\t\t{IncludeRegexp: \"^drivers/pinctrl/pinctrl-cy8c95x0\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pinctrl/pinctrl-k210\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pinctrl/pinctrl-loongson2\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pinctrl/starfive/pinctrl-starfive-jh71[^/]*$|^include/dt-bindings/pinctrl/pinctrl-starfive-jh7100\\\\.h$|^include/dt-bindings/pinctrl/starfive,jh7110-pinctrl\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pinctrl/|^include/dt-bindings/pinctrl/|^include/linux/pinctrl/\"},\n\t\t},\n\t}\n\n\tgreybus = Subsystem{\n\t\tName: \"greybus\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&staging},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/greybus/|^drivers/staging/greybus/|^include/linux/greybus\\\\.h$|^include/linux/greybus/\"},\n\t\t},\n\t}\n\n\thams = Subsystem{\n\t\tName: \"hams\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/net/hamradio/6pack\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/hamradio/[^/]*scc\\\\.c$|^drivers/net/hamradio/z8530\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/hamradio/baycom[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/hamradio/scc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/hamradio/yam[^/]*$|^include/linux/yam\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^include/net/ax25\\\\.h$|^include/uapi/linux/ax25\\\\.h$|^net/ax25/\"},\n\t\t\t{IncludeRegexp: \"^include/net/netrom\\\\.h$|^include/uapi/linux/netrom\\\\.h$|^net/netrom/\"},\n\t\t\t{IncludeRegexp: \"^include/net/rose\\\\.h$|^include/uapi/linux/rose\\\\.h$|^net/rose/\"},\n\t\t\t{IncludeRegexp: \"^net/ax25/af_ax25\\\\.c$|^net/ax25/ax25_dev\\\\.c$|^net/ax25/ax25_ds_[^/]*$|^net/ax25/ax25_in\\\\.c$|^net/ax25/ax25_out\\\\.c$|^net/ax25/ax25_timer\\\\.c$|^net/ax25/sysctl_net_ax25\\\\.c$\"},\n\t\t},\n\t}\n\n\thardening = Subsystem{\n\t\tName: \"hardening\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/acpi/apei/erst\\\\.c$|^drivers/firmware/efi/efi-pstore\\\\.c$|^fs/pstore/|^include/linux/pstore[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/fortify-string\\\\.h$|^lib/fortify_kunit\\\\.c$|^lib/memcpy_kunit\\\\.c$|^lib/strscpy_kunit\\\\.c$|^lib/test_fortify/\"},\n\t\t\t{IncludeRegexp: \"^include/linux/overflow\\\\.h$|^include/linux/randomize_kstack\\\\.h$|^mm/usercopy\\\\.c$\"},\n\t\t},\n\t}\n\n\thexagon = Subsystem{\n\t\tName: \"hexagon\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/hexagon/\"},\n\t\t},\n\t}\n\n\thfs = Subsystem{\n\t\tName: \"hfs\",\n\t\tSyscalls: []string{\"syz_mount_image$hfs\", \"syz_mount_image$hfsplus\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/hfs/\"},\n\t\t\t{IncludeRegexp: \"^fs/hfsplus/\"},\n\t\t},\n\t}\n\n\thippi = Subsystem{\n\t\tName: \"hippi\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/net/hippi/|^include/linux/hippidevice\\\\.h$|^include/uapi/linux/if_hippi\\\\.h$|^net/802/hippi\\\\.c$\"},\n\t\t},\n\t}\n\n\thwmon = Subsystem{\n\t\tName: \"hwmon\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/abituguru3\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/abituguru\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/adm1025\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/adm1029\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/adm1177\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/adt7475\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/applesmc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/aquacomputer_d5next\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/asc7621\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/asus-ec-sensors\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/asus_atk0110\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/asus_wmi_sensors\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/axi-fan-control\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/coretemp\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/corsair-cpro\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/corsair-psu\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/dme1737\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/emc2103\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/f71805f\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/f75375s\\\\.c$|^include/linux/f75375s\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/fam15h_power\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/ina209\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/ina2xx\\\\.c$|^include/linux/platform_data/ina2xx\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/it87\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/jc42\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/k10temp\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/k8temp\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/lm73\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/lm78\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/lm83\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/lm90\\\\.c$|^include/dt-bindings/thermal/lm90\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/lm95234\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/ltc2947-core\\\\.c$|^drivers/hwmon/ltc2947-i2c\\\\.c$|^drivers/hwmon/ltc2947-spi\\\\.c$|^drivers/hwmon/ltc2947\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/ltc4261\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/max16065\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/max6650\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/max6697\\\\.c$|^include/linux/platform_data/max6697\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/nct6775-core\\\\.c$|^drivers/hwmon/nct6775-platform\\\\.c$|^drivers/hwmon/nct6775\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/nct6775-i2c\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/nzxt-kraken2\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/nzxt-smart2\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/oxp-sensors\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/pc87360\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/pc87427\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/peci/\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/pmbus/delta-ahe50dc-fan\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/pmbus/dps920ab\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/pmbus/max15301\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/pmbus/pm6764tr\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/pmbus/stpddc60\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/pmbus/tps546d24\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/pmbus/|^include/linux/pmbus\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/sch5627\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/smm665\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/smsc47b397\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/tmp401\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/tmp464\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/tmp513\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/tps23861\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/vt1211\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/vt8231\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/w83791d\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/w83793\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/w83795\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwmon/|^include/linux/hwmon[^/]*\\\\.h$|^include/trace/events/hwmon[^/]*\\\\.h$\"},\n\t\t},\n\t}\n\n\thyperv = Subsystem{\n\t\tName: \"hyperv\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/arm64/hyperv$|^arch/arm64/include/asm/hyperv-tlfs\\\\.h$|^arch/arm64/include/asm/mshyperv\\\\.h$|^arch/x86/hyperv$|^arch/x86/include/asm/hyperv-tlfs\\\\.h$|^arch/x86/include/asm/mshyperv\\\\.h$|^arch/x86/include/asm/trace/hyperv\\\\.h$|^arch/x86/kernel/cpu/mshyperv\\\\.c$|^drivers/clocksource/hyperv_timer\\\\.c$|^drivers/hid/hid-hyperv\\\\.c$|^drivers/hv/|^drivers/input/serio/hyperv-keyboard\\\\.c$|^drivers/iommu/hyperv-iommu\\\\.c$|^drivers/net/ethernet/microsoft/|^drivers/net/hyperv/|^drivers/pci/controller/pci-hyperv-intf\\\\.c$|^drivers/pci/controller/pci-hyperv\\\\.c$|^drivers/scsi/storvsc_drv\\\\.c$|^drivers/uio/uio_hv_generic\\\\.c$|^drivers/video/fbdev/hyperv_fb\\\\.c$|^include/asm-generic/hyperv-tlfs\\\\.h$|^include/asm-generic/mshyperv\\\\.h$|^include/clocksource/hyperv_timer\\\\.h$|^include/linux/hyperv\\\\.h$|^include/net/mana$|^include/uapi/linux/hyperv\\\\.h$|^net/vmw_vsock/hyperv_transport\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/hyperv$\"},\n\t\t},\n\t}\n\n\ti2c = Subsystem{\n\t\tName: \"i2c\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/hid/hid-ft260\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/hid-mcp2221\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/algos/|^drivers/i2c/busses/|^include/dt-bindings/i2c/\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-ali1535\\\\.c$|^drivers/i2c/busses/i2c-ali1563\\\\.c$|^drivers/i2c/busses/i2c-ali15x3\\\\.c$|^drivers/i2c/busses/i2c-amd756-s4882\\\\.c$|^drivers/i2c/busses/i2c-amd756\\\\.c$|^drivers/i2c/busses/i2c-amd8111\\\\.c$|^drivers/i2c/busses/i2c-i801\\\\.c$|^drivers/i2c/busses/i2c-isch\\\\.c$|^drivers/i2c/busses/i2c-nforce2-s4985\\\\.c$|^drivers/i2c/busses/i2c-nforce2\\\\.c$|^drivers/i2c/busses/i2c-piix4\\\\.c$|^drivers/i2c/busses/i2c-sis5595\\\\.c$|^drivers/i2c/busses/i2c-sis630\\\\.c$|^drivers/i2c/busses/i2c-sis96x\\\\.c$|^drivers/i2c/busses/i2c-via\\\\.c$|^drivers/i2c/busses/i2c-viapro\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-ali1563\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-amd-mp2[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-aspeed\\\\.c$|^drivers/irqchip/irq-aspeed-i2c-ic\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-at91-[^/]*\\\\.c$|^drivers/i2c/busses/i2c-at91\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-axxia\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-brcmstb\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-cht-wc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-cpm\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-designware-[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-diolan-u2c\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-fsi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-hisi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-icy\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-imx-lpi2c\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-imx\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-ismt\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-ls2x\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-mchp-pci1xxxx\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-mlxbf\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-mlxcpld\\\\.c$|^drivers/i2c/muxes/i2c-mux-mlxcpld\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-mpc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-mt65xx\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-mt7621\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-mv64xxx\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-nvidia-gpu\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-ocores\\\\.c$|^include/linux/platform_data/i2c-ocores\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-omap\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-parport\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-qcom-cci\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-stm32[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-synquacer\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-taos-evm\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-tiny-usb\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-virtio\\\\.c$|^include/uapi/linux/virtio_i2c\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-xlp9xx\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/i2c-core-acpi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/i2c-mux\\\\.c$|^drivers/i2c/muxes/|^include/linux/i2c-mux\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/i2c-stub\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/muxes/i2c-mux-gpio\\\\.c$|^include/linux/platform_data/i2c-mux-gpio\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/muxes/i2c-mux-ltc4306\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/muxes/i2c-mux-pca9541\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/|^include/dt-bindings/i2c/i2c\\\\.h$|^include/linux/i2c-dev\\\\.h$|^include/linux/i2c-smbus\\\\.h$|^include/linux/i2c\\\\.h$|^include/uapi/linux/i2c-[^/]*\\\\.h$|^include/uapi/linux/i2c\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/misc/eeprom/at24\\\\.c$\"},\n\t\t},\n\t}\n\n\ti3c = Subsystem{\n\t\tName: \"i3c\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/i3c/master/svc-i3c-master\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i3c/|^include/linux/i3c/\"},\n\t\t},\n\t}\n\n\tia64 = Subsystem{\n\t\tName: \"ia64\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/ia64/\"},\n\t\t},\n\t}\n\n\tide = Subsystem{\n\t\tName: \"ide\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/ata/ahci_dwc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/ata/ahci_platform\\\\.c$|^drivers/ata/libahci_platform\\\\.c$|^include/linux/ahci_platform\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/ata/ata_[^/]*\\\\.c$|^drivers/ata/pata_[^/]*\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/ata/pata_arasan_cf\\\\.c$|^include/linux/pata_arasan_cf_data\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/ata/pata_ftide010\\\\.c$|^drivers/ata/sata_gemini\\\\.c$|^drivers/ata/sata_gemini\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/ata/sata_promise\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/ata/sata_rcar\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/ata/|^include/linux/ata\\\\.h$|^include/linux/libata\\\\.h$\"},\n\t\t},\n\t}\n\n\tiio = Subsystem{\n\t\tName: \"iio\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/counter/104-quad-8\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/counter/ftm-quaddec\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/counter/intel-qep\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/counter/interrupt-cnt\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/counter/microchip-tcb-capture\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/counter/ti-ecap-capture\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/counter/ti-eqep\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/counter/|^include/linux/counter\\\\.h$|^include/uapi/linux/counter\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/hid-sensor-[^/]*$|^drivers/iio/[^/]*/hid-[^/]*$|^include/linux/hid-sensor-[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/accel/adxl355\\\\.h$|^drivers/iio/accel/adxl355_core\\\\.c$|^drivers/iio/accel/adxl355_i2c\\\\.c$|^drivers/iio/accel/adxl355_spi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/accel/adxl367[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/accel/bma400[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/accel/kionix-kx022a[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/accel/msa311\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/adc/ad4130\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/adc/ad7192\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/adc/ad7292\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/adc/ad7768-1\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/adc/ad7780\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/adc/at91-sama5d2_adc\\\\.c$|^include/dt-bindings/iio/adc/at91-sama5d2_adc\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/adc/envelope-detector\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/adc/hx711\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/adc/imx7d_adc\\\\.c$|^drivers/iio/adc/imx93_adc\\\\.c$|^drivers/iio/adc/vf610_adc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/adc/imx8qxp-adc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/adc/max11205\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/adc/mcp3911\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/adc/rcar-gyroadc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/adc/rzg2l_adc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/adc/ti-ads7924\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/adc/ti-lmp92064\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/adc/ti-tsc2046\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/adc/xilinx-ams\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/addac/ad74115\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/addac/ad74413r\\\\.c$|^include/dt-bindings/iio/addac/adi,ad74413r\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/addac/stx104\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/afe/iio-rescale\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/amplifiers/ada4250\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/common/scmi_sensors/scmi_iio\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/dac/ad3552r\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/dac/ad7293\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/dac/cio-dac\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/dac/dpot-dac\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/dac/ltc1660\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/dac/ltc2688\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/dac/ti-dac7612\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/filter/admv8818\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/frequency/adf4377\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/frequency/admv1013\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/frequency/admv1014\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/frequency/adrf6780\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/gyro/adxrs290\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/gyro/fxas21002c\\\\.h$|^drivers/iio/gyro/fxas21002c_core\\\\.c$|^drivers/iio/gyro/fxas21002c_i2c\\\\.c$|^drivers/iio/gyro/fxas21002c_spi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/gyro/mpu3050[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/humidity/hts221[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/imu/adis16460\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/imu/adis16475\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/imu/adis\\\\.c$|^drivers/iio/imu/adis_buffer\\\\.c$|^drivers/iio/imu/adis_trigger\\\\.c$|^include/linux/iio/imu/adis\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/imu/inv_icm42600/\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/imu/st_lsm6dsx/\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/light/as73211\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/light/gp2ap002\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/magnetometer/ak8974\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/magnetometer/rm3100[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/magnetometer/tmag5273\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/multiplexer/iio-mux\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/potentiometer/ad5110\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/potentiometer/mcp4018\\\\.c$|^drivers/iio/potentiometer/mcp4531\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/pressure/dps310\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/proximity/mb1232\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/proximity/ping\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/proximity/srf[^/]*\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/proximity/vl53l0x-i2c\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/temperature/ltc2983\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/temperature/max30208\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/temperature/mlx90614\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/temperature/mlx90632\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/temperature/tmp117\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/|^drivers/staging/iio/|^include/dt-bindings/iio/|^include/linux/iio/\"},\n\t\t\t{IncludeRegexp: \"^drivers/staging/iio/\"},\n\t\t},\n\t}\n\n\timx = Subsystem{\n\t\tName: \"imx\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&arm, &clk},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/clk/imx/|^include/dt-bindings/clock/imx[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-imx-lpi2c\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/adc/imx7d_adc\\\\.c$|^drivers/iio/adc/imx93_adc\\\\.c$|^drivers/iio/adc/vf610_adc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/adc/imx8qxp-adc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/host/sdhci-esdhc-imx\\\\.c$\"},\n\t\t},\n\t}\n\n\tinput = Subsystem{\n\t\tName: \"input\",\n\t\tSyscalls: []string{\"syz_usb_connect$hid\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/hid/amd-sfh-hid/\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/hid-creative-sb0540\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/hid-ft260\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/hid-letsketch\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/hid-lg-g15\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/hid-logitech-[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/hid-logitech-hidpp\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/hid-mcp2221\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/hid-nintendo[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/hid-picolcd[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/hid-playstation\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/hid-pxrc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/hid-sensor-[^/]*$|^drivers/iio/[^/]*/hid-[^/]*$|^include/linux/hid-sensor-[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/hid-udraw-ps3\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/hid-vrc2\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/hid-wiimote[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/intel-ish-hid/\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/surface-hid/\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/uhid\\\\.c$|^include/uapi/linux/uhid\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/wacom\\\\.h$|^drivers/hid/wacom_[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/|^include/linux/hid[^/]*$|^include/uapi/linux/hid[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/input-mt\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/joystick/fsia6b\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/joystick/pxrc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/keyboard/cypress-sf\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/keyboard/dlink-dir685-touchkeys\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/keyboard/sun4i-lradc-keys\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/misc/ibm-panel\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/misc/ideapad_slidebar\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/mouse/bcm5974\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/mouse/vmmouse\\\\.c$|^drivers/input/mouse/vmmouse\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/tablet/wacom_serial4\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/touchscreen/chipone_icn8318\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/touchscreen/chipone_icn8505\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/touchscreen/cy8ctma140\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/touchscreen/cyttsp[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/touchscreen/goodix[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/touchscreen/himax_hx83112b\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/touchscreen/htcpen\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/touchscreen/hycon-hy46xx\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/touchscreen/resistive-adc-touch\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/touchscreen/silead\\\\.c$|^drivers/platform/x86/touchscreen_dmi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/touchscreen/sis_i2c\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/|^include/dt-bindings/input/|^include/linux/input\\\\.h$|^include/linux/input/|^include/uapi/linux/input-event-codes\\\\.h$|^include/uapi/linux/input\\\\.h$\"},\n\t\t},\n\t}\n\n\tintegrity = Subsystem{\n\t\tName: \"integrity\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/char/tpm/\"},\n\t\t\t{IncludeRegexp: \"^include/keys/encrypted-type\\\\.h$|^security/keys/encrypted-keys/\"},\n\t\t\t{IncludeRegexp: \"^include/keys/trusted-type\\\\.h$|^include/keys/trusted_tpm\\\\.h$|^security/keys/trusted-keys/\"},\n\t\t\t{IncludeRegexp: \"^include/keys/trusted_caam\\\\.h$|^security/keys/trusted-keys/trusted_caam\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/keys/trusted_tee\\\\.h$|^security/keys/trusted-keys/trusted_tee\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^security/integrity/evm/|^security/integrity/\"},\n\t\t\t{IncludeRegexp: \"^security/integrity/ima/|^security/integrity/\"},\n\t\t\t{IncludeRegexp: \"^security/integrity/platform_certs$\"},\n\t\t},\n\t}\n\n\tintelgfx = Subsystem{\n\t\tName: \"intel-gfx\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&dri},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/i915/gvt/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/i915/|^include/drm/i915[^/]*$|^include/uapi/drm/i915_drm\\\\.h$\"},\n\t\t},\n\t}\n\n\tintelgvt = Subsystem{\n\t\tName: \"intel-gvt\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&intelgfx},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/i915/gvt/\"},\n\t\t},\n\t}\n\n\tintelwiredlan = Subsystem{\n\t\tName: \"intel-wired-lan\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/intel/|^drivers/net/ethernet/intel/[^/]*/|^include/linux/avf/virtchnl\\\\.h$|^include/linux/net/intel/iidc\\\\.h$\"},\n\t\t},\n\t}\n\n\tiouring = Subsystem{\n\t\tName: \"io-uring\",\n\t\tSyscalls: []string{\"syz_io_uring_setup\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^io_uring/|^include/linux/io_uring\\\\.h$|^include/linux/io_uring_types\\\\.h$|^include/trace/events/io_uring\\\\.h$|^include/uapi/linux/io_uring\\\\.h$\"},\n\t\t},\n\t}\n\n\tiommu = Subsystem{\n\t\tName: \"iommu\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/[^/]*/include/asm/xen/swiotlb-xen\\\\.h$|^drivers/xen/swiotlb-xen\\\\.c$|^include/xen/arm/swiotlb-xen\\\\.h$|^include/xen/swiotlb-xen\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/acpi/viot\\\\.c$|^include/linux/acpi_viot\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iommu/amd/|^include/linux/amd-iommu\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iommu/arm/arm-smmu/qcom_iommu\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iommu/dma-iommu\\\\.c$|^drivers/iommu/dma-iommu\\\\.h$|^drivers/iommu/iova\\\\.c$|^include/linux/iova\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iommu/exynos-iommu\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iommu/intel/\"},\n\t\t\t{IncludeRegexp: \"^drivers/iommu/iommufd/|^include/linux/iommufd\\\\.h$|^include/uapi/linux/iommufd\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iommu/mtk_iommu[^/]*$|^include/dt-bindings/memory/mt[^/]*-port\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iommu/|^include/linux/iommu\\\\.h$|^include/linux/iova\\\\.h$|^include/linux/of_iommu\\\\.h$|^include/uapi/linux/iommu\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^include/asm-generic/dma-mapping\\\\.h$|^include/linux/dma-direct\\\\.h$|^include/linux/dma-mapping\\\\.h$|^include/linux/dma-map-ops\\\\.h$|^include/linux/swiotlb\\\\.h$|^kernel/dma/\"},\n\t\t\t{IncludeRegexp: \"^kernel/dma/map_benchmark\\\\.c$\"},\n\t\t},\n\t}\n\n\tipack = Subsystem{\n\t\tName: \"ipack\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/ipack/\"},\n\t\t},\n\t}\n\n\tisdn4linux = Subsystem{\n\t\tName: \"isdn4linux\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/isdn/Kconfig$|^drivers/isdn/hardware/|^drivers/isdn/mISDN/\"},\n\t\t\t{IncludeRegexp: \"^drivers/isdn/capi/|^include/linux/isdn/|^include/uapi/linux/isdn/|^net/bluetooth/cmtp/\"},\n\t\t},\n\t}\n\n\tisofs = Subsystem{\n\t\tName: \"isofs\",\n\t\tSyscalls: []string{\"syz_mount_image$iso9660\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/isofs/\"},\n\t\t},\n\t}\n\n\tjfs = Subsystem{\n\t\tName: \"jfs\",\n\t\tSyscalls: []string{\"syz_mount_image$jfs\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/jfs/\"},\n\t\t},\n\t}\n\n\tkarma = Subsystem{\n\t\tName: \"karma\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/omfs/\"},\n\t\t},\n\t}\n\n\tkasan = Subsystem{\n\t\tName: \"kasan\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/[^/]*/include/asm/[^/]*kasan\\\\.h$|^arch/[^/]*/mm/kasan_init[^/]*$|^include/linux/kasan[^/]*\\\\.h$|^lib/Kconfig\\\\.kasan$|^mm/kasan/\"},\n\t\t\t{IncludeRegexp: \"^arch/[^/]*/include/asm/kfence\\\\.h$|^include/linux/kfence\\\\.h$|^lib/Kconfig\\\\.kfence$|^mm/kfence/\"},\n\t\t\t{IncludeRegexp: \"^arch/[^/]*/include/asm/kmsan\\\\.h$|^arch/[^/]*/mm/kmsan_[^/]*$|^include/linux/kmsan[^/]*\\\\.h$|^lib/Kconfig\\\\.kmsan$|^mm/kmsan/\"},\n\t\t\t{IncludeRegexp: \"^include/linux/kcov\\\\.h$|^include/uapi/linux/kcov\\\\.h$|^kernel/kcov\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/kcsan[^/]*\\\\.h$|^kernel/kcsan/|^lib/Kconfig\\\\.kcsan$\"},\n\t\t},\n\t}\n\n\tkernel = Subsystem{\n\t\tName: \"kernel\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^[^/]*$|^[^/]*/\"},\n\t\t\t{IncludeRegexp: \"^arch/[^/]*/events/|^arch/[^/]*/events/[^/]*/|^arch/[^/]*/include/asm/perf_event\\\\.h$|^arch/[^/]*/kernel/[^/]*/[^/]*/perf_event[^/]*\\\\.c$|^arch/[^/]*/kernel/[^/]*/perf_event[^/]*\\\\.c$|^arch/[^/]*/kernel/perf_callchain\\\\.c$|^arch/[^/]*/kernel/perf_event[^/]*\\\\.c$|^include/linux/perf_event\\\\.h$|^include/uapi/linux/perf_event\\\\.h$|^kernel/events/\"},\n\t\t\t{IncludeRegexp: \"^arch/[^/]*/include/asm/atomic[^/]*\\\\.h$|^include/[^/]*/atomic[^/]*\\\\.h$|^include/linux/refcount\\\\.h$\"},\n\t\t\t{\"^arch/[^/]*/include/asm/spinlock[^/]*\\\\.h$|^include/linux/lockdep\\\\.h$|^include/linux/mutex[^/]*\\\\.h$|^include/linux/rwlock[^/]*\\\\.h$|^include/linux/rwsem[^/]*\\\\.h$|^include/linux/seqlock\\\\.h$|^include/linux/spinlock[^/]*\\\\.h$|^kernel/locking/|^lib/locking[^/]*\\\\.\\\\[ch\\\\]$\", \"^kernel/locking/locktorture\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/mt2[^/]*$|^arch/arm/boot/dts/mt6[^/]*$|^arch/arm/boot/dts/mt7[^/]*$|^arch/arm/boot/dts/mt8[^/]*$|^arch/arm/mach-mediatek/|^arch/arm64/boot/dts/mediatek/|^drivers/soc/mediatek/|mtk|mt[2678]\"},\n\t\t\t{IncludeRegexp: \"^arch/powerpc/include/asm/membarrier\\\\.h$|^include/uapi/linux/membarrier\\\\.h$|^kernel/sched/membarrier\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/entry/\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/entry/vdso/\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/include/asm/intel-family\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/include/asm/resctrl\\\\.h$|^arch/x86/kernel/cpu/resctrl/\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/kernel/cpu/hygon\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/kernel/cpu/zhaoxin\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/mm/\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/mm/kmmio\\\\.c$|^arch/x86/mm/mmio-mod\\\\.c$|^arch/x86/mm/testmmiotrace\\\\.c$|^include/linux/mmiotrace\\\\.h$|^kernel/trace/trace_mmiotrace\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/[^/]*/[^/]*max77843\\\\.c$|^drivers/[^/]*/max14577[^/]*\\\\.c$|^drivers/[^/]*/max77686[^/]*\\\\.c$|^drivers/[^/]*/max77693[^/]*\\\\.c$|^drivers/clk/clk-max77686\\\\.c$|^drivers/extcon/extcon-max14577\\\\.c$|^drivers/extcon/extcon-max77693\\\\.c$|^drivers/rtc/rtc-max77686\\\\.c$|^include/linux/mfd/max14577[^/]*\\\\.h$|^include/linux/mfd/max77686[^/]*\\\\.h$|^include/linux/mfd/max77693[^/]*\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/android/\"},\n\t\t\t{IncludeRegexp: \"^drivers/base/arch_topology\\\\.c$|^include/linux/arch_topology\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/base/devcoredump\\\\.c$|^include/linux/devcoredump\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/base/firmware_loader/|^include/linux/firmware\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/base/regmap/|^include/linux/regmap\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/block/zram/\"},\n\t\t\t{IncludeRegexp: \"^drivers/bus/fsl-mc/|^include/uapi/linux/fsl_mc\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/char/xillybus/\"},\n\t\t\t{IncludeRegexp: \"^drivers/clk/clk-s2mps11\\\\.c$|^drivers/mfd/sec[^/]*\\\\.c$|^drivers/regulator/s2m[^/]*\\\\.c$|^drivers/regulator/s5m[^/]*\\\\.c$|^drivers/rtc/rtc-s5m\\\\.c$|^include/linux/mfd/samsung/\"},\n\t\t\t{IncludeRegexp: \"^drivers/clk/keystone/\"},\n\t\t\t{IncludeRegexp: \"^drivers/clocksource/\"},\n\t\t\t{IncludeRegexp: \"^drivers/clocksource/timer-keystone\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/extcon/extcon-ptn5150\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/extcon/|^include/linux/extcon\\\\.h$|^include/linux/extcon/\"},\n\t\t\t{IncludeRegexp: \"^drivers/firmware/stratix10-rsu\\\\.c$|^drivers/firmware/stratix10-svc\\\\.c$|^include/linux/firmware/intel/stratix10-smc\\\\.h$|^include/linux/firmware/intel/stratix10-svc-client\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-bd9571mwv\\\\.c$|^drivers/mfd/bd9571mwv\\\\.c$|^drivers/regulator/bd9571mwv-regulator\\\\.c$|^include/linux/mfd/bd9571mwv\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-max77650\\\\.c$|^drivers/input/misc/max77650-onkey\\\\.c$|^drivers/leds/leds-max77650\\\\.c$|^drivers/mfd/max77650\\\\.c$|^drivers/power/supply/max77650-charger\\\\.c$|^drivers/regulator/max77650-regulator\\\\.c$|^include/linux/mfd/max77650\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwtracing/ptt/\"},\n\t\t\t{IncludeRegexp: \"^drivers/irqchip/\"},\n\t\t\t{IncludeRegexp: \"^drivers/mailbox/arm_mhuv2\\\\.c$|^include/linux/mailbox/arm_mhuv2_message\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mailbox/|^include/linux/mailbox_client\\\\.h$|^include/linux/mailbox_controller\\\\.h$|^include/dt-bindings/mailbox/\"},\n\t\t\t{IncludeRegexp: \"^drivers/memory/[^/]*emif[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/memory/|^include/dt-bindings/memory/|^include/memory/\"},\n\t\t\t{IncludeRegexp: \"^drivers/mfd/at91-usart\\\\.c$|^include/dt-bindings/mfd/at91-usart\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mfd/hi6421-spmi-pmic\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/misc/hisi_hikey_usb\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/misc/mei/|^drivers/watchdog/mei_wdt\\\\.c$|^include/linux/mei_aux\\\\.h$|^include/linux/mei_cl_bus\\\\.h$|^include/uapi/linux/mei\\\\.h$|^include/uapi/linux/uuid\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/misc/uacce/|^include/linux/uacce\\\\.h$|^include/uapi/misc/uacce/\"},\n\t\t\t{IncludeRegexp: \"^drivers/misc/vmw_balloon\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/misc/vmw_vmci/|^include/linux/vmw_vmci[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/dsa/sja1105$|^drivers/net/pcs/pcs-xpcs-nxp\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/phy/broadcom/phy-brcm-usb[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/phy/samsung/phy-exynos4210-usb2\\\\.c$|^drivers/phy/samsung/phy-exynos4x12-usb2\\\\.c$|^drivers/phy/samsung/phy-exynos5250-usb2\\\\.c$|^drivers/phy/samsung/phy-s5pv210-usb2\\\\.c$|^drivers/phy/samsung/phy-samsung-usb2\\\\.c$|^drivers/phy/samsung/phy-samsung-usb2\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/phy/xilinx/phy-zynqmp\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/power/reset/keystone-reset\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/regulator/max20086-regulator\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/regulator/max77802-regulator\\\\.c$|^include/dt-bindings/[^/]*/[^/]*max77802\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/regulator/|^include/dt-bindings/regulator/|^include/linux/regulator/\"},\n\t\t\t{IncludeRegexp: \"^drivers/reset/reset-k210\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/soc/fsl/dpio$\"},\n\t\t\t{IncludeRegexp: \"^drivers/soc/ti/\"},\n\t\t\t{IncludeRegexp: \"^drivers/spmi/hisi-spmi-controller\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/spmi/|^include/dt-bindings/spmi/spmi\\\\.h$|^include/linux/spmi\\\\.h$|^include/trace/events/spmi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/staging/vme_user/\"},\n\t\t\t{IncludeRegexp: \"^drivers/virt/nitro_enclaves/|^include/linux/nitro_enclaves\\\\.h$|^include/uapi/linux/nitro_enclaves\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^fs/proc/bootconfig\\\\.c$|^include/linux/bootconfig\\\\.h$|^lib/bootconfig-data\\\\.S$|^lib/bootconfig\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^fs/proc/proc_sysctl\\\\.c$|^include/linux/sysctl\\\\.h$|^kernel/sysctl-test\\\\.c$|^kernel/sysctl\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^fs/proc/|^include/linux/proc_fs\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^fs/timerfd\\\\.c$|^include/linux/time_namespace\\\\.h$|^include/linux/timer[^/]*$|^kernel/time/[^/]*timer[^/]*$|^kernel/time/namespace\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^fs/tracefs/|^include/linux/trace[^/]*\\\\.h$|^include/trace/|^kernel/trace/\"},\n\t\t\t{IncludeRegexp: \"^include/asm-generic/futex\\\\.h$|^include/linux/futex\\\\.h$|^include/uapi/linux/futex\\\\.h$|^kernel/futex/\"},\n\t\t\t{IncludeRegexp: \"^include/asm-generic/kprobes\\\\.h$|^include/linux/kprobes\\\\.h$|^kernel/kprobes\\\\.c$|^lib/test_kprobes\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/asm-generic/vdso/vsyscall\\\\.h$|^include/vdso/|^kernel/time/vsyscall\\\\.c$|^lib/vdso/\"},\n\t\t\t{IncludeRegexp: \"^include/linux/clockchips\\\\.h$|^include/linux/hrtimer\\\\.h$|^kernel/time/clockevents\\\\.c$|^kernel/time/hrtimer\\\\.c$|^kernel/time/timer_[^/]*\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/clocksource\\\\.h$|^include/linux/time\\\\.h$|^include/linux/timex\\\\.h$|^include/uapi/linux/time\\\\.h$|^include/uapi/linux/timex\\\\.h$|^kernel/time/alarmtimer\\\\.c$|^kernel/time/clocksource\\\\.c$|^kernel/time/ntp\\\\.c$|^kernel/time/time[^/]*\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/entry-common\\\\.h$|^include/linux/entry-kvm\\\\.h$|^kernel/entry/\"},\n\t\t\t{IncludeRegexp: \"^include/linux/frontswap\\\\.h$|^mm/frontswap\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/kmod\\\\.h$|^kernel/kmod\\\\.c$|^lib/test_kmod\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/module\\\\.h$|^kernel/module/\"},\n\t\t\t{IncludeRegexp: \"^include/linux/padata\\\\.h$|^kernel/padata\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/preempt\\\\.h$|^include/linux/sched\\\\.h$|^include/linux/wait\\\\.h$|^include/uapi/linux/sched\\\\.h$|^kernel/sched/\"},\n\t\t\t{IncludeRegexp: \"^include/linux/sched/nohz\\\\.h$|^include/linux/tick\\\\.h$|^kernel/time/tick[^/]*\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/umh\\\\.h$|^kernel/umh\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/uuid\\\\.h$|^lib/test_uuid\\\\.c$|^lib/uuid\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/trace/events/rseq\\\\.h$|^include/uapi/linux/rseq\\\\.h$|^kernel/rseq\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^kernel/irq/|^include/linux/group_cpus\\\\.h$|^lib/group_cpus\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^kernel/locking/locktorture\\\\.c$|^kernel/rcu/rcuscale\\\\.c$|^kernel/rcu/rcutorture\\\\.c$|^kernel/rcu/refscale\\\\.c$|^kernel/torture\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^kernel/trace/ftrace[^/]*$|^kernel/trace/fgraph\\\\.c$|^arch/[^/]*/[^/]*/[^/]*/[^/]*ftrace[^/]*$|^arch/[^/]*/[^/]*/[^/]*ftrace[^/]*$|^include/[^/]*/ftrace\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^net/vmw_vsock/vmci_transport[^/]*$\"},\n\t\t\t{IncludeRegexp: \"axp[128]\"},\n\t\t},\n\t\tNoReminders: true,\n\t}\n\n\tkernfs = Subsystem{\n\t\tName: \"kernfs\",\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/kernfs/|^include/linux/kernfs\\\\.h$\"},\n\t\t},\n\t}\n\n\tkexec = Subsystem{\n\t\tName: \"kexec\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/proc/vmcore\\\\.c$|^include/linux/crash_core\\\\.h$|^include/linux/crash_dump\\\\.h$|^include/uapi/linux/vmcore\\\\.h$|^kernel/crash_[^/]*\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/kexec\\\\.h$|^include/uapi/linux/kexec\\\\.h$|^kernel/kexec[^/]*$\"},\n\t\t},\n\t}\n\n\tkeyrings = Subsystem{\n\t\tName: \"keyrings\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^certs/\"},\n\t\t\t{IncludeRegexp: \"^crypto/asymmetric_keys/|^include/crypto/pkcs7\\\\.h$|^include/crypto/public_key\\\\.h$|^include/linux/verification\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^include/keys/encrypted-type\\\\.h$|^security/keys/encrypted-keys/\"},\n\t\t\t{IncludeRegexp: \"^include/keys/trusted-type\\\\.h$|^include/keys/trusted_tpm\\\\.h$|^security/keys/trusted-keys/\"},\n\t\t\t{IncludeRegexp: \"^include/keys/trusted_caam\\\\.h$|^security/keys/trusted-keys/trusted_caam\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/keys/trusted_tee\\\\.h$|^security/keys/trusted-keys/trusted_tee\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/keys/|^include/linux/key-type\\\\.h$|^include/linux/key\\\\.h$|^include/linux/keyctl\\\\.h$|^include/uapi/linux/keyctl\\\\.h$|^security/keys/\"},\n\t\t\t{IncludeRegexp: \"^security/integrity/platform_certs$\"},\n\t\t},\n\t}\n\n\tkgdb = Subsystem{\n\t\tName: \"kgdb\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/misc/kgdbts\\\\.c$|^drivers/tty/serial/kgdboc\\\\.c$|^include/linux/kdb\\\\.h$|^include/linux/kgdb\\\\.h$|^kernel/debug/|^kernel/module/kdb\\\\.c$\"},\n\t\t},\n\t}\n\n\tkunit = Subsystem{\n\t\tName: \"kunit\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^include/kunit/|^lib/kunit/\"},\n\t\t\t{IncludeRegexp: \"^lib/list-test\\\\.c$\"},\n\t\t},\n\t}\n\n\tkvm = Subsystem{\n\t\tName: \"kvm\",\n\t\tSyscalls: []string{\"syz_kvm_setup_cpu\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/mips/include/asm/kvm[^/]*$|^arch/mips/include/uapi/asm/kvm[^/]*$|^arch/mips/kvm/\"},\n\t\t\t{IncludeRegexp: \"^arch/riscv/include/asm/kvm[^/]*$|^arch/riscv/include/uapi/asm/kvm[^/]*$|^arch/riscv/kvm/\"},\n\t\t\t{IncludeRegexp: \"^arch/s390/include/asm/gmap\\\\.h$|^arch/s390/include/asm/kvm[^/]*$|^arch/s390/include/uapi/asm/kvm[^/]*$|^arch/s390/include/uapi/asm/uvdevice\\\\.h$|^arch/s390/kernel/uv\\\\.c$|^arch/s390/kvm/|^arch/s390/mm/gmap\\\\.c$|^drivers/s390/char/uvdevice\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/s390/include/uapi/asm/virtio-ccw\\\\.h$|^drivers/s390/virtio/\"},\n\t\t\t{IncludeRegexp: \"^arch/s390/kvm/pci[^/]*$|^drivers/vfio/pci/vfio_pci_zdev\\\\.c$|^include/uapi/linux/vfio_zdev\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/include/asm/kvm[^/]*$|^arch/x86/include/asm/svm\\\\.h$|^arch/x86/include/asm/vmx[^/]*\\\\.h$|^arch/x86/include/uapi/asm/kvm[^/]*$|^arch/x86/include/uapi/asm/svm\\\\.h$|^arch/x86/include/uapi/asm/vmx\\\\.h$|^arch/x86/kvm/|^arch/x86/kvm/[^/]*/\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/kernel/kvm\\\\.c$|^arch/x86/kernel/kvmclock\\\\.c$|^arch/x86/include/asm/pvclock-abi\\\\.h$|^include/linux/kvm_para\\\\.h$|^include/uapi/linux/kvm_para\\\\.h$|^include/uapi/asm-generic/kvm_para\\\\.h$|^include/asm-generic/kvm_para\\\\.h$|^arch/um/include/asm/kvm_para\\\\.h$|^arch/x86/include/asm/kvm_para\\\\.h$|^arch/x86/include/uapi/asm/kvm_para\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/kvm/hyperv\\\\.[^/]*$|^arch/x86/kvm/kvm_onhyperv\\\\.[^/]*$|^arch/x86/kvm/svm/hyperv\\\\.[^/]*$|^arch/x86/kvm/svm/svm_onhyperv\\\\.[^/]*$|^arch/x86/kvm/vmx/hyperv\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/kvm/xen\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/s390/cio/vfio_ccw[^/]*$|^include/uapi/linux/vfio_ccw\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/uio/uio_pci_generic\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/vfio/fsl-mc/\"},\n\t\t\t{IncludeRegexp: \"^drivers/vfio/mdev/|^include/linux/mdev\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/vfio/pci/[^/]*/\"},\n\t\t\t{IncludeRegexp: \"^drivers/vfio/pci/hisilicon/\"},\n\t\t\t{IncludeRegexp: \"^drivers/vfio/pci/mlx5/\"},\n\t\t\t{IncludeRegexp: \"^drivers/vfio/platform/\"},\n\t\t\t{IncludeRegexp: \"^drivers/vfio/|^include/linux/vfio\\\\.h$|^include/linux/vfio_pci_core\\\\.h$|^include/uapi/linux/vfio\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/vhost/vsock\\\\.c$|^include/linux/virtio_vsock\\\\.h$|^include/uapi/linux/virtio_vsock\\\\.h$|^net/vmw_vsock/virtio_transport\\\\.c$|^net/vmw_vsock/virtio_transport_common\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/vhost/|^include/linux/vhost_iotlb\\\\.h$|^include/uapi/linux/vhost\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^include/asm-generic/kvm[^/]*$|^include/kvm/iodev\\\\.h$|^include/linux/kvm[^/]*$|^include/trace/events/kvm\\\\.h$|^include/uapi/asm-generic/kvm[^/]*$|^include/uapi/linux/kvm[^/]*$|^virt/kvm/\"},\n\t\t\t{IncludeRegexp: \"^virt/lib/\"},\n\t\t},\n\t}\n\n\tkvmriscv = Subsystem{\n\t\tName: \"kvm-riscv\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kvm, &riscv},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/riscv/include/asm/kvm[^/]*$|^arch/riscv/include/uapi/asm/kvm[^/]*$|^arch/riscv/kvm/\"},\n\t\t},\n\t}\n\n\tkvmarm = Subsystem{\n\t\tName: \"kvmarm\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&arm},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/arm64/include/asm/kvm[^/]*$|^arch/arm64/include/uapi/asm/kvm[^/]*$|^arch/arm64/kvm/|^include/kvm/arm_[^/]*$\"},\n\t\t},\n\t}\n\n\tleds = Subsystem{\n\t\tName: \"leds\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/leds/flash/leds-as3645a\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/leds/leds-mlxcpld\\\\.c$|^drivers/leds/leds-mlxreg\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/leds/|^include/dt-bindings/leds/|^include/linux/leds\\\\.h$\"},\n\t\t},\n\t}\n\n\tlibertas = Subsystem{\n\t\tName: \"libertas\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&wireless},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/marvell/libertas/\"},\n\t\t},\n\t}\n\n\tlima = Subsystem{\n\t\tName: \"lima\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&dri},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/lima/|^include/uapi/drm/lima_drm\\\\.h$\"},\n\t\t},\n\t}\n\n\tlinux1394 = Subsystem{\n\t\tName: \"linux1394\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/firewire/|^include/linux/firewire\\\\.h$|^include/uapi/linux/firewire[^/]*\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/firewire/\"},\n\t\t\t{IncludeRegexp: \"^drivers/target/sbp/\"},\n\t\t},\n\t}\n\n\tlinuxppc = Subsystem{\n\t\tName: \"linuxppc\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/powerpc/boot/ps3[^/]*$|^arch/powerpc/include/asm/lv1call\\\\.h$|^arch/powerpc/include/asm/ps3[^/]*\\\\.h$|^arch/powerpc/platforms/ps3/|^drivers/[^/]*/ps3[^/]*$|^drivers/ps3/|^drivers/rtc/rtc-ps3\\\\.c$|^drivers/usb/host/[^/]*ps3\\\\.c$|^sound/ppc/snd_ps3[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/powerpc/include/[^/]*/eeh[^/]*\\\\.h$|^arch/powerpc/kernel/eeh[^/]*\\\\.c$|^arch/powerpc/platforms/[^/]*/eeh[^/]*\\\\.c$|^drivers/pci/pcie/aer\\\\.c$|^drivers/pci/pcie/dpc\\\\.c$|^drivers/pci/pcie/err\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/powerpc/include/asm/cell[^/]*\\\\.h$|^arch/powerpc/include/asm/spu[^/]*\\\\.h$|^arch/powerpc/include/uapi/asm/spu[^/]*\\\\.h$|^arch/powerpc/platforms/cell/\"},\n\t\t\t{IncludeRegexp: \"^arch/powerpc/include/asm/kvm[^/]*$|^arch/powerpc/include/uapi/asm/kvm[^/]*$|^arch/powerpc/kernel/kvm[^/]*$|^arch/powerpc/kvm/\"},\n\t\t\t{IncludeRegexp: \"^arch/powerpc/include/asm/pnv-ocxl\\\\.h$|^arch/powerpc/platforms/powernv/ocxl\\\\.c$|^drivers/misc/ocxl/|^include/misc/ocxl[^/]*$|^include/uapi/misc/ocxl\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^arch/powerpc/platforms/40x/|^arch/powerpc/platforms/44x/\"},\n\t\t\t{IncludeRegexp: \"^arch/powerpc/platforms/512x/|^arch/powerpc/platforms/52xx/\"},\n\t\t\t{IncludeRegexp: \"^arch/powerpc/platforms/83xx/|^arch/powerpc/platforms/85xx/\"},\n\t\t\t{IncludeRegexp: \"^arch/powerpc/platforms/8xx/\"},\n\t\t\t{IncludeRegexp: \"^arch/powerpc/platforms/cell/spufs/\"},\n\t\t\t{IncludeRegexp: \"^arch/powerpc/platforms/powermac/|^drivers/macintosh/\"},\n\t\t\t{IncludeRegexp: \"^arch/powerpc/platforms/powernv/pci-cxl\\\\.c$|^drivers/misc/cxl/|^include/misc/cxl[^/]*$|^include/uapi/misc/cxl\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^arch/powerpc/|^drivers/[^/]*/[^/]*/[^/]*pasemi[^/]*$|^drivers/[^/]*/[^/]*pasemi[^/]*$|^drivers/char/tpm/tpm_ibmvtpm[^/]*$|^drivers/crypto/nx/|^drivers/crypto/vmx/|^drivers/i2c/busses/i2c-opal\\\\.c$|^drivers/net/ethernet/ibm/ibmveth\\\\.[^/]*$|^drivers/net/ethernet/ibm/ibmvnic\\\\.[^/]*$|^drivers/pci/hotplug/pnv_php\\\\.c$|^drivers/pci/hotplug/rpa[^/]*$|^drivers/rtc/rtc-opal\\\\.c$|^drivers/scsi/ibmvscsi/|^drivers/tty/hvc/hvc_opal\\\\.c$|^drivers/watchdog/wdrtas\\\\.c$|/pmac|powermac|powernv|[^a-z0-9]ps3|pseries\"},\n\t\t\t{IncludeRegexp: \"^drivers/block/ps3vram\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/char/powernv-op-panel\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma/fsldma\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-cpm\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/freescale/fs_enet/|^include/linux/fs_enet_pd\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/freescale/ucc_geth[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/toshiba/ps3_gelic_net\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/toshiba/spider_net[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wan/fsl_ucc_hdlc[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/[^/]*layerscape[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/hotplug/rpadlpar[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/hotplug/rpaphp[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/soc/fsl/qe/|^include/soc/fsl/qe/\"},\n\t\t\t{IncludeRegexp: \"^drivers/soc/fsl/|^include/linux/fsl/|^include/soc/fsl/\"},\n\t\t\t{IncludeRegexp: \"^drivers/tty/ehv_bytechan\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/tty/hvc/\"},\n\t\t\t{IncludeRegexp: \"^drivers/tty/serial/ucc_uart\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/gadget/udc/fsl[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/phy/phy-fsl-usb[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^sound/aoa/\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/fsl/fsl[^/]*$|^sound/soc/fsl/imx[^/]*$|^sound/soc/fsl/mpc8610_hpcd\\\\.c$\"},\n\t\t},\n\t}\n\n\tlinuxpps = Subsystem{\n\t\tName: \"linuxpps\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/pps/|^include/linux/pps[^/]*\\\\.h$|^include/uapi/linux/pps\\\\.h$\"},\n\t\t},\n\t}\n\n\tlivepatching = Subsystem{\n\t\tName: \"live-patching\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/powerpc/include/asm/livepatch\\\\.h$|^include/linux/livepatch\\\\.h$|^kernel/livepatch/|^kernel/module/livepatch\\\\.c$|^lib/livepatch/\"},\n\t\t},\n\t}\n\n\tllvm = Subsystem{\n\t\tName: \"llvm\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^include/linux/cfi\\\\.h$|^kernel/cfi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/compiler-clang\\\\.h$\"},\n\t\t},\n\t}\n\n\tloongarch = Subsystem{\n\t\tName: \"loongarch\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/loongarch/|^drivers/[^/]*/[^/]*loongarch[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/soc/loongson/loongson2_guts\\\\.c$\"},\n\t\t},\n\t}\n\n\tlsm = Subsystem{\n\t\tName: \"lsm\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^include/linux/capability\\\\.h$|^include/uapi/linux/capability\\\\.h$|^kernel/capability\\\\.c$|^security/commoncap\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/net/calipso\\\\.h$|^include/net/cipso_ipv4\\\\.h$|^include/net/netlabel\\\\.h$|^include/uapi/linux/netfilter/xt_CONNSECMARK\\\\.h$|^include/uapi/linux/netfilter/xt_SECMARK\\\\.h$|^net/ipv4/cipso_ipv4\\\\.c$|^net/ipv6/calipso\\\\.c$|^net/netfilter/xt_CONNSECMARK\\\\.c$|^net/netfilter/xt_SECMARK\\\\.c$|^net/netlabel/\"},\n\t\t\t{IncludeRegexp: \"^include/uapi/linux/landlock\\\\.h$|^security/landlock/\"},\n\t\t\t{\"^security/\", \"^security/selinux/\"},\n\t\t\t{IncludeRegexp: \"^security/smack/\"},\n\t\t},\n\t}\n\n\tlvs = Subsystem{\n\t\tName: \"lvs\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&netfilter},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^include/net/ip_vs\\\\.h$|^include/uapi/linux/ip_vs\\\\.h$|^net/netfilter/ipvs/\"},\n\t\t},\n\t}\n\n\tm68k = Subsystem{\n\t\tName: \"m68k\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/[^/]*/include/asm/nubus\\\\.h$|^drivers/nubus/|^include/linux/nubus\\\\.h$|^include/uapi/linux/nubus\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^arch/m68k/[^/]*/[^/]*_no\\\\.[^/]*$|^arch/m68k/68[^/]*/|^arch/m68k/coldfire/|^arch/m68k/include/asm/[^/]*_no\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/m68k/mac/|^drivers/macintosh/adb-iop\\\\.c$|^drivers/macintosh/via-macii\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/m68k/|^drivers/zorro/\"},\n\t\t},\n\t}\n\n\tmalidp = Subsystem{\n\t\tName: \"malidp\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&dri},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/arm/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/arm/display/include/|^drivers/gpu/drm/arm/display/komeda/\"},\n\t\t},\n\t}\n\n\tmedia = Subsystem{\n\t\tName: \"media\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/dma-buf/dma-fence[^/]*$|^drivers/dma-buf/sw_sync\\\\.c$|^drivers/dma-buf/sync_[^/]*$|^include/linux/sync_file\\\\.h$|^include/uapi/linux/sync_file\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma-buf/dma-heap\\\\.c$|^drivers/dma-buf/heaps/|^include/linux/dma-heap\\\\.h$|^include/uapi/linux/dma-heap\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma-buf/|^include/linux/[^/]*fence\\\\.h$|^include/linux/dma-buf\\\\.h$|^include/linux/dma-resv\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/cec/i2c/ch7322\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/cec/platform/cec-gpio/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/cec/platform/meson/ao-cec-g12a\\\\.c$|^drivers/media/cec/platform/meson/ao-cec\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/cec/platform/s5p/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/cec/platform/tegra/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/cec/usb/pulse8/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/cec/usb/rainshadow/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/cec/|^drivers/media/rc/keymaps/rc-cec\\\\.c$|^include/media/cec-notifier\\\\.h$|^include/media/cec\\\\.h$|^include/uapi/linux/cec-funcs\\\\.h$|^include/uapi/linux/cec\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/common/cx2341x[^/]*$|^include/media/drv-intf/cx2341x\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/common/cypress_firmware[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/common/saa7146/|^drivers/media/pci/saa7146/|^include/media/drv-intf/saa7146[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/common/siano/|^drivers/media/mmc/siano/|^drivers/media/usb/siano/|^drivers/media/usb/siano/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/common/videobuf2/|^include/media/videobuf2-[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/a8293[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/af9013[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/af9033[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/ascot2e[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/cx24120[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/cxd2099[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/cxd2820r[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/cxd2841er[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/cxd2880/|^drivers/media/spi/cxd2880[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/ec100[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/helene[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/horus3a[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/ix2505v[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/lg2160\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/lgdt3305\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/lnbh25[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/m88ds3103[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/m88rs2000[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/mn88472[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/mn88473[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/mxl5xx[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/rtl2830[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/rtl2832[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/rtl2832_sdr[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/si2165[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/si2168[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/sp2[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/stv0910[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/stv6111[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/tc90522[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/tda10071[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/dvb-frontends/zd1301_demod[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/firewire/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ad5820\\\\.c$|^drivers/media/i2c/et8ek8$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ad9389b[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/adp1653\\\\.c$|^include/media/i2c/adp1653\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/adv7180\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/adv748x/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/adv7511[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/adv7604[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/adv7842[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ak7375\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/aptina-pll\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ar0521\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ccs-pll\\\\.c$|^drivers/media/i2c/ccs-pll\\\\.h$|^drivers/media/i2c/ccs/|^include/uapi/linux/ccs\\\\.h$|^include/uapi/linux/smiapp\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/cs3308\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/dw9714\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/dw9768\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/dw9807-vcm\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/hi556\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/hi846\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/hi847\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/imx208\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/imx214\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/imx219\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/imx258\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/imx274\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/imx290\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/imx296\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/imx319\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/imx334\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/imx335\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/imx355\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/imx412\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/imx415\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/isl7998x\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/m5mols/|^include/media/i2c/m5mols\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/max2175[^/]*$|^include/uapi/linux/max2175\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/max9271\\\\.c$|^drivers/media/i2c/max9271\\\\.h$|^drivers/media/i2c/rdacm20\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/max9271\\\\.c$|^drivers/media/i2c/max9271\\\\.h$|^drivers/media/i2c/rdacm21\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/max9286\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/mt9m032\\\\.c$|^include/media/i2c/mt9m032\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/mt9p031\\\\.c$|^include/media/i2c/mt9p031\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/mt9t001\\\\.c$|^include/media/i2c/mt9t001\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/mt9t112\\\\.c$|^include/media/i2c/mt9t112\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/mt9v032\\\\.c$|^include/media/i2c/mt9v032\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/mt9v111\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/og01a1b\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov02a10\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov08d10\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov08x40\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov13858\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov13b10\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov2659\\\\.c$|^include/media/i2c/ov2659\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov2680\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov2685\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov2740\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov5640\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov5647\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov5647\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov5670\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov5675\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov5693\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov5695\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov7670\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov772x\\\\.c$|^include/media/i2c/ov772x\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov7740\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov8856\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov8858\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov9282\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov9640\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov9650\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/ov9734\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/rj54n1cb0c\\\\.c$|^include/media/i2c/rj54n1cb0c\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/s5c73m3/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/s5k5baf\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/saa6588[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/st-mipid02\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/st-vgxy61\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/tc358743[^/]*$|^include/media/i2c/tc358743\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/tda1997x\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/tda9840[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/tea6415c[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/tea6420[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/tw9910\\\\.c$|^include/media/i2c/tw9910\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/i2c/video-i2c\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/mc/|^include/media/media-[^/]*\\\\.h$|^include/uapi/linux/media\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/pci/bt8xx/bttv[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/pci/cobalt/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/pci/cx18/|^include/uapi/linux/ivtv[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/pci/cx88/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/pci/ddbridge/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/pci/dt3155/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/pci/intel/ipu3/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/pci/ivtv/|^include/uapi/linux/ivtv[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/pci/netup_unidvb/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/pci/pt1/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/pci/pt3/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/pci/saa7134/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/pci/solo6x10/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/pci/tw5864/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/pci/tw68/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/pci/tw686x/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/pci/zoran/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/allegro-dvt/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/amlogic/meson-ge2d/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/amphion/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/aspeed/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/atmel/atmel-isi\\\\.c$|^drivers/media/platform/atmel/atmel-isi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/cadence/cdns-csi2[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/chips-media/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/marvell/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/microchip/microchip-csi2dc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/nvidia/tegra-vde/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/nxp/dw100/|^include/uapi/linux/dw100\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/nxp/imx-jpeg$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/nxp/imx-mipi-csis\\\\.c$|^drivers/media/platform/nxp/imx7-media-csi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/nxp/imx-pxp\\\\.\\\\[ch\\\\]$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/qcom/camss/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/qcom/venus/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/renesas/rcar-fcp\\\\.c$|^include/media/rcar-fcp\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/renesas/rcar-isp\\\\.c$|^drivers/media/platform/renesas/rcar-vin/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/renesas/rcar_drif\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/renesas/rcar_fdp1\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/renesas/rcar_jpu\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/renesas/renesas-ceu\\\\.c$|^include/media/drv-intf/renesas-ceu\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/renesas/sh_vou\\\\.c$|^include/media/drv-intf/sh_vou\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/renesas/vsp1/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/rockchip/rga/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/rockchip/rkisp1$|^include/uapi/linux/rkisp1-config\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/samsung/exynos4-is/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/samsung/s3c-camif/|^include/media/drv-intf/s3c_camif\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/samsung/s5p-g2d/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/samsung/s5p-jpeg/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/samsung/s5p-mfc/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/st/sti/bdisp$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/st/sti/delta$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/st/sti/hva$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/st/stm32/stm32-dcmi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/sunxi/sun4i-csi/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/sunxi/sun6i-csi/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/sunxi/sun6i-mipi-csi2/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/sunxi/sun8i-di/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/sunxi/sun8i-rotate/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/ti/am437x/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/ti/cal/|^drivers/media/platform/ti/vpe/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/ti/davinci/|^include/media/davinci/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/ti/omap3isp/|^drivers/staging/media/omap4iss/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/verisilicon/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/video-mux\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/xilinx/|^include/uapi/linux/xilinx-v4l2-controls\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/radio/dsbr100\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/radio/radio-aimslab[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/radio/radio-aztech[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/radio/radio-cadet[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/radio/radio-gemtek[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/radio/radio-isa[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/radio/radio-keene[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/radio/radio-ma901\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/radio/radio-maxiradio[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/radio/radio-miropcm20[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/radio/radio-mr800\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/radio/radio-raremono\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/radio/radio-shark2\\\\.c$|^drivers/media/radio/radio-tea5777\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/radio/radio-shark\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/radio/si470x/radio-si470x-common\\\\.c$|^drivers/media/radio/si470x/radio-si470x-usb\\\\.c$|^drivers/media/radio/si470x/radio-si470x\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/radio/si470x/radio-si470x-i2c\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/radio/si4713/radio-platform-si4713\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/radio/si4713/radio-usb-si4713\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/radio/si4713/si4713\\\\..$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/rc/gpio-ir-tx\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/rc/igorplugusb\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/rc/iguanair\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/rc/imon\\\\.c$|^drivers/media/rc/imon_raw\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/rc/pwm-ir-tx\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/rc/serial_ir\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/rc/ttusbir\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/rc/|^include/media/rc-map\\\\.h$|^include/media/rc-core\\\\.h$|^include/uapi/linux/lirc\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/spi/gs1662\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/test-drivers/vicodec/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/test-drivers/vidtv/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/test-drivers/vimc/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/test-drivers/visl$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/test-drivers/vivid/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/tuners/e4000[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/tuners/fc0011\\\\.c$|^drivers/media/tuners/fc0011\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/tuners/fc2580[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/tuners/it913x[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/tuners/msi001[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/tuners/mxl301rf[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/tuners/mxl5007t\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/tuners/qm1d1b0004[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/tuners/qm1d1c0042[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/tuners/qt1010[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/tuners/si2157[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/tuners/tda18212[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/tuners/tda18218[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/tuners/tda18250[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/tuners/tda18271[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/tuners/tda8290\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/tuners/tda8290\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/tuners/tea5761\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/tuners/tea5767\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/tuners/tua9001[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/tuners/xc2028\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/airspy/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/dvb-usb-v2/af9015[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/dvb-usb-v2/af9035[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/dvb-usb-v2/anysee[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/dvb-usb-v2/au6610[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/dvb-usb-v2/az6007\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/dvb-usb-v2/ce6230[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/dvb-usb-v2/dvb_usb[^/]*$|^drivers/media/usb/dvb-usb-v2/usb_urb\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/dvb-usb-v2/ec168[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/dvb-usb-v2/gl861[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/dvb-usb-v2/lmedm04[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/dvb-usb-v2/mxl111sf[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/dvb-usb-v2/rtl28xxu[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/dvb-usb-v2/zd1301[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/dvb-usb/cxusb[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/em28xx/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/go7007/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/gspca/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/gspca/finepix\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/gspca/gl860/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/gspca/m5602/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/gspca/pac207\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/gspca/sn9c20x\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/gspca/t613\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/hackrf/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/hdpvr/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/msi2500/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/pvrusb2/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/pwc/|^include/trace/events/pwc\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/stk1160/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/uvc/|^include/uapi/linux/uvcvideo\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/|^drivers/staging/media/|^include/dt-bindings/media/|^include/linux/platform_data/media/|^include/media/|^include/uapi/linux/dvb/|^include/uapi/linux/ivtv[^/]*$|^include/uapi/linux/media\\\\.h$|^include/uapi/linux/uvcvideo\\\\.h$|^include/uapi/linux/v4l2-[^/]*$|^include/uapi/linux/videodev2\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/staging/media/atomisp/\"},\n\t\t\t{IncludeRegexp: \"^drivers/staging/media/deprecated/atmel/atmel-isc[^/]*$|^drivers/staging/media/deprecated/atmel/atmel-sama[^/]*-isc[^/]*$|^drivers/media/platform/microchip/microchip-isc[^/]*$|^drivers/media/platform/microchip/microchip-sama[^/]*-isc[^/]*$|^include/linux/atmel-isc-media\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/staging/media/imx/|^include/linux/imx-media\\\\.h$|^include/media/imx\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/staging/media/ipu3/\"},\n\t\t\t{IncludeRegexp: \"^drivers/staging/media/max96712/max96712\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/staging/media/meson/vdec/\"},\n\t\t\t{IncludeRegexp: \"^drivers/staging/media/rkvdec/\"},\n\t\t\t{IncludeRegexp: \"^drivers/staging/media/sunxi/cedrus/\"},\n\t\t\t{IncludeRegexp: \"^drivers/staging/media/sunxi/sun6i-isp/|^drivers/staging/media/sunxi/sun6i-isp/uapi/sun6i-isp-config\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/staging/media/tegra-video/\"},\n\t\t},\n\t}\n\n\tmediatek = Subsystem{\n\t\tName: \"mediatek\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&arm},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/mt2[^/]*$|^arch/arm/boot/dts/mt6[^/]*$|^arch/arm/boot/dts/mt7[^/]*$|^arch/arm/boot/dts/mt8[^/]*$|^arch/arm/mach-mediatek/|^arch/arm64/boot/dts/mediatek/|^drivers/soc/mediatek/|mtk|mt[2678]\"},\n\t\t\t{IncludeRegexp: \"^drivers/bluetooth/btmtkuart\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/dma/mediatek/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/mediatek/|^drivers/phy/mediatek/phy-mtk-dp\\\\.c$|^drivers/phy/mediatek/phy-mtk-hdmi[^/]*$|^drivers/phy/mediatek/phy-mtk-mipi[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iommu/mtk_iommu[^/]*$|^include/dt-bindings/memory/mt[^/]*-port\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/memory/mtk-smi\\\\.c$|^include/soc/mediatek/smi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/[^/]*mediatek[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/phy/mediatek/\"},\n\t\t\t{IncludeRegexp: \"^drivers/pinctrl/mediatek/\"},\n\t\t\t{IncludeRegexp: \"^drivers/rtc/rtc-mt2712\\\\.c$|^drivers/rtc/rtc-mt6397\\\\.c$|^drivers/rtc/rtc-mt7622\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/ufs/host/ufs-mediatek[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/host/xhci-mtk[^/]*$|^drivers/usb/mtu3/\"},\n\t\t},\n\t}\n\n\tmegaraid = Subsystem{\n\t\tName: \"megaraid\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&scsi},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/scsi/megaraid\\\\.[^/]*$|^drivers/scsi/megaraid/\"},\n\t\t},\n\t}\n\n\tmhi = Subsystem{\n\t\tName: \"mhi\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&armmsm},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/bus/mhi/|^include/linux/mhi\\\\.h$\"},\n\t\t},\n\t}\n\n\tmips = Subsystem{\n\t\tName: \"mips\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/mips/bcm47xx/|^arch/mips/include/asm/mach-bcm47xx/\"},\n\t\t\t{IncludeRegexp: \"^arch/mips/bmips/|^arch/mips/boot/dts/brcm/bcm[^/]*\\\\.dts[^/]*$|^arch/mips/include/asm/mach-bmips/|^arch/mips/kernel/[^/]*bmips[^/]*$|^drivers/soc/bcm/bcm63xx$|^drivers/irqchip/irq-bcm63[^/]*$|^drivers/irqchip/irq-bcm7[^/]*$|^drivers/irqchip/irq-brcmstb[^/]*$|^include/linux/bcm963xx_nvram\\\\.h$|^include/linux/bcm963xx_tag\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^arch/mips/boot/dts/img/boston\\\\.dts$|^arch/mips/configs/generic/board-boston\\\\.config$|^drivers/clk/imgtec/clk-boston\\\\.c$|^include/dt-bindings/clock/boston-clock\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^arch/mips/boot/dts/img/pistachio[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/mips/boot/dts/ingenic/|^arch/mips/generic/board-ingenic\\\\.c$|^arch/mips/include/asm/mach-ingenic/|^arch/mips/ingenic/Kconfig$|^drivers/clk/ingenic/|^drivers/dma/dma-jz4780\\\\.c$|^drivers/gpu/drm/ingenic/|^drivers/i2c/busses/i2c-jz4780\\\\.c$|^drivers/iio/adc/ingenic-adc\\\\.c$|^drivers/irqchip/irq-ingenic\\\\.c$|^drivers/memory/jz4780-nemc\\\\.c$|^drivers/mmc/host/jz4740_mmc\\\\.c$|^drivers/mtd/nand/raw/ingenic/|^drivers/pinctrl/pinctrl-ingenic\\\\.c$|^drivers/power/supply/ingenic-battery\\\\.c$|^drivers/pwm/pwm-jz4740\\\\.c$|^drivers/remoteproc/ingenic_rproc\\\\.c$|^drivers/rtc/rtc-jz4740\\\\.c$|^drivers/tty/serial/8250/8250_ingenic\\\\.c$|^drivers/usb/musb/jz4740\\\\.c$|^drivers/watchdog/jz4740_wdt\\\\.c$|^include/dt-bindings/iio/adc/ingenic,adc\\\\.h$|^include/linux/mfd/ingenic-tcu\\\\.h$|^sound/soc/codecs/jz47[^/]*$|^sound/soc/jz4740/\"},\n\t\t\t{IncludeRegexp: \"^arch/mips/boot/dts/mscc/|^arch/mips/configs/generic/board-ocelot\\\\.config$|^arch/mips/generic/board-ocelot\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/mips/boot/dts/ralink/mt7621[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/mips/boot/dts/ralink/omega2p\\\\.dts$\"},\n\t\t\t{IncludeRegexp: \"^arch/mips/boot/dts/ralink/vocore2\\\\.dts$\"},\n\t\t\t{IncludeRegexp: \"^arch/mips/configs/generic/board-ranchu\\\\.config$|^arch/mips/generic/board-ranchu\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/mips/dec/|^arch/mips/include/asm/dec/|^arch/mips/include/asm/mach-dec/\"},\n\t\t\t{IncludeRegexp: \"^arch/mips/generic/|^arch/mips/tools/generic-board-config\\\\.sh$\"},\n\t\t\t{IncludeRegexp: \"^arch/mips/include/asm/kvm[^/]*$|^arch/mips/include/uapi/asm/kvm[^/]*$|^arch/mips/kvm/\"},\n\t\t\t{IncludeRegexp: \"^arch/mips/include/asm/mach-loongson2ef/|^arch/mips/loongson2ef/|^drivers/cpufreq/loongson2_cpufreq\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/mips/include/asm/mach-loongson32/|^arch/mips/loongson32/|^drivers/[^/]*/[^/]*/[^/]*loongson1[^/]*$|^drivers/[^/]*/[^/]*loongson1[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/mips/include/asm/mach-loongson64/|^arch/mips/loongson64/|^drivers/irqchip/irq-loongson[^/]*$|^drivers/platform/mips/cpu_hwmon\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/mips/lantiq$|^drivers/soc/lantiq$\"},\n\t\t\t{IncludeRegexp: \"^arch/mips/math-emu/dp_rint\\\\.c$|^arch/mips/math-emu/sp_rint\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/mips/ralink$\"},\n\t\t\t{IncludeRegexp: \"^arch/mips/|^drivers/platform/mips/|^include/dt-bindings/mips/\"},\n\t\t\t{IncludeRegexp: \"^drivers/bus/mips_cdmm\\\\.c$|^drivers/clocksource/mips-gic-timer\\\\.c$|^drivers/cpuidle/cpuidle-cps\\\\.c$|^drivers/irqchip/irq-mips-cpu\\\\.c$|^drivers/irqchip/irq-mips-gic\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/edac/octeon_edac[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/firmware/broadcom/\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/nand/raw/ingenic/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/sgi/ioc3-eth\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pinctrl/ralink/\"},\n\t\t\t{IncludeRegexp: \"^drivers/tc/|^include/linux/tc\\\\.h$\"},\n\t\t},\n\t}\n\n\tmjpeg = Subsystem{\n\t\tName: \"mjpeg\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&media},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/media/pci/zoran/\"},\n\t\t},\n\t}\n\n\tmm = Subsystem{\n\t\tName: \"mm\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/[^/]*/include/asm/percpu\\\\.h$|^include/linux/percpu[^/]*\\\\.h$|^lib/percpu[^/]*\\\\.c$|^mm/percpu[^/]*\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/[^/]*/include/asm/tlb\\\\.h$|^include/asm-generic/tlb\\\\.h$|^mm/mmu_gather\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/base/memory\\\\.c$|^include/linux/memory_hotplug\\\\.h$|^mm/memory_hotplug\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^fs/[^/]*binfmt_[^/]*\\\\.c$|^fs/exec\\\\.c$|^include/linux/binfmts\\\\.h$|^include/linux/elf\\\\.h$|^include/uapi/linux/binfmts\\\\.h$|^include/uapi/linux/elf\\\\.h$|asm/elf.h|binfmt\"},\n\t\t\t{IncludeRegexp: \"^fs/hugetlbfs/|^include/linux/hugetlb\\\\.h$|^mm/hugetlb\\\\.c$|^mm/hugetlb_vmemmap\\\\.c$|^mm/hugetlb_vmemmap\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/damon\\\\.h$|^include/trace/events/damon\\\\.h$|^mm/damon/\"},\n\t\t\t{IncludeRegexp: \"^include/linux/gfp\\\\.h$|^include/linux/gfp_types\\\\.h$|^include/linux/memory_hotplug\\\\.h$|^include/linux/mm\\\\.h$|^include/linux/mmzone\\\\.h$|^include/linux/pagewalk\\\\.h$|^mm/\"},\n\t\t\t{IncludeRegexp: \"^include/linux/hmm[^/]*$|^lib/test_hmm[^/]*$|^mm/hmm[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/maple_tree\\\\.h$|^include/trace/events/maple_tree\\\\.h$|^lib/maple_tree\\\\.c$|^lib/test_maple_tree\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/memblock\\\\.h$|^mm/memblock\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/page_table_check\\\\.h$|^mm/page_table_check\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/shmem_fs\\\\.h$|^mm/shmem\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/sl.b[^/]*\\\\.h$|^mm/sl.b[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/vmalloc\\\\.h$|^mm/vmalloc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/zpool\\\\.h$|^mm/zpool\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/zsmalloc\\\\.h$|^mm/zsmalloc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^mm/hwpoison-inject\\\\.c$|^mm/memory-failure\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^mm/memcontrol\\\\.c$|^mm/swap_cgroup\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^mm/z3fold\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^mm/zbud\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^mm/zswap\\\\.c$\"},\n\t\t},\n\t}\n\n\tmmc = Subsystem{\n\t\tName: \"mmc\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/memstick/|^include/linux/memstick\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/host/cqhci[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/host/dw_mmc[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/host/omap_hsmmc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/host/renesas_sdhi[^/]*$|^drivers/mmc/host/tmio_mmc[^/]*$|^include/linux/mfd/tmio\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/host/sdhci-brcmstb[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/host/sdhci-esdhc-imx\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/host/sdhci-esdhc-mcf\\\\.c$|^include/linux/platform_data/mmc-esdhc-mcf\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/host/sdhci-of-aspeed[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/host/sdhci-of-at91\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/host/sdhci-omap\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/host/sdhci-pci-dwc-mshc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/host/sdhci-s3c[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/host/sdhci-spear\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/host/sdhci-xenon[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/host/sdhci[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/host/vub300\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/|^include/linux/mmc/|^include/uapi/linux/mmc/\"},\n\t\t},\n\t}\n\n\tmodules = Subsystem{\n\t\tName: \"modules\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^include/linux/kmod\\\\.h$|^kernel/kmod\\\\.c$|^lib/test_kmod\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/module\\\\.h$|^kernel/module/\"},\n\t\t},\n\t}\n\n\tmpi3 = Subsystem{\n\t\tName: \"mpi3\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&scsi},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/scsi/mpi3mr/\"},\n\t\t},\n\t}\n\n\tmptfusion = Subsystem{\n\t\tName: \"mpt-fusion\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&scsi},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/message/fusion/|^drivers/scsi/mpt3sas/\"},\n\t\t},\n\t}\n\n\tmptcp = Subsystem{\n\t\tName: \"mptcp\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^include/net/mptcp\\\\.h$|^include/trace/events/mptcp\\\\.h$|^include/uapi/linux/mptcp\\\\.h$|^net/mptcp/\"},\n\t\t},\n\t}\n\n\tmtd = Subsystem{\n\t\tName: \"mtd\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/mtd/devices/block2mtd\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/devices/docg3[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/devices/phram\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/hyperbus/|^include/linux/mtd/hyperbus\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/nand/onenand/|^include/linux/mtd/onenand[^/]*\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/nand/raw/arasan-nand-controller\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/nand/raw/atmel/\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/nand/raw/brcmnand/|^include/linux/platform_data/brcmnand\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/nand/raw/cadence-nand-controller\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/nand/raw/denali[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/nand/raw/gpmi-nand/\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/nand/raw/ingenic/\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/nand/raw/marvell_nand\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/nand/raw/meson_[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/nand/raw/mtk_[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/nand/raw/pl35x-nand-controller\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/nand/raw/qcom_nandc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/nand/raw/renesas-nand-controller\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/nand/raw/vf610_nfc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/nand/|^include/linux/mtd/[^/]*nand[^/]*\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/spi-nor/|^include/linux/mtd/spi-nor\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/ubi/|^include/linux/mtd/ubi\\\\.h$|^include/uapi/mtd/ubi-user\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/|^include/linux/mtd/|^include/uapi/mtd/\"},\n\t\t\t{IncludeRegexp: \"^fs/jffs2/|^include/uapi/linux/jffs2\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^fs/ubifs/\"},\n\t\t},\n\t}\n\n\tnbd = Subsystem{\n\t\tName: \"nbd\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&block},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/block/nbd\\\\.c$|^include/trace/events/nbd\\\\.h$|^include/uapi/linux/nbd\\\\.h$\"},\n\t\t},\n\t}\n\n\tnet = Subsystem{\n\t\tName: \"net\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/arm64/boot/dts/amd/amd-seattle-xgbe[^/]*\\\\.dtsi$|^drivers/net/ethernet/amd/xgbe/\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/net/|^include/linux/ip\\\\.h$|^include/linux/ipv6[^/]*$|^include/net/fib[^/]*$|^include/net/ip[^/]*$|^include/net/route\\\\.h$|^net/ipv4/|^net/ipv6/\"},\n\t\t\t{IncludeRegexp: \"^drivers/atm/|^include/linux/atm[^/]*$|^include/uapi/linux/atm[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/connector/|^drivers/net/|^include/dt-bindings/net/|^include/linux/etherdevice\\\\.h$|^include/linux/fcdevice\\\\.h$|^include/linux/fddidevice\\\\.h$|^include/linux/hippidevice\\\\.h$|^include/linux/if_[^/]*$|^include/linux/inetdevice\\\\.h$|^include/linux/netdevice\\\\.h$|^include/uapi/linux/if_[^/]*$|^include/uapi/linux/netdevice\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/firmware/broadcom/tee_bnxt_fw\\\\.c$|^drivers/net/ethernet/broadcom/bnxt/|^include/linux/firmware/broadcom/tee_bnxt_fw\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/isdn/Kconfig$|^drivers/isdn/hardware/|^drivers/isdn/mISDN/\"},\n\t\t\t{IncludeRegexp: \"^drivers/isdn/capi/|^include/linux/isdn/|^include/uapi/linux/isdn/|^net/bluetooth/cmtp/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/amt\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/appletalk/|^include/linux/atalk\\\\.h$|^include/uapi/linux/atalk\\\\.h$|^net/appletalk/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/arcnet/|^include/uapi/linux/if_arcnet\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/bonding/|^include/net/bond[^/]*$|^include/uapi/linux/if_bonding\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/caif/|^include/net/caif/|^include/uapi/linux/caif/|^net/caif/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/dsa/b53/|^drivers/net/dsa/bcm_sf2[^/]*$|^include/linux/dsa/brcm\\\\.h$|^include/linux/platform_data/b53\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/dsa/hirschmann/|^include/linux/platform_data/hirschmann-hellcreek\\\\.h$|^net/dsa/tag_hellcreek\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/dsa/lantiq_gswip\\\\.c$|^drivers/net/dsa/lantiq_pce\\\\.h$|^drivers/net/ethernet/lantiq_xrx200\\\\.c$|^net/dsa/tag_gswip\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/dsa/microchip/|^include/linux/dsa/ksz_common\\\\.h$|^include/linux/platform_data/microchip-ksz\\\\.h$|^net/dsa/tag_ksz\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/dsa/mt7530\\\\.[^/]*$|^net/dsa/tag_mtk\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/dsa/mv88e6xxx/|^include/linux/dsa/mv88e6xxx\\\\.h$|^include/linux/platform_data/mv88e6xxx\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/dsa/ocelot/|^drivers/net/ethernet/mscc/|^include/soc/mscc/ocelot[^/]*$|^net/dsa/tag_ocelot\\\\.c$|^net/dsa/tag_ocelot_8021q\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/dsa/rzn1_a5psw[^/]*$|^drivers/net/pcs/pcs-rzn1-miic\\\\.c$|^include/dt-bindings/net/pcs-rzn1-miic\\\\.h$|^include/linux/pcs-rzn1-miic\\\\.h$|^net/dsa/tag_rzn1_a5psw\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/dsa/xrs700x/|^net/dsa/tag_xrs700x\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/3com/3c59x\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/3com/typhoon[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/8390/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/aeroflex/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/altera/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/amazon/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/amd/pcnet32\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/aquantia/atlantic/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/aquantia/atlantic/aq_ptp[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/atheros/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/broadcom/b44\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/broadcom/bcm4908_enet\\\\.[^/]*$|^drivers/net/ethernet/broadcom/unimac\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/broadcom/bcmsysport\\\\.[^/]*$|^drivers/net/ethernet/broadcom/unimac\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/broadcom/bgmac[^/]*$|^drivers/net/ethernet/broadcom/unimac\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/broadcom/bnx2\\\\.[^/]*$|^drivers/net/ethernet/broadcom/bnx2_[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/broadcom/bnx2x/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/broadcom/genet/|^drivers/net/ethernet/broadcom/unimac\\\\.h$|^drivers/net/mdio/mdio-bcm-unimac\\\\.c$|^include/linux/platform_data/bcmgenet\\\\.h$|^include/linux/platform_data/mdio-bcm-unimac\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/broadcom/tg3\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/brocade/bna/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/cavium/liquidio/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/chelsio/cxgb3/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/chelsio/cxgb4/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/chelsio/cxgb4vf/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/chelsio/inline_crypto/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/cirrus/ep93xx_eth\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/dec/tulip/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/dec/tulip/dmfe\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/dlink/sundance\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/ec_bhf\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/emulex/benet/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/freescale/dpaa$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/freescale/dpaa2/Kconfig$|^drivers/net/ethernet/freescale/dpaa2/dpaa2-eth[^/]*$|^drivers/net/ethernet/freescale/dpaa2/dpaa2-mac[^/]*$|^drivers/net/ethernet/freescale/dpaa2/dpaa2-xsk[^/]*$|^drivers/net/ethernet/freescale/dpaa2/dpkg\\\\.h$|^drivers/net/ethernet/freescale/dpaa2/dpmac[^/]*$|^drivers/net/ethernet/freescale/dpaa2/dpni[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/freescale/dpaa2/dpaa2-ptp[^/]*$|^drivers/net/ethernet/freescale/dpaa2/dprtc[^/]*$|^drivers/net/ethernet/freescale/enetc/enetc_ptp\\\\.c$|^drivers/ptp/ptp_qoriq\\\\.c$|^drivers/ptp/ptp_qoriq_debugfs\\\\.c$|^include/linux/fsl/ptp_qoriq\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/freescale/dpaa2/dpaa2-switch[^/]*$|^drivers/net/ethernet/freescale/dpaa2/dpsw[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/freescale/enetc/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/freescale/fec\\\\.h$|^drivers/net/ethernet/freescale/fec_main\\\\.c$|^drivers/net/ethernet/freescale/fec_ptp\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/freescale/fman$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/freescale/fs_enet/|^include/linux/fs_enet_pd\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/freescale/gianfar[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/freescale/ucc_geth[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/fungible/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/google$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/hisilicon/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/hisilicon/hns3/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/huawei/hinic/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/ibm/ehea/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/ibm/ibmveth\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/ibm/ibmvnic\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/jme\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/marvell/mv643xx_eth\\\\.[^/]*$|^include/linux/mv643xx\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/marvell/mvneta\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/marvell/mvpp2/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/marvell/octeon_ep$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/marvell/octeontx2/af/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/marvell/octeontx2/nic/|^include/linux/soc/marvell/octeontx2/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/marvell/sk[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/mediatek/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/mellanox/mlx4/en_[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/mellanox/mlx4/|^include/linux/mlx4/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/mellanox/mlx5/core/en_[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/mellanox/mlx5/core/en_accel/|^drivers/net/ethernet/mellanox/mlx5/core/fpga/|^include/linux/mlx5/mlx5_ifc_fpga\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/mellanox/mlx5/core/|^include/linux/mlx5/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/mellanox/mlxfw/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/mellanox/mlxsw/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/microchip/lan743x_[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/microchip/lan966x/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/myricom/myri10ge/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/natsemi/sonic\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/neterion/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/nvidia/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/pensando/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/qlogic/netxen/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/qlogic/qed/|^drivers/net/ethernet/qlogic/qede/|^include/linux/qed/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/qlogic/qla3xxx\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/qlogic/qlcnic/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/qualcomm/emac/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/qualcomm/rmnet/|^include/linux/if_rmnet\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/rdc/r6040\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/realtek/r8169[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/renesas/|^include/linux/sh_eth\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/rocker/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/samsung/sxgbe/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/sfc/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/sis/sis190\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/sis/sis900\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/smsc/smsc911x\\\\.[^/]*$|^include/linux/smsc911x\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/smsc/smsc9420\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/socionext/netsec\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/socionext/sni_ave\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/stmicro/stmmac/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/stmicro/stmmac/dwmac-qcom-ethqos\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/sunplus/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/synopsys/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/tehuti/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/ti/cpmac\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/ti/cpsw[^/]*$|^drivers/net/ethernet/ti/davinci[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/toshiba/ps3_gelic_net\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/toshiba/spider_net[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/via/via-velocity\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/wangxun/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ipa/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/mctp/|^include/net/mctp\\\\.h$|^include/net/mctpdevice\\\\.h$|^include/net/netns/mctp\\\\.h$|^net/mctp/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/mdio/mdio-mvusb\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/mdio/|^drivers/net/mdio/acpi_mdio\\\\.c$|^drivers/net/mdio/fwnode_mdio\\\\.c$|^drivers/net/mdio/of_mdio\\\\.c$|^drivers/net/pcs/|^drivers/net/phy/|^include/dt-bindings/net/qca-ar803x\\\\.h$|^include/linux/linkmode\\\\.h$|^include/linux/[^/]*mdio[^/]*\\\\.h$|^include/linux/mdio/[^/]*\\\\.h$|^include/linux/mii\\\\.h$|^include/linux/of_net\\\\.h$|^include/linux/phy\\\\.h$|^include/linux/phy_fixed\\\\.h$|^include/linux/platform_data/mdio-bcm-unimac\\\\.h$|^include/linux/platform_data/mdio-gpio\\\\.h$|^include/trace/events/mdio\\\\.h$|^include/uapi/linux/mdio\\\\.h$|^include/uapi/linux/mii\\\\.h$|^net/core/of_net\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/net_failover\\\\.c$|^include/net/net_failover\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/pcs/pcs-altera-tse\\\\.c$|^include/linux/pcs-altera-tse\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/pcs/pcs-lynx\\\\.c$|^include/linux/pcs-lynx\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/pcs/pcs-xpcs\\\\.c$|^drivers/net/pcs/pcs-xpcs\\\\.h$|^include/linux/pcs/pcs-xpcs\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/phy/adin\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/phy/bcm[^/]*\\\\.\\\\[ch\\\\]$|^drivers/net/phy/broadcom\\\\.c$|^include/linux/brcmphy\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/phy/dp83640[^/]*$|^drivers/ptp/|^include/linux/ptp_cl[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/phy/marvell10g\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/phy/mdio-open-alliance\\\\.h$|^net/ethtool/plca\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/phy/microchip_t1\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/phy/motorcomm\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/phy/mxl-gpy\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/phy/ncn[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/phy/nxp-c45-tja11xx\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/phy/phylink\\\\.c$|^drivers/net/phy/sfp[^/]*$|^include/linux/mdio/mdio-i2c\\\\.h$|^include/linux/phylink\\\\.h$|^include/linux/sfp\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ppp/pptp\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/team/|^include/linux/if_team\\\\.h$|^include/uapi/linux/if_team\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/thunderbolt/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/usb/dm9601\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/usb/lan78xx\\\\.[^/]*$|^include/dt-bindings/net/microchip-lan78xx\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/usb/pegasus\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/usb/qmi_wwan\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/usb/rtl8150\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/usb/smsc75xx\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/usb/smsc95xx\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/usb/usbnet\\\\.c$|^include/linux/usb/usbnet\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/vmxnet3/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/vrf\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/vsockmon\\\\.c$|^include/net/af_vsock\\\\.h$|^include/uapi/linux/vm_sockets\\\\.h$|^include/uapi/linux/vm_sockets_diag\\\\.h$|^include/uapi/linux/vsockmon\\\\.h$|^net/vmw_vsock/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wan/fsl_ucc_hdlc[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireguard/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wwan/iosm/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wwan/qcom_bam_dmux\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wwan/rpmsg_wwan_ctrl\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wwan/t7xx/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wwan/|^include/linux/wwan\\\\.h$|^include/uapi/linux/wwan\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/xen-netback/\"},\n\t\t\t{IncludeRegexp: \"^drivers/nfc/virtual_ncidev\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/nfc/|^include/linux/platform_data/nfcmrvl\\\\.h$|^include/net/nfc/|^include/uapi/linux/nfc\\\\.h$|^net/nfc/\"},\n\t\t\t{IncludeRegexp: \"^drivers/phy/freescale/phy-fsl-lynx-28g\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/ptp/ptp_ocp\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/ptp/ptp_vclock\\\\.c$|^net/ethtool/phc_vclocks\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/ptp/ptp_vmw\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/s390/net/\"},\n\t\t\t{IncludeRegexp: \"^drivers/s390/net/[^/]*iucv[^/]*$|^include/net/iucv/|^net/iucv/\"},\n\t\t\t{IncludeRegexp: \"^drivers/staging/qlge/\"},\n\t\t\t{IncludeRegexp: \"^drivers/vhost/vsock\\\\.c$|^include/linux/virtio_vsock\\\\.h$|^include/uapi/linux/virtio_vsock\\\\.h$|^net/vmw_vsock/virtio_transport\\\\.c$|^net/vmw_vsock/virtio_transport_common\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/vhost/|^include/linux/vhost_iotlb\\\\.h$|^include/uapi/linux/vhost\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/in\\\\.h$|^include/linux/net\\\\.h$|^include/linux/netdevice\\\\.h$|^include/net/|^include/uapi/linux/in\\\\.h$|^include/uapi/linux/net\\\\.h$|^include/uapi/linux/net_namespace\\\\.h$|^include/uapi/linux/netdevice\\\\.h$|^lib/net_utils\\\\.c$|^lib/random32\\\\.c$|^net/\"},\n\t\t\t{IncludeRegexp: \"^include/linux/llc\\\\.h$|^include/net/llc[^/]*$|^include/uapi/linux/llc\\\\.h$|^net/llc/\"},\n\t\t\t{IncludeRegexp: \"^include/linux/netfilter_bridge/|^net/bridge/\"},\n\t\t\t{IncludeRegexp: \"^include/linux/objagg\\\\.h$|^lib/objagg\\\\.c$|^lib/test_objagg\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/packing\\\\.h$|^lib/packing\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/parman\\\\.h$|^lib/parman\\\\.c$|^lib/test_parman\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/rhashtable-types\\\\.h$|^include/linux/rhashtable\\\\.h$|^lib/rhashtable\\\\.c$|^lib/test_rhashtable\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/skmsg\\\\.h$|^net/core/skmsg\\\\.c$|^net/core/sock_map\\\\.c$|^net/ipv4/tcp_bpf\\\\.c$|^net/ipv4/udp_bpf\\\\.c$|^net/unix/unix_bpf\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/tcp\\\\.h$|^include/net/tcp\\\\.h$|^include/trace/events/tcp\\\\.h$|^include/uapi/linux/tcp\\\\.h$|^net/ipv4/syncookies\\\\.c$|^net/ipv4/tcp[^/]*\\\\.c$|^net/ipv6/syncookies\\\\.c$|^net/ipv6/tcp[^/]*\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/net/calipso\\\\.h$|^include/net/cipso_ipv4\\\\.h$|^include/net/netlabel\\\\.h$|^include/uapi/linux/netfilter/xt_CONNSECMARK\\\\.h$|^include/uapi/linux/netfilter/xt_SECMARK\\\\.h$|^net/ipv4/cipso_ipv4\\\\.c$|^net/ipv6/calipso\\\\.c$|^net/netfilter/xt_CONNSECMARK\\\\.c$|^net/netfilter/xt_SECMARK\\\\.c$|^net/netlabel/\"},\n\t\t\t{IncludeRegexp: \"^include/net/devlink\\\\.h$|^include/uapi/linux/devlink\\\\.h$|^net/devlink/\"},\n\t\t\t{IncludeRegexp: \"^include/net/failover\\\\.h$|^net/core/failover\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/net/gre\\\\.h$|^net/ipv4/gre_demux\\\\.c$|^net/ipv4/gre_offload\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/net/ip_vs\\\\.h$|^include/uapi/linux/ip_vs\\\\.h$|^net/netfilter/ipvs/\"},\n\t\t\t{IncludeRegexp: \"^include/net/l3mdev\\\\.h$|^net/l3mdev$\"},\n\t\t\t{IncludeRegexp: \"^include/net/mptcp\\\\.h$|^include/trace/events/mptcp\\\\.h$|^include/uapi/linux/mptcp\\\\.h$|^net/mptcp/\"},\n\t\t\t{IncludeRegexp: \"^include/net/netns/nexthop\\\\.h$|^include/net/nexthop\\\\.h$|^include/uapi/linux/nexthop\\\\.h$|^net/ipv4/nexthop\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/net/page_pool\\\\.h$|^include/trace/events/page_pool\\\\.h$|^net/core/page_pool\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/net/pkt_cls\\\\.h$|^include/net/pkt_sched\\\\.h$|^include/net/tc_act/|^include/uapi/linux/pkt_cls\\\\.h$|^include/uapi/linux/pkt_sched\\\\.h$|^include/uapi/linux/tc_act/|^include/uapi/linux/tc_ematch/|^net/sched/\"},\n\t\t\t{IncludeRegexp: \"^include/net/switchdev\\\\.h$|^net/switchdev/\"},\n\t\t\t{IncludeRegexp: \"^include/net/tls\\\\.h$|^include/uapi/linux/tls\\\\.h$|^net/tls/\"},\n\t\t\t{IncludeRegexp: \"^include/net/xdp\\\\.h$|^include/net/xdp_priv\\\\.h$|^include/trace/events/xdp\\\\.h$|^kernel/bpf/cpumap\\\\.c$|^kernel/bpf/devmap\\\\.c$|^net/core/xdp\\\\.c$|^drivers/net/ethernet/[^/]*/[^/]*/[^/]*/[^/]*/[^/]*xdp[^/]*$|^drivers/net/ethernet/[^/]*/[^/]*/[^/]*xdp[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^include/net/xdp_sock[^/]*$|^include/net/xsk_buff_pool\\\\.h$|^include/uapi/linux/if_xdp\\\\.h$|^include/uapi/linux/xdp_diag\\\\.h$|^include/net/netns/xdp\\\\.h$|^net/xdp/\"},\n\t\t\t{IncludeRegexp: \"^include/net/xfrm\\\\.h$|^include/uapi/linux/xfrm\\\\.h$|^net/ipv4/ah4\\\\.c$|^net/ipv4/esp4[^/]*$|^net/ipv4/ip_vti\\\\.c$|^net/ipv4/ipcomp\\\\.c$|^net/ipv4/xfrm[^/]*$|^net/ipv6/ah6\\\\.c$|^net/ipv6/esp6[^/]*$|^net/ipv6/ip6_vti\\\\.c$|^net/ipv6/ipcomp6\\\\.c$|^net/ipv6/xfrm[^/]*$|^net/key/|^net/xfrm/\"},\n\t\t\t{IncludeRegexp: \"^include/uapi/linux/net_dropmon\\\\.h$|^net/core/drop_monitor\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/uapi/linux/openvswitch\\\\.h$|^net/openvswitch/\"},\n\t\t\t{IncludeRegexp: \"^include/uapi/linux/tipc[^/]*\\\\.h$|^net/tipc/\"},\n\t\t\t{IncludeRegexp: \"^kernel/bpf/bpf_struct[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^net/core/filter\\\\.c$|^net/sched/act_bpf\\\\.c$|^net/sched/cls_bpf\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^net/hsr/\"},\n\t\t\t{IncludeRegexp: \"^net/rds/\"},\n\t\t\t{IncludeRegexp: \"^net/sched/sch_cbs\\\\.c$|^net/sched/sch_etf\\\\.c$|^net/sched/sch_taprio\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^net/sched/sch_netem\\\\.c$\"},\n\t\t},\n\t}\n\n\tnetfilter = Subsystem{\n\t\tName: \"netfilter\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^include/linux/netfilter[^/]*$|^include/linux/netfilter/|^include/net/netfilter/|^include/uapi/linux/netfilter[^/]*$|^include/uapi/linux/netfilter/|^net/[^/]*/netfilter\\\\.c$|^net/[^/]*/netfilter/|^net/bridge/br_netfilter[^/]*\\\\.c$|^net/netfilter/\"},\n\t\t},\n\t}\n\n\tnfc = Subsystem{\n\t\tName: \"nfc\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/nfc/nxp-nci$\"},\n\t\t\t{IncludeRegexp: \"^drivers/nfc/s3fwrn5$\"},\n\t\t\t{IncludeRegexp: \"^drivers/nfc/trf7970a\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/nfc/virtual_ncidev\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/nfc/|^include/linux/platform_data/nfcmrvl\\\\.h$|^include/net/nfc/|^include/uapi/linux/nfc\\\\.h$|^net/nfc/\"},\n\t\t},\n\t}\n\n\tnfs = Subsystem{\n\t\tName: \"nfs\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/exportfs/|^fs/lockd/|^fs/nfs_common/|^fs/nfsd/|^include/linux/lockd/|^include/linux/sunrpc/|^include/trace/events/rpcgss\\\\.h$|^include/trace/events/rpcrdma\\\\.h$|^include/trace/events/sunrpc\\\\.h$|^include/trace/misc/fs\\\\.h$|^include/trace/misc/nfs\\\\.h$|^include/trace/misc/sunrpc\\\\.h$|^include/uapi/linux/nfsd/|^include/uapi/linux/sunrpc/|^net/sunrpc/\"},\n\t\t\t{IncludeRegexp: \"^fs/lockd/|^fs/nfs/|^fs/nfs_common/|^include/linux/lockd/|^include/linux/nfs[^/]*$|^include/linux/sunrpc/|^include/uapi/linux/nfs[^/]*$|^include/uapi/linux/sunrpc/|^net/sunrpc/\"},\n\t\t},\n\t}\n\n\tnilfs = Subsystem{\n\t\tName: \"nilfs\",\n\t\tSyscalls: []string{\"syz_mount_image$nilfs2\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/nilfs2/|^include/trace/events/nilfs2\\\\.h$|^include/uapi/linux/nilfs2_api\\\\.h$|^include/uapi/linux/nilfs2_ondisk\\\\.h$\"},\n\t\t},\n\t}\n\n\tnitro = Subsystem{\n\t\tName: \"nitro\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/virt/nitro_enclaves/|^include/linux/nitro_enclaves\\\\.h$|^include/uapi/linux/nitro_enclaves\\\\.h$\"},\n\t\t},\n\t}\n\n\tnouveau = Subsystem{\n\t\tName: \"nouveau\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&dri},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/x86/mm/kmmio\\\\.c$|^arch/x86/mm/mmio-mod\\\\.c$|^arch/x86/mm/testmmiotrace\\\\.c$|^include/linux/mmiotrace\\\\.h$|^kernel/trace/trace_mmiotrace\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/nouveau/|^include/uapi/drm/nouveau_drm\\\\.h$\"},\n\t\t},\n\t}\n\n\tntb = Subsystem{\n\t\tName: \"ntb\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/net/ntb_netdev\\\\.c$|^drivers/ntb/|^drivers/pci/endpoint/functions/pci-epf-[^/]*ntb\\\\.c$|^include/linux/ntb\\\\.h$|^include/linux/ntb_transport\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/ntb/hw/amd/\"},\n\t\t\t{IncludeRegexp: \"^drivers/ntb/hw/idt/\"},\n\t\t\t{IncludeRegexp: \"^drivers/ntb/hw/intel/\"},\n\t\t},\n\t}\n\n\tntfs = Subsystem{\n\t\tName: \"ntfs\",\n\t\tSyscalls: []string{\"syz_mount_image$ntfs\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^block/partitions/ldm\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^fs/ntfs/\"},\n\t\t},\n\t}\n\n\tntfs3 = Subsystem{\n\t\tName: \"ntfs3\",\n\t\tSyscalls: []string{\"syz_mount_image$ntfs3\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/ntfs3/\"},\n\t\t},\n\t}\n\n\tnvdimm = Subsystem{\n\t\tName: \"nvdimm\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/acpi/nfit/|^drivers/nvdimm/|^include/linux/libnvdimm\\\\.h$|^include/linux/nd\\\\.h$|^include/uapi/linux/ndctl\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/dax/\"},\n\t\t\t{IncludeRegexp: \"^drivers/nvdimm/btt[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/nvdimm/of_pmem\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/nvdimm/pmem[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^fs/dax\\\\.c$|^include/linux/dax\\\\.h$|^include/trace/events/fs_dax\\\\.h$\"},\n\t\t},\n\t}\n\n\tnvme = Subsystem{\n\t\tName: \"nvme\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/nvme/common/|^drivers/nvme/host/|^include/linux/nvme-[^/]*\\\\.h$|^include/linux/nvme\\\\.h$|^include/uapi/linux/nvme_ioctl\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/nvme/host/auth\\\\.c$|^drivers/nvme/target/auth\\\\.c$|^drivers/nvme/target/fabrics-cmd-auth\\\\.c$|^include/linux/nvme-auth\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/nvme/host/fc\\\\.c$|^drivers/nvme/target/fc\\\\.c$|^drivers/nvme/target/fcloop\\\\.c$|^include/linux/nvme-fc-driver\\\\.h$|^include/linux/nvme-fc\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/nvme/host/hwmon\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/nvme/target/\"},\n\t\t},\n\t}\n\n\tocfs2 = Subsystem{\n\t\tName: \"ocfs2\",\n\t\tSyscalls: []string{\"syz_mount_image$ocfs2\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/ocfs2/\"},\n\t\t},\n\t}\n\n\tomap = Subsystem{\n\t\tName: \"omap\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&arm},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/arm/[^/]*omap[^/]*/[^/]*clock[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/[^/]*omap[^/]*/[^/]*pm[^/]*$|^drivers/cpufreq/omap-cpufreq\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/[^/]*omap[^/]*/usb[^/]*$|^drivers/usb/[^/]*/[^/]*omap[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/[^/]*am3[^/]*$|^arch/arm/boot/dts/[^/]*am4[^/]*$|^arch/arm/boot/dts/[^/]*am5[^/]*$|^arch/arm/boot/dts/[^/]*dra7[^/]*$|^arch/arm/boot/dts/[^/]*omap[^/]*$|^arch/arm/boot/dts/logicpd-som-lv[^/]*$|^arch/arm/boot/dts/logicpd-torpedo[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/am335x-nano\\\\.dts$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/omap3-igep[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/configs/omap1_defconfig$|^arch/arm/mach-omap1/|^drivers/i2c/busses/i2c-omap\\\\.c$|^include/linux/platform_data/ams-delta-fiq\\\\.h$|^include/linux/platform_data/i2c-omap\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/configs/omap2plus_defconfig$|^arch/arm/mach-omap2/|^drivers/bus/ti-sysc\\\\.c$|^drivers/i2c/busses/i2c-omap\\\\.c$|^drivers/irqchip/irq-omap-intc\\\\.c$|^drivers/mfd/[^/]*omap[^/]*\\\\.c$|^drivers/mfd/menelaus\\\\.c$|^drivers/mfd/palmas\\\\.c$|^drivers/mfd/tps65217\\\\.c$|^drivers/mfd/tps65218\\\\.c$|^drivers/mfd/tps65219\\\\.c$|^drivers/mfd/tps65910\\\\.c$|^drivers/mfd/twl-core\\\\.\\\\[ch\\\\]$|^drivers/mfd/twl4030[^/]*\\\\.c$|^drivers/mfd/twl6030[^/]*\\\\.c$|^drivers/mfd/twl6040[^/]*\\\\.c$|^drivers/regulator/palmas-regulator[^/]*\\\\.c$|^drivers/regulator/pbias-regulator\\\\.c$|^drivers/regulator/tps65217-regulator\\\\.c$|^drivers/regulator/tps65218-regulator\\\\.c$|^drivers/regulator/tps65219-regulator\\\\.c$|^drivers/regulator/tps65910-regulator\\\\.c$|^drivers/regulator/twl-regulator\\\\.c$|^drivers/regulator/twl6030-regulator\\\\.c$|^include/linux/platform_data/i2c-omap\\\\.h$|^include/linux/platform_data/ti-sysc\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/mach-omap2/[^/]*gpmc[^/]*$|^drivers/memory/omap-gpmc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/mach-omap2/omap_hwmod[^/]*data[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/mach-omap2/omap_hwmod\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/mach-omap2/prm[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/clk/ti/|^include/linux/clk/ti\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/counter/ti-ecap-capture\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-omap\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hwspinlock/omap_hwspinlock\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-omap\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/host/omap\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/host/omap_hsmmc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/ti/cpsw[^/]*$|^drivers/net/ethernet/ti/davinci[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/cadence/pci-j721e\\\\.c$|^drivers/pci/controller/dwc/pci-dra7xx\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pinctrl/pinctrl-single\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/thermal/ti-soc-thermal/\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/omap/\"},\n\t\t\t{IncludeRegexp: \"^drivers/video/fbdev/omap2/\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/ti/n810\\\\.c$|^sound/soc/ti/omap[^/]*$|^sound/soc/ti/rx51\\\\.c$|^sound/soc/ti/sdma-pcm\\\\.[^/]*$\"},\n\t\t},\n\t}\n\n\toptee = Subsystem{\n\t\tName: \"op-tee\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/char/hw_random/optee-rng\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/tee/optee/\"},\n\t\t\t{IncludeRegexp: \"^drivers/tee/|^include/linux/tee_drv\\\\.h$|^include/uapi/linux/tee\\\\.h$\"},\n\t\t},\n\t}\n\n\topeniscsi = Subsystem{\n\t\tName: \"open-iscsi\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&scsi},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/scsi/[^/]*iscsi[^/]*$|^include/scsi/[^/]*iscsi[^/]*$\"},\n\t\t},\n\t}\n\n\topenbmc = Subsystem{\n\t\tName: \"openbmc\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/nuvoton-npcm[^/]*$|^arch/arm/mach-npcm/|^arch/arm64/boot/dts/nuvoton/|^drivers/[^/]*/[^/]*npcm[^/]*$|^drivers/[^/]*/[^/]*/[^/]*npcm[^/]*$|^drivers/rtc/rtc-nct3018y\\\\.c$|^include/dt-bindings/clock/nuvoton,npcm7xx-clock\\\\.h$|^include/dt-bindings/clock/nuvoton,npcm845-clk\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/nuvoton-wpcm450[^/]*$|^arch/arm/configs/wpcm450_defconfig$|^arch/arm/mach-npcm/wpcm450\\\\.c$|^drivers/[^/]*/[^/]*/[^/]*wpcm[^/]*$|^drivers/[^/]*/[^/]*wpcm[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-aspeed\\\\.c$|^drivers/irqchip/irq-aspeed-i2c-ic\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-fsi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/aspeed/\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/host/sdhci-of-aspeed[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/peci/controller/peci-aspeed\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/peci/|^include/linux/peci-cpu\\\\.h$|^include/linux/peci\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pinctrl/aspeed/\"},\n\t\t\t{IncludeRegexp: \"^drivers/spi/spi-aspeed-smc\\\\.c$\"},\n\t\t},\n\t}\n\n\topenipmi = Subsystem{\n\t\tName: \"openipmi\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/char/ipmi/|^include/linux/ipmi[^/]*$|^include/uapi/linux/ipmi[^/]*$\"},\n\t\t},\n\t}\n\n\topenrisc = Subsystem{\n\t\tName: \"openrisc\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/openrisc/|^drivers/irqchip/irq-ompic\\\\.c$|^drivers/irqchip/irq-or1k-[^/]*$\"},\n\t\t},\n\t}\n\n\topenvswitch = Subsystem{\n\t\tName: \"openvswitch\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^include/uapi/linux/openvswitch\\\\.h$|^net/openvswitch/\"},\n\t\t},\n\t}\n\n\topenwrt = Subsystem{\n\t\tName: \"openwrt\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/net/dsa/b53/|^drivers/net/dsa/bcm_sf2[^/]*$|^include/linux/dsa/brcm\\\\.h$|^include/linux/platform_data/b53\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/vlynq/vlynq\\\\.c$|^include/linux/vlynq\\\\.h$\"},\n\t\t},\n\t}\n\n\torangefs = Subsystem{\n\t\tName: \"orangefs\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/orangefs/\"},\n\t\t},\n\t}\n\n\tossdrivers = Subsystem{\n\t\tName: \"oss-drivers\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/netronome/\"},\n\t\t},\n\t}\n\n\toverlayfs = Subsystem{\n\t\tName: \"overlayfs\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/overlayfs/\"},\n\t\t},\n\t}\n\n\toxnas = Subsystem{\n\t\tName: \"oxnas\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&arm},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/ox8[^/]*\\\\.dts[^/]*$|^arch/arm/mach-oxnas/|^drivers/power/reset/oxnas-restart\\\\.c$|oxnas\"},\n\t\t},\n\t}\n\n\tparisc = Subsystem{\n\t\tName: \"parisc\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/parisc/|^drivers/char/agp/parisc-agp\\\\.c$|^drivers/input/misc/hp_sdc_rtc\\\\.c$|^drivers/input/serio/gscps2\\\\.c$|^drivers/input/serio/hp_sdc[^/]*$|^drivers/parisc/|^drivers/parport/parport_gsc\\\\.[^/]*$|^drivers/tty/serial/8250/8250_parisc\\\\.c$|^drivers/video/console/sti[^/]*$|^drivers/video/fbdev/sti[^/]*$|^drivers/video/logo/logo_parisc[^/]*$|^include/linux/hp_sdc\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/dec/tulip/\"},\n\t\t\t{IncludeRegexp: \"^sound/parisc/harmony\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^sound/pci/ad1889\\\\.[^/]*$\"},\n\t\t},\n\t}\n\n\tparport = Subsystem{\n\t\tName: \"parport\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/char/ppdev\\\\.c$|^drivers/parport/|^include/linux/parport[^/]*\\\\.h$|^include/uapi/linux/ppdev\\\\.h$\"},\n\t\t},\n\t}\n\n\tpci = Subsystem{\n\t\tName: \"pci\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/x86/kernel/early-quirks\\\\.c$|^arch/x86/kernel/quirks\\\\.c$|^arch/x86/pci/|^drivers/acpi/pci[^/]*$|^drivers/pci/|^include/asm-generic/pci[^/]*$|^include/linux/of_pci\\\\.h$|^include/linux/pci[^/]*$|^include/uapi/linux/pci[^/]*$|^lib/pci[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/misc/dw-xdata-pcie\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/misc/pci_endpoint_test\\\\.c$|^drivers/pci/endpoint/\"},\n\t\t\t{IncludeRegexp: \"^drivers/ntb/hw/mscc/|^drivers/pci/switch/switchtec[^/]*$|^include/linux/switchtec\\\\.h$|^include/uapi/linux/switchtec_ioctl\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/[^/]*mediatek[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/[^/]*microchip[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/[^/]*mvebu[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/[^/]*rcar[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/cadence/\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/cadence/pci-j721e\\\\.c$|^drivers/pci/controller/dwc/pci-dra7xx\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/[^/]*artpec[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/[^/]*designware[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/[^/]*imx6[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/[^/]*layerscape[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/[^/]*spear[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/pci-exynos\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/pci-meson\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/pcie-al\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/pcie-armada8k\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/pcie-fu740\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/pcie-hisi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/pcie-histb\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/pcie-intel-gw\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/pcie-keembay\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/pcie-kirin\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/pcie-qcom-ep\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/pcie-qcom\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/pcie-uniphier[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/mobiveil/pcie-layerscape-gen4\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/mobiveil/pcie-mobiveil[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pci-aardvark\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pci-host-common\\\\.c$|^drivers/pci/controller/pci-host-generic\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pci-tegra\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pci-thunder-[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pci-v3-semi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pci-versatile\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pci-xgene-msi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pci-xgene\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pcie-altera-msi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pcie-altera\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pcie-apple\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pcie-brcmstb\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pcie-rockchip[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pcie-xilinx-cpm\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/vmd\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/|^drivers/pci/pci-bridge-emul\\\\.c$|^drivers/pci/pci-bridge-emul\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/hotplug/cpci_hotplug[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/hotplug/cpcihp_generic\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/hotplug/cpcihp_zt5550\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/hotplug/rpadlpar[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/hotplug/rpaphp[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/p2pdma\\\\.c$|^include/linux/pci-p2pdma\\\\.h$\"},\n\t\t},\n\t}\n\n\tperf = Subsystem{\n\t\tName: \"perf\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/[^/]*/events/|^arch/[^/]*/events/[^/]*/|^arch/[^/]*/include/asm/perf_event\\\\.h$|^arch/[^/]*/kernel/[^/]*/[^/]*/perf_event[^/]*\\\\.c$|^arch/[^/]*/kernel/[^/]*/perf_event[^/]*\\\\.c$|^arch/[^/]*/kernel/perf_callchain\\\\.c$|^arch/[^/]*/kernel/perf_event[^/]*\\\\.c$|^include/linux/perf_event\\\\.h$|^include/uapi/linux/perf_event\\\\.h$|^kernel/events/\"},\n\t\t},\n\t}\n\n\tphy = Subsystem{\n\t\tName: \"phy\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/phy/|^include/dt-bindings/phy/|^include/linux/phy/\"},\n\t\t},\n\t}\n\n\tpm = Subsystem{\n\t\tName: \"pm\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/[^/]*/include/asm/suspend[^/]*\\\\.h$|^arch/x86/power/|^drivers/base/power/|^include/linux/freezer\\\\.h$|^include/linux/pm\\\\.h$|^include/linux/suspend\\\\.h$|^kernel/power/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/mach-exynos/pm\\\\.c$|^drivers/cpuidle/cpuidle-exynos\\\\.c$|^include/linux/platform_data/cpuidle-exynos\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/kernel/acpi/sleep[^/]*$|^arch/x86/kernel/acpi/wakeup[^/]*$|^drivers/base/power/|^include/linux/freezer\\\\.h$|^include/linux/pm\\\\.h$|^include/linux/suspend\\\\.h$|^kernel/power/\"},\n\t\t\t{IncludeRegexp: \"^drivers/base/power/domain[^/]*\\\\.c$|^include/linux/pm_domain\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/base/power/|^drivers/powercap/|^include/linux/intel_rapl\\\\.h$|^include/linux/pm\\\\.h$|^include/linux/pm_[^/]*$|^include/linux/powercap\\\\.h$|^kernel/configs/nopm\\\\.config$\"},\n\t\t\t{IncludeRegexp: \"^drivers/cpufreq/amd-pstate[^/]*$|^include/linux/amd-pstate\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/cpufreq/bmips-cpufreq\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/cpufreq/brcmstb[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/cpufreq/intel_pstate\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/cpufreq/qcom-cpufreq-nvmem\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/cpufreq/sun50i-cpufreq-nvmem\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/cpufreq/vexpress-spc-cpufreq\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/cpufreq/|^include/linux/cpufreq\\\\.h$|^include/linux/sched/cpufreq\\\\.h$|^kernel/sched/cpufreq[^/]*\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/cpuidle/cpuidle-big_little\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/cpuidle/cpuidle-psci\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/cpuidle/cpuidle-psci\\\\.h$|^drivers/cpuidle/cpuidle-psci-domain\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/cpuidle/cpuidle-riscv-sbi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/cpuidle/dt_idle_genpd\\\\.c$|^drivers/cpuidle/dt_idle_genpd\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/cpuidle/|^include/linux/cpuidle\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/devfreq/devfreq-event\\\\.c$|^drivers/devfreq/event/|^include/dt-bindings/pmu/exynos_ppmu\\\\.h$|^include/linux/devfreq-event\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/devfreq/exynos-bus\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/devfreq/tegra30-devfreq\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/devfreq/|^include/linux/devfreq\\\\.h$|^include/trace/events/devfreq\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/idle/intel_idle\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/interconnect/samsung/\"},\n\t\t\t{IncludeRegexp: \"^drivers/interconnect/|^include/dt-bindings/interconnect/|^include/linux/interconnect-provider\\\\.h$|^include/linux/interconnect\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/memory/samsung/exynos5422-dmc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/opp/|^include/linux/pm_opp\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/power/reset/\"},\n\t\t\t{IncludeRegexp: \"^drivers/power/reset/mt6323-poweroff\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/power/supply/adp5061\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/power/supply/max14577_charger\\\\.c$|^drivers/power/supply/max77693_charger\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/power/supply/max17040_battery\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/power/supply/max17042_battery\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/power/supply/surface_battery\\\\.c$|^drivers/power/supply/surface_charger\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/power/supply/|^include/linux/power/|^include/linux/power_supply\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/powercap/dtpm[^/]*$|^include/linux/dtpm\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/soc/bcm/bcm63xx/bcm-pmb\\\\.c$|^include/dt-bindings/soc/bcm-pmb\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/soc/qcom/cpr\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/soc/ti/smartreflex\\\\.c$|^include/linux/power/smartreflex\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/thermal/amlogic_thermal\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/thermal/broadcom/brcmstb[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/thermal/cpufreq_cooling\\\\.c$|^drivers/thermal/cpuidle_cooling\\\\.c$|^include/linux/cpu_cooling\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/thermal/gov_power_allocator\\\\.c$|^include/trace/events/thermal_power_allocator\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/thermal/intel/intel_menlow\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/thermal/qcom/\"},\n\t\t\t{IncludeRegexp: \"^drivers/thermal/samsung/\"},\n\t\t\t{IncludeRegexp: \"^drivers/thermal/sun8i_thermal\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/thermal/ti-soc-thermal/\"},\n\t\t\t{IncludeRegexp: \"^drivers/thermal/|^include/dt-bindings/thermal/|^include/linux/cpu_cooling\\\\.h$|^include/linux/thermal\\\\.h$|^include/uapi/linux/thermal\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/freezer\\\\.h$|^kernel/freezer\\\\.c$\"},\n\t\t},\n\t}\n\n\tppp = Subsystem{\n\t\tName: \"ppp\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/net/ppp/ppp_[^/]*$\"},\n\t\t},\n\t}\n\n\tpvrusb2 = Subsystem{\n\t\tName: \"pvrusb2\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&media},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/media/usb/pvrusb2/\"},\n\t\t},\n\t}\n\n\tpwm = Subsystem{\n\t\tName: \"pwm\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-mvebu\\\\.c$|^drivers/pwm/|^drivers/video/backlight/pwm_bl\\\\.c$|^include/dt-bindings/pwm/|^include/linux/pwm\\\\.h$|^include/linux/pwm_backlight\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pwm/pwm-atmel\\\\.c$\"},\n\t\t},\n\t}\n\n\tqat = Subsystem{\n\t\tName: \"qat\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&crypto},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/crypto/qat/\"},\n\t\t},\n\t}\n\n\traid = Subsystem{\n\t\tName: \"raid\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/md/Kconfig$|^drivers/md/md[^/]*$|^drivers/md/raid[^/]*$|^include/linux/raid/|^include/uapi/linux/raid/\"},\n\t\t},\n\t}\n\n\trcu = Subsystem{\n\t\tName: \"rcu\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{\"^include/linux/rcu[^/]*$|^kernel/rcu/\", \"^include/linux/srcu[^/]*\\\\.h$|^kernel/rcu/srcu[^/]*\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^include/linux/srcu[^/]*\\\\.h$|^kernel/rcu/srcu[^/]*\\\\.c$\"},\n\t\t},\n\t}\n\n\trdma = Subsystem{\n\t\tName: \"rdma\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/hw/bnxt_re/|^include/uapi/rdma/bnxt_re-abi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/hw/cxgb4/|^include/uapi/rdma/cxgb4-abi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/hw/efa/|^include/uapi/rdma/efa-abi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/hw/erdma$|^include/uapi/rdma/erdma-abi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/hw/hfi1$\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/hw/hns/\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/hw/irdma/|^include/uapi/rdma/irdma-abi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/hw/mana/|^include/net/mana$|^include/uapi/rdma/mana-abi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/hw/mlx4/|^include/linux/mlx4/|^include/uapi/rdma/mlx4-abi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/hw/mlx5/|^include/linux/mlx5/|^include/uapi/rdma/mlx5-abi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/hw/ocrdma/|^include/uapi/rdma/ocrdma-abi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/hw/qedr/|^include/uapi/rdma/qedr-abi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/hw/qib/\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/hw/vmw_pvrdma/\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/sw/rdmavt$\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/sw/rxe/|^include/uapi/rdma/rdma_user_rxe\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/sw/siw/|^include/uapi/rdma/siw-abi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/ulp/iser/\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/ulp/isert$\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/ulp/opa_vnic$\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/ulp/rtrs/\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/ulp/srp/|^include/scsi/srp\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/ulp/srpt/\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/|^include/rdma/|^include/trace/events/ib_mad\\\\.h$|^include/trace/events/ib_umad\\\\.h$|^include/trace/misc/rdma\\\\.h$|^include/uapi/linux/if_infiniband\\\\.h$|^include/uapi/rdma/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/mellanox/mlx4/|^include/linux/mlx4/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/mellanox/mlx5/core/|^include/linux/mlx5/\"},\n\t\t\t{IncludeRegexp: \"^net/rds/\"},\n\t\t},\n\t}\n\n\trds = Subsystem{\n\t\tName: \"rds\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&net, &rdma},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^net/rds/\"},\n\t\t},\n\t}\n\n\treiserfs = Subsystem{\n\t\tName: \"reiserfs\",\n\t\tSyscalls: []string{\"syz_mount_image$reiserfs\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/reiserfs/\"},\n\t\t},\n\t}\n\n\tremoteproc = Subsystem{\n\t\tName: \"remoteproc\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/hwspinlock/|^include/linux/hwspinlock\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wwan/rpmsg_wwan_ctrl\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/remoteproc/|^include/linux/remoteproc\\\\.h$|^include/linux/remoteproc/\"},\n\t\t\t{IncludeRegexp: \"^drivers/rpmsg/|^include/linux/rpmsg\\\\.h$|^include/linux/rpmsg/|^include/uapi/linux/rpmsg\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/tty/rpmsg_tty\\\\.c$\"},\n\t\t},\n\t}\n\n\trenesassoc = Subsystem{\n\t\tName: \"renesas-soc\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/emev2[^/]*$|^arch/arm/boot/dts/gr-peach[^/]*$|^arch/arm/boot/dts/iwg20d-q7[^/]*$|^arch/arm/boot/dts/r7s[^/]*$|^arch/arm/boot/dts/r8a[^/]*$|^arch/arm/boot/dts/r9a[^/]*$|^arch/arm/boot/dts/sh[^/]*$|^arch/arm/configs/shmobile_defconfig$|^arch/arm/include/debug/renesas-scif\\\\.S$|^arch/arm/mach-shmobile/|^arch/arm64/boot/dts/renesas/|^arch/riscv/boot/dts/renesas/|^drivers/soc/renesas/|^include/linux/soc/renesas/\"},\n\t\t\t{IncludeRegexp: \"^drivers/ata/sata_rcar\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/clk/renesas/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-bd9571mwv\\\\.c$|^drivers/mfd/bd9571mwv\\\\.c$|^drivers/regulator/bd9571mwv-regulator\\\\.c$|^include/linux/mfd/bd9571mwv\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/rcar-du/|^drivers/gpu/drm/shmobile/|^include/linux/platform_data/shmob_drm\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-emev2\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-rcar\\\\.c$|^drivers/i2c/busses/i2c-sh_mobile\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-riic\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iio/adc/rzg2l_adc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/renesas/rcar-fcp\\\\.c$|^include/media/rcar-fcp\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/renesas/rcar-isp\\\\.c$|^drivers/media/platform/renesas/rcar-vin/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/renesas/rcar_drif\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/renesas/rcar_fdp1\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/renesas/rcar_jpu\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/renesas/renesas-ceu\\\\.c$|^include/media/drv-intf/renesas-ceu\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/renesas/vsp1/\"},\n\t\t\t{IncludeRegexp: \"^drivers/mmc/host/renesas_sdhi[^/]*$|^drivers/mmc/host/tmio_mmc[^/]*$|^include/linux/mfd/tmio\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/mtd/nand/raw/renesas-nand-controller\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/dsa/rzn1_a5psw[^/]*$|^drivers/net/pcs/pcs-rzn1-miic\\\\.c$|^include/dt-bindings/net/pcs-rzn1-miic\\\\.h$|^include/linux/pcs-rzn1-miic\\\\.h$|^net/dsa/tag_rzn1_a5psw\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ethernet/renesas/|^include/linux/sh_eth\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/[^/]*rcar[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/phy/renesas/phy-rcar-gen3-usb[^/]*\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pinctrl/renesas/\"},\n\t\t\t{IncludeRegexp: \"^drivers/rtc/rtc-rzn1\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/thermal/rcar_gen3_thermal\\\\.c$|^drivers/thermal/rcar_thermal\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/ufs/host/ufs-renesas\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/gadget/udc/renesas_usbf\\\\.c$\"},\n\t\t},\n\t}\n\n\triscv = Subsystem{\n\t\tName: \"riscv\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/riscv/boot/dts/\"},\n\t\t\t{IncludeRegexp: \"^arch/riscv/boot/dts/microchip/|^drivers/char/hw_random/mpfs-rng\\\\.c$|^drivers/clk/microchip/clk-mpfs[^/]*\\\\.c$|^drivers/i2c/busses/i2c-microchip-corei2c\\\\.c$|^drivers/mailbox/mailbox-mpfs\\\\.c$|^drivers/pci/controller/pcie-microchip-host\\\\.c$|^drivers/reset/reset-mpfs\\\\.c$|^drivers/rtc/rtc-mpfs\\\\.c$|^drivers/soc/microchip/mpfs-sys-controller\\\\.c$|^drivers/spi/spi-microchip-core-qspi\\\\.c$|^drivers/spi/spi-microchip-core\\\\.c$|^drivers/usb/musb/mpfs\\\\.c$|^include/soc/microchip/mpfs\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^arch/riscv/include/asm/kvm[^/]*$|^arch/riscv/include/uapi/asm/kvm[^/]*$|^arch/riscv/kvm/\"},\n\t\t\t{IncludeRegexp: \"^arch/riscv/|riscv\"},\n\t\t\t{IncludeRegexp: \"^drivers/cpuidle/cpuidle-riscv-sbi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/perf/riscv_pmu\\\\.c$|^drivers/perf/riscv_pmu_legacy\\\\.c$|^drivers/perf/riscv_pmu_sbi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pinctrl/pinctrl-k210\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/reset/reset-k210\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/soc/canaan/|^include/soc/canaan/\"},\n\t\t\t{IncludeRegexp: \"^drivers/soc/sifive/\"},\n\t\t\t{IncludeRegexp: \"fu540\"},\n\t\t\t{IncludeRegexp: \"sifive\"},\n\t\t},\n\t}\n\n\trockchip = Subsystem{\n\t\tName: \"rockchip\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&arm},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/rk3[^/]*$|^arch/arm/boot/dts/rv11[^/]*$|^arch/arm/mach-rockchip/|^drivers/[^/]*/[^/]*/[^/]*rockchip[^/]*$|^drivers/[^/]*/[^/]*rockchip[^/]*$|^drivers/clk/rockchip/|^drivers/i2c/busses/i2c-rk3x\\\\.c$|^sound/soc/rockchip/|rockchip\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/rockchip/rga/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/rockchip/rkisp1$|^include/uapi/linux/rkisp1-config\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/verisilicon/\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pcie-rockchip[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/staging/media/rkvdec/\"},\n\t\t\t{IncludeRegexp: \"^sound/soc/rockchip/rockchip_i2s_tdm\\\\.[^/]*$\"},\n\t\t},\n\t}\n\n\trpi = Subsystem{\n\t\tName: \"rpi\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&arm},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pcie-brcmstb\\\\.c$|^drivers/staging/vc04_services$|bcm2711|bcm283*|raspberrypi\"},\n\t\t},\n\t}\n\n\trttools = Subsystem{\n\t\tName: \"rt-tools\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&trace},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^include/linux/rv\\\\.h$|^include/rv/|^kernel/trace/rv/\"},\n\t\t},\n\t}\n\n\trtc = Subsystem{\n\t\tName: \"rtc\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/rtc/rtc-optee\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/rtc/rtc-rzn1\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/rtc/rtc-sd3078\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/rtc/rtc-sunplus\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/rtc/|^include/linux/platform_data/rtc-[^/]*$|^include/linux/rtc\\\\.h$|^include/linux/rtc/|^include/uapi/linux/rtc\\\\.h$\"},\n\t\t},\n\t}\n\n\trust = Subsystem{\n\t\tName: \"rust\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^rust/\"},\n\t\t},\n\t}\n\n\ts390 = Subsystem{\n\t\tName: \"s390\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/s390/include/asm/pgtable\\\\.h$|^arch/s390/mm$\"},\n\t\t\t{IncludeRegexp: \"^arch/s390/include/uapi/asm/virtio-ccw\\\\.h$|^drivers/s390/virtio/\"},\n\t\t\t{IncludeRegexp: \"^arch/s390/kvm/pci[^/]*$|^drivers/vfio/pci/vfio_pci_zdev\\\\.c$|^include/uapi/linux/vfio_zdev\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^arch/s390/pci/|^drivers/pci/hotplug/s390_pci_hpc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/s390/|^drivers/s390/|^drivers/watchdog/diag288_wdt\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^block/partitions/ibm\\\\.c$|^drivers/s390/block/dasd[^/]*$|^include/linux/dasd_mod\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iommu/s390-iommu\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/s390/block/scm[^/]*$|^drivers/s390/cio/scm\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/s390/cio/\"},\n\t\t\t{IncludeRegexp: \"^drivers/s390/cio/vfio_ccw[^/]*$|^include/uapi/linux/vfio_ccw\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/s390/crypto/\"},\n\t\t\t{IncludeRegexp: \"^drivers/s390/crypto/vfio_ap[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/s390/net/\"},\n\t\t\t{IncludeRegexp: \"^drivers/s390/net/[^/]*iucv[^/]*$|^include/net/iucv/|^net/iucv/\"},\n\t\t\t{IncludeRegexp: \"^drivers/s390/scsi/zfcp_[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^net/smc/\"},\n\t\t},\n\t}\n\n\tsamsungsoc = Subsystem{\n\t\tName: \"samsung-soc\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&arm},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/exynos[^/]*$|^arch/arm/boot/dts/s3c[^/]*$|^arch/arm/boot/dts/s5p[^/]*$|^arch/arm/mach-exynos[^/]*/|^arch/arm/mach-s3c/|^arch/arm/mach-s5p[^/]*/|^arch/arm64/boot/dts/exynos/|^drivers/[^/]*/[^/]*/[^/]*s3c24[^/]*$|^drivers/[^/]*/[^/]*s3c24[^/]*$|^drivers/[^/]*/[^/]*s3c64xx[^/]*$|^drivers/[^/]*/[^/]*s5pv210[^/]*$|^drivers/clocksource/samsung_pwm_timer\\\\.c$|^drivers/memory/samsung/|^drivers/pwm/pwm-samsung\\\\.c$|^drivers/soc/samsung/|^drivers/tty/serial/samsung[^/]*$|^include/clocksource/samsung_pwm\\\\.h$|^include/linux/platform_data/[^/]*s3c[^/]*$|^include/linux/serial_s3c\\\\.h$|^include/linux/soc/samsung/|exynos|s3c64xx|s5pv210\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/mach-exynos/pm\\\\.c$|^drivers/cpuidle/cpuidle-exynos\\\\.c$|^include/linux/platform_data/cpuidle-exynos\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm64/boot/dts/tesla/\"},\n\t\t\t{IncludeRegexp: \"^drivers/char/hw_random/exynos-trng\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/clk/clk-s2mps11\\\\.c$|^drivers/mfd/sec[^/]*\\\\.c$|^drivers/regulator/s2m[^/]*\\\\.c$|^drivers/regulator/s5m[^/]*\\\\.c$|^drivers/rtc/rtc-s5m\\\\.c$|^include/linux/mfd/samsung/\"},\n\t\t\t{IncludeRegexp: \"^drivers/clk/samsung/|^include/dt-bindings/clock/exynos[^/]*\\\\.h$|^include/dt-bindings/clock/s5p[^/]*\\\\.h$|^include/dt-bindings/clock/samsung,[^/]*\\\\.h$|^include/linux/clk/samsung\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/exynos-rng\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/s5p-sss\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/devfreq/exynos-bus\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/interconnect/samsung/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/cec/platform/s5p/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/samsung/s3c-camif/|^include/media/drv-intf/s3c_camif\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/memory/samsung/exynos5422-dmc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/dwc/pci-exynos\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/pinctrl/samsung/|^include/dt-bindings/pinctrl/samsung\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/spi/spi-s3c[^/]*$|^include/linux/platform_data/spi-s3c64xx\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/thermal/samsung/\"},\n\t\t},\n\t}\n\n\tscsi = Subsystem{\n\t\tName: \"scsi\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^block/bsg\\\\.c$|^include/linux/bsg\\\\.h$|^include/uapi/linux/bsg\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/message/fusion/|^drivers/scsi/mpt3sas/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/3w-[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/53c700[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/53c700[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/BusLogic\\\\.[^/]*$|^drivers/scsi/FlashPoint\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/NCR5380\\\\.[^/]*$|^drivers/scsi/arm/cumana_1\\\\.c$|^drivers/scsi/arm/oak\\\\.c$|^drivers/scsi/atari_scsi\\\\.[^/]*$|^drivers/scsi/dmx3191d\\\\.c$|^drivers/scsi/g_NCR5380\\\\.[^/]*$|^drivers/scsi/mac_scsi\\\\.[^/]*$|^drivers/scsi/sun3_scsi\\\\.[^/]*$|^drivers/scsi/sun3_scsi_vme\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/[^/]*iscsi[^/]*$|^include/scsi/[^/]*iscsi[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/aacraid/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/advansys\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/aha152x[^/]*$|^drivers/scsi/pcmcia/aha152x[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/aic7xxx/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/am53c974\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/be2iscsi/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/bfa/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/bnx2fc/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/bnx2i/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/cxgbi/cxgb3i$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/cxgbi/cxgb4i$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/cxlflash/|^include/uapi/scsi/cxlflash_ioctl\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/elx/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/esas2r$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/fcoe/|^drivers/scsi/libfc/|^include/scsi/fc/|^include/scsi/libfc\\\\.h$|^include/scsi/libfcoe\\\\.h$|^include/uapi/scsi/fc/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/fnic/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/hpsa[^/]*\\\\.\\\\[ch\\\\]$|^include/linux/cciss[^/]*\\\\.h$|^include/uapi/linux/cciss[^/]*\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/ibmvscsi/ibmvfc[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/ibmvscsi/ibmvscsi[^/]*$|^include/scsi/viosrp\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/ibmvscsi_tgt/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/ips[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/isci/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/lpfc/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/megaraid\\\\.[^/]*$|^drivers/scsi/megaraid/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/mpi3mr/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/myrb\\\\.[^/]*$|^drivers/scsi/myrs\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/pm8001/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/pmcraid\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/qedf/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/qedi/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/qla1280\\\\.\\\\[ch\\\\]$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/qla2xxx/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/qla4xxx/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/sg\\\\.c$|^include/scsi/sg\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/smartpqi/Kconfig$|^drivers/scsi/smartpqi/smartpqi[^/]*\\\\.\\\\[ch\\\\]$|^include/linux/cciss[^/]*\\\\.h$|^include/uapi/linux/cciss[^/]*\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/snic/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/st\\\\.[^/]*$|^drivers/scsi/st_[^/]*\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/sym53c8xx_2/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/vmw_pvscsi\\\\.c$|^drivers/scsi/vmw_pvscsi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/xen-scsifront\\\\.c$|^drivers/xen/xen-scsiback\\\\.c$|^include/xen/interface/io/vscsiif\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/|^drivers/ufs/|^include/scsi/\"},\n\t\t\t{IncludeRegexp: \"^drivers/target/sbp/\"},\n\t\t\t{IncludeRegexp: \"^drivers/target/target_core_user\\\\.c$|^include/uapi/linux/target_core_user\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/target/|^include/target/\"},\n\t\t\t{IncludeRegexp: \"^drivers/ufs/core/\"},\n\t\t\t{IncludeRegexp: \"^drivers/ufs/host/[^/]*dwc[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/ufs/host/ufs-exynos[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/ufs/host/ufs-mediatek[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/ufs/host/ufs-qcom[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/ufs/host/ufs-renesas\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/storage/uas\\\\.c$\"},\n\t\t},\n\t}\n\n\tsctp = Subsystem{\n\t\tName: \"sctp\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^include/linux/sctp\\\\.h$|^include/net/sctp/|^include/uapi/linux/sctp\\\\.h$|^net/sctp/\"},\n\t\t},\n\t}\n\n\tselinux = Subsystem{\n\t\tName: \"selinux\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^include/trace/events/avc\\\\.h$|^include/uapi/linux/selinux_netlink\\\\.h$|^security/selinux/\"},\n\t\t},\n\t}\n\n\tserial = Subsystem{\n\t\tName: \"serial\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/tty/serdev/|^include/linux/serdev\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/tty/serial/\"},\n\t\t\t{IncludeRegexp: \"^drivers/tty/serial/8250/8250_bcm7271\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/tty/serial/8250/8250_pci1xxxx\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/tty/serial/8250[^/]*$|^include/linux/serial_8250\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/tty/serial/altera_jtaguart\\\\.c$|^drivers/tty/serial/altera_uart\\\\.c$|^include/linux/altera_jtaguart\\\\.h$|^include/linux/altera_uart\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/tty/serial/jsm/\"},\n\t\t\t{IncludeRegexp: \"^drivers/tty/serial/rp2\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/tty/serial/uartlite\\\\.c$\"},\n\t\t},\n\t}\n\n\tsgx = Subsystem{\n\t\tName: \"sgx\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/x86/entry/vdso/vsgx\\\\.S$|^arch/x86/include/asm/sgx\\\\.h$|^arch/x86/include/uapi/asm/sgx\\\\.h$|^arch/x86/kernel/cpu/sgx/\"},\n\t\t},\n\t}\n\n\tsh = Subsystem{\n\t\tName: \"sh\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/sh/|^drivers/sh/\"},\n\t\t},\n\t}\n\n\tsnpsarc = Subsystem{\n\t\tName: \"snps-arc\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/arc/|^drivers/clocksource/arc_timer\\\\.c$|^drivers/tty/serial/arc_uart\\\\.c$\"},\n\t\t},\n\t}\n\n\tsof = Subsystem{\n\t\tName: \"sof\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\", \"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&alsa},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^sound/soc/sof/\"},\n\t\t},\n\t}\n\n\tsparclinux = Subsystem{\n\t\tName: \"sparclinux\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/sparc/|^drivers/sbus/\"},\n\t\t\t{IncludeRegexp: \"^drivers/tty/serial/suncore\\\\.c$|^drivers/tty/serial/sunhv\\\\.c$|^drivers/tty/serial/sunsab\\\\.c$|^drivers/tty/serial/sunsab\\\\.h$|^drivers/tty/serial/sunsu\\\\.c$|^drivers/tty/serial/sunzilog\\\\.c$|^drivers/tty/serial/sunzilog\\\\.h$|^drivers/tty/vcc\\\\.c$|^include/linux/sunserialcore\\\\.h$\"},\n\t\t},\n\t}\n\n\tspeakup = Subsystem{\n\t\tName: \"speakup\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/accessibility/speakup/\"},\n\t\t},\n\t}\n\n\tspi = Subsystem{\n\t\tName: \"spi\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/spi/spi-aspeed-smc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/spi/spi-at91-usart\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/spi/spi-bcm63xx-hsspi\\\\.c$|^drivers/spi/spi-bcmbca-hsspi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/spi/spi-dw[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/spi/spi-fsi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/spi/spi-fsl-dspi\\\\.c$|^include/linux/spi/spi-fsl-dspi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/spi/spi-fsl-qspi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/spi/spi-hisi-kunpeng\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/spi/spi-nxp-fspi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/spi/spi-s3c[^/]*$|^include/linux/platform_data/spi-s3c64xx\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/spi/spi-stm32\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/spi/spi-sunplus-sp7021\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/spi/spi-synquacer\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/spi/|^include/linux/spi/|^include/uapi/linux/spi/\"},\n\t\t},\n\t}\n\n\tspice = Subsystem{\n\t\tName: \"spice\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&dri, &virt},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/qxl/|^include/uapi/drm/qxl_drm\\\\.h$\"},\n\t\t},\n\t}\n\n\tsquashfs = Subsystem{\n\t\tName: \"squashfs\",\n\t\tSyscalls: []string{\"syz_mount_image$squashfs\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/squashfs/\"},\n\t\t},\n\t}\n\n\tstaging = Subsystem{\n\t\tName: \"staging\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&media},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/staging/\"},\n\t\t},\n\t}\n\n\tstm32 = Subsystem{\n\t\tName: \"stm32\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&arm},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/stm32[^/]*$|^arch/arm/mach-stm32/|^drivers/clocksource/armv7m_systick\\\\.c$|stm32|stm\"},\n\t\t},\n\t}\n\n\tsunxi = Subsystem{\n\t\tName: \"sunxi\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&arm},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/arm/mach-sunxi/|^arch/arm64/boot/dts/allwinner/|^drivers/clk/sunxi-ng/|^drivers/pinctrl/sunxi/|^drivers/soc/sunxi/|allwinner|sun[x456789]i|sun[25]0i\"},\n\t\t},\n\t}\n\n\ttarget = Subsystem{\n\t\tName: \"target\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&scsi},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/ulp/isert$\"},\n\t\t\t{IncludeRegexp: \"^drivers/infiniband/ulp/srpt/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/elx/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/ibmvscsi_tgt/\"},\n\t\t\t{IncludeRegexp: \"^drivers/target/sbp/\"},\n\t\t\t{IncludeRegexp: \"^drivers/target/target_core_user\\\\.c$|^include/uapi/linux/target_core_user\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/target/|^include/target/\"},\n\t\t},\n\t}\n\n\ttegra = Subsystem{\n\t\tName: \"tegra\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"[^a-z]tegra\"},\n\t\t\t{IncludeRegexp: \"^drivers/devfreq/tegra30-devfreq\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/tegra/|^drivers/gpu/host1x/|^include/linux/host1x\\\\.h$|^include/uapi/drm/tegra_drm\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iommu/arm/arm-smmu/arm-smmu-nvidia\\\\.c$|^drivers/iommu/tegra[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/cec/platform/tegra/\"},\n\t\t\t{IncludeRegexp: \"^drivers/media/platform/nvidia/tegra-vde/\"},\n\t\t\t{IncludeRegexp: \"^drivers/pci/controller/pci-tegra\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/spi/spi-tegra210-quad\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/staging/media/tegra-video/\"},\n\t\t\t{IncludeRegexp: \"^drivers/staging/nvec/\"},\n\t\t},\n\t}\n\n\ttipc = Subsystem{\n\t\tName: \"tipc\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^include/uapi/linux/tipc[^/]*\\\\.h$|^net/tipc/\"},\n\t\t},\n\t}\n\n\ttomoyo = Subsystem{\n\t\tName: \"tomoyo\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&lsm},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^security/tomoyo/\"},\n\t\t},\n\t}\n\n\ttrace = Subsystem{\n\t\tName: \"trace\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/proc/bootconfig\\\\.c$|^include/linux/bootconfig\\\\.h$|^lib/bootconfig-data\\\\.S$|^lib/bootconfig\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^fs/tracefs/|^include/linux/trace[^/]*\\\\.h$|^include/trace/|^kernel/trace/\"},\n\t\t\t{IncludeRegexp: \"^include/asm-generic/kprobes\\\\.h$|^include/linux/kprobes\\\\.h$|^kernel/kprobes\\\\.c$|^lib/test_kprobes\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^kernel/trace/ftrace[^/]*$|^kernel/trace/fgraph\\\\.c$|^arch/[^/]*/[^/]*/[^/]*/[^/]*ftrace[^/]*$|^arch/[^/]*/[^/]*/[^/]*ftrace[^/]*$|^include/[^/]*/ftrace\\\\.h$\"},\n\t\t},\n\t}\n\n\tuclinux = Subsystem{\n\t\tName: \"uclinux\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&m68k},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/m68k/[^/]*/[^/]*_no\\\\.[^/]*$|^arch/m68k/68[^/]*/|^arch/m68k/coldfire/|^arch/m68k/include/asm/[^/]*_no\\\\.[^/]*$\"},\n\t\t},\n\t}\n\n\tudf = Subsystem{\n\t\tName: \"udf\",\n\t\tSyscalls: []string{\"syz_mount_image$udf\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/udf/\"},\n\t\t},\n\t}\n\n\tum = Subsystem{\n\t\tName: \"um\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/um/|^arch/x86/um/|^fs/hostfs/\"},\n\t\t},\n\t}\n\n\tunisoc = Subsystem{\n\t\tName: \"unisoc\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&arm},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/arm/boot/dts/rda8810pl-[^/]*$|^drivers/clocksource/timer-rda\\\\.c$|^drivers/gpio/gpio-rda\\\\.c$|^drivers/irqchip/irq-rda-intc\\\\.c$|^drivers/tty/serial/rda-uart\\\\.c$\"},\n\t\t},\n\t}\n\n\tusb = Subsystem{\n\t\tName: \"usb\",\n\t\tSyscalls: []string{\"syz_usb_connect\", \"syz_usb_connect$hid\", \"syz_usb_connect$printer\", \"syz_usb_connect$cdc_ecm\", \"syz_usb_connect$cdc_ncm\", \"syz_usb_connect$uac1\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/arm/[^/]*omap[^/]*/usb[^/]*$|^drivers/usb/[^/]*/[^/]*omap[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/usbhid/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/usb/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/usb/cdc_[^/]*\\\\.c$|^include/uapi/linux/usb/cdc\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/usb/hso\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/usb/pegasus\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/usb/rtl8150\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/phy/hisilicon/phy-hi3660-usb3\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/phy/hisilicon/phy-hi3670-usb3\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/phy/sunplus/Kconfig$|^drivers/phy/sunplus/phy-sunplus-usb2\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/thunderbolt/dma_test\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/thunderbolt/|^include/linux/thunderbolt\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/atm/speedtch\\\\.c$|^drivers/usb/atm/usbatm\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/c67x00/\"},\n\t\t\t{\"^drivers/usb/cdns3/\", \"^drivers/usb/cdns3/cdns3[^/]*$\"},\n\t\t\t{\"^drivers/usb/cdns3/\", \"^drivers/usb/cdns3/cdnsp[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/chipidea/\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/class/cdc-acm\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/class/usblp\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/common/ulpi\\\\.c$|^include/linux/ulpi/\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/common/usb-otg-fsm\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/dwc2/\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/dwc3/\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/fotg210/\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/gadget/function/[^/]*uvc[^/]*$|^drivers/usb/gadget/legacy/webcam\\\\.c$|^include/uapi/linux/usb/g_uvc\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/gadget/legacy/raw_gadget\\\\.c$|^include/uapi/linux/usb/raw_gadget\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/gadget/udc/bcm63xx_udc\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/gadget/udc/bdc/\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/gadget/udc/fsl[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/gadget/udc/renesas_usbf\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/host/ehci-brcm\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/host/ehci[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/host/isp116x[^/]*$|^include/linux/usb/isp116x\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/host/ohci[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/host/pci-quirks[^/]*$|^drivers/usb/host/xhci[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/host/uhci[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/host/xen[^/]*$|^include/xen/interface/io/usbif\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/host/xhci-mtk[^/]*$|^drivers/usb/mtu3/\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/isp1760/\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/misc/apple-mfi-fastcharge\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/misc/brcmstb-usb-pinmap\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/misc/chaoskey\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/misc/onboard_usb_hub\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/misc/usb251xb\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/musb/\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/phy/phy-fsl-usb[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/roles/intel-xhci-usb-role-switch\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/serial/|^include/linux/usb/serial\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/storage/\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/storage/uas\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/typec/altmodes/|^include/linux/usb/typec_altmode\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/typec/mux/intel_pmc_mux\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/typec/mux/pi3usb30532\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/typec/tcpm/\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/typec/|^include/linux/usb/typec\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/usbip/\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/|^include/dt-bindings/usb/|^include/linux/usb\\\\.h$|^include/linux/usb/\"},\n\t\t},\n\t}\n\n\tusbstorage = Subsystem{\n\t\tName: \"usb-storage\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&usb},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/usb/storage/\"},\n\t\t},\n\t}\n\n\tvideo = Subsystem{\n\t\tName: \"video\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/x86/boot/video[^/]*$\"},\n\t\t},\n\t}\n\n\tvirt = Subsystem{\n\t\tName: \"virt\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/[^/]*/include/asm/paravirt[^/]*\\\\.h$|^arch/[^/]*/kernel/paravirt[^/]*$|^include/linux/hypervisor\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^arch/s390/include/uapi/asm/virtio-ccw\\\\.h$|^drivers/s390/virtio/\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/include/asm/vmware\\\\.h$|^arch/x86/kernel/cpu/vmware\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/block/virtio_blk\\\\.c$|^drivers/crypto/virtio/|^drivers/net/virtio_net\\\\.c$|^drivers/vdpa/|^drivers/virtio/|^include/linux/vdpa\\\\.h$|^include/linux/virtio[^/]*\\\\.h$|^include/uapi/linux/virtio_[^/]*\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/block/virtio_blk\\\\.c$|^drivers/scsi/virtio_scsi\\\\.c$|^drivers/vhost/scsi\\\\.c$|^include/uapi/linux/virtio_blk\\\\.h$|^include/uapi/linux/virtio_scsi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/char/virtio_console\\\\.c$|^include/linux/virtio_console\\\\.h$|^include/uapi/linux/virtio_console\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/crypto/virtio/|^include/uapi/linux/virtio_crypto\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpio/gpio-virtio\\\\.c$|^include/uapi/linux/virtio_gpio\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/qxl/|^include/uapi/drm/qxl_drm\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/tiny/bochs\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/tiny/cirrus\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/virtio/|^include/uapi/linux/virtio_gpu\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/i2c/busses/i2c-virtio\\\\.c$|^include/uapi/linux/virtio_i2c\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/iommu/virtio-iommu\\\\.c$|^include/uapi/linux/virtio_iommu\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/vsockmon\\\\.c$|^include/net/af_vsock\\\\.h$|^include/uapi/linux/vm_sockets\\\\.h$|^include/uapi/linux/vm_sockets_diag\\\\.h$|^include/uapi/linux/vsockmon\\\\.h$|^net/vmw_vsock/\"},\n\t\t\t{IncludeRegexp: \"^drivers/nvdimm/virtio_pmem\\\\.c$|^drivers/nvdimm/nd_virtio\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/vhost/vsock\\\\.c$|^include/linux/virtio_vsock\\\\.h$|^include/uapi/linux/virtio_vsock\\\\.h$|^net/vmw_vsock/virtio_transport\\\\.c$|^net/vmw_vsock/virtio_transport_common\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/vhost/|^include/linux/vhost_iotlb\\\\.h$|^include/uapi/linux/vhost\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/virtio/virtio_balloon\\\\.c$|^include/uapi/linux/virtio_balloon\\\\.h$|^include/linux/balloon_compaction\\\\.h$|^mm/balloon_compaction\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/virtio/virtio_mem\\\\.c$|^include/uapi/linux/virtio_mem\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^fs/fuse/virtio_fs\\\\.c$|^include/uapi/linux/virtio_fs\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^include/uapi/linux/virtio_snd\\\\.h$|^sound/virtio/\"},\n\t\t},\n\t}\n\n\twatchdog = Subsystem{\n\t\tName: \"watchdog\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/watchdog/ebc-c384_wdt\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/watchdog/mena21_wdt\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/watchdog/menz69_wdt\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/watchdog/realtek_otto_wdt\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/watchdog/sunplus_wdt\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/watchdog/|^include/linux/watchdog\\\\.h$|^include/uapi/linux/watchdog\\\\.h$|^include/trace/events/watchdog\\\\.h$\"},\n\t\t},\n\t}\n\n\twcn36xx = Subsystem{\n\t\tName: \"wcn36xx\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&wireless},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/ath/wcn36xx/\"},\n\t\t},\n\t}\n\n\twireguard = Subsystem{\n\t\tName: \"wireguard\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/net/wireguard/\"},\n\t\t},\n\t}\n\n\twireless = Subsystem{\n\t\tName: \"wireless\",\n\t\tSyscalls: []string{\"syz_80211_join_ibss\", \"syz_80211_inject_frame\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/bcma/|^include/linux/bcma/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/admtek/adm8211\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/ath/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/ath/ar5523/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/ath/ath5k/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/ath/ath6kl/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/ath/ath9k/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/ath/carl9170/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/ath/wil6210/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/atmel/atmel[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/broadcom/b43/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/broadcom/b43legacy/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/broadcom/brcm80211/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/intel/ipw2x00/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/intel/iwlegacy/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/intel/iwlwifi/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/intersil/hostap/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/intersil/orinoco/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/intersil/p54/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/mac80211_hwsim\\\\.\\\\[ch\\\\]$|^include/net/mac80211\\\\.h$|^net/mac80211/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/marvell/mwifiex/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/marvell/mwl8k\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/mediatek/mt76/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/mediatek/mt7601u/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/microchip/wilc1000/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/purelifi/plfxlc/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/quantenna$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/ralink/rt2x00/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/ray[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/realtek/rtl818x/rtl8180/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/realtek/rtl818x/rtl8187/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/realtek/rtl8xxxu/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/realtek/rtlwifi/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/realtek/rtw88/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/realtek/rtw89/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/rndis_wlan\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/rsi/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/ti/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/wl3501[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/zydas/zd1201\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/zydas/zd1211rw/\"},\n\t\t\t{IncludeRegexp: \"^drivers/nfc/trf7970a\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/ssb/|^include/linux/ssb/\"},\n\t\t\t{IncludeRegexp: \"^include/linux/ieee80211\\\\.h$|^include/net/cfg80211\\\\.h$|^include/net/ieee80211_radiotap\\\\.h$|^include/net/iw_handler\\\\.h$|^include/net/wext\\\\.h$|^include/uapi/linux/nl80211\\\\.h$|^include/uapi/linux/wireless\\\\.h$|^net/wireless/\"},\n\t\t\t{IncludeRegexp: \"^include/linux/rfkill\\\\.h$|^include/uapi/linux/rfkill\\\\.h$|^net/rfkill/\"},\n\t\t},\n\t}\n\n\twpan = Subsystem{\n\t\tName: \"wpan\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/net/ieee802154/adf7242\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ieee802154/at86rf230\\\\.h$|^drivers/net/ieee802154/atusb\\\\.c$|^drivers/net/ieee802154/atusb\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ieee802154/ca8210\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ieee802154/cc2520\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ieee802154/mcr20a\\\\.c$|^drivers/net/ieee802154/mcr20a\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ieee802154/mrf24j40\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/ieee802154/|^include/linux/ieee802154\\\\.h$|^include/linux/nl802154\\\\.h$|^include/net/af_ieee802154\\\\.h$|^include/net/cfg802154\\\\.h$|^include/net/ieee802154_netdev\\\\.h$|^include/net/mac802154\\\\.h$|^include/net/nl802154\\\\.h$|^net/ieee802154/|^net/mac802154/\"},\n\t\t\t{IncludeRegexp: \"^include/net/6lowpan\\\\.h$|^net/6lowpan/\"},\n\t\t},\n\t}\n\n\tx25 = Subsystem{\n\t\tName: \"x25\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&net},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/net/wan/hdlc_x25\\\\.c$|^drivers/net/wan/lapbether\\\\.c$|^include/[^/]*/lapb\\\\.h$|^include/net/x25[^/]*$|^include/uapi/linux/x25\\\\.h$|^net/lapb/|^net/x25/\"},\n\t\t},\n\t}\n\n\tx86 = Subsystem{\n\t\tName: \"x86\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&virt},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/[^/]*/include/asm/paravirt[^/]*\\\\.h$|^arch/[^/]*/kernel/paravirt[^/]*$|^include/linux/hypervisor\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/include/asm/vmware\\\\.h$|^arch/x86/kernel/cpu/vmware\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/platform$\"},\n\t\t},\n\t}\n\n\tx86drivers = Subsystem{\n\t\tName: \"x86-drivers\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/x86/include/asm/amd_hsmp\\\\.h$|^arch/x86/include/uapi/asm/amd_hsmp\\\\.h$|^drivers/platform/x86/amd/hsmp\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/include/asm/intel_punit_ipc\\\\.h$|^drivers/platform/x86/intel/punit_ipc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/include/asm/intel_telemetry\\\\.h$|^drivers/platform/x86/intel/telemetry/\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/platform$\"},\n\t\t\t{IncludeRegexp: \"^drivers/char/sonypi\\\\.c$|^drivers/platform/x86/sony-laptop\\\\.c$|^include/linux/sony-laptop\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/hid/surface-hid/\"},\n\t\t\t{IncludeRegexp: \"^drivers/input/touchscreen/silead\\\\.c$|^drivers/platform/x86/touchscreen_dmi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/mellanox/|^include/linux/platform_data/mlxreg\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/olpc/|^drivers/platform/x86/|^include/linux/platform_data/x86/\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/surface/\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/surface/aggregator/|^drivers/platform/surface/surface_acpi_notify\\\\.c$|^drivers/platform/surface/surface_aggregator_cdev\\\\.c$|^drivers/platform/surface/surface_aggregator_registry\\\\.c$|^include/linux/surface_acpi_notify\\\\.h$|^include/linux/surface_aggregator/|^include/uapi/linux/surface_aggregator/\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/surface/surface_aggregator_hub\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/surface/surface_aggregator_tabletsw\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/surface/surface_dtx\\\\.c$|^include/uapi/linux/surface_aggregator/dtx\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/surface/surface_gpe\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/surface/surface_hotplug\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/surface/surface_platform_profile\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/surface/surfacepro3_button\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/acer-wmi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/acerhdf\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/adv_swbutton\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/amd/pmc\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/amd/pmf/\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/asus-tf103c-dock\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/asus-wireless\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/asus[^/]*\\\\.c$|^drivers/platform/x86/eeepc[^/]*\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/classmate-laptop\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/compal-laptop\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/dell/dcdbas\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/dell/dell-laptop\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/dell/dell-smbios-smm\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/dell/dell-smbios-wmi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/dell/dell-smbios\\\\.[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/dell/dell-wmi-privacy\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/dell/dell-wmi-sysman/\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/dell/dell_rbu\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/fujitsu-laptop\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/fujitsu-tablet\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/gigabyte-wmi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/gpd-pocket-fan\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/hdaps\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/hp/tc1100-wmi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/ideapad-laptop\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/intel/atomisp2/led\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/intel/atomisp2/pm\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/intel/hid\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/intel/int1092/\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/intel/ishtp_eclite\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/intel/pmc/\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/intel/speed_select_if/|^include/uapi/linux/isst_if\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/intel/tpmi\\\\.c$|^include/linux/intel_tpmi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/intel/uncore-frequency/\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/intel/vbtn\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/lg-laptop\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/mlx-platform\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/msi-laptop\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/msi-wmi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/nvidia-wmi-ec-backlight\\\\.c$|^include/linux/platform_data/x86/nvidia-wmi-ec-backlight\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/panasonic-laptop\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/peaq-wmi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/samsung-laptop\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/serial-multi-instantiate\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/system76_acpi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/think-lmi\\\\..$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/thinkpad_acpi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/topstar-laptop\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/toshiba-wmi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/toshiba_acpi\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/toshiba_bluetooth\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/toshiba_haps\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/uv_sysfs\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/wmi\\\\.c$|^include/uapi/linux/wmi\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/platform/x86/x86-android-tablets\\\\.c$\"},\n\t\t\t{IncludeRegexp: \"^drivers/power/supply/surface_battery\\\\.c$|^drivers/power/supply/surface_charger\\\\.c$\"},\n\t\t},\n\t}\n\n\txen = Subsystem{\n\t\tName: \"xen\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/[^/]*/include/asm/xen/swiotlb-xen\\\\.h$|^drivers/xen/swiotlb-xen\\\\.c$|^include/xen/arm/swiotlb-xen\\\\.h$|^include/xen/swiotlb-xen\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^arch/arm/include/asm/xen/|^arch/arm/xen/\"},\n\t\t\t{IncludeRegexp: \"^arch/arm64/include/asm/xen/|^arch/arm64/xen/\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/configs/xen\\\\.config$|^arch/x86/include/asm/pvclock-abi\\\\.h$|^arch/x86/include/asm/xen/|^arch/x86/platform/pvh/|^arch/x86/xen/\"},\n\t\t\t{IncludeRegexp: \"^arch/x86/pci/[^/]*xen[^/]*$|^drivers/pci/[^/]*xen[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/[^/]*/xen-[^/]*front\\\\.c$|^drivers/xen/|^include/uapi/xen/|^include/xen/|^kernel/configs/xen\\\\.config$\"},\n\t\t\t{IncludeRegexp: \"^drivers/block/xen[^/]*$|^drivers/block/xen-blkback/\"},\n\t\t\t{IncludeRegexp: \"^drivers/gpu/drm/xen/\"},\n\t\t\t{IncludeRegexp: \"^drivers/net/xen-netback/\"},\n\t\t\t{IncludeRegexp: \"^drivers/scsi/xen-scsifront\\\\.c$|^drivers/xen/xen-scsiback\\\\.c$|^include/xen/interface/io/vscsiif\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^drivers/usb/host/xen[^/]*$|^include/xen/interface/io/usbif\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^sound/xen/\"},\n\t\t},\n\t}\n\n\txfs = Subsystem{\n\t\tName: \"xfs\",\n\t\tSyscalls: []string{\"syz_mount_image$xfs\"},\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&fs},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^fs/iomap/|^include/linux/iomap\\\\.h$\"},\n\t\t\t{IncludeRegexp: \"^fs/xfs/|^include/uapi/linux/dqblk_xfs\\\\.h$|^include/uapi/linux/fsmap\\\\.h$\"},\n\t\t},\n\t}\n\n\txtensa = Subsystem{\n\t\tName: \"xtensa\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\", \"[email protected]\"},\n\t\tParents: []*Subsystem{&kernel},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^arch/xtensa/|^drivers/irqchip/irq-xtensa-[^/]*$\"},\n\t\t\t{IncludeRegexp: \"^drivers/spi/spi-xtensa-xtfpga\\\\.c$|^sound/soc/xtensa/xtfpga-i2s\\\\.c$\"},\n\t\t},\n\t}\n\n\tzd1211 = Subsystem{\n\t\tName: \"zd1211\",\n\t\tLists: []string{\"[email protected]\"},\n\t\tMaintainers: []string{\"[email protected]\"},\n\t\tParents: []*Subsystem{&wireless},\n\t\tPathRules: []PathRule{\n\t\t\t{IncludeRegexp: \"^drivers/net/wireless/zydas/zd1211rw/\"},\n\t\t},\n\t}\n\n\treturn []*Subsystem{\n\t\t&_9p, &ac100, &accelerators, &acpi, &acpi4asus, &acpica, &acrn, &actions, &afs, &alpha, &alsa, &amdgfx, &amlogic, &apparmor, &arch, &arm, &armmsm, &asahi, &aspeed, &ath10k, &ath11k, &ath12k, &atm, &audit, &autofs, &axis, &b43, &batman, &bcache, &bfs, &block, &bluetooth, &bpf, &brcm80211, &bridge, &btrfs, &cachefs, &can, &ceph, &cgroups, &chrome, &cifs, &cirrus, &clk, &coda, &coresight, &crypto, &csky, &cxl, &damon, &dccp, &dell, &devicetree, &dm, &dmaengine, &drbd, &dri, &ecryptfs, &edac, &efi, &erofs, &etnaviv, &ext4, &f2fs, &fat, &fbdev, &fpga, &freedreno, &fs, &fscrypt, &fsi, &fsverity, &fuse, &geode, &gfs2, &gpio, &greybus, &hams, &hardening, &hexagon, &hfs, &hippi, &hwmon, &hyperv, &i2c, &i3c, &ia64, &ide, &iio, &imx, &input, &integrity, &intelgfx, &intelgvt, &intelwiredlan, &iouring, &iommu, &ipack, &isdn4linux, &isofs, &jfs, &karma, &kasan, &kernel, &kernfs, &kexec, &keyrings, &kgdb, &kunit, &kvm, &kvmriscv, &kvmarm, &leds, &libertas, &lima, &linux1394, &linuxppc, &linuxpps, &livepatching, &llvm, &loongarch, &lsm, &lvs, &m68k, &malidp, &media, &mediatek, &megaraid, &mhi, &mips, &mjpeg, &mm, &mmc, &modules, &mpi3, &mptfusion, &mptcp, &mtd, &nbd, &net, &netfilter, &nfc, &nfs, &nilfs, &nitro, &nouveau, &ntb, &ntfs, &ntfs3, &nvdimm, &nvme, &ocfs2, &omap, &optee, &openiscsi, &openbmc, &openipmi, &openrisc, &openvswitch, &openwrt, &orangefs, &ossdrivers, &overlayfs, &oxnas, &parisc, &parport, &pci, &perf, &phy, &pm, &ppp, &pvrusb2, &pwm, &qat, &raid, &rcu, &rdma, &rds, &reiserfs, &remoteproc, &renesassoc, &riscv, &rockchip, &rpi, &rttools, &rtc, &rust, &s390, &samsungsoc, &scsi, &sctp, &selinux, &serial, &sgx, &sh, &snpsarc, &sof, &sparclinux, &speakup, &spi, &spice, &squashfs, &staging, &stm32, &sunxi, &target, &tegra, &tipc, &tomoyo, &trace, &uclinux, &udf, &um, &unisoc, &usb, &usbstorage, &video, &virt, &watchdog, &wcn36xx, &wireguard, &wireless, &wpan, &x25, &x86, &x86drivers, &xen, &xfs, &xtensa, &zd1211,\n\t}\n\n}", "func IsDev() bool {\n\treturn env == \"dev\"\n}", "func Dev(traceID string, funcName string, format string, a ...interface{}) {\n\tl.DevOffset(traceID, 1, funcName, format, a...)\n}", "func (o KubernetesClusterDefaultNodePoolLinuxOsConfigSysctlConfigOutput) NetCoreNetdevMaxBacklog() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v KubernetesClusterDefaultNodePoolLinuxOsConfigSysctlConfig) *int {\n\t\treturn v.NetCoreNetdevMaxBacklog\n\t}).(pulumi.IntPtrOutput)\n}", "func ideControllerListString(ctlrs []*types.VirtualIDEController) string {\n\tvar l object.VirtualDeviceList\n\tfor _, ctlr := range ctlrs {\n\t\tif ctlr == nil {\n\t\t\tl = append(l, types.BaseVirtualDevice(nil))\n\t\t} else {\n\t\t\tl = append(l, ctlr.GetVirtualDevice())\n\t\t}\n\t}\n\treturn DeviceListString(l)\n}", "func NewPciNetDevice(dev *ghw.PCIDevice, rFactory types.ResourceFactory, rc *types.ResourceConfig) (types.PciNetDevice, error) {\n\tvar ifName string\n\tinfoProviders := make([]types.DeviceInfoProvider, 0)\n\n\tdriverName, err := utils.GetDriverName(dev.Address)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tinfoProviders = append(infoProviders, rFactory.GetDefaultInfoProvider(dev.Address, driverName))\n\trdmaSpec := rFactory.GetRdmaSpec(dev.Address)\n\tnf, ok := rc.SelectorObj.(*types.NetDeviceSelectors)\n\tif ok {\n\t\t// Add InfoProviders based on Selector data\n\t\tif nf.IsRdma {\n\t\t\tif rdmaSpec.IsRdma() {\n\t\t\t\tinfoProviders = append(infoProviders, NewRdmaInfoProvider(rdmaSpec))\n\t\t\t} else {\n\t\t\t\tglog.Warningf(\"RDMA resources for %s not found. Are RDMA modules loaded?\", dev.Address)\n\t\t\t}\n\t\t}\n\t\tif nf.NeedVhostNet {\n\t\t\tif VhostNetDeviceExist() {\n\t\t\t\tinfoProviders = append(infoProviders, NewVhostNetInfoProvider())\n\t\t\t} else {\n\t\t\t\tglog.Errorf(\"GetDeviceSpecs(): vhost-net is required in the configuration but /dev/vhost-net doesn't exist\")\n\t\t\t}\n\t\t}\n\t}\n\n\tpciDev, err := resources.NewPciDevice(dev, rFactory, infoProviders)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tpciAddr := pciDev.GetPciAddr()\n\tnetDevs, _ := utils.GetNetNames(pciAddr)\n\tif len(netDevs) == 0 {\n\t\tifName = \"\"\n\t} else {\n\t\tifName = netDevs[0]\n\t}\n\tpfName, err := utils.GetPfName(pciAddr)\n\tif err != nil {\n\t\tglog.Warningf(\"unable to get PF name %q\", err.Error())\n\t}\n\n\tlinkType := \"\"\n\tif len(ifName) > 0 {\n\t\tla, err := utils.GetLinkAttrs(ifName)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tlinkType = la.EncapType\n\t}\n\n\treturn &pciNetDevice{\n\t\tPciDevice: pciDev,\n\t\tifName: ifName,\n\t\tpfName: pfName,\n\t\tlinkSpeed: \"\", // TO-DO: Get this using utils pkg\n\t\trdmaSpec: rdmaSpec,\n\t\tlinkType: linkType,\n\t}, nil\n}", "func computeDevAddr(device types.BaseVirtualDevice, ctlr types.BaseVirtualController) (string, error) {\n\tvd := device.GetVirtualDevice()\n\tvc := ctlr.GetVirtualController()\n\tctype, err := controllerTypeToClass(ctlr)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tparts := []string{\n\t\tctype,\n\t\tstrconv.Itoa(int(vc.BusNumber)),\n\t\tstrconv.Itoa(int(structure.DeRef(vd.UnitNumber).(int32))),\n\t}\n\treturn strings.Join(parts, \":\"), nil\n}", "func ActionDevices(includedDevices IncludedDevices) carapace.Action {\n\treturn carapace.ActionCallback(func(c carapace.Context) carapace.Action {\n\t\tif _, err := execabs.LookPath(\"nmcli\"); err == nil {\n\t\t\treturn carapace.ActionExecCommand(\"nmcli\", \"--terse\", \"--fields\", \"device,type\", \"device\", \"status\")(func(output []byte) carapace.Action {\n\t\t\t\tlines := strings.Split(string(output), \"\\n\")\n\t\t\t\tvals := make([]string, 0)\n\t\t\t\tfor _, line := range lines[:len(lines)-1] {\n\t\t\t\t\tparts := strings.Split(line, \":\")\n\t\t\t\t\tif includedDevices.Includes(parts[1]) {\n\t\t\t\t\t\tvals = append(vals, parts[0], parts[1])\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\treturn carapace.ActionValuesDescribed(vals...)\n\t\t\t})\n\t\t} else if _, err := execabs.LookPath(\"ifconfig\"); err == nil {\n\t\t\t// fallback to basic ifconfig if nmcli is not available\n\t\t\treturn carapace.ActionExecCommand(\"ifconfig\")(func(output []byte) carapace.Action {\n\t\t\t\tinterfaces := []string{}\n\t\t\t\tr := regexp.MustCompile(\"^[0-9a-zA-Z]\")\n\t\t\t\tfor _, line := range strings.Split(string(output), \"\\n\") {\n\t\t\t\t\tif matches := r.MatchString(line); matches {\n\t\t\t\t\t\tinterfaces = append(interfaces, strings.Split(line, \":\")[0])\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\treturn carapace.ActionValues(interfaces...)\n\t\t\t})\n\t\t}\n\t\treturn carapace.ActionMessage(\"neither nmcli nor ifconfig available\")\n\t}).Tag(\"network devices\")\n}", "func GetNetDevicesFromPci(pciAddress string) ([]string, error) {\n\tpciDir := filepath.Join(PciSysDir, pciAddress, \"net\")\n\treturn getFileNamesFromPath(pciDir)\n}", "func listFilterIfs(filter func(netlink.Link) int) (map[int]netlink.Link, error) {\n\tifs, err := netlink.LinkList()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvethLXCIdxs := map[int]netlink.Link{}\n\tfor _, intf := range ifs {\n\t\tif idx := filter(intf); idx != -1 {\n\t\t\tvethLXCIdxs[idx] = intf\n\t\t}\n\t}\n\treturn vethLXCIdxs, nil\n}", "func DeviceTypesPWinMO6() *DeviceTypes {\n\tv := DeviceTypesVWinMO6\n\treturn &v\n}", "func DeviceTypesPWinMO6() *DeviceTypes {\n\tv := DeviceTypesVWinMO6\n\treturn &v\n}", "func (o KubernetesClusterDefaultNodePoolLinuxOsConfigSysctlConfigPtrOutput) NetCoreNetdevMaxBacklog() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v *KubernetesClusterDefaultNodePoolLinuxOsConfigSysctlConfig) *int {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.NetCoreNetdevMaxBacklog\n\t}).(pulumi.IntPtrOutput)\n}", "func (o KubernetesClusterNodePoolLinuxOsConfigSysctlConfigPtrOutput) NetCoreNetdevMaxBacklog() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v *KubernetesClusterNodePoolLinuxOsConfigSysctlConfig) *int {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.NetCoreNetdevMaxBacklog\n\t}).(pulumi.IntPtrOutput)\n}", "func genMobileNexus10UA() string {\n\tbuild := nexus10Builds[rand.Intn(len(nexus10Builds))]\n\tandroid := nexus10AndroidVersions[rand.Intn(len(nexus10AndroidVersions))]\n\tchrome := chromeVersions[rand.Intn(len(chromeVersions))]\n\treturn fmt.Sprintf(\"Mozilla/5.0 (Linux; Android %s; Nexus 10 Build/%s) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/%s Safari/537.36\", android, build, chrome)\n}", "func (h *DeviceHandler) DeviceList(ctx context.Context, sessionID int) ([]keybase1.Device, error) {\n\tuis := libkb.UIs{\n\t\tLogUI: h.getLogUI(sessionID),\n\t\tSessionID: sessionID,\n\t}\n\teng := engine.NewDevList(h.G())\n\tm := libkb.NewMetaContext(ctx, h.G()).WithUIs(uis)\n\tif err := engine.RunEngine2(m, eng); err != nil {\n\t\treturn nil, err\n\t}\n\treturn eng.List(), nil\n}", "func adaptDeviceToConf(\n\tconf map[string]interface{},\n\tdevice pxapi.QemuDevice,\n) map[string]interface{} {\n\t// Value type should be one of types allowed by Terraform schema types.\n\tfor key, value := range device {\n\t\t// This nested switch is used for nested config like in `net[n]`,\n\t\t// where Proxmox uses `key=<0|1>` in string\" at the same time\n\t\t// a boolean could be used in \".tf\" files.\n\t\tswitch conf[key].(type) {\n\t\tcase bool:\n\t\t\tswitch value := value.(type) {\n\t\t\t// If the key is bool and value is int (which comes from Proxmox API),\n\t\t\t// should be converted to bool (as in \".tf\" conf).\n\t\t\tcase int:\n\t\t\t\tsValue := strconv.Itoa(value)\n\t\t\t\tbValue, err := strconv.ParseBool(sValue)\n\t\t\t\tif err == nil {\n\t\t\t\t\tconf[key] = bValue\n\t\t\t\t}\n\t\t\t// If value is bool, which comes from Terraform conf, add it directly.\n\t\t\tcase bool:\n\t\t\t\tconf[key] = value\n\t\t\t}\n\t\t// Anything else will be added as it is.\n\t\tdefault:\n\t\t\tconf[key] = value\n\t\t}\n\t}\n\n\treturn conf\n}", "func platformHeadersWithAPIToken(apiToken string) map[string]string {\n\tm := make(map[string]string)\n\tm[\"Content-Type\"] = \"application/json\"\n\tm[util.PlatformApiTokenHeader] = apiToken\n\treturn m\n}", "func defaultSolarisIfNameCmd() []string {\n\treturn []string{\"/usr/sbin/route\", \"-n\", \"get\", \"default\"}\n}", "func (h *HeaderfileWriter) writeStaticData(fw io.Writer, e datapath.EndpointConfiguration) {\n\tif e.IsHost() {\n\t\tif option.Config.EnableNodePort {\n\t\t\t// Values defined here are for the host datapath attached to the\n\t\t\t// host device and therefore won't be used. We however need to set\n\t\t\t// non-zero values to prevent the compiler from optimizing them\n\t\t\t// out, because we need to substitute them for host datapaths\n\t\t\t// attached to native devices.\n\t\t\t// When substituting symbols in the object file, we will replace\n\t\t\t// these values with zero for the host device and with the actual\n\t\t\t// values for the native devices.\n\t\t\tfmt.Fprint(fw, \"/* Fake values, replaced by 0 for host device and by actual values for native devices. */\\n\")\n\t\t\tfmt.Fprint(fw, defineUint32(\"NATIVE_DEV_IFINDEX\", 1))\n\t\t\tfmt.Fprint(fw, \"\\n\")\n\t\t}\n\t\tif option.Config.EnableBPFMasquerade {\n\t\t\tif option.Config.EnableIPv4Masquerade {\n\t\t\t\t// NodePort comment above applies to IPV4_MASQUERADE too\n\t\t\t\tplaceholderIPv4 := []byte{1, 1, 1, 1}\n\t\t\t\tfmt.Fprint(fw, defineIPv4(\"IPV4_MASQUERADE\", placeholderIPv4))\n\t\t\t}\n\t\t\tif option.Config.EnableIPv6Masquerade {\n\t\t\t\t// NodePort comment above applies to IPV6_MASQUERADE too\n\t\t\t\tplaceholderIPv6 := []byte{1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1}\n\t\t\t\tfmt.Fprint(fw, defineIPv6(\"IPV6_MASQUERADE\", placeholderIPv6))\n\t\t\t}\n\t\t}\n\t\t// Dummy value to avoid being optimized when 0\n\t\tfmt.Fprint(fw, defineUint32(\"SECCTX_FROM_IPCACHE\", 1))\n\n\t\t// Use templating for ETH_HLEN only if there is any L2-less device\n\t\tif !mac.HaveMACAddrs(option.Config.GetDevices()) {\n\t\t\t// L2 hdr len (for L2-less devices it will be replaced with \"0\")\n\t\t\tfmt.Fprint(fw, defineUint16(\"ETH_HLEN\", mac.EthHdrLen))\n\t\t}\n\t} else {\n\t\t// We want to ensure that the template BPF program always has \"LXC_IP\"\n\t\t// defined and present as a symbol in the resulting object file after\n\t\t// compilation, regardless of whether IPv6 is disabled. Because the type\n\t\t// templateCfg hardcodes a dummy IPv6 address (and adheres to the\n\t\t// datapath.EndpointConfiguration interface), we can rely on it always\n\t\t// having an IPv6 addr. Endpoints however may not have IPv6 addrs if IPv6\n\t\t// is disabled. Hence this check prevents us from omitting the \"LXC_IP\"\n\t\t// symbol from the template BPF program. Without this, the following\n\t\t// scenario is possible:\n\t\t// 1) Enable IPv6 in cilium\n\t\t// 2) Create an endpoint (ensure endpoint has an IPv6 addr)\n\t\t// 3) Disable IPv6 and restart cilium\n\t\t// This results in a template BPF object without an \"LXC_IP\" defined,\n\t\t// __but__ the endpoint still has \"LXC_IP\" defined. This causes a later\n\t\t// call to loader.ELFSubstitutions() to fail on missing a symbol \"LXC_IP\".\n\t\tif ipv6 := e.IPv6Address(); ipv6.IsValid() {\n\t\t\tfmt.Fprint(fw, defineIPv6(\"LXC_IP\", ipv6.AsSlice()))\n\t\t}\n\n\t\tfmt.Fprint(fw, defineIPv4(\"LXC_IPV4\", e.IPv4Address().AsSlice()))\n\t\tfmt.Fprint(fw, defineUint16(\"LXC_ID\", uint16(e.GetID())))\n\t}\n\n\tfmt.Fprint(fw, defineMAC(\"NODE_MAC\", e.GetNodeMAC()))\n\n\tsecID := e.GetIdentityLocked().Uint32()\n\tfmt.Fprint(fw, defineUint32(\"SECLABEL\", secID))\n\tfmt.Fprint(fw, defineUint32(\"SECLABEL_IPV4\", secID))\n\tfmt.Fprint(fw, defineUint32(\"SECLABEL_IPV6\", secID))\n\tfmt.Fprint(fw, defineUint32(\"SECLABEL_NB\", byteorder.HostToNetwork32(secID)))\n\tfmt.Fprint(fw, defineUint32(\"POLICY_VERDICT_LOG_FILTER\", e.GetPolicyVerdictLogFilter()))\n\n\tepID := uint16(e.GetID())\n\tfmt.Fprintf(fw, \"#define POLICY_MAP %s\\n\", bpf.LocalMapName(policymap.MapName, epID))\n\tcallsMapName := callsmap.MapName\n\tif e.IsHost() {\n\t\tcallsMapName = callsmap.HostMapName\n\t}\n\tfmt.Fprintf(fw, \"#define CALLS_MAP %s\\n\", bpf.LocalMapName(callsMapName, epID))\n\tif option.Config.EnableCustomCalls && !e.IsHost() {\n\t\tfmt.Fprintf(fw, \"#define CUSTOM_CALLS_MAP %s\\n\", bpf.LocalMapName(callsmap.CustomCallsMapName, epID))\n\t}\n}", "func makeDeviceObject(m interface{}, d *schema.ResourceData) (output lmv1.RestDevice) {\n\t// if displayname is not there, we can automatically add ipaddr\n\tvar displayname = d.Get(\"display_name\").(string)\n\tif displayname == \"\" {\n\t\tdisplayname = d.Get(\"ip_addr\").(string)\n\t}\n\n\toutput = lmv1.RestDevice{\n\t\tName: d.Get(\"ip_addr\").(string),\n\t\tDisplayName: displayname,\n\t\tDisableAlerting: d.Get(\"disable_alerting\").(bool),\n\t\tHostGroupIds: d.Get(\"hostgroup_id\").(string),\n\t\tPreferredCollectorId: int32(d.Get(\"collector\").(int)),\n\t\tCustomProperties: []lmv1.NameAndValue{},\n\t}\n\n\treturn\n}", "func GetPciFromNetDevice(name string) (string, error) {\n\tdevPath := filepath.Join(NetSysDir, name)\n\n\trealPath, err := utilfs.Fs.Readlink(devPath)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"device %s not found: %s\", name, err)\n\t}\n\n\tparent := filepath.Dir(realPath)\n\tbase := filepath.Base(parent)\n\t// Devices can have their PCI device sysfs entry at different levels:\n\t// PF, VF, SF representor:\n\t// /sys/devices/pci0000:00/.../0000:03:00.0/net/p0\n\t// /sys/devices/pci0000:00/.../0000:03:00.0/net/pf0hpf\n\t// /sys/devices/pci0000:00/.../0000:03:00.0/net/pf0vf0\n\t// /sys/devices/pci0000:00/.../0000:03:00.0/net/pf0sf0\n\t// SF port:\n\t// /sys/devices/pci0000:00/.../0000:03:00.0/mlx5_core.sf.3/net/enp3s0f0s1\n\t// This loop allows detecting any of them.\n\tfor parent != \"/\" && !pciAddressRe.MatchString(base) {\n\t\tparent = filepath.Dir(parent)\n\t\tbase = filepath.Base(parent)\n\t}\n\t// If we stopped on '/' and the base was never a proper PCI address,\n\t// then 'netdev' is not a PCI device.\n\tif !pciAddressRe.MatchString(base) {\n\t\treturn \"\", fmt.Errorf(\"device %s is not a PCI device: %s\", name, realPath)\n\t}\n\treturn base, nil\n}", "func OpenDev(filename string) (BlockDevice, error) {\n\text := strings.ToLower(path.Ext(filename))\n\tswitch ext {\n\tcase \".po\":\n\t\treturn LoadDev(filename)\n\t}\n\treturn nil, fmt.Errorf(\"Unimplemented/unknown device file extension %q\", ext)\n}", "func getIfNameListForPort(\n\tctx *zedrouterContext,\n\tport string) []string {\n\n\tifNameList := adapterToIfNames(ctx, port)\n\tlog.Infof(\"ifNameList: %+v\", ifNameList)\n\n\tfilteredList := make([]string, 0)\n\tfor _, ifName := range ifNameList {\n\t\tdnsPort := ctx.deviceNetworkStatus.GetPortByIfName(ifName)\n\t\tif dnsPort != nil {\n\t\t\t// XXX - We have a bug in MakeDeviceNetworkStatus where we are allowing\n\t\t\t//\ta device without the corresponding linux interface. We can\n\t\t\t//\tremove this check for ifindex here when the MakeDeviceStatus\n\t\t\t//\tis fixed.\n\t\t\t// XXX That bug has been fixed. Retest without this code?\n\t\t\tifIndex, err := devicenetwork.IfnameToIndex(ifName)\n\t\t\tif err == nil {\n\t\t\t\tlog.Infof(\"ifName %s, ifindex: %d added to filteredList\",\n\t\t\t\t\tifName, ifIndex)\n\t\t\t\tfilteredList = append(filteredList, ifName)\n\t\t\t} else {\n\t\t\t\tlog.Infof(\"ifIndex not found for ifName(%s) - err: %s\",\n\t\t\t\t\tifName, err.Error())\n\t\t\t}\n\t\t} else {\n\t\t\tlog.Infof(\"DeviceNetworkStatus not found for port(%s)\", port)\n\t\t}\n\t}\n\tif len(filteredList) > 0 {\n\t\tlog.Infof(\"filteredList: %+v\", filteredList)\n\t\treturn filteredList\n\t}\n\tlog.Infof(\"ifname or ifindex not found for any interface for port(%s).\"+\n\t\t\"Returning the unfiltered list: %+v\", port, ifNameList)\n\treturn ifNameList\n}", "func NewDevSession_List(s *capnp.Segment, sz int32) (DevSession_List, error) {\n\tl, err := capnp.NewPointerList(s, sz)\n\treturn capnp.CapList[DevSession](l), err\n}", "func createIDEController(l *object.VirtualDeviceList) ([]types.BaseVirtualDeviceConfigSpec, error) {\n\tide, _ := l.CreateIDEController()\n\tcspec, err := object.VirtualDeviceList{ide}.ConfigSpec(types.VirtualDeviceConfigSpecOperationAdd)\n\t*l = applyDeviceChange(*l, cspec)\n\treturn cspec, err\n}", "func printConstsDef(w io.Writer, def string, cs []TypedEntry) {\n\tif len(cs) == 0 {\n\t\treturn\n\t}\n\tfmt.Fprintf(w, \"%s(%s\", indent(1), def)\n\tprintTypedNames(w, \"\\n\"+indent(2), cs)\n\tfmt.Fprintln(w, \")\")\n}", "func RuntimeEnvVarDocs(runtime EnvRuntime) string {\n\tw := &bytes.Buffer{}\n\n\tfmt.Fprintln(w, \"Kf provides the following runtime environment variables:\")\n\tfmt.Fprintln(w, \"\")\n\n\tenvVars := getRuntimeEnvVars(runtime)\n\tenvVars = algorithms.Dedupe(envVars).(runtimeEnvVars)\n\tsort.Sort(envVars)\n\n\tfor _, v := range envVars {\n\t\tfmt.Fprintf(w, \" * %s: %s\", v.name, v.description)\n\t\tfmt.Fprintln(w)\n\n\t\tfor _, alias := range v.aliases {\n\t\t\tfmt.Fprintf(w, \" * %s: Alias of %s\", alias, v.name)\n\t\t\tfmt.Fprintln(w)\n\t\t}\n\t}\n\n\treturn w.String()\n}", "func GetSNMPDevices(ctx *Context) {\n\t// swagger:operation GET /cfg/snmpdevice Config_Device GetSNMPDevices\n\t//---\n\t// summary: Get All devices info from DB and Runtime\n\t// description: Get All Devices config info as an array of config and boolean if working in runtime.\n\t// tags:\n\t// - \"Devices Config\"\n\t//\n\t// responses:\n\t// '200':\n\t// description: \"OK\"\n\t// schema:\n\t// \"$ref\": \"#/responses/idOfArrayDeviceStatResp\"\n\t// '404':\n\t// description: unexpected error\n\t// schema:\n\t// \"$ref\": \"#/responses/idOfStringResp\"\n\n\tdevcfgarray, err := agent.MainConfig.Database.GetSnmpDeviceCfgArray(\"\")\n\tif err != nil {\n\t\tctx.JSON(404, err.Error())\n\t\tlog.Errorf(\"Error on get Devices :%+s\", err)\n\t\treturn\n\t}\n\n\tdsmap := []*DeviceStatMap{}\n\tfor _, v := range devcfgarray {\n\t\trt := agent.IsDeviceInRuntime(v.ID)\n\t\tdsmap = append(dsmap, &DeviceStatMap{*v, rt})\n\t}\n\tctx.JSON(200, &dsmap)\n\tlog.Debugf(\"Getting DEVICEs %+v\", &dsmap)\n}", "func LexTypesCfg() {\n\tfmt.Println(\"[start] read lex types config and write totcode\")\n\tdefer fmt.Println(\"[end] read lex types config and write totcode\")\n\twritecv(`type PglaProduct int\n\nconst (\n\tPGLA_PRODUCT_ PglaProduct = iota\n\t`)\n\n\tdatas, err := smn_file.FileReadAll(\"./datas/analysis/lex_pgl/lextypes.cfg\")\n\tcheck(err)\n\n\tconstSet := map[string]bool{\"\": true}\n\tconstList := []string{}\n\tcheckList := []string{} //IsXXX(LexProdcut) bool\n\n\tfor _, line := range strings.Split(string(datas), \"\\n\") {\n\t\tline = strings.Split(line, \"#\")[0]\n\t\tline = strings.TrimSpace(line)\n\t\tline = strings.ToUpper(line)\n\n\t\tif constSet[line] {\n\t\t\tcontinue\n\t\t}\n\n\t\tconstList = append(constList, \"PGLA_PRODUCT_\"+line)\n\t\tcheckList = append(checkList, strings.ToUpper(line[:1])+strings.ToLower(line[1:]))\n\t\tconstSet[line] = true\n\n\t\twritecvf(\"PGLA_PRODUCT_%s\\n\", line)\n\t}\n\n\twritecv(`\n)\n`)\n\twritecv(`var PglaNameMap = map[PglaProduct]string{\n`)\n\twritecv(\"-1 : \\\"EMD\\\",\\n\")\n\n\tfor _, cst := range constList {\n\t\twritecvf(\"%s:\\\"%s\\\",\\n\", cst, cst)\n\t}\n\n\twritecv(`}\n`)\n\n\tfor _, chk := range checkList {\n\t\twritecvf(`//Is%s chack if lex is %s.\nfunc Is%s(lex *LexProduct) bool{\n\treturn lex.ProductType() == int(PGLA_PRODUCT_%s)\n}\n`, chk, chk, chk, strings.ToUpper(chk))\n\t}\n}", "func startDevice(devName string, srcIP net.IP, destIP net.IP) error {\n\tcmd := exec.Command(\"ifconfig\", devName, \"inet\", srcIP.String(), destIP.String(), \"up\")\n\treturn cmd.Run()\n}", "func dockStationMACs(ctx context.Context) ([]string, error) {\n\tmanager, err := shill.NewManager(ctx)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to create shill manager proxy\")\n\t}\n\n\t_, props, err := manager.DevicesByTechnology(ctx, shill.TechnologyEthernet)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to get ethernet devices\")\n\t}\n\n\tvar macs []string\n\n\tfor _, deviceProps := range props {\n\t\tif !deviceProps.Has(shillconst.DevicePropertyEthernetBusType) {\n\t\t\tcontinue\n\t\t}\n\t\tbusType, err := deviceProps.GetString(shillconst.DevicePropertyEthernetBusType)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"failed to get device bus type\")\n\t\t}\n\n\t\tiface, err := deviceProps.GetString(shillconst.DevicePropertyInterface)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"failed to get interface name\")\n\t\t}\n\n\t\tif busType != \"usb\" {\n\t\t\tcontinue\n\t\t}\n\n\t\tifi, err := net.InterfaceByName(iface)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrapf(err, \"failed to get interface by %q name\", iface)\n\t\t}\n\t\tif ifi.HardwareAddr == nil {\n\t\t\treturn nil, errors.New(\"interface MAC address is nil\")\n\t\t}\n\t\tmacs = append(macs, strings.ToLower(ifi.HardwareAddr.String()))\n\t}\n\n\tif len(macs) == 0 {\n\t\treturn nil, errors.New(\"not found USB Ethernet adapter connected to the DUT\")\n\t}\n\n\treturn macs, nil\n}", "func ScanDevices() []FoundDevice {\n\tctx := gousb.NewContext()\n\tctx.Debug(0)\n\tdefer ctx.Close()\n\t// Get the list of device that match TMK's vendor id\n\tdevs, _ := ctx.OpenDevices(isSupported)\n\tvar devices = []FoundDevice{}\n\tfor _, dev := range devs {\n\n\t\tmetadata, _ := getMetdata(dev.Desc)\n\t\tdevices = append(devices, FoundDevice{\n\t\t\tStatus: &DeviceStatus{isDfu: metadata.Name == \"dfu\"},\n\t\t\tDescription: dev.Desc,\n\t\t\tMetadata: metadata,\n\t\t})\n\t}\n\tfor _, d := range devs {\n\t\td.Close()\n\t}\n\treturn devices\n}", "func ListDevices() (list []Device) {\n\tifs, e := pcap.FindAllDevs()\n\tif e != nil {\n\t\treturn\n\t}\n\tfor _, netif := range ifs {\n\t\td := Device{\n\t\t\tName: netif.Name,\n\t\t}\n\t\tfor _, a := range netif.Addresses {\n\t\t\td.Addresses = append(d.Addresses, a.IP)\n\t\t}\n\t\tlist = append(list, d)\n\t}\n\treturn list\n}", "func chipNames() []string {\n\tee, err := ioutil.ReadDir(\"/dev\")\n\tif err != nil {\n\t\treturn nil\n\t}\n\tcc := []string(nil)\n\tfor _, e := range ee {\n\t\tname := e.Name()\n\t\tif strings.HasPrefix(name, \"gpiochip\") {\n\t\t\tcc = append(cc, name)\n\t\t}\n\t}\n\treturn cc\n}", "func (n *NodeInfo) buildDeviceInfo() error {\n\ttotalGPUMem := 0\n\tif n.gpuCount > 0 {\n\t\ttotalGPUMem = n.gpuTotalMemory / n.gpuCount\n\t}\nGPUSearchLoop:\n\tfor _, pod := range n.pods {\n\t\tif gpuMemoryInPod(pod) <= 0 {\n\t\t\tcontinue GPUSearchLoop\n\t\t}\n\t\tfor devID, usedGPUMem := range n.getDeivceInfo(pod) {\n\t\t\tif n.devs[devID] == nil {\n\t\t\t\tn.devs[devID] = &DeviceInfo{\n\t\t\t\t\tpods: []v1.Pod{},\n\t\t\t\t\tidx: devID,\n\t\t\t\t\ttotalGPUMem: totalGPUMem,\n\t\t\t\t\tnode: n.node,\n\t\t\t\t}\n\t\t\t}\n\t\t\tn.devs[devID].usedGPUMem += usedGPUMem\n\t\t\tn.devs[devID].pods = append(n.devs[devID].pods, pod)\n\t\t}\n\t}\n\treturn nil\n}", "func schemaToCreateOpts(d *schema.ResourceData) *nodes.CreateOpts {\n\tproperties := propertiesMerge(d, \"root_device\")\n\treturn &nodes.CreateOpts{\n\t\tBootInterface: d.Get(\"boot_interface\").(string),\n\t\tConductorGroup: d.Get(\"conductor_group\").(string),\n\t\tConsoleInterface: d.Get(\"console_interface\").(string),\n\t\tDeployInterface: d.Get(\"deploy_interface\").(string),\n\t\tDriver: d.Get(\"driver\").(string),\n\t\tDriverInfo: d.Get(\"driver_info\").(map[string]interface{}),\n\t\tExtra: d.Get(\"extra\").(map[string]interface{}),\n\t\tInspectInterface: d.Get(\"inspect_interface\").(string),\n\t\tManagementInterface: d.Get(\"management_interface\").(string),\n\t\tName: d.Get(\"name\").(string),\n\t\tNetworkInterface: d.Get(\"network_interface\").(string),\n\t\tOwner: d.Get(\"owner\").(string),\n\t\tPowerInterface: d.Get(\"power_interface\").(string),\n\t\tProperties: properties,\n\t\tRAIDInterface: d.Get(\"raid_interface\").(string),\n\t\tRescueInterface: d.Get(\"rescue_interface\").(string),\n\t\tResourceClass: d.Get(\"resource_class\").(string),\n\t\tStorageInterface: d.Get(\"storage_interface\").(string),\n\t\tVendorInterface: d.Get(\"vendor_interface\").(string),\n\t}\n}", "func specPlatform(p *regpkg.Platform) *spec.Platform {\n\treturn &spec.Platform{\n\t\tArchitecture: p.Architecture,\n\t\tOS: p.OS,\n\t\tOSVersion: p.OSVersion,\n\t\tOSFeatures: p.OSFeatures,\n\t\tVariant: p.Variant,\n\t}\n}", "func DeviceFilter(rules []*devices.Rule) (asm.Instructions, string, error) {\n\t// Generate the minimum ruleset for the device rules we are given. While we\n\t// don't care about minimum transitions in cgroupv2, using the emulator\n\t// gives us a guarantee that the behaviour of devices filtering is the same\n\t// as cgroupv1, including security hardenings to avoid misconfiguration\n\t// (such as punching holes in wildcard rules).\n\temu := new(devicesemulator.Emulator)\n\tfor _, rule := range rules {\n\t\tif err := emu.Apply(*rule); err != nil {\n\t\t\treturn nil, \"\", err\n\t\t}\n\t}\n\tcleanRules, err := emu.Rules()\n\tif err != nil {\n\t\treturn nil, \"\", err\n\t}\n\n\tp := &program{\n\t\tdefaultAllow: emu.IsBlacklist(),\n\t}\n\tp.init()\n\n\tfor idx, rule := range cleanRules {\n\t\tif rule.Type == devices.WildcardDevice {\n\t\t\t// We can safely skip over wildcard entries because there should\n\t\t\t// only be one (at most) at the very start to instruct cgroupv1 to\n\t\t\t// go into allow-list mode. However we do double-check this here.\n\t\t\tif idx != 0 || rule.Allow != emu.IsBlacklist() {\n\t\t\t\treturn nil, \"\", fmt.Errorf(\"[internal error] emulated cgroupv2 devices ruleset had bad wildcard at idx %v (%s)\", idx, rule.CgroupString())\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\t\tif rule.Allow == p.defaultAllow {\n\t\t\t// There should be no rules which have an action equal to the\n\t\t\t// default action, the emulator removes those.\n\t\t\treturn nil, \"\", fmt.Errorf(\"[internal error] emulated cgroupv2 devices ruleset had no-op rule at idx %v (%s)\", idx, rule.CgroupString())\n\t\t}\n\t\tif err := p.appendRule(rule); err != nil {\n\t\t\treturn nil, \"\", err\n\t\t}\n\t}\n\treturn p.finalize(), license, nil\n}", "func (d *Device) Mkdev() int {\n\treturn int((d.Major << majorNum) | (d.Minor & 0xff) | ((d.Minor & 0xfff00) << minorNum))\n}", "func detectGPUs(visibleGPUs string) ([]device.Device, error) {\n\tflags := detectGPUsArgs[1:]\n\tif visibleGPUs != \"\" {\n\t\tflags = append(flags, fmt.Sprintf(detectGPUsIDFlagTpl, visibleGPUs))\n\t}\n\n\t// #nosec G204\n\tcmd := exec.Command(detectGPUsArgs[0], flags...)\n\tout, err := cmd.Output()\n\n\tif execError, ok := err.(*exec.Error); ok && execError.Err == exec.ErrNotFound {\n\t\treturn nil, nil\n\t} else if err != nil {\n\t\tlog.WithError(err).WithField(\"output\", string(out)).Warnf(\"error while executing nvidia-smi\")\n\t\treturn nil, nil\n\t}\n\n\tdevices := make([]device.Device, 0)\n\n\tr := csv.NewReader(strings.NewReader(string(out)))\n\tfor {\n\t\trecord, err := r.Read()\n\t\tswitch {\n\t\tcase err == io.EOF:\n\t\t\treturn devices, nil\n\t\tcase err != nil:\n\t\t\treturn nil, errors.Wrap(err, \"error parsing output of nvidia-smi as CSV\")\n\t\tcase len(record) != 3:\n\t\t\treturn nil, errors.New(\n\t\t\t\t\"error parsing output of nvidia-smi; GPU record should have exactly 3 fields\")\n\t\t}\n\n\t\tindex, err := strconv.Atoi(strings.TrimSpace(record[0]))\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(\n\t\t\t\terr, \"error parsing output of nvidia-smi; index of GPU cannot be converted to int\")\n\t\t}\n\n\t\tbrand := strings.TrimSpace(record[1])\n\t\tuuid := strings.TrimSpace(record[2])\n\n\t\tdevices = append(devices, device.Device{ID: index, Brand: brand, UUID: uuid, Type: device.GPU})\n\t}\n}", "func main() {\n\t// choose dcgm hostengine running mode\n\t// 1. dcgm.Embedded\n\t// 2. dcgm.Standalone -connect \"addr\", -socket \"isSocket\"\n\t// 3. dcgm.StartHostengine\n\tflag.Parse()\n\tif err := dcgm.Init(dcgm.Standalone, *connectAddr, *isSocket); err != nil {\n\t\tlog.Panicln(err)\n\t}\n\n\tdefer func() {\n\t\tif err := dcgm.Shutdown(); err != nil {\n\t\t\tlog.Panicln(err)\n\t\t}\n\t}()\n\n\tcount, err := dcgm.GetAllDeviceCount()\n\tif err != nil {\n\t\tlog.Panicln(err)\n\t}\n\n\tt := template.Must(template.New(\"Device\").Parse(deviceInfo))\n\n\tfor i := uint(0); i < count; i++ {\n\t\tdeviceInfo, err := dcgm.GetDeviceInfo(i)\n\t\tif err != nil {\n\t\t\tlog.Panicln(err)\n\t\t}\n\n\t\tif err = t.Execute(os.Stdout, deviceInfo); err != nil {\n\t\t\tlog.Panicln(\"Template error:\", err)\n\t\t}\n\t}\n}", "func bindToDev(fd int, devName string) error {\n\treturn fmt.Errorf(\"binding to device is not supported\")\n}", "func usersConfig(port int) Ifx {\n\treturn Ifx{\n\t\tFor: \"users\",\n\t\tName: fmt.Sprintf(\"swp%d\", port),\n\t\tBridgeDefs: []string{\n\t\t\t\"bridge-vids 2002 2003 2005\",\n\t\t\t\"bridge-allow-untagged no\",\n\t\t}}\n}", "func Define2String(def *TableDefine) string {\n\tcolumns := \"\\n\\t---------------+---------------\"\n\tfor _, col := range def.Columns {\n\t\tcolumns += fmt.Sprintf(\"\\n\\t%-15s|%-15s\", col.Name, col.Type)\n\t}\n\tconstraints := \"\\n\\tConstraints:\"\n\tif def.Constraint != nil {\n\t\tif len(def.Constraint.PrimaryKey) > 0 {\n\t\t\tconstraints += fmt.Sprintf(\"\\n\\t\\tPK: %s\", strings.Join(def.Constraint.PrimaryKey, \",\"))\n\t\t}\n\t\tif len(def.Constraint.Uniques) > 0 {\n\t\t\tfor _, unique := range def.Constraint.Uniques {\n\t\t\t\tconstraints += fmt.Sprintf(\"\\n\\t\\tUnique: (%s)\", strings.Join(unique, \",\"))\n\t\t\t}\n\t\t}\n\t}\n\n\treturn fmt.Sprintf(\" Table \\\"%s\\\".\\\"%s\\\" %s\\n%s\\n\", def.Schema, def.Table, columns, constraints)\n}" ]
[ "0.53961873", "0.5309108", "0.52739865", "0.5048785", "0.49038717", "0.48688504", "0.47383785", "0.4717659", "0.4664683", "0.46278298", "0.46216357", "0.46209374", "0.46183315", "0.45500436", "0.4548478", "0.45435837", "0.4460233", "0.4454798", "0.4429953", "0.4396922", "0.43495807", "0.43463606", "0.43440938", "0.4305545", "0.42866492", "0.42779112", "0.4258006", "0.42260116", "0.42190528", "0.42112365", "0.41882446", "0.4184742", "0.41836247", "0.41715014", "0.41578692", "0.4137942", "0.4133384", "0.41325244", "0.41313386", "0.41269937", "0.41206092", "0.4115727", "0.4102666", "0.40774304", "0.40586933", "0.40392867", "0.4036384", "0.40325317", "0.40293097", "0.40253466", "0.40224996", "0.40088874", "0.40060273", "0.40021592", "0.40004268", "0.39978173", "0.3994291", "0.39917022", "0.397596", "0.39676932", "0.3961991", "0.39609057", "0.394686", "0.39354485", "0.39269802", "0.39247492", "0.39247492", "0.39121565", "0.3908119", "0.39034498", "0.38982615", "0.3896603", "0.38961643", "0.3891848", "0.38880128", "0.38854286", "0.38773412", "0.38762704", "0.38760543", "0.38757876", "0.38756606", "0.3875409", "0.3874337", "0.38740933", "0.38666457", "0.38634658", "0.3857718", "0.38529646", "0.38514924", "0.38479877", "0.38369805", "0.38367665", "0.38343528", "0.3833915", "0.38337746", "0.3827474", "0.38246107", "0.38227248", "0.3814806", "0.3814444" ]
0.83545256
0
WriteNetdevConfig writes the BPF configuration for the endpoint to a writer.
func (h *HeaderfileWriter) WriteNetdevConfig(w io.Writer, cfg datapath.DeviceConfiguration) error { fw := bufio.NewWriter(w) h.writeNetdevConfig(fw, cfg) return fw.Flush() }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (mod *EthModule) WriteConfig(config_file string) error {\n\tb, err := json.Marshal(mod.eth.config)\n\tif err != nil {\n\t\tfmt.Println(\"error marshalling config:\", err)\n\t\treturn err\n\t}\n\tvar out bytes.Buffer\n\tjson.Indent(&out, b, \"\", \"\\t\")\n\terr = ioutil.WriteFile(config_file, out.Bytes(), 0600)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (h *HeaderfileWriter) WriteEndpointConfig(w io.Writer, e datapath.EndpointConfiguration) error {\n\tfw := bufio.NewWriter(w)\n\n\twriteIncludes(w)\n\th.writeStaticData(fw, e)\n\n\treturn h.writeTemplateConfig(fw, e)\n}", "func (fes *FrontEndService) WriteConfigKernel(conf *string, hasVIP4 bool, hasVIP6 bool) {\n\teFilter := \"none\"\n\tif hasVIP4 {\n\t\teFilter = \"filter default_v4\"\n\t}\n\n\t*conf += \"protocol kernel {\\n\"\n\t*conf += \"\\tipv4 {\\n\"\n\t*conf += \"\\t\\timport none;\\n\"\n\t*conf += \"\\t\\texport \" + eFilter + \";\\n\"\n\t*conf += \"\\t};\\n\"\n\t*conf += \"\\tkernel table \" + strconv.FormatInt(int64(fes.kernelTableId), 10) + \";\\n\"\n\tif fes.ecmp {\n\t\t*conf += \"\\tmerge paths on;\\n\"\n\t}\n\tif fes.dropIfNoPeer {\n\t\t// Setting the metric for the default blackhole route must be supported,\n\t\t// which requires the kernel proto's metric to be set to zero.\n\t\t//\n\t\t// \"Metric 0 has a special meaning of undefined metric, in which either OS default is used,\n\t\t// or per-route metric can be set using krt_metric attribute. Default: 32. \"\n\t\t*conf += \"\\tmetric 0;\\n\"\n\t}\n\t*conf += \"}\\n\"\n\t*conf += \"\\n\"\n\n\tif hasVIP6 {\n\t\teFilter = \"filter default_v6\"\n\t} else {\n\t\teFilter = \"none\"\n\t}\n\t*conf += \"protocol kernel {\\n\"\n\t*conf += \"\\tipv6 {\\n\"\n\t*conf += \"\\t\\timport none;\\n\"\n\t*conf += \"\\t\\texport \" + eFilter + \";\\n\"\n\t*conf += \"\\t};\\n\"\n\t*conf += \"\\tkernel table \" + strconv.FormatInt(int64(fes.kernelTableId), 10) + \";\\n\"\n\tif fes.ecmp {\n\t\t*conf += \"\\tmerge paths on;\\n\"\n\t}\n\tif fes.dropIfNoPeer {\n\t\t// Setting the metric for the default blackhole route must be supported,\n\t\t// which requires the kernel proto's metric to be set to zero.\n\t\t//\n\t\t// \"Metric 0 has a special meaning of undefined metric, in which either OS default is used,\n\t\t// or per-route metric can be set using krt_metric attribute. Default: 32. \"\n\t\t*conf += \"\\tmetric 0;\\n\"\n\t}\n\t*conf += \"}\\n\"\n\t*conf += \"\\n\"\n}", "func (h *HeaderfileWriter) WriteNodeConfig(w io.Writer, cfg *datapath.LocalNodeConfiguration) error {\n\textraMacrosMap := make(dpdef.Map)\n\tcDefinesMap := make(dpdef.Map)\n\n\tfw := bufio.NewWriter(w)\n\n\twriteIncludes(w)\n\n\trouterIP := node.GetIPv6Router()\n\thostIP := node.GetIPv6()\n\n\tfmt.Fprintf(fw, \"/*\\n\")\n\tif option.Config.EnableIPv6 {\n\t\tfmt.Fprintf(fw, \" cilium.v6.external.str %s\\n\", node.GetIPv6().String())\n\t\tfmt.Fprintf(fw, \" cilium.v6.internal.str %s\\n\", node.GetIPv6Router().String())\n\t\tfmt.Fprintf(fw, \" cilium.v6.nodeport.str %s\\n\", node.GetNodePortIPv6Addrs())\n\t\tfmt.Fprintf(fw, \"\\n\")\n\t}\n\tfmt.Fprintf(fw, \" cilium.v4.external.str %s\\n\", node.GetIPv4().String())\n\tfmt.Fprintf(fw, \" cilium.v4.internal.str %s\\n\", node.GetInternalIPv4Router().String())\n\tfmt.Fprintf(fw, \" cilium.v4.nodeport.str %s\\n\", node.GetNodePortIPv4Addrs())\n\tfmt.Fprintf(fw, \"\\n\")\n\tif option.Config.EnableIPv6 {\n\t\tfw.WriteString(dumpRaw(defaults.RestoreV6Addr, node.GetIPv6Router()))\n\t}\n\tfw.WriteString(dumpRaw(defaults.RestoreV4Addr, node.GetInternalIPv4Router()))\n\tfmt.Fprintf(fw, \" */\\n\\n\")\n\n\tcDefinesMap[\"KERNEL_HZ\"] = fmt.Sprintf(\"%d\", option.Config.KernelHz)\n\n\tif option.Config.EnableIPv6 {\n\t\textraMacrosMap[\"ROUTER_IP\"] = routerIP.String()\n\t\tfw.WriteString(defineIPv6(\"ROUTER_IP\", routerIP))\n\t}\n\n\tif option.Config.EnableIPv4 {\n\t\tipv4GW := node.GetInternalIPv4Router()\n\t\tloopbackIPv4 := node.GetIPv4Loopback()\n\t\tipv4Range := node.GetIPv4AllocRange()\n\t\tcDefinesMap[\"IPV4_GATEWAY\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(ipv4GW))\n\t\tcDefinesMap[\"IPV4_LOOPBACK\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(loopbackIPv4))\n\t\tcDefinesMap[\"IPV4_MASK\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(net.IP(ipv4Range.Mask)))\n\n\t\tif option.Config.EnableIPv4FragmentsTracking {\n\t\t\tcDefinesMap[\"ENABLE_IPV4_FRAGMENTS\"] = \"1\"\n\t\t\tcDefinesMap[\"IPV4_FRAG_DATAGRAMS_MAP\"] = fragmap.MapName\n\t\t\tcDefinesMap[\"CILIUM_IPV4_FRAG_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", option.Config.FragmentsMapEntries)\n\t\t}\n\t}\n\n\tif option.Config.EnableIPv6 {\n\t\textraMacrosMap[\"HOST_IP\"] = hostIP.String()\n\t\tfw.WriteString(defineIPv6(\"HOST_IP\", hostIP))\n\t}\n\n\tfor t, id := range tunnelProtocols {\n\t\tmacroName := fmt.Sprintf(\"TUNNEL_PROTOCOL_%s\", strings.ToUpper(t))\n\t\tcDefinesMap[macroName] = fmt.Sprintf(\"%d\", id)\n\t}\n\n\tencapProto := option.Config.TunnelProtocol\n\tif !option.Config.TunnelingEnabled() &&\n\t\toption.Config.EnableNodePort &&\n\t\toption.Config.NodePortMode != option.NodePortModeSNAT &&\n\t\toption.Config.LoadBalancerDSRDispatch == option.DSRDispatchGeneve {\n\t\tencapProto = option.TunnelGeneve\n\t}\n\n\tcDefinesMap[\"TUNNEL_PROTOCOL\"] = fmt.Sprintf(\"%d\", tunnelProtocols[encapProto])\n\tcDefinesMap[\"TUNNEL_PORT\"] = fmt.Sprintf(\"%d\", option.Config.TunnelPort)\n\n\tif tunnelDev, err := netlink.LinkByName(fmt.Sprintf(\"cilium_%s\", encapProto)); err == nil {\n\t\tcDefinesMap[\"ENCAP_IFINDEX\"] = fmt.Sprintf(\"%d\", tunnelDev.Attrs().Index)\n\t}\n\n\tcDefinesMap[\"HOST_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameHost))\n\tcDefinesMap[\"WORLD_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameWorld))\n\tif option.Config.IsDualStack() {\n\t\tcDefinesMap[\"WORLD_IPV4_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameWorldIPv4))\n\t\tcDefinesMap[\"WORLD_IPV6_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameWorldIPv6))\n\t} else {\n\t\tworldID := identity.GetReservedID(labels.IDNameWorld)\n\t\tcDefinesMap[\"WORLD_IPV4_ID\"] = fmt.Sprintf(\"%d\", worldID)\n\t\tcDefinesMap[\"WORLD_IPV6_ID\"] = fmt.Sprintf(\"%d\", worldID)\n\t}\n\tcDefinesMap[\"HEALTH_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameHealth))\n\tcDefinesMap[\"UNMANAGED_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameUnmanaged))\n\tcDefinesMap[\"INIT_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameInit))\n\tcDefinesMap[\"LOCAL_NODE_ID\"] = fmt.Sprintf(\"%d\", identity.GetLocalNodeID())\n\tcDefinesMap[\"REMOTE_NODE_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameRemoteNode))\n\tcDefinesMap[\"KUBE_APISERVER_NODE_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameKubeAPIServer))\n\tcDefinesMap[\"CILIUM_LB_SERVICE_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.ServiceMapMaxEntries)\n\tcDefinesMap[\"CILIUM_LB_BACKENDS_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.ServiceBackEndMapMaxEntries)\n\tcDefinesMap[\"CILIUM_LB_REV_NAT_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.RevNatMapMaxEntries)\n\tcDefinesMap[\"CILIUM_LB_AFFINITY_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.AffinityMapMaxEntries)\n\tcDefinesMap[\"CILIUM_LB_SOURCE_RANGE_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.SourceRangeMapMaxEntries)\n\tcDefinesMap[\"CILIUM_LB_MAGLEV_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.MaglevMapMaxEntries)\n\n\tcDefinesMap[\"TUNNEL_MAP\"] = tunnel.MapName\n\tcDefinesMap[\"TUNNEL_ENDPOINT_MAP_SIZE\"] = fmt.Sprintf(\"%d\", tunnel.MaxEntries)\n\tcDefinesMap[\"ENDPOINTS_MAP\"] = lxcmap.MapName\n\tcDefinesMap[\"ENDPOINTS_MAP_SIZE\"] = fmt.Sprintf(\"%d\", lxcmap.MaxEntries)\n\tcDefinesMap[\"METRICS_MAP\"] = metricsmap.MapName\n\tcDefinesMap[\"METRICS_MAP_SIZE\"] = fmt.Sprintf(\"%d\", metricsmap.MaxEntries)\n\tcDefinesMap[\"POLICY_MAP_SIZE\"] = fmt.Sprintf(\"%d\", policymap.MaxEntries)\n\tcDefinesMap[\"AUTH_MAP\"] = authmap.MapName\n\tcDefinesMap[\"AUTH_MAP_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.AuthMapEntries)\n\tcDefinesMap[\"CONFIG_MAP\"] = configmap.MapName\n\tcDefinesMap[\"CONFIG_MAP_SIZE\"] = fmt.Sprintf(\"%d\", configmap.MaxEntries)\n\tcDefinesMap[\"IPCACHE_MAP\"] = ipcachemap.Name\n\tcDefinesMap[\"IPCACHE_MAP_SIZE\"] = fmt.Sprintf(\"%d\", ipcachemap.MaxEntries)\n\tcDefinesMap[\"NODE_MAP\"] = nodemap.MapName\n\tcDefinesMap[\"NODE_MAP_SIZE\"] = fmt.Sprintf(\"%d\", nodemap.MaxEntries)\n\tcDefinesMap[\"SRV6_VRF_MAP4\"] = srv6map.VRFMapName4\n\tcDefinesMap[\"SRV6_VRF_MAP6\"] = srv6map.VRFMapName6\n\tcDefinesMap[\"SRV6_POLICY_MAP4\"] = srv6map.PolicyMapName4\n\tcDefinesMap[\"SRV6_POLICY_MAP6\"] = srv6map.PolicyMapName6\n\tcDefinesMap[\"SRV6_SID_MAP\"] = srv6map.SIDMapName\n\tcDefinesMap[\"SRV6_STATE_MAP4\"] = srv6map.StateMapName4\n\tcDefinesMap[\"SRV6_STATE_MAP6\"] = srv6map.StateMapName6\n\tcDefinesMap[\"SRV6_VRF_MAP_SIZE\"] = fmt.Sprintf(\"%d\", srv6map.MaxVRFEntries)\n\tcDefinesMap[\"SRV6_POLICY_MAP_SIZE\"] = fmt.Sprintf(\"%d\", srv6map.MaxPolicyEntries)\n\tcDefinesMap[\"SRV6_SID_MAP_SIZE\"] = fmt.Sprintf(\"%d\", srv6map.MaxSIDEntries)\n\tcDefinesMap[\"SRV6_STATE_MAP_SIZE\"] = fmt.Sprintf(\"%d\", srv6map.MaxStateEntries)\n\tcDefinesMap[\"WORLD_CIDRS4_MAP\"] = worldcidrsmap.MapName4\n\tcDefinesMap[\"WORLD_CIDRS4_MAP_SIZE\"] = fmt.Sprintf(\"%d\", worldcidrsmap.MapMaxEntries)\n\tcDefinesMap[\"POLICY_PROG_MAP_SIZE\"] = fmt.Sprintf(\"%d\", policymap.PolicyCallMaxEntries)\n\tcDefinesMap[\"L2_RESPONSER_MAP4_SIZE\"] = fmt.Sprintf(\"%d\", l2respondermap.DefaultMaxEntries)\n\tcDefinesMap[\"ENCRYPT_MAP\"] = encrypt.MapName\n\tcDefinesMap[\"L2_RESPONDER_MAP4\"] = l2respondermap.MapName\n\tcDefinesMap[\"CT_CONNECTION_LIFETIME_TCP\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutTCP.Seconds()))\n\tcDefinesMap[\"CT_CONNECTION_LIFETIME_NONTCP\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutAny.Seconds()))\n\tcDefinesMap[\"CT_SERVICE_LIFETIME_TCP\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutSVCTCP.Seconds()))\n\tcDefinesMap[\"CT_SERVICE_LIFETIME_NONTCP\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutSVCAny.Seconds()))\n\tcDefinesMap[\"CT_SERVICE_CLOSE_REBALANCE\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutSVCTCPGrace.Seconds()))\n\tcDefinesMap[\"CT_SYN_TIMEOUT\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutSYN.Seconds()))\n\tcDefinesMap[\"CT_CLOSE_TIMEOUT\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutFIN.Seconds()))\n\tcDefinesMap[\"CT_REPORT_INTERVAL\"] = fmt.Sprintf(\"%d\", int64(option.Config.MonitorAggregationInterval.Seconds()))\n\tcDefinesMap[\"CT_REPORT_FLAGS\"] = fmt.Sprintf(\"%#04x\", int64(option.Config.MonitorAggregationFlags))\n\tcDefinesMap[\"CT_TAIL_CALL_BUFFER4\"] = \"cilium_tail_call_buffer4\"\n\tcDefinesMap[\"CT_TAIL_CALL_BUFFER6\"] = \"cilium_tail_call_buffer6\"\n\tcDefinesMap[\"PER_CLUSTER_CT_TCP4\"] = \"cilium_per_cluster_ct_tcp4\"\n\tcDefinesMap[\"PER_CLUSTER_CT_TCP6\"] = \"cilium_per_cluster_ct_tcp6\"\n\tcDefinesMap[\"PER_CLUSTER_CT_ANY4\"] = \"cilium_per_cluster_ct_any4\"\n\tcDefinesMap[\"PER_CLUSTER_CT_ANY6\"] = \"cilium_per_cluster_ct_any6\"\n\tcDefinesMap[\"PER_CLUSTER_SNAT_MAPPING_IPV4\"] = \"cilium_per_cluster_snat_v4_external\"\n\tcDefinesMap[\"PER_CLUSTER_SNAT_MAPPING_IPV6\"] = \"cilium_per_cluster_snat_v6_external\"\n\n\tif option.Config.PreAllocateMaps {\n\t\tcDefinesMap[\"PREALLOCATE_MAPS\"] = \"1\"\n\t}\n\n\tcDefinesMap[\"EVENTS_MAP\"] = eventsmap.MapName\n\tcDefinesMap[\"SIGNAL_MAP\"] = signalmap.MapName\n\tcDefinesMap[\"POLICY_CALL_MAP\"] = policymap.PolicyCallMapName\n\tif option.Config.EnableEnvoyConfig {\n\t\tcDefinesMap[\"POLICY_EGRESSCALL_MAP\"] = policymap.PolicyEgressCallMapName\n\t}\n\tcDefinesMap[\"LB6_REVERSE_NAT_MAP\"] = \"cilium_lb6_reverse_nat\"\n\tcDefinesMap[\"LB6_SERVICES_MAP_V2\"] = \"cilium_lb6_services_v2\"\n\tcDefinesMap[\"LB6_BACKEND_MAP\"] = \"cilium_lb6_backends_v3\"\n\tcDefinesMap[\"LB6_REVERSE_NAT_SK_MAP\"] = lbmap.SockRevNat6MapName\n\tcDefinesMap[\"LB6_REVERSE_NAT_SK_MAP_SIZE\"] = fmt.Sprintf(\"%d\", lbmap.MaxSockRevNat6MapEntries)\n\tcDefinesMap[\"LB4_REVERSE_NAT_MAP\"] = \"cilium_lb4_reverse_nat\"\n\tcDefinesMap[\"LB4_SERVICES_MAP_V2\"] = \"cilium_lb4_services_v2\"\n\tcDefinesMap[\"LB4_BACKEND_MAP\"] = \"cilium_lb4_backends_v3\"\n\tcDefinesMap[\"LB4_REVERSE_NAT_SK_MAP\"] = lbmap.SockRevNat4MapName\n\tcDefinesMap[\"LB4_REVERSE_NAT_SK_MAP_SIZE\"] = fmt.Sprintf(\"%d\", lbmap.MaxSockRevNat4MapEntries)\n\n\tif option.Config.EnableSessionAffinity {\n\t\tcDefinesMap[\"ENABLE_SESSION_AFFINITY\"] = \"1\"\n\t\tcDefinesMap[\"LB_AFFINITY_MATCH_MAP\"] = lbmap.AffinityMatchMapName\n\t\tif option.Config.EnableIPv4 {\n\t\t\tcDefinesMap[\"LB4_AFFINITY_MAP\"] = lbmap.Affinity4MapName\n\t\t}\n\t\tif option.Config.EnableIPv6 {\n\t\t\tcDefinesMap[\"LB6_AFFINITY_MAP\"] = lbmap.Affinity6MapName\n\t\t}\n\t}\n\n\tcDefinesMap[\"TRACE_PAYLOAD_LEN\"] = fmt.Sprintf(\"%dULL\", option.Config.TracePayloadlen)\n\tcDefinesMap[\"MTU\"] = fmt.Sprintf(\"%d\", cfg.MtuConfig.GetDeviceMTU())\n\n\tif option.Config.EnableIPv4 {\n\t\tcDefinesMap[\"ENABLE_IPV4\"] = \"1\"\n\t}\n\n\tif option.Config.EnableIPv6 {\n\t\tcDefinesMap[\"ENABLE_IPV6\"] = \"1\"\n\t}\n\n\tif option.Config.EnableSRv6 {\n\t\tcDefinesMap[\"ENABLE_SRV6\"] = \"1\"\n\t\tif option.Config.SRv6EncapMode != \"reduced\" {\n\t\t\tcDefinesMap[\"ENABLE_SRV6_SRH_ENCAP\"] = \"1\"\n\t\t}\n\t}\n\n\tif option.Config.EnableSCTP {\n\t\tcDefinesMap[\"ENABLE_SCTP\"] = \"1\"\n\t}\n\n\tif option.Config.EnableIPSec {\n\t\tcDefinesMap[\"ENABLE_IPSEC\"] = \"1\"\n\t}\n\n\tif option.Config.EnableWireguard {\n\t\tcDefinesMap[\"ENABLE_WIREGUARD\"] = \"1\"\n\t\tifindex, err := link.GetIfIndex(wgtypes.IfaceName)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcDefinesMap[\"WG_IFINDEX\"] = fmt.Sprintf(\"%d\", ifindex)\n\n\t\tif option.Config.EncryptNode {\n\t\t\tcDefinesMap[\"ENABLE_NODE_ENCRYPTION\"] = \"1\"\n\t\t}\n\t}\n\n\tif option.Config.EnableL2Announcements {\n\t\tcDefinesMap[\"ENABLE_L2_ANNOUNCEMENTS\"] = \"1\"\n\t\t// If the agent is down for longer than the lease duration, stop responding\n\t\tcDefinesMap[\"L2_ANNOUNCEMENTS_MAX_LIVENESS\"] = fmt.Sprintf(\"%dULL\", option.Config.L2AnnouncerLeaseDuration.Nanoseconds())\n\t}\n\n\tif option.Config.EnableEncryptionStrictMode {\n\t\tcDefinesMap[\"ENCRYPTION_STRICT_MODE\"] = \"1\"\n\n\t\t// when parsing the user input we only accept ipv4 addresses\n\t\tcDefinesMap[\"STRICT_IPV4_NET\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPAddrToHost32(option.Config.EncryptionStrictModeCIDR.Addr()))\n\t\tcDefinesMap[\"STRICT_IPV4_NET_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.EncryptionStrictModeCIDR.Bits())\n\n\t\tcDefinesMap[\"IPV4_ENCRYPT_IFACE\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(node.GetIPv4()))\n\n\t\tipv4Interface, ok := netip.AddrFromSlice(node.GetIPv4().To4())\n\t\tif !ok {\n\t\t\treturn fmt.Errorf(\"unable to parse node IPv4 address %s\", node.GetIPv4())\n\t\t}\n\n\t\tif option.Config.EncryptionStrictModeCIDR.Contains(ipv4Interface) {\n\t\t\tif !option.Config.EncryptionStrictModeAllowRemoteNodeIdentities {\n\t\t\t\treturn fmt.Errorf(`encryption strict mode is enabled but the node's IPv4 address is within the strict CIDR range.\n\t\t\t\tThis will cause the node to drop all traffic.\n\t\t\t\tPlease either disable encryption or set --encryption-strict-mode-allow-dynamic-lookup=true`)\n\t\t\t}\n\t\t\tcDefinesMap[\"STRICT_IPV4_OVERLAPPING_CIDR\"] = \"1\"\n\t\t}\n\t}\n\n\tif option.Config.EnableBPFTProxy {\n\t\tcDefinesMap[\"ENABLE_TPROXY\"] = \"1\"\n\t}\n\n\tif option.Config.EnableXDPPrefilter {\n\t\tcDefinesMap[\"ENABLE_PREFILTER\"] = \"1\"\n\t}\n\n\tif option.Config.EnableEndpointRoutes {\n\t\tcDefinesMap[\"ENABLE_ENDPOINT_ROUTES\"] = \"1\"\n\t}\n\n\tif option.Config.EnableEnvoyConfig {\n\t\tcDefinesMap[\"ENABLE_L7_LB\"] = \"1\"\n\t}\n\n\tif option.Config.EnableSocketLB {\n\t\tif option.Config.BPFSocketLBHostnsOnly {\n\t\t\tcDefinesMap[\"ENABLE_SOCKET_LB_HOST_ONLY\"] = \"1\"\n\t\t} else {\n\t\t\tcDefinesMap[\"ENABLE_SOCKET_LB_FULL\"] = \"1\"\n\t\t}\n\t\tif option.Config.EnableSocketLBPeer {\n\t\t\tcDefinesMap[\"ENABLE_SOCKET_LB_PEER\"] = \"1\"\n\t\t}\n\t\tif option.Config.EnableSocketLBTracing {\n\t\t\tcDefinesMap[\"TRACE_SOCK_NOTIFY\"] = \"1\"\n\t\t}\n\n\t\tif cookie, err := netns.GetNetNSCookie(); err == nil {\n\t\t\t// When running in nested environments (e.g. Kind), cilium-agent does\n\t\t\t// not run in the host netns. So, in such cases the cookie comparison\n\t\t\t// based on bpf_get_netns_cookie(NULL) for checking whether a socket\n\t\t\t// belongs to a host netns does not work.\n\t\t\t//\n\t\t\t// To fix this, we derive the cookie of the netns in which cilium-agent\n\t\t\t// runs via getsockopt(...SO_NETNS_COOKIE...) and then use it in the\n\t\t\t// check above. This is based on an assumption that cilium-agent\n\t\t\t// always runs with \"hostNetwork: true\".\n\t\t\tcDefinesMap[\"HOST_NETNS_COOKIE\"] = fmt.Sprintf(\"%d\", cookie)\n\t\t}\n\t}\n\n\tcDefinesMap[\"NAT_46X64_PREFIX_0\"] = \"0\"\n\tcDefinesMap[\"NAT_46X64_PREFIX_1\"] = \"0\"\n\tcDefinesMap[\"NAT_46X64_PREFIX_2\"] = \"0\"\n\tcDefinesMap[\"NAT_46X64_PREFIX_3\"] = \"0\"\n\n\tif option.Config.EnableNodePort {\n\t\tif option.Config.EnableHealthDatapath {\n\t\t\tcDefinesMap[\"ENABLE_HEALTH_CHECK\"] = \"1\"\n\t\t}\n\t\tif option.Config.EnableMKE && option.Config.EnableSocketLB {\n\t\t\tcDefinesMap[\"ENABLE_MKE\"] = \"1\"\n\t\t\tcDefinesMap[\"MKE_HOST\"] = fmt.Sprintf(\"%d\", option.HostExtensionMKE)\n\t\t}\n\t\tif option.Config.EnableRecorder {\n\t\t\tcDefinesMap[\"ENABLE_CAPTURE\"] = \"1\"\n\t\t\tif option.Config.EnableIPv4 {\n\t\t\t\tcDefinesMap[\"CAPTURE4_RULES\"] = recorder.MapNameWcard4\n\t\t\t\tcDefinesMap[\"CAPTURE4_SIZE\"] = fmt.Sprintf(\"%d\", recorder.MapSize)\n\t\t\t}\n\t\t\tif option.Config.EnableIPv6 {\n\t\t\t\tcDefinesMap[\"CAPTURE6_RULES\"] = recorder.MapNameWcard6\n\t\t\t\tcDefinesMap[\"CAPTURE6_SIZE\"] = fmt.Sprintf(\"%d\", recorder.MapSize)\n\t\t\t}\n\t\t}\n\t\tcDefinesMap[\"ENABLE_NODEPORT\"] = \"1\"\n\t\tif option.Config.EnableIPv4 {\n\t\t\tcDefinesMap[\"NODEPORT_NEIGH4\"] = neighborsmap.Map4Name\n\t\t\tcDefinesMap[\"NODEPORT_NEIGH4_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.NeighMapEntriesGlobal)\n\t\t\tif option.Config.EnableHealthDatapath {\n\t\t\t\tcDefinesMap[\"LB4_HEALTH_MAP\"] = lbmap.HealthProbe4MapName\n\t\t\t}\n\t\t}\n\t\tif option.Config.EnableIPv6 {\n\t\t\tcDefinesMap[\"NODEPORT_NEIGH6\"] = neighborsmap.Map6Name\n\t\t\tcDefinesMap[\"NODEPORT_NEIGH6_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.NeighMapEntriesGlobal)\n\t\t\tif option.Config.EnableHealthDatapath {\n\t\t\t\tcDefinesMap[\"LB6_HEALTH_MAP\"] = lbmap.HealthProbe6MapName\n\t\t\t}\n\t\t}\n\t\tif option.Config.EnableNat46X64Gateway {\n\t\t\tcDefinesMap[\"ENABLE_NAT_46X64_GATEWAY\"] = \"1\"\n\t\t\tbase := option.Config.IPv6NAT46x64CIDRBase.AsSlice()\n\t\t\tcDefinesMap[\"NAT_46X64_PREFIX_0\"] = fmt.Sprintf(\"%d\", base[0])\n\t\t\tcDefinesMap[\"NAT_46X64_PREFIX_1\"] = fmt.Sprintf(\"%d\", base[1])\n\t\t\tcDefinesMap[\"NAT_46X64_PREFIX_2\"] = fmt.Sprintf(\"%d\", base[2])\n\t\t\tcDefinesMap[\"NAT_46X64_PREFIX_3\"] = fmt.Sprintf(\"%d\", base[3])\n\t\t}\n\t\tif option.Config.NodePortNat46X64 {\n\t\t\tcDefinesMap[\"ENABLE_NAT_46X64\"] = \"1\"\n\t\t}\n\t\tconst (\n\t\t\tdsrEncapInv = iota\n\t\t\tdsrEncapNone\n\t\t\tdsrEncapIPIP\n\t\t\tdsrEncapGeneve\n\t\t)\n\t\tconst (\n\t\t\tdsrL4XlateInv = iota\n\t\t\tdsrL4XlateFrontend\n\t\t\tdsrL4XlateBackend\n\t\t)\n\t\tcDefinesMap[\"DSR_ENCAP_IPIP\"] = fmt.Sprintf(\"%d\", dsrEncapIPIP)\n\t\tcDefinesMap[\"DSR_ENCAP_GENEVE\"] = fmt.Sprintf(\"%d\", dsrEncapGeneve)\n\t\tcDefinesMap[\"DSR_ENCAP_NONE\"] = fmt.Sprintf(\"%d\", dsrEncapNone)\n\t\tcDefinesMap[\"DSR_XLATE_FRONTEND\"] = fmt.Sprintf(\"%d\", dsrL4XlateFrontend)\n\t\tcDefinesMap[\"DSR_XLATE_BACKEND\"] = fmt.Sprintf(\"%d\", dsrL4XlateBackend)\n\t\tif option.Config.NodePortMode == option.NodePortModeDSR ||\n\t\t\toption.Config.NodePortMode == option.NodePortModeHybrid {\n\t\t\tcDefinesMap[\"ENABLE_DSR\"] = \"1\"\n\t\t\tif option.Config.EnablePMTUDiscovery {\n\t\t\t\tcDefinesMap[\"ENABLE_DSR_ICMP_ERRORS\"] = \"1\"\n\t\t\t}\n\t\t\tif option.Config.NodePortMode == option.NodePortModeHybrid {\n\t\t\t\tcDefinesMap[\"ENABLE_DSR_HYBRID\"] = \"1\"\n\t\t\t}\n\t\t\tif option.Config.LoadBalancerDSRDispatch == option.DSRDispatchOption {\n\t\t\t\tcDefinesMap[\"DSR_ENCAP_MODE\"] = fmt.Sprintf(\"%d\", dsrEncapNone)\n\t\t\t} else if option.Config.LoadBalancerDSRDispatch == option.DSRDispatchIPIP {\n\t\t\t\tcDefinesMap[\"DSR_ENCAP_MODE\"] = fmt.Sprintf(\"%d\", dsrEncapIPIP)\n\t\t\t} else if option.Config.LoadBalancerDSRDispatch == option.DSRDispatchGeneve {\n\t\t\t\tcDefinesMap[\"DSR_ENCAP_MODE\"] = fmt.Sprintf(\"%d\", dsrEncapGeneve)\n\t\t\t}\n\t\t\tif option.Config.LoadBalancerDSRDispatch == option.DSRDispatchIPIP {\n\t\t\t\tif option.Config.LoadBalancerDSRL4Xlate == option.DSRL4XlateFrontend {\n\t\t\t\t\tcDefinesMap[\"DSR_XLATE_MODE\"] = fmt.Sprintf(\"%d\", dsrL4XlateFrontend)\n\t\t\t\t} else if option.Config.LoadBalancerDSRL4Xlate == option.DSRL4XlateBackend {\n\t\t\t\t\tcDefinesMap[\"DSR_XLATE_MODE\"] = fmt.Sprintf(\"%d\", dsrL4XlateBackend)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tcDefinesMap[\"DSR_XLATE_MODE\"] = fmt.Sprintf(\"%d\", dsrL4XlateInv)\n\t\t\t}\n\t\t} else {\n\t\t\tcDefinesMap[\"DSR_ENCAP_MODE\"] = fmt.Sprintf(\"%d\", dsrEncapInv)\n\t\t\tcDefinesMap[\"DSR_XLATE_MODE\"] = fmt.Sprintf(\"%d\", dsrL4XlateInv)\n\t\t}\n\t\tif option.Config.EnableIPv4 {\n\t\t\tif option.Config.LoadBalancerRSSv4CIDR != \"\" {\n\t\t\t\tipv4 := byteorder.NetIPv4ToHost32(option.Config.LoadBalancerRSSv4.IP)\n\t\t\t\tones, _ := option.Config.LoadBalancerRSSv4.Mask.Size()\n\t\t\t\tcDefinesMap[\"IPV4_RSS_PREFIX\"] = fmt.Sprintf(\"%d\", ipv4)\n\t\t\t\tcDefinesMap[\"IPV4_RSS_PREFIX_BITS\"] = fmt.Sprintf(\"%d\", ones)\n\t\t\t} else {\n\t\t\t\tcDefinesMap[\"IPV4_RSS_PREFIX\"] = \"IPV4_DIRECT_ROUTING\"\n\t\t\t\tcDefinesMap[\"IPV4_RSS_PREFIX_BITS\"] = \"32\"\n\t\t\t}\n\t\t}\n\t\tif option.Config.EnableIPv6 {\n\t\t\tif option.Config.LoadBalancerRSSv6CIDR != \"\" {\n\t\t\t\tipv6 := option.Config.LoadBalancerRSSv6.IP\n\t\t\t\tones, _ := option.Config.LoadBalancerRSSv6.Mask.Size()\n\t\t\t\textraMacrosMap[\"IPV6_RSS_PREFIX\"] = ipv6.String()\n\t\t\t\tfw.WriteString(FmtDefineAddress(\"IPV6_RSS_PREFIX\", ipv6))\n\t\t\t\tcDefinesMap[\"IPV6_RSS_PREFIX_BITS\"] = fmt.Sprintf(\"%d\", ones)\n\t\t\t} else {\n\t\t\t\tcDefinesMap[\"IPV6_RSS_PREFIX\"] = \"IPV6_DIRECT_ROUTING\"\n\t\t\t\tcDefinesMap[\"IPV6_RSS_PREFIX_BITS\"] = \"128\"\n\t\t\t}\n\t\t}\n\t\tif option.Config.NodePortAcceleration != option.NodePortAccelerationDisabled {\n\t\t\tcDefinesMap[\"ENABLE_NODEPORT_ACCELERATION\"] = \"1\"\n\t\t}\n\t\tif !option.Config.EnableHostLegacyRouting {\n\t\t\tcDefinesMap[\"ENABLE_HOST_ROUTING\"] = \"1\"\n\t\t}\n\t\tif option.Config.EnableSVCSourceRangeCheck {\n\t\t\tcDefinesMap[\"ENABLE_SRC_RANGE_CHECK\"] = \"1\"\n\t\t\tif option.Config.EnableIPv4 {\n\t\t\t\tcDefinesMap[\"LB4_SRC_RANGE_MAP\"] = lbmap.SourceRange4MapName\n\t\t\t\tcDefinesMap[\"LB4_SRC_RANGE_MAP_SIZE\"] =\n\t\t\t\t\tfmt.Sprintf(\"%d\", lbmap.SourceRange4Map.MaxEntries())\n\t\t\t}\n\t\t\tif option.Config.EnableIPv6 {\n\t\t\t\tcDefinesMap[\"LB6_SRC_RANGE_MAP\"] = lbmap.SourceRange6MapName\n\t\t\t\tcDefinesMap[\"LB6_SRC_RANGE_MAP_SIZE\"] =\n\t\t\t\t\tfmt.Sprintf(\"%d\", lbmap.SourceRange6Map.MaxEntries())\n\t\t\t}\n\t\t}\n\n\t\tcDefinesMap[\"NODEPORT_PORT_MIN\"] = fmt.Sprintf(\"%d\", option.Config.NodePortMin)\n\t\tcDefinesMap[\"NODEPORT_PORT_MAX\"] = fmt.Sprintf(\"%d\", option.Config.NodePortMax)\n\t\tcDefinesMap[\"NODEPORT_PORT_MIN_NAT\"] = fmt.Sprintf(\"%d\", option.Config.NodePortMax+1)\n\t\tcDefinesMap[\"NODEPORT_PORT_MAX_NAT\"] = \"65535\"\n\t}\n\n\tmacByIfIndexMacro, isL3DevMacro, err := devMacros()\n\tif err != nil {\n\t\treturn err\n\t}\n\tcDefinesMap[\"NATIVE_DEV_MAC_BY_IFINDEX(IFINDEX)\"] = macByIfIndexMacro\n\tcDefinesMap[\"IS_L3_DEV(ifindex)\"] = isL3DevMacro\n\n\tconst (\n\t\tselectionRandom = iota + 1\n\t\tselectionMaglev\n\t)\n\tcDefinesMap[\"LB_SELECTION_RANDOM\"] = fmt.Sprintf(\"%d\", selectionRandom)\n\tcDefinesMap[\"LB_SELECTION_MAGLEV\"] = fmt.Sprintf(\"%d\", selectionMaglev)\n\tif option.Config.NodePortAlg == option.NodePortAlgRandom {\n\t\tcDefinesMap[\"LB_SELECTION\"] = fmt.Sprintf(\"%d\", selectionRandom)\n\t} else if option.Config.NodePortAlg == option.NodePortAlgMaglev {\n\t\tcDefinesMap[\"LB_SELECTION\"] = fmt.Sprintf(\"%d\", selectionMaglev)\n\t\tcDefinesMap[\"LB_MAGLEV_LUT_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.MaglevTableSize)\n\t\tif option.Config.EnableIPv6 {\n\t\t\tcDefinesMap[\"LB6_MAGLEV_MAP_OUTER\"] = lbmap.MaglevOuter6MapName\n\t\t}\n\t\tif option.Config.EnableIPv4 {\n\t\t\tcDefinesMap[\"LB4_MAGLEV_MAP_OUTER\"] = lbmap.MaglevOuter4MapName\n\t\t}\n\t}\n\tcDefinesMap[\"HASH_INIT4_SEED\"] = fmt.Sprintf(\"%d\", maglev.SeedJhash0)\n\tcDefinesMap[\"HASH_INIT6_SEED\"] = fmt.Sprintf(\"%d\", maglev.SeedJhash1)\n\n\tif option.Config.DirectRoutingDeviceRequired() {\n\t\tdirectRoutingIface := option.Config.DirectRoutingDevice\n\t\tdirectRoutingIfIndex, err := link.GetIfIndex(directRoutingIface)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcDefinesMap[\"DIRECT_ROUTING_DEV_IFINDEX\"] = fmt.Sprintf(\"%d\", directRoutingIfIndex)\n\t\tif option.Config.EnableIPv4 {\n\t\t\tip, ok := node.GetNodePortIPv4AddrsWithDevices()[directRoutingIface]\n\t\t\tif !ok {\n\t\t\t\tlog.WithFields(logrus.Fields{\n\t\t\t\t\t\"directRoutingIface\": directRoutingIface,\n\t\t\t\t}).Fatal(\"Direct routing device's IPv4 address not found\")\n\t\t\t}\n\n\t\t\tipv4 := byteorder.NetIPv4ToHost32(ip)\n\t\t\tcDefinesMap[\"IPV4_DIRECT_ROUTING\"] = fmt.Sprintf(\"%d\", ipv4)\n\t\t}\n\n\t\tif option.Config.EnableIPv6 {\n\t\t\tdirectRoutingIPv6, ok := node.GetNodePortIPv6AddrsWithDevices()[directRoutingIface]\n\t\t\tif !ok {\n\t\t\t\tlog.WithFields(logrus.Fields{\n\t\t\t\t\t\"directRoutingIface\": directRoutingIface,\n\t\t\t\t}).Fatal(\"Direct routing device's IPv6 address not found\")\n\t\t\t}\n\n\t\t\textraMacrosMap[\"IPV6_DIRECT_ROUTING\"] = directRoutingIPv6.String()\n\t\t\tfw.WriteString(FmtDefineAddress(\"IPV6_DIRECT_ROUTING\", directRoutingIPv6))\n\t\t}\n\t} else {\n\t\tvar directRoutingIPv6 net.IP\n\t\tcDefinesMap[\"DIRECT_ROUTING_DEV_IFINDEX\"] = \"0\"\n\t\tif option.Config.EnableIPv4 {\n\t\t\tcDefinesMap[\"IPV4_DIRECT_ROUTING\"] = \"0\"\n\t\t}\n\t\tif option.Config.EnableIPv6 {\n\t\t\textraMacrosMap[\"IPV6_DIRECT_ROUTING\"] = directRoutingIPv6.String()\n\t\t\tfw.WriteString(FmtDefineAddress(\"IPV6_DIRECT_ROUTING\", directRoutingIPv6))\n\t\t}\n\t}\n\n\tif option.Config.ResetQueueMapping {\n\t\tcDefinesMap[\"RESET_QUEUES\"] = \"1\"\n\t}\n\n\tif option.Config.EnableBandwidthManager {\n\t\tcDefinesMap[\"ENABLE_BANDWIDTH_MANAGER\"] = \"1\"\n\t\tcDefinesMap[\"THROTTLE_MAP\"] = bwmap.MapName\n\t\tcDefinesMap[\"THROTTLE_MAP_SIZE\"] = fmt.Sprintf(\"%d\", bwmap.MapSize)\n\t}\n\n\tif option.Config.EnableHostFirewall {\n\t\tcDefinesMap[\"ENABLE_HOST_FIREWALL\"] = \"1\"\n\t}\n\n\tif option.Config.EnableIPSec {\n\t\tnodeAddress := node.GetIPv4()\n\t\tif nodeAddress == nil {\n\t\t\treturn errors.New(\"external IPv4 node address is required when IPSec is enabled, but none found\")\n\t\t}\n\n\t\ta := byteorder.NetIPv4ToHost32(nodeAddress)\n\t\tcDefinesMap[\"IPV4_ENCRYPT_IFACE\"] = fmt.Sprintf(\"%d\", a)\n\t\tif iface := option.Config.EncryptInterface; len(iface) != 0 {\n\t\t\tlink, err := netlink.LinkByName(iface[0])\n\t\t\tif err == nil {\n\t\t\t\tcDefinesMap[\"ENCRYPT_IFACE\"] = fmt.Sprintf(\"%d\", link.Attrs().Index)\n\t\t\t}\n\t\t}\n\t}\n\n\tif option.Config.EnableNodePort {\n\t\tif option.Config.EnableIPv4 {\n\t\t\tcDefinesMap[\"SNAT_MAPPING_IPV4\"] = nat.MapNameSnat4Global\n\t\t\tcDefinesMap[\"SNAT_MAPPING_IPV4_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.NATMapEntriesGlobal)\n\t\t}\n\n\t\tif option.Config.EnableIPv6 {\n\t\t\tcDefinesMap[\"SNAT_MAPPING_IPV6\"] = nat.MapNameSnat6Global\n\t\t\tcDefinesMap[\"SNAT_MAPPING_IPV6_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.NATMapEntriesGlobal)\n\t\t}\n\n\t\tif option.Config.EnableBPFMasquerade {\n\t\t\tif option.Config.EnableIPv4Masquerade {\n\t\t\t\tcDefinesMap[\"ENABLE_MASQUERADE_IPV4\"] = \"1\"\n\n\t\t\t\t// ip-masq-agent depends on bpf-masq\n\t\t\t\tvar excludeCIDR *cidr.CIDR\n\t\t\t\tif option.Config.EnableIPMasqAgent {\n\t\t\t\t\tcDefinesMap[\"ENABLE_IP_MASQ_AGENT_IPV4\"] = \"1\"\n\t\t\t\t\tcDefinesMap[\"IP_MASQ_AGENT_IPV4\"] = ipmasq.MapNameIPv4\n\n\t\t\t\t\t// native-routing-cidr is optional with ip-masq-agent and may be nil\n\t\t\t\t\texcludeCIDR = option.Config.GetIPv4NativeRoutingCIDR()\n\t\t\t\t} else {\n\t\t\t\t\texcludeCIDR = datapath.RemoteSNATDstAddrExclusionCIDRv4()\n\t\t\t\t}\n\n\t\t\t\tif excludeCIDR != nil {\n\t\t\t\t\tcDefinesMap[\"IPV4_SNAT_EXCLUSION_DST_CIDR\"] =\n\t\t\t\t\t\tfmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(excludeCIDR.IP))\n\t\t\t\t\tones, _ := excludeCIDR.Mask.Size()\n\t\t\t\t\tcDefinesMap[\"IPV4_SNAT_EXCLUSION_DST_CIDR_LEN\"] = fmt.Sprintf(\"%d\", ones)\n\t\t\t\t}\n\t\t\t}\n\t\t\tif option.Config.EnableIPv6Masquerade {\n\t\t\t\tcDefinesMap[\"ENABLE_MASQUERADE_IPV6\"] = \"1\"\n\n\t\t\t\tvar excludeCIDR *cidr.CIDR\n\t\t\t\tif option.Config.EnableIPMasqAgent {\n\t\t\t\t\tcDefinesMap[\"ENABLE_IP_MASQ_AGENT_IPV6\"] = \"1\"\n\t\t\t\t\tcDefinesMap[\"IP_MASQ_AGENT_IPV6\"] = ipmasq.MapNameIPv6\n\n\t\t\t\t\texcludeCIDR = option.Config.GetIPv6NativeRoutingCIDR()\n\t\t\t\t} else {\n\t\t\t\t\texcludeCIDR = datapath.RemoteSNATDstAddrExclusionCIDRv6()\n\t\t\t\t}\n\n\t\t\t\tif excludeCIDR != nil {\n\t\t\t\t\textraMacrosMap[\"IPV6_SNAT_EXCLUSION_DST_CIDR\"] = excludeCIDR.IP.String()\n\t\t\t\t\tfw.WriteString(FmtDefineAddress(\"IPV6_SNAT_EXCLUSION_DST_CIDR\", excludeCIDR.IP))\n\t\t\t\t\textraMacrosMap[\"IPV6_SNAT_EXCLUSION_DST_CIDR_MASK\"] = excludeCIDR.Mask.String()\n\t\t\t\t\tfw.WriteString(FmtDefineAddress(\"IPV6_SNAT_EXCLUSION_DST_CIDR_MASK\", excludeCIDR.Mask))\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\tctmap.WriteBPFMacros(fw, nil)\n\t}\n\n\tif option.Config.AllowICMPFragNeeded {\n\t\tcDefinesMap[\"ALLOW_ICMP_FRAG_NEEDED\"] = \"1\"\n\t}\n\n\tif option.Config.ClockSource == option.ClockSourceJiffies {\n\t\tcDefinesMap[\"ENABLE_JIFFIES\"] = \"1\"\n\t}\n\n\tif option.Config.EnableIdentityMark {\n\t\tcDefinesMap[\"ENABLE_IDENTITY_MARK\"] = \"1\"\n\t}\n\n\tif option.Config.EnableHighScaleIPcache {\n\t\tcDefinesMap[\"ENABLE_HIGH_SCALE_IPCACHE\"] = \"1\"\n\t}\n\n\tif option.Config.EnableCustomCalls {\n\t\tcDefinesMap[\"ENABLE_CUSTOM_CALLS\"] = \"1\"\n\t}\n\n\tif option.Config.EnableVTEP {\n\t\tcDefinesMap[\"ENABLE_VTEP\"] = \"1\"\n\t\tcDefinesMap[\"VTEP_MAP\"] = vtep.Name\n\t\tcDefinesMap[\"VTEP_MAP_SIZE\"] = fmt.Sprintf(\"%d\", vtep.MaxEntries)\n\t\tcDefinesMap[\"VTEP_MASK\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(net.IP(option.Config.VtepCidrMask)))\n\n\t}\n\n\tvlanFilter, err := vlanFilterMacros()\n\tif err != nil {\n\t\treturn err\n\t}\n\tcDefinesMap[\"VLAN_FILTER(ifindex, vlan_id)\"] = vlanFilter\n\n\tif option.Config.EnableICMPRules {\n\t\tcDefinesMap[\"ENABLE_ICMP_RULE\"] = \"1\"\n\t}\n\n\tcDefinesMap[\"CIDR_IDENTITY_RANGE_START\"] = fmt.Sprintf(\"%d\", identity.MinLocalIdentity)\n\tcDefinesMap[\"CIDR_IDENTITY_RANGE_END\"] = fmt.Sprintf(\"%d\", identity.MaxLocalIdentity)\n\n\tif option.Config.TunnelingEnabled() {\n\t\tcDefinesMap[\"TUNNEL_MODE\"] = \"1\"\n\t}\n\n\tciliumNetLink, err := netlink.LinkByName(defaults.SecondHostDevice)\n\tif err != nil {\n\t\treturn err\n\t}\n\tcDefinesMap[\"CILIUM_NET_MAC\"] = fmt.Sprintf(\"{.addr=%s}\", mac.CArrayString(ciliumNetLink.Attrs().HardwareAddr))\n\tcDefinesMap[\"HOST_IFINDEX\"] = fmt.Sprintf(\"%d\", ciliumNetLink.Attrs().Index)\n\n\tciliumHostLink, err := netlink.LinkByName(defaults.HostDevice)\n\tif err != nil {\n\t\treturn err\n\t}\n\tcDefinesMap[\"HOST_IFINDEX_MAC\"] = fmt.Sprintf(\"{.addr=%s}\", mac.CArrayString(ciliumHostLink.Attrs().HardwareAddr))\n\tcDefinesMap[\"CILIUM_IFINDEX\"] = fmt.Sprintf(\"%d\", ciliumHostLink.Attrs().Index)\n\n\tephemeralMin, err := getEphemeralPortRangeMin()\n\tif err != nil {\n\t\treturn err\n\t}\n\tcDefinesMap[\"EPHEMERAL_MIN\"] = fmt.Sprintf(\"%d\", ephemeralMin)\n\n\tif err := cDefinesMap.Merge(h.nodeExtraDefines); err != nil {\n\t\treturn err\n\t}\n\n\tfor _, fn := range h.nodeExtraDefineFns {\n\t\tdefines, err := fn()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := cDefinesMap.Merge(defines); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif option.Config.EnableHealthDatapath {\n\t\tif option.Config.IPv4Enabled() {\n\t\t\tipip4, err := netlink.LinkByName(defaults.IPIPv4Device)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tcDefinesMap[\"ENCAP4_IFINDEX\"] = fmt.Sprintf(\"%d\", ipip4.Attrs().Index)\n\t\t}\n\t\tif option.Config.IPv6Enabled() {\n\t\t\tipip6, err := netlink.LinkByName(defaults.IPIPv6Device)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tcDefinesMap[\"ENCAP6_IFINDEX\"] = fmt.Sprintf(\"%d\", ipip6.Attrs().Index)\n\t\t}\n\t}\n\n\t// Since golang maps are unordered, we sort the keys in the map\n\t// to get a consistent written format to the writer. This maintains\n\t// the consistency when we try to calculate hash for a datapath after\n\t// writing the config.\n\tkeys := make([]string, 0, len(cDefinesMap))\n\tfor key := range cDefinesMap {\n\t\tkeys = append(keys, key)\n\t}\n\tsort.Strings(keys)\n\n\tfor _, key := range keys {\n\t\tfmt.Fprintf(fw, \"#define %s %s\\n\", key, cDefinesMap[key])\n\t}\n\n\t// Populate cDefinesMap with extraMacrosMap to get all the configuration\n\t// in the cDefinesMap itself.\n\tfor key, value := range extraMacrosMap {\n\t\tcDefinesMap[key] = value\n\t}\n\n\t// Write the JSON encoded config as base64 encoded commented string to\n\t// the header file.\n\tjsonBytes, err := json.Marshal(cDefinesMap)\n\tif err == nil {\n\t\t// We don't care if some error occurs while marshaling the map.\n\t\t// In such cases we skip embedding the base64 encoded JSON configuration\n\t\t// to the writer.\n\t\tencodedConfig := base64.StdEncoding.EncodeToString(jsonBytes)\n\t\tfmt.Fprintf(fw, \"\\n// JSON_OUTPUT: %s\\n\", encodedConfig)\n\t}\n\n\treturn fw.Flush()\n}", "func (fes *FrontEndService) WriteConfigDropIfNoPeer(conf *string, hasVIP4 bool, hasVIP6 bool) {\n\tif hasVIP4 {\n\t\t*conf += \"protocol static BH4 {\\n\"\n\t\t*conf += \"\\tipv4 { preference 0; };\\n\"\n\t\t*conf += \"\\troute 0.0.0.0/0 blackhole {\\n\"\n\t\t*conf += \"\\t\\tkrt_metric=4294967295;\\n\"\n\t\t*conf += \"\\t};\\n\"\n\t\t*conf += \"}\\n\"\n\t\t*conf += \"\\n\"\n\t}\n\n\tif hasVIP6 {\n\t\t*conf += \"protocol static BH6 {\\n\"\n\t\t*conf += \"\\tipv6 { preference 0; };\\n\"\n\t\t*conf += \"\\troute 0::/0 blackhole {\\n\"\n\t\t*conf += \"\\t\\tkrt_metric=4294967295;\\n\"\n\t\t*conf += \"\\t};\\n\"\n\t\t*conf += \"}\\n\"\n\t\t*conf += \"\\n\"\n\t}\n\treturn\n}", "func (o KubernetesClusterNodePoolLinuxOsConfigSysctlConfigOutput) NetNetfilterNfConntrackMax() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v KubernetesClusterNodePoolLinuxOsConfigSysctlConfig) *int { return v.NetNetfilterNfConntrackMax }).(pulumi.IntPtrOutput)\n}", "func (o KubernetesClusterDefaultNodePoolLinuxOsConfigSysctlConfigOutput) NetNetfilterNfConntrackMax() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v KubernetesClusterDefaultNodePoolLinuxOsConfigSysctlConfig) *int {\n\t\treturn v.NetNetfilterNfConntrackMax\n\t}).(pulumi.IntPtrOutput)\n}", "func (fes *FrontEndService) WriteConfig(ctx context.Context, cancel context.CancelFunc) {\n\tfile, err := os.Create(fes.birdConfFile)\n\tif err != nil {\n\t\tlogrus.Fatalf(\"FrontEndService: failed to create %v, err: %v\", fes.birdConfFile, err)\n\t\tcancel()\n\t}\n\tdefer file.Close()\n\n\t//conf := \"include \\\"bird-common.conf\\\";\\n\"\n\t//conf += \"\\n\"\n\tconf := \"\"\n\tfes.WriteConfigBase(&conf)\n\thasVIP4, hasVIP6 := fes.WriteConfigVips(&conf)\n\tif len(fes.vrrps) > 0 {\n\t\tfes.WriteConfigVRRPs(&conf, hasVIP4, hasVIP6)\n\t} else if fes.dropIfNoPeer {\n\t\tfes.WriteConfigDropIfNoPeer(&conf, hasVIP4, hasVIP6)\n\t}\n\tfes.WriteConfigKernel(&conf, hasVIP4, hasVIP6)\n\tfes.WriteConfigBGP(&conf)\n\n\tlogrus.Infof(\"FrontEndService: config generated\")\n\tlogrus.Debugf(\"\\n%v\", conf)\n\t_, err = file.WriteString(conf)\n\tif err != nil {\n\t\tlogrus.Fatalf(\"FrontEndService: failed to write %v, err: %v\", fes.birdConfFile, err)\n\t\tcancel()\n\t}\n}", "func WriteConfig(c Config, filename string) {\n\n\tdata, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\terr = ioutil.WriteFile(filename, data, 664)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func (o KubernetesClusterNodePoolLinuxOsConfigSysctlConfigPtrOutput) NetNetfilterNfConntrackMax() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v *KubernetesClusterNodePoolLinuxOsConfigSysctlConfig) *int {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.NetNetfilterNfConntrackMax\n\t}).(pulumi.IntPtrOutput)\n}", "func (b *Bot) WriteConfig(fn configCallback) {\n\tfn(b.conf)\n}", "func (o KubernetesClusterDefaultNodePoolLinuxOsConfigSysctlConfigPtrOutput) NetNetfilterNfConntrackMax() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v *KubernetesClusterDefaultNodePoolLinuxOsConfigSysctlConfig) *int {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.NetNetfilterNfConntrackMax\n\t}).(pulumi.IntPtrOutput)\n}", "func (o KubernetesClusterDefaultNodePoolLinuxOsConfigSysctlConfigOutput) NetNetfilterNfConntrackBuckets() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v KubernetesClusterDefaultNodePoolLinuxOsConfigSysctlConfig) *int {\n\t\treturn v.NetNetfilterNfConntrackBuckets\n\t}).(pulumi.IntPtrOutput)\n}", "func WriteConfig(cfg NetworkMonitorConfig) string {\n\tif cfgBytes, err := yml.Marshal(cfg); err != nil {\n\t\tpanic(err.Error())\n\t} else {\n\t\tfmt.Println(string(cfgBytes))\n\t\treturn string(cfgBytes)\n\t}\n}", "func (o KubernetesClusterNodePoolLinuxOsConfigSysctlConfigOutput) NetNetfilterNfConntrackBuckets() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v KubernetesClusterNodePoolLinuxOsConfigSysctlConfig) *int {\n\t\treturn v.NetNetfilterNfConntrackBuckets\n\t}).(pulumi.IntPtrOutput)\n}", "func WriteChainConfig(db database.KeyValueWriter, hash common.Hash, cfg *config.ChainConfig) {\n\tif cfg == nil {\n\t\treturn\n\t}\n\tdata, err := json.Marshal(cfg)\n\tif err != nil {\n\t\tlog.Critical(\"Failed to JSON encode chain config\", \"err\", err)\n\t}\n\tif err := db.Put(configKey(hash), data, \"config\"); err != nil {\n\t\tlog.Critical(\"Failed to store chain config\", \"err\", err)\n\t}\n}", "func WriteConfig(f string, c *SupportedBranchesConfig) error {\n\treturn util.WithWriteFile(f, func(w io.Writer) error {\n\t\treturn EncodeConfig(w, c)\n\t})\n}", "func (fes *FrontEndService) WriteConfigVips(conf *string) (hasVIP4, hasVIP6 bool) {\n\tv4, v6 := \"\", \"\"\n\thasVIP4, hasVIP6 = false, false\n\n\tfor _, vip := range fes.vips {\n\t\tif isIPv6(vip) {\n\t\t\t// IPv6\n\t\t\t//v6 += \"\\troute \" + vip + \" blackhole;\\n\"\n\t\t\tv6 += \"\\troute \" + vip + \" via \\\"lo\\\";\\n\"\n\t\t} else if isIPv4(vip) {\n\t\t\t// IPv4\n\t\t\t//v4 += \"\\troute \" + vip + \" blackhole;\\n\"\n\t\t\tv4 += \"\\troute \" + vip + \" via \\\"lo\\\";\\n\"\n\t\t}\n\t}\n\n\tif v4 != \"\" {\n\t\thasVIP4 = true\n\t\t*conf += \"protocol static VIP4 {\\n\"\n\t\t*conf += \"\\tipv4 { preference 110; };\\n\"\n\t\t*conf += v4\n\t\t*conf += \"}\\n\"\n\t\t*conf += \"\\n\"\n\t}\n\n\tif v6 != \"\" {\n\t\thasVIP6 = true\n\t\t*conf += \"protocol static VIP6 {\\n\"\n\t\t*conf += \"\\tipv6 { preference 110; };\\n\"\n\t\t*conf += v6\n\t\t*conf += \"}\\n\"\n\t\t*conf += \"\\n\"\n\t}\n\treturn\n}", "func (o KubernetesClusterDefaultNodePoolLinuxOsConfigSysctlConfigPtrOutput) NetNetfilterNfConntrackBuckets() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v *KubernetesClusterDefaultNodePoolLinuxOsConfigSysctlConfig) *int {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.NetNetfilterNfConntrackBuckets\n\t}).(pulumi.IntPtrOutput)\n}", "func (o KubernetesClusterNodePoolLinuxOsConfigSysctlConfigPtrOutput) NetNetfilterNfConntrackBuckets() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v *KubernetesClusterNodePoolLinuxOsConfigSysctlConfig) *int {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.NetNetfilterNfConntrackBuckets\n\t}).(pulumi.IntPtrOutput)\n}", "func (fes *FrontEndService) WriteConfigBGP(conf *string) {\n\tfor _, gw := range fes.gateways {\n\t\tif isIPv6(gw) || isIPv4(gw) {\n\t\t\tipv := \"\"\n\t\t\tif isIPv4(gw) {\n\t\t\t\tipv += \"\\tipv4 {\\n\"\n\t\t\t\tif len(fes.vrrps) > 0 {\n\t\t\t\t\tipv += \"\\t\\timport none;\\n\"\n\t\t\t\t} else {\n\t\t\t\t\tipv += \"\\t\\timport filter default_v4;\\n\"\n\t\t\t\t}\n\t\t\t\tipv += \"\\t\\texport filter cluster_e_static;\\n\"\n\t\t\t\tipv += \"\\t};\\n\"\n\t\t\t} else if isIPv6(gw) {\n\t\t\t\tipv = \"\\tipv6 {\\n\"\n\t\t\t\tif len(fes.vrrps) > 0 {\n\t\t\t\t\tipv += \"\\t\\timport none;\\n\"\n\t\t\t\t} else {\n\t\t\t\t\tipv += \"\\t\\timport filter default_v6;\\n\"\n\t\t\t\t}\n\t\t\t\tipv += \"\\t\\texport filter cluster_e_static;\\n\"\n\t\t\t\tipv += \"\\t};\\n\"\n\t\t\t}\n\t\t\t*conf += \"protocol bgp from LINK {\\n\"\n\t\t\t*conf += \"\\tinterface \\\"\" + fes.extInterface + \"\\\";\\n\"\n\t\t\t*conf += \"\\tlocal port \" + fes.localPortBGP + \" as \" + fes.localAS + \";\\n\"\n\t\t\t*conf += \"\\tneighbor \" + strings.Split(gw, \"/\")[0] + \" port \" + fes.remotePortBGP + \" as \" + fes.remoteAS + \";\\n\"\n\t\t\t*conf += ipv\n\t\t\t*conf += \"}\\n\"\n\t\t\t*conf += \"\\n\"\n\t\t}\n\t}\n}", "func WriteConfigFile(c *cli.Context, cfg *Config) error {\n\tfPath, err := TenetCfgPath(c)\n\tif err != nil {\n\t\treturn errors.Trace(err)\n\t}\n\tvar buf bytes.Buffer\n\tenc := toml.NewEncoder(&buf)\n\terr = enc.Encode(cfg)\n\tif err != nil {\n\t\treturn errors.Trace(err)\n\t}\n\treturn ioutil.WriteFile(fPath, buf.Bytes(), 0644)\n}", "func (c *ClusterController) addNetworkToConfig(name string, config *NetworkConfig) error {\n\tlabelSelector := metav1.LabelSelector{MatchLabels: map[string]string{\"app\": \"onos\", \"type\": \"config\"}}\n\tpods, err := c.kubeclient.CoreV1().Pods(c.clusterID).List(metav1.ListOptions{\n\t\tLabelSelector: labels.Set(labelSelector.MatchLabels).String(),\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor _, pod := range pods.Items {\n\t\tvar port = 50001\n\t\tfor i := 0; i < config.NumDevices; i++ {\n\t\t\tvar buf bytes.Buffer\n\t\t\tbuf.WriteString(name)\n\t\t\tbuf.WriteString(\"-s\")\n\t\t\tbuf.WriteString(strconv.Itoa(i))\n\t\t\tdeviceName := buf.String()\n\t\t\tif err = c.addNetworkToPod(deviceName, port, pod); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tport = port + 1\n\t\t}\n\t}\n\treturn nil\n}", "func (h *HeaderfileWriter) WriteTemplateConfig(w io.Writer, e datapath.EndpointConfiguration) error {\n\tfw := bufio.NewWriter(w)\n\treturn h.writeTemplateConfig(fw, e)\n}", "func WriteConfig(i interface{}, d string) error {\n\t// Convert to yaml\n\tym, err := yaml.Marshal(i)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Write to file\n\treturn ioutil.WriteFile(d, ym, 0755)\n}", "func (c *Config) WriteConfig(filename string) error {\n\tdata, err := yaml.Marshal(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = ioutil.WriteFile(filename, data, 0644)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func WriteThrapConfig(conf *ThrapConfig, filename string) error {\n\tb, err := hclencoder.Encode(conf)\n\tif err == nil {\n\t\terr = ioutil.WriteFile(filename, b, 0644)\n\t}\n\treturn err\n}", "func WriteCNIConfig() error {\n\tnetConf := &ovncnitypes.NetConf{\n\t\tNetConf: types.NetConf{\n\t\t\tCNIVersion: \"0.4.0\",\n\t\t\tName: \"ovn-kubernetes\",\n\t\t\tType: CNI.Plugin,\n\t\t},\n\t\tLogFile: Logging.CNIFile,\n\t\tLogLevel: fmt.Sprintf(\"%d\", Logging.Level),\n\t\tLogFileMaxSize: Logging.LogFileMaxSize,\n\t\tLogFileMaxBackups: Logging.LogFileMaxBackups,\n\t\tLogFileMaxAge: Logging.LogFileMaxAge,\n\t}\n\n\tnewBytes, err := json.Marshal(netConf)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to marshal CNI config JSON: %v\", err)\n\t}\n\n\tconfFile := filepath.Join(CNI.ConfDir, CNIConfFileName)\n\tif existingBytes, err := os.ReadFile(confFile); err == nil {\n\t\tif bytes.Equal(newBytes, existingBytes) {\n\t\t\t// No changes; do nothing\n\t\t\treturn nil\n\t\t}\n\t}\n\n\t// Install the CNI config file after all initialization is done\n\t// MkdirAll() returns no error if the path already exists\n\tif err := os.MkdirAll(CNI.ConfDir, os.ModeDir); err != nil {\n\t\treturn err\n\t}\n\n\tvar f *os.File\n\tf, err = os.CreateTemp(CNI.ConfDir, \"ovnkube-\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif _, err := f.Write(newBytes); err != nil {\n\t\treturn err\n\t}\n\tif err := f.Close(); err != nil {\n\t\treturn err\n\t}\n\n\treturn os.Rename(f.Name(), confFile)\n}", "func writeConfig() error {\n\tconfigFile := filepath.Join(config.ourBinaryDir, config.ourConfigFilename)\n\tlog.Printf(\"Writing YAML file: %s\", configFile)\n\tyamlText, err := yaml.Marshal(&config)\n\tif err != nil {\n\t\tlog.Printf(\"Couldn't generate YAML file: %s\", err)\n\t\treturn err\n\t}\n\terr = writeFileSafe(configFile, yamlText)\n\tif err != nil {\n\t\tlog.Printf(\"Couldn't save YAML config: %s\", err)\n\t\treturn err\n\t}\n\n\tuserFilter := getUserFilter()\n\terr = userFilter.save()\n\tif err != nil {\n\t\tlog.Printf(\"Couldn't save the user filter: %s\", err)\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (parser *Parser) WriteConfig() error {\n\tconfig := structs.Map(parser.Config)\n\tfor _, site := range parser.Sites {\n\t\tif site.Path == \".\" {\n\t\t\tfor key, val := range site.AllParameters {\n\t\t\t\tif _, ok := config[key]; !ok {\n\t\t\t\t\tconfig[key] = val\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tbs, err := yaml.Marshal(config)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn parser.Storage.Write(parser.Storage.GetSourceKey(KeyConfig), bs)\n}", "func (p *Profile) WriteConfigField(field, value string) error {\n\tviper.Set(p.GetConfigField(field), value)\n\treturn viper.WriteConfig()\n}", "func ConfigWrite() error {\n\n\t// Marshal it\n\tconfigJSON, _ := json.MarshalIndent(Config, \"\", \" \")\n\n\t// Write the file\n\tfd, err := os.OpenFile(configSettingsPath(), os.O_WRONLY|os.O_TRUNC|os.O_CREATE, 0666)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfd.Write(configJSON)\n\tfd.Close()\n\n\t// Done\n\treturn err\n\n}", "func (o KubernetesClusterNodePoolLinuxOsConfigSysctlConfigOutput) NetCoreNetdevMaxBacklog() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v KubernetesClusterNodePoolLinuxOsConfigSysctlConfig) *int { return v.NetCoreNetdevMaxBacklog }).(pulumi.IntPtrOutput)\n}", "func EncodeConfig(configFile io.Writer, c *Configuration) error {\n\t// Lock mutex\n\tConfig.rw.Lock()\n\tdefer Config.rw.Unlock()\n\n\t// Encode the given writer with the given interface\n\treturn toml.NewEncoder(configFile).Encode(c)\n}", "func Write(w io.Writer, cfg *Config) error {\n\tif cfg.filename != \"\" {\n\t\tfmt.Fprintf(w, \"# Configuration file location: %v\\n\\n\", cfg.filename)\n\t} else {\n\t\tfmt.Fprintf(w, \"# Cound not find configuration file location.\\n\\n\")\n\t}\n\treturn toml.NewEncoder(w).Encode(cfg.File)\n}", "func (c *Config) ConfigSaveWriter(w io.Writer) error {\n\tout, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = w.Write(out)\n\treturn err\n}", "func WriteConfig(cfg *config.Dendrite, configDir string) error {\n\tdata, err := yaml.Marshal(cfg)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn ioutil.WriteFile(filepath.Join(configDir, ConfigFile), data, 0666)\n}", "func GetNetConfig(key string) ConfigNet {\n\tvar conf ConfigNet\n\tconf.ServerName = GetString(key + \".\" + \"serverName\")\n\tconf.Addr = GetString(key + \".\" + \"addr\")\n\tconf.TimeOutMs = GetUInt32(key + \".\" + \"timeOutMs\")\n\n\treturn conf\n}", "func SerializeConfig(config multichannelfanout.Config) (map[string]string, error) {\n\tjb, err := json.Marshal(config)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn map[string]string{\n\t\tMultiChannelFanoutConfigKey: string(jb),\n\t}, nil\n}", "func writeConfig(f Function) (err error) {\n\tpath := filepath.Join(f.Root, ConfigFile)\n\tc := toConfig(f)\n\tvar bb []byte\n\tif bb, err = yaml.Marshal(&c); err != nil {\n\t\treturn\n\t}\n\treturn ioutil.WriteFile(path, bb, 0644)\n}", "func SafeWriteConfig(mountData bool, unmountData bool) error {\n\t// configFilePath := path.Join(viper.GetString())\n\tdir := viper.GetString(\"cfg_dir\")\n\tfilepath := path.Join(dir, configName+\".\"+configType)\n\n\tdirExists := fs.DirExists(dir)\n\tif !dirExists {\n\t\terr := fs.CreateDir(dir)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"could not create the config directory for sneak: %s\", err)\n\t\t}\n\t}\n\n\texists := fs.FileExists(filepath)\n\tif exists {\n\t\treturn verify(filepath, mountData, unmountData)\n\t}\n\n\tgui.Info(\"popcorn\", \"creating your config file...\", filepath)\n\n\tif _, err := os.OpenFile(filepath, os.O_RDONLY|os.O_CREATE, 0600); err != nil {\n\t\tgui.Warn(\"could not create configuration file\", filepath)\n\t\treturn err\n\t}\n\n\t// set defaults\n\thtbUsername := gui.InputPromptWithResponse(\"what is your hack the box username?\", \"\", true)\n\tviper.Set(\"htb_username\", htbUsername)\n\tviper.Set(\"openvpn_filepath\", fmt.Sprintf(\"%s/%s.ovpn\", dir, viper.Get(\"htb_username\")))\n\tpreferredEditor := gui.GetUsersPreferredEditor(\"\", true)\n\tviper.Set(\"default_editor\", preferredEditor)\n\tviper.Set(\"data\", dir)\n\tviper.Set(\"webshort\", defaultShortcuts)\n\tfmt.Println(htbNetworkIPHelpText)\n\tlabAccessIP := gui.InputPromptWithResponse(\"what is your HTB Lab Network IPv4?\", \"10.10.15.71\", true)\n\tviper.Set(\"htb_network_ip\", labAccessIP)\n\n\t// write config file\n\tgui.Info(\"popcorn\", \"writing sneak defaults...\", filepath)\n\treturn viper.WriteConfigAs(filepath)\n}", "func (o QueueOutput) OutboundCallerConfig() QueueOutboundCallerConfigPtrOutput {\n\treturn o.ApplyT(func(v *Queue) QueueOutboundCallerConfigPtrOutput { return v.OutboundCallerConfig }).(QueueOutboundCallerConfigPtrOutput)\n}", "func BuildConfig(appName string) error {\n\tif !common.IsDeployed(appName) {\n\t\treturn nil\n\t}\n\n\tif staticWebListener := reportStaticWebListener(appName); staticWebListener != \"\" {\n\t\treturn nil\n\t}\n\n\tappRoot := common.AppRoot(appName)\n\ts, err := common.PlugnTriggerOutput(\"ps-current-scale\", []string{appName}...)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tscale, err := common.ParseScaleOutput(s)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif len(scale) == 0 {\n\t\treturn nil\n\t}\n\n\tif common.GetAppScheduler(appName) != \"docker-local\" {\n\t\treturn nil\n\t}\n\n\tcommon.LogInfo1(fmt.Sprintf(\"Ensuring network configuration is in sync for %s\", appName))\n\n\tfor processType, procCount := range scale {\n\t\tcontainerIndex := 0\n\t\tfor containerIndex < procCount {\n\t\t\tcontainerIndex++\n\t\t\tcontainerIndexString := strconv.Itoa(containerIndex)\n\t\t\tcontainerIDFile := fmt.Sprintf(\"%v/CONTAINER.%v.%v\", appRoot, processType, containerIndex)\n\n\t\t\tcontainerID := common.ReadFirstLine(containerIDFile)\n\t\t\tif containerID == \"\" || !common.ContainerIsRunning(containerID) {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tipAddress := GetContainerIpaddress(appName, processType, containerID)\n\t\t\tif ipAddress != \"\" {\n\t\t\t\targs := []string{appName, processType, containerIndexString, ipAddress}\n\t\t\t\t_, err := common.PlugnTriggerOutput(\"network-write-ipaddr\", args...)\n\t\t\t\tif err != nil {\n\t\t\t\t\tcommon.LogWarn(err.Error())\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o KubernetesClusterDefaultNodePoolLinuxOsConfigSysctlConfigOutput) NetCoreNetdevMaxBacklog() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v KubernetesClusterDefaultNodePoolLinuxOsConfigSysctlConfig) *int {\n\t\treturn v.NetCoreNetdevMaxBacklog\n\t}).(pulumi.IntPtrOutput)\n}", "func writeNetconnBlockedMessage(message *convertedCbMessage, kv map[string]interface{}) {\n\tkv[\"event_type\"] = \"blocked_netconn\"\n\t// TODO: need ingress event type for netconn blocks\n\n\tblocked := message.OriginalMessage.NetconnBlocked\n\n\tkv[\"domain\"] = GetUnicodeFromUTF8(blocked.GetUtf8Netpath())\n\tkv[\"ipv4\"] = GetIPv4Address(blocked.GetIpv4Address())\n\tkv[\"port\"] = Ntohs(uint16(blocked.GetPort()))\n\tkv[\"protocol\"] = int32(blocked.GetProtocol())\n\n\tif blocked.GetOutbound() {\n\t\tkv[\"direction\"] = \"outbound\"\n\t} else {\n\t\tkv[\"direction\"] = \"inbound\"\n\t}\n\tif blocked.RemoteIpAddress != nil {\n\t\tkv[\"remote_ip\"] = GetIPv4Address(blocked.GetRemoteIpAddress())\n\t\tkv[\"remote_port\"] = Ntohs(uint16(blocked.GetRemotePort()))\n\t}\n\n\tif blocked.LocalIpAddress != nil {\n\t\tkv[\"local_ip\"] = GetIPv4Address(blocked.GetLocalIpAddress())\n\t\tkv[\"local_port\"] = Ntohs(uint16(blocked.GetLocalPort()))\n\t}\n}", "func NewNetworkConfigSerde(configType string, modelPtr interface{}) serde.Serde {\n\tmodelFactory, err := createModelFactory(modelPtr)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn &binarySerde{\n\t\tdomain: NetworkConfigSerdeDomain,\n\t\tserdeType: configType,\n\t\tmodelFactory: modelFactory,\n\t}\n}", "func (c *Config) Write() error {\n\trawConfig, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = os.MkdirAll(configDirPath, 0755)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn ioutil.WriteFile(configPath, rawConfig, 0644)\n}", "func WriteConnectionConfig(context *clusterd.Context, clusterInfo *cephclient.ClusterInfo) error {\n\t// write the latest config to the config dir\n\tif _, err := cephclient.GenerateConnectionConfig(context, clusterInfo); err != nil {\n\t\treturn errors.Wrap(err, \"failed to write connection config\")\n\t}\n\n\treturn nil\n}", "func (c *Config) Write(filename string) (err error) {\n\tb, err := json.Marshal(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfile, err := os.Create(filename)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif _, err := file.Write(b); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func WriteCredsConfig(cc *CredsConfig, fpath string) error {\n\tb, err := hclencoder.Encode(cc)\n\tif err == nil {\n\t\terr = ioutil.WriteFile(fpath, b, 0644)\n\t}\n\treturn err\n}", "func EncodeConfig(w io.Writer, c *SupportedBranchesConfig) error {\n\tc.Sort()\n\tb, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = w.Write(b)\n\treturn err\n}", "func AppendConfigToDisk(specObj *model.SpecObject, filename string) error {\n\t// Marshal spec object to yaml\n\tdata, err := yaml.Marshal(specObj)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Check if file exists. We need to ammend the file if it does.\n\tif fileExists(filename) {\n\t\tf, err := os.OpenFile(filename, os.O_APPEND|os.O_WRONLY, 0600)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tdefer func() {\n\t\t\t_ = f.Close()\n\t\t}()\n\n\t\t_, err = f.Write(append([]byte(\"---\\n\"), data...))\n\t\treturn err\n\t}\n\n\t// Create a new file with out specs\n\treturn ioutil.WriteFile(filename, data, 0755)\n}", "func (o WorkerPoolOutput) NetworkConfig() NetworkConfigResponseOutput {\n\treturn o.ApplyT(func(v *WorkerPool) NetworkConfigResponseOutput { return v.NetworkConfig }).(NetworkConfigResponseOutput)\n}", "func WriteCephConfig(context *clusterd.Context, clusterInfo *ClusterInfo) error {\n\t// create the ceph.conf with the default settings\n\tcephConfig, err := CreateDefaultCephConfig(context, clusterInfo)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to create default ceph config\")\n\t}\n\n\t// write the latest config to the config dir\n\tconfFilePath, err := GenerateConnectionConfigWithSettings(context, clusterInfo, cephConfig)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to write connection config\")\n\t}\n\tsrc, err := os.ReadFile(filepath.Clean(confFilePath))\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to copy connection config to /etc/ceph. failed to read the connection config\")\n\t}\n\terr = os.WriteFile(DefaultConfigFilePath(), src, 0600)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"failed to copy connection config to /etc/ceph. failed to write %q\", DefaultConfigFilePath())\n\t}\n\tdst, err := os.ReadFile(DefaultConfigFilePath())\n\tif err == nil {\n\t\tlogger.Debugf(\"config file @ %s:\\n%s\", DefaultConfigFilePath(), dst)\n\t} else {\n\t\tlogger.Warningf(\"wrote and copied config file but failed to read it back from %s for logging. %v\", DefaultConfigFilePath(), err)\n\t}\n\treturn nil\n}", "func (o KubernetesClusterDefaultNodePoolLinuxOsConfigSysctlConfigPtrOutput) NetCoreNetdevMaxBacklog() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v *KubernetesClusterDefaultNodePoolLinuxOsConfigSysctlConfig) *int {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.NetCoreNetdevMaxBacklog\n\t}).(pulumi.IntPtrOutput)\n}", "func (fes *FrontEndService) WriteConfigVRRPs(conf *string, hasVIP4, hasVIP6 bool) {\n\tfor _, ip := range fes.vrrps {\n\t\tif isIPv6(ip) || isIPv4(ip) {\n\t\t\t*conf += \"protocol static {\\n\"\n\t\t\tif isIPv4(ip) {\n\t\t\t\t*conf += \"\\tipv4;\\n\"\n\t\t\t\t*conf += \"\\troute 0.0.0.0/0 via \" + strings.Split(ip, \"/\")[0] + \"%'\" + fes.extInterface + \"' onlink;\\n\"\n\t\t\t} else if isIPv6(ip) {\n\t\t\t\t*conf += \"\\tipv6;\\n\"\n\t\t\t\t*conf += \"\\troute 0::/0 via \" + strings.Split(ip, \"/\")[0] + \"%'\" + fes.extInterface + \"' onlink;\\n\"\n\t\t\t}\n\t\t\t*conf += \"}\\n\"\n\t\t\t*conf += \"\\n\"\n\t\t}\n\t}\n}", "func (o KubernetesClusterNodePoolLinuxOsConfigSysctlConfigPtrOutput) NetCoreNetdevMaxBacklog() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v *KubernetesClusterNodePoolLinuxOsConfigSysctlConfig) *int {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.NetCoreNetdevMaxBacklog\n\t}).(pulumi.IntPtrOutput)\n}", "func (c *Client) PutNetconf(url string, payloadBody bytes.Buffer) error {\n\t_, err := c.httpRequest(url, \"PUT\", payloadBody)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func NewNetSocketCollector(cfgBaseName, procFSPath string) (collector.Collector, error) {\n\tprocFile := filepath.Join(\"net\", \"dev\")\n\n\tc := NetSocket{\n\t\tcommon: newCommon(NameNetSocket, procFSPath, procFile, tags.FromList(tags.GetBaseTags())),\n\t}\n\n\tc.include = defaultIncludeRegex\n\tc.exclude = regexp.MustCompile(fmt.Sprintf(regexPat, `lo`))\n\n\tif cfgBaseName == \"\" {\n\t\tif _, err := os.Stat(c.file); os.IsNotExist(err) {\n\t\t\treturn nil, fmt.Errorf(\"%s procfile: %w\", c.pkgID, err)\n\t\t}\n\t\treturn &c, nil\n\t}\n\n\tvar opts netSocketOptions\n\terr := config.LoadConfigFile(cfgBaseName, &opts)\n\tif err != nil {\n\t\tif !strings.Contains(err.Error(), \"no config found matching\") {\n\t\t\tc.logger.Warn().Err(err).Str(\"file\", cfgBaseName).Msg(\"loading config file\")\n\t\t\treturn nil, fmt.Errorf(\"%s config: %w\", c.pkgID, err)\n\t\t}\n\t} else {\n\t\tc.logger.Debug().Str(\"base\", cfgBaseName).Interface(\"config\", opts).Msg(\"loaded config\")\n\t}\n\n\tif opts.IncludeRegex != \"\" {\n\t\trx, err := regexp.Compile(fmt.Sprintf(regexPat, opts.IncludeRegex))\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"%s compile include rx: %w\", c.pkgID, err)\n\t\t}\n\t\tc.include = rx\n\t}\n\n\tif opts.ExcludeRegex != \"\" {\n\t\trx, err := regexp.Compile(fmt.Sprintf(regexPat, opts.ExcludeRegex))\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"%s compile exclude rx: %w\", c.pkgID, err)\n\t\t}\n\t\tc.exclude = rx\n\t}\n\n\tif opts.ID != \"\" {\n\t\tc.id = opts.ID\n\t}\n\n\tif opts.ProcFSPath != \"\" {\n\t\tc.procFSPath = opts.ProcFSPath\n\t\tc.file = filepath.Join(c.procFSPath, procFile)\n\t}\n\n\tif opts.RunTTL != \"\" {\n\t\tdur, err := time.ParseDuration(opts.RunTTL)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"%s parsing run_ttl: %w\", c.pkgID, err)\n\t\t}\n\t\tc.runTTL = dur\n\t}\n\n\tif _, err := os.Stat(c.file); os.IsNotExist(err) {\n\t\treturn nil, fmt.Errorf(\"%s procfile: %w\", c.pkgID, err)\n\t}\n\n\treturn &c, nil\n}", "func (c *ClusterController) removeNetworkFromConfig(name string, configMap *corev1.ConfigMapList) error {\n\tlabelSelector := metav1.LabelSelector{MatchLabels: map[string]string{\"app\": \"onos\", \"type\": \"config\"}}\n\tpods, err := c.kubeclient.CoreV1().Pods(c.clusterID).List(metav1.ListOptions{\n\t\tLabelSelector: labels.Set(labelSelector.MatchLabels).String(),\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\tdataMap := configMap.Items[0].BinaryData[\"config\"]\n\tm := make(map[string]interface{})\n\terr = yaml.Unmarshal(dataMap, &m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tnumDevices := m[\"numdevices\"].(int)\n\n\tfor _, pod := range pods.Items {\n\t\tfor i := 0; i < numDevices; i++ {\n\t\t\tvar buf bytes.Buffer\n\t\t\tbuf.WriteString(name)\n\t\t\tbuf.WriteString(\"-s\")\n\t\t\tbuf.WriteString(strconv.Itoa(i))\n\t\t\tdeviceName := buf.String()\n\t\t\tif err = c.removeNetworkFromPod(deviceName, pod); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func (g *Generator) AddConfigExposedPort(port string) {\n\tg.image.Config.ExposedPorts[port] = struct{}{}\n}", "func (ec *ExecutionContext) WriteConfig(config *Config) error {\n\tvar op errors.Op = \"cli.ExecutionContext.WriteConfig\"\n\tvar cfg *Config\n\tif config != nil {\n\t\tcfg = config\n\t} else {\n\t\tcfg = ec.Config\n\t}\n\tbuf := new(bytes.Buffer)\n\tencoder := yaml.NewEncoder(buf)\n\tencoder.SetIndent(2)\n\terr := encoder.Encode(cfg)\n\tif err != nil {\n\t\treturn errors.E(op, err)\n\t}\n\terr = ioutil.WriteFile(ec.ConfigFile, buf.Bytes(), 0644)\n\tif err != nil {\n\t\treturn errors.E(op, err)\n\t}\n\treturn nil\n}", "func (in *NetDeviceConfig) DeepCopy() *NetDeviceConfig {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(NetDeviceConfig)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func writeKubeconfig(cfg *rest.Config) {\n\ttemplate := `# Kubeconfig file for Calico CNI plugin. Installed by calico/node.\napiVersion: v1\nkind: Config\nclusters:\n- name: local\n cluster:\n server: %s\n certificate-authority-data: \"%s\"\nusers:\n- name: calico\n user:\n token: %s\ncontexts:\n- name: calico-context\n context:\n cluster: local\n user: calico\ncurrent-context: calico-context`\n\n\t// Replace the placeholders.\n\tdata := fmt.Sprintf(template, cfg.Host, string(base64.StdEncoding.EncodeToString(cfg.CAData)), cfg.BearerToken)\n\n\t// Write the filled out config to disk.\n\tif err := ioutil.WriteFile(kubeconfigPath, []byte(data), 0600); err != nil {\n\t\tlogrus.WithError(err).Error(\"Failed to write CNI plugin kubeconfig file\")\n\t\treturn\n\t}\n\tlogrus.WithField(\"path\", kubeconfigPath).Info(\"Wrote updated CNI kubeconfig file.\")\n}", "func (r *Resource) WriteConfig(filename string) error {\n\tr.Lock()\n\tdefer r.Unlock()\n\n\tvar b bytes.Buffer\n\n\tb.WriteString(fmt.Sprintf(\"# meta-data-json:{\\\"updated\\\": \\\"%s\\\"}\\n\", time.Now().UTC()))\n\tb.WriteString(fmt.Sprintf(\"resource %s {\\n\", r.name))\n\n\tvar hosts []string\n\tfor _, h := range r.host {\n\t\thosts = append(hosts, h.Name)\n\n\t\tb.WriteString(indentf(1, \"on %s {\\n\", h.Name))\n\t\tb.WriteString(indentf(2, \"node-id %d;\\n\", h.ID))\n\t\tb.WriteString(indentf(2, \"address %s:%d;\\n\", h.IP, r.port))\n\t\tfor _, v := range h.volume {\n\t\t\tb.WriteString(indentf(2, \"volume %d {\\n\", v.id))\n\t\t\tb.WriteString(indentf(3, \"device minor %d;\\n\", v.minor))\n\t\t\tb.WriteString(indentf(3, \"disk %s;\\n\", v.backingDevice))\n\t\t\tb.WriteString(indentf(3, \"meta-disk internal;\\n\"))\n\t\t\tb.WriteString(indentf(2, \"}\\n\")) // end volume section\n\t\t}\n\t\tb.WriteString(indentf(1, \"}\\n\")) // end on section\n\t\tb.WriteString(\"\\n\")\n\t}\n\n\tb.WriteString(indentf(1, \"connection-mesh {\\n\"))\n\tb.WriteString(indentf(2, \"hosts %s;\\n\", strings.Join(hosts, \" \")))\n\tb.WriteString(indentf(1, \"}\\n\"))\n\n\tb.WriteString(\"}\") // end resource section\n\n\treturn ioutil.WriteFile(filename, b.Bytes(), 0644)\n}", "func (runner *Runner) writeRuntimeConfig(cfg *ini.File) {\n\n\tvar err error\n\n\trunner.runtimeCfgFile, err = ioutil.TempFile(\"\", \"frpok\")\n\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\t// save to tmp file\n\terr = cfg.SaveTo(runner.runtimeCfgFile.Name())\n\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n}", "func (m *VpnConfiguration) Serialize(writer i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.SerializationWriter)(error) {\n err := m.DeviceConfiguration.Serialize(writer)\n if err != nil {\n return err\n }\n if m.GetAuthenticationMethod() != nil {\n cast := (*m.GetAuthenticationMethod()).String()\n err = writer.WriteStringValue(\"authenticationMethod\", &cast)\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"connectionName\", m.GetConnectionName())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"realm\", m.GetRealm())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"role\", m.GetRole())\n if err != nil {\n return err\n }\n }\n if m.GetServers() != nil {\n cast := make([]i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, len(m.GetServers()))\n for i, v := range m.GetServers() {\n if v != nil {\n cast[i] = v.(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable)\n }\n }\n err = writer.WriteCollectionOfObjectValues(\"servers\", cast)\n if err != nil {\n return err\n }\n }\n return nil\n}", "func (p *ProviderConfigs) writeProviderConfig() error {\n\t// Get file.\n\tconfig := getProviderConfigFile()\n\n\t// Convert object to json.\n\tdata, err := json.MarshalIndent(p, \"\", \" \")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\t// Write json output to file.\n\terr = ioutil.WriteFile(config, data, 0644)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn nil\n}", "func renderOVNFlowsConfig(bootstrapResult *bootstrap.BootstrapResult, data *render.RenderData) {\n\tflows := bootstrapResult.OVN.FlowsConfig\n\tif flows == nil {\n\t\treturn\n\t}\n\tif flows.Target == \"\" {\n\t\tklog.Warningf(\"ovs-flows-config configmap 'target' field can't be empty. Ignoring configuration: %+v\", flows)\n\t\treturn\n\t}\n\t// if IPFIX collectors are provided by means of both the operator configuration and the\n\t// ovs-flows-config ConfigMap, we will merge both targets\n\tif colls, ok := data.Data[\"IPFIXCollectors\"].(string); !ok || colls == \"\" {\n\t\tdata.Data[\"IPFIXCollectors\"] = flows.Target\n\t} else {\n\t\tdata.Data[\"IPFIXCollectors\"] = colls + \",\" + flows.Target\n\t}\n\tif flows.CacheMaxFlows != nil {\n\t\tdata.Data[\"IPFIXCacheMaxFlows\"] = *flows.CacheMaxFlows\n\t}\n\tif flows.Sampling != nil {\n\t\tdata.Data[\"IPFIXSampling\"] = *flows.Sampling\n\t}\n\tif flows.CacheActiveTimeout != nil {\n\t\tdata.Data[\"IPFIXCacheActiveTimeout\"] = *flows.CacheActiveTimeout\n\t}\n}", "func dumpConfig(ctx *cli.Context) error {\n\tcfg := makeAllConfigs(ctx)\n\tcomment := \"\"\n\n\tout, err := tomlSettings.Marshal(&cfg)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tlog.Info(\"Dumping legacy Lachesis config file\")\n\n\tdump := os.Stdout\n\tif ctx.NArg() > 0 {\n\t\tdump, err = os.OpenFile(ctx.Args().Get(0), os.O_RDWR|os.O_CREATE|os.O_TRUNC, 0644)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer dump.Close()\n\t}\n\tdump.WriteString(comment)\n\tdump.Write(out)\n\n\tlog.Info(\"Dumping Opera config file\")\n\n\toperaDatadir := path.Join(cfg.Node.DataDir, \"opera\")\n\toperaArgs := excludeArg(os.Args, \"--\"+validatorFlag.Name, true)\n\toperaArgs = excludeArg(operaArgs, \"--\"+configFileFlag.Name, true)\n\toperaArgs = excludeArg(operaArgs, \"--\"+DataDirFlag.Name, true)\n\toperaArgs = addFrontArgs(operaArgs, []string{\"--\" + DataDirFlag.Name, operaDatadir})\n\toperaArgs = excludeArg(operaArgs, \"--\"+utils.LegacyTestnetFlag.Name, false)\n\n\tmetrics.Enabled = false\n\treturn launcher.Launch(operaArgs)\n}", "func WriteServerConfig(filePath string, cfg *ServerConfig) error {\n\tbts, err := toml.Marshal(*cfg)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn files.WriteAndBackup(filePath, bts)\n}", "func GenerateNetlinxCompileCfg(w http.ResponseWriter, r *http.Request) {\n\t// Get Body as Bytes Array\n\tbody, err := ioutil.ReadAll(r.Body)\n\tif err != nil {\n\t\tw.WriteHeader(500)\n\t\treturn\n\t}\n\n\t// Load this into an APW Workspace\n\ta, err := apw.NewAPW(\"myWorkspace.apw\", body)\n\n\t// Get URL Variables\n\troot := r.URL.Query().Get(\"root\")\n\tlogfile := r.URL.Query().Get(\"logfile\")\n\tlogconsole, _ := strconv.ParseBool(r.URL.Query().Get(\"logconsole\"))\n\t// Process and generate the .cfg\n\tb := compilecfg.Generate(*a, root, logfile, logconsole)\n\n\tw.Write(b)\n}", "func (s *Syncthing) SaveConfig(dev *model.Dev) error {\n\tmarshalled, err := yaml.Marshal(s)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tsyncthingInfoFile := getInfoFile(dev.Namespace, dev.Name)\n\tif err := os.WriteFile(syncthingInfoFile, marshalled, 0600); err != nil {\n\t\treturn fmt.Errorf(\"failed to write syncthing info file: %w\", err)\n\t}\n\n\treturn nil\n}", "func GenerateIptablesConfig(path string, privnet *model.Network) error {\n\tvar (\n\t\tf *os.File\n\t\terr error\n\t)\n\tt := template.Must(template.New(\"iptablesConfig\").Parse(iptablesTmpl))\n\tif f, err = os.OpenFile(path, os.O_RDWR|os.O_CREATE|os.O_TRUNC, 0644); err != nil {\n\t\treturn fmt.Errorf(\"error opening output file at \\\"%s\\\": %w\", path, err)\n\t}\n\tdefer f.Close()\n\tif err = t.Execute(f, struct {\n\t\tPrivateNetwork string\n\t}{privnet.IPv4Addresses[0].Net.String()}); err != nil {\n\t\treturn fmt.Errorf(\"error executing template: %w\", err)\n\t}\n\treturn nil\n}", "func (c *Config) WriteScopeConfig(path string) error {\n\tscb, err := c.ScopeConfigYaml()\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = ioutil.WriteFile(path, scb, 0644)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error writing ScopeConfig to file %s: %v\", path, err)\n\t}\n\treturn nil\n}", "func (o *CreateRouteRequest) SetNetPeeringId(v string) {\n\to.NetPeeringId = &v\n}", "func (o ServiceResponseOutput) NetworkConfig() NetworkConfigResponseOutput {\n\treturn o.ApplyT(func(v ServiceResponse) NetworkConfigResponse { return v.NetworkConfig }).(NetworkConfigResponseOutput)\n}", "func (g *Generator) ConfigExposedPorts() map[string]struct{} {\n\t// We have to make a copy to preserve the privacy of g.image.Config.\n\tcopy := map[string]struct{}{}\n\tfor k, v := range g.image.Config.ExposedPorts {\n\t\tcopy[k] = v\n\t}\n\treturn copy\n}", "func ExportEmptyNetworkConfig(w io.Writer) error {\n\t_, err := fmt.Fprintln(w, \"version: 2\\nethernets: {}\")\n\treturn err\n}", "func (o HttpFilterConfigOutput) ConfigTypeUrl() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v HttpFilterConfig) *string { return v.ConfigTypeUrl }).(pulumi.StringPtrOutput)\n}", "func writeConfig(o options, clientset kubernetes.Interface) error {\n\t// kubeconfig is a kubernetes config.\n\tvar kubeconfig []byte\n\n\tdir, file := filepath.Split(o.output)\n\n\terr := os.MkdirAll(dir, os.ModePerm)\n\tif err != nil {\n\t\treturn &util.ExitError{Message: fmt.Sprintf(\"unable to create output directory %v: %v.\", dir, err), Code: 1}\n\t}\n\n\tif o.certificate {\n\t\tif kubeconfig, err = certificate.CreateKubeConfigWithCertificateCredentials(clientset, o.name); err != nil {\n\t\t\treturn &util.ExitError{Message: fmt.Sprintf(\"unable to create kubeconfig file with cert and key for %v: %v.\", o.name, err), Code: 1}\n\t\t}\n\t} else {\n\t\t// Service account credentials are the default if unspecified.\n\t\tif kubeconfig, err = serviceaccount.CreateKubeConfigWithServiceAccountCredentials(clientset, o.name); err != nil {\n\t\t\treturn &util.ExitError{Message: fmt.Sprintf(\"unable to create kubeconfig file with service account for %v: %v.\", o.name, err), Code: 1}\n\t\t}\n\t}\n\n\tif !o.overwrite && util.FileExists(o.output) {\n\t\tif kubeconfig, err = mergeConfigs(o, kubeconfig); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif err = ioutil.WriteFile(o.output, kubeconfig, 0644); err != nil {\n\t\treturn &util.ExitError{Message: fmt.Sprintf(\"unable to write to file %v: %v.\", file, err), Code: 1}\n\t}\n\n\treturn nil\n}", "func (m *BgpConfiguration) Serialize(writer i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.SerializationWriter)(error) {\n {\n err := writer.WriteInt32Value(\"asn\", m.GetAsn())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"ipAddress\", m.GetIpAddress())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"localIpAddress\", m.GetLocalIpAddress())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"@odata.type\", m.GetOdataType())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"peerIpAddress\", m.GetPeerIpAddress())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteAdditionalData(m.GetAdditionalData())\n if err != nil {\n return err\n }\n }\n return nil\n}", "func (c *Config) WritePayload(w io.Writer, payload interface{}) error {\n\tjsonConf, err := json.Marshal(payload)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to marshal payload: %s\", err)\n\t}\n\n\tc.config.jsonConfSize = C.uint(len(jsonConf))\n\tc.config.nsPathSize = C.uint(len(c.nsPath))\n\tcconfPayload := C.GoBytes(unsafe.Pointer(c.config), C.sizeof_struct_cConfig)\n\tcconfPayload = append(cconfPayload, c.nsPath...)\n\tcconfPayload = append(cconfPayload, jsonConf...)\n\n\tif n, err := w.Write(cconfPayload); err != nil || n != len(cconfPayload) {\n\t\treturn fmt.Errorf(\"failed to write payload: %s\", err)\n\t}\n\treturn nil\n}", "func (f *FileConfigWriter) Write(config *auth.Config) error {\n\tif f.filename == \"\" {\n\t\treturn errors.New(\"No config file name defined\")\n\t}\n\tcontent, err := yaml.Marshal(config)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"marshaling the config to yaml\")\n\t}\n\terr = ioutil.WriteFile(f.filename, content, util.DefaultWritePermissions)\n\treturn nil\n}", "func (m *AospDeviceOwnerDeviceConfiguration) SetWifiBlockEditConfigurations(value *bool)() {\n err := m.GetBackingStore().Set(\"wifiBlockEditConfigurations\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *LogicAppTriggerEndpointConfiguration) Serialize(writer i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.SerializationWriter)(error) {\n err := m.CustomExtensionEndpointConfiguration.Serialize(writer)\n if err != nil {\n return err\n }\n {\n err = writer.WriteStringValue(\"logicAppWorkflowName\", m.GetLogicAppWorkflowName())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"resourceGroupName\", m.GetResourceGroupName())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"subscriptionId\", m.GetSubscriptionId())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"url\", m.GetUrl())\n if err != nil {\n return err\n }\n }\n return nil\n}", "func WriteConfigDeprecated(configPath vfs.Path, config interface{}, writeOptions ...vfs.WriteOption) error {\n\tdata, err := utils.YamlMarshal(config)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error marshalling configuration: %v\", err)\n\t}\n\n\tcreate := false\n\tfor _, writeOption := range writeOptions {\n\t\tswitch writeOption {\n\t\tcase vfs.WriteOptionCreate:\n\t\t\tcreate = true\n\t\tcase vfs.WriteOptionOnlyIfExists:\n\t\t\t_, err = configPath.ReadFile()\n\t\t\tif err != nil {\n\t\t\t\tif os.IsNotExist(err) {\n\t\t\t\t\treturn fmt.Errorf(\"cannot update configuration file %s: does not exist\", configPath)\n\t\t\t\t}\n\t\t\t\treturn fmt.Errorf(\"error checking if configuration file %s exists already: %v\", configPath, err)\n\t\t\t}\n\t\tdefault:\n\t\t\treturn fmt.Errorf(\"unknown write option: %q\", writeOption)\n\t\t}\n\t}\n\n\tif create {\n\t\terr = configPath.CreateFile(data)\n\t} else {\n\t\terr = configPath.WriteFile(data)\n\t}\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error writing configuration file %s: %v\", configPath, err)\n\t}\n\treturn nil\n}", "func foldersConfigWrite(file string, folders []folder.FolderConfig) error {\n\tffMutex.Lock()\n\tdefer ffMutex.Unlock()\n\n\tfd, err := os.OpenFile(file, os.O_CREATE|os.O_TRUNC|os.O_WRONLY, 0666)\n\tdefer fd.Close()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdata := &xmlFolders{\n\t\tVersion: \"1\",\n\t\tFolders: folders,\n\t}\n\n\tenc := xml.NewEncoder(fd)\n\tenc.Indent(\"\", \" \")\n\treturn enc.Encode(data)\n}", "func WriteConfig() {\n\tf, err := os.Create(\"config.json\")\n\tif err != nil {\n\t\tlog.Fatalf(\"Can't write configuration file: %v\", err)\n\t}\n\te := json.NewEncoder(f)\n\te.SetIndent(\"\", \" \")\n\te.Encode(defaultConfig)\n\tf.Close()\n\tos.Exit(0)\n}", "func (client *XenClient) PBDSetDeviceConfig(self string, value map[string]string) (err error) {\n\t_, err = client.APICall(\"PBD.set_device_config\", self, value)\n\tif err != nil {\n\t\treturn\n\t}\n\t// no return result\n\treturn\n}", "func NewNetworkEntityConfigSerde(configType string, modelPtr interface{}) serde.Serde {\n\tmodelFactory, err := createModelFactory(modelPtr)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn &binarySerde{\n\t\tdomain: NetworkEntitySerdeDomain,\n\t\tserdeType: configType,\n\t\tmodelFactory: modelFactory,\n\t}\n}", "func (c DQLConfig) Write() error {\n\tf := filepath.Join(helpers.GetProjectPath(c.ProjectPath), \"dql.conf.json\")\n\n\tjson, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn ioutil.WriteFile(f, json, 0644)\n}", "func (c *ServiceConfig) GetNetNetwork() string {\n\tif c != nil && c.NetNetwork != \"\" {\n\t\treturn c.NetNetwork\n\t}\n\treturn \"tcp\"\n}", "func writeScorecardConfig(dir string, cfg v1alpha3.Configuration) error {\n\t// Skip writing if config is empty.\n\tif cfg.Metadata.Name == \"\" {\n\t\treturn nil\n\t}\n\n\tb, err := yaml.Marshal(cfg)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tcfgDir := filepath.Join(dir, filepath.FromSlash(scorecard.DefaultConfigDir))\n\tif err := os.MkdirAll(cfgDir, 0755); err != nil {\n\t\treturn err\n\t}\n\tscorecardConfigPath := filepath.Join(cfgDir, scorecard.ConfigFileName)\n\treturn os.WriteFile(scorecardConfigPath, b, 0666)\n}", "func dumpConfig(ctx *cli.Context) error {\n\tcfg := makeAllConfigs(ctx)\n\tcomment := \"\"\n\n\tout, err := toml.Settings.Marshal(&cfg)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdump := os.Stdout\n\tif ctx.NArg() > 0 {\n\t\tdump, err = os.OpenFile(ctx.Args().Get(0), os.O_RDWR|os.O_CREATE|os.O_TRUNC, 0644)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer dump.Close()\n\t}\n\tdump.WriteString(comment)\n\tdump.Write(out)\n\n\treturn nil\n}", "func (*EdgeDevConfig) Descriptor() ([]byte, []int) {\n\treturn file_config_devconfig_proto_rawDescGZIP(), []int{1}\n}", "func WebsocketNetDialContext(dialContext func(ctx context.Context, network string, addr string) (net.Conn, error)) optSetter {\n\treturn func(f *Forwarder) error {\n\t\tf.websocketDialer.NetDialContext = dialContext\n\t\treturn nil\n\t}\n}", "func (opt Options) PrintConfig(w io.Writer) error {\n\tt1, _ := template.New(\"webhook\").Funcs(sprig.TxtFuncMap()).Parse(webhook)\n\terr := t1.Execute(w, opt)\n\tif err != nil {\n\t\treturn err\n\t}\n\tt2, _ := template.New(\"csr\").Funcs(sprig.TxtFuncMap()).Parse(csr)\n\terr = t2.Execute(w, opt)\n\tif err != nil {\n\t\treturn err\n\t}\n\tt3, _ := template.New(\"deployment\").Funcs(sprig.TxtFuncMap()).Parse(webhookDeployment)\n\terr = t3.Execute(w, opt)\n\treturn err\n}", "func writeConfig(args *common.Parameters) {\n\n\t//Create the config file and open it for writing.\n\tconfigWrite, err := os.Create(\"./data/cluster/config.csv\")\n\tif err != nil {\n\t\targs.ErrorLogger.Println(\"entity=\" + entityKind + \" message=\" + err.Error())\n\t\tfmt.Println(\"[ERROR] entity=\" + entityKind + \" message=\" + err.Error())\n\t\treturn\n\t}\n\n\t//Write out the header.\n\tfmt.Fprintln(configWrite, \"AuditTime,Name\")\n\tfmt.Fprintf(configWrite, \"%s,%s\\n\", common.Format(args.CurrentTime), *args.ClusterName)\n\tconfigWrite.Close()\n}", "func SetBpfDatalink(fd, t int) (int, error) {\n\terr := ioctlPtr(fd, BIOCSDLT, unsafe.Pointer(&t))\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn t, nil\n}" ]
[ "0.56793135", "0.5659518", "0.5651362", "0.5474038", "0.5373798", "0.534663", "0.52460337", "0.5223805", "0.52133596", "0.5210043", "0.5181914", "0.51706094", "0.51527107", "0.5151398", "0.5138834", "0.5118219", "0.5109886", "0.50933087", "0.5050451", "0.5049241", "0.50281286", "0.49867433", "0.49394113", "0.4932857", "0.49321076", "0.4864468", "0.48024076", "0.47854388", "0.4759463", "0.47515175", "0.4748615", "0.46364936", "0.46359292", "0.4592744", "0.45858085", "0.4584669", "0.45798188", "0.45778075", "0.4560068", "0.4550801", "0.45503756", "0.45483285", "0.4538859", "0.4529097", "0.45203206", "0.4517039", "0.45137328", "0.4509893", "0.4500501", "0.44983342", "0.44970438", "0.44941103", "0.44825512", "0.44690347", "0.44689232", "0.4466727", "0.44628116", "0.44410682", "0.4435314", "0.44251004", "0.44191936", "0.44134426", "0.44129542", "0.44008958", "0.43981054", "0.4396826", "0.4395622", "0.43867895", "0.4375431", "0.43447715", "0.4313932", "0.43067026", "0.42910436", "0.42905095", "0.4290182", "0.4271889", "0.42632493", "0.42630404", "0.4254137", "0.42526665", "0.4249646", "0.4248219", "0.42466345", "0.42412743", "0.42355418", "0.42298597", "0.42139992", "0.42133835", "0.42050177", "0.4201294", "0.4193603", "0.41865557", "0.4185826", "0.41822594", "0.41817066", "0.41804048", "0.41796434", "0.41761354", "0.41747367", "0.4174364" ]
0.73389
0
writeStaticData writes the endpointspecific static data defines to the specified writer. This must be kept in sync with loader.ELFSubstitutions().
func (h *HeaderfileWriter) writeStaticData(fw io.Writer, e datapath.EndpointConfiguration) { if e.IsHost() { if option.Config.EnableNodePort { // Values defined here are for the host datapath attached to the // host device and therefore won't be used. We however need to set // non-zero values to prevent the compiler from optimizing them // out, because we need to substitute them for host datapaths // attached to native devices. // When substituting symbols in the object file, we will replace // these values with zero for the host device and with the actual // values for the native devices. fmt.Fprint(fw, "/* Fake values, replaced by 0 for host device and by actual values for native devices. */\n") fmt.Fprint(fw, defineUint32("NATIVE_DEV_IFINDEX", 1)) fmt.Fprint(fw, "\n") } if option.Config.EnableBPFMasquerade { if option.Config.EnableIPv4Masquerade { // NodePort comment above applies to IPV4_MASQUERADE too placeholderIPv4 := []byte{1, 1, 1, 1} fmt.Fprint(fw, defineIPv4("IPV4_MASQUERADE", placeholderIPv4)) } if option.Config.EnableIPv6Masquerade { // NodePort comment above applies to IPV6_MASQUERADE too placeholderIPv6 := []byte{1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1} fmt.Fprint(fw, defineIPv6("IPV6_MASQUERADE", placeholderIPv6)) } } // Dummy value to avoid being optimized when 0 fmt.Fprint(fw, defineUint32("SECCTX_FROM_IPCACHE", 1)) // Use templating for ETH_HLEN only if there is any L2-less device if !mac.HaveMACAddrs(option.Config.GetDevices()) { // L2 hdr len (for L2-less devices it will be replaced with "0") fmt.Fprint(fw, defineUint16("ETH_HLEN", mac.EthHdrLen)) } } else { // We want to ensure that the template BPF program always has "LXC_IP" // defined and present as a symbol in the resulting object file after // compilation, regardless of whether IPv6 is disabled. Because the type // templateCfg hardcodes a dummy IPv6 address (and adheres to the // datapath.EndpointConfiguration interface), we can rely on it always // having an IPv6 addr. Endpoints however may not have IPv6 addrs if IPv6 // is disabled. Hence this check prevents us from omitting the "LXC_IP" // symbol from the template BPF program. Without this, the following // scenario is possible: // 1) Enable IPv6 in cilium // 2) Create an endpoint (ensure endpoint has an IPv6 addr) // 3) Disable IPv6 and restart cilium // This results in a template BPF object without an "LXC_IP" defined, // __but__ the endpoint still has "LXC_IP" defined. This causes a later // call to loader.ELFSubstitutions() to fail on missing a symbol "LXC_IP". if ipv6 := e.IPv6Address(); ipv6.IsValid() { fmt.Fprint(fw, defineIPv6("LXC_IP", ipv6.AsSlice())) } fmt.Fprint(fw, defineIPv4("LXC_IPV4", e.IPv4Address().AsSlice())) fmt.Fprint(fw, defineUint16("LXC_ID", uint16(e.GetID()))) } fmt.Fprint(fw, defineMAC("NODE_MAC", e.GetNodeMAC())) secID := e.GetIdentityLocked().Uint32() fmt.Fprint(fw, defineUint32("SECLABEL", secID)) fmt.Fprint(fw, defineUint32("SECLABEL_IPV4", secID)) fmt.Fprint(fw, defineUint32("SECLABEL_IPV6", secID)) fmt.Fprint(fw, defineUint32("SECLABEL_NB", byteorder.HostToNetwork32(secID))) fmt.Fprint(fw, defineUint32("POLICY_VERDICT_LOG_FILTER", e.GetPolicyVerdictLogFilter())) epID := uint16(e.GetID()) fmt.Fprintf(fw, "#define POLICY_MAP %s\n", bpf.LocalMapName(policymap.MapName, epID)) callsMapName := callsmap.MapName if e.IsHost() { callsMapName = callsmap.HostMapName } fmt.Fprintf(fw, "#define CALLS_MAP %s\n", bpf.LocalMapName(callsMapName, epID)) if option.Config.EnableCustomCalls && !e.IsHost() { fmt.Fprintf(fw, "#define CUSTOM_CALLS_MAP %s\n", bpf.LocalMapName(callsmap.CustomCallsMapName, epID)) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *Site) writeStatic() error {\n\n\tfor _, file := range s.files {\n\t\tfrom := filepath.Join(s.Src, file)\n\t\tto := filepath.Join(s.Dest, file)\n\t\tlogf(MsgCopyingFile, file)\n\t\tif err := copyTo(from, to); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (sddww SeriesDataDatasourceWriterWrapper) WriteData(w io.Writer, sf SeriesFooter) (n int, err error) {\n defer func() {\n if state := recover(); state != nil {\n err = log.Wrap(state.(error))\n }\n }()\n\n count, err := io.Copy(w, sddww.r)\n log.PanicIf(err)\n\n return int(count), nil\n}", "func handleStatic(writer http.ResponseWriter, request *http.Request) {\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\twriter.WriteHeader(500)\n\t\t\t_, err := writer.Write([]byte(fmt.Sprint(r)))\n\t\t\tif err != nil {\n\t\t\t\tfmt.Println(\"Error writing response: \", err)\n\t\t\t}\n\t\t}\n\t}()\n\n\t// First, check if the path matches any in handlerMap, otherwise serve from static.\n\tpath := request.URL.Path\n\tfor k, v := range handlerMap {\n\t\tif strings.Index(path, k) == 0 {\n\t\t\tv(writer, request)\n\t\t\treturn\n\t\t}\n\t}\n\n\tfmt.Printf(\"%s %s\\n\", request.Method, request.URL.String())\n\n\t// Only supports get requests\n\tif strings.ToUpper(request.Method) != \"GET\" {\n\t\tpanic(\"Unsupported HTTP Method\")\n\t}\n\n\twd, err := os.Getwd()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tfilePath := filepath.Join(wd, \"client\", path)\n\n\t// If the file does not exist, return 404\n\tif _, err := os.Stat(filePath); err != nil {\n\t\tif os.IsNotExist(err) {\n\t\t\twriter.WriteHeader(404)\n\t\t\treturn\n\t\t} else {\n\t\t\tpanic(err)\n\t\t}\n\t}\n\n\t// Otherwise, return the file\n\tb, err := ioutil.ReadFile(filePath)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\t// Get the mime type from the filename\n\tvar mimeType string\n\text := filepath.Ext(filePath)\n\tif mt, ok := mimeMap[ext]; ok {\n\t\tmimeType = mt\n\t} else {\n\t\tmimeType = \"text/plain\"\n\t}\n\n\twriter.Header().Set(\"Content-Type\", mimeType)\n\n\twriter.WriteHeader(200)\n\n\t_, err = writer.Write(b)\n\tif err != nil {\n\t\tfmt.Println(\"ERROR: Failed to write to response: \", err)\n\t\treturn\n\t}\n\n\treturn\n}", "func (ws *Server) AddStaticDirectory(endpoint string, path string) {\n\tws.staticFiles[endpoint] = path\n}", "func Static(w http.ResponseWriter, r *http.Request, params httprouter.Params) {\n var mime_type string\n\n static_path := \"/src/github.com/mrgirlyman/watcher/frontman/build\" + params.ByName(\"filepath\")\n p, _ := loadPage(static_path)\n\n // Determining the MIME type of the file based on file extension\n switch slice := strings.Split(static_path, \".\"); slice[len(slice) - 1] {\n case \"css\":\n mime_type = \"text/css\"\n case \"js\":\n mime_type = \"application/javascript\"\n default:\n mime_type = \"text/plain\"\n }\n \n w.Header().Set(\"Content-Type\", mime_type)\n\n // Writing content to the http.ResponseWriter\n fmt.Fprintf(w, string(p.Body))\n}", "func WriteStaticPng(slm *sunlightmap, pathfileext string) (err error) {\n\n\tif pathfileext == \"\" {\n\t\tpathfileext = \"./slmbg.png\"\n\t}\n\n\tslm.visualization = \"static\"\n\tdaycolor := PngGetColorAt(slm.DaylightImageFilename, slm.Width/2, slm.Height-10)\n\tnightcolor := PngGetColorAt(slm.NighttimeImageFilename, 10, slm.Height/2)\n\tfmt.Println(daycolor, nightcolor)\n\tsslm := newSingle(slm, slm.zeitpunkte[0])\n\t//img, err := mergeImages2(slm, &sslm)\n\timg, err := mergeImages2(slm, &sslm)\n\tif err != nil {\n\t\tpanic(err) //fixme\n\t}\n\n\toverheight := 0\n\tdst := image.NewRGBA(image.Rect(0, 0, slm.Width, slm.Height+overheight))\n\tgreen := color.RGBA{0, 100, 0, 255}\n\tdraw.Draw(dst, dst.Bounds(), &image.Uniform{green}, image.ZP, draw.Src)\n\tdraw.Draw(dst, image.Rect(0, 0, slm.Width, overheight/2+2), &image.Uniform{daycolor}, image.ZP, draw.Src)\n\tdraw.Draw(dst, image.Rect(0, slm.Height-2, slm.Width, slm.Height+overheight), &image.Uniform{nightcolor}, image.ZP, draw.Src)\n\t//draw.Draw(dst, dst.Bounds(), &image.Uniform{color.RGBA{30, 30, 30, 30}}, image.ZP, draw.Src)\n\tfmt.Println(slm.CenterLongitude)\n\n\t//with the assumption that draw.Draw is optimized\n\t//so no scrolling/wrapping/... implemented but left-or-right\n\tcenterAt := 999\n\tif centerAt == 999 {\n\t\tcenterAt = 44\n\t} else if centerAt > 180 {\n\t\tcenterAt = 0\n\t}\n\tif centerAt < -180 {\n\t\tcenterAt = 0\n\t}\n\tx := (slm.Width / 2) + (centerAt * slm.Width / 2 / 180) //return int\n\toffset := 0\n\tif x > slm.Width/2 {\n\t\tfmt.Println(\"easting\")\n\t\toffset = x - slm.Width/2\n\t\tslidedFill := image.Rect(slm.Width-offset, overheight/2, slm.Width, slm.Height+overheight)\n\t\tslidedCenter := image.Rect(0, overheight/2, slm.Width-offset, slm.Height+overheight)\n\n\t\t// Draw aligns slided*.Min in dst with Pt in src and then replaces the\n\t\t// rectangle slided* in dst with the result of drawing src on dst.\n\t\tdraw.Draw(dst, slidedFill, img, image.Pt(0, 0), draw.Src)\n\t\tdraw.Draw(dst, slidedCenter, img, image.Pt(offset, 0), draw.Src)\n\t} else if x < slm.Width/2 {\n\t\tfmt.Println(\"westing\")\n\t\toffset = slm.Width/2 - x\n\t\tslidedFill := image.Rect(0, overheight/2, offset, slm.Height+overheight)\n\t\tslidedCenter := image.Rect(offset, overheight/2, slm.Width, slm.Height+overheight)\n\t\tdraw.Draw(dst, slidedFill, img, image.Pt(slm.Width-offset, 0), draw.Src)\n\t\tdraw.Draw(dst, slidedCenter, img, image.Pt(0, 0), draw.Src)\n\t} else {\n\t\tdraw.Draw(dst, dst.Bounds(), img, image.Pt(0, 0), draw.Src)\n\t}\n\tfmt.Println(\"x:\", x, \"offset:\", offset)\n\n\tf, err := os.OpenFile(pathfileext, os.O_WRONLY|os.O_CREATE, 0600)\n\tdefer f.Close()\n\tpng.Encode(f, dst)\n\n\treturn\n\n}", "func static(w http.ResponseWriter, r *http.Request) {\n\thttp.ServeFile(w, r, r.URL.Path[1:])\n\treturn\n}", "func (_DevUtils *DevUtilsCaller) EncodeStaticCallAssetData(opts *bind.CallOpts, staticCallTargetAddress common.Address, staticCallData []byte, expectedReturnDataHash [32]byte) ([]byte, error) {\n\tvar (\n\t\tret0 = new([]byte)\n\t)\n\tout := ret0\n\terr := _DevUtils.contract.Call(opts, out, \"encodeStaticCallAssetData\", staticCallTargetAddress, staticCallData, expectedReturnDataHash)\n\treturn *ret0, err\n}", "func Write(dir, source string, data HashSet) error {\n\ttarget := filepath.Join(dir, PatchFile)\n\tjoinBuffer, size := bytes.Buffer{}, len(data)\n\t// append source if it exists\n\tif source != \"\" {\n\t\tjoinBuffer.WriteString(SourceMarker + source + LineSeperator)\n\t\tlog.Notice(\"generated patch with source\", source)\n\t}\n\t// store all datasets in the file\n\tfor i := 0; i < size; i++ {\n\t\tjoinBuffer.WriteString(data[i].String())\n\t\tjoinBuffer.WriteString(LineSeperator)\n\t}\n\toutputFile, err := os.Create(target)\n\tif err != nil {\n\t\tlog.Critical(\"failed to open patch file:\", err)\n\t\treturn err\n\t}\n\tdefer outputFile.Close()\n\t// copy byte buffer into file\n\tjoinBuffer.WriteTo(outputFile)\n\treturn nil\n}", "func (ged GobSingleObjectEncoderDatasource) WriteData(w io.Writer, sf SeriesFooter) (n int, err error) {\n defer func() {\n if state := recover(); state != nil {\n err = log.Wrap(state.(error))\n }\n }()\n\n wc := rifs.NewWriteCounter(w)\n e := gob.NewEncoder(wc)\n\n err = e.Encode(ged.inputValue)\n log.PanicIf(err)\n\n return wc.Count(), nil\n}", "func (d *Dispatcher) Static(w http.ResponseWriter, r *http.Request) {\n\tfor prefix, staticDir := range d.Router.StaticRoutes {\n\t\tif strings.HasPrefix(r.URL.Path, prefix) {\n\t\t\tfile := staticDir + r.URL.Path[len(prefix):]\n\t\t\thttp.ServeFile(w, r, file)\n\t\t\td.found = true\n\t\t}\n\t}\n}", "func Static(next http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t// Serves any requested file from dist/. These are immutable & cached.\n\t\t// Served with the dist prefix, e.g. GET /dist/js/foo-HASH.js.\n\t\tif file := r.URL.Path[1:]; dist[file] {\n\t\t\tw.Header().Set(\"Cache-Control\", \"max-age=31536000, public, immutable\")\n\t\t\tw.Header().Set(\"Vary\", \"Accept-Encoding\")\n\n\t\t\t// Identify with original filename, not pre-compressed filename.\n\t\t\tif ctype := mime.TypeByExtension(filepath.Ext(file)); ctype != \"\" {\n\t\t\t\tw.Header().Set(\"Content-Type\", ctype)\n\t\t\t}\n\n\t\t\t// TODO Content negotiation github.com/golang/go/issues/19307.\n\t\t\taccept := r.Header.Get(\"Accept-Encoding\")\n\t\t\tif strings.Contains(accept, \"br\") && dist[file+\".br\"] {\n\t\t\t\tw.Header().Set(\"Content-Encoding\", \"br\")\n\t\t\t\tfile += \".br\"\n\t\t\t} else if strings.Contains(accept, \"gzip\") && dist[file+\".gz\"] {\n\t\t\t\tw.Header().Set(\"Content-Encoding\", \"gzip\")\n\t\t\t\tfile += \".gz\"\n\t\t\t}\n\n\t\t\thttp.ServeFile(w, r, file)\n\t\t\treturn\n\t\t}\n\n\t\t// Serves any requested file from public/. These aren't cached.\n\t\t// Served without a prefix, e.g. GET /robots.txt.\n\t\tif file := path.Join(\"public\", r.URL.Path); public[file] {\n\t\t\thttp.ServeFile(w, r, file)\n\t\t\treturn\n\t\t}\n\n\t\tnext.ServeHTTP(w, r)\n\t})\n}", "func (a *Api) serveStatic(res http.ResponseWriter, req *http.Request) {\n\tdata, err := ioutil.ReadFile(\"index.html\")\n\tif err != nil {\n\t\tlog.Printf(\"templates - failure to read index.html\")\n\t}\n\tres.Header().Set(\"content-type\", \"text/html\")\n\tres.WriteHeader(200)\n\tres.Write(data)\n\treturn\n}", "func Write(w io.Writer, dt DataType, data []float64) error {\n\tbuf := &bytes.Buffer{}\n\tvar err error\n\tswitch dt {\n\tcase DSA:\n\t\terr = writeDSA(buf, Float64sToInt16s(data))\n\tcase DFA:\n\t\terr = writeDFA(buf, Float64sToFloat32s(data))\n\tcase DDA:\n\t\terr = writeDDA(buf, data)\n\tcase DSB:\n\t\terr = writeDSB(buf, Float64sToInt16s(data))\n\tcase DFB:\n\t\terr = writeDFB(buf, Float64sToFloat32s(data))\n\tcase DDB:\n\t\terr = writeDDB(buf, data)\n\tdefault:\n\t\terr = ErrUnknownDataType\n\t}\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = io.Copy(w, buf)\n\treturn err\n}", "func Static(fs embed.FS) echo.MiddlewareFunc {\n\treturn func(next echo.HandlerFunc) echo.HandlerFunc {\n\t\treturn func(c echo.Context) (err error) {\n\t\t\turi := c.Request().RequestURI\n\n\t\t\tif strings.HasPrefix(uri, \"/api\") {\n\t\t\t\treturn next(c)\n\t\t\t}\n\n\t\t\tif uri == \"/\" {\n\t\t\t\turi = \"index.html\"\n\t\t\t}\n\n\t\t\turi = filepath.Join(\"assets\", uri)\n\n\t\t\tdata, err := fs.ReadFile(uri)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Fprintf(os.Stderr, \"ERROR: reading file from embeded FS: %s\\n\", err.Error())\n\t\t\t\treturn next(c)\n\t\t\t}\n\n\t\t\tmime := mimetype.Detect(data).String()\n\t\t\tif strings.HasSuffix(uri, \"wasm\") {\n\t\t\t\tmime = \"application/wasm\"\n\t\t\t}\n\t\t\tif strings.HasSuffix(uri, \"html\") {\n\t\t\t\tmime = \"text/html; charset=utf-8\"\n\t\t\t}\n\t\t\tif strings.HasSuffix(uri, \"js\") {\n\t\t\t\tmime = \"application/js; charset=utf-8\"\n\t\t\t}\n\t\t\tif strings.HasSuffix(uri, \"png\") {\n\t\t\t\tmime = \"image/png\"\n\t\t\t}\n\n\t\t\treturn c.Blob(http.StatusOK, mime, data)\n\t\t}\n\t}\n}", "func staticHandler(c *RequestContext, w http.ResponseWriter, r *http.Request) (int, error) {\n\tif r.URL.Path != \"/static/manifest.json\" {\n\t\thttp.FileServer(c.FM.assets.HTTPBox()).ServeHTTP(w, r)\n\t\treturn 0, nil\n\t}\n\n\treturn renderFile(\n\t\tw,\n\t\tc.FM.assets.MustString(\"static/manifest.json\"),\n\t\t\"application/json\",\n\t\tc,\n\t)\n}", "func (m *SiteCollection) Serialize(writer i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.SerializationWriter)(error) {\n {\n err := writer.WriteStringValue(\"dataLocationCode\", m.GetDataLocationCode())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"hostname\", m.GetHostname())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"@odata.type\", m.GetOdataType())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteObjectValue(\"root\", m.GetRoot())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteAdditionalData(m.GetAdditionalData())\n if err != nil {\n return err\n }\n }\n return nil\n}", "func Static(opts *Options) macaron.Handler {\n\treturn macaron.Static(\n\t\topts.Directory,\n\t\tmacaron.StaticOptions{\n\t\t\tSkipLogging: opts.SkipLogging,\n\t\t\tFileSystem: bindata.Static(bindata.Options{\n\t\t\t\tAsset: Asset,\n\t\t\t\tAssetDir: AssetDir,\n\t\t\t\tAssetInfo: AssetInfo,\n\t\t\t\tAssetNames: AssetNames,\n\t\t\t\tPrefix: \"\",\n\t\t\t}),\n\t\t},\n\t)\n}", "func (t *staticTemplateHandler) SetData(key string, value interface{}) templateHandler {\n\tt.data[key] = value\n\treturn t\n}", "func Static(w http.ResponseWriter, r *http.Request) {\n http.ServeFile(w, r, r.URL.Path[1:])\n}", "func newStaticTemplateHandler(filename string, data map[string]interface{}) *staticTemplateHandler {\n\tif data == nil {\n\t\tdata = map[string]interface{}{}\n\t}\n\n\tt := &staticTemplateHandler{\n\t\tfilename: filename,\n\t\tdata: data,\n\t}\n\n\tt.once.Do(func() {\n\t\tfuncs := template.FuncMap{\n\t\t\t\"include\": func(part string) (template.HTML, error) {\n\t\t\t\tbuf := bytes.NewBuffer(nil)\n\t\t\t\tpt := template.Must(template.ParseFiles(filepath.Join(partialDir, part)))\n\t\t\t\terr := pt.Execute(buf, t.data)\n\t\t\t\treturn template.HTML(buf.String()), err\n\t\t\t},\n\t\t}\n\t\tt.templ = template.Must(template.New(t.filename).Funcs(funcs).\n\t\t\tParseFiles(filepath.Join(templateDir, t.filename)))\n\t})\n\treturn t\n}", "func (_DevUtils *DevUtilsSession) EncodeStaticCallAssetData(staticCallTargetAddress common.Address, staticCallData []byte, expectedReturnDataHash [32]byte) ([]byte, error) {\n\treturn _DevUtils.Contract.EncodeStaticCallAssetData(&_DevUtils.CallOpts, staticCallTargetAddress, staticCallData, expectedReturnDataHash)\n}", "func (_DevUtils *DevUtilsCallerSession) EncodeStaticCallAssetData(staticCallTargetAddress common.Address, staticCallData []byte, expectedReturnDataHash [32]byte) ([]byte, error) {\n\treturn _DevUtils.Contract.EncodeStaticCallAssetData(&_DevUtils.CallOpts, staticCallTargetAddress, staticCallData, expectedReturnDataHash)\n}", "func ReturnStaticPngBase64(slm *sunlightmap) (retval string, err error) {\n\n\tbuff := bytes.Buffer{}\n\tslm.visualization = \"static\"\n\t//for _, value := range slm.zeitpunkte {\n\n\tsslm := newSingle(slm, slm.zeitpunkte[0])\n\timg, err := mergeImages2(slm, &sslm)\n\tif err != nil {\n\t\tpanic(err) //fixme\n\t}\n\tpng.Encode(&buff, img)\n\tretval = base64.StdEncoding.EncodeToString(buff.Bytes())\n\treturn\n\n}", "func RegisterStaticHandlers(cfg *Config, db *gorm.DB, logger *logrus.Logger) {\n\thttp.HandleFunc(\"/\", handleDynamicContent(cfg, db, logger))\n}", "func (sh ServerHandler) DefineStaticRoute(r *mux.Router) {\n\tr.Methods(\"GET\").PathPrefix(\"/static/\").HandlerFunc(sh.serveStatic)\n}", "func staticHandler(w http.ResponseWriter, r *http.Request) {\n\tfs := http.FileServer(http.Dir(\"webapp/build\"))\n\tfs.ServeHTTP(w, r)\n}", "func staticFile(name string) string {\n\treturn filepath.Join(Config.StaticPath, name)\n}", "func setupStatic(mux *goji.Mux) {\n\tif len(opts.DevServerUri) > 0 {\n\t\tlog.Printf(\"Proxying static files to development server %v.\",\n\t\t\topts.DevServerUri)\n\t\tdevServerProxyUrl, err := url.Parse(opts.DevServerUri)\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t\tdevServerProxy :=\n\t\t\thttputil.NewSingleHostReverseProxy(devServerProxyUrl)\n\t\tmux.Handle(pat.Get(\"/*\"), devServerProxy)\n\t} else {\n\t\tpublic := http.FileServer(\n\t\t\trice.MustFindBox(\"./public\").HTTPBox())\n\t\tmux.Handle(pat.Get(\"/*\"), public)\n\t}\n}", "func buildStaticFile(rel string) func(src, dst string, _ winter.Config) error {\n\treturn func(src, dst string, _ winter.Config) error {\n\t\trel, err := filepath.Rel(rel, src)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"can't get `%s` relative to `%s`: %w\", src, rel, err)\n\t\t}\n\n\t\tif err := os.MkdirAll(filepath.Dir(filepath.Join(dst, rel)), 0755); err != nil {\n\t\t\treturn fmt.Errorf(\"can't make static asset dir `%s`: %w\", dst, err)\n\t\t}\n\n\t\ts, err := os.Open(src)\n\t\tif err != nil {\n\t\t\tlog.Printf(\"can't read static file `%s`: %s\", src, err)\n\t\t\t// TODO: This happens sometimes\n\t\t\treturn nil\n\t\t}\n\t\tdefer s.Close()\n\n\t\td, err := os.Create(filepath.Join(dst, rel))\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\n\t\t\t\t\"can't write static file `%s` to `%s`: %w\",\n\t\t\t\tsrc,\n\t\t\t\tfilepath.Join(dst, rel),\n\t\t\t\terr,\n\t\t\t)\n\t\t}\n\t\tdefer d.Close()\n\n\t\tif _, err := io.Copy(d, s); err != nil {\n\t\t\treturn fmt.Errorf(\"can't build static asset %s: %w\", src, err)\n\t\t}\n\n\t\treturn nil\n\t}\n}", "func StaticContentHandler(data []byte, contentType string) http.Handler {\n\tif len(data) == 0 {\n\t\treturn errorHandler(http.StatusNoContent)\n\t}\n\tmodtime := time.Now()\n\n\tmodtimeStr := modtime.UTC().Format(TimeFormat)\n\treturn http.HandlerFunc(func(res http.ResponseWriter, req *http.Request) {\n\t\tif t, err := time.Parse(TimeFormat, req.Header.Get(ifModifiedSince)); err == nil && modtime.Before(t.Add(StaticCacheDuration)) {\n\t\t\tres.Header().Del(contentTypeHeader)\n\t\t\tres.Header().Del(contentLength)\n\t\t\tres.WriteHeader(http.StatusNotModified)\n\t\t\treturn\n\t\t}\n\t\tsetContentType(res, contentType, contentBinary)\n\t\tres.Header().Set(lastModified, modtimeStr)\n\t\tres.Write(data)\n\t})\n}", "func Write(fn string, d map[string]interface{}, f DataFmt) error {\n\tb, err := Marshal(d, f)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tw, err := os.Create(fn)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif _, err = w.Write(b); err != nil {\n\t\treturn err\n\t}\n\n\tw.Close()\n\treturn nil\n}", "func registerStatic(mux *http.ServeMux, staticDir string) {\n\tif mux == nil {\n\t\tmux = http.DefaultServeMux\n\t}\n\tmux.Handle(\"/\", hostRedirectHandler(wrappedHandler(http.FileServer(http.Dir(staticDir)))))\n}", "func (h *HeaderfileWriter) WriteEndpointConfig(w io.Writer, e datapath.EndpointConfiguration) error {\n\tfw := bufio.NewWriter(w)\n\n\twriteIncludes(w)\n\th.writeStaticData(fw, e)\n\n\treturn h.writeTemplateConfig(fw, e)\n}", "func (group *RouterGroup) Static(p, root string) {\n\tp = path.Join(p, \"/*filepath\")\n\tfileServer := http.FileServer(http.Dir(root))\n\n\tgroup.GET(p, func(c *Context) {\n\t\toriginal := c.Request.URL.Path\n\t\tc.Request.URL.Path = c.Params.ByName(\"filepath\")\n\n\t\twriter := c.Writer\n\n\t\theaders := writer.Header()\n\n\t\t// in production environment static content needs to be cached by browsers & proxies\n\t\tif Env == Prod {\n\t\t\t// cache for 3 months\n\t\t\theaders.Set(\"Expires\", time.Now().AddDate(0, 3, 0).Format(http.TimeFormat))\n\n\t\t\theaders.Add(\"Cache-Control\", \"public\")\n\t\t\theaders.Add(\"Cache-Control\", \"max-age=28771200\")\n\t\t}\n\n\t\tfileServer.ServeHTTP(writer, c.Request)\n\t\tc.Request.URL.Path = original\n\t})\n}", "func RegisterStaticHandler(r *mux.Router, logger *zap.Logger, qOpts *QueryOptions) {\n\tstaticHandler, err := NewStaticAssetsHandler(qOpts.StaticAssets, StaticAssetsHandlerOptions{\n\t\tBasePath: qOpts.BasePath,\n\t\tUIConfigPath: qOpts.UIConfig,\n\t\tLogger: logger,\n\t\tLogAccess: qOpts.LogStaticAssetsAccess,\n\t})\n\tif err != nil {\n\t\tlogger.Panic(\"Could not create static assets handler\", zap.Error(err))\n\t}\n\n\tstaticHandler.RegisterRoutes(r)\n}", "func (s *StaticDataElement) ServeHTTP(w http.ResponseWriter, r *http.Request) {\n\tswitch r.Method {\n\tcase \"GET\":\n\t\tfallthrough\n\tcase \"HEAD\":\n\t\tw.Header().Set(\"Last-Modified\", s.Time().UTC().Format(http.TimeFormat))\n\t\tif ims := r.Header.Get(\"If-Modified-Since\"); ims != \"\" {\n\t\t\tif t, e := time.Parse(http.TimeFormat, ims); e == nil && !t.Before(s.Time()) {\n\t\t\t\tw.WriteHeader(http.StatusNotModified)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\tif r.Method == \"GET\" {\n\t\t\tw.Header().Set(\"Content-type\", s.Mime)\n\t\t\t_, _ = w.Write(s.Data)\n\t\t}\n\t\treturn\n\tcase \"POST\":\n\t\tw.Header().Set(\"Allow\", \"Get\")\n\t\tw.Header().Add(\"Allow\", \"Head\")\n\t\tw.WriteHeader(http.StatusMethodNotAllowed)\n\tdefault:\n\t\tw.WriteHeader(http.StatusNotImplemented)\n\t\treturn\n\t}\n\treturn\n}", "func WriteHandler(ctx *routing.Context) error {\n\n\t//Get the Path values\n\tindexPath := strings.Replace(string(ctx.Path()), \"/write/\", \"\", -1)\n\tif indexPath == \"\" {\n\t\tindexPath = \"default\"\n\t}\n\ttmp := strings.SplitAfter(indexPath, \"/\")\n\tindexHead := strings.TrimRight(tmp[0], \"/\")\n\tindexPath = strings.TrimRight(indexPath, \"/\")\n\n\t//Ensure pathing is proper\n\tisValidPathChar := regexp.MustCompile(`^[A-Za-z0-9\\/\\.\\-\\_]+$`).MatchString\n\tfor _, pathchar := range []string{indexPath} {\n\t\tif !isValidPathChar(pathchar) {\n\t\t\tctx.Error(\"invalid path detected\", 400)\n\t\t\treturn nil\n\t\t}\n\t}\n\n\t//Get the Header, validate type, push to Ingestion Enging\n\trawHeaders := string(ctx.Request.Header.Peek(\"Content-Type\"))\n\tif rawHeaders == \"application/x-ndjson\" || rawHeaders == \"application/json\" {\n\t\t//Send the data onward to the Ingestion Engine for indexing\n\t\tdatabase.IngestionEngine(database.LogPacket{\n\t\t\tTimeAtIndex: time.Now().Unix(),\n\t\t\tIndexHead: indexHead,\n\t\t\tIndexPath: indexPath,\n\t\t\tDataBlob: ctx.PostBody(),\n\t\t\tDataType: rawHeaders})\n\t} else {\n\t\tctx.Error(\"invalid headers detected\", 415)\n\t}\n\n\treturn nil\n}", "func NewStaticHandler(cnf Config) *StaticHandler {\n\tfs := intStaticFS\n\tif cnf.Dir != \"\" {\n\t\tfs = http.Dir(path.Join(cnf.Dir, \"static\"))\n\t}\n\treturn &StaticHandler{fs: fs}\n}", "func staticHandler(w http.ResponseWriter, r *http.Request) {\n\tvar path string\n\tif strings.HasSuffix(r.URL.Path, \"/\") {\n\t\tpath = r.URL.Path + \"index.html\"\n\t} else {\n\t\tpath = r.URL.Path\n\t}\n\thttp.ServeFile(w, r, \"static/\"+path)\n}", "func NewStaticProvider(data interface{}) Provider {\n\tb, err := yaml.Marshal(data)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn staticProvider{NewYAMLProviderFromBytes(b)}\n}", "func StaticSite(app, env, svc string, rg bucketResourceGetter, emptier bucketEmptier) *StaticSiteCleaner {\n\treturn &StaticSiteCleaner{\n\t\tapp: app,\n\t\tenv: env,\n\t\tsvc: svc,\n\t\tbucketResourceGetter: rg,\n\t\tbucketEmptier: emptier,\n\t}\n}", "func Static(name, contentType string) gin.HandlerFunc {\n\tdata := Assets.Files[name].Data\n\treturn func(c *gin.Context) {\n\t\tc.Data(http.StatusOK, contentType, data)\n\t}\n}", "func AddStatic(mux Mux, path, content string) {\n\tAddStaticBinary(mux, path, []byte(content))\n}", "func Static(opts *Options) macaron.Handler {\n\topts.FileSystem = Assets\n\t// we don't need to pass the directory, because the directory var is only\n\t// used when in the options there is no FileSystem.\n\treturn opts.staticHandler(\"\")\n}", "func NewStatic() filters.Spec { return &static{} }", "func NewStatic() filters.Spec { return &static{} }", "func (b *Baa) Static(prefix string, dir string, index bool, h HandlerFunc) {\n\tif prefix == \"\" {\n\t\tpanic(\"baa.Static prefix can not be empty\")\n\t}\n\tif dir == \"\" {\n\t\tpanic(\"baa.Static dir can not be empty\")\n\t}\n\tb.Get(prefix+\"*\", newStatic(prefix, dir, index, h))\n}", "func (l *Loader) SetAttrCgoExportStatic(i Sym, v bool) {\n\tif v {\n\t\tl.attrCgoExportStatic[i] = struct{}{}\n\t} else {\n\t\tdelete(l.attrCgoExportStatic, i)\n\t}\n}", "func WriteResource(w http.ResponseWriter, status int, resource *Resource, links *DocumentLinks, included ...*Resource) error {\n\treturn WriteResponse(w, status, &Document{\n\t\tData: &HybridResource{\n\t\t\tOne: resource,\n\t\t},\n\t\tLinks: links,\n\t\tIncluded: included,\n\t})\n}", "func (s *sitesController) UpdateStaticSites() {\n\tfor _, config := range s.configs {\n\t\tlog.Debug(\"sites.Controller.UpdateStaticSites - Creating Site:\" + config.Domain)\n\t\tsiteCreator := NewSiteCreator(config)\n\t\tsiteCreator.addSite()\n\t\tsiteCreator.addSources()\n\t\tsiteCreator.addContainers()\n\t\tsiteCreator.addLocations()\n\t\tsiteCreator.addContexts()\n\t\tsiteCreator.fillFileContainers(config)\n\t\tsiteCreator.writeFiles()\n\t}\n}", "func WriteRaw(out io.Writer, d Data) error {\n\treturn write(out, d)\n}", "func (group *RouterGroup) createStaticHandler(relativePath string, fs http.FileSystem) HandlerFunc {\n\tabsolutePath := path.Join(group.prefix, relativePath) // client path\n\tfileServer := http.StripPrefix(absolutePath, http.FileServer(fs)) // server path\n\n\treturn func(ctx *Context) {\n\t\t// Check whether the file exists\n\t\tfile := ctx.Param(\"filepath\") // custom parameter name\n\t\tif _, err := fs.Open(file); err != nil {\n\t\t\tctx.Fail(http.StatusNotFound, \"Status Not Found: \"+ctx.Pattern)\n\t\t\treturn\n\t\t}\n\n\t\t// Serve file content\n\t\tfileServer.ServeHTTP(ctx.RespWriter, ctx.Req)\n\t}\n}", "func Static(w http.ResponseWriter, r *http.Request) {\n\tpath, _ := os.Getwd()\n\tpath = filepath.Join(path, \"web/static\")\n\n\thttp.StripPrefix(\n\t\t\"/static/\",\n\t\thttp.FileServer(http.Dir(path)),\n\t).ServeHTTP(w, r)\n}", "func setupStaticFilesMiddleware(next http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t// Returns static files under /static\n\t\tif strings.Index(r.URL.Path, \"/assets/\") == 0 {\n\t\t\tw.Header().Set(\"Cache-Control\", \"public, max-age=7776000\")\n\t\t\tfs := http.FileServer(http.Dir(charthelper.DataDirBase()))\n\t\t\tfs = http.StripPrefix(\"/assets/\", fs)\n\t\t\tfs.ServeHTTP(w, r)\n\t\t} else {\n\t\t\t// Fallbacks to chained hander\n\t\t\tnext.ServeHTTP(w, r)\n\t\t}\n\t})\n}", "func (io *JSONWSIO) WriteData(d interface{}) error {\n\treturn io.WSConn.WriteJSON(d)\n}", "func setStaticIP(ifaceName string) error {\n\tip := getFullIP(ifaceName)\n\tif len(ip) == 0 {\n\t\treturn errors.New(\"Can't get IP address\")\n\t}\n\n\tbody, err := ioutil.ReadFile(\"/etc/dhcpcd.conf\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tip4, _, err := net.ParseCIDR(ip)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tadd := fmt.Sprintf(\"\\ninterface %s\\nstatic ip_address=%s\\n\",\n\t\tifaceName, ip)\n\tbody = append(body, []byte(add)...)\n\n\tgatewayIP := getGatewayIP(ifaceName)\n\tif len(gatewayIP) != 0 {\n\t\tadd = fmt.Sprintf(\"static routers=%s\\n\",\n\t\t\tgatewayIP)\n\t\tbody = append(body, []byte(add)...)\n\t}\n\n\tadd = fmt.Sprintf(\"static domain_name_servers=%s\\n\\n\",\n\t\tip4)\n\tbody = append(body, []byte(add)...)\n\n\terr = file.SafeWrite(\"/etc/dhcpcd.conf\", body)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (s *AppServerV3) setStaticFields() {\n\ts.Kind = KindAppServer\n\ts.Version = V3\n}", "func ServeStaticFS(c *gin.Context, path string) {\n\temb.ServeFile(subStatic, path, c.Writer, c.Request)\n}", "func dResponseWriter(w http.ResponseWriter, data interface{}, HStat int) error {\n\tdataType := reflect.TypeOf(data)\n\tif dataType.Kind() == reflect.String {\n\t\tw.WriteHeader(HStat)\n\t\tw.Header().Set(\"Content-Type\", \"application/text\")\n\n\t\t_, err := w.Write([]byte(data.(string)))\n\t\treturn err\n\t} else if reflect.PtrTo(dataType).Kind() == dataType.Kind() {\n\t\tw.WriteHeader(HStat)\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\n\t\toutData, err := json.MarshalIndent(data, \"\", \"\\t\")\n\t\tif err != nil {\n\t\t\tzerolog.Error().Msg(err.Error())\n\t\t\tw.Write([]byte(err.Error()))\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = w.Write(outData)\n\t\treturn err\n\t} else if reflect.Struct == dataType.Kind() {\n\t\tw.WriteHeader(HStat)\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\n\t\toutData, err := json.MarshalIndent(data, \"\", \"\\t\")\n\t\tif err != nil {\n\t\t\tzerolog.Error().Msg(err.Error())\n\t\t\tw.Write([]byte(err.Error()))\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = w.Write(outData)\n\t\treturn err\n\t} else if reflect.Slice == dataType.Kind() {\n\t\tw.WriteHeader(HStat)\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\n\t\toutData, err := json.MarshalIndent(data, \"\", \"\\t\")\n\t\tif err != nil {\n\t\t\tzerolog.Error().Msg(err.Error())\n\t\t\tw.Write([]byte(err.Error()))\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = w.Write(outData)\n\t\treturn err\n\t}\n\n\treturn errors.New(\"we could not be able to support data type that you passed\")\n}", "func (s *Dataset) Write(data interface{}, dtype *Datatype) error {\n\tvar addr uintptr\n\tv := reflect.ValueOf(data)\n\tswitch v.Kind() {\n\n\tcase reflect.Array:\n\t\taddr = v.UnsafeAddr()\n\n\tcase reflect.String:\n\t\tstr := (*reflect.StringHeader)(unsafe.Pointer(v.UnsafeAddr()))\n\t\taddr = str.Data\n\n\tcase reflect.Ptr:\n\t\taddr = v.Pointer()\n\n\tdefault:\n\t\taddr = v.Pointer()\n\t}\n\n\trc := C.H5Dwrite(s.id, dtype.id, 0, 0, 0, unsafe.Pointer(addr))\n\terr := h5err(rc)\n\treturn err\n}", "func (gateway *EgressGateway) AddStaticIptablesRule(setName string, sourceIPs []string, destinationIP, egressIP string) error {\n\n\t// create IPset from sourceIP's\n\tset, err := ipset.New(setName, \"hash:ip\", &ipset.Params{})\n\tif err != nil {\n\t\treturn errors.New(\"Failed to create ipset with name \" + setName + \" due to %\" + err.Error())\n\t}\n\tglog.Infof(\"Created ipset name: %s\", setName)\n\n\t// add IP's that need to be part of the ipset\n\tfor _, ip := range sourceIPs {\n\t\terr = set.Add(ip, 0)\n\t\tif err != nil {\n\t\t\treturn errors.New(\"Failed to add an ip \" + ip + \" into ipset with name \" + setName + \" due to %\" + err.Error())\n\t\t}\n\t}\n\tglog.Infof(\"Added ips %v to the ipset name: %s\", sourceIPs, setName)\n\n\truleSpec := []string{\"-m\", \"set\", \"--set\", setName, \"src\", \"-d\", destinationIP, \"-j\", \"ACCEPT\"}\n\thasRule, err := gateway.ipt.Exists(\"filter\", egressGatewayFWChainName, ruleSpec...)\n\tif err != nil {\n\t\treturn errors.New(\"Failed to verify rule exists in \" + egressGatewayFWChainName + \" chain of filter table\" + err.Error())\n\t}\n\tif !hasRule {\n\t\terr = gateway.ipt.Append(\"filter\", egressGatewayFWChainName, ruleSpec...)\n\t\tif err != nil {\n\t\t\treturn errors.New(\"Failed to add iptables command to ACCEPT traffic from director nodes to get forrwarded\" + err.Error())\n\t\t}\n\t}\n\tglog.Infof(\"Added rules in filter table FORWARD chain to permit traffic\")\n\n\truleSpec = []string{\"-m\", \"set\", \"--match-set\", setName, \"src\", \"-d\", destinationIP, \"-j\", \"SNAT\", \"--to-source\", egressIP}\n\tif err := gateway.insertRule(defaultNATIptable, egressGatewayNATChainName, 1, ruleSpec...); err != nil {\n\t\treturn fmt.Errorf(\"failed to insert rule to chain %v err %v\", defaultPostRoutingChain, err)\n\t}\n\n\treturn nil\n}", "func WriteSEN(w io.Writer, data interface{}, args ...interface{}) (err error) {\n\tpw := Writer{\n\t\tOptions: ojg.DefaultOptions,\n\t\tWidth: 80,\n\t\tMaxDepth: 3,\n\t\tSEN: true,\n\t}\n\tpw.w = w\n\tpw.config(args)\n\t_, err = pw.encode(data)\n\n\treturn\n}", "func vizFileWrite(v *VizceralGraph) {\n\tvJson, _ := json.Marshal(*v)\n\tsJson := fmt.Sprintf(\"%s\", vJson)\n\n\tdf := os.Getenv(\"TRAFFIC_URL\")\n\tdataFile := string(\"/usr/src/app/dist/\" + df)\n\tcreateFile(dataFile)\n\tWriteFile(dataFile, sJson)\n}", "func Static(v map[string]string) Source {\n\treturn SourceFunc(func() (map[string]string, error) {\n\t\treturn v, nil\n\t})\n}", "func (r *Registry) staticSaveEntry(v *value, used bool) error {\n\tvar entry persistedEntry\n\tvar err error\n\tif used {\n\t\tentry, err = newPersistedEntry(v)\n\t}\n\tif err != nil {\n\t\treturn errors.AddContext(err, \"Save: failed to get persistedEntry from key-value pair\")\n\t}\n\tb, err := entry.Marshal()\n\tif err != nil {\n\t\treturn errors.AddContext(err, \"Save: failed to marshal persistedEntry\")\n\t}\n\t_, err = r.staticFile.WriteAt(b, v.staticIndex*PersistedEntrySize)\n\tif err != nil {\n\t\treturn errors.AddContext(err, \"failed to save entry\")\n\t}\n\treturn nil\n}", "func Static(val interface{}) YAMLOption {\n\tbs, err := yaml.Marshal(val)\n\tif err != nil {\n\t\treturn failed(err)\n\t}\n\treturn optionFunc(func(c *config) {\n\t\tc.sources = append(c.sources, source{bytes: bs})\n\t})\n}", "func AddStaticBinary(mux Mux, path string, content []byte) {\n\tmux.Handle(\n\t\tpath,\n\t\thttp.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\thttp.ServeContent(w, r, path, kAppStart, bytes.NewReader(content))\n\t\t}))\n}", "func staticFile(mux *http.ServeMux, name string) {\n\tabs := filepath.Join(service.Static.DiskPath, name)\n\tmux.HandleFunc(name, func(w http.ResponseWriter, r *http.Request) {\n\t\thttp.ServeFile(w, r, abs)\n\t})\n}", "func NewWriteHandler(logger log.Logger, reg prometheus.Registerer, appendable storage.Appendable) http.Handler {\n\th := &writeHandler{\n\t\tlogger: logger,\n\t\tappendable: appendable,\n\n\t\tsamplesWithInvalidLabelsTotal: prometheus.NewCounter(prometheus.CounterOpts{\n\t\t\tNamespace: \"prometheus\",\n\t\t\tSubsystem: \"api\",\n\t\t\tName: \"remote_write_invalid_labels_samples_total\",\n\t\t\tHelp: \"The total number of remote write samples which contains invalid labels.\",\n\t\t}),\n\t}\n\tif reg != nil {\n\t\treg.MustRegister(h.samplesWithInvalidLabelsTotal)\n\t}\n\treturn h\n}", "func (t colType) dataWrite(e *binary.Encoder, data interface{}) error {\n\tif !t.valid {\n\t\treturn errors.New(\"typeInfo.dataWrite: missing type properties\")\n\t}\n\n\t// 0 length for null value\n\tif data == nil {\n\t\tif t.writeOptions&isLong != 0 && t.writeOptions&isLob == 0 {\n\t\t\te.WriteUint32(0)\n\t\t} else {\n\t\t\te.WriteUint8(0)\n\t\t}\n\t\treturn nil\n\t}\n\n\t// write data length for nullable fixed length types\n\tif t.encodingProps.numBytes != 0 && t.writeOptions&isNullable != 0 {\n\t\te.WriteInt8(t.encodingProps.numBytes)\n\t}\n\n\treturn t.encodingProps.writer(e, data, t)\n}", "func (cfw *CoverageDataWriter) Write(metaFileHash [16]byte, args map[string]string, visitor CounterVisitor) error {\n\tif err := cfw.writeHeader(metaFileHash); err != nil {\n\t\treturn err\n\t}\n\treturn cfw.AppendSegment(args, visitor)\n}", "func (sf *File) writeData(w io.Writer) error {\n\tif sf.NoObs == 0 {\n\t\treturn nil\n\t}\n\tif len(sf.fields) == 0 {\n\t\treturn fmt.Errorf(\"No fields\")\n\t}\n\tbs := make([]byte, sf.recordSize)\n\tfor i := int32(0); i < sf.NoObs; i++ {\n\t\toffset := 0\n\t\tfor _, f := range sf.fields {\n\t\t\tswitch f.FieldType {\n\t\t\tcase StataByteId:\n\t\t\t\tv := f.data.([]Byte)[i]\n\t\t\t\tbs[offset] = byte(v)\n\t\t\t\toffset++\n\t\t\tcase StataIntId:\n\t\t\t\tv := f.data.([]Int)[i]\n\t\t\t\tbs[offset] = byte(v)\n\t\t\t\toffset++ //incrementing the offset instead of using bs[offset+1] to avoid doing the addition twice\n\t\t\t\tbs[offset] = byte(v >> 8)\n\t\t\t\toffset++\n\t\t\tcase StataLongId:\n\t\t\t\tbase := *(*[4]byte)(unsafe.Pointer(&f.data.([]Long)[i]))\n\t\t\t\tcopy(bs[offset:], base[:])\n\t\t\t\toffset += 4\n\t\t\tcase StataFloatId:\n\t\t\t\tbase := *(*[4]byte)(unsafe.Pointer(&f.data.([]Float)[i]))\n\t\t\t\tcopy(bs[offset:], base[:])\n\t\t\t\toffset += 4\n\t\t\tcase StataDoubleId:\n\t\t\t\tbase := *(*[8]byte)(unsafe.Pointer(&f.data.([]Double)[i]))\n\t\t\t\tcopy(bs[offset:], base[:])\n\t\t\t\toffset += 8\n\t\t\tdefault:\n\t\t\t\treturn fmt.Errorf(\"Field type [%d] not supported in field %s\", f.FieldType, f.Name)\n\t\t\t}\n\t\t}\n\t\tif _, err := w.Write(bs); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (s *BaseSyslParserListener) ExitHttp_path_static(ctx *Http_path_staticContext) {}", "func Write(w io.Writer, byteorder binary.ByteOrder, data interface{}) error {\n\n\tswitch data.(type) {\n\tcase uint8, uint16, uint32, uint64, int8, int16, int32, int64, float32, float64:\n\t\treturn binary.Write(w, byteorder, data)\n\t}\n\n\tv := reflect.ValueOf(data)\n\n\tswitch v.Kind() {\n\tcase reflect.Array, reflect.Slice:\n\t\tif v.Kind() == reflect.Slice && v.Type().Elem().Kind() == reflect.Uint8 {\n\t\t\t_, err := w.Write(v.Bytes())\n\t\t\treturn err\n\t\t}\n\n\t\tl := v.Len()\n\t\tfor i := 0; i < l; i++ {\n\t\t\terr := Write(w, byteorder, v.Index(i).Interface())\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\tcase reflect.Struct:\n\t\t// write all public fields in order\n\t\ttyp := v.Type()\n\t\tl := typ.NumField()\n\t\tfor i := 0; i < l; i++ {\n\t\t\tf := typ.Field(i)\n\t\t\tif f.PkgPath != \"\" {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tfval := v.Field(i)\n\n\t\t\topts := parseTag(f.Tag)\n\t\t\tvar bOrder binary.ByteOrder = byteorder\n\t\t\tif opts.endian != nil {\n\t\t\t\tbOrder = opts.endian\n\t\t\t}\n\n\t\t\t// if we have a slice embedded in a struct, get the struct tag that tells us how to write the (unknown) length before the contents\n\t\t\tif f.Type.Kind() == reflect.Slice {\n\t\t\t\tslen := uint64(fval.Len())\n\n\t\t\t\tif opts.lenprefix == \"\" {\n\t\t\t\t\treturn ErrMissingLenPrefix\n\t\t\t\t}\n\n\t\t\t\tmaxlen, ok := maxSliceLen[opts.lenprefix]\n\t\t\t\tif !ok {\n\t\t\t\t\treturn ErrUnknownLenPrefix\n\t\t\t\t}\n\n\t\t\t\tif slen > maxlen {\n\t\t\t\t\treturn ErrSliceTooLarge\n\t\t\t\t}\n\n\t\t\t\tvar err error\n\t\t\t\tswitch opts.lenprefix {\n\t\t\t\tcase \"uint8\":\n\t\t\t\t\terr = binary.Write(w, bOrder, uint8(slen))\n\n\t\t\t\tcase \"uint16\":\n\t\t\t\t\terr = binary.Write(w, bOrder, uint16(slen))\n\n\t\t\t\tcase \"uint32\":\n\t\t\t\t\terr = binary.Write(w, bOrder, uint32(slen))\n\n\t\t\t\tcase \"uint64\":\n\t\t\t\t\terr = binary.Write(w, bOrder, slen)\n\n\t\t\t\tcase \"int8\":\n\t\t\t\t\terr = binary.Write(w, bOrder, int8(slen))\n\n\t\t\t\tcase \"int16\":\n\t\t\t\t\terr = binary.Write(w, bOrder, int16(slen))\n\n\t\t\t\tcase \"int32\":\n\t\t\t\t\terr = binary.Write(w, bOrder, int32(slen))\n\n\t\t\t\tcase \"int64\":\n\t\t\t\t\terr = binary.Write(w, bOrder, int64(slen))\n\t\t\t\t}\n\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t}\n\n\t\t\terr := Write(w, bOrder, v.Field(i).Interface())\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n\n}", "func WriteCoinData(slug, data string, d interface{}) bool {\n\tdc := mod.Cache{Data: d}\n\treturn DB.Write(cfg.Web+\"/data/\"+slug, data, dc) == nil\n}", "func tryStaticHTML(ctx *fasthttp.RequestCtx, p IPage) error {\n\tif !p.Attr().CacheOption.IsStaticHTML {\n\t\treturn nil\n\t}\n\n\tfilename := util.Bytes2Str(ctx.RequestURI())\n\tl := len(filename)\n\tif l < 1 {\n\t\treturn nil\n\t}\n\n\tif filename[l-1] == '/' {\n\t\tfilename += \"index.html\"\n\t} else {\n\t\tfilename += \".html\"\n\t}\n\n\tfile, err := os.OpenFile(filepath.Join(HomeDir, WEBROOT, filename), os.O_CREATE|os.O_TRUNC, os.ModePerm)\n\tdefer file.Close()\n\tfile.Write(ctx.Response.Body())\n\n\treturn err\n}", "func (c *digisparkI2cConnection) WriteWordData(reg uint8, val uint16) error {\n\tc.mtx.Lock()\n\tdefer c.mtx.Unlock()\n\n\tlow := uint8(val & 0xff)\n\thigh := uint8((val >> 8) & 0xff)\n\tbuf := []byte{reg, low, high}\n\treturn c.writeAndCheckCount(buf, true)\n}", "func (s *DatabaseServerV3) setStaticFields() {\n\ts.Kind = KindDatabaseServer\n\ts.Version = V3\n}", "func StaticProvider(data map[string]interface{}) ProviderFunc {\n\treturn func() (Provider, error) {\n\t\treturn NewStaticProvider(data), nil\n\t}\n}", "func (w *DataFileWriter) Write(v interface{}) error {\n\tw.blockCount++\n\terr := w.datumWriter.Write(v, w.blockEnc)\n\treturn err\n}", "func (d *SQLdataloader) Write(s string) error {\n\tif len(s) == 0 { // don't try to write zero bytes\n\t\treturn nil\n\t}\n\tvar err error // needed on multiple paths\n\tif d.reccount >= d.recmax { // if enough recs for a LOAD DATA\n\t\terr = d.doload() // load the data into the database\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t} // do it\n\tif d.fd == nil { // if need to start a file\n\t\td.fd, err = ioutil.TempFile(\"\", \"SQLBULKLOAD-\")\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\td.buf = bufio.NewWriter(d.fd) // create a new buffered writer\n\t\td.totalcount = d.totalcount + int64(d.reccount) // tally\n\t\td.reccount = 0 // it's empty\n\t}\n\t_, err = d.buf.WriteString(s) // write the string to the file\n\td.reccount++ // tally\n\treturn err // return status\n}", "func (w *win) writeData(data []byte) {\n\tconst maxWrite = 512\n\tfor len(data) > 0 {\n\t\tsz := len(data)\n\t\tif sz > maxWrite {\n\t\t\tsz = maxWrite\n\t\t}\n\t\tn, err := w.Write(\"data\", data[:sz])\n\t\tif err != nil {\n\t\t\tpanic(\"Failed to write to window: \" + err.Error())\n\t\t}\n\t\tdata = data[n:]\n\t}\n}", "func write(w io.Writer, data map[string]specs, pkg string) error {\n\ttmpl, err := template.ParseFiles(\"cpu_table.go.template\")\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn tmpl.Execute(w, Template{\n\t\tPackage: pkg,\n\t\tData: data,\n\t})\n}", "func (s *BaseSyslParserListener) EnterHttp_path_static(ctx *Http_path_staticContext) {}", "func StaticFileServer(r chi.Router, public string, static string) {\n\n\t// everything up to the r.Get call is executed the first time the function is called\n\tif strings.ContainsAny(public, \"{}*\") {\n\t\tpanic(\"FileServer does not permit URL parameters.\")\n\t}\n\n\troot, _ := filepath.Abs(static)\n\tif _, err := os.Stat(root); os.IsNotExist(err) {\n\t\tpanic(\"Static Documents Directory Not Found\")\n\t}\n\n\tfs := http.StripPrefix(public, http.FileServer(http.Dir(root)))\n\n\tif public != \"/\" && public[len(public)-1] != '/' {\n\t\tr.Get(public, http.RedirectHandler(public+\"/\", 301).ServeHTTP)\n\t\tpublic += \"/\"\n\t}\n\n\tlog.Printf(\"Serving spa index.html from: %s\", http.Dir(root))\n\n\t// Register the Get request for the specified path, most likely /*\n\tr.Get(public+\"*\", func(w http.ResponseWriter, r *http.Request) {\n\t\tfile := strings.Replace(r.RequestURI, public, \"/\", 1)\n\t\t// if the requested resource was not found, pass the request to the client\n\t\tif _, err := os.Stat(root + file); os.IsNotExist(err) {\n\t\t\thttp.ServeFile(w, r, path.Join(root, \"index.html\"))\n\t\t\treturn\n\t\t}\n\t\t// if the requested resource was found, serve it\n\t\tfs.ServeHTTP(w, r)\n\t})\n}", "func (d *DriverFile) Write(folder string) {\n\ttag := \"NikuDataBus\"\n\tif folder == constant.FolderDebug {\n\t\ttag = \"XOGOutput\"\n\t}\n\tr, _ := regexp.Compile(\"(?s)<\" + tag + \"(.*)</\" + tag + \">\")\n\tstr := r.FindString(d.xogXML)\n\tif str == constant.Undefined {\n\t\tstr = d.xogXML\n\t}\n\tioutil.WriteFile(folder+d.Type+\"/\"+d.Path, []byte(str), os.ModePerm)\n}", "func WriteDataset(ctx context.Context, store cafs.Filestore, ds *dataset.Dataset, pin bool) (string, error) {\n\n\tif ds == nil || ds.IsEmpty() {\n\t\treturn \"\", fmt.Errorf(\"cannot save empty dataset\")\n\t}\n\tname := ds.Name // preserve name for body file\n\tbodyFile := ds.BodyFile()\n\tfileTasks := 0\n\taddedDataset := false\n\tadder, err := store.NewAdder(pin, true)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"error creating new adder: %s\", err.Error())\n\t}\n\n\tif ds.Viz != nil {\n\t\tds.Viz.DropTransientValues()\n\t\tvizScript := ds.Viz.ScriptFile()\n\t\tvizRendered := ds.Viz.RenderedFile()\n\t\t// add task for the viz.json\n\t\tfileTasks++\n\t\tif vizRendered != nil {\n\t\t\t// add the rendered visualization\n\t\t\t// and add working group for adding the viz script file\n\t\t\tfileTasks += 2\n\t\t\tvrFile := qfs.NewMemfileReader(PackageFileRenderedViz.String(), vizRendered)\n\t\t\tdefer vrFile.Close()\n\t\t\tadder.AddFile(ctx, vrFile)\n\t\t} else if vizScript != nil {\n\t\t\t// add the vizScript\n\t\t\tfileTasks++\n\t\t\tvsFile := qfs.NewMemfileReader(vizScriptFilename, vizScript)\n\t\t\tdefer vsFile.Close()\n\t\t\tadder.AddFile(ctx, vsFile)\n\t\t} else {\n\t\t\tvizdata, err := json.Marshal(ds.Viz)\n\t\t\tif err != nil {\n\t\t\t\treturn \"\", fmt.Errorf(\"error marshalling dataset viz to json: %s\", err.Error())\n\t\t\t}\n\t\t\tadder.AddFile(ctx, qfs.NewMemfileBytes(PackageFileViz.String(), vizdata))\n\t\t}\n\t}\n\n\tif ds.Meta != nil {\n\t\tmdf, err := JSONFile(PackageFileMeta.String(), ds.Meta)\n\t\tif err != nil {\n\t\t\treturn \"\", fmt.Errorf(\"error marshaling metadata to json: %s\", err.Error())\n\t\t}\n\t\tfileTasks++\n\t\tadder.AddFile(ctx, mdf)\n\t}\n\n\tif ds.Transform != nil {\n\t\t// TODO (b5): this is validation logic, should happen before WriteDataset is ever called\n\t\t// all resources must be references\n\t\tfor key, r := range ds.Transform.Resources {\n\t\t\tif r.Path == \"\" {\n\t\t\t\treturn \"\", fmt.Errorf(\"transform resource %s requires a path to save\", key)\n\t\t\t}\n\t\t}\n\n\t\tsr := ds.Transform.ScriptFile()\n\t\tds.Transform.DropTransientValues()\n\t\tif sr != nil {\n\t\t\tfileTasks++\n\t\t\ttsFile := qfs.NewMemfileReader(transformScriptFilename, sr)\n\t\t\tdefer tsFile.Close()\n\t\t\tadder.AddFile(ctx, tsFile)\n\t\t\t// NOTE - add wg for the transform.json file ahead of time, which isn't completed\n\t\t\t// until after scriptPath has been added\n\t\t\tfileTasks++\n\t\t} else {\n\t\t\ttfdata, err := json.Marshal(ds.Transform)\n\t\t\tif err != nil {\n\t\t\t\treturn \"\", fmt.Errorf(\"error marshalling dataset transform to json: %s\", err.Error())\n\t\t\t}\n\n\t\t\tfileTasks++\n\t\t\tadder.AddFile(ctx, qfs.NewMemfileBytes(PackageFileTransform.String(), tfdata))\n\t\t}\n\t}\n\n\tif ds.Commit != nil {\n\t\tds.Commit.DropTransientValues()\n\t\tcmf, err := JSONFile(PackageFileCommit.String(), ds.Commit)\n\t\tif err != nil {\n\t\t\treturn \"\", fmt.Errorf(\"error marshilng dataset commit message to json: %s\", err.Error())\n\t\t}\n\t\tfileTasks++\n\t\tadder.AddFile(ctx, cmf)\n\t}\n\n\tif ds.Structure != nil {\n\t\tds.Structure.DropTransientValues()\n\t\tstf, err := JSONFile(PackageFileStructure.String(), ds.Structure)\n\t\tif err != nil {\n\t\t\treturn \"\", fmt.Errorf(\"error marshaling dataset structure to json: %s\", err.Error())\n\t\t}\n\t\tfileTasks++\n\t\tadder.AddFile(ctx, stf)\n\t}\n\n\tfileTasks++\n\tadder.AddFile(ctx, bodyFile)\n\n\tvar path string\n\tdone := make(chan error, 0)\n\tgo func() {\n\t\tfor ao := range adder.Added() {\n\t\t\tpath = ao.Path\n\t\t\tswitch ao.Name {\n\t\t\tcase PackageFileStructure.String():\n\t\t\t\tds.Structure = dataset.NewStructureRef(ao.Path)\n\t\t\tcase PackageFileTransform.String():\n\t\t\t\tds.Transform = dataset.NewTransformRef(ao.Path)\n\t\t\tcase PackageFileMeta.String():\n\t\t\t\tds.Meta = dataset.NewMetaRef(ao.Path)\n\t\t\tcase PackageFileCommit.String():\n\t\t\t\tds.Commit = dataset.NewCommitRef(ao.Path)\n\t\t\tcase PackageFileViz.String():\n\t\t\t\tds.Viz = dataset.NewVizRef(ao.Path)\n\t\t\tcase bodyFile.FileName():\n\t\t\t\tds.BodyPath = ao.Path\n\t\t\t\t// ds.SetBodyFile(qfs.NewMemfileBytes(bodyFile.FileName(), bodyBytesBuf.Bytes()))\n\t\t\tcase transformScriptFilename:\n\t\t\t\tds.Transform.ScriptPath = ao.Path\n\t\t\t\ttfdata, err := json.Marshal(ds.Transform)\n\t\t\t\tif err != nil {\n\t\t\t\t\tdone <- err\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\t// Add the encoded transform file, decrementing the stray fileTasks from above\n\t\t\t\tadder.AddFile(ctx, qfs.NewMemfileBytes(PackageFileTransform.String(), tfdata))\n\t\t\tcase PackageFileRenderedViz.String():\n\t\t\t\tds.Viz.RenderedPath = ao.Path\n\t\t\t\tvsFile := qfs.NewMemfileReader(vizScriptFilename, ds.Viz.ScriptFile())\n\t\t\t\tdefer vsFile.Close()\n\t\t\t\tadder.AddFile(ctx, vsFile)\n\t\t\tcase vizScriptFilename:\n\t\t\t\tds.Viz.ScriptPath = ao.Path\n\t\t\t\tvizdata, err := json.Marshal(ds.Viz)\n\t\t\t\tif err != nil {\n\t\t\t\t\tdone <- err\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\t// Add the encoded transform file, decrementing the stray fileTasks from above\n\t\t\t\tadder.AddFile(ctx, qfs.NewMemfileBytes(PackageFileViz.String(), vizdata))\n\t\t\t}\n\n\t\t\tfileTasks--\n\t\t\tif fileTasks == 0 {\n\t\t\t\tif !addedDataset {\n\t\t\t\t\tds.DropTransientValues()\n\t\t\t\t\tdsdata, err := json.Marshal(ds)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tdone <- err\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\n\t\t\t\t\tadder.AddFile(ctx, qfs.NewMemfileBytes(PackageFileDataset.String(), dsdata))\n\t\t\t\t}\n\t\t\t\t//\n\t\t\t\tif err := adder.Close(); err != nil {\n\t\t\t\t\tdone <- err\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tdone <- nil\n\t}()\n\n\terr = <-done\n\tif err != nil {\n\t\treturn path, err\n\t}\n\t// TODO (b5): currently we're loading to keep the ds pointer hydrated post-write\n\t// we should remove that assumption, allowing callers to skip this load step, which may\n\t// be unnecessary\n\tvar loaded *dataset.Dataset\n\tloaded, err = LoadDataset(ctx, store, path)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tloaded.Name = name\n\t*ds = *loaded\n\treturn path, nil\n}", "func (app *Application) setupStatic() {\n\tapp.App.Static(\n\t\tapp.Config.Static.Prefix,\n\t\tapp.Config.Static.Root,\n\t\tfiber.Static{\n\t\t\tIndex: app.Config.Static.Index,\n\t\t},\n\t)\n}", "func writeListOfEndpoints(w http.ResponseWriter, r *http.Request, funcMap map[string]*RPCFunc) {\n\thasArgs := make(map[string]string)\n\tnoArgs := make(map[string]string)\n\tfor name, rf := range funcMap {\n\t\tbase := fmt.Sprintf(\"//%s/%s\", r.Host, name)\n\t\tif len(rf.args) == 0 {\n\t\t\tnoArgs[name] = base\n\t\t\tcontinue\n\t\t}\n\t\tvar query []string\n\t\tfor _, arg := range rf.args {\n\t\t\tquery = append(query, arg.name+\"=_\")\n\t\t}\n\t\thasArgs[name] = base + \"?\" + strings.Join(query, \"&\")\n\t}\n\tw.Header().Set(\"Content-Type\", \"text/html\")\n\t_ = listOfEndpoints.Execute(w, map[string]map[string]string{\n\t\t\"NoArgs\": noArgs,\n\t\t\"HasArgs\": hasArgs,\n\t})\n}", "func (f WriteHandlerFunc) ServeWrite(req Request, rsp ResponseWriter) {\n\tf(req, rsp)\n}", "func (group *RouterGroup) Static(relativePath, root string) IRoutes {\n\treturn group.StaticFS(relativePath, Dir(root, false))\n}", "func StaticDir(staticDir string) Option {\n\treturn func(s *Service) {\n\t\ts.staticDir = staticDir\n\t}\n}", "func WriteData(text interface{}) {\n\tfmt.Printf(\" %s\\n\", text)\n}", "func NewStaticProvider(data map[string]interface{}) Provider {\n\treturn &staticProvider{\n\t\tdata: data,\n\t}\n}", "func (c *CrawlerState) WriteSiteMap(f io.Writer) {\n\txmlTemplate := `<?xml version=\"1.0\" encoding=\"UTF-8\"?>\n\n<urlset xmlns=\"http://www.sitemaps.org/schemas/sitemap/0.9\">\n{{range $element := . }}\n <url>\n <loc>{{$element}}</loc>\n </url>{{end}}\n</urlset>\n`\n\ttmpl, err := template.New(\"test\").Parse(xmlTemplate)\n\tif err != nil {\n\t\tlog.Error(err)\n\t\treturn\n\t}\n\terr = tmpl.Execute(f, c.urls)\n\tif err != nil {\n\t\tlog.Error(err)\n\t}\n}", "func (c *logCommonBlock) WriteDataEntry(buf *bytes.Buffer) *bytes.Buffer {\n\tbuf.WriteByte('{')\n\tif c.attributes != nil {\n\t\tw := internal.JSONFieldsWriter{Buf: buf}\n\t\tw.AddKey(c.attributes.DataTypeKey())\n\t\tc.attributes.WriteDataEntry(buf)\n\t}\n\tbuf.WriteByte('}')\n\treturn buf\n}", "func (hg *HostGroup) GetStaticInfo(ctx context.Context, params StaticInfoParams) ([]byte, error) {\n\treturn hg.client.PostInOut(ctx, \"/api/v1.0/HostGroup.GetStaticInfo\", params, nil)\n}", "func (statics *AssestStruct) FileHandlerFunc(name string) http.HandlerFunc {\n\tif strings.Contains(name, \"private\") {\n\t\treturn http.NotFound\n\t}\n\tstatic, err := statics.GetAssestFile(name)\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tif err != nil {\n\t\t\thttp.NotFound(w, r)\n\t\t\treturn\n\t\t}\n\t\tmodtime := time.Unix(static.Mtime, 0)\n\t\tmodifiedSince := r.Header.Get(\"If-Modified-Since\")\n\t\tif modifiedSince != \"\" {\n\t\t\tt, err := time.Parse(http.TimeFormat, modifiedSince)\n\t\t\tif err == nil && modtime.Before(t.Add(1*time.Second)) {\n\t\t\t\tw.Header().Del(\"Content-Type\")\n\t\t\t\tw.Header().Del(\"Content-Length\")\n\t\t\t\tw.Header().Set(\"Last-Modified\", modtime.UTC().Format(http.TimeFormat))\n\t\t\t\tw.WriteHeader(http.StatusNotModified)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t\tmimeType := mime.TypeByExtension(filepath.Ext(static.Name))\n\t\tif mimeType != \"\" {\n\t\t\tw.Header().Set(\"Content-Type\", mimeType)\n\t\t}\n\t\tw.Header().Set(\"Last-Modified\", modtime.UTC().Format(http.TimeFormat))\n\t\tw.Write([]byte(static.Content))\n\t}\n}", "func (group *RouterGroup) Static(relativePath string, root string) {\n\tpattern := path.Join(relativePath, \"/*filepath\") // custom parameter name\n\thandler := group.createStaticHandler(relativePath, http.Dir(root))\n\tgroup.GET(pattern, handler)\n}" ]
[ "0.58449274", "0.5381385", "0.5305445", "0.5223065", "0.50529706", "0.49200627", "0.48813868", "0.47990835", "0.47695982", "0.47620812", "0.47569618", "0.47276986", "0.46729904", "0.46655488", "0.46359763", "0.46199363", "0.4617461", "0.46172822", "0.46101165", "0.45811358", "0.4579009", "0.4577851", "0.45597348", "0.45487136", "0.454523", "0.45402896", "0.45279714", "0.45221418", "0.44773087", "0.44678453", "0.4459626", "0.44552088", "0.4450907", "0.44494727", "0.44489554", "0.44472837", "0.44466916", "0.44421065", "0.44416866", "0.44400194", "0.44396147", "0.44271177", "0.44245115", "0.4420229", "0.44179666", "0.4416892", "0.4416892", "0.44104874", "0.44056794", "0.44049582", "0.4383708", "0.43812412", "0.4374331", "0.43606973", "0.4357442", "0.43553627", "0.43469378", "0.43369365", "0.43177846", "0.43161845", "0.43127462", "0.4309921", "0.43080384", "0.42811725", "0.42759258", "0.42746058", "0.42705053", "0.4270384", "0.42656147", "0.42466497", "0.4241888", "0.4241114", "0.4235419", "0.42192137", "0.42163622", "0.42123166", "0.42081732", "0.42043832", "0.4202332", "0.4197023", "0.4193788", "0.41929066", "0.4190281", "0.4185829", "0.4182778", "0.41766936", "0.41757494", "0.41753662", "0.41743663", "0.41719916", "0.41711515", "0.41534218", "0.41455585", "0.41395426", "0.41384262", "0.41319132", "0.41267073", "0.41201138", "0.41138577", "0.41095656" ]
0.7692826
0
WriteEndpointConfig writes the BPF configuration for the endpoint to a writer.
func (h *HeaderfileWriter) WriteEndpointConfig(w io.Writer, e datapath.EndpointConfiguration) error { fw := bufio.NewWriter(w) writeIncludes(w) h.writeStaticData(fw, e) return h.writeTemplateConfig(fw, e) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (h *HeaderfileWriter) WriteTemplateConfig(w io.Writer, e datapath.EndpointConfiguration) error {\n\tfw := bufio.NewWriter(w)\n\treturn h.writeTemplateConfig(fw, e)\n}", "func (m *LogicAppTriggerEndpointConfiguration) Serialize(writer i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.SerializationWriter)(error) {\n err := m.CustomExtensionEndpointConfiguration.Serialize(writer)\n if err != nil {\n return err\n }\n {\n err = writer.WriteStringValue(\"logicAppWorkflowName\", m.GetLogicAppWorkflowName())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"resourceGroupName\", m.GetResourceGroupName())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"subscriptionId\", m.GetSubscriptionId())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"url\", m.GetUrl())\n if err != nil {\n return err\n }\n }\n return nil\n}", "func (mod *EthModule) WriteConfig(config_file string) error {\n\tb, err := json.Marshal(mod.eth.config)\n\tif err != nil {\n\t\tfmt.Println(\"error marshalling config:\", err)\n\t\treturn err\n\t}\n\tvar out bytes.Buffer\n\tjson.Indent(&out, b, \"\", \"\\t\")\n\terr = ioutil.WriteFile(config_file, out.Bytes(), 0600)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (fes *FrontEndService) WriteConfig(ctx context.Context, cancel context.CancelFunc) {\n\tfile, err := os.Create(fes.birdConfFile)\n\tif err != nil {\n\t\tlogrus.Fatalf(\"FrontEndService: failed to create %v, err: %v\", fes.birdConfFile, err)\n\t\tcancel()\n\t}\n\tdefer file.Close()\n\n\t//conf := \"include \\\"bird-common.conf\\\";\\n\"\n\t//conf += \"\\n\"\n\tconf := \"\"\n\tfes.WriteConfigBase(&conf)\n\thasVIP4, hasVIP6 := fes.WriteConfigVips(&conf)\n\tif len(fes.vrrps) > 0 {\n\t\tfes.WriteConfigVRRPs(&conf, hasVIP4, hasVIP6)\n\t} else if fes.dropIfNoPeer {\n\t\tfes.WriteConfigDropIfNoPeer(&conf, hasVIP4, hasVIP6)\n\t}\n\tfes.WriteConfigKernel(&conf, hasVIP4, hasVIP6)\n\tfes.WriteConfigBGP(&conf)\n\n\tlogrus.Infof(\"FrontEndService: config generated\")\n\tlogrus.Debugf(\"\\n%v\", conf)\n\t_, err = file.WriteString(conf)\n\tif err != nil {\n\t\tlogrus.Fatalf(\"FrontEndService: failed to write %v, err: %v\", fes.birdConfFile, err)\n\t\tcancel()\n\t}\n}", "func (s *Service) WriteToEndpointFile() (path string, err error) {\n\tdata, err := json.Marshal(s)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tpath = defaultEndpointFile()\n\tif err := os.MkdirAll(filepath.Dir(path), 0755); err != nil {\n\t\treturn \"\", err\n\t}\n\tif err := ioutil.WriteFile(path, data, 0755); err != nil {\n\t\treturn \"\", err\n\t}\n\treturn path, nil\n}", "func (s *PendingDeploymentSummary) SetEndpointConfigName(v string) *PendingDeploymentSummary {\n\ts.EndpointConfigName = &v\n\treturn s\n}", "func editEndpointInConfig(t *testing.T, configFilePath, endpoint string) {\n\tvar config cli.Config\n\tb, err := ioutil.ReadFile(configFilePath)\n\trequire.NoError(t, err)\n\n\terr = yaml.Unmarshal(b, &config)\n\trequire.NoError(t, err)\n\n\tconfig.Endpoint = endpoint\n\n\tb, err = yaml.Marshal(&config)\n\trequire.NoError(t, err)\n\n\terr = ioutil.WriteFile(configFilePath, b, 0655)\n\trequire.NoError(t, err)\n\n}", "func WriteConfig(f string, c *SupportedBranchesConfig) error {\n\treturn util.WithWriteFile(f, func(w io.Writer) error {\n\t\treturn EncodeConfig(w, c)\n\t})\n}", "func WriteConfig(c Config, filename string) {\n\n\tdata, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\terr = ioutil.WriteFile(filename, data, 664)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func (b *Bot) WriteConfig(fn configCallback) {\n\tfn(b.conf)\n}", "func (s *InferenceRecommendation) SetEndpointConfiguration(v *EndpointOutputConfiguration) *InferenceRecommendation {\n\ts.EndpointConfiguration = v\n\treturn s\n}", "func (fes *FrontEndService) WriteConfigVips(conf *string) (hasVIP4, hasVIP6 bool) {\n\tv4, v6 := \"\", \"\"\n\thasVIP4, hasVIP6 = false, false\n\n\tfor _, vip := range fes.vips {\n\t\tif isIPv6(vip) {\n\t\t\t// IPv6\n\t\t\t//v6 += \"\\troute \" + vip + \" blackhole;\\n\"\n\t\t\tv6 += \"\\troute \" + vip + \" via \\\"lo\\\";\\n\"\n\t\t} else if isIPv4(vip) {\n\t\t\t// IPv4\n\t\t\t//v4 += \"\\troute \" + vip + \" blackhole;\\n\"\n\t\t\tv4 += \"\\troute \" + vip + \" via \\\"lo\\\";\\n\"\n\t\t}\n\t}\n\n\tif v4 != \"\" {\n\t\thasVIP4 = true\n\t\t*conf += \"protocol static VIP4 {\\n\"\n\t\t*conf += \"\\tipv4 { preference 110; };\\n\"\n\t\t*conf += v4\n\t\t*conf += \"}\\n\"\n\t\t*conf += \"\\n\"\n\t}\n\n\tif v6 != \"\" {\n\t\thasVIP6 = true\n\t\t*conf += \"protocol static VIP6 {\\n\"\n\t\t*conf += \"\\tipv6 { preference 110; };\\n\"\n\t\t*conf += v6\n\t\t*conf += \"}\\n\"\n\t\t*conf += \"\\n\"\n\t}\n\treturn\n}", "func (s *RecommendationJobInferenceBenchmark) SetEndpointConfiguration(v *EndpointOutputConfiguration) *RecommendationJobInferenceBenchmark {\n\ts.EndpointConfiguration = v\n\treturn s\n}", "func (c *Config) WriteConfig(filename string) error {\n\tdata, err := yaml.Marshal(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = ioutil.WriteFile(filename, data, 0644)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (r *Resource) WriteConfig(filename string) error {\n\tr.Lock()\n\tdefer r.Unlock()\n\n\tvar b bytes.Buffer\n\n\tb.WriteString(fmt.Sprintf(\"# meta-data-json:{\\\"updated\\\": \\\"%s\\\"}\\n\", time.Now().UTC()))\n\tb.WriteString(fmt.Sprintf(\"resource %s {\\n\", r.name))\n\n\tvar hosts []string\n\tfor _, h := range r.host {\n\t\thosts = append(hosts, h.Name)\n\n\t\tb.WriteString(indentf(1, \"on %s {\\n\", h.Name))\n\t\tb.WriteString(indentf(2, \"node-id %d;\\n\", h.ID))\n\t\tb.WriteString(indentf(2, \"address %s:%d;\\n\", h.IP, r.port))\n\t\tfor _, v := range h.volume {\n\t\t\tb.WriteString(indentf(2, \"volume %d {\\n\", v.id))\n\t\t\tb.WriteString(indentf(3, \"device minor %d;\\n\", v.minor))\n\t\t\tb.WriteString(indentf(3, \"disk %s;\\n\", v.backingDevice))\n\t\t\tb.WriteString(indentf(3, \"meta-disk internal;\\n\"))\n\t\t\tb.WriteString(indentf(2, \"}\\n\")) // end volume section\n\t\t}\n\t\tb.WriteString(indentf(1, \"}\\n\")) // end on section\n\t\tb.WriteString(\"\\n\")\n\t}\n\n\tb.WriteString(indentf(1, \"connection-mesh {\\n\"))\n\tb.WriteString(indentf(2, \"hosts %s;\\n\", strings.Join(hosts, \" \")))\n\tb.WriteString(indentf(1, \"}\\n\"))\n\n\tb.WriteString(\"}\") // end resource section\n\n\treturn ioutil.WriteFile(filename, b.Bytes(), 0644)\n}", "func (c *Provider) EndpointConfig() fab.EndpointConfig {\n\treturn c.endpointConfig\n}", "func WriteThrapConfig(conf *ThrapConfig, filename string) error {\n\tb, err := hclencoder.Encode(conf)\n\tif err == nil {\n\t\terr = ioutil.WriteFile(filename, b, 0644)\n\t}\n\treturn err\n}", "func WithEndpoint(endpoint string) CollectorEndpointOption {\n\treturn collectorEndpointOptionFunc(func(o *collectorEndpointConfig) {\n\t\to.endpoint = endpoint\n\t})\n}", "func (fes *FrontEndService) WriteConfigBGP(conf *string) {\n\tfor _, gw := range fes.gateways {\n\t\tif isIPv6(gw) || isIPv4(gw) {\n\t\t\tipv := \"\"\n\t\t\tif isIPv4(gw) {\n\t\t\t\tipv += \"\\tipv4 {\\n\"\n\t\t\t\tif len(fes.vrrps) > 0 {\n\t\t\t\t\tipv += \"\\t\\timport none;\\n\"\n\t\t\t\t} else {\n\t\t\t\t\tipv += \"\\t\\timport filter default_v4;\\n\"\n\t\t\t\t}\n\t\t\t\tipv += \"\\t\\texport filter cluster_e_static;\\n\"\n\t\t\t\tipv += \"\\t};\\n\"\n\t\t\t} else if isIPv6(gw) {\n\t\t\t\tipv = \"\\tipv6 {\\n\"\n\t\t\t\tif len(fes.vrrps) > 0 {\n\t\t\t\t\tipv += \"\\t\\timport none;\\n\"\n\t\t\t\t} else {\n\t\t\t\t\tipv += \"\\t\\timport filter default_v6;\\n\"\n\t\t\t\t}\n\t\t\t\tipv += \"\\t\\texport filter cluster_e_static;\\n\"\n\t\t\t\tipv += \"\\t};\\n\"\n\t\t\t}\n\t\t\t*conf += \"protocol bgp from LINK {\\n\"\n\t\t\t*conf += \"\\tinterface \\\"\" + fes.extInterface + \"\\\";\\n\"\n\t\t\t*conf += \"\\tlocal port \" + fes.localPortBGP + \" as \" + fes.localAS + \";\\n\"\n\t\t\t*conf += \"\\tneighbor \" + strings.Split(gw, \"/\")[0] + \" port \" + fes.remotePortBGP + \" as \" + fes.remoteAS + \";\\n\"\n\t\t\t*conf += ipv\n\t\t\t*conf += \"}\\n\"\n\t\t\t*conf += \"\\n\"\n\t\t}\n\t}\n}", "func WriteConfig(i interface{}, d string) error {\n\t// Convert to yaml\n\tym, err := yaml.Marshal(i)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Write to file\n\treturn ioutil.WriteFile(d, ym, 0755)\n}", "func (f *FileConfigWriter) Write(config *auth.Config) error {\n\tif f.filename == \"\" {\n\t\treturn errors.New(\"No config file name defined\")\n\t}\n\tcontent, err := yaml.Marshal(config)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"marshaling the config to yaml\")\n\t}\n\terr = ioutil.WriteFile(f.filename, content, util.DefaultWritePermissions)\n\treturn nil\n}", "func (s *EndpointConfigSummary) SetEndpointConfigName(v string) *EndpointConfigSummary {\n\ts.EndpointConfigName = &v\n\treturn s\n}", "func (s *Endpoint) SetEndpointConfigName(v string) *Endpoint {\n\ts.EndpointConfigName = &v\n\treturn s\n}", "func (s *UpdateEndpointInput) SetEndpointConfigName(v string) *UpdateEndpointInput {\n\ts.EndpointConfigName = &v\n\treturn s\n}", "func (fes *FrontEndService) WriteConfigKernel(conf *string, hasVIP4 bool, hasVIP6 bool) {\n\teFilter := \"none\"\n\tif hasVIP4 {\n\t\teFilter = \"filter default_v4\"\n\t}\n\n\t*conf += \"protocol kernel {\\n\"\n\t*conf += \"\\tipv4 {\\n\"\n\t*conf += \"\\t\\timport none;\\n\"\n\t*conf += \"\\t\\texport \" + eFilter + \";\\n\"\n\t*conf += \"\\t};\\n\"\n\t*conf += \"\\tkernel table \" + strconv.FormatInt(int64(fes.kernelTableId), 10) + \";\\n\"\n\tif fes.ecmp {\n\t\t*conf += \"\\tmerge paths on;\\n\"\n\t}\n\tif fes.dropIfNoPeer {\n\t\t// Setting the metric for the default blackhole route must be supported,\n\t\t// which requires the kernel proto's metric to be set to zero.\n\t\t//\n\t\t// \"Metric 0 has a special meaning of undefined metric, in which either OS default is used,\n\t\t// or per-route metric can be set using krt_metric attribute. Default: 32. \"\n\t\t*conf += \"\\tmetric 0;\\n\"\n\t}\n\t*conf += \"}\\n\"\n\t*conf += \"\\n\"\n\n\tif hasVIP6 {\n\t\teFilter = \"filter default_v6\"\n\t} else {\n\t\teFilter = \"none\"\n\t}\n\t*conf += \"protocol kernel {\\n\"\n\t*conf += \"\\tipv6 {\\n\"\n\t*conf += \"\\t\\timport none;\\n\"\n\t*conf += \"\\t\\texport \" + eFilter + \";\\n\"\n\t*conf += \"\\t};\\n\"\n\t*conf += \"\\tkernel table \" + strconv.FormatInt(int64(fes.kernelTableId), 10) + \";\\n\"\n\tif fes.ecmp {\n\t\t*conf += \"\\tmerge paths on;\\n\"\n\t}\n\tif fes.dropIfNoPeer {\n\t\t// Setting the metric for the default blackhole route must be supported,\n\t\t// which requires the kernel proto's metric to be set to zero.\n\t\t//\n\t\t// \"Metric 0 has a special meaning of undefined metric, in which either OS default is used,\n\t\t// or per-route metric can be set using krt_metric attribute. Default: 32. \"\n\t\t*conf += \"\\tmetric 0;\\n\"\n\t}\n\t*conf += \"}\\n\"\n\t*conf += \"\\n\"\n}", "func WithEndpoint(endpoint string) {\n\tcfg.endpoint = strings.TrimRight(endpoint, \"/\")\n}", "func (parser *Parser) WriteConfig() error {\n\tconfig := structs.Map(parser.Config)\n\tfor _, site := range parser.Sites {\n\t\tif site.Path == \".\" {\n\t\t\tfor key, val := range site.AllParameters {\n\t\t\t\tif _, ok := config[key]; !ok {\n\t\t\t\t\tconfig[key] = val\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tbs, err := yaml.Marshal(config)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn parser.Storage.Write(parser.Storage.GetSourceKey(KeyConfig), bs)\n}", "func (s *EndpointMetadata) SetEndpointConfigName(v string) *EndpointMetadata {\n\ts.EndpointConfigName = &v\n\treturn s\n}", "func MakeUpdateConfigEndpoint(svc service.Service) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (response interface{}, err error) {\n\t\treq := request.(UpdateConfigRequest)\n\t\terr = svc.UpdateConfig(ctx, req.SiteID, req.Config)\n\t\treturn UpdateConfigResponse{Err: err}, nil\n\t}\n}", "func (endpointCluster *EndpointCluster) SetEndpointsConfig(endpointInfos []EndpointInfo) error {\n\tif len(endpointInfos) == 0 {\n\t\treturn nil\n\t}\n\tif endpointCluster.Config == nil {\n\t\tendpointCluster.Config = &EndpointConfig{}\n\t}\n\t// Set timeout\n\tif endpointCluster.Config.TimeoutInMillis == 0 {\n\t\ttimeout := endpointInfos[0].Config.ActionDuration\n\t\tif timeout != \"\" {\n\t\t\trouteTimeout, err := strconv.ParseInt(timeout, 10, 32)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tendpointCluster.Config.TimeoutInMillis = uint32(routeTimeout)\n\t\t}\n\t}\n\n\t// retry\n\tif endpointCluster.Config.RetryConfig == nil {\n\t\tretryCount := endpointInfos[0].Config.RetryTimeOut\n\t\tif retryCount != \"\" {\n\t\t\tcount, err := strconv.ParseInt(retryCount, 10, 32)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tconf, _ := config.ReadConfigs()\n\t\t\tretryConfig := &RetryConfig{\n\t\t\t\tCount: int32(count),\n\t\t\t\tStatusCodes: conf.Envoy.Upstream.Retry.StatusCodes,\n\t\t\t}\n\t\t\tendpointCluster.Config.RetryConfig = retryConfig\n\t\t}\n\t}\n\treturn nil\n}", "func (s *DeleteEndpointConfigInput) SetEndpointConfigName(v string) *DeleteEndpointConfigInput {\n\ts.EndpointConfigName = &v\n\treturn s\n}", "func (s *CreateEndpointConfigInput) SetEndpointConfigName(v string) *CreateEndpointConfigInput {\n\ts.EndpointConfigName = &v\n\treturn s\n}", "func WriteConnectionConfig(context *clusterd.Context, clusterInfo *cephclient.ClusterInfo) error {\n\t// write the latest config to the config dir\n\tif _, err := cephclient.GenerateConnectionConfig(context, clusterInfo); err != nil {\n\t\treturn errors.Wrap(err, \"failed to write connection config\")\n\t}\n\n\treturn nil\n}", "func (s *DescribeEndpointConfigInput) SetEndpointConfigName(v string) *DescribeEndpointConfigInput {\n\ts.EndpointConfigName = &v\n\treturn s\n}", "func (s *CreateEndpointInput) SetEndpointConfigName(v string) *CreateEndpointInput {\n\ts.EndpointConfigName = &v\n\treturn s\n}", "func (s *DescribeEndpointConfigOutput) SetEndpointConfigName(v string) *DescribeEndpointConfigOutput {\n\ts.EndpointConfigName = &v\n\treturn s\n}", "func WithEndpointConfig(endpointConfig fab.EndpointConfig) SDKContextParams {\n\treturn func(ctx *Provider) {\n\t\tctx.endpointConfig = endpointConfig\n\t}\n}", "func writeConfig(f Function) (err error) {\n\tpath := filepath.Join(f.Root, ConfigFile)\n\tc := toConfig(f)\n\tvar bb []byte\n\tif bb, err = yaml.Marshal(&c); err != nil {\n\t\treturn\n\t}\n\treturn ioutil.WriteFile(path, bb, 0644)\n}", "func (c *Config) WritePayload(w io.Writer, payload interface{}) error {\n\tjsonConf, err := json.Marshal(payload)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to marshal payload: %s\", err)\n\t}\n\n\tc.config.jsonConfSize = C.uint(len(jsonConf))\n\tc.config.nsPathSize = C.uint(len(c.nsPath))\n\tcconfPayload := C.GoBytes(unsafe.Pointer(c.config), C.sizeof_struct_cConfig)\n\tcconfPayload = append(cconfPayload, c.nsPath...)\n\tcconfPayload = append(cconfPayload, jsonConf...)\n\n\tif n, err := w.Write(cconfPayload); err != nil || n != len(cconfPayload) {\n\t\treturn fmt.Errorf(\"failed to write payload: %s\", err)\n\t}\n\treturn nil\n}", "func (s *DescribeEndpointOutput) SetEndpointConfigName(v string) *DescribeEndpointOutput {\n\ts.EndpointConfigName = &v\n\treturn s\n}", "func (fes *FrontEndService) WriteConfigDropIfNoPeer(conf *string, hasVIP4 bool, hasVIP6 bool) {\n\tif hasVIP4 {\n\t\t*conf += \"protocol static BH4 {\\n\"\n\t\t*conf += \"\\tipv4 { preference 0; };\\n\"\n\t\t*conf += \"\\troute 0.0.0.0/0 blackhole {\\n\"\n\t\t*conf += \"\\t\\tkrt_metric=4294967295;\\n\"\n\t\t*conf += \"\\t};\\n\"\n\t\t*conf += \"}\\n\"\n\t\t*conf += \"\\n\"\n\t}\n\n\tif hasVIP6 {\n\t\t*conf += \"protocol static BH6 {\\n\"\n\t\t*conf += \"\\tipv6 { preference 0; };\\n\"\n\t\t*conf += \"\\troute 0::/0 blackhole {\\n\"\n\t\t*conf += \"\\t\\tkrt_metric=4294967295;\\n\"\n\t\t*conf += \"\\t};\\n\"\n\t\t*conf += \"}\\n\"\n\t\t*conf += \"\\n\"\n\t}\n\treturn\n}", "func ConfigWrite() error {\n\n\t// Marshal it\n\tconfigJSON, _ := json.MarshalIndent(Config, \"\", \" \")\n\n\t// Write the file\n\tfd, err := os.OpenFile(configSettingsPath(), os.O_WRONLY|os.O_TRUNC|os.O_CREATE, 0666)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfd.Write(configJSON)\n\tfd.Close()\n\n\t// Done\n\treturn err\n\n}", "func (mr *MockProvidersMockRecorder) EndpointConfig() *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"EndpointConfig\", reflect.TypeOf((*MockProviders)(nil).EndpointConfig))\n}", "func (c *Config) Write() error {\n\trawConfig, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = os.MkdirAll(configDirPath, 0755)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn ioutil.WriteFile(configPath, rawConfig, 0644)\n}", "func (self *PolicyAgent) AddEndpoint(endpoint *OfnetEndpoint) error {\n\n\tself.mutex.RLock()\n\tif self.dstGrpFlow[endpoint.EndpointID] != nil {\n\t\t// FIXME: handle this as Update\n\t\tlog.Warnf(\"DstGroup for endpoint %+v already exists\", endpoint)\n\t\tself.mutex.RUnlock()\n\t\treturn nil\n\t}\n\tself.mutex.RUnlock()\n\n\tlog.Infof(\"Adding dst group entry for endpoint: %+v\", endpoint)\n\n\tvrf := self.agent.getvlanVrf(endpoint.Vlan)\n\n\tif vrf == nil {\n\t\tlog.Errorf(\"Error finding vrf for vlan %s\", endpoint.Vlan)\n\t\treturn errors.New(\"Error finding vrf for vlan\")\n\t}\n\tlog.Infof(\"Recevied add endpoint for vrf %v\", *vrf)\n\n\tself.agent.vrfMutex.RLock()\n\tvrfid := self.agent.vrfNameIdMap[*vrf]\n\tself.agent.vrfMutex.RUnlock()\n\tvrfMetadata, vrfMetadataMask := Vrfmetadata(*vrfid)\n\t// Install the Dst group lookup flow\n\tdstGrpFlow, err := self.dstGrpTable.NewFlow(ofctrl.FlowMatch{\n\t\tPriority: FLOW_MATCH_PRIORITY,\n\t\tEthertype: 0x0800,\n\t\tIpDa: &endpoint.IpAddr,\n\t\tMetadata: &vrfMetadata,\n\t\tMetadataMask: &vrfMetadataMask,\n\t})\n\tif err != nil {\n\t\tlog.Errorf(\"Error adding dstGroup flow for %v. Err: %v\", endpoint.IpAddr, err)\n\t\treturn err\n\t}\n\n\t// Format the metadata\n\tmetadata, metadataMask := DstGroupMetadata(endpoint.EndpointGroup)\n\n\t// Set dst GroupId\n\terr = dstGrpFlow.SetMetadata(metadata, metadataMask)\n\tif err != nil {\n\t\tlog.Errorf(\"Error setting metadata %v for flow {%+v}. Err: %v\", metadata, dstGrpFlow, err)\n\t\treturn err\n\t}\n\t// Go to policy Table\n\terr = dstGrpFlow.Next(self.policyTable)\n\tif err != nil {\n\t\tlog.Errorf(\"Error installing flow {%+v}. Err: %v\", dstGrpFlow, err)\n\t\treturn err\n\t}\n\n\t// save the Flow\n\tself.mutex.Lock()\n\tself.dstGrpFlow[endpoint.EndpointID] = dstGrpFlow\n\tself.mutex.Unlock()\n\treturn nil\n}", "func (ec *ExecutionContext) WriteConfig(config *Config) error {\n\tvar op errors.Op = \"cli.ExecutionContext.WriteConfig\"\n\tvar cfg *Config\n\tif config != nil {\n\t\tcfg = config\n\t} else {\n\t\tcfg = ec.Config\n\t}\n\tbuf := new(bytes.Buffer)\n\tencoder := yaml.NewEncoder(buf)\n\tencoder.SetIndent(2)\n\terr := encoder.Encode(cfg)\n\tif err != nil {\n\t\treturn errors.E(op, err)\n\t}\n\terr = ioutil.WriteFile(ec.ConfigFile, buf.Bytes(), 0644)\n\tif err != nil {\n\t\treturn errors.E(op, err)\n\t}\n\treturn nil\n}", "func Write(w io.Writer, cfg *Config) error {\n\tif cfg.filename != \"\" {\n\t\tfmt.Fprintf(w, \"# Configuration file location: %v\\n\\n\", cfg.filename)\n\t} else {\n\t\tfmt.Fprintf(w, \"# Cound not find configuration file location.\\n\\n\")\n\t}\n\treturn toml.NewEncoder(w).Encode(cfg.File)\n}", "func (h *HeaderfileWriter) WriteNetdevConfig(w io.Writer, cfg datapath.DeviceConfiguration) error {\n\tfw := bufio.NewWriter(w)\n\th.writeNetdevConfig(fw, cfg)\n\treturn fw.Flush()\n}", "func (options *UpdateNotificationChannelOptions) SetEndpoint(endpoint string) *UpdateNotificationChannelOptions {\n\toptions.Endpoint = core.StringPtr(endpoint)\n\treturn options\n}", "func (d *Driver) CreateEndpoint(r *sdk.CreateEndpointRequest) (*sdk.CreateEndpointResponse, error) {\n\tendID := r.EndpointID\n\tnetID := r.NetworkID\n\teInfo := r.Interface\n\tlog.Debugf(\"CreateEndpoint called :%v\", r)\n\t// Get the network handler and make sure it exists\n\td.Lock()\n\tnetwork, ok := d.networks[r.NetworkID]\n\td.Unlock()\n\n\tif !ok {\n\t\treturn nil, types.NotFoundErrorf(\"network %s does not exist\", netID)\n\t}\n\n\t// Try to convert the options to endpoint configuration\n\tepConfig, err := parseEndpointOptions(r.Options)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Create and add the endpoint\n\tnetwork.Lock()\n\tendpoint := &bridgeEndpoint{id: endID, nid: netID, config: epConfig}\n\tnetwork.endpoints[endID] = endpoint\n\tnetwork.Unlock()\n\n\t// On failure make sure to remove the endpoint\n\tdefer func() {\n\t\tif err != nil {\n\t\t\tnetwork.Lock()\n\t\t\tdelete(network.endpoints, endID)\n\t\t\tnetwork.Unlock()\n\t\t}\n\t}()\n\n\t// Generate a name for what will be the host side pipe interface\n\thostIfName, err := netutils.GenerateIfaceName(d.nlh, vethPrefix, vethLen)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Generate a name for what will be the sandbox side pipe interface\n\tcontainerIfName := network.config.ContainerIfName\n\n\t// Generate and add the interface pipe host <-> sandbox\n\tveth := &netlink.Veth{\n\t\tLinkAttrs: netlink.LinkAttrs{Name: hostIfName, TxQLen: 0},\n\t\tPeerName: containerIfName}\n\tif err = d.nlh.LinkAdd(veth); err != nil {\n\t\treturn nil, types.InternalErrorf(\"failed to add the host (%s) <=> sandbox (%s) pair interfaces: %v\", hostIfName, containerIfName, err)\n\t}\n\n\t// Get the host side pipe interface handler\n\thost, err := d.nlh.LinkByName(hostIfName)\n\tif err != nil {\n\t\treturn nil, types.InternalErrorf(\"failed to find host side interface %s: %v\", hostIfName, err)\n\t}\n\tdefer func() {\n\t\tif err != nil {\n\t\t\td.nlh.LinkDel(host)\n\t\t}\n\t}()\n\n\t// Get the sandbox side pipe interface handler\n\tsbox, err := d.nlh.LinkByName(containerIfName)\n\tif err != nil {\n\t\treturn nil, types.InternalErrorf(\"failed to find sandbox side interface %s: %v\", containerIfName, err)\n\t}\n\tdefer func() {\n\t\tif err != nil {\n\t\t\td.nlh.LinkDel(sbox)\n\t\t}\n\t}()\n\n\tnetwork.Lock()\n\tconfig := network.config\n\tnetwork.Unlock()\n\n\t// Add bridge inherited attributes to pipe interfaces\n\tif config.Mtu != 0 {\n\t\terr = d.nlh.LinkSetMTU(host, config.Mtu)\n\t\tif err != nil {\n\t\t\treturn nil, types.InternalErrorf(\"failed to set MTU on host interface %s: %v\", hostIfName, err)\n\t\t}\n\t\terr = d.nlh.LinkSetMTU(sbox, config.Mtu)\n\t\tif err != nil {\n\t\t\treturn nil, types.InternalErrorf(\"failed to set MTU on sandbox interface %s: %v\", containerIfName, err)\n\t\t}\n\t}\n\n\t// Attach host side pipe interface into the bridge\n\tif err = addToBridge(d.nlh, hostIfName, config.BridgeName); err != nil {\n\t\treturn nil, fmt.Errorf(\"adding interface %s to bridge %s failed: %v\", hostIfName, config.BridgeName, err)\n\t}\n\n\t// Store the sandbox side pipe interface parameters\n\tendpoint.srcName = containerIfName\n\tendpoint.macAddress = eInfo.MacAddress\n\tendpoint.addr = eInfo.Address\n\tendpoint.addrv6 = eInfo.AddressIPv6\n\n\t// Up the host interface after finishing all netlink configuration\n\tif err = d.nlh.LinkSetUp(host); err != nil {\n\t\treturn nil, fmt.Errorf(\"could not set link up for host interface %s: %v\", hostIfName, err)\n\t}\n\n\tres := &sdk.CreateEndpointResponse{\n\t\tInterface: &sdk.EndpointInterface{\n\t\t\tAddress: endpoint.addr,\n\t\t\tMacAddress: endpoint.macAddress,\n\t\t},\n\t}\n\n\tlog.Debugf(\"Create endpoint response: %+v\", res)\n\tlog.Debugf(\"Create endpoint %s %+v\", endID, res)\n\treturn res, nil\n}", "func (s *MonitoringJobDefinitionSummary) SetEndpointName(v string) *MonitoringJobDefinitionSummary {\n\ts.EndpointName = &v\n\treturn s\n}", "func (options *CreateNotificationChannelOptions) SetEndpoint(endpoint string) *CreateNotificationChannelOptions {\n\toptions.Endpoint = core.StringPtr(endpoint)\n\treturn options\n}", "func NewEndpointConfig(ctx *pulumi.Context,\n\tname string, args *EndpointConfigArgs, opts ...pulumi.ResourceOption) (*EndpointConfig, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.ProductionVariants == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'ProductionVariants'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource EndpointConfig\n\terr := ctx.RegisterResource(\"aws-native:sagemaker:EndpointConfig\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (c *Config) Write(filename string) (err error) {\n\tb, err := json.Marshal(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfile, err := os.Create(filename)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif _, err := file.Write(b); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (mr *MockClientMockRecorder) EndpointConfig() *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"EndpointConfig\", reflect.TypeOf((*MockClient)(nil).EndpointConfig))\n}", "func (s *MonitoringScheduleSummary) SetEndpointName(v string) *MonitoringScheduleSummary {\n\ts.EndpointName = &v\n\treturn s\n}", "func (s *MonitoringExecutionSummary) SetEndpointName(v string) *MonitoringExecutionSummary {\n\ts.EndpointName = &v\n\treturn s\n}", "func WriteCredsConfig(cc *CredsConfig, fpath string) error {\n\tb, err := hclencoder.Encode(cc)\n\tif err == nil {\n\t\terr = ioutil.WriteFile(fpath, b, 0644)\n\t}\n\treturn err\n}", "func (in *DomainName_EndpointConfiguration) DeepCopy() *DomainName_EndpointConfiguration {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(DomainName_EndpointConfiguration)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (s *ListModelBiasJobDefinitionsInput) SetEndpointName(v string) *ListModelBiasJobDefinitionsInput {\n\ts.EndpointName = &v\n\treturn s\n}", "func (b *NutanixPrismElementEndpointApplyConfiguration) WithEndpoint(value *NutanixPrismEndpointApplyConfiguration) *NutanixPrismElementEndpointApplyConfiguration {\n\tb.Endpoint = value\n\treturn b\n}", "func (s DescribeEndpointConfigOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s *GetScalingConfigurationRecommendationOutput) SetEndpointName(v string) *GetScalingConfigurationRecommendationOutput {\n\ts.EndpointName = &v\n\treturn s\n}", "func (s *DescribeMonitoringScheduleOutput) SetEndpointName(v string) *DescribeMonitoringScheduleOutput {\n\ts.EndpointName = &v\n\treturn s\n}", "func (s *CreateInferenceExperimentInput) SetEndpointName(v string) *CreateInferenceExperimentInput {\n\ts.EndpointName = &v\n\treturn s\n}", "func (o GetVpcEndpointsEndpointOutput) EndpointId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetVpcEndpointsEndpoint) string { return v.EndpointId }).(pulumi.StringOutput)\n}", "func (p *ProviderConfigs) writeProviderConfig() error {\n\t// Get file.\n\tconfig := getProviderConfigFile()\n\n\t// Convert object to json.\n\tdata, err := json.MarshalIndent(p, \"\", \" \")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\t// Write json output to file.\n\terr = ioutil.WriteFile(config, data, 0644)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn nil\n}", "func (s *ListMonitoringExecutionsInput) SetEndpointName(v string) *ListMonitoringExecutionsInput {\n\ts.EndpointName = &v\n\treturn s\n}", "func (s *GetScalingConfigurationRecommendationInput) SetEndpointName(v string) *GetScalingConfigurationRecommendationInput {\n\ts.EndpointName = &v\n\treturn s\n}", "func (s *ModelDashboardMonitoringSchedule) SetEndpointName(v string) *ModelDashboardMonitoringSchedule {\n\ts.EndpointName = &v\n\treturn s\n}", "func (app *App) registerEndpoint() {\n\tendpoint, err := vulcan.NewEndpoint(app.Config.Name, app.Config.ListenIP, app.Config.ListenPort)\n\tif err != nil {\n\t\tlog.Errorf(\"Failed to create an endpoint: %v\", err)\n\t\treturn\n\t}\n\n\tif err := app.registry.RegisterEndpoint(endpoint); err != nil {\n\t\tlog.Errorf(\"Failed to register an endpoint: %v %v\", endpoint, err)\n\t\treturn\n\t}\n\n\tlog.Infof(\"Registered: %v\", endpoint)\n}", "func (r *templateRouter) writeConfig() error {\n\t//write out any certificate files that don't exist\n\tfor k, cfg := range r.state {\n\t\tcfg := cfg // avoid implicit memory aliasing (gosec G601)\n\t\tif err := r.writeCertificates(&cfg); err != nil {\n\t\t\treturn fmt.Errorf(\"error writing certificates for %s: %v\", k, err)\n\t\t}\n\n\t\t// calculate the server weight for the endpoints in each service\n\t\t// called here to make sure we have the actual number of endpoints.\n\t\tcfg.ServiceUnitNames = r.calculateServiceWeights(cfg.ServiceUnits, cfg.PreferPort)\n\n\t\t// Calculate the number of active endpoints for the route.\n\t\tcfg.ActiveEndpoints = r.getActiveEndpoints(cfg.ServiceUnits, cfg.PreferPort)\n\n\t\tcfg.Status = ServiceAliasConfigStatusSaved\n\t\tr.state[k] = cfg\n\t}\n\n\tlog.V(4).Info(\"committing router certificate manager changes...\")\n\tif err := r.certManager.Commit(); err != nil {\n\t\treturn fmt.Errorf(\"error committing certificate changes: %v\", err)\n\t}\n\n\tlog.V(4).Info(\"router certificate manager config committed\")\n\n\tdisableHTTP2, _ := strconv.ParseBool(os.Getenv(\"ROUTER_DISABLE_HTTP2\"))\n\n\tfor name, template := range r.templates {\n\t\tfilename := filepath.Join(r.dir, name)\n\t\tif err := os.MkdirAll(filepath.Dir(filename), 0777); err != nil {\n\t\t\treturn fmt.Errorf(\"error creating path %q: %v\", filepath.Dir(filename), err)\n\t\t}\n\t\tfile, err := os.Create(filename)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"error creating config file %s: %v\", filename, err)\n\t\t}\n\n\t\tdata := templateData{\n\t\t\tWorkingDir: r.dir,\n\t\t\tState: r.state,\n\t\t\tServiceUnits: r.serviceUnits,\n\t\t\tDefaultCertificate: r.defaultCertificatePath,\n\t\t\tDefaultDestinationCA: r.defaultDestinationCAPath,\n\t\t\tStatsUser: r.statsUser,\n\t\t\tStatsPassword: r.statsPassword,\n\t\t\tStatsPort: r.statsPort,\n\t\t\tBindPorts: !r.bindPortsAfterSync || r.synced,\n\t\t\tDynamicConfigManager: r.dynamicConfigManager,\n\t\t\tDisableHTTP2: disableHTTP2,\n\t\t\tCaptureHTTPRequestHeaders: r.captureHTTPRequestHeaders,\n\t\t\tCaptureHTTPResponseHeaders: r.captureHTTPResponseHeaders,\n\t\t\tCaptureHTTPCookie: r.captureHTTPCookie,\n\t\t\tHTTPHeaderNameCaseAdjustments: r.httpHeaderNameCaseAdjustments,\n\t\t\tHaveClientCA: r.haveClientCA,\n\t\t\tHaveCRLs: r.haveCRLs,\n\t\t\tHTTPResponseHeaders: r.httpResponseHeaders,\n\t\t\tHTTPRequestHeaders: r.httpRequestHeaders,\n\t\t}\n\t\tif err := template.Execute(file, data); err != nil {\n\t\t\tfile.Close()\n\t\t\treturn fmt.Errorf(\"error executing template for file %s: %v\", filename, err)\n\t\t}\n\t\tfile.Close()\n\t}\n\n\treturn nil\n}", "func (x *XVerify) SetEndpoint(e string) {\n\tendpoint = e\n}", "func WriteConfig(cfg *config.Dendrite, configDir string) error {\n\tdata, err := yaml.Marshal(cfg)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn ioutil.WriteFile(filepath.Join(configDir, ConfigFile), data, 0666)\n}", "func (s *MonitoringSchedule) SetEndpointName(v string) *MonitoringSchedule {\n\ts.EndpointName = &v\n\treturn s\n}", "func EncodeConfig(configFile io.Writer, c *Configuration) error {\n\t// Lock mutex\n\tConfig.rw.Lock()\n\tdefer Config.rw.Unlock()\n\n\t// Encode the given writer with the given interface\n\treturn toml.NewEncoder(configFile).Encode(c)\n}", "func writeConfig() error {\n\tconfigFile := filepath.Join(config.ourBinaryDir, config.ourConfigFilename)\n\tlog.Printf(\"Writing YAML file: %s\", configFile)\n\tyamlText, err := yaml.Marshal(&config)\n\tif err != nil {\n\t\tlog.Printf(\"Couldn't generate YAML file: %s\", err)\n\t\treturn err\n\t}\n\terr = writeFileSafe(configFile, yamlText)\n\tif err != nil {\n\t\tlog.Printf(\"Couldn't save YAML config: %s\", err)\n\t\treturn err\n\t}\n\n\tuserFilter := getUserFilter()\n\terr = userFilter.save()\n\tif err != nil {\n\t\tlog.Printf(\"Couldn't save the user filter: %s\", err)\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (p *Profile) WriteConfigField(field, value string) error {\n\tviper.Set(p.GetConfigField(field), value)\n\treturn viper.WriteConfig()\n}", "func (s *ListMonitoringSchedulesInput) SetEndpointName(v string) *ListMonitoringSchedulesInput {\n\ts.EndpointName = &v\n\treturn s\n}", "func (h *HeaderfileWriter) WriteNodeConfig(w io.Writer, cfg *datapath.LocalNodeConfiguration) error {\n\textraMacrosMap := make(dpdef.Map)\n\tcDefinesMap := make(dpdef.Map)\n\n\tfw := bufio.NewWriter(w)\n\n\twriteIncludes(w)\n\n\trouterIP := node.GetIPv6Router()\n\thostIP := node.GetIPv6()\n\n\tfmt.Fprintf(fw, \"/*\\n\")\n\tif option.Config.EnableIPv6 {\n\t\tfmt.Fprintf(fw, \" cilium.v6.external.str %s\\n\", node.GetIPv6().String())\n\t\tfmt.Fprintf(fw, \" cilium.v6.internal.str %s\\n\", node.GetIPv6Router().String())\n\t\tfmt.Fprintf(fw, \" cilium.v6.nodeport.str %s\\n\", node.GetNodePortIPv6Addrs())\n\t\tfmt.Fprintf(fw, \"\\n\")\n\t}\n\tfmt.Fprintf(fw, \" cilium.v4.external.str %s\\n\", node.GetIPv4().String())\n\tfmt.Fprintf(fw, \" cilium.v4.internal.str %s\\n\", node.GetInternalIPv4Router().String())\n\tfmt.Fprintf(fw, \" cilium.v4.nodeport.str %s\\n\", node.GetNodePortIPv4Addrs())\n\tfmt.Fprintf(fw, \"\\n\")\n\tif option.Config.EnableIPv6 {\n\t\tfw.WriteString(dumpRaw(defaults.RestoreV6Addr, node.GetIPv6Router()))\n\t}\n\tfw.WriteString(dumpRaw(defaults.RestoreV4Addr, node.GetInternalIPv4Router()))\n\tfmt.Fprintf(fw, \" */\\n\\n\")\n\n\tcDefinesMap[\"KERNEL_HZ\"] = fmt.Sprintf(\"%d\", option.Config.KernelHz)\n\n\tif option.Config.EnableIPv6 {\n\t\textraMacrosMap[\"ROUTER_IP\"] = routerIP.String()\n\t\tfw.WriteString(defineIPv6(\"ROUTER_IP\", routerIP))\n\t}\n\n\tif option.Config.EnableIPv4 {\n\t\tipv4GW := node.GetInternalIPv4Router()\n\t\tloopbackIPv4 := node.GetIPv4Loopback()\n\t\tipv4Range := node.GetIPv4AllocRange()\n\t\tcDefinesMap[\"IPV4_GATEWAY\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(ipv4GW))\n\t\tcDefinesMap[\"IPV4_LOOPBACK\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(loopbackIPv4))\n\t\tcDefinesMap[\"IPV4_MASK\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(net.IP(ipv4Range.Mask)))\n\n\t\tif option.Config.EnableIPv4FragmentsTracking {\n\t\t\tcDefinesMap[\"ENABLE_IPV4_FRAGMENTS\"] = \"1\"\n\t\t\tcDefinesMap[\"IPV4_FRAG_DATAGRAMS_MAP\"] = fragmap.MapName\n\t\t\tcDefinesMap[\"CILIUM_IPV4_FRAG_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", option.Config.FragmentsMapEntries)\n\t\t}\n\t}\n\n\tif option.Config.EnableIPv6 {\n\t\textraMacrosMap[\"HOST_IP\"] = hostIP.String()\n\t\tfw.WriteString(defineIPv6(\"HOST_IP\", hostIP))\n\t}\n\n\tfor t, id := range tunnelProtocols {\n\t\tmacroName := fmt.Sprintf(\"TUNNEL_PROTOCOL_%s\", strings.ToUpper(t))\n\t\tcDefinesMap[macroName] = fmt.Sprintf(\"%d\", id)\n\t}\n\n\tencapProto := option.Config.TunnelProtocol\n\tif !option.Config.TunnelingEnabled() &&\n\t\toption.Config.EnableNodePort &&\n\t\toption.Config.NodePortMode != option.NodePortModeSNAT &&\n\t\toption.Config.LoadBalancerDSRDispatch == option.DSRDispatchGeneve {\n\t\tencapProto = option.TunnelGeneve\n\t}\n\n\tcDefinesMap[\"TUNNEL_PROTOCOL\"] = fmt.Sprintf(\"%d\", tunnelProtocols[encapProto])\n\tcDefinesMap[\"TUNNEL_PORT\"] = fmt.Sprintf(\"%d\", option.Config.TunnelPort)\n\n\tif tunnelDev, err := netlink.LinkByName(fmt.Sprintf(\"cilium_%s\", encapProto)); err == nil {\n\t\tcDefinesMap[\"ENCAP_IFINDEX\"] = fmt.Sprintf(\"%d\", tunnelDev.Attrs().Index)\n\t}\n\n\tcDefinesMap[\"HOST_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameHost))\n\tcDefinesMap[\"WORLD_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameWorld))\n\tif option.Config.IsDualStack() {\n\t\tcDefinesMap[\"WORLD_IPV4_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameWorldIPv4))\n\t\tcDefinesMap[\"WORLD_IPV6_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameWorldIPv6))\n\t} else {\n\t\tworldID := identity.GetReservedID(labels.IDNameWorld)\n\t\tcDefinesMap[\"WORLD_IPV4_ID\"] = fmt.Sprintf(\"%d\", worldID)\n\t\tcDefinesMap[\"WORLD_IPV6_ID\"] = fmt.Sprintf(\"%d\", worldID)\n\t}\n\tcDefinesMap[\"HEALTH_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameHealth))\n\tcDefinesMap[\"UNMANAGED_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameUnmanaged))\n\tcDefinesMap[\"INIT_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameInit))\n\tcDefinesMap[\"LOCAL_NODE_ID\"] = fmt.Sprintf(\"%d\", identity.GetLocalNodeID())\n\tcDefinesMap[\"REMOTE_NODE_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameRemoteNode))\n\tcDefinesMap[\"KUBE_APISERVER_NODE_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameKubeAPIServer))\n\tcDefinesMap[\"CILIUM_LB_SERVICE_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.ServiceMapMaxEntries)\n\tcDefinesMap[\"CILIUM_LB_BACKENDS_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.ServiceBackEndMapMaxEntries)\n\tcDefinesMap[\"CILIUM_LB_REV_NAT_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.RevNatMapMaxEntries)\n\tcDefinesMap[\"CILIUM_LB_AFFINITY_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.AffinityMapMaxEntries)\n\tcDefinesMap[\"CILIUM_LB_SOURCE_RANGE_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.SourceRangeMapMaxEntries)\n\tcDefinesMap[\"CILIUM_LB_MAGLEV_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.MaglevMapMaxEntries)\n\n\tcDefinesMap[\"TUNNEL_MAP\"] = tunnel.MapName\n\tcDefinesMap[\"TUNNEL_ENDPOINT_MAP_SIZE\"] = fmt.Sprintf(\"%d\", tunnel.MaxEntries)\n\tcDefinesMap[\"ENDPOINTS_MAP\"] = lxcmap.MapName\n\tcDefinesMap[\"ENDPOINTS_MAP_SIZE\"] = fmt.Sprintf(\"%d\", lxcmap.MaxEntries)\n\tcDefinesMap[\"METRICS_MAP\"] = metricsmap.MapName\n\tcDefinesMap[\"METRICS_MAP_SIZE\"] = fmt.Sprintf(\"%d\", metricsmap.MaxEntries)\n\tcDefinesMap[\"POLICY_MAP_SIZE\"] = fmt.Sprintf(\"%d\", policymap.MaxEntries)\n\tcDefinesMap[\"AUTH_MAP\"] = authmap.MapName\n\tcDefinesMap[\"AUTH_MAP_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.AuthMapEntries)\n\tcDefinesMap[\"CONFIG_MAP\"] = configmap.MapName\n\tcDefinesMap[\"CONFIG_MAP_SIZE\"] = fmt.Sprintf(\"%d\", configmap.MaxEntries)\n\tcDefinesMap[\"IPCACHE_MAP\"] = ipcachemap.Name\n\tcDefinesMap[\"IPCACHE_MAP_SIZE\"] = fmt.Sprintf(\"%d\", ipcachemap.MaxEntries)\n\tcDefinesMap[\"NODE_MAP\"] = nodemap.MapName\n\tcDefinesMap[\"NODE_MAP_SIZE\"] = fmt.Sprintf(\"%d\", nodemap.MaxEntries)\n\tcDefinesMap[\"SRV6_VRF_MAP4\"] = srv6map.VRFMapName4\n\tcDefinesMap[\"SRV6_VRF_MAP6\"] = srv6map.VRFMapName6\n\tcDefinesMap[\"SRV6_POLICY_MAP4\"] = srv6map.PolicyMapName4\n\tcDefinesMap[\"SRV6_POLICY_MAP6\"] = srv6map.PolicyMapName6\n\tcDefinesMap[\"SRV6_SID_MAP\"] = srv6map.SIDMapName\n\tcDefinesMap[\"SRV6_STATE_MAP4\"] = srv6map.StateMapName4\n\tcDefinesMap[\"SRV6_STATE_MAP6\"] = srv6map.StateMapName6\n\tcDefinesMap[\"SRV6_VRF_MAP_SIZE\"] = fmt.Sprintf(\"%d\", srv6map.MaxVRFEntries)\n\tcDefinesMap[\"SRV6_POLICY_MAP_SIZE\"] = fmt.Sprintf(\"%d\", srv6map.MaxPolicyEntries)\n\tcDefinesMap[\"SRV6_SID_MAP_SIZE\"] = fmt.Sprintf(\"%d\", srv6map.MaxSIDEntries)\n\tcDefinesMap[\"SRV6_STATE_MAP_SIZE\"] = fmt.Sprintf(\"%d\", srv6map.MaxStateEntries)\n\tcDefinesMap[\"WORLD_CIDRS4_MAP\"] = worldcidrsmap.MapName4\n\tcDefinesMap[\"WORLD_CIDRS4_MAP_SIZE\"] = fmt.Sprintf(\"%d\", worldcidrsmap.MapMaxEntries)\n\tcDefinesMap[\"POLICY_PROG_MAP_SIZE\"] = fmt.Sprintf(\"%d\", policymap.PolicyCallMaxEntries)\n\tcDefinesMap[\"L2_RESPONSER_MAP4_SIZE\"] = fmt.Sprintf(\"%d\", l2respondermap.DefaultMaxEntries)\n\tcDefinesMap[\"ENCRYPT_MAP\"] = encrypt.MapName\n\tcDefinesMap[\"L2_RESPONDER_MAP4\"] = l2respondermap.MapName\n\tcDefinesMap[\"CT_CONNECTION_LIFETIME_TCP\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutTCP.Seconds()))\n\tcDefinesMap[\"CT_CONNECTION_LIFETIME_NONTCP\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutAny.Seconds()))\n\tcDefinesMap[\"CT_SERVICE_LIFETIME_TCP\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutSVCTCP.Seconds()))\n\tcDefinesMap[\"CT_SERVICE_LIFETIME_NONTCP\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutSVCAny.Seconds()))\n\tcDefinesMap[\"CT_SERVICE_CLOSE_REBALANCE\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutSVCTCPGrace.Seconds()))\n\tcDefinesMap[\"CT_SYN_TIMEOUT\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutSYN.Seconds()))\n\tcDefinesMap[\"CT_CLOSE_TIMEOUT\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutFIN.Seconds()))\n\tcDefinesMap[\"CT_REPORT_INTERVAL\"] = fmt.Sprintf(\"%d\", int64(option.Config.MonitorAggregationInterval.Seconds()))\n\tcDefinesMap[\"CT_REPORT_FLAGS\"] = fmt.Sprintf(\"%#04x\", int64(option.Config.MonitorAggregationFlags))\n\tcDefinesMap[\"CT_TAIL_CALL_BUFFER4\"] = \"cilium_tail_call_buffer4\"\n\tcDefinesMap[\"CT_TAIL_CALL_BUFFER6\"] = \"cilium_tail_call_buffer6\"\n\tcDefinesMap[\"PER_CLUSTER_CT_TCP4\"] = \"cilium_per_cluster_ct_tcp4\"\n\tcDefinesMap[\"PER_CLUSTER_CT_TCP6\"] = \"cilium_per_cluster_ct_tcp6\"\n\tcDefinesMap[\"PER_CLUSTER_CT_ANY4\"] = \"cilium_per_cluster_ct_any4\"\n\tcDefinesMap[\"PER_CLUSTER_CT_ANY6\"] = \"cilium_per_cluster_ct_any6\"\n\tcDefinesMap[\"PER_CLUSTER_SNAT_MAPPING_IPV4\"] = \"cilium_per_cluster_snat_v4_external\"\n\tcDefinesMap[\"PER_CLUSTER_SNAT_MAPPING_IPV6\"] = \"cilium_per_cluster_snat_v6_external\"\n\n\tif option.Config.PreAllocateMaps {\n\t\tcDefinesMap[\"PREALLOCATE_MAPS\"] = \"1\"\n\t}\n\n\tcDefinesMap[\"EVENTS_MAP\"] = eventsmap.MapName\n\tcDefinesMap[\"SIGNAL_MAP\"] = signalmap.MapName\n\tcDefinesMap[\"POLICY_CALL_MAP\"] = policymap.PolicyCallMapName\n\tif option.Config.EnableEnvoyConfig {\n\t\tcDefinesMap[\"POLICY_EGRESSCALL_MAP\"] = policymap.PolicyEgressCallMapName\n\t}\n\tcDefinesMap[\"LB6_REVERSE_NAT_MAP\"] = \"cilium_lb6_reverse_nat\"\n\tcDefinesMap[\"LB6_SERVICES_MAP_V2\"] = \"cilium_lb6_services_v2\"\n\tcDefinesMap[\"LB6_BACKEND_MAP\"] = \"cilium_lb6_backends_v3\"\n\tcDefinesMap[\"LB6_REVERSE_NAT_SK_MAP\"] = lbmap.SockRevNat6MapName\n\tcDefinesMap[\"LB6_REVERSE_NAT_SK_MAP_SIZE\"] = fmt.Sprintf(\"%d\", lbmap.MaxSockRevNat6MapEntries)\n\tcDefinesMap[\"LB4_REVERSE_NAT_MAP\"] = \"cilium_lb4_reverse_nat\"\n\tcDefinesMap[\"LB4_SERVICES_MAP_V2\"] = \"cilium_lb4_services_v2\"\n\tcDefinesMap[\"LB4_BACKEND_MAP\"] = \"cilium_lb4_backends_v3\"\n\tcDefinesMap[\"LB4_REVERSE_NAT_SK_MAP\"] = lbmap.SockRevNat4MapName\n\tcDefinesMap[\"LB4_REVERSE_NAT_SK_MAP_SIZE\"] = fmt.Sprintf(\"%d\", lbmap.MaxSockRevNat4MapEntries)\n\n\tif option.Config.EnableSessionAffinity {\n\t\tcDefinesMap[\"ENABLE_SESSION_AFFINITY\"] = \"1\"\n\t\tcDefinesMap[\"LB_AFFINITY_MATCH_MAP\"] = lbmap.AffinityMatchMapName\n\t\tif option.Config.EnableIPv4 {\n\t\t\tcDefinesMap[\"LB4_AFFINITY_MAP\"] = lbmap.Affinity4MapName\n\t\t}\n\t\tif option.Config.EnableIPv6 {\n\t\t\tcDefinesMap[\"LB6_AFFINITY_MAP\"] = lbmap.Affinity6MapName\n\t\t}\n\t}\n\n\tcDefinesMap[\"TRACE_PAYLOAD_LEN\"] = fmt.Sprintf(\"%dULL\", option.Config.TracePayloadlen)\n\tcDefinesMap[\"MTU\"] = fmt.Sprintf(\"%d\", cfg.MtuConfig.GetDeviceMTU())\n\n\tif option.Config.EnableIPv4 {\n\t\tcDefinesMap[\"ENABLE_IPV4\"] = \"1\"\n\t}\n\n\tif option.Config.EnableIPv6 {\n\t\tcDefinesMap[\"ENABLE_IPV6\"] = \"1\"\n\t}\n\n\tif option.Config.EnableSRv6 {\n\t\tcDefinesMap[\"ENABLE_SRV6\"] = \"1\"\n\t\tif option.Config.SRv6EncapMode != \"reduced\" {\n\t\t\tcDefinesMap[\"ENABLE_SRV6_SRH_ENCAP\"] = \"1\"\n\t\t}\n\t}\n\n\tif option.Config.EnableSCTP {\n\t\tcDefinesMap[\"ENABLE_SCTP\"] = \"1\"\n\t}\n\n\tif option.Config.EnableIPSec {\n\t\tcDefinesMap[\"ENABLE_IPSEC\"] = \"1\"\n\t}\n\n\tif option.Config.EnableWireguard {\n\t\tcDefinesMap[\"ENABLE_WIREGUARD\"] = \"1\"\n\t\tifindex, err := link.GetIfIndex(wgtypes.IfaceName)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcDefinesMap[\"WG_IFINDEX\"] = fmt.Sprintf(\"%d\", ifindex)\n\n\t\tif option.Config.EncryptNode {\n\t\t\tcDefinesMap[\"ENABLE_NODE_ENCRYPTION\"] = \"1\"\n\t\t}\n\t}\n\n\tif option.Config.EnableL2Announcements {\n\t\tcDefinesMap[\"ENABLE_L2_ANNOUNCEMENTS\"] = \"1\"\n\t\t// If the agent is down for longer than the lease duration, stop responding\n\t\tcDefinesMap[\"L2_ANNOUNCEMENTS_MAX_LIVENESS\"] = fmt.Sprintf(\"%dULL\", option.Config.L2AnnouncerLeaseDuration.Nanoseconds())\n\t}\n\n\tif option.Config.EnableEncryptionStrictMode {\n\t\tcDefinesMap[\"ENCRYPTION_STRICT_MODE\"] = \"1\"\n\n\t\t// when parsing the user input we only accept ipv4 addresses\n\t\tcDefinesMap[\"STRICT_IPV4_NET\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPAddrToHost32(option.Config.EncryptionStrictModeCIDR.Addr()))\n\t\tcDefinesMap[\"STRICT_IPV4_NET_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.EncryptionStrictModeCIDR.Bits())\n\n\t\tcDefinesMap[\"IPV4_ENCRYPT_IFACE\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(node.GetIPv4()))\n\n\t\tipv4Interface, ok := netip.AddrFromSlice(node.GetIPv4().To4())\n\t\tif !ok {\n\t\t\treturn fmt.Errorf(\"unable to parse node IPv4 address %s\", node.GetIPv4())\n\t\t}\n\n\t\tif option.Config.EncryptionStrictModeCIDR.Contains(ipv4Interface) {\n\t\t\tif !option.Config.EncryptionStrictModeAllowRemoteNodeIdentities {\n\t\t\t\treturn fmt.Errorf(`encryption strict mode is enabled but the node's IPv4 address is within the strict CIDR range.\n\t\t\t\tThis will cause the node to drop all traffic.\n\t\t\t\tPlease either disable encryption or set --encryption-strict-mode-allow-dynamic-lookup=true`)\n\t\t\t}\n\t\t\tcDefinesMap[\"STRICT_IPV4_OVERLAPPING_CIDR\"] = \"1\"\n\t\t}\n\t}\n\n\tif option.Config.EnableBPFTProxy {\n\t\tcDefinesMap[\"ENABLE_TPROXY\"] = \"1\"\n\t}\n\n\tif option.Config.EnableXDPPrefilter {\n\t\tcDefinesMap[\"ENABLE_PREFILTER\"] = \"1\"\n\t}\n\n\tif option.Config.EnableEndpointRoutes {\n\t\tcDefinesMap[\"ENABLE_ENDPOINT_ROUTES\"] = \"1\"\n\t}\n\n\tif option.Config.EnableEnvoyConfig {\n\t\tcDefinesMap[\"ENABLE_L7_LB\"] = \"1\"\n\t}\n\n\tif option.Config.EnableSocketLB {\n\t\tif option.Config.BPFSocketLBHostnsOnly {\n\t\t\tcDefinesMap[\"ENABLE_SOCKET_LB_HOST_ONLY\"] = \"1\"\n\t\t} else {\n\t\t\tcDefinesMap[\"ENABLE_SOCKET_LB_FULL\"] = \"1\"\n\t\t}\n\t\tif option.Config.EnableSocketLBPeer {\n\t\t\tcDefinesMap[\"ENABLE_SOCKET_LB_PEER\"] = \"1\"\n\t\t}\n\t\tif option.Config.EnableSocketLBTracing {\n\t\t\tcDefinesMap[\"TRACE_SOCK_NOTIFY\"] = \"1\"\n\t\t}\n\n\t\tif cookie, err := netns.GetNetNSCookie(); err == nil {\n\t\t\t// When running in nested environments (e.g. Kind), cilium-agent does\n\t\t\t// not run in the host netns. So, in such cases the cookie comparison\n\t\t\t// based on bpf_get_netns_cookie(NULL) for checking whether a socket\n\t\t\t// belongs to a host netns does not work.\n\t\t\t//\n\t\t\t// To fix this, we derive the cookie of the netns in which cilium-agent\n\t\t\t// runs via getsockopt(...SO_NETNS_COOKIE...) and then use it in the\n\t\t\t// check above. This is based on an assumption that cilium-agent\n\t\t\t// always runs with \"hostNetwork: true\".\n\t\t\tcDefinesMap[\"HOST_NETNS_COOKIE\"] = fmt.Sprintf(\"%d\", cookie)\n\t\t}\n\t}\n\n\tcDefinesMap[\"NAT_46X64_PREFIX_0\"] = \"0\"\n\tcDefinesMap[\"NAT_46X64_PREFIX_1\"] = \"0\"\n\tcDefinesMap[\"NAT_46X64_PREFIX_2\"] = \"0\"\n\tcDefinesMap[\"NAT_46X64_PREFIX_3\"] = \"0\"\n\n\tif option.Config.EnableNodePort {\n\t\tif option.Config.EnableHealthDatapath {\n\t\t\tcDefinesMap[\"ENABLE_HEALTH_CHECK\"] = \"1\"\n\t\t}\n\t\tif option.Config.EnableMKE && option.Config.EnableSocketLB {\n\t\t\tcDefinesMap[\"ENABLE_MKE\"] = \"1\"\n\t\t\tcDefinesMap[\"MKE_HOST\"] = fmt.Sprintf(\"%d\", option.HostExtensionMKE)\n\t\t}\n\t\tif option.Config.EnableRecorder {\n\t\t\tcDefinesMap[\"ENABLE_CAPTURE\"] = \"1\"\n\t\t\tif option.Config.EnableIPv4 {\n\t\t\t\tcDefinesMap[\"CAPTURE4_RULES\"] = recorder.MapNameWcard4\n\t\t\t\tcDefinesMap[\"CAPTURE4_SIZE\"] = fmt.Sprintf(\"%d\", recorder.MapSize)\n\t\t\t}\n\t\t\tif option.Config.EnableIPv6 {\n\t\t\t\tcDefinesMap[\"CAPTURE6_RULES\"] = recorder.MapNameWcard6\n\t\t\t\tcDefinesMap[\"CAPTURE6_SIZE\"] = fmt.Sprintf(\"%d\", recorder.MapSize)\n\t\t\t}\n\t\t}\n\t\tcDefinesMap[\"ENABLE_NODEPORT\"] = \"1\"\n\t\tif option.Config.EnableIPv4 {\n\t\t\tcDefinesMap[\"NODEPORT_NEIGH4\"] = neighborsmap.Map4Name\n\t\t\tcDefinesMap[\"NODEPORT_NEIGH4_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.NeighMapEntriesGlobal)\n\t\t\tif option.Config.EnableHealthDatapath {\n\t\t\t\tcDefinesMap[\"LB4_HEALTH_MAP\"] = lbmap.HealthProbe4MapName\n\t\t\t}\n\t\t}\n\t\tif option.Config.EnableIPv6 {\n\t\t\tcDefinesMap[\"NODEPORT_NEIGH6\"] = neighborsmap.Map6Name\n\t\t\tcDefinesMap[\"NODEPORT_NEIGH6_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.NeighMapEntriesGlobal)\n\t\t\tif option.Config.EnableHealthDatapath {\n\t\t\t\tcDefinesMap[\"LB6_HEALTH_MAP\"] = lbmap.HealthProbe6MapName\n\t\t\t}\n\t\t}\n\t\tif option.Config.EnableNat46X64Gateway {\n\t\t\tcDefinesMap[\"ENABLE_NAT_46X64_GATEWAY\"] = \"1\"\n\t\t\tbase := option.Config.IPv6NAT46x64CIDRBase.AsSlice()\n\t\t\tcDefinesMap[\"NAT_46X64_PREFIX_0\"] = fmt.Sprintf(\"%d\", base[0])\n\t\t\tcDefinesMap[\"NAT_46X64_PREFIX_1\"] = fmt.Sprintf(\"%d\", base[1])\n\t\t\tcDefinesMap[\"NAT_46X64_PREFIX_2\"] = fmt.Sprintf(\"%d\", base[2])\n\t\t\tcDefinesMap[\"NAT_46X64_PREFIX_3\"] = fmt.Sprintf(\"%d\", base[3])\n\t\t}\n\t\tif option.Config.NodePortNat46X64 {\n\t\t\tcDefinesMap[\"ENABLE_NAT_46X64\"] = \"1\"\n\t\t}\n\t\tconst (\n\t\t\tdsrEncapInv = iota\n\t\t\tdsrEncapNone\n\t\t\tdsrEncapIPIP\n\t\t\tdsrEncapGeneve\n\t\t)\n\t\tconst (\n\t\t\tdsrL4XlateInv = iota\n\t\t\tdsrL4XlateFrontend\n\t\t\tdsrL4XlateBackend\n\t\t)\n\t\tcDefinesMap[\"DSR_ENCAP_IPIP\"] = fmt.Sprintf(\"%d\", dsrEncapIPIP)\n\t\tcDefinesMap[\"DSR_ENCAP_GENEVE\"] = fmt.Sprintf(\"%d\", dsrEncapGeneve)\n\t\tcDefinesMap[\"DSR_ENCAP_NONE\"] = fmt.Sprintf(\"%d\", dsrEncapNone)\n\t\tcDefinesMap[\"DSR_XLATE_FRONTEND\"] = fmt.Sprintf(\"%d\", dsrL4XlateFrontend)\n\t\tcDefinesMap[\"DSR_XLATE_BACKEND\"] = fmt.Sprintf(\"%d\", dsrL4XlateBackend)\n\t\tif option.Config.NodePortMode == option.NodePortModeDSR ||\n\t\t\toption.Config.NodePortMode == option.NodePortModeHybrid {\n\t\t\tcDefinesMap[\"ENABLE_DSR\"] = \"1\"\n\t\t\tif option.Config.EnablePMTUDiscovery {\n\t\t\t\tcDefinesMap[\"ENABLE_DSR_ICMP_ERRORS\"] = \"1\"\n\t\t\t}\n\t\t\tif option.Config.NodePortMode == option.NodePortModeHybrid {\n\t\t\t\tcDefinesMap[\"ENABLE_DSR_HYBRID\"] = \"1\"\n\t\t\t}\n\t\t\tif option.Config.LoadBalancerDSRDispatch == option.DSRDispatchOption {\n\t\t\t\tcDefinesMap[\"DSR_ENCAP_MODE\"] = fmt.Sprintf(\"%d\", dsrEncapNone)\n\t\t\t} else if option.Config.LoadBalancerDSRDispatch == option.DSRDispatchIPIP {\n\t\t\t\tcDefinesMap[\"DSR_ENCAP_MODE\"] = fmt.Sprintf(\"%d\", dsrEncapIPIP)\n\t\t\t} else if option.Config.LoadBalancerDSRDispatch == option.DSRDispatchGeneve {\n\t\t\t\tcDefinesMap[\"DSR_ENCAP_MODE\"] = fmt.Sprintf(\"%d\", dsrEncapGeneve)\n\t\t\t}\n\t\t\tif option.Config.LoadBalancerDSRDispatch == option.DSRDispatchIPIP {\n\t\t\t\tif option.Config.LoadBalancerDSRL4Xlate == option.DSRL4XlateFrontend {\n\t\t\t\t\tcDefinesMap[\"DSR_XLATE_MODE\"] = fmt.Sprintf(\"%d\", dsrL4XlateFrontend)\n\t\t\t\t} else if option.Config.LoadBalancerDSRL4Xlate == option.DSRL4XlateBackend {\n\t\t\t\t\tcDefinesMap[\"DSR_XLATE_MODE\"] = fmt.Sprintf(\"%d\", dsrL4XlateBackend)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tcDefinesMap[\"DSR_XLATE_MODE\"] = fmt.Sprintf(\"%d\", dsrL4XlateInv)\n\t\t\t}\n\t\t} else {\n\t\t\tcDefinesMap[\"DSR_ENCAP_MODE\"] = fmt.Sprintf(\"%d\", dsrEncapInv)\n\t\t\tcDefinesMap[\"DSR_XLATE_MODE\"] = fmt.Sprintf(\"%d\", dsrL4XlateInv)\n\t\t}\n\t\tif option.Config.EnableIPv4 {\n\t\t\tif option.Config.LoadBalancerRSSv4CIDR != \"\" {\n\t\t\t\tipv4 := byteorder.NetIPv4ToHost32(option.Config.LoadBalancerRSSv4.IP)\n\t\t\t\tones, _ := option.Config.LoadBalancerRSSv4.Mask.Size()\n\t\t\t\tcDefinesMap[\"IPV4_RSS_PREFIX\"] = fmt.Sprintf(\"%d\", ipv4)\n\t\t\t\tcDefinesMap[\"IPV4_RSS_PREFIX_BITS\"] = fmt.Sprintf(\"%d\", ones)\n\t\t\t} else {\n\t\t\t\tcDefinesMap[\"IPV4_RSS_PREFIX\"] = \"IPV4_DIRECT_ROUTING\"\n\t\t\t\tcDefinesMap[\"IPV4_RSS_PREFIX_BITS\"] = \"32\"\n\t\t\t}\n\t\t}\n\t\tif option.Config.EnableIPv6 {\n\t\t\tif option.Config.LoadBalancerRSSv6CIDR != \"\" {\n\t\t\t\tipv6 := option.Config.LoadBalancerRSSv6.IP\n\t\t\t\tones, _ := option.Config.LoadBalancerRSSv6.Mask.Size()\n\t\t\t\textraMacrosMap[\"IPV6_RSS_PREFIX\"] = ipv6.String()\n\t\t\t\tfw.WriteString(FmtDefineAddress(\"IPV6_RSS_PREFIX\", ipv6))\n\t\t\t\tcDefinesMap[\"IPV6_RSS_PREFIX_BITS\"] = fmt.Sprintf(\"%d\", ones)\n\t\t\t} else {\n\t\t\t\tcDefinesMap[\"IPV6_RSS_PREFIX\"] = \"IPV6_DIRECT_ROUTING\"\n\t\t\t\tcDefinesMap[\"IPV6_RSS_PREFIX_BITS\"] = \"128\"\n\t\t\t}\n\t\t}\n\t\tif option.Config.NodePortAcceleration != option.NodePortAccelerationDisabled {\n\t\t\tcDefinesMap[\"ENABLE_NODEPORT_ACCELERATION\"] = \"1\"\n\t\t}\n\t\tif !option.Config.EnableHostLegacyRouting {\n\t\t\tcDefinesMap[\"ENABLE_HOST_ROUTING\"] = \"1\"\n\t\t}\n\t\tif option.Config.EnableSVCSourceRangeCheck {\n\t\t\tcDefinesMap[\"ENABLE_SRC_RANGE_CHECK\"] = \"1\"\n\t\t\tif option.Config.EnableIPv4 {\n\t\t\t\tcDefinesMap[\"LB4_SRC_RANGE_MAP\"] = lbmap.SourceRange4MapName\n\t\t\t\tcDefinesMap[\"LB4_SRC_RANGE_MAP_SIZE\"] =\n\t\t\t\t\tfmt.Sprintf(\"%d\", lbmap.SourceRange4Map.MaxEntries())\n\t\t\t}\n\t\t\tif option.Config.EnableIPv6 {\n\t\t\t\tcDefinesMap[\"LB6_SRC_RANGE_MAP\"] = lbmap.SourceRange6MapName\n\t\t\t\tcDefinesMap[\"LB6_SRC_RANGE_MAP_SIZE\"] =\n\t\t\t\t\tfmt.Sprintf(\"%d\", lbmap.SourceRange6Map.MaxEntries())\n\t\t\t}\n\t\t}\n\n\t\tcDefinesMap[\"NODEPORT_PORT_MIN\"] = fmt.Sprintf(\"%d\", option.Config.NodePortMin)\n\t\tcDefinesMap[\"NODEPORT_PORT_MAX\"] = fmt.Sprintf(\"%d\", option.Config.NodePortMax)\n\t\tcDefinesMap[\"NODEPORT_PORT_MIN_NAT\"] = fmt.Sprintf(\"%d\", option.Config.NodePortMax+1)\n\t\tcDefinesMap[\"NODEPORT_PORT_MAX_NAT\"] = \"65535\"\n\t}\n\n\tmacByIfIndexMacro, isL3DevMacro, err := devMacros()\n\tif err != nil {\n\t\treturn err\n\t}\n\tcDefinesMap[\"NATIVE_DEV_MAC_BY_IFINDEX(IFINDEX)\"] = macByIfIndexMacro\n\tcDefinesMap[\"IS_L3_DEV(ifindex)\"] = isL3DevMacro\n\n\tconst (\n\t\tselectionRandom = iota + 1\n\t\tselectionMaglev\n\t)\n\tcDefinesMap[\"LB_SELECTION_RANDOM\"] = fmt.Sprintf(\"%d\", selectionRandom)\n\tcDefinesMap[\"LB_SELECTION_MAGLEV\"] = fmt.Sprintf(\"%d\", selectionMaglev)\n\tif option.Config.NodePortAlg == option.NodePortAlgRandom {\n\t\tcDefinesMap[\"LB_SELECTION\"] = fmt.Sprintf(\"%d\", selectionRandom)\n\t} else if option.Config.NodePortAlg == option.NodePortAlgMaglev {\n\t\tcDefinesMap[\"LB_SELECTION\"] = fmt.Sprintf(\"%d\", selectionMaglev)\n\t\tcDefinesMap[\"LB_MAGLEV_LUT_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.MaglevTableSize)\n\t\tif option.Config.EnableIPv6 {\n\t\t\tcDefinesMap[\"LB6_MAGLEV_MAP_OUTER\"] = lbmap.MaglevOuter6MapName\n\t\t}\n\t\tif option.Config.EnableIPv4 {\n\t\t\tcDefinesMap[\"LB4_MAGLEV_MAP_OUTER\"] = lbmap.MaglevOuter4MapName\n\t\t}\n\t}\n\tcDefinesMap[\"HASH_INIT4_SEED\"] = fmt.Sprintf(\"%d\", maglev.SeedJhash0)\n\tcDefinesMap[\"HASH_INIT6_SEED\"] = fmt.Sprintf(\"%d\", maglev.SeedJhash1)\n\n\tif option.Config.DirectRoutingDeviceRequired() {\n\t\tdirectRoutingIface := option.Config.DirectRoutingDevice\n\t\tdirectRoutingIfIndex, err := link.GetIfIndex(directRoutingIface)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcDefinesMap[\"DIRECT_ROUTING_DEV_IFINDEX\"] = fmt.Sprintf(\"%d\", directRoutingIfIndex)\n\t\tif option.Config.EnableIPv4 {\n\t\t\tip, ok := node.GetNodePortIPv4AddrsWithDevices()[directRoutingIface]\n\t\t\tif !ok {\n\t\t\t\tlog.WithFields(logrus.Fields{\n\t\t\t\t\t\"directRoutingIface\": directRoutingIface,\n\t\t\t\t}).Fatal(\"Direct routing device's IPv4 address not found\")\n\t\t\t}\n\n\t\t\tipv4 := byteorder.NetIPv4ToHost32(ip)\n\t\t\tcDefinesMap[\"IPV4_DIRECT_ROUTING\"] = fmt.Sprintf(\"%d\", ipv4)\n\t\t}\n\n\t\tif option.Config.EnableIPv6 {\n\t\t\tdirectRoutingIPv6, ok := node.GetNodePortIPv6AddrsWithDevices()[directRoutingIface]\n\t\t\tif !ok {\n\t\t\t\tlog.WithFields(logrus.Fields{\n\t\t\t\t\t\"directRoutingIface\": directRoutingIface,\n\t\t\t\t}).Fatal(\"Direct routing device's IPv6 address not found\")\n\t\t\t}\n\n\t\t\textraMacrosMap[\"IPV6_DIRECT_ROUTING\"] = directRoutingIPv6.String()\n\t\t\tfw.WriteString(FmtDefineAddress(\"IPV6_DIRECT_ROUTING\", directRoutingIPv6))\n\t\t}\n\t} else {\n\t\tvar directRoutingIPv6 net.IP\n\t\tcDefinesMap[\"DIRECT_ROUTING_DEV_IFINDEX\"] = \"0\"\n\t\tif option.Config.EnableIPv4 {\n\t\t\tcDefinesMap[\"IPV4_DIRECT_ROUTING\"] = \"0\"\n\t\t}\n\t\tif option.Config.EnableIPv6 {\n\t\t\textraMacrosMap[\"IPV6_DIRECT_ROUTING\"] = directRoutingIPv6.String()\n\t\t\tfw.WriteString(FmtDefineAddress(\"IPV6_DIRECT_ROUTING\", directRoutingIPv6))\n\t\t}\n\t}\n\n\tif option.Config.ResetQueueMapping {\n\t\tcDefinesMap[\"RESET_QUEUES\"] = \"1\"\n\t}\n\n\tif option.Config.EnableBandwidthManager {\n\t\tcDefinesMap[\"ENABLE_BANDWIDTH_MANAGER\"] = \"1\"\n\t\tcDefinesMap[\"THROTTLE_MAP\"] = bwmap.MapName\n\t\tcDefinesMap[\"THROTTLE_MAP_SIZE\"] = fmt.Sprintf(\"%d\", bwmap.MapSize)\n\t}\n\n\tif option.Config.EnableHostFirewall {\n\t\tcDefinesMap[\"ENABLE_HOST_FIREWALL\"] = \"1\"\n\t}\n\n\tif option.Config.EnableIPSec {\n\t\tnodeAddress := node.GetIPv4()\n\t\tif nodeAddress == nil {\n\t\t\treturn errors.New(\"external IPv4 node address is required when IPSec is enabled, but none found\")\n\t\t}\n\n\t\ta := byteorder.NetIPv4ToHost32(nodeAddress)\n\t\tcDefinesMap[\"IPV4_ENCRYPT_IFACE\"] = fmt.Sprintf(\"%d\", a)\n\t\tif iface := option.Config.EncryptInterface; len(iface) != 0 {\n\t\t\tlink, err := netlink.LinkByName(iface[0])\n\t\t\tif err == nil {\n\t\t\t\tcDefinesMap[\"ENCRYPT_IFACE\"] = fmt.Sprintf(\"%d\", link.Attrs().Index)\n\t\t\t}\n\t\t}\n\t}\n\n\tif option.Config.EnableNodePort {\n\t\tif option.Config.EnableIPv4 {\n\t\t\tcDefinesMap[\"SNAT_MAPPING_IPV4\"] = nat.MapNameSnat4Global\n\t\t\tcDefinesMap[\"SNAT_MAPPING_IPV4_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.NATMapEntriesGlobal)\n\t\t}\n\n\t\tif option.Config.EnableIPv6 {\n\t\t\tcDefinesMap[\"SNAT_MAPPING_IPV6\"] = nat.MapNameSnat6Global\n\t\t\tcDefinesMap[\"SNAT_MAPPING_IPV6_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.NATMapEntriesGlobal)\n\t\t}\n\n\t\tif option.Config.EnableBPFMasquerade {\n\t\t\tif option.Config.EnableIPv4Masquerade {\n\t\t\t\tcDefinesMap[\"ENABLE_MASQUERADE_IPV4\"] = \"1\"\n\n\t\t\t\t// ip-masq-agent depends on bpf-masq\n\t\t\t\tvar excludeCIDR *cidr.CIDR\n\t\t\t\tif option.Config.EnableIPMasqAgent {\n\t\t\t\t\tcDefinesMap[\"ENABLE_IP_MASQ_AGENT_IPV4\"] = \"1\"\n\t\t\t\t\tcDefinesMap[\"IP_MASQ_AGENT_IPV4\"] = ipmasq.MapNameIPv4\n\n\t\t\t\t\t// native-routing-cidr is optional with ip-masq-agent and may be nil\n\t\t\t\t\texcludeCIDR = option.Config.GetIPv4NativeRoutingCIDR()\n\t\t\t\t} else {\n\t\t\t\t\texcludeCIDR = datapath.RemoteSNATDstAddrExclusionCIDRv4()\n\t\t\t\t}\n\n\t\t\t\tif excludeCIDR != nil {\n\t\t\t\t\tcDefinesMap[\"IPV4_SNAT_EXCLUSION_DST_CIDR\"] =\n\t\t\t\t\t\tfmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(excludeCIDR.IP))\n\t\t\t\t\tones, _ := excludeCIDR.Mask.Size()\n\t\t\t\t\tcDefinesMap[\"IPV4_SNAT_EXCLUSION_DST_CIDR_LEN\"] = fmt.Sprintf(\"%d\", ones)\n\t\t\t\t}\n\t\t\t}\n\t\t\tif option.Config.EnableIPv6Masquerade {\n\t\t\t\tcDefinesMap[\"ENABLE_MASQUERADE_IPV6\"] = \"1\"\n\n\t\t\t\tvar excludeCIDR *cidr.CIDR\n\t\t\t\tif option.Config.EnableIPMasqAgent {\n\t\t\t\t\tcDefinesMap[\"ENABLE_IP_MASQ_AGENT_IPV6\"] = \"1\"\n\t\t\t\t\tcDefinesMap[\"IP_MASQ_AGENT_IPV6\"] = ipmasq.MapNameIPv6\n\n\t\t\t\t\texcludeCIDR = option.Config.GetIPv6NativeRoutingCIDR()\n\t\t\t\t} else {\n\t\t\t\t\texcludeCIDR = datapath.RemoteSNATDstAddrExclusionCIDRv6()\n\t\t\t\t}\n\n\t\t\t\tif excludeCIDR != nil {\n\t\t\t\t\textraMacrosMap[\"IPV6_SNAT_EXCLUSION_DST_CIDR\"] = excludeCIDR.IP.String()\n\t\t\t\t\tfw.WriteString(FmtDefineAddress(\"IPV6_SNAT_EXCLUSION_DST_CIDR\", excludeCIDR.IP))\n\t\t\t\t\textraMacrosMap[\"IPV6_SNAT_EXCLUSION_DST_CIDR_MASK\"] = excludeCIDR.Mask.String()\n\t\t\t\t\tfw.WriteString(FmtDefineAddress(\"IPV6_SNAT_EXCLUSION_DST_CIDR_MASK\", excludeCIDR.Mask))\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\tctmap.WriteBPFMacros(fw, nil)\n\t}\n\n\tif option.Config.AllowICMPFragNeeded {\n\t\tcDefinesMap[\"ALLOW_ICMP_FRAG_NEEDED\"] = \"1\"\n\t}\n\n\tif option.Config.ClockSource == option.ClockSourceJiffies {\n\t\tcDefinesMap[\"ENABLE_JIFFIES\"] = \"1\"\n\t}\n\n\tif option.Config.EnableIdentityMark {\n\t\tcDefinesMap[\"ENABLE_IDENTITY_MARK\"] = \"1\"\n\t}\n\n\tif option.Config.EnableHighScaleIPcache {\n\t\tcDefinesMap[\"ENABLE_HIGH_SCALE_IPCACHE\"] = \"1\"\n\t}\n\n\tif option.Config.EnableCustomCalls {\n\t\tcDefinesMap[\"ENABLE_CUSTOM_CALLS\"] = \"1\"\n\t}\n\n\tif option.Config.EnableVTEP {\n\t\tcDefinesMap[\"ENABLE_VTEP\"] = \"1\"\n\t\tcDefinesMap[\"VTEP_MAP\"] = vtep.Name\n\t\tcDefinesMap[\"VTEP_MAP_SIZE\"] = fmt.Sprintf(\"%d\", vtep.MaxEntries)\n\t\tcDefinesMap[\"VTEP_MASK\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(net.IP(option.Config.VtepCidrMask)))\n\n\t}\n\n\tvlanFilter, err := vlanFilterMacros()\n\tif err != nil {\n\t\treturn err\n\t}\n\tcDefinesMap[\"VLAN_FILTER(ifindex, vlan_id)\"] = vlanFilter\n\n\tif option.Config.EnableICMPRules {\n\t\tcDefinesMap[\"ENABLE_ICMP_RULE\"] = \"1\"\n\t}\n\n\tcDefinesMap[\"CIDR_IDENTITY_RANGE_START\"] = fmt.Sprintf(\"%d\", identity.MinLocalIdentity)\n\tcDefinesMap[\"CIDR_IDENTITY_RANGE_END\"] = fmt.Sprintf(\"%d\", identity.MaxLocalIdentity)\n\n\tif option.Config.TunnelingEnabled() {\n\t\tcDefinesMap[\"TUNNEL_MODE\"] = \"1\"\n\t}\n\n\tciliumNetLink, err := netlink.LinkByName(defaults.SecondHostDevice)\n\tif err != nil {\n\t\treturn err\n\t}\n\tcDefinesMap[\"CILIUM_NET_MAC\"] = fmt.Sprintf(\"{.addr=%s}\", mac.CArrayString(ciliumNetLink.Attrs().HardwareAddr))\n\tcDefinesMap[\"HOST_IFINDEX\"] = fmt.Sprintf(\"%d\", ciliumNetLink.Attrs().Index)\n\n\tciliumHostLink, err := netlink.LinkByName(defaults.HostDevice)\n\tif err != nil {\n\t\treturn err\n\t}\n\tcDefinesMap[\"HOST_IFINDEX_MAC\"] = fmt.Sprintf(\"{.addr=%s}\", mac.CArrayString(ciliumHostLink.Attrs().HardwareAddr))\n\tcDefinesMap[\"CILIUM_IFINDEX\"] = fmt.Sprintf(\"%d\", ciliumHostLink.Attrs().Index)\n\n\tephemeralMin, err := getEphemeralPortRangeMin()\n\tif err != nil {\n\t\treturn err\n\t}\n\tcDefinesMap[\"EPHEMERAL_MIN\"] = fmt.Sprintf(\"%d\", ephemeralMin)\n\n\tif err := cDefinesMap.Merge(h.nodeExtraDefines); err != nil {\n\t\treturn err\n\t}\n\n\tfor _, fn := range h.nodeExtraDefineFns {\n\t\tdefines, err := fn()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := cDefinesMap.Merge(defines); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif option.Config.EnableHealthDatapath {\n\t\tif option.Config.IPv4Enabled() {\n\t\t\tipip4, err := netlink.LinkByName(defaults.IPIPv4Device)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tcDefinesMap[\"ENCAP4_IFINDEX\"] = fmt.Sprintf(\"%d\", ipip4.Attrs().Index)\n\t\t}\n\t\tif option.Config.IPv6Enabled() {\n\t\t\tipip6, err := netlink.LinkByName(defaults.IPIPv6Device)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tcDefinesMap[\"ENCAP6_IFINDEX\"] = fmt.Sprintf(\"%d\", ipip6.Attrs().Index)\n\t\t}\n\t}\n\n\t// Since golang maps are unordered, we sort the keys in the map\n\t// to get a consistent written format to the writer. This maintains\n\t// the consistency when we try to calculate hash for a datapath after\n\t// writing the config.\n\tkeys := make([]string, 0, len(cDefinesMap))\n\tfor key := range cDefinesMap {\n\t\tkeys = append(keys, key)\n\t}\n\tsort.Strings(keys)\n\n\tfor _, key := range keys {\n\t\tfmt.Fprintf(fw, \"#define %s %s\\n\", key, cDefinesMap[key])\n\t}\n\n\t// Populate cDefinesMap with extraMacrosMap to get all the configuration\n\t// in the cDefinesMap itself.\n\tfor key, value := range extraMacrosMap {\n\t\tcDefinesMap[key] = value\n\t}\n\n\t// Write the JSON encoded config as base64 encoded commented string to\n\t// the header file.\n\tjsonBytes, err := json.Marshal(cDefinesMap)\n\tif err == nil {\n\t\t// We don't care if some error occurs while marshaling the map.\n\t\t// In such cases we skip embedding the base64 encoded JSON configuration\n\t\t// to the writer.\n\t\tencodedConfig := base64.StdEncoding.EncodeToString(jsonBytes)\n\t\tfmt.Fprintf(fw, \"\\n// JSON_OUTPUT: %s\\n\", encodedConfig)\n\t}\n\n\treturn fw.Flush()\n}", "func (o ClusterOutput) ConfigurationEndpoint() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Cluster) pulumi.StringOutput { return v.ConfigurationEndpoint }).(pulumi.StringOutput)\n}", "func (r *templateRouter) writeConfig() error {\n\t//write out any certificate files that don't exist\n\t//TODO: better way so this doesn't need to create lots of files every time state is written, probably too expensive\n\tfor _, serviceUnit := range r.state {\n\t\tfor _, cfg := range serviceUnit.ServiceAliasConfigs {\n\t\t\tr.certManager.writeCertificatesForConfig(&cfg)\n\t\t}\n\t}\n\n\tfor path, template := range r.templates {\n\t\tfile, err := os.Create(path)\n\t\tif err != nil {\n\t\t\tglog.Errorf(\"Error creating config file %v: %v\", path, err)\n\t\t\treturn err\n\t\t}\n\n\t\terr = template.Execute(file, r.state)\n\t\tif err != nil {\n\t\t\tglog.Errorf(\"Error executing template for file %v: %v\", path, err)\n\t\t\treturn err\n\t\t}\n\n\t\tfile.Close()\n\t}\n\n\treturn nil\n}", "func setupSink(config map[string]any) (*os.File, error) {\n\taddrOpaque, ok := config[\"endpoint\"]\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"endpoint not present in configuration\")\n\t}\n\taddr, ok := addrOpaque.(string)\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"endpoint %q is not a string\", addrOpaque)\n\t}\n\treturn setup(addr)\n}", "func (s *ListModelExplainabilityJobDefinitionsInput) SetEndpointName(v string) *ListModelExplainabilityJobDefinitionsInput {\n\ts.EndpointName = &v\n\treturn s\n}", "func (s *ListDataQualityJobDefinitionsInput) SetEndpointName(v string) *ListDataQualityJobDefinitionsInput {\n\ts.EndpointName = &v\n\treturn s\n}", "func (s *ModelDeployResult) SetEndpointName(v string) *ModelDeployResult {\n\ts.EndpointName = &v\n\treturn s\n}", "func (s *ListModelQualityJobDefinitionsInput) SetEndpointName(v string) *ListModelQualityJobDefinitionsInput {\n\ts.EndpointName = &v\n\treturn s\n}", "func (vl *VlanBridge) AddEndpoint(endpoint *OfnetEndpoint) error {\n\tlog.Infof(\"Received endpoint: %+v\", endpoint)\n\n\t// Install dst group entry for the endpoint\n\terr := vl.policyAgent.AddEndpoint(endpoint)\n\tif err != nil {\n\t\tlog.Errorf(\"Error adding endpoint to policy agent{%+v}. Err: %v\", endpoint, err)\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func WriteConfigFile(c *cli.Context, cfg *Config) error {\n\tfPath, err := TenetCfgPath(c)\n\tif err != nil {\n\t\treturn errors.Trace(err)\n\t}\n\tvar buf bytes.Buffer\n\tenc := toml.NewEncoder(&buf)\n\terr = enc.Encode(cfg)\n\tif err != nil {\n\t\treturn errors.Trace(err)\n\t}\n\treturn ioutil.WriteFile(fPath, buf.Bytes(), 0644)\n}", "func (in *EndpointConfigSummary) DeepCopy() *EndpointConfigSummary {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(EndpointConfigSummary)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (m *MockProviders) EndpointConfig() fab.EndpointConfig {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"EndpointConfig\")\n\tret0, _ := ret[0].(fab.EndpointConfig)\n\treturn ret0\n}", "func (r *DnsEndpointReconciler) endpointsForDnsEndpoint(d *k8sv1.DnsEndpoint, addresses []corev1.EndpointAddress) *corev1.Endpoints {\n\tls := labelsForDnsEndpoint(d.Name)\n\n\tendpoints := &corev1.Endpoints{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tName: d.Name,\n\t\t\tNamespace: d.Namespace,\n\t\t\tLabels: ls,\n\t\t},\n\t\tSubsets: []corev1.EndpointSubset{{\n\t\t\tAddresses: addresses,\n\t\t\tPorts: []corev1.EndpointPort{{\n\t\t\t\tName: d.Spec.Name,\n\t\t\t\tPort: d.Spec.Port,\n\t\t\t}},\n\t\t}},\n\t}\n\t// Set DnsEndpoint instance as the owner and controller\n\tctrl.SetControllerReference(d, endpoints, r.Scheme)\n\treturn endpoints\n}", "func (s *HttpsNotificationConfiguration) SetEndpoint(v string) *HttpsNotificationConfiguration {\n\ts.Endpoint = &v\n\treturn s\n}", "func WithCollectorEndpoint(options ...CollectorEndpointOption) EndpointOption {\n\treturn endpointOptionFunc(func() (batchUploader, error) {\n\t\tcfg := &collectorEndpointConfig{\n\t\t\tendpoint: envOr(envEndpoint, \"http://localhost:14268/api/traces\"),\n\t\t\tusername: envOr(envUser, \"\"),\n\t\t\tpassword: envOr(envPassword, \"\"),\n\t\t\thttpClient: http.DefaultClient,\n\t\t}\n\n\t\tfor _, opt := range options {\n\t\t\topt.apply(cfg)\n\t\t}\n\n\t\treturn &collectorUploader{\n\t\t\tendpoint: cfg.endpoint,\n\t\t\tusername: cfg.username,\n\t\t\tpassword: cfg.password,\n\t\t\thttpClient: cfg.httpClient,\n\t\t}, nil\n\t})\n}", "func (cfg *Config) Write() error {\n\tif err := os.MkdirAll(filepath.Dir(cfg.path), 0755); err != nil {\n\t\treturn errors.Wrap(err, \"failed to create config path\")\n\t}\n\n\tfh, err := os.Create(cfg.path)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to open config file\")\n\t}\n\tdefer fh.Close()\n\n\tbuf, err := yaml.Marshal(cfg)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to marshal config\")\n\t}\n\n\t_, err = fh.Write(buf)\n\treturn errors.Wrap(err, \"failed to write config\")\n}", "func (s *EndpointOutputConfiguration) SetEndpointName(v string) *EndpointOutputConfiguration {\n\ts.EndpointName = &v\n\treturn s\n}", "func (s *EndpointSummary) SetEndpointName(v string) *EndpointSummary {\n\ts.EndpointName = &v\n\treturn s\n}", "func (s CreateEndpointConfigOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s CreateEndpointConfigOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (b *PodNetworkConnectivityCheckSpecApplyConfiguration) WithTargetEndpoint(value string) *PodNetworkConnectivityCheckSpecApplyConfiguration {\n\tb.TargetEndpoint = &value\n\treturn b\n}" ]
[ "0.6244771", "0.57506377", "0.5709975", "0.5645815", "0.5617773", "0.5518858", "0.5429672", "0.5404889", "0.54006004", "0.5368511", "0.53134185", "0.5267684", "0.52165043", "0.519649", "0.5177079", "0.5149827", "0.5149628", "0.5138795", "0.5119497", "0.51023805", "0.5093861", "0.50864387", "0.5080077", "0.5033053", "0.50286686", "0.50237256", "0.4995644", "0.49917302", "0.49711153", "0.4969318", "0.4968494", "0.49665183", "0.49590477", "0.4950771", "0.49482164", "0.49466527", "0.4939549", "0.4937399", "0.49365023", "0.4923569", "0.49185506", "0.49113464", "0.48823583", "0.48738202", "0.48481005", "0.48362368", "0.4830733", "0.4822381", "0.4812781", "0.4804096", "0.47951066", "0.479319", "0.4766198", "0.47567618", "0.47494623", "0.47482562", "0.4746566", "0.47387442", "0.47365785", "0.4735561", "0.47350982", "0.47286174", "0.47280225", "0.47269472", "0.47021627", "0.4698423", "0.46975508", "0.4692552", "0.46898103", "0.46816018", "0.46814418", "0.46774313", "0.46622294", "0.4656007", "0.4651735", "0.46415865", "0.46397936", "0.4630313", "0.4627595", "0.4622736", "0.46212396", "0.46195054", "0.4616351", "0.46156633", "0.46155724", "0.46139434", "0.46065313", "0.46021345", "0.45882717", "0.4578785", "0.45772263", "0.45705312", "0.45555145", "0.45548916", "0.45476082", "0.4546762", "0.45424587", "0.45292175", "0.45292175", "0.45241636" ]
0.80347174
0
WriteTemplateConfig writes the BPF configuration for the template to a writer.
func (h *HeaderfileWriter) WriteTemplateConfig(w io.Writer, e datapath.EndpointConfiguration) error { fw := bufio.NewWriter(w) return h.writeTemplateConfig(fw, e) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (r *templateRouter) writeConfig() error {\n\t//write out any certificate files that don't exist\n\t//TODO: better way so this doesn't need to create lots of files every time state is written, probably too expensive\n\tfor _, serviceUnit := range r.state {\n\t\tfor _, cfg := range serviceUnit.ServiceAliasConfigs {\n\t\t\tr.certManager.writeCertificatesForConfig(&cfg)\n\t\t}\n\t}\n\n\tfor path, template := range r.templates {\n\t\tfile, err := os.Create(path)\n\t\tif err != nil {\n\t\t\tglog.Errorf(\"Error creating config file %v: %v\", path, err)\n\t\t\treturn err\n\t\t}\n\n\t\terr = template.Execute(file, r.state)\n\t\tif err != nil {\n\t\t\tglog.Errorf(\"Error executing template for file %v: %v\", path, err)\n\t\t\treturn err\n\t\t}\n\n\t\tfile.Close()\n\t}\n\n\treturn nil\n}", "func (t *TemplateConfig) Write() error {\n\tfp := filepath.Join(t.ProjectPath, \"template.yml\")\n\t// make sure directory exists\n\tif _, err := os.Stat(t.ProjectPath); os.IsNotExist(err) {\n\t\tif err := os.MkdirAll(t.ProjectPath, 0755); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tyml, err := yaml.Marshal(t)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn ioutil.WriteFile(fp, yml, 0644)\n}", "func CreateConfigTemplate(filename string) error {\n\tvar config Config\n\tmarshaled, err := json.Marshal(config)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err := ioutil.WriteFile(filename, marshaled, 0600); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (mod *EthModule) WriteConfig(config_file string) error {\n\tb, err := json.Marshal(mod.eth.config)\n\tif err != nil {\n\t\tfmt.Println(\"error marshalling config:\", err)\n\t\treturn err\n\t}\n\tvar out bytes.Buffer\n\tjson.Indent(&out, b, \"\", \"\\t\")\n\terr = ioutil.WriteFile(config_file, out.Bytes(), 0600)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func WriteConfig(c Config, filename string) {\n\n\tdata, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\terr = ioutil.WriteFile(filename, data, 664)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func WriteThrapConfig(conf *ThrapConfig, filename string) error {\n\tb, err := hclencoder.Encode(conf)\n\tif err == nil {\n\t\terr = ioutil.WriteFile(filename, b, 0644)\n\t}\n\treturn err\n}", "func WriteTemplateToFile(tpl string, config interface{}, writepath string, filemode os.FileMode) error {\n\tvar tplbuffer bytes.Buffer\n\tvar packageTemplate = template.Must(template.New(\"\").Parse(tpl))\n\terr := packageTemplate.Execute(&tplbuffer, config)\n\tif err != nil {\n\t\tlog.Warnf(\"Unable to translate template %q to string using the data %v\", tpl, config)\n\t\treturn err\n\t}\n\terr = ioutil.WriteFile(writepath, tplbuffer.Bytes(), filemode)\n\tif err != nil {\n\t\tlog.Warnf(\"Error writing file at %s : %s\", writepath, err)\n\t\treturn err\n\t}\n\treturn nil\n}", "func WriteConfigFile(configFilePath string, config *Config) {\r\n\tvar buffer bytes.Buffer\r\n\r\n\tif err := configTemplate.Execute(&buffer, config); err != nil {\r\n\t\tpanic(err)\r\n\t}\r\n\r\n\ttmos.MustWriteFile(configFilePath, buffer.Bytes(), 0644)\r\n}", "func (r *templateRouter) writeConfig() error {\n\t//write out any certificate files that don't exist\n\tfor k, cfg := range r.state {\n\t\tcfg := cfg // avoid implicit memory aliasing (gosec G601)\n\t\tif err := r.writeCertificates(&cfg); err != nil {\n\t\t\treturn fmt.Errorf(\"error writing certificates for %s: %v\", k, err)\n\t\t}\n\n\t\t// calculate the server weight for the endpoints in each service\n\t\t// called here to make sure we have the actual number of endpoints.\n\t\tcfg.ServiceUnitNames = r.calculateServiceWeights(cfg.ServiceUnits, cfg.PreferPort)\n\n\t\t// Calculate the number of active endpoints for the route.\n\t\tcfg.ActiveEndpoints = r.getActiveEndpoints(cfg.ServiceUnits, cfg.PreferPort)\n\n\t\tcfg.Status = ServiceAliasConfigStatusSaved\n\t\tr.state[k] = cfg\n\t}\n\n\tlog.V(4).Info(\"committing router certificate manager changes...\")\n\tif err := r.certManager.Commit(); err != nil {\n\t\treturn fmt.Errorf(\"error committing certificate changes: %v\", err)\n\t}\n\n\tlog.V(4).Info(\"router certificate manager config committed\")\n\n\tdisableHTTP2, _ := strconv.ParseBool(os.Getenv(\"ROUTER_DISABLE_HTTP2\"))\n\n\tfor name, template := range r.templates {\n\t\tfilename := filepath.Join(r.dir, name)\n\t\tif err := os.MkdirAll(filepath.Dir(filename), 0777); err != nil {\n\t\t\treturn fmt.Errorf(\"error creating path %q: %v\", filepath.Dir(filename), err)\n\t\t}\n\t\tfile, err := os.Create(filename)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"error creating config file %s: %v\", filename, err)\n\t\t}\n\n\t\tdata := templateData{\n\t\t\tWorkingDir: r.dir,\n\t\t\tState: r.state,\n\t\t\tServiceUnits: r.serviceUnits,\n\t\t\tDefaultCertificate: r.defaultCertificatePath,\n\t\t\tDefaultDestinationCA: r.defaultDestinationCAPath,\n\t\t\tStatsUser: r.statsUser,\n\t\t\tStatsPassword: r.statsPassword,\n\t\t\tStatsPort: r.statsPort,\n\t\t\tBindPorts: !r.bindPortsAfterSync || r.synced,\n\t\t\tDynamicConfigManager: r.dynamicConfigManager,\n\t\t\tDisableHTTP2: disableHTTP2,\n\t\t\tCaptureHTTPRequestHeaders: r.captureHTTPRequestHeaders,\n\t\t\tCaptureHTTPResponseHeaders: r.captureHTTPResponseHeaders,\n\t\t\tCaptureHTTPCookie: r.captureHTTPCookie,\n\t\t\tHTTPHeaderNameCaseAdjustments: r.httpHeaderNameCaseAdjustments,\n\t\t\tHaveClientCA: r.haveClientCA,\n\t\t\tHaveCRLs: r.haveCRLs,\n\t\t\tHTTPResponseHeaders: r.httpResponseHeaders,\n\t\t\tHTTPRequestHeaders: r.httpRequestHeaders,\n\t\t}\n\t\tif err := template.Execute(file, data); err != nil {\n\t\t\tfile.Close()\n\t\t\treturn fmt.Errorf(\"error executing template for file %s: %v\", filename, err)\n\t\t}\n\t\tfile.Close()\n\t}\n\n\treturn nil\n}", "func WriteConfiguration(config *Config, checked, checkedIgnore bool) error {\n\tif config.IgnoreURL == \"\" {\n\t\tconfig.IgnoreURL = \"https://raw.githubusercontent.com/projectdiscovery/nuclei-templates/master/.nuclei-ignore\"\n\t}\n\tif checked {\n\t\tconfig.LastChecked = time.Now()\n\t}\n\tif checkedIgnore {\n\t\tconfig.LastCheckedIgnore = time.Now()\n\t}\n\tconfig.NucleiVersion = Version\n\n\ttemplatesConfigFile, err := getConfigDetails()\n\tif err != nil {\n\t\treturn err\n\t}\n\tfile, err := os.OpenFile(templatesConfigFile, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, 0777)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer file.Close()\n\n\terr = jsoniter.NewEncoder(file).Encode(config)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func WriteConfig(i interface{}, d string) error {\n\t// Convert to yaml\n\tym, err := yaml.Marshal(i)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Write to file\n\treturn ioutil.WriteFile(d, ym, 0755)\n}", "func (c *Config) WriteConfig(filename string) error {\n\tdata, err := yaml.Marshal(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = ioutil.WriteFile(filename, data, 0644)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (r *Resource) WriteConfig(filename string) error {\n\tr.Lock()\n\tdefer r.Unlock()\n\n\tvar b bytes.Buffer\n\n\tb.WriteString(fmt.Sprintf(\"# meta-data-json:{\\\"updated\\\": \\\"%s\\\"}\\n\", time.Now().UTC()))\n\tb.WriteString(fmt.Sprintf(\"resource %s {\\n\", r.name))\n\n\tvar hosts []string\n\tfor _, h := range r.host {\n\t\thosts = append(hosts, h.Name)\n\n\t\tb.WriteString(indentf(1, \"on %s {\\n\", h.Name))\n\t\tb.WriteString(indentf(2, \"node-id %d;\\n\", h.ID))\n\t\tb.WriteString(indentf(2, \"address %s:%d;\\n\", h.IP, r.port))\n\t\tfor _, v := range h.volume {\n\t\t\tb.WriteString(indentf(2, \"volume %d {\\n\", v.id))\n\t\t\tb.WriteString(indentf(3, \"device minor %d;\\n\", v.minor))\n\t\t\tb.WriteString(indentf(3, \"disk %s;\\n\", v.backingDevice))\n\t\t\tb.WriteString(indentf(3, \"meta-disk internal;\\n\"))\n\t\t\tb.WriteString(indentf(2, \"}\\n\")) // end volume section\n\t\t}\n\t\tb.WriteString(indentf(1, \"}\\n\")) // end on section\n\t\tb.WriteString(\"\\n\")\n\t}\n\n\tb.WriteString(indentf(1, \"connection-mesh {\\n\"))\n\tb.WriteString(indentf(2, \"hosts %s;\\n\", strings.Join(hosts, \" \")))\n\tb.WriteString(indentf(1, \"}\\n\"))\n\n\tb.WriteString(\"}\") // end resource section\n\n\treturn ioutil.WriteFile(filename, b.Bytes(), 0644)\n}", "func writeServiceTemplate(byter *bytes.Buffer, serviceConfig structs.ServiceConfig, deviceType, designation, deviceID string) error {\n\tenvMap, err := retrieveEnvironmentVariables(deviceType, designation)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor k, v := range serviceConfig.Data {\n\t\tbyter.WriteString(fmt.Sprintf(\"[%s]\\n\", k))\n\t\tfor key, value := range v {\n\t\t\tif isEnvironment := strings.Split(key, \"=\"); len(isEnvironment) == 2 {\n\t\t\t\tbyter.WriteString(fmt.Sprintf(\"%s=%s\\n\", key, serviceTemplateEnvSwap(value, envMap, deviceID)))\n\t\t\t} else {\n\t\t\t\tbyter.WriteString(fmt.Sprintf(\"%s=%s\\n\", key, value))\n\t\t\t}\n\t\t}\n\t\tbyter.WriteString(\"\\n\")\n\t}\n\treturn nil\n}", "func WriteConfigFile(configFilePath string, config *Configuration) {\n\tvar buffer bytes.Buffer\n\n\tif err := configTemplate.Execute(&buffer, config); err != nil {\n\t\tpanic(err)\n\t}\n\tcmn.MustWriteFile(configFilePath, buffer.Bytes(), 0644)\n}", "func (c *CommandWriter) WriteTemplate(data TemplateModel) error {\n\tif c.config.PreCommand != \"\" {\n\t\terr := preCheck(c.config, data)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"pre check failed\")\n\t\t}\n\t}\n\n\terr := write(c.config, data)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to write data\")\n\t}\n\n\tif c.config.PostCommand != \"\" {\n\t\terr = post(c.config.PostCommand)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"post command failed\")\n\t\t}\n\t}\n\treturn nil\n}", "func (h *HeaderfileWriter) WriteEndpointConfig(w io.Writer, e datapath.EndpointConfiguration) error {\n\tfw := bufio.NewWriter(w)\n\n\twriteIncludes(w)\n\th.writeStaticData(fw, e)\n\n\treturn h.writeTemplateConfig(fw, e)\n}", "func (fes *FrontEndService) WriteConfig(ctx context.Context, cancel context.CancelFunc) {\n\tfile, err := os.Create(fes.birdConfFile)\n\tif err != nil {\n\t\tlogrus.Fatalf(\"FrontEndService: failed to create %v, err: %v\", fes.birdConfFile, err)\n\t\tcancel()\n\t}\n\tdefer file.Close()\n\n\t//conf := \"include \\\"bird-common.conf\\\";\\n\"\n\t//conf += \"\\n\"\n\tconf := \"\"\n\tfes.WriteConfigBase(&conf)\n\thasVIP4, hasVIP6 := fes.WriteConfigVips(&conf)\n\tif len(fes.vrrps) > 0 {\n\t\tfes.WriteConfigVRRPs(&conf, hasVIP4, hasVIP6)\n\t} else if fes.dropIfNoPeer {\n\t\tfes.WriteConfigDropIfNoPeer(&conf, hasVIP4, hasVIP6)\n\t}\n\tfes.WriteConfigKernel(&conf, hasVIP4, hasVIP6)\n\tfes.WriteConfigBGP(&conf)\n\n\tlogrus.Infof(\"FrontEndService: config generated\")\n\tlogrus.Debugf(\"\\n%v\", conf)\n\t_, err = file.WriteString(conf)\n\tif err != nil {\n\t\tlogrus.Fatalf(\"FrontEndService: failed to write %v, err: %v\", fes.birdConfFile, err)\n\t\tcancel()\n\t}\n}", "func WriteConfig(f string, c *SupportedBranchesConfig) error {\n\treturn util.WithWriteFile(f, func(w io.Writer) error {\n\t\treturn EncodeConfig(w, c)\n\t})\n}", "func (fes *FrontEndService) WriteConfigVips(conf *string) (hasVIP4, hasVIP6 bool) {\n\tv4, v6 := \"\", \"\"\n\thasVIP4, hasVIP6 = false, false\n\n\tfor _, vip := range fes.vips {\n\t\tif isIPv6(vip) {\n\t\t\t// IPv6\n\t\t\t//v6 += \"\\troute \" + vip + \" blackhole;\\n\"\n\t\t\tv6 += \"\\troute \" + vip + \" via \\\"lo\\\";\\n\"\n\t\t} else if isIPv4(vip) {\n\t\t\t// IPv4\n\t\t\t//v4 += \"\\troute \" + vip + \" blackhole;\\n\"\n\t\t\tv4 += \"\\troute \" + vip + \" via \\\"lo\\\";\\n\"\n\t\t}\n\t}\n\n\tif v4 != \"\" {\n\t\thasVIP4 = true\n\t\t*conf += \"protocol static VIP4 {\\n\"\n\t\t*conf += \"\\tipv4 { preference 110; };\\n\"\n\t\t*conf += v4\n\t\t*conf += \"}\\n\"\n\t\t*conf += \"\\n\"\n\t}\n\n\tif v6 != \"\" {\n\t\thasVIP6 = true\n\t\t*conf += \"protocol static VIP6 {\\n\"\n\t\t*conf += \"\\tipv6 { preference 110; };\\n\"\n\t\t*conf += v6\n\t\t*conf += \"}\\n\"\n\t\t*conf += \"\\n\"\n\t}\n\treturn\n}", "func EncodeConfig(configFile io.Writer, c *Configuration) error {\n\t// Lock mutex\n\tConfig.rw.Lock()\n\tdefer Config.rw.Unlock()\n\n\t// Encode the given writer with the given interface\n\treturn toml.NewEncoder(configFile).Encode(c)\n}", "func (b *Bot) WriteConfig(fn configCallback) {\n\tfn(b.conf)\n}", "func WriteConfigFile(c *cli.Context, cfg *Config) error {\n\tfPath, err := TenetCfgPath(c)\n\tif err != nil {\n\t\treturn errors.Trace(err)\n\t}\n\tvar buf bytes.Buffer\n\tenc := toml.NewEncoder(&buf)\n\terr = enc.Encode(cfg)\n\tif err != nil {\n\t\treturn errors.Trace(err)\n\t}\n\treturn ioutil.WriteFile(fPath, buf.Bytes(), 0644)\n}", "func Write(w io.Writer, cfg *Config) error {\n\tif cfg.filename != \"\" {\n\t\tfmt.Fprintf(w, \"# Configuration file location: %v\\n\\n\", cfg.filename)\n\t} else {\n\t\tfmt.Fprintf(w, \"# Cound not find configuration file location.\\n\\n\")\n\t}\n\treturn toml.NewEncoder(w).Encode(cfg.File)\n}", "func (opt Options) PrintConfig(w io.Writer) error {\n\tt1, _ := template.New(\"webhook\").Funcs(sprig.TxtFuncMap()).Parse(webhook)\n\terr := t1.Execute(w, opt)\n\tif err != nil {\n\t\treturn err\n\t}\n\tt2, _ := template.New(\"csr\").Funcs(sprig.TxtFuncMap()).Parse(csr)\n\terr = t2.Execute(w, opt)\n\tif err != nil {\n\t\treturn err\n\t}\n\tt3, _ := template.New(\"deployment\").Funcs(sprig.TxtFuncMap()).Parse(webhookDeployment)\n\terr = t3.Execute(w, opt)\n\treturn err\n}", "func writeConfig(f Function) (err error) {\n\tpath := filepath.Join(f.Root, ConfigFile)\n\tc := toConfig(f)\n\tvar bb []byte\n\tif bb, err = yaml.Marshal(&c); err != nil {\n\t\treturn\n\t}\n\treturn ioutil.WriteFile(path, bb, 0644)\n}", "func WriteTemplate(path string, t *template.Template, funcs *map[string]Function) {\n\n\tvar buf bytes.Buffer\n\tif err := t.Execute(&buf, funcs); err != nil {\n\t\tlog.Fatalf(\"Cannot write template to buffer: %x\", err)\n\t}\n\tutils.WriteFile(path, buf.String())\n}", "func writeConfig() error {\n\tconfigFile := filepath.Join(config.ourBinaryDir, config.ourConfigFilename)\n\tlog.Printf(\"Writing YAML file: %s\", configFile)\n\tyamlText, err := yaml.Marshal(&config)\n\tif err != nil {\n\t\tlog.Printf(\"Couldn't generate YAML file: %s\", err)\n\t\treturn err\n\t}\n\terr = writeFileSafe(configFile, yamlText)\n\tif err != nil {\n\t\tlog.Printf(\"Couldn't save YAML config: %s\", err)\n\t\treturn err\n\t}\n\n\tuserFilter := getUserFilter()\n\terr = userFilter.save()\n\tif err != nil {\n\t\tlog.Printf(\"Couldn't save the user filter: %s\", err)\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func ConfigWrite() error {\n\n\t// Marshal it\n\tconfigJSON, _ := json.MarshalIndent(Config, \"\", \" \")\n\n\t// Write the file\n\tfd, err := os.OpenFile(configSettingsPath(), os.O_WRONLY|os.O_TRUNC|os.O_CREATE, 0666)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfd.Write(configJSON)\n\tfd.Close()\n\n\t// Done\n\treturn err\n\n}", "func (fes *FrontEndService) WriteConfigKernel(conf *string, hasVIP4 bool, hasVIP6 bool) {\n\teFilter := \"none\"\n\tif hasVIP4 {\n\t\teFilter = \"filter default_v4\"\n\t}\n\n\t*conf += \"protocol kernel {\\n\"\n\t*conf += \"\\tipv4 {\\n\"\n\t*conf += \"\\t\\timport none;\\n\"\n\t*conf += \"\\t\\texport \" + eFilter + \";\\n\"\n\t*conf += \"\\t};\\n\"\n\t*conf += \"\\tkernel table \" + strconv.FormatInt(int64(fes.kernelTableId), 10) + \";\\n\"\n\tif fes.ecmp {\n\t\t*conf += \"\\tmerge paths on;\\n\"\n\t}\n\tif fes.dropIfNoPeer {\n\t\t// Setting the metric for the default blackhole route must be supported,\n\t\t// which requires the kernel proto's metric to be set to zero.\n\t\t//\n\t\t// \"Metric 0 has a special meaning of undefined metric, in which either OS default is used,\n\t\t// or per-route metric can be set using krt_metric attribute. Default: 32. \"\n\t\t*conf += \"\\tmetric 0;\\n\"\n\t}\n\t*conf += \"}\\n\"\n\t*conf += \"\\n\"\n\n\tif hasVIP6 {\n\t\teFilter = \"filter default_v6\"\n\t} else {\n\t\teFilter = \"none\"\n\t}\n\t*conf += \"protocol kernel {\\n\"\n\t*conf += \"\\tipv6 {\\n\"\n\t*conf += \"\\t\\timport none;\\n\"\n\t*conf += \"\\t\\texport \" + eFilter + \";\\n\"\n\t*conf += \"\\t};\\n\"\n\t*conf += \"\\tkernel table \" + strconv.FormatInt(int64(fes.kernelTableId), 10) + \";\\n\"\n\tif fes.ecmp {\n\t\t*conf += \"\\tmerge paths on;\\n\"\n\t}\n\tif fes.dropIfNoPeer {\n\t\t// Setting the metric for the default blackhole route must be supported,\n\t\t// which requires the kernel proto's metric to be set to zero.\n\t\t//\n\t\t// \"Metric 0 has a special meaning of undefined metric, in which either OS default is used,\n\t\t// or per-route metric can be set using krt_metric attribute. Default: 32. \"\n\t\t*conf += \"\\tmetric 0;\\n\"\n\t}\n\t*conf += \"}\\n\"\n\t*conf += \"\\n\"\n}", "func WriteTemplate(w io.Writer, t *template.Template, v interface{}) {\n\tif err := t.Execute(w, v); err != nil {\n\t\tfmt.Fprintln(w, \"Error in template.\")\n\t\tkLog.Printf(\"Error in template: %v\\n\", err)\n\t}\n}", "func (f *FileConfigWriter) Write(config *auth.Config) error {\n\tif f.filename == \"\" {\n\t\treturn errors.New(\"No config file name defined\")\n\t}\n\tcontent, err := yaml.Marshal(config)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"marshaling the config to yaml\")\n\t}\n\terr = ioutil.WriteFile(f.filename, content, util.DefaultWritePermissions)\n\treturn nil\n}", "func WriteCredsConfig(cc *CredsConfig, fpath string) error {\n\tb, err := hclencoder.Encode(cc)\n\tif err == nil {\n\t\terr = ioutil.WriteFile(fpath, b, 0644)\n\t}\n\treturn err\n}", "func (c *Config) Write() error {\n\trawConfig, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = os.MkdirAll(configDirPath, 0755)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn ioutil.WriteFile(configPath, rawConfig, 0644)\n}", "func (h *HeaderfileWriter) WriteNodeConfig(w io.Writer, cfg *datapath.LocalNodeConfiguration) error {\n\textraMacrosMap := make(dpdef.Map)\n\tcDefinesMap := make(dpdef.Map)\n\n\tfw := bufio.NewWriter(w)\n\n\twriteIncludes(w)\n\n\trouterIP := node.GetIPv6Router()\n\thostIP := node.GetIPv6()\n\n\tfmt.Fprintf(fw, \"/*\\n\")\n\tif option.Config.EnableIPv6 {\n\t\tfmt.Fprintf(fw, \" cilium.v6.external.str %s\\n\", node.GetIPv6().String())\n\t\tfmt.Fprintf(fw, \" cilium.v6.internal.str %s\\n\", node.GetIPv6Router().String())\n\t\tfmt.Fprintf(fw, \" cilium.v6.nodeport.str %s\\n\", node.GetNodePortIPv6Addrs())\n\t\tfmt.Fprintf(fw, \"\\n\")\n\t}\n\tfmt.Fprintf(fw, \" cilium.v4.external.str %s\\n\", node.GetIPv4().String())\n\tfmt.Fprintf(fw, \" cilium.v4.internal.str %s\\n\", node.GetInternalIPv4Router().String())\n\tfmt.Fprintf(fw, \" cilium.v4.nodeport.str %s\\n\", node.GetNodePortIPv4Addrs())\n\tfmt.Fprintf(fw, \"\\n\")\n\tif option.Config.EnableIPv6 {\n\t\tfw.WriteString(dumpRaw(defaults.RestoreV6Addr, node.GetIPv6Router()))\n\t}\n\tfw.WriteString(dumpRaw(defaults.RestoreV4Addr, node.GetInternalIPv4Router()))\n\tfmt.Fprintf(fw, \" */\\n\\n\")\n\n\tcDefinesMap[\"KERNEL_HZ\"] = fmt.Sprintf(\"%d\", option.Config.KernelHz)\n\n\tif option.Config.EnableIPv6 {\n\t\textraMacrosMap[\"ROUTER_IP\"] = routerIP.String()\n\t\tfw.WriteString(defineIPv6(\"ROUTER_IP\", routerIP))\n\t}\n\n\tif option.Config.EnableIPv4 {\n\t\tipv4GW := node.GetInternalIPv4Router()\n\t\tloopbackIPv4 := node.GetIPv4Loopback()\n\t\tipv4Range := node.GetIPv4AllocRange()\n\t\tcDefinesMap[\"IPV4_GATEWAY\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(ipv4GW))\n\t\tcDefinesMap[\"IPV4_LOOPBACK\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(loopbackIPv4))\n\t\tcDefinesMap[\"IPV4_MASK\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(net.IP(ipv4Range.Mask)))\n\n\t\tif option.Config.EnableIPv4FragmentsTracking {\n\t\t\tcDefinesMap[\"ENABLE_IPV4_FRAGMENTS\"] = \"1\"\n\t\t\tcDefinesMap[\"IPV4_FRAG_DATAGRAMS_MAP\"] = fragmap.MapName\n\t\t\tcDefinesMap[\"CILIUM_IPV4_FRAG_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", option.Config.FragmentsMapEntries)\n\t\t}\n\t}\n\n\tif option.Config.EnableIPv6 {\n\t\textraMacrosMap[\"HOST_IP\"] = hostIP.String()\n\t\tfw.WriteString(defineIPv6(\"HOST_IP\", hostIP))\n\t}\n\n\tfor t, id := range tunnelProtocols {\n\t\tmacroName := fmt.Sprintf(\"TUNNEL_PROTOCOL_%s\", strings.ToUpper(t))\n\t\tcDefinesMap[macroName] = fmt.Sprintf(\"%d\", id)\n\t}\n\n\tencapProto := option.Config.TunnelProtocol\n\tif !option.Config.TunnelingEnabled() &&\n\t\toption.Config.EnableNodePort &&\n\t\toption.Config.NodePortMode != option.NodePortModeSNAT &&\n\t\toption.Config.LoadBalancerDSRDispatch == option.DSRDispatchGeneve {\n\t\tencapProto = option.TunnelGeneve\n\t}\n\n\tcDefinesMap[\"TUNNEL_PROTOCOL\"] = fmt.Sprintf(\"%d\", tunnelProtocols[encapProto])\n\tcDefinesMap[\"TUNNEL_PORT\"] = fmt.Sprintf(\"%d\", option.Config.TunnelPort)\n\n\tif tunnelDev, err := netlink.LinkByName(fmt.Sprintf(\"cilium_%s\", encapProto)); err == nil {\n\t\tcDefinesMap[\"ENCAP_IFINDEX\"] = fmt.Sprintf(\"%d\", tunnelDev.Attrs().Index)\n\t}\n\n\tcDefinesMap[\"HOST_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameHost))\n\tcDefinesMap[\"WORLD_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameWorld))\n\tif option.Config.IsDualStack() {\n\t\tcDefinesMap[\"WORLD_IPV4_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameWorldIPv4))\n\t\tcDefinesMap[\"WORLD_IPV6_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameWorldIPv6))\n\t} else {\n\t\tworldID := identity.GetReservedID(labels.IDNameWorld)\n\t\tcDefinesMap[\"WORLD_IPV4_ID\"] = fmt.Sprintf(\"%d\", worldID)\n\t\tcDefinesMap[\"WORLD_IPV6_ID\"] = fmt.Sprintf(\"%d\", worldID)\n\t}\n\tcDefinesMap[\"HEALTH_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameHealth))\n\tcDefinesMap[\"UNMANAGED_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameUnmanaged))\n\tcDefinesMap[\"INIT_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameInit))\n\tcDefinesMap[\"LOCAL_NODE_ID\"] = fmt.Sprintf(\"%d\", identity.GetLocalNodeID())\n\tcDefinesMap[\"REMOTE_NODE_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameRemoteNode))\n\tcDefinesMap[\"KUBE_APISERVER_NODE_ID\"] = fmt.Sprintf(\"%d\", identity.GetReservedID(labels.IDNameKubeAPIServer))\n\tcDefinesMap[\"CILIUM_LB_SERVICE_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.ServiceMapMaxEntries)\n\tcDefinesMap[\"CILIUM_LB_BACKENDS_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.ServiceBackEndMapMaxEntries)\n\tcDefinesMap[\"CILIUM_LB_REV_NAT_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.RevNatMapMaxEntries)\n\tcDefinesMap[\"CILIUM_LB_AFFINITY_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.AffinityMapMaxEntries)\n\tcDefinesMap[\"CILIUM_LB_SOURCE_RANGE_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.SourceRangeMapMaxEntries)\n\tcDefinesMap[\"CILIUM_LB_MAGLEV_MAP_MAX_ENTRIES\"] = fmt.Sprintf(\"%d\", lbmap.MaglevMapMaxEntries)\n\n\tcDefinesMap[\"TUNNEL_MAP\"] = tunnel.MapName\n\tcDefinesMap[\"TUNNEL_ENDPOINT_MAP_SIZE\"] = fmt.Sprintf(\"%d\", tunnel.MaxEntries)\n\tcDefinesMap[\"ENDPOINTS_MAP\"] = lxcmap.MapName\n\tcDefinesMap[\"ENDPOINTS_MAP_SIZE\"] = fmt.Sprintf(\"%d\", lxcmap.MaxEntries)\n\tcDefinesMap[\"METRICS_MAP\"] = metricsmap.MapName\n\tcDefinesMap[\"METRICS_MAP_SIZE\"] = fmt.Sprintf(\"%d\", metricsmap.MaxEntries)\n\tcDefinesMap[\"POLICY_MAP_SIZE\"] = fmt.Sprintf(\"%d\", policymap.MaxEntries)\n\tcDefinesMap[\"AUTH_MAP\"] = authmap.MapName\n\tcDefinesMap[\"AUTH_MAP_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.AuthMapEntries)\n\tcDefinesMap[\"CONFIG_MAP\"] = configmap.MapName\n\tcDefinesMap[\"CONFIG_MAP_SIZE\"] = fmt.Sprintf(\"%d\", configmap.MaxEntries)\n\tcDefinesMap[\"IPCACHE_MAP\"] = ipcachemap.Name\n\tcDefinesMap[\"IPCACHE_MAP_SIZE\"] = fmt.Sprintf(\"%d\", ipcachemap.MaxEntries)\n\tcDefinesMap[\"NODE_MAP\"] = nodemap.MapName\n\tcDefinesMap[\"NODE_MAP_SIZE\"] = fmt.Sprintf(\"%d\", nodemap.MaxEntries)\n\tcDefinesMap[\"SRV6_VRF_MAP4\"] = srv6map.VRFMapName4\n\tcDefinesMap[\"SRV6_VRF_MAP6\"] = srv6map.VRFMapName6\n\tcDefinesMap[\"SRV6_POLICY_MAP4\"] = srv6map.PolicyMapName4\n\tcDefinesMap[\"SRV6_POLICY_MAP6\"] = srv6map.PolicyMapName6\n\tcDefinesMap[\"SRV6_SID_MAP\"] = srv6map.SIDMapName\n\tcDefinesMap[\"SRV6_STATE_MAP4\"] = srv6map.StateMapName4\n\tcDefinesMap[\"SRV6_STATE_MAP6\"] = srv6map.StateMapName6\n\tcDefinesMap[\"SRV6_VRF_MAP_SIZE\"] = fmt.Sprintf(\"%d\", srv6map.MaxVRFEntries)\n\tcDefinesMap[\"SRV6_POLICY_MAP_SIZE\"] = fmt.Sprintf(\"%d\", srv6map.MaxPolicyEntries)\n\tcDefinesMap[\"SRV6_SID_MAP_SIZE\"] = fmt.Sprintf(\"%d\", srv6map.MaxSIDEntries)\n\tcDefinesMap[\"SRV6_STATE_MAP_SIZE\"] = fmt.Sprintf(\"%d\", srv6map.MaxStateEntries)\n\tcDefinesMap[\"WORLD_CIDRS4_MAP\"] = worldcidrsmap.MapName4\n\tcDefinesMap[\"WORLD_CIDRS4_MAP_SIZE\"] = fmt.Sprintf(\"%d\", worldcidrsmap.MapMaxEntries)\n\tcDefinesMap[\"POLICY_PROG_MAP_SIZE\"] = fmt.Sprintf(\"%d\", policymap.PolicyCallMaxEntries)\n\tcDefinesMap[\"L2_RESPONSER_MAP4_SIZE\"] = fmt.Sprintf(\"%d\", l2respondermap.DefaultMaxEntries)\n\tcDefinesMap[\"ENCRYPT_MAP\"] = encrypt.MapName\n\tcDefinesMap[\"L2_RESPONDER_MAP4\"] = l2respondermap.MapName\n\tcDefinesMap[\"CT_CONNECTION_LIFETIME_TCP\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutTCP.Seconds()))\n\tcDefinesMap[\"CT_CONNECTION_LIFETIME_NONTCP\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutAny.Seconds()))\n\tcDefinesMap[\"CT_SERVICE_LIFETIME_TCP\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutSVCTCP.Seconds()))\n\tcDefinesMap[\"CT_SERVICE_LIFETIME_NONTCP\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutSVCAny.Seconds()))\n\tcDefinesMap[\"CT_SERVICE_CLOSE_REBALANCE\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutSVCTCPGrace.Seconds()))\n\tcDefinesMap[\"CT_SYN_TIMEOUT\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutSYN.Seconds()))\n\tcDefinesMap[\"CT_CLOSE_TIMEOUT\"] = fmt.Sprintf(\"%d\", int64(option.Config.CTMapEntriesTimeoutFIN.Seconds()))\n\tcDefinesMap[\"CT_REPORT_INTERVAL\"] = fmt.Sprintf(\"%d\", int64(option.Config.MonitorAggregationInterval.Seconds()))\n\tcDefinesMap[\"CT_REPORT_FLAGS\"] = fmt.Sprintf(\"%#04x\", int64(option.Config.MonitorAggregationFlags))\n\tcDefinesMap[\"CT_TAIL_CALL_BUFFER4\"] = \"cilium_tail_call_buffer4\"\n\tcDefinesMap[\"CT_TAIL_CALL_BUFFER6\"] = \"cilium_tail_call_buffer6\"\n\tcDefinesMap[\"PER_CLUSTER_CT_TCP4\"] = \"cilium_per_cluster_ct_tcp4\"\n\tcDefinesMap[\"PER_CLUSTER_CT_TCP6\"] = \"cilium_per_cluster_ct_tcp6\"\n\tcDefinesMap[\"PER_CLUSTER_CT_ANY4\"] = \"cilium_per_cluster_ct_any4\"\n\tcDefinesMap[\"PER_CLUSTER_CT_ANY6\"] = \"cilium_per_cluster_ct_any6\"\n\tcDefinesMap[\"PER_CLUSTER_SNAT_MAPPING_IPV4\"] = \"cilium_per_cluster_snat_v4_external\"\n\tcDefinesMap[\"PER_CLUSTER_SNAT_MAPPING_IPV6\"] = \"cilium_per_cluster_snat_v6_external\"\n\n\tif option.Config.PreAllocateMaps {\n\t\tcDefinesMap[\"PREALLOCATE_MAPS\"] = \"1\"\n\t}\n\n\tcDefinesMap[\"EVENTS_MAP\"] = eventsmap.MapName\n\tcDefinesMap[\"SIGNAL_MAP\"] = signalmap.MapName\n\tcDefinesMap[\"POLICY_CALL_MAP\"] = policymap.PolicyCallMapName\n\tif option.Config.EnableEnvoyConfig {\n\t\tcDefinesMap[\"POLICY_EGRESSCALL_MAP\"] = policymap.PolicyEgressCallMapName\n\t}\n\tcDefinesMap[\"LB6_REVERSE_NAT_MAP\"] = \"cilium_lb6_reverse_nat\"\n\tcDefinesMap[\"LB6_SERVICES_MAP_V2\"] = \"cilium_lb6_services_v2\"\n\tcDefinesMap[\"LB6_BACKEND_MAP\"] = \"cilium_lb6_backends_v3\"\n\tcDefinesMap[\"LB6_REVERSE_NAT_SK_MAP\"] = lbmap.SockRevNat6MapName\n\tcDefinesMap[\"LB6_REVERSE_NAT_SK_MAP_SIZE\"] = fmt.Sprintf(\"%d\", lbmap.MaxSockRevNat6MapEntries)\n\tcDefinesMap[\"LB4_REVERSE_NAT_MAP\"] = \"cilium_lb4_reverse_nat\"\n\tcDefinesMap[\"LB4_SERVICES_MAP_V2\"] = \"cilium_lb4_services_v2\"\n\tcDefinesMap[\"LB4_BACKEND_MAP\"] = \"cilium_lb4_backends_v3\"\n\tcDefinesMap[\"LB4_REVERSE_NAT_SK_MAP\"] = lbmap.SockRevNat4MapName\n\tcDefinesMap[\"LB4_REVERSE_NAT_SK_MAP_SIZE\"] = fmt.Sprintf(\"%d\", lbmap.MaxSockRevNat4MapEntries)\n\n\tif option.Config.EnableSessionAffinity {\n\t\tcDefinesMap[\"ENABLE_SESSION_AFFINITY\"] = \"1\"\n\t\tcDefinesMap[\"LB_AFFINITY_MATCH_MAP\"] = lbmap.AffinityMatchMapName\n\t\tif option.Config.EnableIPv4 {\n\t\t\tcDefinesMap[\"LB4_AFFINITY_MAP\"] = lbmap.Affinity4MapName\n\t\t}\n\t\tif option.Config.EnableIPv6 {\n\t\t\tcDefinesMap[\"LB6_AFFINITY_MAP\"] = lbmap.Affinity6MapName\n\t\t}\n\t}\n\n\tcDefinesMap[\"TRACE_PAYLOAD_LEN\"] = fmt.Sprintf(\"%dULL\", option.Config.TracePayloadlen)\n\tcDefinesMap[\"MTU\"] = fmt.Sprintf(\"%d\", cfg.MtuConfig.GetDeviceMTU())\n\n\tif option.Config.EnableIPv4 {\n\t\tcDefinesMap[\"ENABLE_IPV4\"] = \"1\"\n\t}\n\n\tif option.Config.EnableIPv6 {\n\t\tcDefinesMap[\"ENABLE_IPV6\"] = \"1\"\n\t}\n\n\tif option.Config.EnableSRv6 {\n\t\tcDefinesMap[\"ENABLE_SRV6\"] = \"1\"\n\t\tif option.Config.SRv6EncapMode != \"reduced\" {\n\t\t\tcDefinesMap[\"ENABLE_SRV6_SRH_ENCAP\"] = \"1\"\n\t\t}\n\t}\n\n\tif option.Config.EnableSCTP {\n\t\tcDefinesMap[\"ENABLE_SCTP\"] = \"1\"\n\t}\n\n\tif option.Config.EnableIPSec {\n\t\tcDefinesMap[\"ENABLE_IPSEC\"] = \"1\"\n\t}\n\n\tif option.Config.EnableWireguard {\n\t\tcDefinesMap[\"ENABLE_WIREGUARD\"] = \"1\"\n\t\tifindex, err := link.GetIfIndex(wgtypes.IfaceName)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcDefinesMap[\"WG_IFINDEX\"] = fmt.Sprintf(\"%d\", ifindex)\n\n\t\tif option.Config.EncryptNode {\n\t\t\tcDefinesMap[\"ENABLE_NODE_ENCRYPTION\"] = \"1\"\n\t\t}\n\t}\n\n\tif option.Config.EnableL2Announcements {\n\t\tcDefinesMap[\"ENABLE_L2_ANNOUNCEMENTS\"] = \"1\"\n\t\t// If the agent is down for longer than the lease duration, stop responding\n\t\tcDefinesMap[\"L2_ANNOUNCEMENTS_MAX_LIVENESS\"] = fmt.Sprintf(\"%dULL\", option.Config.L2AnnouncerLeaseDuration.Nanoseconds())\n\t}\n\n\tif option.Config.EnableEncryptionStrictMode {\n\t\tcDefinesMap[\"ENCRYPTION_STRICT_MODE\"] = \"1\"\n\n\t\t// when parsing the user input we only accept ipv4 addresses\n\t\tcDefinesMap[\"STRICT_IPV4_NET\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPAddrToHost32(option.Config.EncryptionStrictModeCIDR.Addr()))\n\t\tcDefinesMap[\"STRICT_IPV4_NET_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.EncryptionStrictModeCIDR.Bits())\n\n\t\tcDefinesMap[\"IPV4_ENCRYPT_IFACE\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(node.GetIPv4()))\n\n\t\tipv4Interface, ok := netip.AddrFromSlice(node.GetIPv4().To4())\n\t\tif !ok {\n\t\t\treturn fmt.Errorf(\"unable to parse node IPv4 address %s\", node.GetIPv4())\n\t\t}\n\n\t\tif option.Config.EncryptionStrictModeCIDR.Contains(ipv4Interface) {\n\t\t\tif !option.Config.EncryptionStrictModeAllowRemoteNodeIdentities {\n\t\t\t\treturn fmt.Errorf(`encryption strict mode is enabled but the node's IPv4 address is within the strict CIDR range.\n\t\t\t\tThis will cause the node to drop all traffic.\n\t\t\t\tPlease either disable encryption or set --encryption-strict-mode-allow-dynamic-lookup=true`)\n\t\t\t}\n\t\t\tcDefinesMap[\"STRICT_IPV4_OVERLAPPING_CIDR\"] = \"1\"\n\t\t}\n\t}\n\n\tif option.Config.EnableBPFTProxy {\n\t\tcDefinesMap[\"ENABLE_TPROXY\"] = \"1\"\n\t}\n\n\tif option.Config.EnableXDPPrefilter {\n\t\tcDefinesMap[\"ENABLE_PREFILTER\"] = \"1\"\n\t}\n\n\tif option.Config.EnableEndpointRoutes {\n\t\tcDefinesMap[\"ENABLE_ENDPOINT_ROUTES\"] = \"1\"\n\t}\n\n\tif option.Config.EnableEnvoyConfig {\n\t\tcDefinesMap[\"ENABLE_L7_LB\"] = \"1\"\n\t}\n\n\tif option.Config.EnableSocketLB {\n\t\tif option.Config.BPFSocketLBHostnsOnly {\n\t\t\tcDefinesMap[\"ENABLE_SOCKET_LB_HOST_ONLY\"] = \"1\"\n\t\t} else {\n\t\t\tcDefinesMap[\"ENABLE_SOCKET_LB_FULL\"] = \"1\"\n\t\t}\n\t\tif option.Config.EnableSocketLBPeer {\n\t\t\tcDefinesMap[\"ENABLE_SOCKET_LB_PEER\"] = \"1\"\n\t\t}\n\t\tif option.Config.EnableSocketLBTracing {\n\t\t\tcDefinesMap[\"TRACE_SOCK_NOTIFY\"] = \"1\"\n\t\t}\n\n\t\tif cookie, err := netns.GetNetNSCookie(); err == nil {\n\t\t\t// When running in nested environments (e.g. Kind), cilium-agent does\n\t\t\t// not run in the host netns. So, in such cases the cookie comparison\n\t\t\t// based on bpf_get_netns_cookie(NULL) for checking whether a socket\n\t\t\t// belongs to a host netns does not work.\n\t\t\t//\n\t\t\t// To fix this, we derive the cookie of the netns in which cilium-agent\n\t\t\t// runs via getsockopt(...SO_NETNS_COOKIE...) and then use it in the\n\t\t\t// check above. This is based on an assumption that cilium-agent\n\t\t\t// always runs with \"hostNetwork: true\".\n\t\t\tcDefinesMap[\"HOST_NETNS_COOKIE\"] = fmt.Sprintf(\"%d\", cookie)\n\t\t}\n\t}\n\n\tcDefinesMap[\"NAT_46X64_PREFIX_0\"] = \"0\"\n\tcDefinesMap[\"NAT_46X64_PREFIX_1\"] = \"0\"\n\tcDefinesMap[\"NAT_46X64_PREFIX_2\"] = \"0\"\n\tcDefinesMap[\"NAT_46X64_PREFIX_3\"] = \"0\"\n\n\tif option.Config.EnableNodePort {\n\t\tif option.Config.EnableHealthDatapath {\n\t\t\tcDefinesMap[\"ENABLE_HEALTH_CHECK\"] = \"1\"\n\t\t}\n\t\tif option.Config.EnableMKE && option.Config.EnableSocketLB {\n\t\t\tcDefinesMap[\"ENABLE_MKE\"] = \"1\"\n\t\t\tcDefinesMap[\"MKE_HOST\"] = fmt.Sprintf(\"%d\", option.HostExtensionMKE)\n\t\t}\n\t\tif option.Config.EnableRecorder {\n\t\t\tcDefinesMap[\"ENABLE_CAPTURE\"] = \"1\"\n\t\t\tif option.Config.EnableIPv4 {\n\t\t\t\tcDefinesMap[\"CAPTURE4_RULES\"] = recorder.MapNameWcard4\n\t\t\t\tcDefinesMap[\"CAPTURE4_SIZE\"] = fmt.Sprintf(\"%d\", recorder.MapSize)\n\t\t\t}\n\t\t\tif option.Config.EnableIPv6 {\n\t\t\t\tcDefinesMap[\"CAPTURE6_RULES\"] = recorder.MapNameWcard6\n\t\t\t\tcDefinesMap[\"CAPTURE6_SIZE\"] = fmt.Sprintf(\"%d\", recorder.MapSize)\n\t\t\t}\n\t\t}\n\t\tcDefinesMap[\"ENABLE_NODEPORT\"] = \"1\"\n\t\tif option.Config.EnableIPv4 {\n\t\t\tcDefinesMap[\"NODEPORT_NEIGH4\"] = neighborsmap.Map4Name\n\t\t\tcDefinesMap[\"NODEPORT_NEIGH4_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.NeighMapEntriesGlobal)\n\t\t\tif option.Config.EnableHealthDatapath {\n\t\t\t\tcDefinesMap[\"LB4_HEALTH_MAP\"] = lbmap.HealthProbe4MapName\n\t\t\t}\n\t\t}\n\t\tif option.Config.EnableIPv6 {\n\t\t\tcDefinesMap[\"NODEPORT_NEIGH6\"] = neighborsmap.Map6Name\n\t\t\tcDefinesMap[\"NODEPORT_NEIGH6_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.NeighMapEntriesGlobal)\n\t\t\tif option.Config.EnableHealthDatapath {\n\t\t\t\tcDefinesMap[\"LB6_HEALTH_MAP\"] = lbmap.HealthProbe6MapName\n\t\t\t}\n\t\t}\n\t\tif option.Config.EnableNat46X64Gateway {\n\t\t\tcDefinesMap[\"ENABLE_NAT_46X64_GATEWAY\"] = \"1\"\n\t\t\tbase := option.Config.IPv6NAT46x64CIDRBase.AsSlice()\n\t\t\tcDefinesMap[\"NAT_46X64_PREFIX_0\"] = fmt.Sprintf(\"%d\", base[0])\n\t\t\tcDefinesMap[\"NAT_46X64_PREFIX_1\"] = fmt.Sprintf(\"%d\", base[1])\n\t\t\tcDefinesMap[\"NAT_46X64_PREFIX_2\"] = fmt.Sprintf(\"%d\", base[2])\n\t\t\tcDefinesMap[\"NAT_46X64_PREFIX_3\"] = fmt.Sprintf(\"%d\", base[3])\n\t\t}\n\t\tif option.Config.NodePortNat46X64 {\n\t\t\tcDefinesMap[\"ENABLE_NAT_46X64\"] = \"1\"\n\t\t}\n\t\tconst (\n\t\t\tdsrEncapInv = iota\n\t\t\tdsrEncapNone\n\t\t\tdsrEncapIPIP\n\t\t\tdsrEncapGeneve\n\t\t)\n\t\tconst (\n\t\t\tdsrL4XlateInv = iota\n\t\t\tdsrL4XlateFrontend\n\t\t\tdsrL4XlateBackend\n\t\t)\n\t\tcDefinesMap[\"DSR_ENCAP_IPIP\"] = fmt.Sprintf(\"%d\", dsrEncapIPIP)\n\t\tcDefinesMap[\"DSR_ENCAP_GENEVE\"] = fmt.Sprintf(\"%d\", dsrEncapGeneve)\n\t\tcDefinesMap[\"DSR_ENCAP_NONE\"] = fmt.Sprintf(\"%d\", dsrEncapNone)\n\t\tcDefinesMap[\"DSR_XLATE_FRONTEND\"] = fmt.Sprintf(\"%d\", dsrL4XlateFrontend)\n\t\tcDefinesMap[\"DSR_XLATE_BACKEND\"] = fmt.Sprintf(\"%d\", dsrL4XlateBackend)\n\t\tif option.Config.NodePortMode == option.NodePortModeDSR ||\n\t\t\toption.Config.NodePortMode == option.NodePortModeHybrid {\n\t\t\tcDefinesMap[\"ENABLE_DSR\"] = \"1\"\n\t\t\tif option.Config.EnablePMTUDiscovery {\n\t\t\t\tcDefinesMap[\"ENABLE_DSR_ICMP_ERRORS\"] = \"1\"\n\t\t\t}\n\t\t\tif option.Config.NodePortMode == option.NodePortModeHybrid {\n\t\t\t\tcDefinesMap[\"ENABLE_DSR_HYBRID\"] = \"1\"\n\t\t\t}\n\t\t\tif option.Config.LoadBalancerDSRDispatch == option.DSRDispatchOption {\n\t\t\t\tcDefinesMap[\"DSR_ENCAP_MODE\"] = fmt.Sprintf(\"%d\", dsrEncapNone)\n\t\t\t} else if option.Config.LoadBalancerDSRDispatch == option.DSRDispatchIPIP {\n\t\t\t\tcDefinesMap[\"DSR_ENCAP_MODE\"] = fmt.Sprintf(\"%d\", dsrEncapIPIP)\n\t\t\t} else if option.Config.LoadBalancerDSRDispatch == option.DSRDispatchGeneve {\n\t\t\t\tcDefinesMap[\"DSR_ENCAP_MODE\"] = fmt.Sprintf(\"%d\", dsrEncapGeneve)\n\t\t\t}\n\t\t\tif option.Config.LoadBalancerDSRDispatch == option.DSRDispatchIPIP {\n\t\t\t\tif option.Config.LoadBalancerDSRL4Xlate == option.DSRL4XlateFrontend {\n\t\t\t\t\tcDefinesMap[\"DSR_XLATE_MODE\"] = fmt.Sprintf(\"%d\", dsrL4XlateFrontend)\n\t\t\t\t} else if option.Config.LoadBalancerDSRL4Xlate == option.DSRL4XlateBackend {\n\t\t\t\t\tcDefinesMap[\"DSR_XLATE_MODE\"] = fmt.Sprintf(\"%d\", dsrL4XlateBackend)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tcDefinesMap[\"DSR_XLATE_MODE\"] = fmt.Sprintf(\"%d\", dsrL4XlateInv)\n\t\t\t}\n\t\t} else {\n\t\t\tcDefinesMap[\"DSR_ENCAP_MODE\"] = fmt.Sprintf(\"%d\", dsrEncapInv)\n\t\t\tcDefinesMap[\"DSR_XLATE_MODE\"] = fmt.Sprintf(\"%d\", dsrL4XlateInv)\n\t\t}\n\t\tif option.Config.EnableIPv4 {\n\t\t\tif option.Config.LoadBalancerRSSv4CIDR != \"\" {\n\t\t\t\tipv4 := byteorder.NetIPv4ToHost32(option.Config.LoadBalancerRSSv4.IP)\n\t\t\t\tones, _ := option.Config.LoadBalancerRSSv4.Mask.Size()\n\t\t\t\tcDefinesMap[\"IPV4_RSS_PREFIX\"] = fmt.Sprintf(\"%d\", ipv4)\n\t\t\t\tcDefinesMap[\"IPV4_RSS_PREFIX_BITS\"] = fmt.Sprintf(\"%d\", ones)\n\t\t\t} else {\n\t\t\t\tcDefinesMap[\"IPV4_RSS_PREFIX\"] = \"IPV4_DIRECT_ROUTING\"\n\t\t\t\tcDefinesMap[\"IPV4_RSS_PREFIX_BITS\"] = \"32\"\n\t\t\t}\n\t\t}\n\t\tif option.Config.EnableIPv6 {\n\t\t\tif option.Config.LoadBalancerRSSv6CIDR != \"\" {\n\t\t\t\tipv6 := option.Config.LoadBalancerRSSv6.IP\n\t\t\t\tones, _ := option.Config.LoadBalancerRSSv6.Mask.Size()\n\t\t\t\textraMacrosMap[\"IPV6_RSS_PREFIX\"] = ipv6.String()\n\t\t\t\tfw.WriteString(FmtDefineAddress(\"IPV6_RSS_PREFIX\", ipv6))\n\t\t\t\tcDefinesMap[\"IPV6_RSS_PREFIX_BITS\"] = fmt.Sprintf(\"%d\", ones)\n\t\t\t} else {\n\t\t\t\tcDefinesMap[\"IPV6_RSS_PREFIX\"] = \"IPV6_DIRECT_ROUTING\"\n\t\t\t\tcDefinesMap[\"IPV6_RSS_PREFIX_BITS\"] = \"128\"\n\t\t\t}\n\t\t}\n\t\tif option.Config.NodePortAcceleration != option.NodePortAccelerationDisabled {\n\t\t\tcDefinesMap[\"ENABLE_NODEPORT_ACCELERATION\"] = \"1\"\n\t\t}\n\t\tif !option.Config.EnableHostLegacyRouting {\n\t\t\tcDefinesMap[\"ENABLE_HOST_ROUTING\"] = \"1\"\n\t\t}\n\t\tif option.Config.EnableSVCSourceRangeCheck {\n\t\t\tcDefinesMap[\"ENABLE_SRC_RANGE_CHECK\"] = \"1\"\n\t\t\tif option.Config.EnableIPv4 {\n\t\t\t\tcDefinesMap[\"LB4_SRC_RANGE_MAP\"] = lbmap.SourceRange4MapName\n\t\t\t\tcDefinesMap[\"LB4_SRC_RANGE_MAP_SIZE\"] =\n\t\t\t\t\tfmt.Sprintf(\"%d\", lbmap.SourceRange4Map.MaxEntries())\n\t\t\t}\n\t\t\tif option.Config.EnableIPv6 {\n\t\t\t\tcDefinesMap[\"LB6_SRC_RANGE_MAP\"] = lbmap.SourceRange6MapName\n\t\t\t\tcDefinesMap[\"LB6_SRC_RANGE_MAP_SIZE\"] =\n\t\t\t\t\tfmt.Sprintf(\"%d\", lbmap.SourceRange6Map.MaxEntries())\n\t\t\t}\n\t\t}\n\n\t\tcDefinesMap[\"NODEPORT_PORT_MIN\"] = fmt.Sprintf(\"%d\", option.Config.NodePortMin)\n\t\tcDefinesMap[\"NODEPORT_PORT_MAX\"] = fmt.Sprintf(\"%d\", option.Config.NodePortMax)\n\t\tcDefinesMap[\"NODEPORT_PORT_MIN_NAT\"] = fmt.Sprintf(\"%d\", option.Config.NodePortMax+1)\n\t\tcDefinesMap[\"NODEPORT_PORT_MAX_NAT\"] = \"65535\"\n\t}\n\n\tmacByIfIndexMacro, isL3DevMacro, err := devMacros()\n\tif err != nil {\n\t\treturn err\n\t}\n\tcDefinesMap[\"NATIVE_DEV_MAC_BY_IFINDEX(IFINDEX)\"] = macByIfIndexMacro\n\tcDefinesMap[\"IS_L3_DEV(ifindex)\"] = isL3DevMacro\n\n\tconst (\n\t\tselectionRandom = iota + 1\n\t\tselectionMaglev\n\t)\n\tcDefinesMap[\"LB_SELECTION_RANDOM\"] = fmt.Sprintf(\"%d\", selectionRandom)\n\tcDefinesMap[\"LB_SELECTION_MAGLEV\"] = fmt.Sprintf(\"%d\", selectionMaglev)\n\tif option.Config.NodePortAlg == option.NodePortAlgRandom {\n\t\tcDefinesMap[\"LB_SELECTION\"] = fmt.Sprintf(\"%d\", selectionRandom)\n\t} else if option.Config.NodePortAlg == option.NodePortAlgMaglev {\n\t\tcDefinesMap[\"LB_SELECTION\"] = fmt.Sprintf(\"%d\", selectionMaglev)\n\t\tcDefinesMap[\"LB_MAGLEV_LUT_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.MaglevTableSize)\n\t\tif option.Config.EnableIPv6 {\n\t\t\tcDefinesMap[\"LB6_MAGLEV_MAP_OUTER\"] = lbmap.MaglevOuter6MapName\n\t\t}\n\t\tif option.Config.EnableIPv4 {\n\t\t\tcDefinesMap[\"LB4_MAGLEV_MAP_OUTER\"] = lbmap.MaglevOuter4MapName\n\t\t}\n\t}\n\tcDefinesMap[\"HASH_INIT4_SEED\"] = fmt.Sprintf(\"%d\", maglev.SeedJhash0)\n\tcDefinesMap[\"HASH_INIT6_SEED\"] = fmt.Sprintf(\"%d\", maglev.SeedJhash1)\n\n\tif option.Config.DirectRoutingDeviceRequired() {\n\t\tdirectRoutingIface := option.Config.DirectRoutingDevice\n\t\tdirectRoutingIfIndex, err := link.GetIfIndex(directRoutingIface)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcDefinesMap[\"DIRECT_ROUTING_DEV_IFINDEX\"] = fmt.Sprintf(\"%d\", directRoutingIfIndex)\n\t\tif option.Config.EnableIPv4 {\n\t\t\tip, ok := node.GetNodePortIPv4AddrsWithDevices()[directRoutingIface]\n\t\t\tif !ok {\n\t\t\t\tlog.WithFields(logrus.Fields{\n\t\t\t\t\t\"directRoutingIface\": directRoutingIface,\n\t\t\t\t}).Fatal(\"Direct routing device's IPv4 address not found\")\n\t\t\t}\n\n\t\t\tipv4 := byteorder.NetIPv4ToHost32(ip)\n\t\t\tcDefinesMap[\"IPV4_DIRECT_ROUTING\"] = fmt.Sprintf(\"%d\", ipv4)\n\t\t}\n\n\t\tif option.Config.EnableIPv6 {\n\t\t\tdirectRoutingIPv6, ok := node.GetNodePortIPv6AddrsWithDevices()[directRoutingIface]\n\t\t\tif !ok {\n\t\t\t\tlog.WithFields(logrus.Fields{\n\t\t\t\t\t\"directRoutingIface\": directRoutingIface,\n\t\t\t\t}).Fatal(\"Direct routing device's IPv6 address not found\")\n\t\t\t}\n\n\t\t\textraMacrosMap[\"IPV6_DIRECT_ROUTING\"] = directRoutingIPv6.String()\n\t\t\tfw.WriteString(FmtDefineAddress(\"IPV6_DIRECT_ROUTING\", directRoutingIPv6))\n\t\t}\n\t} else {\n\t\tvar directRoutingIPv6 net.IP\n\t\tcDefinesMap[\"DIRECT_ROUTING_DEV_IFINDEX\"] = \"0\"\n\t\tif option.Config.EnableIPv4 {\n\t\t\tcDefinesMap[\"IPV4_DIRECT_ROUTING\"] = \"0\"\n\t\t}\n\t\tif option.Config.EnableIPv6 {\n\t\t\textraMacrosMap[\"IPV6_DIRECT_ROUTING\"] = directRoutingIPv6.String()\n\t\t\tfw.WriteString(FmtDefineAddress(\"IPV6_DIRECT_ROUTING\", directRoutingIPv6))\n\t\t}\n\t}\n\n\tif option.Config.ResetQueueMapping {\n\t\tcDefinesMap[\"RESET_QUEUES\"] = \"1\"\n\t}\n\n\tif option.Config.EnableBandwidthManager {\n\t\tcDefinesMap[\"ENABLE_BANDWIDTH_MANAGER\"] = \"1\"\n\t\tcDefinesMap[\"THROTTLE_MAP\"] = bwmap.MapName\n\t\tcDefinesMap[\"THROTTLE_MAP_SIZE\"] = fmt.Sprintf(\"%d\", bwmap.MapSize)\n\t}\n\n\tif option.Config.EnableHostFirewall {\n\t\tcDefinesMap[\"ENABLE_HOST_FIREWALL\"] = \"1\"\n\t}\n\n\tif option.Config.EnableIPSec {\n\t\tnodeAddress := node.GetIPv4()\n\t\tif nodeAddress == nil {\n\t\t\treturn errors.New(\"external IPv4 node address is required when IPSec is enabled, but none found\")\n\t\t}\n\n\t\ta := byteorder.NetIPv4ToHost32(nodeAddress)\n\t\tcDefinesMap[\"IPV4_ENCRYPT_IFACE\"] = fmt.Sprintf(\"%d\", a)\n\t\tif iface := option.Config.EncryptInterface; len(iface) != 0 {\n\t\t\tlink, err := netlink.LinkByName(iface[0])\n\t\t\tif err == nil {\n\t\t\t\tcDefinesMap[\"ENCRYPT_IFACE\"] = fmt.Sprintf(\"%d\", link.Attrs().Index)\n\t\t\t}\n\t\t}\n\t}\n\n\tif option.Config.EnableNodePort {\n\t\tif option.Config.EnableIPv4 {\n\t\t\tcDefinesMap[\"SNAT_MAPPING_IPV4\"] = nat.MapNameSnat4Global\n\t\t\tcDefinesMap[\"SNAT_MAPPING_IPV4_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.NATMapEntriesGlobal)\n\t\t}\n\n\t\tif option.Config.EnableIPv6 {\n\t\t\tcDefinesMap[\"SNAT_MAPPING_IPV6\"] = nat.MapNameSnat6Global\n\t\t\tcDefinesMap[\"SNAT_MAPPING_IPV6_SIZE\"] = fmt.Sprintf(\"%d\", option.Config.NATMapEntriesGlobal)\n\t\t}\n\n\t\tif option.Config.EnableBPFMasquerade {\n\t\t\tif option.Config.EnableIPv4Masquerade {\n\t\t\t\tcDefinesMap[\"ENABLE_MASQUERADE_IPV4\"] = \"1\"\n\n\t\t\t\t// ip-masq-agent depends on bpf-masq\n\t\t\t\tvar excludeCIDR *cidr.CIDR\n\t\t\t\tif option.Config.EnableIPMasqAgent {\n\t\t\t\t\tcDefinesMap[\"ENABLE_IP_MASQ_AGENT_IPV4\"] = \"1\"\n\t\t\t\t\tcDefinesMap[\"IP_MASQ_AGENT_IPV4\"] = ipmasq.MapNameIPv4\n\n\t\t\t\t\t// native-routing-cidr is optional with ip-masq-agent and may be nil\n\t\t\t\t\texcludeCIDR = option.Config.GetIPv4NativeRoutingCIDR()\n\t\t\t\t} else {\n\t\t\t\t\texcludeCIDR = datapath.RemoteSNATDstAddrExclusionCIDRv4()\n\t\t\t\t}\n\n\t\t\t\tif excludeCIDR != nil {\n\t\t\t\t\tcDefinesMap[\"IPV4_SNAT_EXCLUSION_DST_CIDR\"] =\n\t\t\t\t\t\tfmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(excludeCIDR.IP))\n\t\t\t\t\tones, _ := excludeCIDR.Mask.Size()\n\t\t\t\t\tcDefinesMap[\"IPV4_SNAT_EXCLUSION_DST_CIDR_LEN\"] = fmt.Sprintf(\"%d\", ones)\n\t\t\t\t}\n\t\t\t}\n\t\t\tif option.Config.EnableIPv6Masquerade {\n\t\t\t\tcDefinesMap[\"ENABLE_MASQUERADE_IPV6\"] = \"1\"\n\n\t\t\t\tvar excludeCIDR *cidr.CIDR\n\t\t\t\tif option.Config.EnableIPMasqAgent {\n\t\t\t\t\tcDefinesMap[\"ENABLE_IP_MASQ_AGENT_IPV6\"] = \"1\"\n\t\t\t\t\tcDefinesMap[\"IP_MASQ_AGENT_IPV6\"] = ipmasq.MapNameIPv6\n\n\t\t\t\t\texcludeCIDR = option.Config.GetIPv6NativeRoutingCIDR()\n\t\t\t\t} else {\n\t\t\t\t\texcludeCIDR = datapath.RemoteSNATDstAddrExclusionCIDRv6()\n\t\t\t\t}\n\n\t\t\t\tif excludeCIDR != nil {\n\t\t\t\t\textraMacrosMap[\"IPV6_SNAT_EXCLUSION_DST_CIDR\"] = excludeCIDR.IP.String()\n\t\t\t\t\tfw.WriteString(FmtDefineAddress(\"IPV6_SNAT_EXCLUSION_DST_CIDR\", excludeCIDR.IP))\n\t\t\t\t\textraMacrosMap[\"IPV6_SNAT_EXCLUSION_DST_CIDR_MASK\"] = excludeCIDR.Mask.String()\n\t\t\t\t\tfw.WriteString(FmtDefineAddress(\"IPV6_SNAT_EXCLUSION_DST_CIDR_MASK\", excludeCIDR.Mask))\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\tctmap.WriteBPFMacros(fw, nil)\n\t}\n\n\tif option.Config.AllowICMPFragNeeded {\n\t\tcDefinesMap[\"ALLOW_ICMP_FRAG_NEEDED\"] = \"1\"\n\t}\n\n\tif option.Config.ClockSource == option.ClockSourceJiffies {\n\t\tcDefinesMap[\"ENABLE_JIFFIES\"] = \"1\"\n\t}\n\n\tif option.Config.EnableIdentityMark {\n\t\tcDefinesMap[\"ENABLE_IDENTITY_MARK\"] = \"1\"\n\t}\n\n\tif option.Config.EnableHighScaleIPcache {\n\t\tcDefinesMap[\"ENABLE_HIGH_SCALE_IPCACHE\"] = \"1\"\n\t}\n\n\tif option.Config.EnableCustomCalls {\n\t\tcDefinesMap[\"ENABLE_CUSTOM_CALLS\"] = \"1\"\n\t}\n\n\tif option.Config.EnableVTEP {\n\t\tcDefinesMap[\"ENABLE_VTEP\"] = \"1\"\n\t\tcDefinesMap[\"VTEP_MAP\"] = vtep.Name\n\t\tcDefinesMap[\"VTEP_MAP_SIZE\"] = fmt.Sprintf(\"%d\", vtep.MaxEntries)\n\t\tcDefinesMap[\"VTEP_MASK\"] = fmt.Sprintf(\"%#x\", byteorder.NetIPv4ToHost32(net.IP(option.Config.VtepCidrMask)))\n\n\t}\n\n\tvlanFilter, err := vlanFilterMacros()\n\tif err != nil {\n\t\treturn err\n\t}\n\tcDefinesMap[\"VLAN_FILTER(ifindex, vlan_id)\"] = vlanFilter\n\n\tif option.Config.EnableICMPRules {\n\t\tcDefinesMap[\"ENABLE_ICMP_RULE\"] = \"1\"\n\t}\n\n\tcDefinesMap[\"CIDR_IDENTITY_RANGE_START\"] = fmt.Sprintf(\"%d\", identity.MinLocalIdentity)\n\tcDefinesMap[\"CIDR_IDENTITY_RANGE_END\"] = fmt.Sprintf(\"%d\", identity.MaxLocalIdentity)\n\n\tif option.Config.TunnelingEnabled() {\n\t\tcDefinesMap[\"TUNNEL_MODE\"] = \"1\"\n\t}\n\n\tciliumNetLink, err := netlink.LinkByName(defaults.SecondHostDevice)\n\tif err != nil {\n\t\treturn err\n\t}\n\tcDefinesMap[\"CILIUM_NET_MAC\"] = fmt.Sprintf(\"{.addr=%s}\", mac.CArrayString(ciliumNetLink.Attrs().HardwareAddr))\n\tcDefinesMap[\"HOST_IFINDEX\"] = fmt.Sprintf(\"%d\", ciliumNetLink.Attrs().Index)\n\n\tciliumHostLink, err := netlink.LinkByName(defaults.HostDevice)\n\tif err != nil {\n\t\treturn err\n\t}\n\tcDefinesMap[\"HOST_IFINDEX_MAC\"] = fmt.Sprintf(\"{.addr=%s}\", mac.CArrayString(ciliumHostLink.Attrs().HardwareAddr))\n\tcDefinesMap[\"CILIUM_IFINDEX\"] = fmt.Sprintf(\"%d\", ciliumHostLink.Attrs().Index)\n\n\tephemeralMin, err := getEphemeralPortRangeMin()\n\tif err != nil {\n\t\treturn err\n\t}\n\tcDefinesMap[\"EPHEMERAL_MIN\"] = fmt.Sprintf(\"%d\", ephemeralMin)\n\n\tif err := cDefinesMap.Merge(h.nodeExtraDefines); err != nil {\n\t\treturn err\n\t}\n\n\tfor _, fn := range h.nodeExtraDefineFns {\n\t\tdefines, err := fn()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := cDefinesMap.Merge(defines); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif option.Config.EnableHealthDatapath {\n\t\tif option.Config.IPv4Enabled() {\n\t\t\tipip4, err := netlink.LinkByName(defaults.IPIPv4Device)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tcDefinesMap[\"ENCAP4_IFINDEX\"] = fmt.Sprintf(\"%d\", ipip4.Attrs().Index)\n\t\t}\n\t\tif option.Config.IPv6Enabled() {\n\t\t\tipip6, err := netlink.LinkByName(defaults.IPIPv6Device)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tcDefinesMap[\"ENCAP6_IFINDEX\"] = fmt.Sprintf(\"%d\", ipip6.Attrs().Index)\n\t\t}\n\t}\n\n\t// Since golang maps are unordered, we sort the keys in the map\n\t// to get a consistent written format to the writer. This maintains\n\t// the consistency when we try to calculate hash for a datapath after\n\t// writing the config.\n\tkeys := make([]string, 0, len(cDefinesMap))\n\tfor key := range cDefinesMap {\n\t\tkeys = append(keys, key)\n\t}\n\tsort.Strings(keys)\n\n\tfor _, key := range keys {\n\t\tfmt.Fprintf(fw, \"#define %s %s\\n\", key, cDefinesMap[key])\n\t}\n\n\t// Populate cDefinesMap with extraMacrosMap to get all the configuration\n\t// in the cDefinesMap itself.\n\tfor key, value := range extraMacrosMap {\n\t\tcDefinesMap[key] = value\n\t}\n\n\t// Write the JSON encoded config as base64 encoded commented string to\n\t// the header file.\n\tjsonBytes, err := json.Marshal(cDefinesMap)\n\tif err == nil {\n\t\t// We don't care if some error occurs while marshaling the map.\n\t\t// In such cases we skip embedding the base64 encoded JSON configuration\n\t\t// to the writer.\n\t\tencodedConfig := base64.StdEncoding.EncodeToString(jsonBytes)\n\t\tfmt.Fprintf(fw, \"\\n// JSON_OUTPUT: %s\\n\", encodedConfig)\n\t}\n\n\treturn fw.Flush()\n}", "func (c *Config) Write(filename string) (err error) {\n\tb, err := json.Marshal(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfile, err := os.Create(filename)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif _, err := file.Write(b); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (parser *Parser) WriteConfig() error {\n\tconfig := structs.Map(parser.Config)\n\tfor _, site := range parser.Sites {\n\t\tif site.Path == \".\" {\n\t\t\tfor key, val := range site.AllParameters {\n\t\t\t\tif _, ok := config[key]; !ok {\n\t\t\t\t\tconfig[key] = val\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tbs, err := yaml.Marshal(config)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn parser.Storage.Write(parser.Storage.GetSourceKey(KeyConfig), bs)\n}", "func writeTemplate(templatePath string, writer io.Writer, ctx *Context) error {\n\ttmpl, err := template.New(filepath.Base(templatePath)).Funcs(template.FuncMap{\n\t\t\"test\": tplFuncTest,\n\t}).ParseFiles(templatePath)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to prepare template %s, error: %v\", templatePath, err)\n\t}\n\n\terr = tmpl.Execute(writer, &ctx)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to generate template %s, error: %v\", templatePath, err)\n\t}\n\treturn nil\n}", "func (ec *ExecutionContext) WriteConfig(config *Config) error {\n\tvar op errors.Op = \"cli.ExecutionContext.WriteConfig\"\n\tvar cfg *Config\n\tif config != nil {\n\t\tcfg = config\n\t} else {\n\t\tcfg = ec.Config\n\t}\n\tbuf := new(bytes.Buffer)\n\tencoder := yaml.NewEncoder(buf)\n\tencoder.SetIndent(2)\n\terr := encoder.Encode(cfg)\n\tif err != nil {\n\t\treturn errors.E(op, err)\n\t}\n\terr = ioutil.WriteFile(ec.ConfigFile, buf.Bytes(), 0644)\n\tif err != nil {\n\t\treturn errors.E(op, err)\n\t}\n\treturn nil\n}", "func (h *HeaderfileWriter) WriteNetdevConfig(w io.Writer, cfg datapath.DeviceConfiguration) error {\n\tfw := bufio.NewWriter(w)\n\th.writeNetdevConfig(fw, cfg)\n\treturn fw.Flush()\n}", "func WriteConfig(config Config, path string) error {\n\tb, err := yaml.Marshal(config)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = ioutil.WriteFile(path, b, 0644)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func writeConfig(o options, clientset kubernetes.Interface) error {\n\t// kubeconfig is a kubernetes config.\n\tvar kubeconfig []byte\n\n\tdir, file := filepath.Split(o.output)\n\n\terr := os.MkdirAll(dir, os.ModePerm)\n\tif err != nil {\n\t\treturn &util.ExitError{Message: fmt.Sprintf(\"unable to create output directory %v: %v.\", dir, err), Code: 1}\n\t}\n\n\tif o.certificate {\n\t\tif kubeconfig, err = certificate.CreateKubeConfigWithCertificateCredentials(clientset, o.name); err != nil {\n\t\t\treturn &util.ExitError{Message: fmt.Sprintf(\"unable to create kubeconfig file with cert and key for %v: %v.\", o.name, err), Code: 1}\n\t\t}\n\t} else {\n\t\t// Service account credentials are the default if unspecified.\n\t\tif kubeconfig, err = serviceaccount.CreateKubeConfigWithServiceAccountCredentials(clientset, o.name); err != nil {\n\t\t\treturn &util.ExitError{Message: fmt.Sprintf(\"unable to create kubeconfig file with service account for %v: %v.\", o.name, err), Code: 1}\n\t\t}\n\t}\n\n\tif !o.overwrite && util.FileExists(o.output) {\n\t\tif kubeconfig, err = mergeConfigs(o, kubeconfig); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif err = ioutil.WriteFile(o.output, kubeconfig, 0644); err != nil {\n\t\treturn &util.ExitError{Message: fmt.Sprintf(\"unable to write to file %v: %v.\", file, err), Code: 1}\n\t}\n\n\treturn nil\n}", "func WriteConfig(cfg NetworkMonitorConfig) string {\n\tif cfgBytes, err := yml.Marshal(cfg); err != nil {\n\t\tpanic(err.Error())\n\t} else {\n\t\tfmt.Println(string(cfgBytes))\n\t\treturn string(cfgBytes)\n\t}\n}", "func WriteTemplate(appCtx *config.AppContext, res http.ResponseWriter, t Template, data interface{}) {\n\tres.WriteHeader(http.StatusOK)\n\tres.Header().Set(\"Content-Type\", \"text/html\")\n\terr := t.T.ExecuteTemplate(res, t.Name, data)\n\tif err != nil && appCtx != nil {\n\t\t//Error while writing the template\n\t\tappCtx.Log.Error(\"Error while rendering the template\", err.Error())\n\t} else if appCtx == nil {\n\t\t//Error while writing the template\n\t\tlog.Error(\"No app context found\")\n\t}\n}", "func WriteConnectionConfig(context *clusterd.Context, clusterInfo *cephclient.ClusterInfo) error {\n\t// write the latest config to the config dir\n\tif _, err := cephclient.GenerateConnectionConfig(context, clusterInfo); err != nil {\n\t\treturn errors.Wrap(err, \"failed to write connection config\")\n\t}\n\n\treturn nil\n}", "func GenerateConfig(configPath string) error {\n\ttemplate := &TemporalConfig{}\n\ttemplate.setDefaults()\n\tb, err := json.Marshal(template)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar pretty bytes.Buffer\n\tif err = json.Indent(&pretty, b, \"\", \"\\t\"); err != nil {\n\t\treturn err\n\t}\n\treturn ioutil.WriteFile(configPath, pretty.Bytes(), os.ModePerm)\n}", "func (cfg *Config) Write() error {\n\tif err := os.MkdirAll(filepath.Dir(cfg.path), 0755); err != nil {\n\t\treturn errors.Wrap(err, \"failed to create config path\")\n\t}\n\n\tfh, err := os.Create(cfg.path)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to open config file\")\n\t}\n\tdefer fh.Close()\n\n\tbuf, err := yaml.Marshal(cfg)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to marshal config\")\n\t}\n\n\t_, err = fh.Write(buf)\n\treturn errors.Wrap(err, \"failed to write config\")\n}", "func (c *DMMasterScript) ConfigToFile(tmplFile, outFile string) error {\n\ttmpl, err := template.ParseFiles(tmplFile)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"DMMasterScript template new failed: %v\", err)\n\t}\n\n\tf, err := os.OpenFile(outFile, os.O_WRONLY|os.O_CREATE, 0755)\n\tdefer f.Close()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"DMMasterScript open file failed: %v\", err)\n\t}\n\tif err := tmpl.Execute(f, c); err != nil {\n\t\treturn fmt.Errorf(\"DMMasterScript template execute failed: %v\", err)\n\t}\n\treturn nil\n}", "func WriteConfig(cfg *config.Dendrite, configDir string) error {\n\tdata, err := yaml.Marshal(cfg)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn ioutil.WriteFile(filepath.Join(configDir, ConfigFile), data, 0666)\n}", "func WriteConfig(config SlackConfiguration) error {\n\texpanded, err := homedir.Expand(configPath)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tconfigFolder := filepath.Dir(expanded)\n\tif err := os.MkdirAll(configFolder, 0777); err != nil {\n\t\treturn err\n\t}\n\tconfigFile, err := os.Create(expanded)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer configFile.Close()\n\n\tif err := json.NewEncoder(configFile).Encode(config); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *Config) WriteScopeConfig(path string) error {\n\tscb, err := c.ScopeConfigYaml()\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = ioutil.WriteFile(path, scb, 0644)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error writing ScopeConfig to file %s: %v\", path, err)\n\t}\n\treturn nil\n}", "func (cfg *Config) GenerateSampleConfig(dstDir string) error {\n\n\t// get a *Template\n\tat, err := prepareTemplate(\"/templates/config.json.gotmpl\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// check the destination file-path and create if required\n\t_, err = os.Stat(dstDir)\n\tif err != nil {\n\t\tos.Mkdir(dstDir, 0755)\n\t}\n\n\tvar tfDir string\n\n\tfor i := 0; i < 2; i++ {\n\n\t\tswitch i {\n\t\tcase 0:\n\t\t\ttfDir = dstDir + \"/.dev.config.json\"\n\t\t\tcfg.Env = \"dev\"\n\t\tcase 1:\n\t\t\ttfDir = dstDir + \"/.prd.config.json\"\n\t\t\tcfg.Env = \"prod\"\n\t\tdefault:\n\n\t\t}\n\n\t\t// create the .xxx.config.json file\n\t\tf, err := os.Create(tfDir)\n\t\tif err != nil {\n\t\t\tlog.Printf(\"GenerateSampleConfig: %v\\n\", err)\n\t\t\treturn err\n\t\t}\n\t\tdefer f.Close()\n\n\t\t// set permissions\n\t\terr = f.Chmod(0755)\n\t\tif err != nil {\n\t\t\tlog.Printf(\"GenerateSampleConfig: %v\\n\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t// execute the config.json.gotmpl template using new file .xxx.config.json as a target\n\t\terr = at.Execute(f, cfg)\n\t\tif err != nil {\n\t\t\tlog.Printf(\"GenerateSampleConfig: %v\\n\", err)\n\t\t\treturn err\n\t\t}\n\t\tlog.Println(\"generated:\", tfDir)\n\t}\n\treturn nil\n}", "func (c *CNIConfigWriter) Write(inputs CNIConfigInputs) error {\n\tif inputs == c.lastInputs {\n\t\treturn nil\n\t}\n\n\t// use an extension not recognized by CNI to write the contents initially\n\t// https://github.com/containerd/go-cni/blob/891c2a41e18144b2d7921f971d6c9789a68046b2/opts.go#L170\n\t// then we can rename to atomically make the file appear\n\tf, err := os.Create(c.path + \".temp\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// actually write the config\n\tif err := writeCNIConfig(f, c.template, inputs); err != nil {\n\t\tf.Close()\n\t\tos.Remove(f.Name())\n\t\treturn err\n\t}\n\tf.Sync()\n\tf.Close()\n\n\t// then we can rename to the target config path\n\tif err := os.Rename(f.Name(), c.path); err != nil {\n\t\treturn err\n\t}\n\n\t// we're safely done now, record the inputs\n\tc.lastInputs = inputs\n\treturn nil\n}", "func SaveConfigToFile(name string, data interface{}) error {\n\tvar (\n\t\tbuffer bytes.Buffer\n\t\tsaveErr error\n\t)\n\tpath := filepath.Join(ConfigDir, name)\n\t// Create our buffer and encoder\n\twriter := bufio.NewWriter(&buffer)\n\tencoder := toml.NewEncoder(writer)\n\t// Encode the struct as toml\n\tif saveErr = encoder.Encode(data); saveErr == nil {\n\t\t// Write to the blacklist file\n\t\tsaveErr = ioutil.WriteFile(path, buffer.Bytes(), 0644)\n\t}\n\treturn saveErr\n}", "func SafeWriteConfig(mountData bool, unmountData bool) error {\n\t// configFilePath := path.Join(viper.GetString())\n\tdir := viper.GetString(\"cfg_dir\")\n\tfilepath := path.Join(dir, configName+\".\"+configType)\n\n\tdirExists := fs.DirExists(dir)\n\tif !dirExists {\n\t\terr := fs.CreateDir(dir)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"could not create the config directory for sneak: %s\", err)\n\t\t}\n\t}\n\n\texists := fs.FileExists(filepath)\n\tif exists {\n\t\treturn verify(filepath, mountData, unmountData)\n\t}\n\n\tgui.Info(\"popcorn\", \"creating your config file...\", filepath)\n\n\tif _, err := os.OpenFile(filepath, os.O_RDONLY|os.O_CREATE, 0600); err != nil {\n\t\tgui.Warn(\"could not create configuration file\", filepath)\n\t\treturn err\n\t}\n\n\t// set defaults\n\thtbUsername := gui.InputPromptWithResponse(\"what is your hack the box username?\", \"\", true)\n\tviper.Set(\"htb_username\", htbUsername)\n\tviper.Set(\"openvpn_filepath\", fmt.Sprintf(\"%s/%s.ovpn\", dir, viper.Get(\"htb_username\")))\n\tpreferredEditor := gui.GetUsersPreferredEditor(\"\", true)\n\tviper.Set(\"default_editor\", preferredEditor)\n\tviper.Set(\"data\", dir)\n\tviper.Set(\"webshort\", defaultShortcuts)\n\tfmt.Println(htbNetworkIPHelpText)\n\tlabAccessIP := gui.InputPromptWithResponse(\"what is your HTB Lab Network IPv4?\", \"10.10.15.71\", true)\n\tviper.Set(\"htb_network_ip\", labAccessIP)\n\n\t// write config file\n\tgui.Info(\"popcorn\", \"writing sneak defaults...\", filepath)\n\treturn viper.WriteConfigAs(filepath)\n}", "func writeConfig(args *common.Parameters) {\n\n\t//Create the config file and open it for writing.\n\tconfigWrite, err := os.Create(\"./data/cluster/config.csv\")\n\tif err != nil {\n\t\targs.ErrorLogger.Println(\"entity=\" + entityKind + \" message=\" + err.Error())\n\t\tfmt.Println(\"[ERROR] entity=\" + entityKind + \" message=\" + err.Error())\n\t\treturn\n\t}\n\n\t//Write out the header.\n\tfmt.Fprintln(configWrite, \"AuditTime,Name\")\n\tfmt.Fprintf(configWrite, \"%s,%s\\n\", common.Format(args.CurrentTime), *args.ClusterName)\n\tconfigWrite.Close()\n}", "func (mailConf *MailConfig) WriteConf() error {\n\n\ttype Write struct {\n\t\tHeader *MailConfig `yaml:\"mail\"`\n\t}\n\n\tw := Write{Header: mailConf}\n\n\tdata, err := yaml.Marshal(w)\n\tif err != nil {\n\t\treturn errors.New(\"Error to marshal mail configuration: \"+err.Error())\n\t}\n\n\terr = os.WriteFile(\"config/mail.tmp.yml\", data, 0644)\n\tif err != nil {\n\t\treturn errors.New(\"Error to write mail configuration file: \"+err.Error())\n\t}\n\n\treturn nil\n}", "func (c *Config) ConfigSaveWriter(w io.Writer) error {\n\tout, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = w.Write(out)\n\treturn err\n}", "func (c *Config) Write(path string) error {\r\n\tbytes, err := yaml.Marshal(c)\r\n\tif err == nil {\r\n\t\treturn ioutil.WriteFile(path, bytes, 0777)\r\n\t}\r\n\treturn err\r\n}", "func (fes *FrontEndService) WriteConfigDropIfNoPeer(conf *string, hasVIP4 bool, hasVIP6 bool) {\n\tif hasVIP4 {\n\t\t*conf += \"protocol static BH4 {\\n\"\n\t\t*conf += \"\\tipv4 { preference 0; };\\n\"\n\t\t*conf += \"\\troute 0.0.0.0/0 blackhole {\\n\"\n\t\t*conf += \"\\t\\tkrt_metric=4294967295;\\n\"\n\t\t*conf += \"\\t};\\n\"\n\t\t*conf += \"}\\n\"\n\t\t*conf += \"\\n\"\n\t}\n\n\tif hasVIP6 {\n\t\t*conf += \"protocol static BH6 {\\n\"\n\t\t*conf += \"\\tipv6 { preference 0; };\\n\"\n\t\t*conf += \"\\troute 0::/0 blackhole {\\n\"\n\t\t*conf += \"\\t\\tkrt_metric=4294967295;\\n\"\n\t\t*conf += \"\\t};\\n\"\n\t\t*conf += \"}\\n\"\n\t\t*conf += \"\\n\"\n\t}\n\treturn\n}", "func (cfg *Config) Write() error {\n\tyamlBytes, err := yaml.Marshal(cfg)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn ioutil.WriteFile(FilePath(), yamlBytes, 0644)\n}", "func EncodeConfig(w io.Writer, c *SupportedBranchesConfig) error {\n\tc.Sort()\n\tb, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = w.Write(b)\n\treturn err\n}", "func WriteConfigJSON(c Configuration, filename string) error {\n\tbytes, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn ioutil.WriteFile(filename, bytes, 0644)\n}", "func (fes *FrontEndService) WriteConfigBGP(conf *string) {\n\tfor _, gw := range fes.gateways {\n\t\tif isIPv6(gw) || isIPv4(gw) {\n\t\t\tipv := \"\"\n\t\t\tif isIPv4(gw) {\n\t\t\t\tipv += \"\\tipv4 {\\n\"\n\t\t\t\tif len(fes.vrrps) > 0 {\n\t\t\t\t\tipv += \"\\t\\timport none;\\n\"\n\t\t\t\t} else {\n\t\t\t\t\tipv += \"\\t\\timport filter default_v4;\\n\"\n\t\t\t\t}\n\t\t\t\tipv += \"\\t\\texport filter cluster_e_static;\\n\"\n\t\t\t\tipv += \"\\t};\\n\"\n\t\t\t} else if isIPv6(gw) {\n\t\t\t\tipv = \"\\tipv6 {\\n\"\n\t\t\t\tif len(fes.vrrps) > 0 {\n\t\t\t\t\tipv += \"\\t\\timport none;\\n\"\n\t\t\t\t} else {\n\t\t\t\t\tipv += \"\\t\\timport filter default_v6;\\n\"\n\t\t\t\t}\n\t\t\t\tipv += \"\\t\\texport filter cluster_e_static;\\n\"\n\t\t\t\tipv += \"\\t};\\n\"\n\t\t\t}\n\t\t\t*conf += \"protocol bgp from LINK {\\n\"\n\t\t\t*conf += \"\\tinterface \\\"\" + fes.extInterface + \"\\\";\\n\"\n\t\t\t*conf += \"\\tlocal port \" + fes.localPortBGP + \" as \" + fes.localAS + \";\\n\"\n\t\t\t*conf += \"\\tneighbor \" + strings.Split(gw, \"/\")[0] + \" port \" + fes.remotePortBGP + \" as \" + fes.remoteAS + \";\\n\"\n\t\t\t*conf += ipv\n\t\t\t*conf += \"}\\n\"\n\t\t\t*conf += \"\\n\"\n\t\t}\n\t}\n}", "func (srvConf *ServerConfig) WriteConf() error {\n\n\ttype Write struct {\n\t\tHeader *ServerConfig `yaml:\"server\"`\n\t}\n\n\tw := Write{Header: srvConf}\n\n\tdata, err := yaml.Marshal(w)\n\tif err != nil {\n\t\treturn errors.New(\"Error to marshal server configuration: \"+err.Error())\n\t}\n\n\terr = os.WriteFile(\"config/server.tmp.yml\", data, 0644)\n\tif err != nil {\n\t\treturn errors.New(\"Error to write server configuration file: \"+err.Error())\n\t}\n\n\treturn nil\n}", "func (c *DMMasterScaleScript) ConfigToFile(tmplFile, outFile string) error {\n\ttmpl, err := template.ParseFiles(tmplFile)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"DMMasterScript template new failed: %v\", err)\n\t}\n\n\tf, err := os.OpenFile(outFile, os.O_WRONLY|os.O_CREATE, 0755)\n\tdefer f.Close()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"DMMasterScript open file failed: %v\", err)\n\t}\n\tif err := tmpl.Execute(f, c); err != nil {\n\t\treturn fmt.Errorf(\"DMMasterScript template execute failed: %v\", err)\n\t}\n\treturn nil\n}", "func WriteKubeConfig(configRaw *clientcmdapi.Config) (string, error) {\n\tdata, err := clientcmd.Write(*configRaw)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Create temp file\n\ttempFile, err := os.CreateTemp(\"\", \"\")\n\tif err != nil {\n\t\treturn \"\", errors.Wrap(err, \"create temp file\")\n\t}\n\n\t// Write to temp file\n\t_, err = tempFile.Write(data)\n\tif err != nil {\n\t\tos.Remove(tempFile.Name())\n\t\treturn \"\", errors.Wrap(err, \"write temp file\")\n\t}\n\n\t// Close temp file\n\ttempFile.Close()\n\n\t// Okay sometimes the file is written so quickly that helm somehow\n\t// cannot read it immediately which causes errors\n\t// so we wait here till the file is ready\n\tnow := time.Now()\n\tfor time.Since(now) < time.Minute {\n\t\t_, err = os.Stat(tempFile.Name())\n\t\tif err != nil {\n\t\t\tif os.IsNotExist(err) {\n\t\t\t\ttime.Sleep(time.Millisecond * 50)\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tos.Remove(tempFile.Name())\n\t\t\treturn \"\", err\n\t\t}\n\n\t\tbreak\n\t}\n\n\treturn tempFile.Name(), nil\n}", "func (p *ProviderConfigs) writeProviderConfig() error {\n\t// Get file.\n\tconfig := getProviderConfigFile()\n\n\t// Convert object to json.\n\tdata, err := json.MarshalIndent(p, \"\", \" \")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\t// Write json output to file.\n\terr = ioutil.WriteFile(config, data, 0644)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn nil\n}", "func (c *PumpScript) ConfigWithTemplate(tpl string) (string, error) {\n\ttmpl, err := template.New(\"Pump\").Parse(tpl)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tcontent := bytes.NewBufferString(\"\")\n\tif err := tmpl.Execute(content, c); err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn content.String(), nil\n}", "func GenerateConfig(f string) error {\n\texists, err := pathExists(f)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif exists {\n\t\treturn errors.New(\"output config file already exists: \" + f)\n\t}\n\text := filepath.Ext(f)\n\n\tif len(ext) > 1 {\n\t\text = ext[1:]\n\t}\n\n\tsupportedExts := []string{\"json\", \"toml\", \"yaml\", \"yml\"}\n\n\tif !stringInSlice(ext, supportedExts) {\n\t\treturn errors.New(\"output config file must have one of supported extensions: \" + strings.Join(supportedExts, \", \"))\n\t}\n\n\tvar t *template.Template\n\n\tswitch ext {\n\tcase \"json\":\n\t\tt, err = template.New(\"config\").Parse(jsonConfigTemplate)\n\tcase \"toml\":\n\t\tt, err = template.New(\"config\").Parse(tomlConfigTemplate)\n\tcase \"yaml\", \"yml\":\n\t\tt, err = template.New(\"config\").Parse(yamlConfigTemplate)\n\t}\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar output bytes.Buffer\n\t_ = t.Execute(&output, struct {\n\t\tTokenSecret string\n\t\tAdminPassword string\n\t\tAdminSecret string\n\t\tAPIKey string\n\t}{\n\t\tuuid.New().String(),\n\t\tuuid.New().String(),\n\t\tuuid.New().String(),\n\t\tuuid.New().String(),\n\t})\n\n\treturn ioutil.WriteFile(f, output.Bytes(), 0644)\n}", "func WriteSysConfig(fileName, value string) error {\n\tf, err := os.OpenFile(fileName, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, 0666)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to open configuration file: %s\", err)\n\t}\n\t_, err = f.WriteString(value)\n\tif err != nil {\n\t\tf.Close()\n\t\treturn fmt.Errorf(\"unable to write value: %s\", err)\n\t}\n\terr = f.Close()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to close configuration file: %s\", err)\n\t}\n\treturn nil\n}", "func (sm *ShardMaster) writeConfig(config Config) {\n\ttoWrite := JSONConfig{Num: config.Num, Shards: config.Shards}\n\tnewGroups := make(map[string][]string)\n\n\tfor gid, vals := range config.Groups {\n\t\tnewGroups[strconv.FormatInt(gid, 10)] = vals\n\t}\n\ttoWrite.Groups = newGroups\n\n\tb, err := json.Marshal(toWrite)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tf, err2 := os.OpenFile(sm.configFile, os.O_APPEND|os.O_WRONLY|os.O_CREATE, 0666)\n\tif err2 != nil {\n\t\tlog.Fatal(err2)\n\t}\n\tdefer f.Close()\n\t_, err = f.WriteString(string(b) + \"\\n\")\n}", "func (c *Config) Write(filename string, flags int) error {\n\t// Avoid including the auth parameters in the configuration file\n\ttemp := c.Get(\"auth\")\n\tdefer c.Set(\"auth\", temp)\n\tdelete(config.mp, \"auth\")\n\n\tcontent, err := c.marshal(filepath.Ext(filename))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tf, err := os.OpenFile(filename, flags, 0600)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif _, err := f.Write(content); err != nil {\n\t\treturn errors.Wrap(err, \"writing config\")\n\t}\n\n\tif err := f.Close(); err != nil {\n\t\treturn errors.Wrap(err, \"closing file\")\n\t}\n\n\treturn nil\n}", "func writeConfiguration(conf configuration) error {\n\thomeDir, err := homedir.Dir()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tconfigFilePath := filepath.Join(homeDir, \".telelog.conf\")\n\n\tf, err := os.Create(configFilePath)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer f.Close()\n\n\tencoder := json.NewEncoder(f)\n\treturn encoder.Encode(conf)\n}", "func WriteChainConfig(db database.KeyValueWriter, hash common.Hash, cfg *config.ChainConfig) {\n\tif cfg == nil {\n\t\treturn\n\t}\n\tdata, err := json.Marshal(cfg)\n\tif err != nil {\n\t\tlog.Critical(\"Failed to JSON encode chain config\", \"err\", err)\n\t}\n\tif err := db.Put(configKey(hash), data, \"config\"); err != nil {\n\t\tlog.Critical(\"Failed to store chain config\", \"err\", err)\n\t}\n}", "func WriteJSONConfig(conf *Config) {\n\tfile, err := os.Create(settingsFile)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdefer file.Close()\n\tstr, err := json.MarshalIndent(conf, \"\", \" \")\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tfile.Write(str)\n\tfile.WriteString(\"\\n\")\n}", "func (cfg *Configuration) WriteConfiguration(path string) error {\n\n\tfile, err := os.Create(path)\n\tif err != nil {\n\t\tlog.WithFields(log.Fields{\"err\": err}).Error(\"os.Create\")\n\t\treturn err\n\t}\n\tdefer file.Close()\n\n\te := json.NewEncoder(file)\n\tif err := e.Encode(cfg); err != nil {\n\t\tlog.WithFields(log.Fields{\"err\": err}).Error(\"json.NewEncoder.Encode\")\n\t\treturn err\n\t}\n\treturn nil\n}", "func writeConfigFile() {\n f, err := os.Create(\"../src/config/config.toml\")\n if err != nil {\n log.Panic(\"Could not open config.toml\")\n }\n defer f.Close()\n\n var api = \"http://localhost:3000\" // Placeholder\n var rpc = \"http://localhost:8545\" // Placeholder\n dir, _ := filepath.Abs(filepath.Dir(os.Args[0]))\n\n var s = fmt.Sprintf(`[development]\ngridplus_api = \"%s\"\nrpc_provider = \"%s\"\nserial_no = \"ABCD0101\"\n[wallet]\nkey_path = \"%s/../src/config\"`, api, rpc, dir)\n\n _, err2 := f.WriteString(s)\n if err2 != nil {\n log.Panic(\"Could not write config file\")\n }\n return\n}", "func WriteServerConfig(filePath string, cfg *ServerConfig) error {\n\tbts, err := toml.Marshal(*cfg)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn files.WriteAndBackup(filePath, bts)\n}", "func (p *BaseProvider) CreateConfiguration(tmplContent string, funcMap template.FuncMap, templateObjects interface{}) (*types.Configuration, error) {\n\tvar defaultFuncMap = sprig.TxtFuncMap()\n\t// tolower is deprecated in favor of sprig's lower function\n\tdefaultFuncMap[\"tolower\"] = strings.ToLower\n\tdefaultFuncMap[\"normalize\"] = Normalize\n\tdefaultFuncMap[\"split\"] = split\n\tfor funcID, funcElement := range funcMap {\n\t\tdefaultFuncMap[funcID] = funcElement\n\t}\n\n\ttmpl := template.New(p.Filename).Funcs(defaultFuncMap)\n\n\t_, err := tmpl.Parse(tmplContent)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar buffer bytes.Buffer\n\terr = tmpl.Execute(&buffer, templateObjects)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar renderedTemplate = buffer.String()\n\tif p.DebugLogGeneratedTemplate {\n\t\tlog.Debugf(\"Template content: %s\", tmplContent)\n\t\tlog.Debugf(\"Rendering results: %s\", renderedTemplate)\n\t}\n\treturn p.DecodeConfiguration(renderedTemplate)\n}", "func (c DQLConfig) Write() error {\n\tf := filepath.Join(helpers.GetProjectPath(c.ProjectPath), \"dql.conf.json\")\n\n\tjson, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn ioutil.WriteFile(f, json, 0644)\n}", "func (c *Config) WriteTo(w io.Writer) (n int64, err error) {\n\n\tvar buf bytes.Buffer\n\tenc := toml.NewEncoder(&buf)\n\terr = enc.Encode(&c.Settings)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\twn, err := w.Write(buf.Bytes())\n\treturn int64(wn), err\n}", "func ApplyTemplate(props interface{}, properties *configProperties) {\n\tstringvalues := properties.StringMap()\n\tfuncmap := make(map[string]interface{})\n\tfuncmap[\"to_upper\"] = strings.ToUpper\n\tfuncmap[\"to_lower\"] = strings.ToLower\n\tfuncmap[\"split\"] = strings.Split\n\tfuncmap[\"reg_match\"] = regMatch\n\tfuncmap[\"reg_replace\"] = regReplace\n\tfuncmap[\"match_srcs\"] = matchSrcs\n\tfuncmap[\"add_if_supported\"] = filter_compiler_flags\n\tpropsVal := reflect.Indirect(reflect.ValueOf(props))\n\n\tapplyTemplateRecursive(propsVal, stringvalues, funcmap)\n}", "func Write(tx Transport, host string, data, info []string, options ...TransportOption) error {\n\t// the Kind should configure the transport parameters before\n\n\terr := tx.Connect(host, options...)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"%s: %s\", host, err)\n\t}\n\n\tdefer tx.Close()\n\n\tfor i1, d1 := range data {\n\t\terr := tx.Write(&d1, &info[i1])\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"could not write config %s: %s\", d1, err)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (c Config) WriteTo(w io.Writer) (int64, error) {\n\tdata, err := json.Marshal(c)\n\tif err != nil {\n\t\treturn int64(len(data)), err\n\t}\n\treturn io.Copy(w, bytes.NewReader(data))\n}", "func renderConfig(templateName, configTemplate string, config interface{}) (string, error) {\n\tvar configuration bytes.Buffer\n\ttmpl, err := template.New(templateName).Parse(configTemplate)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif err := tmpl.Execute(&configuration, config); err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn configuration.String(), nil\n}", "func GenerateIptablesConfig(path string, privnet *model.Network) error {\n\tvar (\n\t\tf *os.File\n\t\terr error\n\t)\n\tt := template.Must(template.New(\"iptablesConfig\").Parse(iptablesTmpl))\n\tif f, err = os.OpenFile(path, os.O_RDWR|os.O_CREATE|os.O_TRUNC, 0644); err != nil {\n\t\treturn fmt.Errorf(\"error opening output file at \\\"%s\\\": %w\", path, err)\n\t}\n\tdefer f.Close()\n\tif err = t.Execute(f, struct {\n\t\tPrivateNetwork string\n\t}{privnet.IPv4Addresses[0].Net.String()}); err != nil {\n\t\treturn fmt.Errorf(\"error executing template: %w\", err)\n\t}\n\treturn nil\n}", "func TestConfigTemplate(t *testing.T) {\n\n\tassert := assert.New(t)\n\n\tprepareClusterZone(t, assert)\n\n\t// create config template set\n\tnewFpath := randName(\"/e2e-fpath\")\n\tnewConfigTemplateSetID := \"\"\n\tnewConfigTemplateSetData, _ := e2edata.CreateConfigTemplateSetTestData(newBid, newFpath)\n\t{\n\t\tt.Logf(\"Case: create config template set\")\n\t\tbody, err := httpFunc(\"POST\", testhost(templatesetInterfaceV1), newConfigTemplateSetData)\n\t\tassert.Nil(err)\n\n\t\tassert.EqualValues(pbcommon.ErrCode_E_OK, gjson.Get(body, \"errCode\").Int(), body)\n\t\tassert.Equal(\"OK\", gjson.Get(body, \"errMsg\").String(), body)\n\t\tassert.NotEmpty(gjson.Get(body, \"setid\").String(), body)\n\t\tnewConfigTemplateSetID = gjson.Get(body, \"setid\").String()\n\t}\n\n\t// create config template\n\tnewConfigTemplateData, _ := e2edata.CreateConfigTemplateTestData(newBid, newConfigTemplateSetID)\n\tnewConfigTemplateID := \"\"\n\t{\n\t\tt.Logf(\"Case: create config template\")\n\t\tbody, err := httpFunc(\"POST\", testhost(templateInterfaceV1), newConfigTemplateData)\n\t\tassert.Nil(err)\n\n\t\tassert.EqualValues(pbcommon.ErrCode_E_OK, gjson.Get(body, \"errCode\").Int(), body)\n\t\tassert.Equal(\"OK\", gjson.Get(body, \"errMsg\").String(), body)\n\t\tassert.NotEmpty(gjson.Get(body, \"templateid\").String(), body)\n\t\tnewConfigTemplateID = gjson.Get(body, \"templateid\").String()\n\t}\n\n\t// update config template\n\tupdateConfigTemplateData, _ := e2edata.UpdateConfigTemplateTestData(newBid, newConfigTemplateID)\n\t{\n\t\tt.Logf(\"Case: update config template\")\n\t\tbody, err := httpFunc(\"PUT\", testhost(templateInterfaceV1), updateConfigTemplateData)\n\t\tassert.Nil(err)\n\n\t\tassert.EqualValues(pbcommon.ErrCode_E_OK, gjson.Get(body, \"errCode\").Int(), body)\n\t\tassert.Equal(\"OK\", gjson.Get(body, \"errMsg\").String(), body)\n\t}\n\n\t// query config template\n\t{\n\t\tt.Logf(\"Case: query config template\")\n\t\tbody, err := httpGet(testhost(templateInterfaceV1), map[string]string{\n\t\t\t\"seq\": \"0\",\n\t\t\t\"bid\": newBid,\n\t\t\t\"templateid\": newConfigTemplateID,\n\t\t})\n\t\tassert.Nil(err)\n\n\t\tassert.EqualValues(pbcommon.ErrCode_E_OK, gjson.Get(body, \"errCode\").Int(), body)\n\t\tassert.Equal(\"OK\", gjson.Get(body, \"errMsg\").String(), body)\n\t\tassert.Equal(gjson.Get(updateConfigTemplateData, \"name\").String(), gjson.Get(body, \"configTemplate.name\").String(), body)\n\t\tassert.Equal(gjson.Get(updateConfigTemplateData, \"memo\").String(), gjson.Get(body, \"configTemplate.memo\").String(), body)\n\t\tassert.Equal(gjson.Get(updateConfigTemplateData, \"user\").String(), gjson.Get(body, \"configTemplate.user\").String(), body)\n\t\tassert.Equal(gjson.Get(updateConfigTemplateData, \"group\").String(), gjson.Get(body, \"configTemplate.group\").String(), body)\n\t\tassert.Equal(gjson.Get(updateConfigTemplateData, \"fileEncoding\").String(), gjson.Get(body, \"configTemplate.fileEncoding\").String(), body)\n\t\tassert.Equal(gjson.Get(updateConfigTemplateData, \"permission\").Int(), gjson.Get(body, \"configTemplate.permission\").Int(), body)\n\t\tassert.Equal(newConfigTemplateSetID, gjson.Get(body, \"configTemplate.setid\").String(), body)\n\t\tassert.Equal(gjson.Get(newConfigTemplateSetData, \"fpath\").String(), gjson.Get(body, \"configTemplate.fpath\").String(), body)\n\t\tassert.Equal(gjson.Get(newConfigTemplateData, \"creator\").String(), gjson.Get(body, \"configTemplate.creator\").String(), body)\n\t\tassert.Equal(gjson.Get(updateConfigTemplateData, \"operator\").String(), gjson.Get(body, \"configTemplate.lastModifyBy\").String(), body)\n\t}\n\n\t// list config template\n\t{\n\t\tt.Logf(\"Case: list config template\")\n\t\tbody, err := httpGet(testhost(templateListInterfaceV1), map[string]string{\n\t\t\t\"seq\": \"0\",\n\t\t\t\"bid\": newBid,\n\t\t\t\"setid\": newConfigTemplateSetID,\n\t\t\t\"limit\": \"10\",\n\t\t})\n\t\tassert.Nil(err)\n\n\t\tarrs := gjson.Get(body, \"configTemplates\").Array()\n\t\tassert.NotEqual(0, len(arrs), arrs)\n\n\t\tfor _, e := range arrs {\n\t\t\tif e.Get(\"templateid\").String() == newConfigTemplateID {\n\t\t\t\tassert.Equal(gjson.Get(updateConfigTemplateData, \"name\").String(), e.Get(\"name\").String(), body)\n\t\t\t\tassert.Equal(gjson.Get(updateConfigTemplateData, \"memo\").String(), e.Get(\"memo\").String(), body)\n\t\t\t\tassert.Equal(gjson.Get(updateConfigTemplateData, \"user\").String(), e.Get(\"user\").String(), body)\n\t\t\t\tassert.Equal(gjson.Get(updateConfigTemplateData, \"group\").String(), e.Get(\"group\").String(), body)\n\t\t\t\tassert.Equal(gjson.Get(updateConfigTemplateData, \"fileEncoding\").String(), e.Get(\"fileEncoding\").String(), body)\n\t\t\t\tassert.Equal(gjson.Get(updateConfigTemplateData, \"permission\").Int(), e.Get(\"permission\").Int(), body)\n\t\t\t\tassert.Equal(newConfigTemplateSetID, e.Get(\"setid\").String(), body)\n\t\t\t\tassert.Equal(gjson.Get(newConfigTemplateSetData, \"fpath\").String(), e.Get(\"fpath\").String(), body)\n\t\t\t\tassert.Equal(gjson.Get(newConfigTemplateData, \"creator\").String(), e.Get(\"creator\").String(), body)\n\t\t\t\tassert.Equal(gjson.Get(updateConfigTemplateData, \"operator\").String(), e.Get(\"lastModifyBy\").String(), body)\n\t\t\t}\n\t\t}\n\t}\n\n\t// delete config template\n\t{\n\t\tt.Logf(\"Case: delete config template\")\n\t\tbody, err := httpDelete(testhost(templateInterfaceV1), map[string]string{\n\t\t\t\"seq\": \"0\",\n\t\t\t\"bid\": newBid,\n\t\t\t\"templateid\": newConfigTemplateID,\n\t\t\t\"operator\": \"e2e\",\n\t\t})\n\t\tassert.Nil(err)\n\n\t\tassert.EqualValues(pbcommon.ErrCode_E_OK, gjson.Get(body, \"errCode\").Int(), body)\n\t\tassert.Equal(\"OK\", gjson.Get(body, \"errMsg\").String(), body)\n\t}\n\n\t// delete config template set\n\t{\n\t\tt.Logf(\"Case: delete config template set\")\n\t\tbody, err := httpDelete(testhost(templatesetInterfaceV1), map[string]string{\n\t\t\t\"seq\": \"0\",\n\t\t\t\"bid\": newBid,\n\t\t\t\"setid\": newConfigTemplateSetID,\n\t\t\t\"operator\": \"e2e\",\n\t\t})\n\t\tassert.Nil(err)\n\n\t\tassert.EqualValues(pbcommon.ErrCode_E_OK, gjson.Get(body, \"errCode\").Int(), body)\n\t\tassert.Equal(\"OK\", gjson.Get(body, \"errMsg\").String(), body)\n\t}\n\n\tdeleteClusterZone(t, assert)\n}", "func (tp *TemplateProvider) TemplateToConfig(path string, mapping map[string]interface{}) (*api.Config, error) {\n\tb, err := tp.getMappedTemplate(path, mapping)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn convertToConfig(b)\n}", "func (m DeploymentOptionModel) ApplyTemplate(name string, rawtemplate string, replacement []OptionTemplate) bytes.Buffer {\n\tdynstructBuilder := dynamicstruct.NewStruct()\n\t// valarray := make(map[int]interface{})\n\n\tfor _, r := range replacement {\n\t\tdynstructBuilder.AddField(r.Key, r.Value, \"\")\n\t\t// valarray[i] = r.Value\n\t}\n\tdynstruct := dynstructBuilder.Build().New()\n\telem := reflect.ValueOf(&dynstruct).Elem()\n\tfor _, r := range replacement {\n\t\tvalue := reflect.ValueOf(r.Value)\n\t\telem.FieldByName(r.Key).Set(value)\n\t}\n\n\tt := template.Must(template.New(name).Parse(rawtemplate))\n\tvar buf bytes.Buffer\n\tw := io.Writer(&buf)\n\terr := t.Execute(w, dynstruct)\n\tif err != nil {\n\t\tlog.Fatal(\"Execute template error:\", err)\n\t}\n\treturn buf\n}", "func WriteConfigVals(dir string, vals map[string]string) error {\n\tdata := \"\"\n\tfor k, v := range vals {\n\t\tdata += fmt.Sprintf(\"%s = \\\"%s\\\"\\n\", k, v)\n\t}\n\tcfile := filepath.Join(dir, \"config.toml\")\n\treturn os.WriteFile(cfile, []byte(data), 0o600)\n}", "func Write(config *api.Config, path string) error {\n\treturn clientcmd.WriteToFile(*config, path)\n}", "func (m *Config) WriteToFile(fname string) {\n\tif len((*m).Images) == 0 {\n\t\tif err := DeleteYaml(fname); err != nil {\n\t\t\tfmt.Println(err)\n\t\t\texit(1)\n\t\t}\n\t\tif isV {\n\t\t\tfmt.Println(\"Delete config file. :\", fname)\n\t\t}\n\t\treturn\n\t}\n\t//renewal commands\n\tcc := make(map[string]string)\n\tfor key, val := range (*m).Images {\n\t\tfor com := range val.Commands {\n\t\t\tif cnt, ok := cc[com]; ok {\n\t\t\t\tfmt.Println(\"Confrict command name.:\", com)\n\t\t\t\tfmt.Printf(\" image [%s] and [%s]\\n\", cnt, key)\n\t\t\t\tfmt.Printf(\" in file:%s\\n\", fname)\n\t\t\t\texit(1)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tcc[com] = key\n\t\t}\n\t}\n\t(*m).Commands = cc\n\n\t//write config file\n\tif err := SaveYaml(m, fname); err != nil {\n\t\tfmt.Println(err)\n\t\texit(1)\n\t\treturn\n\t}\n\tif isV {\n\t\tfmt.Println(\"Wrote config file. :\", fname)\n\t}\n}", "func (fes *FrontEndService) WriteConfigVRRPs(conf *string, hasVIP4, hasVIP6 bool) {\n\tfor _, ip := range fes.vrrps {\n\t\tif isIPv6(ip) || isIPv4(ip) {\n\t\t\t*conf += \"protocol static {\\n\"\n\t\t\tif isIPv4(ip) {\n\t\t\t\t*conf += \"\\tipv4;\\n\"\n\t\t\t\t*conf += \"\\troute 0.0.0.0/0 via \" + strings.Split(ip, \"/\")[0] + \"%'\" + fes.extInterface + \"' onlink;\\n\"\n\t\t\t} else if isIPv6(ip) {\n\t\t\t\t*conf += \"\\tipv6;\\n\"\n\t\t\t\t*conf += \"\\troute 0::/0 via \" + strings.Split(ip, \"/\")[0] + \"%'\" + fes.extInterface + \"' onlink;\\n\"\n\t\t\t}\n\t\t\t*conf += \"}\\n\"\n\t\t\t*conf += \"\\n\"\n\t\t}\n\t}\n}", "func SaveConfig(file string, conf AppConfig) error {\n\tvar confBuf bytes.Buffer\n\n\te := toml.NewEncoder(&confBuf)\n\tif err := e.Encode(conf); err != nil {\n\t\treturn err\n\t}\n\treturn utils.WriteFile(file, confBuf.Bytes(), 0644)\n}", "func writeConfig(absolutePath string, content interface{}, perm os.FileMode) error {\n\ttask := \"Write a configuration file\"\n\n\t// Check the configuration directory and make sure it exists.\n\tconfigDir := filepath.Dir(absolutePath)\n\tinfo, err := os.Stat(configDir)\n\tif err != nil {\n\t\tif !os.IsNotExist(err) {\n\t\t\treturn errs.NewError(task, err)\n\t\t}\n\n\t\t// The directory doesn't exist.\n\t\tif err := os.MkdirAll(configDir, 0750); err != nil {\n\t\t\treturn errs.NewError(task, err)\n\t\t}\n\t}\n\tif !info.IsDir() {\n\t\treturn errs.NewError(task, errors.New(\"not a directory: \"+configDir))\n\t}\n\n\t// Marshal the content.\n\traw, err := Marshal(content)\n\tif err != nil {\n\t\treturn errs.NewError(task, err)\n\t}\n\n\t// Write the raw content into the file.\n\tif err := ioutil.WriteFile(absolutePath, raw, perm); err != nil {\n\t\treturn errs.NewError(task, err)\n\t}\n\treturn nil\n}", "func CreateFileConfig(conf model.Configuration) bool {\n\n\tconf.RunIntervalMin = interval\n\tconf.SleepIntervalSec = sleep\n\n\tf, err := os.Create(path)\n\tif err != nil {\n\t\tfmt.Println(err.Error())\n\t\treturn false\n\t}\n\tdefer f.Close()\n\n\tjsonConfig, err := json.MarshalIndent(conf, \"\", \" \")\n\tif err != nil {\n\t\tfmt.Println(err.Error())\n\t\treturn false\n\t}\n\n\tf.WriteString(string(jsonConfig))\n\treturn true\n}", "func WriteMerged(kindConfig *Config, explicitConfigPath string) error {\n\t// figure out what filepath we should use\n\tconfigPath := pathForMerge(explicitConfigPath, os.Getenv)\n\n\t// lock config file the same as client-go\n\tif err := lockFile(configPath); err != nil {\n\t\treturn errors.Wrap(err, \"failed to lock config file\")\n\t}\n\tdefer func() {\n\t\t_ = unlockFile(configPath)\n\t}()\n\n\t// read in existing\n\texisting, err := read(configPath)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to get kubeconfig to merge\")\n\t}\n\n\t// merge with kind kubeconfig\n\tif err := merge(existing, kindConfig); err != nil {\n\t\treturn err\n\t}\n\n\t// write back out\n\treturn write(existing, configPath)\n}", "func (p *Profile) WriteConfigField(field, value string) error {\n\tviper.Set(p.GetConfigField(field), value)\n\treturn viper.WriteConfig()\n}", "func createTincConf(log *logging.Logger, i ClusterInstance, vpnName string, connectTo []string) error {\n\tlines := []string{\n\t\tfmt.Sprintf(\"Name = %s\", tincName(i)),\n\t\t\"AddressFamily = ipv4\",\n\t\t\"Interface = tun0\",\n\t}\n\tfor _, name := range connectTo {\n\t\tlines = append(lines, fmt.Sprintf(\"ConnectTo = %s\", name))\n\t}\n\tconfDir := path.Join(\"/etc/tinc\", vpnName)\n\tconfPath := path.Join(confDir, \"tinc.conf\")\n\tif _, err := i.runRemoteCommand(log, fmt.Sprintf(\"sudo mkdir -p %s\", confDir), \"\", false); err != nil {\n\t\treturn maskAny(err)\n\t}\n\tif _, err := i.runRemoteCommand(log, fmt.Sprintf(\"sudo tee %s\", confPath), strings.Join(lines, \"\\n\"), false); err != nil {\n\t\treturn maskAny(err)\n\t}\n\treturn nil\n}" ]
[ "0.6178704", "0.6177519", "0.58227545", "0.58178943", "0.5744848", "0.57286936", "0.57203996", "0.5636844", "0.5603397", "0.5517643", "0.5504188", "0.5494074", "0.54918796", "0.5486956", "0.54737705", "0.5466237", "0.5462646", "0.53718305", "0.5367005", "0.5348312", "0.53151506", "0.5313349", "0.5306713", "0.53027874", "0.52922523", "0.5266352", "0.525707", "0.52082133", "0.52017033", "0.51392376", "0.51379114", "0.51238537", "0.51161766", "0.5076442", "0.5067454", "0.5006473", "0.49973357", "0.49951857", "0.49871406", "0.49548754", "0.49425402", "0.49331436", "0.4921166", "0.49064785", "0.48738542", "0.48736247", "0.48544052", "0.4846351", "0.4841564", "0.4841517", "0.48193365", "0.48067966", "0.48009443", "0.47928047", "0.4792045", "0.478377", "0.47601834", "0.4750569", "0.47456384", "0.47396418", "0.47242367", "0.47117928", "0.47056884", "0.46871063", "0.467861", "0.46742165", "0.46588218", "0.46573806", "0.46487775", "0.46264726", "0.4622611", "0.46194264", "0.46144783", "0.46143717", "0.4596491", "0.45884612", "0.45762312", "0.45757094", "0.4567145", "0.4564424", "0.45346156", "0.45337567", "0.45276728", "0.4522583", "0.45136794", "0.45095423", "0.450813", "0.45054698", "0.44960165", "0.4491892", "0.448215", "0.44723532", "0.4470028", "0.44691598", "0.44622296", "0.44597596", "0.4457824", "0.44506934", "0.44497094", "0.44453347" ]
0.7395368
0
Slightly modified from widget.RunWindow
func (p *ControlPanel) RunWindow(opts *widget.RunWindowOptions) error { var ( nwo *screen.NewWindowOptions t *theme.Theme ) if opts != nil { nwo = &opts.NewWindowOptions t = &opts.Theme } var err error p.w, err = p.s.NewWindow(nwo) if err != nil { return err } defer p.w.Release() paintPending := false gef := gesture.EventFilter{EventDeque: p.w} for { e := p.w.NextEvent() if e = gef.Filter(e); e == nil { continue } switch e := e.(type) { case lifecycle.Event: p.root.OnLifecycleEvent(e) if e.To == lifecycle.StageDead { return nil } case gesture.Event, mouse.Event: p.root.OnInputEvent(e, image.Point{}) case paint.Event: ctx := &node.PaintContext{ Theme: t, Screen: p.s, Drawer: p.w, Src2Dst: f64.Aff3{ 1, 0, 0, 0, 1, 0, }, } if err := p.root.Paint(ctx, image.Point{}); err != nil { return err } p.w.Publish() paintPending = false case size.Event: if dpi := float64(e.PixelsPerPt) * unit.PointsPerInch; dpi != t.GetDPI() { newT := new(theme.Theme) if t != nil { *newT = *t } newT.DPI = dpi t = newT } windowSize := e.Size() p.root.Measure(t, windowSize.X, windowSize.Y) p.root.Wrappee().Rect = e.Bounds() p.root.Layout(t) // TODO: call Mark(node.MarkNeedsPaint)? case panelUpdate: case error: return e } if !paintPending && p.root.Wrappee().Marks.NeedsPaint() { paintPending = true p.w.Send(paint.Event{}) } } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func AbsEndRun(window fyne.Window, c *fyne.Container, w, h int) {\n\twindow.Resize(fyne.NewSize(float32(w), float32(h)))\n\twindow.SetFixedSize(true)\n\twindow.SetPadded(false)\n\twindow.SetContent(c)\n\twindow.ShowAndRun()\n}", "func (w *Window) Run() {\n\tw.readEvents(w.nativeWin.EventChan())\n}", "func runStatusWindow() error {\n\tif nbmRunStatusWindow.Lock() {\n\t\tdefer nbmRunStatusWindow.Unlock()\n\n\t\tmw := declarative.MainWindow{\n\t\t\tAssignTo: &statusWindow,\n\t\t\tName: \"statusmw\",\n\t\t\tTitle: \"Status Data\",\n\t\t\tIcon: appIcon,\n\t\t\tSize: declarative.Size{Width: 300, Height: 250},\n\t\t\tLayout: declarative.VBox{MarginsZero: true},\n\t\t\tChildren: []declarative.Widget{\n\t\t\t\tdeclarative.Composite{\n\t\t\t\t\tLayout: declarative.Grid{Rows: 2},\n\t\t\t\t\tStretchFactor: 4,\n\t\t\t\t\tChildren: []declarative.Widget{\n\t\t\t\t\t\tdeclarative.TableView{\n\t\t\t\t\t\t\tName: \"statustv\",\n\t\t\t\t\t\t\tColumnsOrderable: true,\n\t\t\t\t\t\t\tAlternatingRowBG: true,\n\t\t\t\t\t\t\tHeaderHidden: true,\n\t\t\t\t\t\t\tLastColumnStretched: true,\n\t\t\t\t\t\t\tColumns: []declarative.TableViewColumn{\n\t\t\t\t\t\t\t\t{Name: \"Index\", Hidden: true},\n\t\t\t\t\t\t\t\t{Name: \"Name\"},\n\t\t\t\t\t\t\t\t{Name: \"Value\"},\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t\tModel: newStatusTableDataModel(),\n\t\t\t\t\t\t},\n\t\t\t\t\t\tdeclarative.PushButton{\n\t\t\t\t\t\t\tText: \"OK\",\n\t\t\t\t\t\t\tOnClicked: func() {\n\t\t\t\t\t\t\t\tstatusWindow.Close()\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t}\n\n\t\t// create window\n\t\terr := mw.Create()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// disable maximize, minimize, and resizing\n\t\thwnd := statusWindow.Handle()\n\t\twin.SetWindowLong(hwnd, win.GWL_STYLE, win.GetWindowLong(hwnd, win.GWL_STYLE) & ^(win.WS_MAXIMIZEBOX|win.WS_MINIMIZEBOX|win.WS_SIZEBOX))\n\n\t\t// start message loop\n\t\tstatusWindow.Run()\n\t} else {\n\t\t// bring already running status window to top\n\t\tstatusWindow.Show()\n\t}\n\n\treturn nil\n}", "func (w *windowImpl) GoRunOnWin(f func()) {\n\tif w.IsClosed() {\n\t\treturn\n\t}\n\tgo func() {\n\t\tw.runQueue <- funcRun{f: f, done: nil}\n\t}()\n}", "func (w *windowImpl) GoRunOnWin(f func()) {\n\tif w.IsClosed() {\n\t\treturn\n\t}\n\tgo func() {\n\t\tw.runQueue <- funcRun{f: f, done: nil}\n\t}()\n}", "func Run(app func(w *Window)) {\n\tw := &Window{\n\t\tRedraw: make(chan window.Event, 128),\n\t\tapp: app,\n\t\trender: make(chan struct{}),\n\t}\n\twindow.Run(w.gfxLoop, Props)\n}", "func (c *Canvas) EndRun() {\n\twindow := c.Window\n\twindow.Resize(fyne.NewSize(float32(c.Width), float32(c.Height)))\n\twindow.SetFixedSize(true)\n\twindow.SetPadded(false)\n\twindow.SetContent(c.Container)\n\twindow.ShowAndRun()\n}", "func (w *windowImpl) RunOnWin(f func()) {\n\tif w.IsClosed() {\n\t\treturn\n\t}\n\tdone := make(chan bool)\n\tw.runQueue <- funcRun{f: f, done: done}\n\t<-done\n}", "func (w *windowImpl) RunOnWin(f func()) {\n\tif w.IsClosed() {\n\t\treturn\n\t}\n\tdone := make(chan bool)\n\tw.runQueue <- funcRun{f: f, done: done}\n\t<-done\n}", "func Run() {\n\tlog.Printf(\"[%s] starting ui\", tag)\n\n\t// stream list\n\tstreamList := cview.NewList()\n\tstreamList.\n\t\tClear().\n\t\tSetHighlightFullLine(true).\n\t\tShowSecondaryText(false).\n\t\tSetBorder(true).\n\t\tSetBorderColor(tcell.ColorBlue).\n\t\tSetTitle(\" 📻 streams \").\n\t\tSetTitleAlign(cview.AlignLeft)\n\n\tfor _, stationID := range streams.StreamStationIDs {\n\t\tstreamList.AddItem(streams.Streams[stationID].Stream, \"\", 0, nil)\n\t}\n\n\t// now playing\n\tnowplaying := cview.NewTextView()\n\tnowplaying.\n\t\tSetText(\"...\").\n\t\tSetTextAlign(cview.AlignCenter).\n\t\tSetTitle(\" 🎵 now playing \").\n\t\tSetTitleAlign(cview.AlignLeft).\n\t\tSetBorderColor(tcell.ColorOrange).\n\t\tSetBorder(true)\n\n\tstreamList.SetSelectedFunc(func(idx int, maintext string, secondarytext string, shortcut rune) {\n\t\tlog.Printf(\"[%s] selected stream changed\", tag)\n\n\t\tstationID := streams.StreamStationIDs[idx]\n\t\turl := \"http:\" + streams.Streams[stationID].URLHigh\n\t\tnowplaying.SetText(streams.Streams[stationID].Stream)\n\n\t\tlog.Printf(\"[%s] playing %s from url %s\", tag, streams.Streams[stationID].Stream, url)\n\t\tplayer.Stop()\n\t\tplayer.Play(url)\n\t})\n\n\t// main layout\n\tflex := cview.NewFlex().\n\t\tSetDirection(cview.FlexRow).\n\t\tAddItem(streamList, 0, 1, true).\n\t\tAddItem(nowplaying, 3, 1, false)\n\n\tapp.SetInputCapture(handleKeyEvent)\n\tif err := app.SetRoot(flex, true).Run(); err != nil {\n\t\tlog.Fatalf(\"[%s] ui initialization failed: %d\", tag, err)\n\t}\n}", "func (m *Maker) Run(t *testing.T, userCalls ...func(*testing.T, gtk.Widgetter)) {\n\tvar w gtk.Widgetter // stores the widget to provide for all calls.\n\tcalls := make([]interface{}, len(userCalls)+1)\n\tcalls[0] = func() gtk.Widgetter {\n\t\tw = m.newW()\n\t\treturn w\n\t}\n\tfor i, call := range userCalls {\n\t\tcalls[i+1] = func() { call(t, w) }\n\t}\n\tm.app.ID = fmt.Sprintf(\"%s%d\", m.baseID, m.counter)\n\tm.counter++\n\tm.app.Run(calls...)\n}", "func run() {\n\tw, h := 800, 550\n\n\tvar err error\n\twin, err := wde.NewWindow(w, h)\n\tif err != nil {\n\t\tlog.Printf(\"Create window error: %v\", err)\n\t\treturn\n\t}\n\n\twin.SetTitle(title)\n\twin.LockSize(true)\n\twin.Show()\n\n\teng = engine.NewEngine(win, w, h)\n\tgo eng.Run()\n\n\tfor event := range win.EventChan() {\n\t\tif quit := handleEvent(event); quit {\n\t\t\tbreak\n\t\t}\n\t}\n\teng.Stop()\n\n\twde.Stop()\n}", "func (ws *WindowSurface) Run() {\n\t// log.Println(\"Starting viewer polling\")\n\tws.running = true\n\t// var simStatus = \"\"\n\tvar frameStart time.Time\n\tvar elapsedTime float64\n\tvar loopTime float64\n\n\tsleepDelay := 0.0\n\n\t// Get a reference to SDL's internal keyboard state. It is updated\n\t// during sdl.PollEvent()\n\tkeyState := sdl.GetKeyboardState()\n\n\trasterizer := renderer.NewBresenHamRasterizer()\n\n\tsdl.SetEventFilterFunc(ws.filterEvent, nil)\n\n\tfor ws.running {\n\t\tframeStart = time.Now()\n\n\t\tsdl.PumpEvents()\n\n\t\tif keyState[sdl.SCANCODE_Z] != 0 {\n\t\t\tws.mod--\n\t\t}\n\t\tif keyState[sdl.SCANCODE_X] != 0 {\n\t\t\tws.mod++\n\t\t}\n\n\t\tws.clearDisplay()\n\n\t\tws.render(rasterizer)\n\n\t\t// This takes on average 5-7ms\n\t\t// ws.texture.Update(nil, ws.pixels.Pix, ws.pixels.Stride)\n\t\tws.texture.Update(nil, ws.rasterBuffer.Pixels().Pix, ws.rasterBuffer.Pixels().Stride)\n\t\tws.renderer.Copy(ws.texture, nil, nil)\n\n\t\tws.txtFPSLabel.DrawAt(10, 10)\n\t\tf := fmt.Sprintf(\"%2.2f\", 1.0/elapsedTime*1000.0)\n\t\tws.dynaTxt.DrawAt(ws.txtFPSLabel.Bounds.W+10, 10, f)\n\n\t\t// ws.mx, ws.my, _ = sdl.GetMouseState()\n\t\tws.txtMousePos.DrawAt(10, 25)\n\t\tf = fmt.Sprintf(\"<%d, %d>\", ws.mx, ws.my)\n\t\tws.dynaTxt.DrawAt(ws.txtMousePos.Bounds.W+10, 25, f)\n\n\t\tws.txtLoopLabel.DrawAt(10, 40)\n\t\tf = fmt.Sprintf(\"%2.2f\", loopTime)\n\t\tws.dynaTxt.DrawAt(ws.txtLoopLabel.Bounds.W+10, 40, f)\n\n\t\tws.renderer.Present()\n\n\t\t// time.Sleep(time.Millisecond * 5)\n\t\tloopTime = float64(time.Since(frameStart).Nanoseconds() / 1000000.0)\n\t\t// elapsedTime = float64(time.Since(frameStart).Seconds())\n\n\t\tsleepDelay = math.Floor(framePeriod - loopTime)\n\t\t// fmt.Printf(\"%3.5f ,%3.5f, %3.5f \\n\", framePeriod, elapsedTime, sleepDelay)\n\t\tif sleepDelay > 0 {\n\t\t\tsdl.Delay(uint32(sleepDelay))\n\t\t\telapsedTime = framePeriod\n\t\t} else {\n\t\t\telapsedTime = loopTime\n\t\t}\n\t}\n}", "func (tui *TUI) Run() error {\n\treturn tui.app.SetRoot(tui.window, true).Run()\n}", "func (s *Service) renderMainWindow() error {\n\tif s.mainwin == nil {\n\t\tvar err error\n\t\ts.mainwin, err = gc.NewWindow(s.screenRows, s.screenCols, 0, 0)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\ts.mainwin.Clear()\n\ts.mainwin.ColorOn(2)\n\ts.mainwin.MoveWindow(0, 0)\n\ts.mainwin.Resize(s.screenRows, s.screenCols)\n\ts.mainwin.Box(0, 0)\n\ts.mainwin.Refresh()\n\treturn nil\n}", "func setupWindow(title string) *gtk.Window {\n\twin, err := gtk.WindowNew(gtk.WINDOW_TOPLEVEL)\n\tif err != nil {\n\t\tlog.Fatal(\"Unable to create window:\", err)\n\t}\n\n\twin.SetTitle(title)\n\twin.Connect(\"destroy\", func() {\n\t\tgtk.MainQuit()\n\t})\n\twin.SetPosition(gtk.WIN_POS_CENTER)\n\twidth, height := 600, 300\n\twin.SetDefaultSize(width, height)\n\n\tbox, _ := gtk.BoxNew(gtk.ORIENTATION_VERTICAL, 0)\n\tbtn, _ := gtk.ButtonNew()\n\tbtn.Connect(\"clicked\", ButtonClicked)\n\tbtn.SetLabel(\"Stop timeout\")\n\n\tbox.Add(btn)\n\twin.Add(box)\n\n\treturn win\n}", "func Window(title string) *WindowWidget {\n\treturn &WindowWidget{\n\t\ttitle: title,\n\t}\n}", "func appConsolatoryWin(win *gtk.Window) {\n\tlbl, err := gtk.LabelNew(\"There were no tags to be displayed, sorry!\")\n\tif err != nil {\n\t\tlog(ERR, fmt.Sprintf(\" creating consolatory label: %s\", err.Error()))\n\t\treturn\n\t}\n\n\tbtn, err := gtk.ButtonNewWithLabel(\"Try again\")\n\tif err != nil {\n\t\tlog(ERR, fmt.Sprintf(\" creating file reload button: %s\\n\", err.Error()))\n\t\treturn\n\t}\n\tbtn.Connect(\"clicked\", handleFiles, win)\n\n\tvbox, err := gtk.BoxNew(gtk.ORIENTATION_VERTICAL, 10)\n\tif err != nil {\n\t\tlog(ERR, fmt.Sprintf(\" creating vbox: %s\\n\", err.Error()))\n\t\treturn\n\t}\n\n\tvbox.Add(lbl)\n\tvbox.Add(btn)\n\twin.Add(vbox)\n\twin.ShowAll()\n}", "func windowsShowSettingsUI(_ *cagent.Cagent, _ bool) {\n\n}", "func NewWindow() *Window {\n\tfile := ui.NewFileWithName(\":/widget.ui\")\n\tloader := ui.NewUiLoader()\n\twidget := loader.Load(file)\n\n\t// Init main window\n\twindow := ui.NewMainWindow()\n\twindow.SetCentralWidget(widget)\n\twindow.SetWindowTitle(\"DFSS Demonstrator v\" + dfss.Version)\n\n\tw := &Window{\n\t\tQMainWindow: window,\n\t\tscene: &Scene{},\n\t}\n\tw.InstallEventFilter(w)\n\n\t// Load dynamic elements from driver\n\tw.logField = ui.NewTextEditFromDriver(widget.FindChild(\"logField\"))\n\tw.graphics = ui.NewGraphicsViewFromDriver(widget.FindChild(\"graphicsView\"))\n\tw.progress = ui.NewLabelFromDriver(widget.FindChild(\"progressLabel\"))\n\n\tw.playButton = ui.NewPushButtonFromDriver(widget.FindChild(\"playButton\"))\n\tw.stopButton = ui.NewPushButtonFromDriver(widget.FindChild(\"stopButton\"))\n\tw.replayButton = ui.NewPushButtonFromDriver(widget.FindChild(\"replayButton\"))\n\n\tw.quantumField = ui.NewSpinBoxFromDriver(widget.FindChild(\"quantumField\"))\n\tw.speedSlider = ui.NewSliderFromDriver(widget.FindChild(\"speedSlider\"))\n\n\t// Load pixmaps\n\tw.pixmaps = map[string]*ui.QPixmap{\n\t\t\"ttp\": ui.NewPixmapWithFilenameFormatFlags(\":/images/server_key.png\", \"\", ui.Qt_AutoColor),\n\t\t\"platform\": ui.NewPixmapWithFilenameFormatFlags(\":/images/server_connect.png\", \"\", ui.Qt_AutoColor),\n\t}\n\n\t// Load icons\n\tw.addIcons()\n\n\t// Add actions\n\tw.addActions()\n\tw.initScene()\n\tw.initTimer()\n\n\tw.StatusBar().ShowMessage(\"Ready\")\n\tw.PrintQuantumInformation()\n\treturn w\n}", "func textWinBtnDoneClicked() {\n\tobj.textWinChkShowModifications.SetActive(false)\n\n\topt.SourceWinWidth, opt.SourceWinHeight = obj.textWin.GetSize()\n\topt.SourceWinPosX, opt.SourceWinPosY = obj.textWin.GetPosition()\n\topt.PanedWidth = opt.SourceWinWidth - obj.Paned.GetPosition()\n\n\tgenericHideWindow(obj.textWin)\n\n\tcurrentText = \"\"\n\tcurrentLine = -1\n}", "func (ui *jobCreatorUI) ShowAndRun() {\n\tif ui.window == nil {\n\t\tfmt.Errorf(\"Window of jobCreatorUI is nil!\")\n\t} else {\n\t\tui.window.ShowAll()\n\t}\n\tgtk.Main()\n}", "func (w *windowImpl) winLoop() {\nouter:\n\tfor {\n\t\tselect {\n\t\tcase <-w.winClose:\n\t\t\tbreak outer\n\t\tcase f := <-w.runQueue:\n\t\t\tif w.glw == nil {\n\t\t\t\tbreak outer\n\t\t\t}\n\t\t\tf.f()\n\t\t\tif f.done != nil {\n\t\t\t\tf.done <- true\n\t\t\t}\n\t\tcase <-w.publish:\n\t\t\tif w.glw == nil {\n\t\t\t\tbreak outer\n\t\t\t}\n\t\t\tif !theApp.noScreens {\n\t\t\t\ttheApp.RunOnMain(func() {\n\t\t\t\t\tif !w.Activate() {\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\t\t\t\t\tw.glw.SwapBuffers() // note: implicitly does a flush\n\t\t\t\t\t// note: generally don't need this:\n\t\t\t\t\t// gpu.Draw.Clear(true, true)\n\t\t\t\t})\n\t\t\t\tw.publishDone <- struct{}{}\n\t\t\t}\n\t\t}\n\t}\n}", "func (w *MainWindow) Run() {\n\t// https://www.iditect.com/how-to/53890601.html\n\tglfw.SwapInterval(0)\n\n\t//TODO detect best sleeping times using the old frame time and desired refresh rate\n\n\tbegin := time.Now()\n\tlast := time.Duration(0)\n\tfor !w.glfwWindow.ShouldClose() {\n\t\ttime.Sleep(w.FixedPreFrameSleep)\n\n\t\tt := time.Since(begin)\n\t\tdt := (t - last)\n\t\tif w.maxSimStep > 0 && dt > w.maxSimStep {\n\t\t\tdt = w.maxSimStep\n\t\t}\n\t\tw.totalSimTime += dt\n\t\tlast = t\n\n\t\tglfw.WaitEventsTimeout(w.FixedPollEventsTimeout.Seconds())\n\n\t\tgl.Clear(gl.COLOR_BUFFER_BIT | gl.DEPTH_BUFFER_BIT)\n\n\t\tfor _, c := range w.layers {\n\t\t\tc.Update(w, dt.Seconds())\n\t\t}\n\n\t\tfor _, c := range w.layers {\n\t\t\tc.Render(w)\n\t\t}\n\n\t\tw.glfwWindow.SwapBuffers()\n\t}\n\n\t// now gracefully close remaining contexts:\n\tfor len(w.layers) > 0 {\n\t\tw.LeaveUppermostLayer()\n\t}\n}", "func CreateMainWindow() {\n\n\tvBox := tui.NewVBox()\n\tvBox.SetSizePolicy(tui.Minimum, tui.Minimum)\n\tSidebar := tui.NewVBox()\n\tSidebar.SetSizePolicy(tui.Minimum, tui.Minimum)\n\n\tfor _, cmd := range strings.Split(libs.Cmds, \",\") {\n\t\tSidebar.Append(tui.NewLabel(wordwrap.WrapString(cmd, 50)))\n\t}\n\n\tSidebar.SetBorder(true)\n\tSidebar.Prepend(tui.NewLabel(\"***COMMANDS***\"))\n\n\tInput.SetFocused(true)\n\tInput.SetSizePolicy(tui.Expanding, tui.Maximum)\n\n\tinputBox := tui.NewHBox(Input)\n\tinputBox.SetBorder(true)\n\tinputBox.SetSizePolicy(tui.Expanding, tui.Maximum)\n\n\thistoryScroll := tui.NewScrollArea(History)\n\thistoryScroll.SetAutoscrollToBottom(true)\n\thistoryBox := tui.NewVBox(historyScroll)\n\thistoryBox.SetBorder(true)\n\n\tchat := tui.NewVBox(historyBox, inputBox)\n\tchat.SetSizePolicy(tui.Expanding, tui.Expanding)\n\n\t// create root window and add all windows\n\troot := tui.NewHBox(Sidebar, chat)\n\tui, err := tui.New(root)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tui.SetKeybinding(\"Esc\", func() { ui.Quit() })\n\n\tInput.OnSubmit(func(e *tui.Entry) {\n\t\t// this is just to see what command given\n\t\tuserCommand := e.Text()\n\t\tif userCommand == \"\" {\n\t\t\tHistory.Append(tui.NewLabel(\"that is not acceptable command\"))\n\t\t\tHistory.Append(tui.NewLabel(libs.PrintHelp()))\n\t\t} else {\n\t\t\tHistory.Append(tui.NewHBox(\n\t\t\t\ttui.NewLabel(\"Your Command: \" + userCommand),\n\t\t\t))\n\t\t\tHistory.Append(tui.NewHBox(tui.NewLabel(\"\")))\n\n\t\t\tif strings.HasPrefix(userCommand, \"\\\\\") {\n\t\t\t\t// then this is command ..\n\t\t\t\tswitch userCommand {\n\t\t\t\tcase \"\\\\help\":\n\t\t\t\t\tHistory.Append(tui.NewLabel(libs.PrintHelp()))\n\t\t\t\tcase \"\\\\monitor\":\n\t\t\t\t\tHistory.Append(tui.NewLabel(\"Switching to MONITOR mode for device \" + DeviceName))\n\t\t\t\t\tChangeToMonitorMode()\n\t\t\t\tcase \"\\\\managed\":\n\t\t\t\t\tHistory.Append(tui.NewLabel(\"Switching to MANAGED mode for device \" + DeviceName))\n\t\t\t\t\tChangeToManagedMode()\n\t\t\t\tcase \"\\\\exit\":\n\t\t\t\t\tHistory.Append(tui.NewHBox(tui.NewLabel(\"quitting...\")))\n\t\t\t\t\ttime.Sleep(1000 * time.Millisecond)\n\t\t\t\t\t// os.Exit(0)\n\n\t\t\t\t}\n\t\t\t} else if strings.Contains(userCommand, \":\") {\n\t\t\t\t// then this is declaration\n\t\t\t\tcmdSplit := strings.Split(userCommand, \":\")\n\t\t\t\tif cmdSplit[1] == \"\" {\n\t\t\t\t\tHistory.Append(tui.NewLabel(\"that is not acceptable command\"))\n\t\t\t\t\tHistory.Append(tui.NewLabel(libs.PrintHelp()))\n\t\t\t\t} else {\n\t\t\t\t\tswitch cmdSplit[0] {\n\t\t\t\t\tcase \"device\":\n\t\t\t\t\t\tSetDeviceName(cmdSplit[1])\n\t\t\t\t\tdefault:\n\t\t\t\t\t\tHistory.Append(tui.NewLabel(\"there is no such declaration or command\"))\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t} else {\n\t\t\t\tHistory.Append(tui.NewHBox(tui.NewLabel(userCommand + \" is not command or a declaration\")))\n\t\t\t}\n\t\t}\n\t\tInput.SetText(\"\")\n\t})\n\n\tif err := ui.Run(); err != nil {\n\t\tlog.Fatal(err)\n\t}\n}", "func MainWindow(plots plot.Plots) (*walk.MainWindow, *Plot, error) {\n\twithCaption := false\n\tfor _, p := range plots {\n\t\tif p.Caption != nil {\n\t\t\twithCaption = true\n\t\t\tbreak\n\t\t}\n\t}\n\n\tvar ui Plot\n\tvar children []declarative.Widget\n\tif withCaption {\n\t\tchildren = append(children, declarative.VSplitter{\n\t\t\tChildren: []declarative.Widget{\n\t\t\t\tui.BuildPlot(mainMenu(&ui)),\n\t\t\t\tdeclarative.Composite{\n\t\t\t\t\tLayout: declarative.VBox{MarginsZero: true, SpacingZero: true},\n\t\t\t\t\tChildren: []declarative.Widget{\n\t\t\t\t\t\tui.BuildSlider(),\n\t\t\t\t\t\tui.BuildCaption(nil),\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t})\n\t} else {\n\t\tchildren = append(children, declarative.Composite{\n\t\t\tLayout: declarative.VBox{MarginsZero: true, SpacingZero: true},\n\t\t\tChildren: []declarative.Widget{\n\t\t\t\tui.BuildPlot(mainMenu(&ui)),\n\t\t\t\tui.BuildSlider(),\n\t\t\t},\n\t\t})\n\t}\n\n\terr := declarative.MainWindow{\n\t\tAssignTo: &mw,\n\t\tTitle: \"Plot\",\n\t\tSize: declarative.Size{800, 800},\n\t\tOnKeyDown: func(key walk.Key) {\n\t\t\tif key == walk.KeyQ {\n\t\t\t\tos.Exit(0)\n\t\t\t}\n\t\t},\n\t\tLayout: declarative.VBox{MarginsZero: true},\n\t\tChildren: children,\n\t}.Create()\n\treturn mw, &ui, err\n}", "func (sh *Shell) runCommandWindows(line string, timeout time.Duration) (out string, err error) {\n\treturn\n}", "func (w *WindowWidget) Layout(widgets ...Widget) {\n\tif widgets == nil {\n\t\treturn\n\t}\n\n\tws := w.getState()\n\n\tif w.flags&imgui.WindowFlagsNoMove != 0 && w.flags&imgui.WindowFlagsNoResize != 0 {\n\t\timgui.SetNextWindowPos(imgui.Vec2{X: w.x, Y: w.y})\n\t\timgui.SetNextWindowSize(imgui.Vec2{X: w.width, Y: w.height})\n\t} else {\n\t\timgui.SetNextWindowPosV(imgui.Vec2{X: w.x, Y: w.y}, imgui.ConditionFirstUseEver, imgui.Vec2{X: 0, Y: 0})\n\t\timgui.SetNextWindowSizeV(imgui.Vec2{X: w.width, Y: w.height}, imgui.ConditionFirstUseEver)\n\t}\n\n\tif w.bringToFront {\n\t\timgui.SetNextWindowFocus()\n\n\t\tw.bringToFront = false\n\t}\n\n\twidgets = append(widgets,\n\t\tCustom(func() {\n\t\t\thasFocus := IsWindowFocused(0)\n\t\t\tif !hasFocus && ws.hasFocus {\n\t\t\t\tContext.InputHandler.UnregisterWindowShortcuts()\n\t\t\t}\n\n\t\t\tws.hasFocus = hasFocus\n\n\t\t\tws.currentPosition = imgui.WindowPos()\n\t\t\tws.currentSize = imgui.WindowSize()\n\t\t}),\n\t)\n\n\tshowed := imgui.BeginV(Context.FontAtlas.RegisterString(w.title), w.open, int(w.flags))\n\n\tif showed {\n\t\tLayout(widgets).Build()\n\t}\n\n\timgui.End()\n}", "func run() {\n\tglobal.gVariables.Load(wConfigFile)\n\n\t// Initialize window\n\tcfg := pixelgl.WindowConfig{\n\t\tTitle: wWindowTitle,\n\t\tBounds: pixel.R(0, 0, global.gVariables.WindowWidth, global.gVariables.WindowHeight),\n\t\tVSync: true,\n\t}\n\tgWin, err := pixelgl.NewWindow(cfg)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tgWin.SetCursorVisible(false)\n\tglobal.gWin = gWin\n\n\tsetup()\n\n\tgameLoop()\n}", "func main() {\n\tlog.Println(\"Done.\")\n\tapp.Run(func() error {\n\t\tw1 := NewWindow()\n\t\t//w1.Self = w1\n\t\tw1.SetObjID(\"window1\")\n\t\tw1.SetHints(gui.HintResizable)\n\t\tw1.Create(0, 0)\n\t\tw1.SetTitle(\"w1 resizable\")\n\t\tw1.Show()\n\t\t// w1.ToggleFullScreen()\n\t\tw1.Destroy()\n\t\tw2 := NewWindow()\n\t\tw2.SetObjID(\"window2\")\n\t\tw2.Create(200, 200)\n\t\tw2.SetTitle(\"w2 fixed\")\n\t\tw2.Show()\n\t\t// w2.ToggleFullScreen()\n\t\t// runtime.GC()\n\t\treturn nil\n\t}, nil)\n}", "func (n *DesktopNotificator) Run(ctx context.Context) (string, error) {\n\tin := ctx.Value(prevTaskOutputKey).(string)\n\treturn in, n.Send(ctx, in)\n}", "func (v *Viewer) Run(ctx context.Context) {\n\tv.ui.Run()\n}", "func (opt *MainOpt) UpdateObjects() {\n\n\t// With GtkApplicationWindow (does not happen with GtkWindow) I have strange behavior\n\t// when updating window size and position, sometimes width, height is not restored\n\t// successfully, I have tried to figure it out but after a few (long) times I resigned\n\t// myself to using a workaround method, right now I am using a timer that runs 'count'\n\t// times the same commands to finally get the desired result (set window's size with\n\t// the previously saved values).\n\t// count := 5\n\t// glib.TimeoutAdd(uint(64), func() bool {\n\n\tmainObjects.MainWindow.Resize(opt.MainWinWidth, opt.MainWinHeight)\n\tmainObjects.MainWindow.Move(opt.MainWinPosX, opt.MainWinPosY)\n\n\t// count--\n\t// return count > 0\n\t// })\n\n\tmainObjects.CheckbuttonAddReminder.SetActive(opt.Reminder)\n\tmainObjects.CheckbuttonMd4.SetActive(opt.Md4)\n\tmainObjects.CheckbuttonMd5.SetActive(opt.Md5)\n\tmainObjects.CheckbuttonSha1.SetActive(opt.Sha1)\n\tmainObjects.CheckbuttonSha256.SetActive(opt.Sha256)\n\tmainObjects.CheckbuttonSha384.SetActive(opt.Sha384)\n\tmainObjects.CheckbuttonSha512.SetActive(opt.Sha512)\n\tmainObjects.CheckbuttonSha3_256.SetActive(opt.Sha3_256)\n\tmainObjects.CheckbuttonSha3_384.SetActive(opt.Sha3_384)\n\tmainObjects.CheckbuttonSha3_512.SetActive(opt.Sha3_512)\n\tmainObjects.CheckbuttonBlake2b256.SetActive(opt.Blake2b256)\n\tmainObjects.CheckbuttonBlake2b384.SetActive(opt.Blake2b384)\n\tmainObjects.CheckbuttonBlake2b512.SetActive(opt.Blake2b512)\n\tmainObjects.CheckbuttonShowFilename.SetActive(opt.ShowFilename)\n\tmainObjects.CheckbuttonAppendFiles.SetActive(opt.AppendDroppedFiles)\n\tmainObjects.CheckbuttonUseDecimal.SetActive(opt.UseDecimal)\n\tmainObjects.CheckbuttonConcurrentOp.SetActive(opt.ConcurrentOp)\n\tmainObjects.CheckbuttonRecursiveScan.SetActive(opt.RecursiveScan)\n\tmainObjects.CheckbuttonCreateFile.SetActive(opt.MakeOutputFile)\n\n\tif len(opt.CurrentStackPage) > 0 {\n\t\tmainObjects.Stack.SetVisibleChildName(opt.CurrentStackPage)\n\t}\n\tmainObjects.SwitchTreeView.SetActive(opt.SwitchStackPage)\n\tmainObjects.SwitchExpand.SetActive(opt.SwitchExpandState)\n\tSwitchExpandStateSet(mainObjects.SwitchExpand)\n\n\tmainObjects.CheckbuttonShowSplash.SetActive(opt.ShowSplash)\n}", "func (w *Window) Start() error {\n\tnewui, err := lorca.New(\"data:text/html,\"+url.PathEscape(fmt.Sprintf(\"%s\", w)), w.ProfileDir, w.Width, w.Height, w.Args...)\n\tif err != nil {\n\t\treturn err\n\t}\n\tw.ui = newui\n\n\tw.BindChildren(nil)\n\n\t//Apply Bindings\n\tfor _, bound := range w.Bindings {\n\t\terr = newui.Bind(bound.FunctionName, bound.BoundFunction)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (ac *Activity) ResizeWindow(ctx context.Context, border BorderType, to Point, t time.Duration) error {\n\ttask, err := ac.getTaskInfo(ctx)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"could not get task info\")\n\t}\n\n\tif task.windowState != WindowStateNormal && task.windowState != WindowStatePIP {\n\t\treturn errors.Errorf(\"cannot move window in state %d\", int(task.windowState))\n\t}\n\n\t// Default value: center of window.\n\tbounds, err := ac.WindowBounds(ctx)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"could not get activity bounds\")\n\t}\n\tsrc := Point{\n\t\tbounds.Left + bounds.Width/2,\n\t\tbounds.Top + bounds.Height/2,\n\t}\n\n\tborderOffset := borderOffsetForNormal\n\tif task.windowState == WindowStatePIP {\n\t\tborderOffset = borderOffsetForPIP\n\t}\n\n\t// Top & Bottom are exclusive.\n\tif border&BorderTop != 0 {\n\t\tsrc.Y = bounds.Top - borderOffset\n\t} else if border&BorderBottom != 0 {\n\t\tsrc.Y = bounds.Top + bounds.Height + borderOffset\n\t}\n\n\t// Left & Right are exclusive.\n\tif border&BorderLeft != 0 {\n\t\tsrc.X = bounds.Left - borderOffset\n\t} else if border&BorderRight != 0 {\n\t\tsrc.X = bounds.Left + bounds.Width + borderOffset\n\t}\n\n\t// After updating src, clamp it to valid display bounds.\n\tds, err := ac.disp.Size(ctx)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"could not get display size\")\n\t}\n\tsrc.X = int(math.Max(0, math.Min(float64(ds.W-1), float64(src.X))))\n\tsrc.Y = int(math.Max(0, math.Min(float64(ds.H-1), float64(src.Y))))\n\n\treturn ac.swipe(ctx, src, to, t)\n}", "func (l *List) Window() sparta.Window {\n\treturn l.win\n}", "func Window(append ...bool) vecty.Markup {\n\treturn AddClass(window, append...)\n}", "func termboxRun() {\n\tif err := termbox.Init(); err != nil {\n\t\tfmt.Println(err)\n\t\treturn\n\t}\n\tdefer termbox.Close()\n\n\t// create main window\n\tmw := newHomeMVC()\n\tw, h := termbox.Size()\n\tmw.SetDims(Rectangle{0, 0, w, h})\n\tmw.Focus()\n\n\tfor {\n\t\t// update view\n\t\ttermbox.Clear(termbox.ColorDefault, termbox.ColorDefault)\n\t\tmw.Draw()\n\t\ttermbox.Flush()\n\n\t\t// handle next event\n\t\tevent := termbox.PollEvent()\n\t\tswitch event.Type {\n\t\tcase termbox.EventKey:\n\t\t\tswitch {\n\t\t\tcase event.Ch != 0:\n\t\t\t\tmw.HandleRune(event.Ch)\n\t\t\tcase event.Key == termbox.KeyEsc:\n\t\t\t\treturn\n\t\t\tdefault:\n\t\t\t\tmw.HandleKey(event.Key)\n\t\t\t}\n\n\t\tcase termbox.EventResize:\n\t\t\tw, h = event.Width, event.Height\n\t\t\tmw.SetDims(Rectangle{0, 0, w, h})\n\n\t\tcase termbox.EventMouse:\n\t\t\t// mouse events not yet supported\n\n\t\tcase termbox.EventError:\n\t\t\tdrawError(\"Input error:\", event.Err)\n\t\t\ttermbox.Flush()\n\t\t\treturn\n\t\t}\n\t}\n}", "func minWindow(s string, t string) string {\n \n}", "func winGUI() {\n\tinst := w32.GetModuleHandle(\"\")\n\t// Parse arguments from command line.\n\tvar s int64\n\tflag.Int64Var(&s, \"r\", 0, \"initial signed 32-bit seed for dungeon generation\")\n\tflag.Parse()\n\tswitch {\n\tcase s >= -2147483648 && s <= 2147483647:\n\t\t*diablo.FlagRSeed = int32(s)\n\tdefault:\n\t\tpanic(fmt.Errorf(\"invalid seed; expected >= -2147483648 and <= 2147483647; got %d\", s))\n\t}\n\targs := strings.Join(flag.Args(), \" \")\n\tfmt.Println(\"args:\", args)\n\tshow := w32.SW_SHOWDEFAULT\n\tdiablo.WinMain(inst, 0, args, show)\n}", "func (me TxsdPresentationAttributesGraphicsDisplay) IsRunIn() bool { return me.String() == \"run-in\" }", "func (app *App) Run() error {\n\tgo func() {\n\t\t<-app.Quit\n\t\tapp.app.Stop()\n\t}()\n\n\tui.SetConsoleTitle(\"RazTracer\")\n\n\treturn app.app.SetFocus(app).Run()\n}", "func makeWindow(wmctrl_out_line string) []Window {\n\tparts := strings.Split(wmctrl_out_line, \" \")\n\tcomponents := make([]string, 9)\n\tcurrent_component := 0\n\tfor _, text := range parts {\n\t\tif current_component == 8 {\n\t\t\tif text == \"\" {\n\t\t\t\tcomponents[8] += \" \"\n\t\t\t} else {\n\t\t\t\tcomponents[8] += text\n\t\t\t}\n\t\t} else if text != \"\" {\n\t\t\tcomponents[current_component] = text\n\t\t\tcurrent_component += 1\n\t\t}\n\t}\n\n\twid, err0 := strconv.ParseInt(strings.TrimLeft(components[0], \"x0\"), 16, 64)\n\txmin, err1 := strconv.Atoi(components[3])\n\tymin, err2 := strconv.Atoi(components[4])\n\twidth, err3 := strconv.Atoi(components[5])\n\theight, err4 := strconv.Atoi(components[6])\n\tgravity, err5 := strconv.Atoi(components[1])\n\t\n\tproblem := err0 != nil || err1 != nil || err2 != nil || err3 != nil || err4 != nil || err5 != nil\n\t\n\tif gravity == -1 {\n\t\t// Most people have a Desktop, which *is* a window, but we don't want to switch to it\n\t\treturn []Window{}\n\t} else if problem {\n\t\tfmt.Println(\"Something did not parse correctly in this line of output from `wmctrl`. This window will be ignored: \")\n\t\tfmt.Println(wmctrl_out_line)\n\t\treturn []Window{}\n\t} else {\n\t\tw := Window{int(wid), xmin, width, xmin + width, xmin + width/2, ymin, height, ymin + height, ymin + height/2}\n\t\treturn []Window{w}\n\t}\n\n}", "func (m *display) ClrWindow(id int) (int) {\n\tn := m.Send([]byte{0xFE, 0x2C, byte(id)})\n\treturn n\n}", "func mainStartGtk(winTitle string, width, height int, center bool) {\n\tobj = new(MainControlsObj)\n\tgtk.Init(nil)\n\tif newBuilder(mainGlade) == nil {\n\t\t// Init tempDir and Remove it on quit if requested.\n\t\tif doTempDir {\n\t\t\ttempDir = tempMake(Name)\n\t\t\tdefer os.RemoveAll(tempDir)\n\t\t}\n\t\t// Parse Gtk objects\n\t\tgladeObjParser()\n\t\t// Objects Signals initialisations\n\t\tsignalsPropHandler()\n\t\t/* Fill control with images */\n\t\tassignImages()\n\t\t// Set Window Properties\n\t\tif center {\n\t\t\tobj.MainWindow.SetPosition(gtk.WIN_POS_CENTER)\n\t\t}\n\t\tobj.MainWindow.SetTitle(winTitle)\n\t\tobj.MainWindow.SetDefaultSize(width, height)\n\t\tobj.MainWindow.Connect(\"delete-event\", windowDestroy)\n\t\t// Start main application ...\n\t\tmainApplication()\n\t\t//\tStart Gui loop\n\t\tobj.MainWindow.ShowAll()\n\t\tgtk.Main()\n\t} else {\n\t\tlog.Fatal(\"Builder initialisation error.\")\n\t}\n}", "func Main(ctx context.Context, cancelFunc func(), title string, showMainWindow bool) error {\n\tlog.WithFields(log.Fields{\"title\": title, \"showMainWindow\": showMainWindow}).Info(\"Initializing GUI.\")\n\t// Note: ui.Main() calls any functions queued with ui.QueueMain() before the one we provide via parameter.\n\treturn ui.Main(func() {\n\t\twindowTitle = title\n\t\twindow = ui.NewWindow(windowTitle, 600, 50, false)\n\t\tapplyIconToWindow(window.Handle())\n\t\tapplyWindowStyle(window.Handle())\n\n\t\twindow.OnClosing(func(*ui.Window) bool {\n\t\t\tlog.Info(\"User tries to close the window.\")\n\t\t\tcancelFunc()\n\t\t\treturn false\n\t\t})\n\n\t\tpanelDownloadStatus = makeContent()\n\t\twindow.SetChild(panelDownloadStatus)\n\t\twindow.SetMargined(true)\n\n\t\tui.OnShouldQuit(func() bool {\n\t\t\tlog.Info(\"OnShouldQuit().\")\n\t\t\tcancelFunc()\n\t\t\treturn false\n\t\t})\n\n\t\tif showMainWindow {\n\t\t\tcenterWindow(window.Handle())\n\t\t\twindow.Show()\n\t\t\tcenterWindow(window.Handle())\n\t\t}\n\n\t\tgo updateProgressPeriodically(ctx)\n\n\t\tguiInitWaitGroup.Done()\n\t})\n}", "func runTUI() error {\n\n\t// Set function to manage all views and keybindings\n\tclientGui.SetManagerFunc(layout)\n\n\t// Bind keys with functions\n\t_ = clientGui.SetKeybinding(\"\", gocui.KeyCtrlC, gocui.ModNone, quit)\n\t_ = clientGui.SetKeybinding(\"input\", gocui.KeyEnter, gocui.ModNone, send)\n\n\t// Start main event loop of the TUI\n\treturn clientGui.MainLoop()\n}", "func (gn *Gen) ConfigGui() *gi.Window {\n\twidth := 1600\n\theight := 1200\n\n\tgi.SetAppName(\"Gen\")\n\tgi.SetAppAbout(`Gen concatenated strings of syllables`)\n\n\twin := gi.NewMainWindow(\"one\", \"Gen ...\", width, height)\n\n\tvp := win.WinViewport2D()\n\tupdt := vp.UpdateStart()\n\n\tmfr := win.SetMainFrame()\n\n\ttbar := gi.AddNewToolBar(mfr, \"tbar\")\n\ttbar.SetStretchMaxWidth()\n\t// vi.ToolBar = tbar\n\n\tsplit := gi.AddNewSplitView(mfr, \"split\")\n\tsplit.Dim = gi.X\n\tsplit.SetStretchMaxWidth()\n\tsplit.SetStretchMaxHeight()\n\n\tsv := giv.AddNewStructView(split, \"sv\")\n\tsv.SetStruct(gn)\n\tgn.StructView = sv\n\n\t// tv := gi.AddNewTabView(split, \"tv\")\n\n\ttbar.AddAction(gi.ActOpts{Label: \"Gen cat string\", Icon: \"new\", Tooltip: \"Generate a new initial random seed to get different results. By default, Init re-establishes the same initial seed every time.\"}, win.This(),\n\t\tfunc(recv, send ki.Ki, sig int64, data interface{}) {\n\t\t\tgn.CatNoRepeat(gn.syls1)\n\t\t})\n\n\tvp.UpdateEndNoSig(updt)\n\n\t// main menu\n\tappnm := gi.AppName()\n\tmmen := win.MainMenu\n\tmmen.ConfigMenus([]string{appnm, \"File\", \"Edit\", \"Window\"})\n\n\tamen := win.MainMenu.ChildByName(appnm, 0).(*gi.Action)\n\tamen.Menu.AddAppMenu(win)\n\n\temen := win.MainMenu.ChildByName(\"Edit\", 1).(*gi.Action)\n\temen.Menu.AddCopyCutPaste(win)\n\n\tvp.UpdateEndNoSig(updt)\n\n\twin.MainMenuUpdated()\n\treturn win\n}", "func (gn *Gen) ConfigGui() *gi.Window {\n\twidth := 1600\n\theight := 1200\n\n\tgi.SetAppName(\"Gen\")\n\tgi.SetAppAbout(`Gen concatenated strings of syllables`)\n\n\twin := gi.NewMainWindow(\"one\", \"Gen ...\", width, height)\n\n\tvp := win.WinViewport2D()\n\tupdt := vp.UpdateStart()\n\n\tmfr := win.SetMainFrame()\n\n\ttbar := gi.AddNewToolBar(mfr, \"tbar\")\n\ttbar.SetStretchMaxWidth()\n\t// vi.ToolBar = tbar\n\n\tsplit := gi.AddNewSplitView(mfr, \"split\")\n\tsplit.Dim = gi.X\n\tsplit.SetStretchMaxWidth()\n\tsplit.SetStretchMaxHeight()\n\n\tsv := giv.AddNewStructView(split, \"sv\")\n\tsv.SetStruct(gn)\n\tgn.StructView = sv\n\n\t// tv := gi.AddNewTabView(split, \"tv\")\n\n\ttbar.AddAction(gi.ActOpts{Label: \"Reset\", Icon: \"new\", Tooltip: \"\"}, win.This(),\n\t\tfunc(recv, send ki.Ki, sig int64, data interface{}) {\n\t\t\tgn.Reset()\n\t\t})\n\n\ttbar.AddAction(gi.ActOpts{Label: \"Load Params\", Icon: \"new\", Tooltip: \"\"}, win.This(),\n\t\tfunc(recv, send ki.Ki, sig int64, data interface{}) {\n\t\t\tgn.LoadParams()\n\t\t})\n\n\ttbar.AddAction(gi.ActOpts{Label: \"Gen Wavs\", Icon: \"new\", Tooltip: \"Generate the .wav files\"}, win.This(),\n\t\tfunc(recv, send ki.Ki, sig int64, data interface{}) {\n\t\t\tgn.GenWavs()\n\t\t})\n\n\ttbar.AddAction(gi.ActOpts{Label: \"Split Wavs\", Icon: \"new\", Tooltip: \"\"}, win.This(),\n\t\tfunc(recv, send ki.Ki, sig int64, data interface{}) {\n\t\t\tgn.SplitWavs()\n\t\t})\n\n\tvp.UpdateEndNoSig(updt)\n\n\t// main menu\n\tappnm := gi.AppName()\n\tmmen := win.MainMenu\n\tmmen.ConfigMenus([]string{appnm, \"File\", \"Edit\", \"Window\"})\n\n\tamen := win.MainMenu.ChildByName(appnm, 0).(*gi.Action)\n\tamen.Menu.AddAppMenu(win)\n\n\temen := win.MainMenu.ChildByName(\"Edit\", 1).(*gi.Action)\n\temen.Menu.AddCopyCutPaste(win)\n\n\tvp.UpdateEndNoSig(updt)\n\n\twin.MainMenuUpdated()\n\treturn win\n}", "func main() {\n\tdomTarget := dom.GetWindow().Document().GetElementByID(\"app\")\n\n\tr.Render(container.Container(), domTarget)\n}", "func Run(onStart func() error, onExit func(err error)) {\n\tif started {\n\t\tpanic(\"re-enter func Run()\")\n\t}\n\tstarted = true\n\tgOnStart = onStart\n\tgOnExit = onExit\n\tcode := C.winl_event_loop()\n\twinl_on_exit(code)\n}", "func (a *App) Run() error {\n\tdirs := generateDirectories(a.Config.Directories, a.Config.Depth)\n\tif a.Config.QuickMode {\n\t\treturn a.execQuickMode(dirs)\n\t}\n\t// create a gui.Gui struct and run the gui\n\tgui, err := gui.New(a.Config.Mode, dirs)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn gui.Run()\n}", "func TestAddDifferentWindows(t *testing.T) {\n\tp, err := NewProgram()\n\tif err != nil {\n\t\tt.Error(\"Unable to start termbox\")\n\t}\n\n\tdefer p.Close()\n\n\twinNumberBefore := len(p.windows)\n\tshowNumberBefore := len(p.showed)\n\n\tw := NewWindow()\n\terr2 := p.AddWindow(w)\n\n\tif err2 != nil {\n\t\tt.Errorf(\"An error occured (1st win) %s\", fmt.Sprint(err2))\n\t}\n\n\tw2 := NewWindow()\n\terr3 := p.AddWindow(w2)\n\tif err3 != nil {\n\t\tt.Errorf(\"An error occured (2nd win) %s\", fmt.Sprint(err2))\n\t}\n\n\tif winNumberBefore+2 != len(p.windows) {\n\t\tt.Error(\"Window not successfully added to windows's map\")\n\t}\n\n\tif showNumberBefore+2 != len(p.showed) {\n\t\tt.Error(\"Window not successfully added to showed windows map\")\n\t}\n}", "func (w *windowImpl) winLoop() {\n\twinShow := time.NewTimer(time.Second)\nouter:\n\tfor {\n\t\tlog.Println(\"mobile window loop iteration\")\n\t\tselect {\n\t\tcase <-w.winClose:\n\t\t\tbreak outer\n\t\tcase f := <-w.runQueue:\n\t\t\tif w.app.gpu == nil {\n\t\t\t\tbreak outer\n\t\t\t}\n\t\t\tf.f()\n\t\t\tif f.done != nil {\n\t\t\t\tf.done <- true\n\t\t\t}\n\t\tcase <-winShow.C:\n\t\t\tif w.app.gpu == nil {\n\t\t\t\tbreak outer\n\t\t\t}\n\t\t\tw.sendWindowEvent(window.Show)\n\t\t}\n\t}\n}", "func (ref *UIElement) Window() *UIElement {\n\tret, _ := ref.UIElementAttr(WindowAttribute)\n\treturn ret\n}", "func (win *Window) WindowPresent() {\n\twin.Candy().Guify(\"gtk_window_present\", win)\n}", "func (rb *RabinKarp) CurrentWindowText() string {\n\treturn rb.Text[rb.Start:rb.End]\n}", "func TestAddWindow(t *testing.T) {\n\tp, err := NewProgram()\n\tif err != nil {\n\t\tt.Error(\"Unable to start termbox\")\n\t}\n\n\tdefer p.Close()\n\n\twinNumberBefore := len(p.windows)\n\tshowNumberBefore := len(p.showed)\n\n\tw := NewWindow()\n\terr2 := p.AddWindow(w)\n\n\tif err2 != nil {\n\t\tt.Errorf(\"An error occured %s\", fmt.Sprint(err2))\n\t}\n\n\tif winNumberBefore+1 != len(p.windows) {\n\t\tt.Error(\"Window not successfully added to windows's map\")\n\t}\n\n\tif showNumberBefore+1 != len(p.showed) {\n\t\tt.Error(\"Window not successfully added to showed windows map\")\n\t}\n\n\tif w.parent == nil {\n\t\tt.Error(\"The window must have a parent when added to a program\")\n\t}\n}", "func IsWin() bool { return false }", "func Run(app App) error {\n\t// -------------------------------------------------------------------- //\n\t// Create\n\t// -------------------------------------------------------------------- //\n\tsettings := defaultSettings\n\terr := app.OnCreate(&settings)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\tpanic(err)\n\t}\n\n\t// -------------------------------------------------------------------- //\n\t// Init\n\t// -------------------------------------------------------------------- //\n\tjsTge := js.Global().Get(\"tge\")\n\tif settings.Fullscreen {\n\t\tjsTge.Call(\"setFullscreen\", settings.Fullscreen)\n\t} else {\n\t\tjsTge.Call(\"resize\", settings.Width, settings.Height)\n\t}\n\n\tcanvas := jsTge.Call(\"init\")\n\n\t// Instanciate Runtime\n\tbrowserRuntime := _runtimeInstance.(*browserRuntime)\n\tbrowserRuntime.app = app\n\tbrowserRuntime.canvas = &canvas\n\tbrowserRuntime.jsTge = &jsTge\n\tbrowserRuntime.settings = settings\n\tbrowserRuntime.isPaused = true\n\tbrowserRuntime.isStopped = true\n\tbrowserRuntime.done = make(chan bool)\n\n\t// Init plugins\n\tinitPlugins()\n\n\t// Start App\n\terr = app.OnStart(browserRuntime)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\tpanic(err)\n\t}\n\tbrowserRuntime.isStopped = false\n\n\t// Resume App\n\tapp.OnResume()\n\tbrowserRuntime.isPaused = false\n\n\t// Resize App\n\tpublish(ResizeEvent{int32(browserRuntime.canvas.Get(\"clientWidth\").Int()),\n\t\tint32(browserRuntime.canvas.Get(\"clientHeight\").Int())})\n\n\t// -------------------------------------------------------------------- //\n\t// Ticker Loop\n\t// -------------------------------------------------------------------- //\n\tsyncChan := make(chan interface{})\n\telapsedTpsTime := time.Duration(0)\n\tgo func() {\n\t\tfor !browserRuntime.isStopped {\n\t\t\tif !browserRuntime.isPaused {\n\t\t\t\tnow := time.Now()\n\t\t\t\tapp.OnTick(elapsedTpsTime, syncChan)\n\t\t\t\telapsedTpsTime = time.Since(now)\n\t\t\t}\n\t\t}\n\t}()\n\n\t// -------------------------------------------------------------------- //\n\t// Callbacks\n\t// -------------------------------------------------------------------- //\n\n\t// Resize\n\tresizeEvtCb := js.FuncOf(func(this js.Value, args []js.Value) interface{} {\n\t\tif !browserRuntime.isStopped {\n\t\t\tw := int32(browserRuntime.canvas.Get(\"clientWidth\").Int())\n\t\t\th := int32(browserRuntime.canvas.Get(\"clientHeight\").Int())\n\t\t\tjsTge.Call(\"resize\", w, h)\n\t\t\tpublish(ResizeEvent{\n\t\t\t\tWidth: w,\n\t\t\t\tHeight: h,\n\t\t\t})\n\t\t}\n\t\treturn false\n\t})\n\tdefer resizeEvtCb.Release()\n\tjs.Global().Call(\"addEventListener\", \"resize\", resizeEvtCb)\n\n\t// Focus\n\tblurEvtCb := js.FuncOf(func(this js.Value, args []js.Value) interface{} {\n\t\tif !browserRuntime.isStopped && !browserRuntime.isPaused {\n\t\t\tgo func() {\n\t\t\t\tbrowserRuntime.isPaused = true\n\t\t\t\tbrowserRuntime.app.OnPause()\n\t\t\t}()\n\t\t}\n\t\treturn false\n\t})\n\tdefer blurEvtCb.Release()\n\tbrowserRuntime.canvas.Call(\"addEventListener\", \"blur\", blurEvtCb)\n\n\tfocuseEvtCb := js.FuncOf(func(this js.Value, args []js.Value) interface{} {\n\t\tif !browserRuntime.isStopped && browserRuntime.isPaused {\n\t\t\t//Called in go routine in case of asset loading in resume (blocking)\n\t\t\tgo func() {\n\t\t\t\tbrowserRuntime.app.OnResume()\n\t\t\t\tbrowserRuntime.isPaused = false\n\t\t\t}()\n\t\t}\n\t\treturn false\n\t})\n\tdefer focuseEvtCb.Release()\n\tbrowserRuntime.canvas.Call(\"addEventListener\", \"focus\", focuseEvtCb)\n\n\t// Destroy\n\tbeforeunloadEvtCb := js.FuncOf(func(this js.Value, args []js.Value) interface{} {\n\t\tif !browserRuntime.isStopped {\n\t\t\tbrowserRuntime.Stop()\n\t\t}\n\t\treturn false\n\t})\n\tdefer beforeunloadEvtCb.Release()\n\tjs.Global().Call(\"addEventListener\", \"beforeunload\", beforeunloadEvtCb)\n\n\t// MouseButtonEvent\n\tif (settings.EventMask & MouseButtonEventEnabled) != 0 {\n\t\tmouseDownEvtCb := js.FuncOf(func(this js.Value, args []js.Value) interface{} {\n\t\t\tif !browserRuntime.isStopped && !browserRuntime.isPaused {\n\t\t\t\tevent := args[0]\n\t\t\t\tbutton := ButtonNone\n\t\t\t\tswitch event.Get(\"button\").Int() {\n\t\t\t\tcase 0:\n\t\t\t\t\tbutton = ButtonLeft\n\t\t\t\tcase 1:\n\t\t\t\t\tbutton = ButtonMiddle\n\t\t\t\tcase 2:\n\t\t\t\t\tbutton = ButtonRight\n\t\t\t\t}\n\t\t\t\tpublish(MouseEvent{\n\t\t\t\t\tX: int32(event.Get(\"offsetX\").Int()),\n\t\t\t\t\tY: int32(event.Get(\"offsetY\").Int()),\n\t\t\t\t\tButton: button,\n\t\t\t\t\tType: TypeDown,\n\t\t\t\t})\n\t\t\t}\n\t\t\treturn false\n\t\t})\n\t\tdefer mouseDownEvtCb.Release()\n\t\tbrowserRuntime.canvas.Call(\"addEventListener\", \"mousedown\", mouseDownEvtCb)\n\n\t\ttouchDownEvtCb := js.FuncOf(func(this js.Value, args []js.Value) interface{} {\n\t\t\tif !browserRuntime.isStopped && !browserRuntime.isPaused {\n\t\t\t\tevent := args[0]\n\t\t\t\tevent.Call(\"preventDefault\")\n\t\t\t\ttouchList := event.Get(\"touches\")\n\t\t\t\ttouchListLen := touchList.Get(\"length\").Int()\n\t\t\t\tfor i := 0; i < touchListLen; i++ {\n\t\t\t\t\tbutton := ButtonNone\n\t\t\t\t\tswitch i {\n\t\t\t\t\tcase 0:\n\t\t\t\t\t\tbutton = TouchFirst\n\t\t\t\t\tcase 1:\n\t\t\t\t\t\tbutton = TouchSecond\n\t\t\t\t\tcase 2:\n\t\t\t\t\t\tbutton = TouchThird\n\t\t\t\t\t}\n\t\t\t\t\ttouch := touchList.Index(i)\n\t\t\t\t\tpublish(MouseEvent{\n\t\t\t\t\t\tX: int32(touch.Get(\"clientX\").Int()),\n\t\t\t\t\t\tY: int32(touch.Get(\"clientY\").Int()),\n\t\t\t\t\t\tButton: button,\n\t\t\t\t\t\tType: TypeDown,\n\t\t\t\t\t})\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn false\n\t\t})\n\t\tdefer touchDownEvtCb.Release()\n\t\tbrowserRuntime.canvas.Call(\"addEventListener\", \"touchstart\", touchDownEvtCb)\n\n\t\tmouseUpEvtCb := js.FuncOf(func(this js.Value, args []js.Value) interface{} {\n\t\t\tif !browserRuntime.isStopped && !browserRuntime.isPaused {\n\t\t\t\tevent := args[0]\n\t\t\t\tbutton := ButtonNone\n\t\t\t\tswitch event.Get(\"button\").Int() {\n\t\t\t\tcase 0:\n\t\t\t\t\tbutton = ButtonLeft\n\t\t\t\tcase 1:\n\t\t\t\t\tbutton = ButtonMiddle\n\t\t\t\tcase 2:\n\t\t\t\t\tbutton = ButtonRight\n\t\t\t\t}\n\t\t\t\tpublish(MouseEvent{\n\t\t\t\t\tX: int32(event.Get(\"offsetX\").Int()),\n\t\t\t\t\tY: int32(event.Get(\"offsetY\").Int()),\n\t\t\t\t\tButton: button,\n\t\t\t\t\tType: TypeUp,\n\t\t\t\t})\n\t\t\t}\n\t\t\treturn false\n\t\t})\n\t\tdefer mouseUpEvtCb.Release()\n\t\tbrowserRuntime.canvas.Call(\"addEventListener\", \"mouseup\", mouseUpEvtCb)\n\n\t\ttouchUpEvtCb := js.FuncOf(func(this js.Value, args []js.Value) interface{} {\n\t\t\tif !browserRuntime.isStopped && !browserRuntime.isPaused {\n\t\t\t\tevent := args[0]\n\t\t\t\tevent.Call(\"preventDefault\")\n\t\t\t\ttouchList := event.Get(\"changedTouches\")\n\t\t\t\ttouchListLen := touchList.Get(\"length\").Int()\n\t\t\t\tfor i := 0; i < touchListLen; i++ {\n\t\t\t\t\tbutton := ButtonNone\n\t\t\t\t\tswitch i {\n\t\t\t\t\tcase 0:\n\t\t\t\t\t\tbutton = TouchFirst\n\t\t\t\t\tcase 1:\n\t\t\t\t\t\tbutton = TouchSecond\n\t\t\t\t\tcase 2:\n\t\t\t\t\t\tbutton = TouchThird\n\t\t\t\t\t}\n\t\t\t\t\ttouch := touchList.Index(i)\n\t\t\t\t\tpublish(MouseEvent{\n\t\t\t\t\t\tX: int32(touch.Get(\"clientX\").Int()),\n\t\t\t\t\t\tY: int32(touch.Get(\"clientY\").Int()),\n\t\t\t\t\t\tButton: button,\n\t\t\t\t\t\tType: TypeUp,\n\t\t\t\t\t})\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn false\n\t\t})\n\t\tdefer touchUpEvtCb.Release()\n\t\tbrowserRuntime.canvas.Call(\"addEventListener\", \"touchend\", touchUpEvtCb)\n\t}\n\n\t// MouseMotionEventEnabled\n\tif (settings.EventMask & MouseMotionEventEnabled) != 0 {\n\t\tmouseMoveEvtCb := js.FuncOf(func(this js.Value, args []js.Value) interface{} {\n\t\t\tif !browserRuntime.isStopped && !browserRuntime.isPaused {\n\t\t\t\tevent := args[0]\n\t\t\t\tpublish(MouseEvent{\n\t\t\t\t\tX: int32(event.Get(\"clientX\").Int()),\n\t\t\t\t\tY: int32(event.Get(\"clientY\").Int()),\n\t\t\t\t\tButton: ButtonNone,\n\t\t\t\t\tType: TypeMove,\n\t\t\t\t})\n\t\t\t}\n\t\t\treturn false\n\t\t})\n\t\tdefer mouseMoveEvtCb.Release()\n\t\tbrowserRuntime.canvas.Call(\"addEventListener\", \"mousemove\", mouseMoveEvtCb)\n\n\t\ttouchMoveEvtCb := js.FuncOf(func(this js.Value, args []js.Value) interface{} {\n\t\t\tif !browserRuntime.isStopped && !browserRuntime.isPaused {\n\t\t\t\tevent := args[0]\n\t\t\t\tevent.Call(\"preventDefault\")\n\t\t\t\ttouchList := event.Get(\"touches\")\n\t\t\t\ttouchListLen := touchList.Get(\"length\").Int()\n\t\t\t\tfor i := 0; i < touchListLen; i++ {\n\t\t\t\t\tbutton := ButtonNone\n\t\t\t\t\tswitch i {\n\t\t\t\t\tcase 0:\n\t\t\t\t\t\tbutton = TouchFirst\n\t\t\t\t\tcase 1:\n\t\t\t\t\t\tbutton = TouchSecond\n\t\t\t\t\tcase 2:\n\t\t\t\t\t\tbutton = TouchThird\n\t\t\t\t\t}\n\t\t\t\t\ttouch := touchList.Index(i)\n\t\t\t\t\tpublish(MouseEvent{\n\t\t\t\t\t\tX: int32(touch.Get(\"clientX\").Int()),\n\t\t\t\t\t\tY: int32(touch.Get(\"clientY\").Int()),\n\t\t\t\t\t\tButton: button,\n\t\t\t\t\t\tType: TypeMove,\n\t\t\t\t\t})\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn false\n\t\t})\n\t\tdefer touchMoveEvtCb.Release()\n\t\tbrowserRuntime.canvas.Call(\"addEventListener\", \"touchmove\", touchMoveEvtCb)\n\t}\n\n\t// ScrollEvent\n\tif (settings.EventMask & ScrollEventEnabled) != 0 {\n\t\twheelEvtCb := js.FuncOf(func(this js.Value, args []js.Value) interface{} {\n\t\t\tif !browserRuntime.isStopped && !browserRuntime.isPaused {\n\t\t\t\tevent := args[0]\n\t\t\t\tevent.Call(\"preventDefault\")\n\t\t\t\tx := float64(event.Get(\"deltaX\").Int())\n\t\t\t\ty := float64(event.Get(\"deltaY\").Int())\n\t\t\t\tif x != 0 {\n\t\t\t\t\tx = x / math.Abs(x)\n\t\t\t\t}\n\t\t\t\tif y != 0 {\n\t\t\t\t\ty = y / math.Abs(y)\n\t\t\t\t}\n\t\t\t\tpublish(ScrollEvent{\n\t\t\t\t\tX: int32(x),\n\t\t\t\t\tY: -int32(y),\n\t\t\t\t})\n\t\t\t}\n\t\t\treturn false\n\t\t})\n\t\tdefer wheelEvtCb.Release()\n\t\tbrowserRuntime.canvas.Call(\"addEventListener\", \"wheel\", wheelEvtCb)\n\t}\n\n\t// KeyEvent\n\tif (settings.EventMask & KeyEventEnabled) != 0 {\n\t\tkeyDownEvtCb := js.FuncOf(func(this js.Value, args []js.Value) interface{} {\n\t\t\tif !browserRuntime.isStopped && !browserRuntime.isPaused {\n\t\t\t\tevent := args[0]\n\t\t\t\tevent.Call(\"preventDefault\")\n\t\t\t\tkeyCode := event.Get(\"key\").String()\n\t\t\t\tpublish(KeyEvent{\n\t\t\t\t\tKey: keyMap[keyCode],\n\t\t\t\t\tValue: keyCode,\n\t\t\t\t\tType: TypeDown,\n\t\t\t\t})\n\t\t\t}\n\t\t\treturn false\n\t\t})\n\t\tdefer keyDownEvtCb.Release()\n\t\tbrowserRuntime.canvas.Call(\"addEventListener\", \"keydown\", keyDownEvtCb)\n\n\t\tkeyUpEvtCb := js.FuncOf(func(this js.Value, args []js.Value) interface{} {\n\t\t\tif !browserRuntime.isStopped && !browserRuntime.isPaused {\n\t\t\t\tevent := args[0]\n\t\t\t\tevent.Call(\"preventDefault\")\n\t\t\t\tkeyCode := event.Get(\"key\").String()\n\t\t\t\tpublish(KeyEvent{\n\t\t\t\t\tKey: keyMap[keyCode],\n\t\t\t\t\tValue: keyCode,\n\t\t\t\t\tType: TypeUp,\n\t\t\t\t})\n\t\t\t}\n\t\t\treturn false\n\t\t})\n\t\tdefer keyUpEvtCb.Release()\n\t\tbrowserRuntime.canvas.Call(\"addEventListener\", \"keyup\", keyUpEvtCb)\n\t}\n\n\t// -------------------------------------------------------------------- //\n\t// Render Loop\n\t// -------------------------------------------------------------------- //\n\tvar renderFrame js.Func\n\telapsedFpsTime := time.Duration(0)\n\trenderFrame = js.FuncOf(func(this js.Value, args []js.Value) interface{} {\n\t\tif !browserRuntime.isPaused {\n\t\t\tnow := time.Now()\n\t\t\tapp.OnRender(elapsedFpsTime, syncChan)\n\t\t\telapsedFpsTime = time.Since(now)\n\t\t}\n\t\tif !browserRuntime.isStopped {\n\t\t\tjs.Global().Call(\"requestAnimationFrame\", renderFrame)\n\t\t} else {\n\t\t\tbrowserRuntime.done <- true\n\t\t}\n\t\treturn false\n\t})\n\tjs.Global().Call(\"requestAnimationFrame\", renderFrame)\n\n\t<-browserRuntime.done\n\n\trenderFrame.Release()\n\tjsTge.Call(\"stop\")\n\n\tnoExit := make(chan int)\n\t<-noExit\n\n\treturn nil\n}", "func (lw *listWin) refresh() {\n\tvar inds []int\n\tfor i, task := range file.Tasks {\n\t\tok := true\n\t\tfor _, filter := range lw.filters {\n\t\t\tif !task.HasTag(filter) {\n\t\t\t\tok = false\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif ok {\n\t\t\tinds = append(inds, i)\n\t\t}\n\t}\n\n\tsort.Sort(sorter{inds, file.Tasks, lw.less})\n\n\tprojs := make(map[string]bool)\n\tctxs := make(map[string]bool)\n\n\tif err := lw.Addr(\",\"); err != nil {\n\t\tdie(1, \"Failed to set address for %s: %s\", lw.title, err)\n\t}\n\n\tfor _, i := range inds {\n\t\ttask := file.Tasks[i]\n\t\tif _, err := fmt.Fprintf(lw.Data, \"%5d. %s\\n\", i+1, task.String()); err != nil {\n\t\t\tdie(1, \"Failed to refresh window %s: %s\", lw.title, err)\n\t\t}\n\t\tfor _, t := range task.Tags(todotxt.ProjectTag) {\n\t\t\tprojs[t] = true\n\t\t}\n\t\tfor _, t := range task.Tags(todotxt.ContextTag) {\n\t\t\tctxs[t] = true\n\t\t}\n\t}\n\n\tif err := lw.Addr(\"#0\"); err != nil {\n\t\tdie(1, \"Failed to write address to %s: %s\", lw.title, err)\n\t}\n\tif err := lw.Ctl(\"dot=addr\"); err != nil {\n\t\tdie(1, \"Failed to write dot=addr to %s ctl: %s\", lw.title, err)\n\t}\n\tif err := lw.Ctl(\"show\"); err != nil {\n\t\tdie(1, \"Failed to write show to %s ctl: %s\", lw.title, err)\n\t}\n\tif err := lw.Ctl(\"clean\"); err != nil {\n\t\tdie(1, \"Failed to write clean to %s ctl: %s\", lw.title, err)\n\t}\n}", "func main() {\n\tgo func() {\n\t\tw := app.NewWindow(\n\t\t\tapp.Title(\"Gopher-Garden\"),\n\t\t\tapp.Size(unit.Dp(ui.WidthPx+500), unit.Dp(ui.HeightPx)))\n\t\tu := ui.NewUi(w)\n\t\tif err := u.Loop(); err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t\tos.Exit(0)\n\t}()\n\tapp.Main()\n}", "func (d *Dashboard) Run() {\n\terr := termbox.Init()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tdefer termbox.Close()\n\ttermbox.SetInputMode(termbox.InputEsc)\n\ttermbox.SetOutputMode(termbox.Output256)\n\tif err := d.redraw(); err != nil {\n\t\tfmt.Println(\"Error: %s\", err)\n\t\treturn\n\t}\n\teventChannel := make(chan termbox.Event, 10)\n\tgo d.termboxEventPoller(eventChannel)\n\nmainloop:\n\tfor {\n\t\tselect {\n\t\tcase ev, ok := <-eventChannel:\n\t\t\tif !ok {\n\t\t\t\tbreak mainloop\n\t\t\t}\n\t\t\tswitch ev.Type {\n\t\t\tcase termbox.EventKey:\n\t\t\t\tswitch ev.Key {\n\t\t\t\tcase termbox.KeyEsc:\n\t\t\t\t\tbreak mainloop\n\t\t\t\tdefault:\n\t\t\t\t\tif ev.Ch == 'q' {\n\t\t\t\t\t\tbreak mainloop\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\tcase termbox.EventError:\n\t\t\t\tfmt.Printf(\"Error: %s\\n\", ev.Err)\n\t\t\t\tbreak mainloop\n\t\t\tcase termbox.EventResize:\n\t\t\t\ttermbox.Clear(termbox.ColorDefault, termbox.ColorDefault)\n\t\t\t\tif err := d.redraw(); err != nil {\n\t\t\t\t\tfmt.Println(\"Error: %s\", err)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t\tif err := d.redraw(); err != nil {\n\t\t\t\tfmt.Println(\"Error: %s\", err)\n\t\t\t\treturn\n\t\t\t}\n\t\tcase buildUpdate := <-d.fetcher.BuildChannel():\n\t\t\td.builds = buildUpdate.builds\n\t\t\td.err = buildUpdate.err\n\t\t\tif err := d.redraw(); err != nil {\n\t\t\t\tfmt.Println(\"Error: %s\", err)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}\n}", "func (a *Application) Run(update func(rend *renderer.Renderer, deltaTime time.Duration)) {\r\n\r\n\t// Initialize start and frame time\r\n\ta.startTime = time.Now()\r\n\ta.frameStart = time.Now()\r\n\r\n\t// Set up recurring calls to user's update function\r\n\tfor true {\r\n\t\t// If Exit() was called or there was an attempt to close the window dispatch OnExit event for subscribers.\r\n\t\t// If no subscriber cancelled the event, terminate the application.\r\n\t\tif a.IWindow.(*window.GlfwWindow).ShouldClose() {\r\n\t\t\ta.Dispatch(OnExit, nil)\r\n\t\t\t// TODO allow for cancelling exit e.g. showing dialog asking the user if he/she wants to save changes\r\n\t\t\t// if exit was cancelled {\r\n\t\t\t// a.IWindow.(*window.GlfwWindow).SetShouldClose(false)\r\n\t\t\t// } else {\r\n\t\t\tbreak\r\n\t\t\t// }\r\n\t\t}\r\n\t\t// Update frame start and frame delta\r\n\t\tnow := time.Now()\r\n\t\ta.frameDelta = now.Sub(a.frameStart)\r\n\t\ta.frameStart = now\r\n\t\t// Call user's update function\r\n\t\tupdate(a.renderer, a.frameDelta)\r\n\t\t// Swap buffers and poll events\r\n\t\ta.IWindow.(*window.GlfwWindow).SwapBuffers()\r\n\t\ta.IWindow.(*window.GlfwWindow).PollEvents()\r\n\t}\r\n\r\n\t// Close default audio device\r\n\tif a.audioDev != nil {\r\n\t\tal.CloseDevice(a.audioDev)\r\n\t}\r\n\t// Destroy window\r\n\ta.Destroy()\r\n}", "func WindowMessageName(msg UINT) string {\n\tvar messages = map[UINT]string{\n\t\tWM_ACTIVATE: \"WM_ACTIVATE\",\n\t\tWM_CHAR: \"WM_CHAR\",\n\t\tWM_COMMAND: \"WM_COMMAND\",\n\t\tWM_COPYDATA: \"WM_COPYDATA\",\n\t\tWM_CREATE: \"WM_CREATE\",\n\t\tWM_DESTROY: \"WM_DESTROY\",\n\t\tWM_DROPFILES: \"WM_DROPFILES\",\n\t\tWM_INITDIALOG: \"WM_INITDIALOG\",\n\t\tWM_KEYDOWN: \"WM_KEYDOWN\",\n\t\tWM_KILLFOCUS: \"WM_KILLFOCUS\",\n\t\tWM_LBUTTONDOWN: \"WM_LBUTTONDOWN\",\n\t\tWM_LBUTTONUP: \"WM_LBUTTONUP\",\n\t\tWM_MOUSEMOVE: \"WM_MOUSEMOVE\",\n\t\tWM_MOUSEWHEEL: \"WM_MOUSEWHEEL\",\n\t\tWM_MOVE: \"WM_MOVE\",\n\t\tWM_NULL: \"WM_NULL\",\n\t\tWM_PAINT: \"WM_PAINT\",\n\t\tWM_SETFOCUS: \"WM_SETFOCUS\",\n\t\tWM_SETFONT: \"WM_SETFONT\",\n\t\tWM_SIZE: \"WM_SIZE\",\n\t\t// other messages:\n\t\tWM_ENABLE: \"WM_ENABLE\",\n\t\tWM_SETREDRAW: \"WM_SETREDRAW\",\n\t\tWM_SETTEXT: \"WM_SETTEXT\",\n\t\tWM_GETTEXT: \"WM_GETTEXT\",\n\t\tWM_GETTEXTLENGTH: \"WM_GETTEXTLENGTH\",\n\t\tWM_CLOSE: \"WM_CLOSE\",\n\t\tWM_QUERYENDSESSION: \"WM_QUERYENDSESSION\",\n\t\tWM_QUIT: \"WM_QUIT\",\n\t\tWM_QUERYOPEN: \"WM_QUERYOPEN\",\n\t\tWM_ERASEBKGND: \"WM_ERASEBKGND\",\n\t\tWM_SYSCOLORCHANGE: \"WM_SYSCOLORCHANGE\",\n\t\tWM_ENDSESSION: \"WM_ENDSESSION\",\n\t\tWM_SHOWWINDOW: \"WM_SHOWWINDOW\",\n\t\tWM_WININICHANGE: \"WM_WININICHANGE\",\n\t\tWM_DEVMODECHANGE: \"WM_DEVMODECHANGE\",\n\t\tWM_ACTIVATEAPP: \"WM_ACTIVATEAPP\",\n\t\tWM_FONTCHANGE: \"WM_FONTCHANGE\",\n\t\tWM_TIMECHANGE: \"WM_TIMECHANGE\",\n\t\tWM_CANCELMODE: \"WM_CANCELMODE\",\n\t\tWM_SETCURSOR: \"WM_SETCURSOR\",\n\t\tWM_MOUSEACTIVATE: \"WM_MOUSEACTIVATE\",\n\t\tWM_CHILDACTIVATE: \"WM_CHILDACTIVATE\",\n\t\tWM_QUEUESYNC: \"WM_QUEUESYNC\",\n\t\tWM_GETMINMAXINFO: \"WM_GETMINMAXINFO\",\n\t\tWM_PAINTICON: \"WM_PAINTICON\",\n\t\tWM_ICONERASEBKGND: \"WM_ICONERASEBKGND\",\n\t\tWM_NEXTDLGCTL: \"WM_NEXTDLGCTL\",\n\t\tWM_SPOOLERSTATUS: \"WM_SPOOLERSTATUS\",\n\t\tWM_DRAWITEM: \"WM_DRAWITEM\",\n\t\tWM_MEASUREITEM: \"WM_MEASUREITEM\",\n\t\tWM_DELETEITEM: \"WM_DELETEITEM\",\n\t\tWM_VKEYTOITEM: \"WM_VKEYTOITEM\",\n\t\tWM_CHARTOITEM: \"WM_CHARTOITEM\",\n\t\tWM_GETFONT: \"WM_GETFONT\",\n\t\tWM_SETHOTKEY: \"WM_SETHOTKEY\",\n\t\tWM_GETHOTKEY: \"WM_GETHOTKEY\",\n\t\tWM_QUERYDRAGICON: \"WM_QUERYDRAGICON\",\n\t\tWM_COMPAREITEM: \"WM_COMPAREITEM\",\n\t\tWM_GETOBJECT: \"WM_GETOBJECT\",\n\t\tWM_COMPACTING: \"WM_COMPACTING\",\n\t\tWM_COMMNOTIFY: \"WM_COMMNOTIFY\",\n\t\tWM_WINDOWPOSCHANGING: \"WM_WINDOWPOSCHANGING\",\n\t\tWM_WINDOWPOSCHANGED: \"WM_WINDOWPOSCHANGED\",\n\t\tWM_POWER: \"WM_POWER\",\n\t\tWM_CANCELJOURNAL: \"WM_CANCELJOURNAL\",\n\t\tWM_NOTIFY: \"WM_NOTIFY\",\n\t\tWM_INPUTLANGCHANGEREQUEST: \"WM_INPUTLANGCHANGEREQUEST\",\n\t\tWM_INPUTLANGCHANGE: \"WM_INPUTLANGCHANGE\",\n\t\tWM_TCARD: \"WM_TCARD\",\n\t\tWM_HELP: \"WM_HELP\",\n\t\tWM_USERCHANGED: \"WM_USERCHANGED\",\n\t\tWM_NOTIFYFORMAT: \"WM_NOTIFYFORMAT\",\n\t\tWM_CONTEXTMENU: \"WM_CONTEXTMENU\",\n\t\tWM_STYLECHANGING: \"WM_STYLECHANGING\",\n\t\tWM_STYLECHANGED: \"WM_STYLECHANGED\",\n\t\tWM_DISPLAYCHANGE: \"WM_DISPLAYCHANGE\",\n\t\tWM_GETICON: \"WM_GETICON\",\n\t\tWM_SETICON: \"WM_SETICON\",\n\t\tWM_NCCREATE: \"WM_NCCREATE\",\n\t\tWM_NCDESTROY: \"WM_NCDESTROY\",\n\t\tWM_NCCALCSIZE: \"WM_NCCALCSIZE\",\n\t\tWM_NCHITTEST: \"WM_NCHITTEST\",\n\t\tWM_NCPAINT: \"WM_NCPAINT\",\n\t\tWM_NCACTIVATE: \"WM_NCACTIVATE\",\n\t\tWM_GETDLGCODE: \"WM_GETDLGCODE\",\n\t\tWM_SYNCPAINT: \"WM_SYNCPAINT\",\n\t\tWM_NCMOUSEMOVE: \"WM_NCMOUSEMOVE\",\n\t\tWM_NCLBUTTONDOWN: \"WM_NCLBUTTONDOWN\",\n\t\tWM_NCLBUTTONUP: \"WM_NCLBUTTONUP\",\n\t\tWM_NCLBUTTONDBLCLK: \"WM_NCLBUTTONDBLCLK\",\n\t\tWM_NCRBUTTONDOWN: \"WM_NCRBUTTONDOWN\",\n\t\tWM_NCRBUTTONUP: \"WM_NCRBUTTONUP\",\n\t\tWM_NCRBUTTONDBLCLK: \"WM_NCRBUTTONDBLCLK\",\n\t\tWM_NCMBUTTONDOWN: \"WM_NCMBUTTONDOWN\",\n\t\tWM_NCMBUTTONUP: \"WM_NCMBUTTONUP\",\n\t\tWM_NCMBUTTONDBLCLK: \"WM_NCMBUTTONDBLCLK\",\n\t\tWM_NCXBUTTONDOWN: \"WM_NCXBUTTONDOWN\",\n\t\tWM_NCXBUTTONUP: \"WM_NCXBUTTONUP\",\n\t\tWM_NCXBUTTONDBLCLK: \"WM_NCXBUTTONDBLCLK\",\n\t\tWM_INPUT_DEVICE_CHANGE: \"WM_INPUT_DEVICE_CHANGE\",\n\t\tWM_INPUT: \"WM_INPUT\",\n\t\tWM_KEYUP: \"WM_KEYUP\",\n\t\tWM_DEADCHAR: \"WM_DEADCHAR\",\n\t\tWM_SYSKEYDOWN: \"WM_SYSKEYDOWN\",\n\t\tWM_SYSKEYUP: \"WM_SYSKEYUP\",\n\t\tWM_SYSCHAR: \"WM_SYSCHAR\",\n\t\tWM_SYSDEADCHAR: \"WM_SYSDEADCHAR\",\n\t\tWM_UNICHAR: \"WM_UNICHAR\",\n\t\tWM_IME_STARTCOMPOSITION: \"WM_IME_STARTCOMPOSITION\",\n\t\tWM_IME_ENDCOMPOSITION: \"WM_IME_ENDCOMPOSITION\",\n\t\tWM_IME_COMPOSITION: \"WM_IME_COMPOSITION\",\n\t\tWM_SYSCOMMAND: \"WM_SYSCOMMAND\",\n\t\tWM_TIMER: \"WM_TIMER\",\n\t\tWM_HSCROLL: \"WM_HSCROLL\",\n\t\tWM_VSCROLL: \"WM_VSCROLL\",\n\t\tWM_INITMENU: \"WM_INITMENU\",\n\t\tWM_INITMENUPOPUP: \"WM_INITMENUPOPUP\",\n\t\tWM_GESTURE: \"WM_GESTURE\",\n\t\tWM_GESTURENOTIFY: \"WM_GESTURENOTIFY\",\n\t\tWM_MENUSELECT: \"WM_MENUSELECT\",\n\t\tWM_MENUCHAR: \"WM_MENUCHAR\",\n\t\tWM_ENTERIDLE: \"WM_ENTERIDLE\",\n\t\tWM_MENURBUTTONUP: \"WM_MENURBUTTONUP\",\n\t\tWM_MENUDRAG: \"WM_MENUDRAG\",\n\t\tWM_MENUGETOBJECT: \"WM_MENUGETOBJECT\",\n\t\tWM_UNINITMENUPOPUP: \"WM_UNINITMENUPOPUP\",\n\t\tWM_MENUCOMMAND: \"WM_MENUCOMMAND\",\n\t\tWM_CHANGEUISTATE: \"WM_CHANGEUISTATE\",\n\t\tWM_UPDATEUISTATE: \"WM_UPDATEUISTATE\",\n\t\tWM_QUERYUISTATE: \"WM_QUERYUISTATE\",\n\t\tWM_CTLCOLORMSGBOX: \"WM_CTLCOLORMSGBOX\",\n\t\tWM_CTLCOLOREDIT: \"WM_CTLCOLOREDIT\",\n\t\tWM_CTLCOLORLISTBOX: \"WM_CTLCOLORLISTBOX\",\n\t\tWM_CTLCOLORBTN: \"WM_CTLCOLORBTN\",\n\t\tWM_CTLCOLORDLG: \"WM_CTLCOLORDLG\",\n\t\tWM_CTLCOLORSCROLLBAR: \"WM_CTLCOLORSCROLLBAR\",\n\t\tWM_CTLCOLORSTATIC: \"WM_CTLCOLORSTATIC\",\n\t\tWM_LBUTTONDBLCLK: \"WM_LBUTTONDBLCLK\",\n\t\tWM_RBUTTONDOWN: \"WM_RBUTTONDOWN\",\n\t\tWM_RBUTTONUP: \"WM_RBUTTONUP\",\n\t\tWM_RBUTTONDBLCLK: \"WM_RBUTTONDBLCLK\",\n\t\tWM_MBUTTONDOWN: \"WM_MBUTTONDOWN\",\n\t\tWM_MBUTTONUP: \"WM_MBUTTONUP\",\n\t\tWM_MBUTTONDBLCLK: \"WM_MBUTTONDBLCLK\",\n\t\tWM_XBUTTONDOWN: \"WM_XBUTTONDOWN\",\n\t\tWM_XBUTTONUP: \"WM_XBUTTONUP\",\n\t\tWM_XBUTTONDBLCLK: \"WM_XBUTTONDBLCLK\",\n\t\tWM_MOUSEHWHEEL: \"WM_MOUSEHWHEEL\",\n\t\tWM_PARENTNOTIFY: \"WM_PARENTNOTIFY\",\n\t\tWM_ENTERMENULOOP: \"WM_ENTERMENULOOP\",\n\t\tWM_EXITMENULOOP: \"WM_EXITMENULOOP\",\n\t\tWM_NEXTMENU: \"WM_NEXTMENU\",\n\t\tWM_SIZING: \"WM_SIZING\",\n\t\tWM_CAPTURECHANGED: \"WM_CAPTURECHANGED\",\n\t\tWM_MOVING: \"WM_MOVING\",\n\t\tWM_POWERBROADCAST: \"WM_POWERBROADCAST\",\n\t\tWM_DEVICECHANGE: \"WM_DEVICECHANGE\",\n\t\tWM_MDICREATE: \"WM_MDICREATE\",\n\t\tWM_MDIDESTROY: \"WM_MDIDESTROY\",\n\t\tWM_MDIACTIVATE: \"WM_MDIACTIVATE\",\n\t\tWM_MDIRESTORE: \"WM_MDIRESTORE\",\n\t\tWM_MDINEXT: \"WM_MDINEXT\",\n\t\tWM_MDIMAXIMIZE: \"WM_MDIMAXIMIZE\",\n\t\tWM_MDITILE: \"WM_MDITILE\",\n\t\tWM_MDICASCADE: \"WM_MDICASCADE\",\n\t\tWM_MDIICONARRANGE: \"WM_MDIICONARRANGE\",\n\t\tWM_MDIGETACTIVE: \"WM_MDIGETACTIVE\",\n\t\tWM_MDISETMENU: \"WM_MDISETMENU\",\n\t\tWM_ENTERSIZEMOVE: \"WM_ENTERSIZEMOVE\",\n\t\tWM_EXITSIZEMOVE: \"WM_EXITSIZEMOVE\",\n\t\tWM_MDIREFRESHMENU: \"WM_MDIREFRESHMENU\",\n\t\tWM_TOUCH: \"WM_TOUCH\",\n\t\tWM_IME_SETCONTEXT: \"WM_IME_SETCONTEXT\",\n\t\tWM_IME_NOTIFY: \"WM_IME_NOTIFY\",\n\t\tWM_IME_CONTROL: \"WM_IME_CONTROL\",\n\t\tWM_IME_COMPOSITIONFULL: \"WM_IME_COMPOSITIONFULL\",\n\t\tWM_IME_SELECT: \"WM_IME_SELECT\",\n\t\tWM_IME_CHAR: \"WM_IME_CHAR\",\n\t\tWM_IME_REQUEST: \"WM_IME_REQUEST\",\n\t\tWM_IME_KEYDOWN: \"WM_IME_KEYDOWN\",\n\t\tWM_IME_KEYUP: \"WM_IME_KEYUP\",\n\t\tWM_NCMOUSEHOVER: \"WM_NCMOUSEHOVER\",\n\t\tWM_MOUSEHOVER: \"WM_MOUSEHOVER\",\n\t\tWM_NCMOUSELEAVE: \"WM_NCMOUSELEAVE\",\n\t\tWM_MOUSELEAVE: \"WM_MOUSELEAVE\",\n\t\tWM_WTSSESSION_CHANGE: \"WM_WTSSESSION_CHANGE\",\n\t\tWM_TABLET_FIRST: \"WM_TABLET_FIRST\",\n\t\tWM_TABLET_LAST: \"WM_TABLET_LAST\",\n\t\tWM_CUT: \"WM_CUT\",\n\t\tWM_COPY: \"WM_COPY\",\n\t\tWM_PASTE: \"WM_PASTE\",\n\t\tWM_CLEAR: \"WM_CLEAR\",\n\t\tWM_UNDO: \"WM_UNDO\",\n\t\tWM_RENDERFORMAT: \"WM_RENDERFORMAT\",\n\t\tWM_RENDERALLFORMATS: \"WM_RENDERALLFORMATS\",\n\t\tWM_DESTROYCLIPBOARD: \"WM_DESTROYCLIPBOARD\",\n\t\tWM_DRAWCLIPBOARD: \"WM_DRAWCLIPBOARD\",\n\t\tWM_PAINTCLIPBOARD: \"WM_PAINTCLIPBOARD\",\n\t\tWM_VSCROLLCLIPBOARD: \"WM_VSCROLLCLIPBOARD\",\n\t\tWM_SIZECLIPBOARD: \"WM_SIZECLIPBOARD\",\n\t\tWM_ASKCBFORMATNAME: \"WM_ASKCBFORMATNAME\",\n\t\tWM_CHANGECBCHAIN: \"WM_CHANGECBCHAIN\",\n\t\tWM_HSCROLLCLIPBOARD: \"WM_HSCROLLCLIPBOARD\",\n\t\tWM_QUERYNEWPALETTE: \"WM_QUERYNEWPALETTE\",\n\t\tWM_PALETTEISCHANGING: \"WM_PALETTEISCHANGING\",\n\t\tWM_PALETTECHANGED: \"WM_PALETTECHANGED\",\n\t\tWM_HOTKEY: \"WM_HOTKEY\",\n\t\tWM_PRINT: \"WM_PRINT\",\n\t\tWM_PRINTCLIENT: \"WM_PRINTCLIENT\",\n\t\tWM_APPCOMMAND: \"WM_APPCOMMAND\",\n\t\tWM_THEMECHANGED: \"WM_THEMECHANGED\",\n\t\tWM_CLIPBOARDUPDATE: \"WM_CLIPBOARDUPDATE\",\n\t\tWM_DWMCOMPOSITIONCHANGED: \"WM_DWMCOMPOSITIONCHANGED\",\n\t\tWM_DWMNCRENDERINGCHANGED: \"WM_DWMNCRENDERINGCHANGED\",\n\t\tWM_DWMCOLORIZATIONCOLORCHANGED: \"WM_DWMCOLORIZATIONCOLORCHANGED\",\n\t\tWM_DWMWINDOWMAXIMIZEDCHANGE: \"WM_DWMWINDOWMAXIMIZEDCHANGE\",\n\t\tWM_DWMSENDICONICTHUMBNAIL: \"WM_DWMSENDICONICTHUMBNAIL\",\n\t\tWM_DWMSENDICONICLIVEPREVIEWBITMAP: \"WM_DWMSENDICONICLIVEPREVIEWBITMAP\",\n\t\tWM_GETTITLEBARINFOEX: \"WM_GETTITLEBARINFOEX\",\n\t\tWM_HANDHELDFIRST: \"WM_HANDHELDFIRST\",\n\t\tWM_HANDHELDLAST: \"WM_HANDHELDLAST\",\n\t\tWM_AFXFIRST: \"WM_AFXFIRST\",\n\t\tWM_AFXLAST: \"WM_AFXLAST\",\n\t\tWM_PENWINFIRST: \"WM_PENWINFIRST\",\n\t\tWM_PENWINLAST: \"WM_PENWINLAST\",\n\t\tWM_USER: \"WM_USER\",\n\t\tWM_APP: \"WM_APP\",\n\t}\n\t// messages that equal to existing messages:\n\t// WM_IME_KEYLAST\n\t// WM_SETTINGCHANGE\n\t// WM_MOUSEFIRST\n\t// WM_MOUSELAST\n\tif s, ok := messages[msg]; ok {\n\t\treturn s\n\t}\n\treturn \"WM_...UNKNOWN\"\n}", "func (s *Scroll) Window() sparta.Window {\n\treturn s.win\n}", "func (m *MainWindow) checkPositionAndSize() {\n\tglib.IdleAdd(func() bool {\n\t\tm.window_pos_x, m.window_pos_y = m.window.GetPosition()\n\t\tm.window_width, m.window_height = m.window.GetSize()\n\n\t\tm.hpane.SetPosition(m.window_width - m.pane_negative_position)\n\t\treturn false\n\t})\n}", "func GuiWindowBox(bounds Rectangle, title string) bool {\n\tcbounds, _ := *(*C.Rectangle)(unsafe.Pointer(&bounds)), cgoAllocsUnknown\n\ttitle = safeString(title)\n\tctitle, _ := unpackPCharString(title)\n\t__ret := C.GuiWindowBox(cbounds, ctitle)\n\truntime.KeepAlive(title)\n\t__v := (bool)(__ret)\n\treturn __v\n}", "func (wc *WebClient) WindowFullscreen() {}", "func (c *CloseFolderList) Run(w *lime.Window) error {\n\tfor _, folder := range w.Project().Folders() {\n\t\tw.Project().RemoveFolder(folder)\n\t}\n\treturn nil\n}", "func TestAddSameWindow(t *testing.T) {\n\tp, err := NewProgram()\n\tif err != nil {\n\t\tt.Error(\"Unable to start termbox\")\n\t}\n\n\tdefer p.Close()\n\n\tw := NewWindow()\n\terr2 := p.AddWindow(w)\n\tif err2 != nil {\n\t\tt.Errorf(\"An error occured %s\", fmt.Sprint(err2))\n\t}\n\n\terr3 := p.AddWindow(w)\n\tif err3 == nil {\n\t\tt.Error(\"An error was expected\")\n\t}\n}", "func CreateWindow(title string, width, height uint32, r Framer) (*Window, error) {\n\tconst className = \"goui.wingui.window\"\n\n\tres := &Window{\n\t\tHandle: tHWND(0),\n\t\tRunning: true,\n\t\tRoot: r,\n\t}\n\n\tinstance, err := getModuleHandle()\n\tif err != nil {\n\t\treturn res, err\n\t}\n\n\tcursor, err := loadCursor(cIDC_ARROW)\n\tif err != nil {\n\t\treturn res, err\n\t}\n\n\tcallback := func(hwnd tHWND, msg uint32, wparam, lparam uintptr) uintptr {\n\t\treturn mainLoop(res, hwnd, msg, wparam, lparam)\n\t}\n\n\twndclass := tWNDCLASSEXW{\n\t\twndProc: syscall.NewCallback(callback),\n\t\tinstance: instance,\n\t\tcursor: cursor,\n\t\tbackground: cCOLOR_WINDOW,\n\t\tclassName: syscall.StringToUTF16Ptr(className),\n\t}\n\twndclass.size = uint32(unsafe.Sizeof(wndclass))\n\n\tif _, err = registerClassEx(&wndclass); err != nil {\n\t\treturn res, err\n\t}\n\n\tres.Handle, err = createWindow(\n\t\tclassName,\n\t\ttitle,\n\t\tcWS_VISIBLE|cWS_OVERLAPPEDWINDOW,\n\t\tcSW_USE_DEFAULT,\n\t\tcSW_USE_DEFAULT,\n\t\tint32(width),\n\t\tint32(height),\n\t\t0,\n\t\t0,\n\t\tinstance,\n\t)\n\tif err != nil {\n\t\treturn res, err\n\t}\n\n\treturn res, nil\n}", "func (a *App) Run() error {\n\t// Run state updater\n\tctx, cancel := context.WithCancel(context.Background())\n\tgo a.updater.Run(ctx)\n\ta.cancelStateUpdate = cancel\n\n\t// run gui\n\treturn a.gui.Run()\n}", "func WindowContent(append ...bool) vecty.Markup {\n\treturn AddClass(wContent, append...)\n}", "func NewWindow(width, height int) (w *Window, err error) {\n\n\tw = new(Window)\n\tw.width, w.height = width, height\n\tw.showed = false\n\n\tw.xu, err = xgbutil.NewConn()\n\tif err != nil {\n\t\treturn\n\t}\n\n\tw.conn = w.xu.Conn()\n\t// screen := w.xu.Screen()\n\n\tw.win, err = xwindow.Generate(w.xu)\n\tif err != nil {\n\t\tlog.Printf(\"ERROR: %#+v\\n\", err)\n\t\treturn\n\t}\n\n\tkeybind.Initialize(w.xu)\n\n\terr = w.win.CreateChecked(w.xu.RootWin(), 0, 0, width, height, xproto.CwBackPixel, 0x606060ff)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tw.win.Listen(AllEventsMask)\n\n\terr = icccm.WmProtocolsSet(w.xu, w.win.Id, []string{\"WM_DELETE_WINDOW\"})\n\tif err != nil {\n\t\tfmt.Fprintln(os.Stderr, err)\n\t\terr = nil\n\t}\n\n\tw.bufferLck = &sync.Mutex{}\n\tw.buffer = xgraphics.New(w.xu, image.Rect(0, 0, width, height))\n\tw.buffer.XSurfaceSet(w.win.Id)\n\n\t// I /think/ XDraw actually sends data to server?\n\tw.buffer.XDraw()\n\t// I /think/ XPaint tells the server to paint image to window\n\tw.buffer.XPaint(w.win.Id)\n\n\tkeyMap, modMap := keybind.MapsGet(w.xu)\n\tkeybind.KeyMapSet(w.xu, keyMap)\n\tkeybind.ModMapSet(w.xu, modMap)\n\n\tw.events = make(chan interface{})\n\n\tw.SetIcon(Gordon)\n\tw.SetIconName(\"Go\")\n\n\tgo w.handleEvents()\n\n\treturn\n}", "func (e *Emulator) Run() {\n\tdriver.Main(func(s screen.Screen) {\n\t\te.s = s\n\t\t// Calculate initial window size based on whatever our gutter/pixel pitch currently is.\n\t\tdims := e.matrixWithMarginsRect()\n\t\twopts := &screen.NewWindowOptions{\n\t\t\tTitle: \"RGB LED Matrix Emulator\",\n\t\t\tWidth: dims.Max.X,\n\t\t\tHeight: dims.Max.Y,\n\t\t}\n\t\tw, err := s.NewWindow(wopts)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\te.w = w\n\t\tfirstRender := true\n\t\tfor {\n\t\t\tevn := w.NextEvent()\n\t\t\tswitch evn := evn.(type) {\n\t\t\tcase key.Event:\n\t\t\t\tif evn.Code == key.CodeEscape {\n\t\t\t\t\te.Close()\n\t\t\t\t}\n\t\t\tcase paint.Event:\n\t\t\t\te.Render()\n\t\t\tcase size.Event:\n\t\t\t\tif evn.WidthPx == 0 && evn.HeightPx == 0 {\n\t\t\t\t\te.Close()\n\t\t\t\t}\n\t\t\t\te.sz = evn\n\t\t\t\tif firstRender {\n\t\t\t\t\te.Render()\n\t\t\t\t\tfirstRender = false\n\t\t\t\t}\n\t\t\tcase error:\n\t\t\t\tfmt.Println(\"render:\", err)\n\t\t\t}\n\t\t}\n\t})\n}", "func (p *Window) Class() string {\n\treturn (`winl.Window`)\n}", "func TestRemoveWindow(t *testing.T) {\n\tp, err := NewProgram()\n\tif err != nil {\n\t\tt.Error(\"Unable to start termbox\")\n\t}\n\n\tdefer p.Close()\n\n\tw := NewWindow()\n\terr2 := p.AddWindow(w)\n\n\tif err2 != nil {\n\t\tt.Errorf(\"An error occured (1st win) %s\", fmt.Sprint(err2))\n\t}\n\n\twinNumberBefore := len(p.windows)\n\tshowNumberBefore := len(p.showed)\n\n\terr3 := p.RemoveWindow(w)\n\tif err3 != nil {\n\t\tt.Errorf(\"An error occured (1st win) %s\", fmt.Sprint(err2))\n\t}\n\n\tif winNumberBefore-1 != len(p.windows) {\n\t\tt.Error(\"Window not successfully removed from windows's map\")\n\t}\n\n\tif showNumberBefore-1 != len(p.showed) {\n\t\tt.Error(\"Window not successfully removed from showed windows map\")\n\t}\n\n\tif w.parent != nil {\n\t\tt.Error(\"Window must not have a parent when it's removed from the program\")\n\t}\n}", "func takeWindowScreenshot(ctx context.Context, tconn *chrome.TestConn) error {\n\tui := uiauto.New(tconn)\n\n\tif _, err := filesapp.Launch(ctx, tconn); err != nil {\n\t\treturn errors.Wrap(err, \"failed to launch the app\")\n\t}\n\n\tactiveWindow, err := ash.GetActiveWindow(ctx, tconn)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to find active window\")\n\t}\n\tcenterPoint := activeWindow.BoundsInRoot.CenterPoint()\n\n\tif err = uiauto.Combine(\"take window screenshot\",\n\t\tui.LeftClick(nodewith.Role(role.ToggleButton).Name(\"Screenshot\")),\n\t\tui.LeftClick(nodewith.Role(role.ToggleButton).Name(\"Take window screenshot\")),\n\t\tmouse.Move(tconn, centerPoint, time.Second), // Different names for clamshell/tablet mode.\n\t\tui.LeftClick(nodewith.Role(role.Window).First()), // Click on the center of root window to take the screenshot.\n\t)(ctx); err != nil {\n\t\treturn errors.Wrap(err, \"failed to take window screenshot\")\n\t}\n\n\treturn nil\n}", "func (w *WidgetImplement) FindWindow() IWindow {\n\tparent := w.Parent()\n\tif parent == nil {\n\t\tpanic(\"Widget:internal error (could not find parent window)\")\n\t}\n\treturn parent.FindWindow()\n}", "func (ref *UIElement) MainWindow() *UIElement {\n\tret, _ := ref.UIElementAttr(MainWindowAttribute)\n\treturn ret\n}", "func RunTablet(ctx, closeCtx context.Context, tconn *chrome.TestConn, ui *uiauto.Context, pc pointer.Context) (retErr error) {\n\tconst (\n\t\ttimeout = 10 * time.Second\n\t\tduration = 2 * time.Second\n\t\tdoubleTapInterval = 100 * time.Millisecond\n\t)\n\n\t// Gets primary display info and interesting drag points.\n\tinfo, err := display.GetPrimaryInfo(ctx, tconn)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to get the primary display info\")\n\t}\n\n\tsplitViewDragPoints := []coords.Point{\n\t\tinfo.WorkArea.CenterPoint(),\n\t\tcoords.NewPoint(info.WorkArea.Left+info.WorkArea.Width/4, info.WorkArea.CenterY()),\n\t\tcoords.NewPoint(info.WorkArea.Left+info.WorkArea.Width-1, info.WorkArea.CenterY()),\n\t}\n\tsnapRightPoint := coords.NewPoint(info.WorkArea.Right()-1, info.WorkArea.CenterY())\n\n\ttabStripButton := nodewith.Role(role.Button).ClassName(\"WebUITabCounterButton\").First()\n\tif err := pc.Click(tabStripButton)(ctx); err != nil {\n\t\treturn errors.Wrap(err, \"failed to click the tab strip button\")\n\t}\n\n\t// Get the first tab location with a polling interval of 2 seconds (meaning\n\t// wait until the location is stable for 2 seconds) to work around a\n\t// glitchy animation that sometimes happens when bringing up the tab strip.\n\tfirstTab := nodewith.Role(role.Tab).First()\n\tfirstTabRect, err := ui.WithInterval(2*time.Second).Location(ctx, firstTab)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to get the location of the first tab\")\n\t}\n\n\t// Drag the first tab in the tab strip and snap it to the right.\n\tdefer cleanUp(ctx, action.Named(\n\t\t\"maximize the window\",\n\t\tfunc(ctx context.Context) error {\n\t\t\tws, err := getAllNonPipWindows(ctx, tconn)\n\t\t\tif err != nil {\n\t\t\t\treturn errors.Wrap(err, \"failed to get windows\")\n\t\t\t}\n\t\t\tif len(ws) != 1 {\n\t\t\t\treturn errors.Errorf(\"unexpected number of windows: got %d; want 1\", len(ws))\n\t\t\t}\n\t\t\tif err := ash.SetWindowStateAndWait(ctx, tconn, ws[0].ID, ash.WindowStateMaximized); err != nil {\n\t\t\t\treturn errors.Wrap(err, \"failed to set browser window state to \\\"Maximized\\\"\")\n\t\t\t}\n\t\t\treturn nil\n\t\t},\n\t), &retErr)\n\tif err := pc.Drag(firstTabRect.CenterPoint(),\n\t\tuiauto.Sleep(time.Second),\n\t\tpc.DragTo(snapRightPoint, 3*time.Second),\n\t)(ctx); err != nil {\n\t\treturn errors.Wrap(err, \"failed to drag a tab to snap to the right\")\n\t}\n\tdefer cleanUp(ctx, action.Named(\n\t\t\"recombine the browser tabs\",\n\t\tfunc(ctx context.Context) error {\n\t\t\treturn combineTabs(ctx, tconn, ui, pc, duration)\n\t\t},\n\t), &retErr)\n\n\tif err := testing.Poll(ctx, func(ctx context.Context) error {\n\t\tws, err := getAllNonPipWindows(ctx, tconn)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"failed to obtain the window list\")\n\t\t}\n\t\tif len(ws) != 2 {\n\t\t\treturn errors.Errorf(\"should be 2 windows, got %v\", len(ws))\n\t\t}\n\t\tif (ws[1].State == ash.WindowStateLeftSnapped && ws[0].State == ash.WindowStateRightSnapped) ||\n\t\t\t(ws[0].State == ash.WindowStateLeftSnapped && ws[1].State == ash.WindowStateRightSnapped) {\n\t\t\treturn nil\n\t\t}\n\t\treturn errors.New(\"browser windows are not snapped yet\")\n\t}, &testing.PollOptions{Timeout: timeout}); err != nil {\n\t\treturn errors.Wrap(err, \"failed to wait for browser windows to be snapped correctly\")\n\t}\n\n\t// Create a second virtual desk.\n\tif err := ash.CreateNewDesk(ctx, tconn); err != nil {\n\t\treturn errors.Wrap(err, \"failed to create a new desk\")\n\t}\n\tdefer cleanUp(closeCtx, action.Named(\n\t\t\"remove extra desk\",\n\t\tfunc(ctx context.Context) error {\n\t\t\treturn removeExtraDesk(ctx, tconn)\n\t\t},\n\t), &retErr)\n\n\tkw, err := input.Keyboard(ctx)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to open the keyboard\")\n\t}\n\tdefer cleanUp(closeCtx, action.Named(\n\t\t\"close the keyboard\",\n\t\tfunc(ctx context.Context) error {\n\t\t\treturn kw.Close()\n\t\t},\n\t), &retErr)\n\ttopRow, err := input.KeyboardTopRowLayout(ctx, kw)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to obtain the top-row layout\")\n\t}\n\tenterOverview := kw.AccelAction(topRow.SelectTask)\n\t// Exercise split view resize functionality.\n\tif err := exerciseSplitViewResize(ctx, tconn, ui, pc, enterOverview, splitViewDragPoints...); err != nil {\n\t\treturn errors.Wrap(err, \"failed to exercise split view resize functionality with two browser windows\")\n\t}\n\n\treturn nil\n}", "func doShowMeshWindow(compMesh *component.Mesh) {\n\tmeshWindow := uiman.GetWindow(fmt.Sprintf(\"%s%s\", compMeshWindowID, compMesh.Name))\n\tif meshWindow == nil {\n\t\tcreateMeshWindow(compMesh, meshWndX, meshWndY)\n\t}\n}", "func (win *Window) Width() int {\n\tsize := C.sfRenderWindow_getSize(win.win)\n\treturn int(size.x)\n}", "func (o *NewWindowOptions) Fixup() {\n\tsc := TheApp.Screen(0)\n\tscsz := sc.Geometry.Size() // window coords size\n\n\tif o.Size.X <= 0 {\n\t\to.StdPixels = false\n\t\to.Size.X = int(0.8 * float32(scsz.X) * sc.DevicePixelRatio)\n\t}\n\tif o.Size.Y <= 0 {\n\t\to.StdPixels = false\n\t\to.Size.Y = int(0.8 * float32(scsz.Y) * sc.DevicePixelRatio)\n\t}\n\n\to.Size, o.Pos = sc.ConstrainWinGeom(o.Size, o.Pos)\n\tif o.Pos.X == 0 && o.Pos.Y == 0 {\n\t\twsz := sc.WinSizeFmPix(o.Size)\n\t\tdialog, modal, _, _ := WindowFlagsToBool(o.Flags)\n\t\tnw := TheApp.NWindows()\n\t\tif nw > 0 {\n\t\t\tlastw := TheApp.Window(nw - 1)\n\t\t\tlsz := lastw.WinSize()\n\t\t\tlp := lastw.Position()\n\n\t\t\tnwbig := wsz.X > lsz.X || wsz.Y > lsz.Y\n\n\t\t\tif modal || dialog || !nwbig { // place centered on top of current\n\t\t\t\tctrx := lp.X + (lsz.X / 2)\n\t\t\t\tctry := lp.Y + (lsz.Y / 2)\n\t\t\t\to.Pos.X = ctrx - wsz.X/2\n\t\t\t\to.Pos.Y = ctry - wsz.Y/2\n\t\t\t} else { // cascade to right\n\t\t\t\to.Pos.X = lp.X + lsz.X // tile to right -- could depend on orientation\n\t\t\t\to.Pos.Y = lp.Y + 72 // and move down a bit\n\t\t\t}\n\t\t} else { // center in screen\n\t\t\to.Pos.X = scsz.X/2 - wsz.X/2\n\t\t\to.Pos.Y = scsz.Y/2 - wsz.Y/2\n\t\t}\n\t\to.Size, o.Pos = sc.ConstrainWinGeom(o.Size, o.Pos) // make sure ok\n\t}\n}", "func (c *PromptAddFolder) Run(w *lime.Window) error {\n\tdir := viewDirectory(w.ActiveView())\n\tfe := lime.GetEditor().Frontend()\n\tfolders := fe.Prompt(\"Open file\", dir, lime.PROMPT_ONLY_FOLDER|lime.PROMPT_SELECT_MULTIPLE)\n\tfor _, folder := range folders {\n\t\tw.Project().AddFolder(folder)\n\t}\n\treturn nil\n}", "func (c *thickClient) Run() {\n\tpixelgl.Run(c.run)\n}", "func (win *Window) Display() {\n\tC.sfRenderWindow_display(win.win)\n}", "func Run() {\n\troot.Execute()\n}", "func (wc *WebClient) WindowSetColour(colour int) {\n}", "func Start(hwnd W.HWND) uintptr {\n\tvar message W.MSG\n//\tW.ShowWindow(hwnd,W.SW_SHOW)\n\tfor{\n\t\tif W.GetMessage(&message, 0, 0, 0) == 0 {break}\n\t\tW.TranslateMessage(&message)\n\t\tW.DispatchMessage(&message)\n\t}\n\treturn 0\n}", "func (m *display) SetTextWindow(id int) (int) {\n\tdata := []byte{0xFE, 0x2A, byte(id)}\n\tn1 := m.Send(data)\n return n1\n}", "func RunMinUI(n *node.Node) {\n\ts := minui.New(Config, n)\n\ts.Run()\n}", "func GuiWindowBox(bounds Rectangle, title string) bool {\n\tctitle := C.CString(title)\n\tdefer C.free(unsafe.Pointer(ctitle))\n\tcbounds := *bounds.cptr()\n\tres := C.GuiWindowBox(cbounds, ctitle)\n\treturn bool(res)\n}", "func (l *List) RemoveWindow() {\n\tl.win = nil\n}", "func (m *display) InitTextWindow(id, x1, y1, x2, y2, font, charSpace, lineSpace, scroll int) int {\n\tdata := []byte{0xFE, 0x2B, byte(id), byte(x1), byte(y1), byte(x2), byte(y2)}\n\tfontb := make([]byte,2)\n\tbinary.LittleEndian.PutUint16(fontb, uint16(font))\n\tdata = append(data, fontb...)\n\tdata = append(data, byte(charSpace))\n\tdata = append(data, byte(lineSpace))\n\tdata = append(data, byte(scroll))\n\n\tn1 := m.Send(data)\n\treturn n1\n}", "func (v *App) Run() {\n\t// log.Println(\"Starting App polling\")\n\tv.running = true\n\tsdl.SetEventFilterFunc(v.filterEvent, nil)\n\n\t// sdl.SetHint(sdl.HINT_RENDER_SCALE_QUALITY, \"linear\")\n\tv.renderer.SetDrawColor(64, 64, 64, 255)\n\tv.renderer.Clear()\n\n\t// v.pointsGraph.SetSeries(v.pointsGraph.Accessor())\n\tv.spikeGraph.SetSeries(nil)\n\n\tfor v.running {\n\t\tsdl.PumpEvents()\n\n\t\tv.renderer.Clear()\n\n\t\t// v.pointsGraph.MarkDirty(true)\n\t\tv.spikeGraph.MarkDirty(true)\n\n\t\tif samples.PoiSamples != nil {\n\t\t\tdraw := v.spikeGraph.Check()\n\t\t\tif draw {\n\t\t\t\tv.spikeGraph.DrawAt(0, 100)\n\t\t\t}\n\t\t}\n\n\t\tv.expoGraph.DrawAt(0, 300)\n\n\t\tv.txtSimStatus.Draw()\n\t\tv.txtActiveProperty.Draw()\n\n\t\tv.window.UpdateSurface()\n\n\t\t// sdl.Delay(17)\n\t\ttime.Sleep(time.Millisecond * 100)\n\t}\n\n\tv.shutdown()\n}", "func (w *Window) Base() app.Window {\n\treturn w\n}", "func WindowSize() (w, h int) {\n\treturn windowSize()\n}", "func win() {\n\tfmt.Println(\"win\", i, j)\n\tos.Exit(1)\n}" ]
[ "0.60856587", "0.60131717", "0.5949291", "0.5947787", "0.5947787", "0.58854175", "0.5827745", "0.5742636", "0.5742636", "0.5660663", "0.5650461", "0.5602295", "0.5563611", "0.55624324", "0.5503847", "0.5464583", "0.54509115", "0.5449197", "0.5443993", "0.5436137", "0.5410743", "0.54017246", "0.5372587", "0.53571385", "0.53401357", "0.5275388", "0.5238067", "0.52235425", "0.5211092", "0.520786", "0.5207466", "0.518148", "0.516912", "0.51556784", "0.514169", "0.5140647", "0.51256156", "0.5099033", "0.5092412", "0.50837505", "0.507117", "0.5062216", "0.50545573", "0.5043169", "0.5023831", "0.5016801", "0.5009283", "0.5002626", "0.50002277", "0.49921337", "0.49872065", "0.49754694", "0.49749616", "0.49733326", "0.49695525", "0.49695417", "0.4968981", "0.4968907", "0.495739", "0.49477452", "0.49427846", "0.49143493", "0.4912616", "0.48981178", "0.48954073", "0.48932472", "0.48916095", "0.48883158", "0.4869405", "0.48623306", "0.4859348", "0.4858011", "0.4843883", "0.4841292", "0.4839711", "0.48395708", "0.48388514", "0.48276478", "0.4821412", "0.48177513", "0.48161757", "0.48102772", "0.48093203", "0.48075676", "0.47998267", "0.47881734", "0.47867584", "0.47860363", "0.47787893", "0.47787553", "0.4773595", "0.4771242", "0.4769398", "0.47608843", "0.47547024", "0.47522575", "0.47509262", "0.4749381", "0.474207", "0.47412777" ]
0.6676458
0
updateMaps recomputes the sample rate based on the countList.
func (t *WindowedThroughput) updateMaps() { currentIndex := t.indexGenerator.GetCurrentIndex() lookbackIndexes := t.indexGenerator.DurationToIndexes(t.LookbackFrequencyDuration) aggregateCounts := t.countList.AggregateCounts(currentIndex, lookbackIndexes) // Apply the same aggregation algorithm as total throughput // Short circuit if no traffic t.numKeys = len(aggregateCounts) if t.numKeys == 0 { // no traffic during the last period. t.lock.Lock() defer t.lock.Unlock() t.savedSampleRates = make(map[string]int) return } // figure out our target throughput per key over the lookback window. totalGoalThroughput := t.GoalThroughputPerSec * t.LookbackFrequencyDuration.Seconds() // floor the throughput but min should be 1 event per bucket per time period throughputPerKey := math.Max(1, float64(totalGoalThroughput)/float64(t.numKeys)) // for each key, calculate sample rate by dividing counted events by the // desired number of events newSavedSampleRates := make(map[string]int) for k, v := range aggregateCounts { rate := int(math.Max(1, (float64(v) / float64(throughputPerKey)))) newSavedSampleRates[k] = rate } // save newly calculated sample rates t.lock.Lock() defer t.lock.Unlock() t.savedSampleRates = newSavedSampleRates }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *CountMinSketch) Update(key []byte, count uint64) {\n\tfor r, c := range s.locations(key) {\n\t\ts.count[r][c] += count\n\t}\n}", "func UpdateCount(key Type, val int) {\n\tmutex.Lock()\n\tdefer mutex.Unlock()\n\tCountStats[key] = CountStats[key] + val\n}", "func (g GaugeMaps) Update(resources Resources) {\n\tg[cpu].Update(resources.GetCPU())\n\tg[mem].Update(resources.GetMem())\n\tg[disk].Update(resources.GetDisk())\n\tg[gpu].Update(resources.GetGPU())\n}", "func updateCounters(t *db.TestDetails, counters map[string]map[string]pair, isBlocked bool) {\n\tvar category string\n\tvar typ string\n\n\tif isApiTest(t.TestSet) {\n\t\tcategory = \"api\"\n\t} else {\n\t\tcategory = \"app\"\n\t}\n\n\tif t.Type == \"\" {\n\t\ttyp = \"unknown\"\n\t} else {\n\t\ttyp = strings.ToLower(t.Type)\n\t}\n\n\tif _, ok := counters[category]; !ok {\n\t\tcounters[category] = make(map[string]pair)\n\t}\n\n\tval := counters[category][typ]\n\tif isBlocked {\n\t\tval.blocked++\n\t} else {\n\t\tval.bypassed++\n\t}\n\tcounters[category][typ] = val\n}", "func mapCount(itr Iterator, m *mapper) {\n\tn := 0\n\tfor k, _ := itr.Next(); k != 0; k, _ = itr.Next() {\n\t\tn++\n\t}\n\tm.emit(itr.Time(), float64(n))\n}", "func (self *IoStatsBuilder) Update(current map[int]*IoAmount, now time.Time) {\n\tseconds := now.Sub(self.lastMeasurement).Seconds()\n\tvar total IoAmount\n\tfor pid, amt := range current {\n\t\ttotal.Increment(amt)\n\t\tdelete(self.lastPids, pid)\n\t}\n\tfor _, amt := range self.lastPids {\n\t\tself.deadUsage.Increment(amt)\n\t}\n\ttotal.Increment(&self.deadUsage)\n\tself.lastPids = current\n\tself.lastMeasurement = now\n\tdiff := self.Total.update(&total)\n\tif seconds > 0 {\n\t\trate := diff.rate(seconds)\n\t\tself.RateMax.TakeMax(rate)\n\t\trate.weightSquared(seconds)\n\t\tself.weightedSumSquared.Increment(rate)\n\t\tif t := now.Sub(self.start).Seconds(); t > 0 {\n\t\t\tself.RateDev = self.weightedSumSquared.computeStdDev(\n\t\t\t\t&self.Total, t)\n\t\t}\n\t}\n}", "func BenchmarkAddMap(b *testing.B) {\n\taddInit(b)\n\tfor n := 0; n < b.N; n++ {\n\t\tb.StopTimer()\n\t\tdata := addIntMapData.Dup()\n\t\tb.StartTimer()\n\t\tdata.Add(addValue)\n\t}\n}", "func (q *QueryFeedback) Update(startKey kv.Key, counts []int64) {\n\t// Older version do not have the counts info.\n\tif len(counts) == 0 {\n\t\tq.Invalidate()\n\t\treturn\n\t}\n\tsum := int64(0)\n\tfor _, count := range counts {\n\t\tsum += count\n\t}\n\t// remove metrics\n\tq.actual += sum\n\tif !q.Valid || q.Hist == nil {\n\t\treturn\n\t}\n\n\tif q.Tp == IndexType {\n\t\tstartKey = tablecodec.CutIndexPrefix(startKey)\n\t} else {\n\t\tstartKey = tablecodec.CutRowKeyPrefix(startKey)\n\t}\n\t// Find the range that startKey falls in.\n\tidx := sort.Search(len(q.Feedback), func(i int) bool {\n\t\treturn bytes.Compare(q.Feedback[i].Lower.GetBytes(), startKey) > 0\n\t})\n\tidx--\n\tif idx < 0 {\n\t\treturn\n\t}\n\t// If the desc is true, the counts is reversed, so here we need to reverse it back.\n\tif q.desc {\n\t\tfor i := 0; i < len(counts)/2; i++ {\n\t\t\tj := len(counts) - i - 1\n\t\t\tcounts[i], counts[j] = counts[j], counts[i]\n\t\t}\n\t}\n\t// Update the feedback count info.\n\tfor i, count := range counts {\n\t\tif i+idx >= len(q.Feedback) {\n\t\t\tq.Invalidate()\n\t\t\tbreak\n\t\t}\n\t\tq.Feedback[i+idx].Count += count\n\t}\n\treturn\n}", "func (cg *CandlesGroup) mapUpdate(symbol string, data []interface{}) []schemas.Candle {\n\ts, _, _ := parseSymbol(symbol)\n\treturn []schemas.Candle{schemas.Candle{\n\t\tSymbol: s,\n\t\tOpen: data[1].(float64),\n\t\tClose: data[2].(float64),\n\t\tHigh: data[3].(float64),\n\t\tLow: data[4].(float64),\n\t\tVolume: data[5].(float64),\n\t\tTimestamp: int64(data[0].(float64)),\n\t\tDiscretization: 60,\n\t},\n\t}\n}", "func (bg *bufferedGroup) updateLocationIndex() {\n bg.locationIndex = make(map[string]int)\n for i, bi := range bg.images {\n if _, found := bg.locationIndex[bi.nearestCityKey]; found == false {\n locationTimekey := bi.LocationTimekey()\n bg.locationIndex[locationTimekey] = i\n }\n }\n}", "func (a *StandardEWMA) Update(n int64) {\n\ta.uncounted += n\n}", "func (m *MLP) Update(s int64, f float64) {\n\tp := Rand(s, len(m.Weights))\n\tfor i := range p {\n\t\tm.Weights[i] += rate * p[i] * f / scale\n\t}\n}", "func (g *Generator) updateCounter() {\n\t// Loop over each position and increase it by 1. If position already max\n\t// value, then drop it to 0 and increase next one by 1.\n\tfor i, _ := range g.position {\n\t\tg.position[i]++\n\t\tif g.position[i] < int(len(g.elements)) {\n\t\t\tbreak\n\t\t}\n\t\tg.position[i] = 0\n\t}\n\tg.current++\n}", "func (counter *Counter) Update(request map[string]string) {\n\th := sha256.New()\n\tfor _, f := range []string{\"remote\", \"path\", \"agent\", \"http_x_forwarded_for\"} {\n\t\th.Write([]byte(request[f]))\n\t}\n\tvar k [sha256.Size]byte\n\tcopy(k[:], h.Sum(nil))\n\n\tcounter.mutex.Lock()\n\tdefer counter.mutex.Unlock()\n\n\tcounter.entries[k] = time.Now()\n\tcounter.callback(float64(len(counter.entries)))\n}", "func MapCount(itr Iterator) interface{} {\n\tn := float64(0)\n\tfor k, _ := itr.Next(); k != -1; k, _ = itr.Next() {\n\t\tn++\n\t}\n\tif n > 0 {\n\t\treturn n\n\t}\n\treturn nil\n}", "func (e *exemplarSampler) updateAggregations(val float64) {\n\te.count++\n\tdelta := val - e.mean\n\te.mean += delta / float64(e.count)\n\tdelta2 := val - e.mean\n\te.m2 += delta * delta2\n}", "func fakeKongAdminUpdateCount(newcounts ...int) int {\n\tcountLock.Lock()\n\tdefer countLock.Unlock()\n\tif len(newcounts) < 1 {\n\t\treturn updateCount\n\t}\n\tif len(newcounts) == 1 && newcounts[0] == 0 {\n\t\tupdateCount = 0\n\t\treturn 0\n\t}\n\tfor _, count := range newcounts {\n\t\tupdateCount = updateCount + count\n\t}\n\treturn updateCount\n}", "func TestMapCount(t *testing.T) {\n\tm := map[Key]interface{}{}\n\ttestMapCountN(testN, m)\n}", "func (s *storage) UpdateMaps(maps pb.SyncMapping) {\n\tsrcMaps := s.GetMaps()\n\tmappings := make(pb.SyncMapping, 0, len(maps))\n\tfor _, val := range maps {\n\t\tdata, err := proto.Marshal(val)\n\t\tif err != nil {\n\t\t\tlog.Errorf(err, \"Proto marshal failed: %s\", err)\n\t\t\tcontinue\n\t\t}\n\n\t\tputOp := putMappingOp(val.ClusterName, val.OrgInstanceID, data)\n\t\t_, err = s.engine.Do(context.Background(), putOp)\n\t\tif err != nil {\n\t\t\tlog.Errorf(err, \"Save mapping to etcd failed: %s\", err)\n\t\t\tcontinue\n\t\t}\n\t\tmappings = append(mappings, val)\n\t}\n\ts.cleanExpired(srcMaps, mappings)\n}", "func (c *PrometheusCollector) Update(r metricCollector.MetricResult) {\n\tc.RWMutex.Lock()\n\tdefer c.RWMutex.Unlock()\n\n\t// check circuit open\n\tif r.Successes > 0 {\n\t\tgauge := c.gauges[metricCircuitOpen]\n\t\tgauge.Set(0)\n\n\t\tcounter := c.counters[metricSuccesses]\n\t\tcounter.Add(r.Successes)\n\t}\n\tif r.ShortCircuits > 0 {\n\t\tgauge := c.gauges[metricCircuitOpen]\n\t\tgauge.Set(1)\n\n\t\tcounter := c.counters[metricShortCircuits]\n\t\tcounter.Add(r.ShortCircuits)\n\t}\n\t// update metric\n\tif r.Attempts > 0 {\n\t\tcounter := c.counters[metricAttempts]\n\t\tcounter.Add(r.Attempts)\n\t}\n\tif r.Errors > 0 {\n\t\tcounter := c.counters[metricErrors]\n\t\tcounter.Add(r.Errors)\n\t}\n\tif r.Failures > 0 {\n\t\tcounter := c.counters[metricFailures]\n\t\tcounter.Add(r.Failures)\n\t}\n\tif r.Rejects > 0 {\n\t\tcounter := c.counters[metricRejects]\n\t\tcounter.Add(r.Rejects)\n\t}\n\tif r.Timeouts > 0 {\n\t\tcounter := c.counters[metricTimeouts]\n\t\tcounter.Add(r.Timeouts)\n\t}\n\tif r.FallbackSuccesses > 0 {\n\t\tcounter := c.counters[metricFallbackSuccesses]\n\t\tcounter.Add(r.FallbackSuccesses)\n\t}\n\tif r.FallbackFailures > 0 {\n\t\tcounter := c.counters[metricFallbackFailures]\n\t\tcounter.Add(r.FallbackFailures)\n\t}\n\n\tgauge := c.gauges[metricTotalDuration]\n\tgauge.Set(r.TotalDuration.Seconds())\n\n\tgauge = c.gauges[metricRunDuration]\n\tgauge.Set(r.RunDuration.Seconds())\n\n\tgauge = c.gauges[metricConcurrencyInUse]\n\tgauge.Set(r.ConcurrencyInUse)\n}", "func (stat *NetworkStat) Update() {\n\tfor i, node := range stat.bn.nodeIndex {\n\t\tassignment := node.GetAssignment()\n\t\tif assignment == \"F\" {\n\t\t\tcontinue\n\t\t}\n\t\tstat.count[i] += 1\n\t}\n\tstat.total += 1\n}", "func (s *SliceOfUint64) Map(mapFunc func(uint64) uint64) *SliceOfUint64 {\n\tfor index, value := range s.items {\n\t\ts.items[index] = mapFunc(value)\n\t}\n\treturn s\n}", "func (m ConcurrentRoomInfoMap) Count() int {\n\tcount := 0\n\tfor i := 0; i < shardCount; i++ {\n\t\tshard := m[i]\n\t\tshard.RLock()\n\t\tcount += len(shard.items)\n\t\tshard.RUnlock()\n\t}\n\treturn count\n}", "func updateMetrics(interval time.Duration, diskRoot string) {\n\tlogger := logrus.WithField(\"sync-loop\", \"updateMetrics\")\n\tticker := time.NewTicker(interval)\n\tfor ; true; <-ticker.C {\n\t\tlogger.Info(\"tick\")\n\t\t_, bytesFree, bytesUsed, _, _, _, err := diskutil.GetDiskUsage(diskRoot)\n\t\tif err != nil {\n\t\t\tlogger.WithError(err).Error(\"Failed to get disk metrics\")\n\t\t} else {\n\t\t\tpromMetrics.DiskFree.Set(float64(bytesFree) / 1e9)\n\t\t\tpromMetrics.DiskUsed.Set(float64(bytesUsed) / 1e9)\n\t\t\tpromMetrics.DiskTotal.Set(float64(bytesFree+bytesUsed) / 1e9)\n\t\t}\n\t}\n}", "func (b *Batcher) Update(matched, total int) {\n\tb.ratio = float64(matched) / float64(total)\n}", "func (c *Cache) recordUpdate(p *partition, bytesAdded, bytesGuessed, entriesAdded int32) {\n\t// This method is always called while p.mu is held.\n\t// The below code takes care to ensure that all bytes in c due to p are\n\t// updated appropriately.\n\n\t// NB: The loop and atomics are used because p.size can be modified\n\t// concurrently to calls to recordUpdate. In all cases where p.size is updated\n\t// outside of this function occur while c.mu is held inside of c.Add. These\n\t// occur when either:\n\t//\n\t// 1) a new write adds its guessed write size to p\n\t// 2) p is evicted to make room for a write\n\t//\n\t// Thus p.size is either increasing or becomes evicted while we attempt to\n\t// record the update to p. Once p is evicted it stays evicted forever.\n\t// These facts combine to ensure that p.size never becomes negative from the\n\t// below call to add.\n\n\tdelta := bytesAdded - bytesGuessed\n\tfor {\n\t\tcurSize := p.loadSize()\n\t\tif curSize == evicted {\n\t\t\treturn\n\t\t}\n\t\tnewSize := curSize.add(delta, entriesAdded)\n\t\tif updated := p.setSize(curSize, newSize); updated {\n\t\t\tc.updateGauges(c.addBytes(delta), c.addEntries(entriesAdded))\n\t\t\treturn\n\t\t}\n\t}\n}", "func (s *StashList) updateStashListMap(ctx context.Context, vw types.ValueWriter) (hash.Hash, error) {\n\t// update stash map data and reset the stash map's hash\n\tdata := stashlist_flatbuffer(s.am)\n\tr, err := vw.WriteValue(ctx, types.SerialMessage(data))\n\tif err != nil {\n\t\treturn hash.Hash{}, err\n\t}\n\ts.addr = r.TargetHash()\n\n\treturn s.addr, nil\n}", "func UpdateCount(w http.ResponseWriter, r *http.Request) {\n\n\tvar message Message\n\t_ = json.NewDecoder(r.Body).Decode(&message)\n\tid := message.ID\n\t\n\twrite(id, read(id) + message.Count) \n\n\t//broadcast in background routine\n\tgo func(msg Message) {\n broadcast(msg)\n\t}(message)\n\t\n\tfmt.Println(\"Broadcast done\")\n\tjson.NewEncoder(w).Encode(Message{ID : id, Count : read(id)})\n}", "func (esc *BgMetadataElasticSearchConnector) updateInternalMetrics(res *esapi.Response) {\n\tdefer func() {\n\t\tif err := recover(); err != nil {\n\t\t\tesc.logger.Warn(\"malformed bulk response\", zap.Error(err.(error)))\n\t\t}\n\t}()\n\tvar mapResp map[string]interface{}\n\tjson.NewDecoder(res.Body).Decode(&mapResp)\n\tesc.WriteDurationMs.Observe(mapResp[\"took\"].(float64))\n\tfor _, item := range mapResp[\"items\"].([]interface{}) {\n\t\tmapCreate := item.(map[string]interface{})[\"create\"].(map[string]interface{})\n\t\t// protected by esc.Mux currentIndex may not change while looping\n\t\tif int(mapCreate[\"status\"].(float64)) == http.StatusCreated {\n\t\t\tif strings.HasPrefix(mapCreate[\"_index\"].(string), esc.currentIndex) {\n\t\t\t\tesc.UpdatedDocuments.WithLabelValues(\"created\", \"metric\").Inc()\n\t\t\t} else {\n\t\t\t\tesc.UpdatedDocuments.WithLabelValues(\"created\", \"directory\").Inc()\n\t\t\t}\n\t\t}\n\t}\n}", "func mergeMaps(m1 *map[string]int, m2 *map[string]int) {\n\n\tif *m1 == nil && m2 != nil {\n\t\t*m1 = *m2\n\t}\n\n\tif m1 != nil && *m2 == nil {\n\t\treturn\n\t}\n\n\tfor k, v := range *m2 {\n\t\texistingWordCount, ok := (*m1)[k]\n\t\tif !ok {\n\t\t\t(*m1)[k] = v\n\t\t} else {\n\t\t\t(*m1)[k] = existingWordCount + v\n\t\t}\n\t}\n}", "func BenchmarkReadWriteMap(b *testing.B) {\n\tvalue := newValue(123)\n\tfor i := 0; i <= 10; i++ {\n\t\treadFrac := float32(i) / 10.0\n\t\tb.Run(fmt.Sprintf(\"frac_%d\", i), func(b *testing.B) {\n\t\t\tm := make(map[string][]byte)\n\t\t\tvar mutex sync.RWMutex\n\t\t\tb.ResetTimer()\n\t\t\tvar count int\n\t\t\tb.RunParallel(func(pb *testing.PB) {\n\t\t\t\trng := rand.New(rand.NewSource(time.Now().UnixNano()))\n\t\t\t\tfor pb.Next() {\n\t\t\t\t\tif rng.Float32() < readFrac {\n\t\t\t\t\t\tmutex.RLock()\n\t\t\t\t\t\t_, ok := m[string(randomKey(rng))]\n\t\t\t\t\t\tmutex.RUnlock()\n\t\t\t\t\t\tif ok {\n\t\t\t\t\t\t\tcount++\n\t\t\t\t\t\t}\n\t\t\t\t\t} else {\n\t\t\t\t\t\tmutex.Lock()\n\t\t\t\t\t\tm[string(randomKey(rng))] = value\n\t\t\t\t\t\tmutex.Unlock()\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t})\n\t\t})\n\t}\n}", "func (s *Stats) Update(code int) {\n\ts.Lock()\n\ts.StatusCodes[code]++\n\ts.Unlock()\n}", "func updateMetrics(metrics map[string]*metricData) {\n\n\tfor _, metricClass := range mqmetric.Metrics.Classes {\n\t\tfor _, metricType := range metricClass.Types {\n\t\t\tif !strings.Contains(metricType.ObjectTopic, \"%s\") {\n\t\t\t\tfor _, metricElement := range metricType.Elements {\n\n\t\t\t\t\t// Unexpected metric elements (with no defined mapping) are handled in 'initialiseMetrics'\n\t\t\t\t\t// - if any exist, they are logged as errors and skipped (they are not added to the metrics map)\n\t\t\t\t\t// Therefore we can ignore handling any unexpected metric elements found here\n\t\t\t\t\t// - this avoids us logging excessive errors, as this function is called frequently\n\t\t\t\t\tmetric, ok := metrics[makeKey(metricElement)]\n\t\t\t\t\tif ok {\n\t\t\t\t\t\t// Clear existing metric values\n\t\t\t\t\t\tmetric.values = make(map[string]float64)\n\n\t\t\t\t\t\t// Update metric with cached values of publication data\n\t\t\t\t\t\tfor label, value := range metricElement.Values {\n\t\t\t\t\t\t\tnormalisedValue := mqmetric.Normalise(metricElement, label, value)\n\t\t\t\t\t\t\tmetric.values[label] = normalisedValue\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\n\t\t\t\t\t// Reset cached values of publication data for this metric\n\t\t\t\t\tmetricElement.Values = make(map[string]int64)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n}", "func UpdateMetrics(result *Results) {\n\n\t// Publish system variables\n\tupTimeGauge.Set(float64(result.SysMonitorInfo.Uptime))\n\tcpuUsageGauge.Set(float64(result.SysMonitorInfo.CpuUsagePercent))\n\n\t// Memory\n\tmemUsagePercentGauge.Set(result.SysMonitorInfo.MemUsagePercent)\n\tmemTotalGauge.Set(float64(result.SysMonitorInfo.MemTotal))\n\tmemAvailableGauge.Set(float64(result.SysMonitorInfo.MemAvailable))\n\n\t// Bandwidth\n\tbandwidthUsageTotalGauge.Set(float64(result.SysMonitorInfo.BandwidthUsageTotal))\n\tbandwidthUsageSentGauge.Set(float64(result.SysMonitorInfo.BandwidthUsageSent))\n\tbandwidthUsageRecvGauge.Set(float64(result.SysMonitorInfo.BandwidthUsageRecv))\n\n\tfor _, driveUsage := range result.SysMonitorInfo.DriveUsage {\n\t\t// \"drive_path\", \"available\", \"growth_rate\", \"full_in\", \"physical_drive\"\n\n\t\tdays := strconv.FormatFloat(driveUsage.DaysTillFull, 'f', 3, 64)\n\n\t\tif math.IsInf(driveUsage.DaysTillFull, 0) {\n\t\t\tdays = \"10 years\"\n\t\t}\n\n\t\tdriveSpace.WithLabelValues(driveUsage.Path,\n\t\t\tstrconv.FormatFloat(driveUsage.PercentUsed, 'f', 3, 64),\n\t\t\tstrconv.FormatUint(driveUsage.GrowthPerDayBytes, 10),\n\t\t\tdays,\n\t\t\tdriveUsage.VolumeName).Set(driveUsage.PercentUsed)\n\t}\n\n\t// Publish endpoints being monitored\n\tfor _, uptimeResponse := range result.UptimeList {\n\n\t\tif uptimeResponse.ResponseCode == 200 {\n\t\t\tendpointAvailable.WithLabelValues(uptimeResponse.Endpoint).Set(1)\n\t\t} else {\n\t\t\tendpointAvailable.WithLabelValues(uptimeResponse.Endpoint).Set(0)\n\t\t}\n\n\t\tendpointDuration.WithLabelValues(uptimeResponse.Endpoint).Set(uptimeResponse.ResponseTime.Seconds())\n\t}\n\n\tfor _, backupInfo := range result.BackupInfoList {\n\n\t\t/*\n\t\t\tif backupInfo.WasBackedUp {\n\t\t\t\tbackupsDone.WithLabelValues(backupInfo.Folder).Set(1)\n\t\t\t} else {\n\t\t\t\tbackupsDone.WithLabelValues(backupInfo.Folder).Set(0)\n\t\t\t}\n\t\t*/\n\n\t\t// {\"backup_directory\", \"backup_in_last_24_hours\", \"last_backup_size\", \"last_backup_date\", \"last_backup_time\"})\n\n\t\t// backupsSize.WithLabelValues(backupInfo.Folder).Set(float64(backupInfo.BackupFileSize))\n\n\t\tbackupInfoGauge.WithLabelValues(backupInfo.Folder,\n\t\t\tbtoa(backupInfo.WasBackedUp),\n\t\t\titoa(backupInfo.LastBackupSize),\n\t\t\tttoa(backupInfo.LastBackupTime),\n\t\t\tbackupInfo.LastBackupFile).Set(btof(backupInfo.WasBackedUp))\n\t}\n\n\t// TODO: This loop is not needed, you can build the summary on the first loop\n\tvar too_many_lines = 500\n\tfor _, logLine := range result.LoglineList {\n\n\t\tsummary, ok := result.LogSummary[logLine.LogPath]\n\n\t\tif ok == false {\n\t\t\tsummary = LogSummary{}\n\t\t\tsummary.StatusCount = make(map[string]int64)\n\t\t\tsummary.SeverityLevelCount = make(map[string]int64)\n\t\t}\n\n\t\tsummary.StatusCount[logLine.StatusCode] = summary.StatusCount[logLine.StatusCode] + 1\n\n\t\tif len(logLine.Severity) > 0 {\n\t\t\tsummary.SeverityLevelCount[logLine.Severity] = summary.SeverityLevelCount[logLine.Severity] + 1\n\t\t}\n\n\t\tresult.LogSummary[logLine.LogPath] = summary\n\n\t\tif too_many_lines <= 0 {\n\t\t\t// Pending a better solution, let's not allow the processing\n\t\t\t// of too many lines, to not kill the server\n\t\t\tlLog.Print(\"Too many lines for a single tick to process\")\n\t\t\tbreak\n\t\t}\n\n\t}\n\n\t// Set the values for the logs. We use two labels (logpath, code)\n\tfor logFilePath, logSummary := range result.LogSummary {\n\n\t\tfor s, value := range logSummary.StatusCount {\n\t\t\tstatusCodes.WithLabelValues(logFilePath, s).Set(float64(value))\n\t\t}\n\n\t\tfor s, value := range logSummary.SeverityLevelCount {\n\t\t\tseverity.WithLabelValues(logFilePath, s).Set(float64(value))\n\t\t}\n\n\t}\n}", "func (t *Telemetry) UpdateTelemetryLists() {\n\tt.init(false)\n}", "func (m NMap) Count() int {\n\tcount := 0\n\tfor _, inMap := range m {\n\t\tinMap.RLock()\n\t\tcount += len(inMap.objs)\n\t\tinMap.RUnlock()\n\t}\n\treturn count\n}", "func (tb *Table) UpdateMetrics(m *TableMetrics) {\n\tm.ActiveInmemoryMerges += atomic.LoadUint64(&tb.activeInmemoryMerges)\n\tm.ActiveFileMerges += atomic.LoadUint64(&tb.activeFileMerges)\n\n\tm.InmemoryMergesCount += atomic.LoadUint64(&tb.inmemoryMergesCount)\n\tm.FileMergesCount += atomic.LoadUint64(&tb.fileMergesCount)\n\n\tm.InmemoryItemsMerged += atomic.LoadUint64(&tb.inmemoryItemsMerged)\n\tm.FileItemsMerged += atomic.LoadUint64(&tb.fileItemsMerged)\n\n\tm.InmemoryAssistedMerges += atomic.LoadUint64(&tb.inmemoryAssistedMerges)\n\tm.FileAssistedMerges += atomic.LoadUint64(&tb.fileAssistedMerges)\n\n\tm.ItemsAdded += atomic.LoadUint64(&tb.itemsAdded)\n\tm.ItemsAddedSizeBytes += atomic.LoadUint64(&tb.itemsAddedSizeBytes)\n\n\tm.PendingItems += uint64(tb.rawItems.Len())\n\n\ttb.partsLock.Lock()\n\n\tm.InmemoryPartsCount += uint64(len(tb.inmemoryParts))\n\tfor _, pw := range tb.inmemoryParts {\n\t\tp := pw.p\n\t\tm.InmemoryBlocksCount += p.ph.blocksCount\n\t\tm.InmemoryItemsCount += p.ph.itemsCount\n\t\tm.InmemorySizeBytes += p.size\n\t\tm.PartsRefCount += uint64(atomic.LoadUint32(&pw.refCount))\n\t}\n\n\tm.FilePartsCount += uint64(len(tb.fileParts))\n\tfor _, pw := range tb.fileParts {\n\t\tp := pw.p\n\t\tm.FileBlocksCount += p.ph.blocksCount\n\t\tm.FileItemsCount += p.ph.itemsCount\n\t\tm.FileSizeBytes += p.size\n\t\tm.PartsRefCount += uint64(atomic.LoadUint32(&pw.refCount))\n\t}\n\ttb.partsLock.Unlock()\n\n\tm.DataBlocksCacheSize = uint64(ibCache.Len())\n\tm.DataBlocksCacheSizeBytes = uint64(ibCache.SizeBytes())\n\tm.DataBlocksCacheSizeMaxBytes = uint64(ibCache.SizeMaxBytes())\n\tm.DataBlocksCacheRequests = ibCache.Requests()\n\tm.DataBlocksCacheMisses = ibCache.Misses()\n\n\tm.IndexBlocksCacheSize = uint64(idxbCache.Len())\n\tm.IndexBlocksCacheSizeBytes = uint64(idxbCache.SizeBytes())\n\tm.IndexBlocksCacheSizeMaxBytes = uint64(idxbCache.SizeMaxBytes())\n\tm.IndexBlocksCacheRequests = idxbCache.Requests()\n\tm.IndexBlocksCacheMisses = idxbCache.Misses()\n}", "func (h *batchHistogram) update(his *metrics.Float64Histogram, sum float64) {\n\tcounts, buckets := his.Counts, his.Buckets\n\n\th.mu.Lock()\n\tdefer h.mu.Unlock()\n\n\t// Clear buckets.\n\tfor i := range h.counts {\n\t\th.counts[i] = 0\n\t}\n\t// Copy and reduce buckets.\n\tvar j int\n\tfor i, count := range counts {\n\t\th.counts[j] += count\n\t\tif buckets[i+1] == h.buckets[j+1] {\n\t\t\tj++\n\t\t}\n\t}\n\tif h.hasSum {\n\t\th.sum = sum\n\t}\n}", "func UpdateFromUserDefinedMappings(m map[string]string) (err error) {\n\tlimitsOnce.Do(populateStaticENILimits)\n\n\tlimits.Lock()\n\tdefer limits.Unlock()\n\n\tfor instanceType, limitString := range m {\n\t\tlimit, err := parseLimitString(limitString)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t// Add or overwrite limits\n\t\tlimits.m[instanceType] = limit\n\t}\n\treturn nil\n}", "func (c *Aggregator) Update(_ context.Context, number number.Number, desc *sdkapi.Descriptor) error {\n\tnow := time.Now()\n\tc.lock.Lock()\n\tdefer c.lock.Unlock()\n\tc.samples = append(c.samples, Point{\n\t\tNumber: number,\n\t\tTime: now,\n\t})\n\n\treturn nil\n}", "func (s *SliceOfInt64) Map(mapFunc func(int64) int64) *SliceOfInt64 {\n\tfor index, value := range s.items {\n\t\ts.items[index] = mapFunc(value)\n\t}\n\treturn s\n}", "func (m *Cmap) Count() int {\n\treturn int(atomic.LoadInt64(&m.count))\n}", "func (rb *reportBucket) UpdateKindMap() {\n\toSlice := objectSlice(rb.objects)\n\tsort.Sort(oSlice)\n\tkindMap := make(map[string][]*reportObject)\n\tfor _, object := range rb.objects {\n\t\tif object.kind == \"\" {\n\t\t\tcontinue\n\t\t}\n\t\tkindMap[object.kind] = append(kindMap[object.kind], object)\n\t}\n\trb.kindMap = kindMap\n}", "func (r *DeviceRates) Update(maj, min, val int64) {\n\tfor index, devRate := range *r {\n\t\tif devRate.Major == maj && devRate.Minor == min {\n\t\t\t(*r)[index].Rate = val\n\t\t\treturn\n\t\t}\n\t}\n\tr.Append(maj, min, val)\n}", "func (mc *MockCluster) UpdateRegionCount(storeID uint64, regionCount int) {\n\tstore := mc.GetStore(storeID)\n\tstore.RegionCount = regionCount\n\tstore.RegionSize = int64(regionCount) * 10\n\tmc.PutStore(store)\n}", "func (c *Counter) Update(timestamp time.Time, value int64, annotation []byte) {\n\tif c.lastAt.IsZero() || timestamp.After(c.lastAt) {\n\t\t// NB(r): Only set the last value if this value arrives\n\t\t// after the wall clock timestamp of previous values, not\n\t\t// the arrival time (i.e. order received).\n\t\tc.lastAt = timestamp\n\t} else {\n\t\tc.Options.Metrics.Counter.IncValuesOutOfOrder()\n\t}\n\n\tc.sum += value\n\n\tc.count++\n\tif c.max < value {\n\t\tc.max = value\n\t}\n\tif c.min > value {\n\t\tc.min = value\n\t}\n\n\tif c.HasExpensiveAggregations {\n\t\tc.sumSq += value * value\n\t}\n\n\tc.annotation = MaybeReplaceAnnotation(c.annotation, annotation)\n}", "func (c Counter) Update(token Token) {\n\tif count, ok := c.Data[token.Datum]; ok {\n\t\tc.Data[token.Datum] = count + 1\n\t} else {\n\t\tc.Data[token.Datum] = 1\n\t}\n}", "func (s *UniformSample) Update(v int64) {\n\ts.mutex.Lock()\n\tdefer s.mutex.Unlock()\n\ts.count++\n\tif len(s.values) < s.reservoirSize {\n\t\ts.values = append(s.values, v)\n\t} else {\n\t\t// Use circle buffer to eliminate the oldest value\n\t\tidx := s.count % int64(s.reservoirSize)\n\t\ts.values[idx] = v\n\t}\n}", "func (c *Aggregator) Update(_ context.Context, number number.Number, desc *metric.Descriptor) error {\n\tkind := desc.NumberKind()\n\tasFloat := number.CoerceToFloat64(kind)\n\n\tbucketID := len(c.boundaries)\n\tfor i, boundary := range c.boundaries {\n\t\tif asFloat < boundary {\n\t\t\tbucketID = i\n\t\t\tbreak\n\t\t}\n\t}\n\t// Note: Binary-search was compared using the benchmarks. The following\n\t// code is equivalent to the linear search above:\n\t//\n\t// bucketID := sort.Search(len(c.boundaries), func(i int) bool {\n\t// return asFloat < c.boundaries[i]\n\t// })\n\t//\n\t// The binary search wins for very large boundary sets, but\n\t// the linear search performs better up through arrays between\n\t// 256 and 512 elements, which is a relatively large histogram, so we\n\t// continue to prefer linear search.\n\n\tc.lock.Lock()\n\tdefer c.lock.Unlock()\n\n\tc.state.count++\n\tc.state.sum.AddNumber(kind, number)\n\tc.state.bucketCounts[bucketID]++\n\n\treturn nil\n}", "func processMapsets() {\n\tfor i := 0; i < 2; i++ { // 1000 mapsets.\n\t\tmapsets, err := getMapsets(i)\n\t\tif err != nil {\n\t\t\tlogMsg(\"Getting data from osusearch.com failed: %s\", err)\n\t\t}\n\n\t\tmapsets = dedup(mapsets)\n\t\tlog.Printf(\"Retrieved %d mapset(s)\\n\", len(mapsets))\n\n\t\tfor _, mapset := range mapsets {\n\t\t\t// The osu! API separates date from time with a space, and osusearch does not.\n\t\t\tmapset.Updated = strings.Replace(mapset.Updated, \"T\", \" \", 1)\n\t\t\texisting := new(gosubscribe.Mapset)\n\t\t\tgosubscribe.DB.Where(\"id = ?\", mapset.ID).First(existing)\n\n\t\t\tif existing.ID == 0 {\n\t\t\t\tprocessMapset(mapset, \"new\")\n\t\t\t} else if existing.Status != mapset.Status {\n\t\t\t\tprocessMapset(mapset, \"status\")\n\t\t\t} else if existing.Updated != mapset.Updated {\n\t\t\t\tprocessMapset(mapset, \"update\")\n\t\t\t}\n\t\t}\n\t}\n}", "func (cs *consumerSummary) update(acknowledgments []*magmasc.Acknowledgment) {\n\tcs.totalSessions = len(acknowledgments)\n\n\ttotalRewards := big.NewFloat(0)\n\tfor _, ackn := range acknowledgments {\n\t\tcs.totalQoSDB += float64(ackn.Provider.Terms.QoS.DownloadMbps)\n\t\tcs.totalQoSUB += float64(ackn.Provider.Terms.QoS.UploadMbps)\n\n\t\tdOff, ok := cs.DataOffloaded[ackn.Provider.ExtID]\n\t\tif !ok {\n\t\t\tdOff = new(dataOffloadedStat)\n\t\t}\n\t\tdOff.DataUploaded += ackn.Billing.DataUsage.UploadBytes\n\t\tdOff.DataDownloaded += ackn.Billing.DataUsage.DownloadBytes\n\t\tcs.DataOffloaded[ackn.Provider.ExtID] = dOff\n\n\t\tif ackn.Billing.CompletedAt > 0 {\n\t\t\ttotalRewards.Add(\n\t\t\t\ttotalRewards,\n\t\t\t\tbig.NewFloat(float64(ackn.Billing.Amount)),\n\t\t\t)\n\t\t}\n\t}\n\n\t// convert to ZCN\n\ttotalRewards.Quo(\n\t\ttotalRewards,\n\t\tbig.NewFloat(billion),\n\t)\n\tcs.TotalRewardsGiven = totalRewards.String()\n\n\t// computing average fields\n\tvar (\n\t\taverageQoSUB, averageQoSDB float32\n\t\taverageRewards = big.NewFloat(0)\n\t)\n\tif cs.totalSessions != 0 {\n\t\taverageQoSUB = float32(cs.totalQoSUB / float64(cs.totalSessions))\n\t\taverageQoSDB = float32(cs.totalQoSDB / float64(cs.totalSessions))\n\n\t\taverageRewards = totalRewards.Quo(\n\t\t\ttotalRewards,\n\t\t\tbig.NewFloat(float64(cs.totalSessions)),\n\t\t)\n\t}\n\n\tcs.AverageQoS = magma.QoS{\n\t\tUploadMbps: averageQoSUB,\n\t\tDownloadMbps: averageQoSDB,\n\t}\n\n\tcs.AverageRewards = averageRewards.String()\n}", "func init() {\n MapCrawlerSpeed = make(map[string]int)\n for i := 0; i < 86400; i++ {\n listi := list.New()\n listi.PushBack(i)\n AllTimeStamp[i] = listi\n }\n \n ReadConfig()\n}", "func (mm *MetricManager) addToMaps(metricPath string) *rollup {\n\tvar currentRollup *rollup\n\n\texpr := mm.getExpression(metricPath)\n\tcurrentRollup = new(rollup)\n\tcurrentRollup.expr = expr\n\tcurrentRollup.count = make([]uint64, len(mm.rollup[expr].Windows))\n\tcurrentRollup.value = make([]float64, len(mm.rollup[expr].Windows))\n\tmm.byPath[metricPath] = currentRollup\n\tmm.byExpr[expr].path[metricPath] = currentRollup\n\n\treturn currentRollup\n}", "func (s *Segment) initializeDict(results []*index.AnalysisResult) {\n\tvar numPostingsLists int\n\n\tnumTermsPerPostingsList := make([]int, 0, 64) // Keyed by postings list id.\n\tnumLocsPerPostingsList := make([]int, 0, 64) // Keyed by postings list id.\n\n\tvar numTokenFrequencies int\n\tvar totLocs int\n\n\t// initial scan for all fieldID's to sort them\n\tfor _, result := range results {\n\t\tfor _, field := range result.Document.CompositeFields {\n\t\t\ts.getOrDefineField(field.Name())\n\t\t}\n\t\tfor _, field := range result.Document.Fields {\n\t\t\ts.getOrDefineField(field.Name())\n\t\t}\n\t}\n\tsort.Strings(s.FieldsInv[1:]) // keep _id as first field\n\ts.FieldsMap = make(map[string]uint16, len(s.FieldsInv))\n\tfor fieldID, fieldName := range s.FieldsInv {\n\t\ts.FieldsMap[fieldName] = uint16(fieldID + 1)\n\t}\n\n\tprocessField := func(fieldID uint16, tfs analysis.TokenFrequencies) {\n\t\tdict := s.Dicts[fieldID]\n\t\tdictKeys := s.DictKeys[fieldID]\n\t\tfor term, tf := range tfs {\n\t\t\tpidPlus1, exists := dict[term]\n\t\t\tif !exists {\n\t\t\t\tnumPostingsLists++\n\t\t\t\tpidPlus1 = uint64(numPostingsLists)\n\t\t\t\tdict[term] = pidPlus1\n\t\t\t\tdictKeys = append(dictKeys, term)\n\t\t\t\tnumTermsPerPostingsList = append(numTermsPerPostingsList, 0)\n\t\t\t\tnumLocsPerPostingsList = append(numLocsPerPostingsList, 0)\n\t\t\t}\n\t\t\tpid := pidPlus1 - 1\n\t\t\tnumTermsPerPostingsList[pid] += 1\n\t\t\tnumLocsPerPostingsList[pid] += len(tf.Locations)\n\t\t\ttotLocs += len(tf.Locations)\n\t\t}\n\t\tnumTokenFrequencies += len(tfs)\n\t\ts.DictKeys[fieldID] = dictKeys\n\t}\n\n\tfor _, result := range results {\n\t\t// walk each composite field\n\t\tfor _, field := range result.Document.CompositeFields {\n\t\t\tfieldID := uint16(s.getOrDefineField(field.Name()))\n\t\t\t_, tf := field.Analyze()\n\t\t\tprocessField(fieldID, tf)\n\t\t}\n\n\t\t// walk each field\n\t\tfor i, field := range result.Document.Fields {\n\t\t\tfieldID := uint16(s.getOrDefineField(field.Name()))\n\t\t\ttf := result.Analyzed[i]\n\t\t\tprocessField(fieldID, tf)\n\t\t}\n\t}\n\n\ts.Postings = make([]*roaring.Bitmap, numPostingsLists)\n\tfor i := 0; i < numPostingsLists; i++ {\n\t\ts.Postings[i] = roaring.New()\n\t}\n\ts.PostingsLocs = make([]*roaring.Bitmap, numPostingsLists)\n\tfor i := 0; i < numPostingsLists; i++ {\n\t\ts.PostingsLocs[i] = roaring.New()\n\t}\n\n\t// Preallocate big, contiguous backing arrays.\n\tauint64Backing := make([][]uint64, numPostingsLists*4+totLocs) // For Freqs, Locstarts, Locends, Locpos, sub-Locarraypos.\n\tuint64Backing := make([]uint64, numTokenFrequencies+totLocs*3) // For sub-Freqs, sub-Locstarts, sub-Locends, sub-Locpos.\n\tfloat32Backing := make([]float32, numTokenFrequencies) // For sub-Norms.\n\tuint16Backing := make([]uint16, totLocs) // For sub-Locfields.\n\n\t// Point top-level slices to the backing arrays.\n\ts.Freqs = auint64Backing[0:numPostingsLists]\n\tauint64Backing = auint64Backing[numPostingsLists:]\n\n\ts.Norms = make([][]float32, numPostingsLists)\n\n\ts.Locfields = make([][]uint16, numPostingsLists)\n\n\ts.Locstarts = auint64Backing[0:numPostingsLists]\n\tauint64Backing = auint64Backing[numPostingsLists:]\n\n\ts.Locends = auint64Backing[0:numPostingsLists]\n\tauint64Backing = auint64Backing[numPostingsLists:]\n\n\ts.Locpos = auint64Backing[0:numPostingsLists]\n\tauint64Backing = auint64Backing[numPostingsLists:]\n\n\ts.Locarraypos = make([][][]uint64, numPostingsLists)\n\n\t// Point sub-slices to the backing arrays.\n\tfor pid, numTerms := range numTermsPerPostingsList {\n\t\ts.Freqs[pid] = uint64Backing[0:0]\n\t\tuint64Backing = uint64Backing[numTerms:]\n\n\t\ts.Norms[pid] = float32Backing[0:0]\n\t\tfloat32Backing = float32Backing[numTerms:]\n\t}\n\n\tfor pid, numLocs := range numLocsPerPostingsList {\n\t\ts.Locfields[pid] = uint16Backing[0:0]\n\t\tuint16Backing = uint16Backing[numLocs:]\n\n\t\ts.Locstarts[pid] = uint64Backing[0:0]\n\t\tuint64Backing = uint64Backing[numLocs:]\n\n\t\ts.Locends[pid] = uint64Backing[0:0]\n\t\tuint64Backing = uint64Backing[numLocs:]\n\n\t\ts.Locpos[pid] = uint64Backing[0:0]\n\t\tuint64Backing = uint64Backing[numLocs:]\n\n\t\ts.Locarraypos[pid] = auint64Backing[0:0]\n\t\tauint64Backing = auint64Backing[numLocs:]\n\t}\n}", "func BenchmarkUnionMap(b *testing.B) {\n\tunionWithInit(b)\n\tfor n := 0; n < b.N; n++ {\n\t\tb.StopTimer()\n\t\tdata := unionIntMapData1.Dup()\n\t\tb.StartTimer()\n\t\tdata.UnionWith(unionIntMapData2)\n\t}\n}", "func Update(key Type, val time.Duration) {\n\tif GlobalStats != nil {\n\t\tmutex.Lock()\n\t\tdefer mutex.Unlock()\n\t\tGlobalStats[key] = GlobalStats[key] + val\n\t}\n}", "func (f *falconMeter) Update() {\n\tfalcon.SetMeterCount(f.name, 1)\n}", "func reloadInstrumentsMap(instruments []Instrument) {\n\n\t// here we translate the api output format (instruments with all their attributes)\n\t// into a frendlier format with what we are trying to achieve: a map of instruments slices by exchange\n\t// exampe of filling:\n\t// res := append(m[\"binance\"], \"BTCUSD\")\n\t// m[\"bitfinex\"] = []string{\"BTC-USD\"}\n\t// example of usage:\n\t// binanceInstruments = m[\"binance\"]\n\n\t// first we instantiate/reset the map\n\tinstrumentsByExchange = make(map[string][]string)\n\n\t// now we loop through the api results to match our map format\n\tfor _, instrument := range instruments {\n\t\tif exchangeExists(instrument.ExchangeCode) {\n\t\t\t// key (exchange) already exists\n\t\t\tinstrumentsByExchange[instrument.ExchangeCode] = append(instrumentsByExchange[instrument.ExchangeCode], instrument.ExchangePairCode)\n\t\t} else {\n\t\t\t// new key (exchange)\n\t\t\tinstrumentsByExchange[instrument.ExchangeCode] = []string{instrument.ExchangePairCode}\n\t\t}\n\t}\n\t// once this is over, we can safely save our map in a persistent cache\n\t// so that we don't have to do this again if the server is restarted\n\t// or RAM is flushed\n\tif len(instrumentsByExchange) > 1 {\n\t\t// let's build the new cache from the fresh data\n\t\terr := instrumentsCache.writeSerialize(instrumentsByExchange)\n\t\tif err != nil {\n\t\t\tprintln(\"instruments cache could not be updated:\", err)\n\t\t}\n\t}\n}", "func (sub *subState) updateRedeliveryCount(m *pb.MsgProto) {\n\tvar rdlvCountMap *map[uint64]uint32\n\tif sub.qstate != nil {\n\t\trdlvCountMap = &sub.qstate.rdlvCount\n\t} else {\n\t\trdlvCountMap = &sub.rdlvCount\n\t}\n\tif *rdlvCountMap == nil {\n\t\t*rdlvCountMap = make(map[uint64]uint32)\n\t}\n\tfor {\n\t\t(*rdlvCountMap)[m.Sequence]++\n\t\tm.RedeliveryCount = (*rdlvCountMap)[m.Sequence]\n\t\t// Just in case we rolled over...\n\t\tif m.RedeliveryCount != 0 {\n\t\t\tbreak\n\t\t}\n\t}\n}", "func (pm *PmMetrics) UpdateFrequency(frequency uint32) {\n\tpm.frequency = frequency\n}", "func (m *ConcurrentMap) Count() int {\n\tcount := 0\n\tfor i := 0; i < SHARD_COUNT; i++ {\n\t\tshard := m.Shareds[i]\n\t\tshard.RLock()\n\t\tcount += len(shard.items)\n\t\tshard.RUnlock()\n\t}\n\treturn count\n}", "func (c *Cache) updateStats() error {\n\tvar newUsed int64\n\terr := c.walk(func(osPath string, fi os.FileInfo, name string) error {\n\t\tif !fi.IsDir() {\n\t\t\t// Update the atime with that of the file\n\t\t\tatime := times.Get(fi).AccessTime()\n\t\t\tc.updateStat(name, atime, fi.Size())\n\t\t\tnewUsed += fi.Size()\n\t\t} else {\n\t\t\tc.cacheDir(name)\n\t\t}\n\t\treturn nil\n\t})\n\tc.itemMu.Lock()\n\tc.used = newUsed\n\tc.itemMu.Unlock()\n\treturn err\n}", "func (pc *NginxProcessesMetricsCollector) updateWorkerProcessCount() {\n\tcurrWorkerProcesses, prevWorkerProcesses, err := getWorkerProcesses()\n\tif err != nil {\n\t\tglog.Errorf(\"unable to collect process metrics : %v\", err)\n\t\treturn\n\t}\n\tpc.workerProcessTotal.WithLabelValues(\"current\").Set(float64(currWorkerProcesses))\n\tpc.workerProcessTotal.WithLabelValues(\"old\").Set(float64(prevWorkerProcesses))\n}", "func (b *BPFMap) Update(value string) error {\n\n\tlog.Debug().Msgf(\"update map name %s ID %d\", b.Name, b.MapID)\n\tebpfMap, err := ebpf.NewMapFromID(b.MapID)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"access new map from ID failed %v\", err)\n\t}\n\tdefer ebpfMap.Close()\n\n\t// check values are single or multiple\n\ts := strings.Split(value, \",\")\n\n\tif b.Type == ebpf.Hash {\n\t\t// clear map elements\n\t\tkey := 0\n\t\tval := 0\n\t\tentries := ebpfMap.Iterate()\n\t\tfor entries.Next(unsafe.Pointer(&key), unsafe.Pointer(&val)) {\n\t\t\t// Order of keys is non-deterministic due to randomized map seed\n\t\t\tif err := ebpfMap.Delete(unsafe.Pointer(&key)); err != nil {\n\t\t\t\tlog.Warn().Err(err).Msgf(\"delete hash map for key %d failed\", key)\n\t\t\t}\n\t\t}\n\n\t\tfor key, val := range s {\n\t\t\tv, _ := strconv.ParseInt(val, 10, 64)\n\t\t\tx := 1\n\t\t\tlog.Info().Msgf(\"updating map %s key %d mapid %d\", b.Name, v, b.MapID)\n\t\t\tif err := ebpfMap.Update(unsafe.Pointer(&v), unsafe.Pointer(&x), 0); err != nil {\n\t\t\t\treturn fmt.Errorf(\"update hash map element failed for key %d error %v\", key, err)\n\t\t\t}\n\t\t}\n\t} else if b.Type == ebpf.Array {\n\t\tfor key, val := range s {\n\t\t\tv, _ := strconv.ParseInt(val, 10, 64)\n\t\t\tlog.Info().Msgf(\"updating map %s key %d mapid %d\", b.Name, v, b.MapID)\n\t\t\tif err := ebpfMap.Update(unsafe.Pointer(&key), unsafe.Pointer(&v), 0); err != nil {\n\t\t\t\treturn fmt.Errorf(\"update array map index %d %v\\n\", key, err)\n\t\t\t}\n\t\t}\n\t} else {\n\t\treturn fmt.Errorf(\"unsupported map type\")\n\t}\n\treturn nil\n}", "func (ds *Dataset) update(value float64) {\n\tds.total += value\n\tds.product *= value\n\tds.recipsum += 1 / value\n\tds.min = math.Min(ds.min, value)\n\tds.max = math.Max(ds.max, value)\n}", "func (m *MAF) Update(classification map[string]float64) (map[string]float64, error) {\n\tif m.state == nil {\n\t\treturn nil, fmt.Errorf(\"invalid MAF, use NewMAF\")\n\t}\n\tif len(classification) == 0 {\n\t\treturn nil, fmt.Errorf(\"classification must not be empty\")\n\t}\n\n\t// todo: check that all labels from initialization are present?\n\t// todo: for the first \"size\" updates, only take present values into account?\n\n\tr := map[string]float64{}\n\tfor label, value := range classification {\n\t\tls, ok := m.state[label]\n\t\tif !ok {\n\t\t\treturn nil, fmt.Errorf(\"unknown label %q\", label)\n\t\t}\n\t\tls.sum -= ls.values[ls.index]\n\t\tls.sum += value\n\t\tls.values[ls.index] = value\n\t\tr[label] = ls.sum / float64(len(ls.values))\n\t\tls.index++\n\t\tif ls.index >= len(ls.values) {\n\t\t\tls.index = 0\n\t\t}\n\t}\n\treturn r, nil\n}", "func (m *Metric) Update(v float64) *Metric {\n\tm.Lock()\n\tm.LastValue = v\n\tm.Sum += v\n\tif v < m.Min {\n\t\tm.Min = v\n\t}\n\tif v > m.Max {\n\t\tm.Max = v\n\t}\n\tm.Samples++\n\tm.Unlock()\n\treturn m\n}", "func (v *verifiableMetric) update(data *Data, fieldValues string, packer *numberPacker) {\n\tswitch v.wantMetric.Type {\n\tcase TypeCounter:\n\t\tv.lastCounterValue[v.verifier.internMap.Intern(fieldValues)] = packer.pack(data.Number)\n\tcase TypeHistogram:\n\t\tlastDistributionSnapshot := v.lastDistributionSnapshot[v.verifier.internMap.Intern(fieldValues)]\n\t\tlastBucketSamples := lastDistributionSnapshot.numSamples\n\t\tvar count uint64\n\t\tfor i, b := range data.HistogramValue.Buckets {\n\t\t\tlastBucketSamples[i] = packer.packInt(int64(b.Samples))\n\t\t\tcount += b.Samples\n\t\t}\n\t\tlastDistributionSnapshot.sum = packer.pack(&data.HistogramValue.Total)\n\t\tlastDistributionSnapshot.count = packer.packInt(int64(count))\n\t\tlastDistributionSnapshot.min = packer.pack(&data.HistogramValue.Min)\n\t\tlastDistributionSnapshot.max = packer.pack(&data.HistogramValue.Max)\n\t\tlastDistributionSnapshot.ssd = packer.pack(&data.HistogramValue.SumOfSquaredDeviations)\n\t}\n}", "func (m *ConcurrentMap) Count() int {\n\tcount := 0\n\tfor i := 0; i < SHARD_COUNT; i++ {\n\t\tshard := m.shards[i]\n\t\tshard.RLock()\n\t\tcount += len(shard.items)\n\t\tshard.RUnlock()\n\t}\n\treturn count\n}", "func (set *Beatmaps) Update() error {\n\ttemp, err := set.session.FetchBeatmaps(set.apiCall)\n\t*set = temp\n\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n\n\t// Allows for the updating of individual beatmaps\n\tfor i := 0; i < len(set.Beatmaps); i++ {\n\t\tv := url.Values{}\n\t\tv.Add(endpointAPIKey, set.session.key)\n\n\t\tif set.Converted != \"\" {\n\t\t\tv.Add(endpointParamConverted, set.Converted)\n\t\t}\n\n\t\tv.Add(endpointParamBeatmapID, set.Beatmaps[i].BeatmapID)\n\t\tv.Add(endpointParamMode, set.Beatmaps[i].Mode)\n\t\tset.Beatmaps[i].apiURL = set.session.buildCall(endpointBeatmaps, v)\n\t\tset.Beatmaps[i].session = set.session\n\t}\n\n\treturn nil\n}", "func add(finalMap dict, m dict) { // redfn\n\tfor key, value := range m {\n\t\tfinalMap[key] += value\n\t}\n}", "func (tfm *trxFlowMonitor) updateCount() {\n\t// pull the value from DB\n\tval, err := repo.TransactionsCount()\n\tif err != nil {\n\t\tlog.Errorf(\"can not update trx count estimation; %s\", err.Error())\n\t\treturn\n\t}\n\n\t// update the estimate\n\trepo.UpdateTrxCountEstimate(val)\n}", "func (s *SweepPruneSet[T]) updateMarkers() {\n\tfor i := range s.seriesX {\n\t\tmarker := &s.seriesX[i]\n\t\titemIndex := marker.ItemIndex\n\t\tif _, ok := s.dirtyItemIDs[itemIndex]; ok {\n\t\t\tmarker.Coord = s.items[itemIndex].BoundaryX(marker.IsStart)\n\t\t}\n\t}\n\tfor i := range s.seriesY {\n\t\tmarker := &s.seriesY[i]\n\t\titemIndex := marker.ItemIndex\n\t\tif _, ok := s.dirtyItemIDs[itemIndex]; ok {\n\t\t\tmarker.Coord = s.items[itemIndex].BoundaryY(marker.IsStart)\n\t\t}\n\t}\n\tfor i := range s.seriesZ {\n\t\tmarker := &s.seriesZ[i]\n\t\titemIndex := marker.ItemIndex\n\t\tif _, ok := s.dirtyItemIDs[itemIndex]; ok {\n\t\t\tmarker.Coord = s.items[itemIndex].BoundaryZ(marker.IsStart)\n\t\t}\n\t}\n\tmaps.Clear(s.dirtyItemIDs)\n}", "func newMapCache(dss map[string]rrd.DataSourcer) *mapCache {\n\tmc := &mapCache{make(map[string]int64), make(map[int64]rrd.DataSourcer)}\n\tvar n int64\n\tfor name, ds := range dss {\n\t\tmc.byName[name] = n\n\t\tmc.byId[n] = ds\n\t\tn++\n\t}\n\treturn mc\n}", "func update(weight, mean, sumSq, sampleWeight, sample, sampleSumSq float64) (newWeight, newMean, newSumSq float64) {\n\tnewWeight = weight + sampleWeight\n\tnewMean = (weight*mean + sample*sampleWeight) / newWeight\n\tnewSumSq = sumSq + sampleSumSq + (sample-mean)*(sample-mean)*weight*sampleWeight/newWeight\n\treturn newWeight, newMean, newSumSq\n}", "func (settings Settings) UpdateDistances(lat float64, lon float64) {\n\tfor i, server := range settings.Servers {\n\t\tsettings.Servers[i].Distance = Distance(\n\t\t\tserver.Lat*degToRad, server.Lon*degToRad,\n\t\t\tlat*degToRad, lon*degToRad)\n\t}\n}", "func updateExplore(loc Location, steps int, m *Map) {\n\tm.ExploreVal[loc] += 1.0 / float64(steps)\n\treturn\n}", "func (d *Demo) UpdateMap(g *gom.Gom) {\n\ttoolkit.Println(\"===== Update With Map =====\")\n\thero := bson.M{\n\t\t\"Name\": \"Hawkeye\",\n\t\t\"RealName\": \"Jeremy Renner\",\n\t\t\"Age\": 33,\n\t}\n\n\tvar err error\n\tif d.useParams {\n\t\t_, err = g.Set(&gom.SetParams{\n\t\t\tTableName: \"hero\",\n\t\t\tFilter: gom.Eq(\"Name\", \"Wolverine\"),\n\t\t\tTimeout: 10,\n\t\t}).Cmd().Update(&hero)\n\t} else {\n\t\t_, err = g.Set(nil).Table(\"hero\").Timeout(10).Filter(gom.Eq(\"Name\", \"Wolverine\")).Cmd().Update(&hero)\n\t}\n\n\tif err != nil {\n\t\ttoolkit.Println(err.Error())\n\t\treturn\n\t}\n}", "func (h *Handler) UpdateCount() {\n\th.Client.Do(\"INCR\", \"visitors\")\n}", "func updateConfigMapFunc(f *framework.Framework, tc *nodeConfigTestCase) error {\n\t// Clear ResourceVersion from the ConfigMap objects we use to initiate mutations\n\t// so that we don't get 409 (conflict) responses. ConfigMaps always allow updates\n\t// (with respect to concurrency control) when you omit ResourceVersion.\n\t// We know that we won't perform concurrent updates during this test.\n\ttc.configMap.ResourceVersion = \"\"\n\tcm, err := f.ClientSet.CoreV1().ConfigMaps(tc.configMap.Namespace).Update(tc.configMap)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// update tc.configMap's ResourceVersion to match the updated ConfigMap, this makes\n\t// sure our derived status checks have up-to-date information\n\ttc.configMap.ResourceVersion = cm.ResourceVersion\n\treturn nil\n}", "func (mc *MapChanges) AccumulateMapChanges(cs CachedSelector, adds, deletes []identity.NumericIdentity,\n\tport uint16, proto uint8, direction trafficdirection.TrafficDirection,\n\tredirect, isDeny bool, derivedFrom labels.LabelArrayList) {\n\tkey := Key{\n\t\t// The actual identity is set in the loops below\n\t\tIdentity: 0,\n\t\t// NOTE: Port is in host byte-order!\n\t\tDestPort: port,\n\t\tNexthdr: proto,\n\t\tTrafficDirection: direction.Uint8(),\n\t}\n\n\tvalue := NewMapStateEntry(cs, derivedFrom, redirect, isDeny)\n\n\tif option.Config.Debug {\n\t\tlog.WithFields(logrus.Fields{\n\t\t\tlogfields.EndpointSelector: cs,\n\t\t\tlogfields.AddedPolicyID: adds,\n\t\t\tlogfields.DeletedPolicyID: deletes,\n\t\t\tlogfields.Port: port,\n\t\t\tlogfields.Protocol: proto,\n\t\t\tlogfields.TrafficDirection: direction,\n\t\t\tlogfields.IsRedirect: redirect,\n\t\t}).Debug(\"AccumulateMapChanges\")\n\t}\n\n\tmc.mutex.Lock()\n\tfor _, id := range adds {\n\t\tkey.Identity = id.Uint32()\n\t\tmc.changes = append(mc.changes, MapChange{Add: true, Key: key, Value: value})\n\t}\n\tfor _, id := range deletes {\n\t\tkey.Identity = id.Uint32()\n\t\tmc.changes = append(mc.changes, MapChange{Add: false, Key: key, Value: value})\n\t}\n\tmc.mutex.Unlock()\n}", "func (s *SliceOfFloat64) Map(mapFunc func(float64) float64) *SliceOfFloat64 {\n\tfor index, value := range s.items {\n\t\ts.items[index] = mapFunc(value)\n\t}\n\treturn s\n}", "func calculateCount(cMap map[string]int) int {\n\tcount := 0\n\tfor _, val := range cMap {\n\t\tif val > 0 {\n\t\t\tcount++\n\t\t}\n\t\tif count == 5 {\n\t\t\tbreak\n\t\t}\n\t}\n\treturn count\n}", "func ThrottleMap(throttlers []*Throttler) map[string]*HTTPThrottler {\n\tthrottledMap := make(map[string]*HTTPThrottler)\n\tfor _, throttle := range throttlers {\n\t\tdelay, err := throttle.GetDuration()\n\t\tif err != nil {\n\t\t\tcontinue // Error is logged in Duration function.\n\t\t}\n\t\t// Initialize the latest fetch to yesterday.\n\t\tthrottledMap[throttle.Host] = &HTTPThrottler{delay: delay, latestFetch: time.Now().AddDate(0, 0, -1)}\n\t}\n\treturn throttledMap\n}", "func (m *metricActiveDirectoryDsNameCacheHitRate) updateCapacity() {\n\tif m.data.Gauge().DataPoints().Len() > m.capacity {\n\t\tm.capacity = m.data.Gauge().DataPoints().Len()\n\t}\n}", "func (d DestinationCounters) Update(duration time.Duration, failed bool) {\n\tif failed {\n\t\td.failedTotalCount.Inc()\n\t\td.failedDuration.Observe(duration.Seconds())\n\t} else {\n\t\td.totalCount.Inc()\n\t\td.duration.Observe(duration.Seconds())\n\t}\n}", "func BenchmarkRoundRobinUpdate(b *testing.B) {\n\tfor k := 0.; k <= 10; k++ {\n\t\tn := int(math.Pow(2, k))\n\t\tb.Run(fmt.Sprintf(\"update/%d\", n), func(b *testing.B) {\n\t\t\tb.ReportAllocs()\n\t\t\tbalancer := NewRoundRobinBalancer()\n\t\t\tb.ResetTimer()\n\n\t\t\tfor i := 0; i < b.N; i++ {\n\t\t\t\tb.StopTimer()\n\t\t\t\troutes := make([]Route, n)\n\t\t\t\tfor index := 0; index < n; index++ {\n\t\t\t\t\troutes[index] = testRoute(IPv4Address(int64(index)))\n\t\t\t\t}\n\t\t\t\tb.StartTimer()\n\n\t\t\t\tbalancer.Update(routes)\n\t\t\t}\n\t\t})\n\t}\n}", "func (s *samplerBackendRateCounter) Count() {\n\ts.backend.countSample()\n}", "func mapBottles(bottles []Bottle) map[uint32]int {\n\tbottleMap := make(map[uint32]int, len(bottles))\n\tfor _, bottle := range bottles {\n\t\t_, ok := bottleMap[bottle.VolumeML]\n\t\tif !ok {\n\t\t\tbottleMap[bottle.VolumeML] = 1\n\t\t\tcontinue\n\t\t}\n\t\tbottleMap[bottle.VolumeML]++\n\t}\n\treturn bottleMap\n}", "func (h *healthcheckManager) buildMaps() {\n\tallChecks := make(map[CheckKey]*check)\n\tfor _, vchecks := range h.vserverChecks {\n\t\tfor k, c := range vchecks {\n\t\t\tif allChecks[k] == nil {\n\t\t\t\tallChecks[k] = c\n\t\t\t} else {\n\t\t\t\tlog.Warningf(\"Duplicate key: %v\", k)\n\t\t\t}\n\t\t}\n\t}\n\n\th.lock.RLock()\n\tids := h.ids\n\tcfgs := h.cfgs\n\th.lock.RUnlock()\n\tnewIDs := make(map[checkerKey]healthcheck.Id)\n\tnewCfgs := make(map[healthcheck.Id]*healthcheck.Config)\n\tnewChecks := make(map[healthcheck.Id][]*check)\n\n\tfor key, c := range allChecks {\n\t\tcKey := checkerKey{\n\t\t\tkey: dedup(key),\n\t\t\tcfg: *c.healthcheck,\n\t\t}\n\t\tid, ok := ids[cKey]\n\t\tif !ok {\n\t\t\tid = h.next\n\t\t\th.next++\n\t\t}\n\t\tcfg, ok := cfgs[id]\n\t\tif !ok {\n\t\t\tnewCfg, err := h.newConfig(id, cKey.key, c.healthcheck)\n\t\t\tif err != nil {\n\t\t\t\tlog.Error(err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tcfg = newCfg\n\t\t}\n\n\t\tnewIDs[cKey] = id\n\t\tnewCfgs[id] = cfg\n\t\tnewChecks[id] = append(newChecks[id], c)\n\t}\n\n\th.lock.Lock()\n\th.ids = newIDs\n\th.cfgs = newCfgs\n\th.checks = newChecks\n\th.lock.Unlock()\n\n\th.pruneMarks()\n}", "func (vpr *valueProposerRoot) updatePolicyRefForResources(newMap map[string]string) {\n\tlogger.Debugf(\"Updating policy ref map for %p\", vpr)\n\tvpr.resourceToPolicyRefMap.Store(newMap)\n}", "func (ms *metricsStore) update(obj runtime.Object, rms []*resourceMetrics) error {\n\tms.Lock()\n\tdefer ms.Unlock()\n\n\tkey, err := utils.GetUIDForObject(obj)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\torigMds := make([]agentmetricspb.ExportMetricsServiceRequest, len(rms))\n\tmds := make([]*agentmetricspb.ExportMetricsServiceRequest, len(rms))\n\tfor i, rm := range rms {\n\t\tmds[i] = &origMds[i]\n\t\tmds[i].Resource = rm.resource\n\t\tmds[i].Metrics = rm.metrics\n\t}\n\n\tms.metricsCache[key] = mds\n\treturn nil\n}", "func UpdateNodesCount(ready, unready, starting, longUnregistered, unregistered int) {\n\tnodesCount.WithLabelValues(readyLabel).Set(float64(ready))\n\tnodesCount.WithLabelValues(unreadyLabel).Set(float64(unready))\n\tnodesCount.WithLabelValues(startingLabel).Set(float64(starting))\n\tnodesCount.WithLabelValues(longUnregisteredLabel).Set(float64(longUnregistered))\n\tnodesCount.WithLabelValues(unregisteredLabel).Set(float64(unregistered))\n}", "func (tc *LegacyTabletStatsCache) StatsUpdate(ts *LegacyTabletStats) {\n\tif ts.Target.TabletType != topodatapb.TabletType_MASTER &&\n\t\tts.Tablet.Alias.Cell != tc.cell &&\n\t\ttc.getAliasByCell(ts.Tablet.Alias.Cell) != tc.getAliasByCell(tc.cell) {\n\t\t// this is for a non-master tablet in a different cell and a different alias, drop it\n\t\treturn\n\t}\n\n\te := tc.getOrCreateEntry(ts.Target)\n\te.mu.Lock()\n\tdefer e.mu.Unlock()\n\n\t// Update our full map.\n\ttrivialNonMasterUpdate := false\n\tif existing, ok := e.all[ts.Key]; ok {\n\t\tif ts.Up {\n\t\t\t// We have an existing entry, and a new entry.\n\t\t\t// Remember if they are both good (most common case).\n\t\t\ttrivialNonMasterUpdate = existing.LastError == nil && existing.Serving && ts.LastError == nil &&\n\t\t\t\tts.Serving && ts.Target.TabletType != topodatapb.TabletType_MASTER && existing.TrivialStatsUpdate(ts)\n\n\t\t\t// We already have the entry, update the\n\t\t\t// values if necessary. (will update both\n\t\t\t// 'all' and 'healthy' as they use pointers).\n\t\t\tif !trivialNonMasterUpdate {\n\t\t\t\t*existing = *ts\n\t\t\t}\n\t\t} else {\n\t\t\t// We have an entry which we shouldn't. Remove it.\n\t\t\tdelete(e.all, ts.Key)\n\t\t}\n\t} else {\n\t\tif ts.Up {\n\t\t\t// Add the entry.\n\t\t\te.all[ts.Key] = ts\n\t\t} else {\n\t\t\t// We were told to remove an entry which we\n\t\t\t// didn't have anyway, nothing should happen.\n\t\t\treturn\n\t\t}\n\t}\n\n\t// Update our healthy list.\n\tvar allArray []*LegacyTabletStats\n\tif ts.Target.TabletType == topodatapb.TabletType_MASTER {\n\t\t// The healthy list is different for TabletType_MASTER: we\n\t\t// only keep the most recent one.\n\t\te.updateHealthyMapForMaster(ts)\n\t} else {\n\t\t// For non-master, if it is a trivial update,\n\t\t// we just skip everything else. We don't even update the\n\t\t// aggregate stats.\n\t\tif trivialNonMasterUpdate {\n\t\t\treturn\n\t\t}\n\n\t\t// Now we need to do some work. Recompute our healthy list.\n\t\tallArray = make([]*LegacyTabletStats, 0, len(e.all))\n\t\tfor _, s := range e.all {\n\t\t\tallArray = append(allArray, s)\n\t\t}\n\t\te.healthy = FilterLegacyStatsByReplicationLag(allArray)\n\t}\n}", "func (m Map) Count() int {\n\treturn m.Imm.Len()\n}", "func updateFood(loc Location, steps int, m *Map) {\n\tm.FoodVal[loc] += 1.0 / float64(steps)\n\treturn\n}", "func (s HelpAppUpdateArray) FillMap(to map[int]HelpAppUpdate) {\n\tfor _, value := range s {\n\t\tto[value.GetID()] = value\n\t}\n}", "func (sr *ServicedStatsReporter) updateStats() {\n\t// Stats for host.\n\tsr.updateHostStats()\n\t// Stats for the containers.\n\tstates, err := zkservice.GetHostStates(sr.conn, \"\", sr.hostID)\n\tif err != nil {\n\t\tplog.WithFields(logrus.Fields{\n\t\t\t\"conn\": sr.conn,\n\t\t\t\"hostID\": sr.hostID,\n\t\t}).WithError(err).Error(\"Could not get host states from Zookeeper\")\n\t}\n\n\tfor _, rs := range states {\n\t\tif rs.ContainerID != \"\" {\n\n\t\t\tcontainerRegistry := sr.getOrCreateContainerRegistry(rs.ServiceID, rs.InstanceID)\n\t\t\tstats, err := sr.docker.GetContainerStats(rs.ContainerID, 30*time.Second)\n\t\t\tif err != nil || stats == nil { //stats may be nil if service is shutting down\n\t\t\t\tplog.WithFields(logrus.Fields{\n\t\t\t\t\t\"serviceID\": rs.ServiceID,\n\t\t\t\t\t\"instanceID\": rs.InstanceID,\n\t\t\t\t}).WithError(err).Warn(\"Couldn't get stats from docker\")\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// Check to see if we have the previous stats for this running instance\n\t\t\tusePreviousStats := true\n\t\t\tkey := rs.ContainerID\n\t\t\tif _, found := sr.previousStats[key]; !found {\n\t\t\t\tsr.previousStats[key] = make(map[string]uint64)\n\t\t\t\tusePreviousStats = false\n\t\t\t}\n\n\t\t\t// CPU Stats\n\t\t\t// TODO: Consolidate this into a single object that both ISVCS and non-ISVCS can use\n\t\t\tvar (\n\t\t\t\tkernelCPUPercent float64\n\t\t\t\tuserCPUPercent float64\n\t\t\t\ttotalCPUChange uint64\n\t\t\t)\n\n\t\t\tkernelCPU := stats.CPUStats.CPUUsage.UsageInKernelmode\n\t\t\tuserCPU := stats.CPUStats.CPUUsage.UsageInUsermode\n\t\t\ttotalCPU := stats.CPUStats.SystemCPUUsage\n\n\t\t\t// Total CPU Cycles\n\t\t\tpreviousTotalCPU, found := sr.previousStats[key][\"totalCPU\"]\n\t\t\tif found {\n\t\t\t\tif totalCPU <= previousTotalCPU {\n\t\t\t\t\tplog.WithFields(logrus.Fields{\n\t\t\t\t\t\t\"totalCPU\": totalCPU,\n\t\t\t\t\t\t\"previousTotalCPU\": previousTotalCPU,\n\t\t\t\t\t}).Debug(\"Change in total CPU usage was nonpositive, skipping CPU stats update\")\n\t\t\t\t\tusePreviousStats = false\n\t\t\t\t} else {\n\t\t\t\t\ttotalCPUChange = totalCPU - previousTotalCPU\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tusePreviousStats = false\n\t\t\t}\n\t\t\tsr.previousStats[key][\"totalCPU\"] = totalCPU\n\n\t\t\t// CPU Cycles in Kernel mode\n\t\t\tif previousKernelCPU, found := sr.previousStats[key][\"kernelCPU\"]; found && usePreviousStats {\n\t\t\t\tkernelCPUChange := kernelCPU - previousKernelCPU\n\t\t\t\tkernelCPUPercent = (float64(kernelCPUChange) / float64(totalCPUChange)) * float64(len(stats.CPUStats.CPUUsage.PercpuUsage)) * 100.0\n\t\t\t} else {\n\t\t\t\tusePreviousStats = false\n\t\t\t}\n\t\t\tsr.previousStats[key][\"kernelCPU\"] = kernelCPU\n\n\t\t\t// CPU Cycles in User mode\n\t\t\tif previousUserCPU, found := sr.previousStats[key][\"userCPU\"]; found && usePreviousStats {\n\t\t\t\tuserCPUChange := userCPU - previousUserCPU\n\t\t\t\tuserCPUPercent = (float64(userCPUChange) / float64(totalCPUChange)) * float64(len(stats.CPUStats.CPUUsage.PercpuUsage)) * 100.0\n\t\t\t} else {\n\t\t\t\tusePreviousStats = false\n\t\t\t}\n\t\t\tsr.previousStats[key][\"userCPU\"] = userCPU\n\n\t\t\t// Update CPU metrics\n\t\t\tif usePreviousStats {\n\t\t\t\tmetrics.GetOrRegisterGaugeFloat64(\"docker.usageinkernelmode\", containerRegistry).Update(kernelCPUPercent)\n\t\t\t\tmetrics.GetOrRegisterGaugeFloat64(\"docker.usageinusermode\", containerRegistry).Update(userCPUPercent)\n\t\t\t} else {\n\t\t\t\tplog.WithFields(logrus.Fields{\n\t\t\t\t\t\"serviceID\": rs.ServiceID,\n\t\t\t\t\t\"instanceID\": rs.InstanceID,\n\t\t\t\t}).Debug(\"Skipping CPU stats, no previous values to compare to\")\n\t\t\t}\n\n\t\t\t// Memory Stats\n\t\t\tpgFault := int64(stats.MemoryStats.Stats.Pgfault)\n\t\t\ttotalRSS := int64(stats.MemoryStats.Stats.TotalRss)\n\t\t\tcache := int64(stats.MemoryStats.Stats.Cache)\n\t\t\tif pgFault < 0 || totalRSS < 0 || cache < 0 {\n\t\t\t\tplog.WithFields(logrus.Fields{\n\t\t\t\t\t\"serviceID\": rs.ServiceID,\n\t\t\t\t\t\"instanceID\": rs.InstanceID,\n\t\t\t\t}).Debug(\"Memory metric value too big for int64\")\n\t\t\t}\n\t\t\tmetrics.GetOrRegisterGauge(\"cgroup.memory.pgmajfault\", containerRegistry).Update(pgFault)\n\t\t\tmetrics.GetOrRegisterGauge(\"cgroup.memory.totalrss\", containerRegistry).Update(totalRSS)\n\t\t\tmetrics.GetOrRegisterGauge(\"cgroup.memory.cache\", containerRegistry).Update(cache)\n\n\t\t} else {\n\t\t\tplog.WithFields(logrus.Fields{\n\t\t\t\t\"serviceID\": rs.ServiceID,\n\t\t\t\t\"instanceID\": rs.InstanceID,\n\t\t\t}).Debug(\"Skipping stats update, no container ID exists\")\n\t\t}\n\t}\n\t// Clean out old container registries\n\tsr.removeStaleRegistries(states)\n}", "func (m candidateMap) update(ticket []uint64, rules ruleSet) {\n\tindexMap := map[int][]string{}\n\tfor idx, value := range ticket {\n\t\tvalidFields := rules.validRules(value)\n\t\tif len(validFields) == 0 {\n\t\t\t// Ignore the whole ticket\n\t\t\tfmt.Printf(\"Ignoring ticket %+v\\n\", ticket)\n\t\t\treturn\n\t\t}\n\t\tindexMap[idx] = validFields\n\t}\n\t// Now remove each index from the fields not listed in its candidate list\n\tfor idx, fieldList := range indexMap {\n\t\tfor fieldName, candidates := range m {\n\t\t\tif !contains(fieldList, fieldName) {\n\t\t\t\t// Remove our index from the candidate list\n\t\t\t\tvar c []int\n\t\t\t\tfor _, index := range candidates {\n\t\t\t\t\tif index != idx {\n\t\t\t\t\t\tc = append(c, index)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tm[fieldName] = c\n\t\t\t}\n\t\t}\n\t}\n}", "func (m *metricFlinkJvmGcCollectionsCount) updateCapacity() {\n\tif m.data.Sum().DataPoints().Len() > m.capacity {\n\t\tm.capacity = m.data.Sum().DataPoints().Len()\n\t}\n}" ]
[ "0.62803847", "0.60310405", "0.5673678", "0.56716883", "0.56316656", "0.5610204", "0.52624536", "0.52363783", "0.52031183", "0.51712793", "0.51409674", "0.50864744", "0.50659347", "0.5059397", "0.5045441", "0.50371087", "0.50339305", "0.5005483", "0.49625853", "0.49607196", "0.49600342", "0.4938792", "0.4925258", "0.49193507", "0.49137855", "0.48994312", "0.48970008", "0.4889318", "0.4888421", "0.48873535", "0.48822016", "0.48765555", "0.48762468", "0.48757654", "0.48639584", "0.48622337", "0.4855081", "0.4852194", "0.48517513", "0.4844805", "0.4844364", "0.48441607", "0.48351148", "0.48234737", "0.48165035", "0.48126507", "0.47964242", "0.47844", "0.4778738", "0.47753295", "0.47652784", "0.47614282", "0.47574702", "0.47568592", "0.47522113", "0.4750254", "0.47366062", "0.47332633", "0.47282016", "0.4726216", "0.47260496", "0.47246662", "0.47192433", "0.47164693", "0.47163242", "0.47138914", "0.4712338", "0.47086892", "0.4708235", "0.47067636", "0.4706646", "0.46985176", "0.46907288", "0.4681468", "0.4671516", "0.4669671", "0.46627116", "0.4658646", "0.46561274", "0.46520978", "0.46517026", "0.46508157", "0.46457306", "0.46402484", "0.46335343", "0.4629985", "0.4622704", "0.46172783", "0.46167222", "0.4613053", "0.46110082", "0.46081054", "0.46045944", "0.46001828", "0.45989627", "0.45958367", "0.45955738", "0.45885608", "0.4588555", "0.45873493" ]
0.76042026
0
GetSampleRate takes a key and returns the appropriate sample rate for that key.
func (t *WindowedThroughput) GetSampleRate(key string) int { return t.GetSampleRateMulti(key, 1) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *Sampler) GetSampleRate(trace pb.Trace, root *pb.Span, signature Signature) float64 {\n\treturn s.loadRate(s.GetSignatureSampleRate(signature) * s.extraRate)\n}", "func (c *Context) GetSampleRate() (rate int) {\n\treturn int(C.rtlsdr_get_sample_rate((*C.rtlsdr_dev_t)(c.dev)))\n}", "func (cfg Config) GetSampleRate() (sampleRate float64) {\n\treturn cfg.SampleRate\n}", "func (t *WindowedThroughput) GetSampleRateMulti(key string, count int) int {\n\tt.requestCount++\n\tt.eventCount += int64(count)\n\n\t// Insert the new key into the map.\n\tcurrent := t.indexGenerator.GetCurrentIndex()\n\terr := t.countList.IncrementKey(key, current, count)\n\n\t// We've reached MaxKeys, return 0.\n\tif err != nil {\n\t\treturn 0\n\t}\n\n\tt.lock.Lock()\n\tdefer t.lock.Unlock()\n\tif rate, found := t.savedSampleRates[key]; found {\n\t\treturn rate\n\t}\n\treturn 0\n}", "func (f *Format) SampleRate() freq.T {\n\treturn f.freq\n}", "func (t *Sampler) SampleRate() Rate {\n\treturn t.sampleRate\n}", "func (a *AudioData) SampleRate() int {\n\treturn a.format.SampleRate\n}", "func (audio *Audio) SampleRate() int {\n\treturn audio.sampleRate\n}", "func (e *Engine) SampleRate() int { return e.backend.SampleRate() }", "func (s *Sink) Samplerate() uint {\n\treturn s.samplerate\n}", "func (s *Source) Samplerate() (n uint) {\n\ts.ifOpen(func() {\n\t\tn = uint(C.aubio_source_get_samplerate(s.s))\n\t})\n\treturn\n}", "func (info TrackInfo) SampleRate() int {\n\t// bits 23-0 - audio sample rate\n\treturn int(info & 0xFFFFFF)\n}", "func (t Track) SampleRate() int {\n\tpanic(\"implement me\")\n}", "func (s *maxEPSSampler) getSampleRate() float64 {\n\trate := 1.0\n\tcurrentEPS := s.rateCounter.GetRate()\n\tif currentEPS > s.maxEPS {\n\t\trate = s.maxEPS / currentEPS\n\t}\n\treturn rate\n}", "func (o BackendServiceLogConfigOutput) SampleRate() pulumi.Float64PtrOutput {\n\treturn o.ApplyT(func(v BackendServiceLogConfig) *float64 { return v.SampleRate }).(pulumi.Float64PtrOutput)\n}", "func (o BackendServiceLogConfigResponseOutput) SampleRate() pulumi.Float64Output {\n\treturn o.ApplyT(func(v BackendServiceLogConfigResponse) float64 { return v.SampleRate }).(pulumi.Float64Output)\n}", "func (pipe *pipe) SampleRate() uint {\n\treturn pipe.samplesPerSecond\n}", "func (c *Context) SetSampleRate(rate int) (err int) {\n\treturn int(C.rtlsdr_set_sample_rate((*C.rtlsdr_dev_t)(c.dev),\n\t\tC.uint32_t(rate)))\n}", "func (m *Model) GetModelSampleRate() int {\n\treturn int(C.GetModelSampleRate(m.w))\n}", "func (r *RecordStream) SampleRate() int {\n\treturn int(r.createReply.Rate)\n}", "func (cfg *Config) SetSampleRate(sampleRate float64) {\n\tcfg.SampleRate = sampleRate\n}", "func (r *rateSampler) Rate() float64 {\n\tr.RLock()\n\tdefer r.RUnlock()\n\treturn r.rate\n}", "func (o BackendServiceLogConfigPtrOutput) SampleRate() pulumi.Float64PtrOutput {\n\treturn o.ApplyT(func(v *BackendServiceLogConfig) *float64 {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.SampleRate\n\t}).(pulumi.Float64PtrOutput)\n}", "func (s *samplerBackendRateCounter) GetRate() float64 {\n\treturn s.backend.getUpperSampledScore()\n}", "func RecordSampleRate(rate int) RecordOption {\n\treturn func(r *RecordStream) {\n\t\tr.createRequest.Rate = uint32(rate)\n\t}\n}", "func (this *fileStruct) SampleRate() uint32 {\n\treturn this.sampleRate\n}", "func (o BackendServiceLogConfigResponsePtrOutput) SampleRate() pulumi.Float64PtrOutput {\n\treturn o.ApplyT(func(v *BackendServiceLogConfigResponse) *float64 {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.SampleRate\n\t}).(pulumi.Float64PtrOutput)\n}", "func GetGenlockSampleRateI3D(hDC unsafe.Pointer, uRate unsafe.Pointer) unsafe.Pointer {\n\tret, _, _ := syscall.Syscall(gpGetGenlockSampleRateI3D, 2, uintptr(hDC), uintptr(uRate), 0)\n\treturn (unsafe.Pointer)(ret)\n}", "func (rl *limiter) GetRate() int64 {\n\trl.lock.RLock()\n\tqps := rl.qps\n\trl.lock.RUnlock()\n\n\treturn qps\n}", "func extractSampleRate(data []byte) uint32 {\n\t// Skip 40 bits (28 of the OGG header + 12 of the Opus header) and read the\n\t// next 4 bits to extract the sample rate as an int.\n\tif len(data) < 44 {\n\t\tlog.Panicf(\"unexpected data length: %v\", len(data))\n\t}\n\n\treturn binary.LittleEndian.Uint32(data[40:44])\n}", "func (d *Detector) SetSampleRate(x int) error {\n\terrno := C.fvad_set_sample_rate(d.fvad, C.int(x))\n\tif errno != 0 {\n\t\treturn fmt.Errorf(\"invalid sample rate: %v\", x)\n\t}\n\treturn nil\n}", "func (ss *Rate) GetSingle(key string, nowTs time.Time) (float64, float64) {\n\tnow := nowTs.UnixNano()\n\tvar bucket *bucket\n\tif bucketno, found := ss.keytobucketno[key]; found {\n\t\tbucket = &ss.buckets[bucketno]\n\t\trate := ss.recount(bucket.rate, bucket.lastTs, now)\n\t\terrRate := ss.recount(bucket.errRate, bucket.errLastTs, now)\n\t\treturn rate - errRate, rate\n\t} else {\n\t\tbucketno = uint32(ss.sh.h[0])\n\t\tbucket = &ss.buckets[bucketno]\n\t\terrRate := ss.recount(bucket.rate, bucket.lastTs, now)\n\t\treturn 0, errRate\n\t}\n\n}", "func (p *Pump) WavSampleRate() phono.SampleRate {\n\treturn p.wavSampleRate\n}", "func (r *rateSampler) Sample(spn ddtrace.Span) bool {\n\tif r.rate == 1 {\n\t\t// fast path\n\t\treturn true\n\t}\n\ts, ok := spn.(*span)\n\tif !ok {\n\t\treturn false\n\t}\n\tr.RLock()\n\tdefer r.RUnlock()\n\treturn sampledByRate(s.TraceID, r.rate)\n}", "func sampledByRate(n uint64, rate float64) bool {\n\tif rate < 1 {\n\t\treturn n*knuthFactor < uint64(rate*math.MaxUint64)\n\t}\n\treturn true\n}", "func NewRateSampler(rate float64) RateSampler {\n\treturn &rateSampler{rate: rate}\n}", "func (o *AudioStreamSample) GetMixRate() gdnative.Int {\n\t//log.Println(\"Calling AudioStreamSample.GetMixRate()\")\n\n\t// Build out the method's arguments\n\tptrArguments := make([]gdnative.Pointer, 0, 0)\n\n\t// Get the method bind\n\tmethodBind := gdnative.NewMethodBind(\"AudioStreamSample\", \"get_mix_rate\")\n\n\t// Call the parent method.\n\t// int\n\tretPtr := gdnative.NewEmptyInt()\n\tgdnative.MethodBindPtrCall(methodBind, o.GetBaseObject(), ptrArguments, retPtr)\n\n\t// If we have a return type, convert it from a pointer into its actual object.\n\tret := gdnative.NewIntFromPointer(retPtr)\n\treturn ret\n}", "func (s *AudioParameters) SetSampleRate(v string) *AudioParameters {\n\ts.SampleRate = &v\n\treturn s\n}", "func (ps *prioritySampler) getRate(spn *span) float64 {\n\tkey := \"service:\" + spn.Service + \",env:\" + spn.Meta[ext.Environment]\n\tps.mu.RLock()\n\tdefer ps.mu.RUnlock()\n\tif rate, ok := ps.rates[key]; ok {\n\t\treturn rate\n\t}\n\treturn ps.defaultRate\n}", "func NewSampler(sampleRate Rate) (*Sampler, error) {\n\tif err := sampleRate.Validate(); err != nil {\n\t\treturn nil, err\n\t}\n\tif sampleRate == 0 {\n\t\treturn &Sampler{\n\t\t\tsampleRate: sampleRate,\n\t\t\tsampleEvery: 0,\n\t\t}, nil\n\t}\n\treturn &Sampler{\n\t\tsampleRate: sampleRate,\n\t\tnumTried: atomic.NewInt32(0),\n\t\tsampleEvery: int32(1.0 / sampleRate),\n\t}, nil\n}", "func (e *Engine) SetSampleRate(sr float64) error {\n\tif !e.initialized {\n\t\treturn errorEngineNotInitialized\n\t}\n\t// update the stream parameters\n\te.streamParameters.SampleRate = sr\n\treturn nil\n}", "func (m *MetricsProvider) SignerGetKey(value time.Duration) {\n}", "func (s *RateBasedRule) SetRateKey(v string) *RateBasedRule {\n\ts.RateKey = &v\n\treturn s\n}", "func (s *CreateRateBasedRuleInput) SetRateKey(v string) *CreateRateBasedRuleInput {\n\ts.RateKey = &v\n\treturn s\n}", "func ParseSamplingRate(s string) (float64, error) {\n\tv, err := ParseFloat64(s)\n\tif err != nil {\n\t\treturn 0.0, err\n\t}\n\tif v < 0.0 {\n\t\treturn -1.0 / v, nil\n\t}\n\treturn v, nil\n}", "func (client *Client) isSendAcceptedBySampleRate(sampleRate float32) bool {\n\tif sampleRate >= 1 {\n\t\treturn true\n\t}\n\trandomNumber := client.rand.Float32()\n\treturn randomNumber <= sampleRate\n}", "func (r *rateSampler) SetRate(rate float64) {\n\tr.Lock()\n\tr.rate = rate\n\tr.Unlock()\n}", "func (rc *RateCalculator) GetRate(name string, attributes map[string]interface{}, val float64, now time.Time) (gauge telemetry.Gauge, valid bool) {\n\tvar attributesJSON []byte\n\tif nil != attributes {\n\t\tattributesJSON = internal.MarshalOrderedAttributes(attributes)\n\t}\n\trc.lock.Lock()\n\tdefer rc.lock.Unlock()\n\n\tid := internal.MetricIdentity{Name: name, AttributesJSON: string(attributesJSON)}\n\n\tlast, found := rc.datapoints[id]\n\tif found {\n\t\t// don't accept timestamps older that the last one for this metric\n\t\tif last.When.Before(now) {\n\t\t\telapsedSeconds := now.Sub(last.When).Seconds()\n\t\t\trate := val / elapsedSeconds\n\n\t\t\tgauge.Name = name\n\t\t\tgauge.Timestamp = now\n\t\t\tgauge.Value = rate\n\t\t\tgauge.Attributes = attributes\n\t\t\tgauge.AttributesJSON = attributesJSON\n\n\t\t\tvalid = true\n\t\t}\n\t} else {\n\t\trc.datapoints[id] = internal.LastValue{When: now}\n\t}\n\n\treturn\n}", "func (client *Client) Timing(key string, time int64, sampleRate float32) {\n\tmetricValue := fmt.Sprintf(\"%d|%s\", time, metricTypeTiming)\n\tif sampleRate < 1 {\n\t\tif client.isSendAcceptedBySampleRate(sampleRate) {\n\t\t\tmetricValue = fmt.Sprintf(\"%s|@%g\", metricValue, sampleRate)\n\t\t} else {\n\t\t\treturn\n\t\t}\n\t}\n\n\tclient.addToBuffer(key, metricValue)\n}", "func (rt *Rates) Get() (rateMap map[string][]float64) {\n\trt.mu.Lock()\n\tdefer rt.mu.Unlock()\n\n\trateMap = make(map[string][]float64)\n\ttimeStamps := rt.timeStamps.Values()\n\tif len(timeStamps) <= 1 {\n\t\treturn\n\t}\n\tfor k, v := range rt.counts {\n\t\trateMap[k] = make([]float64, len(timeStamps)-1)\n\t\tvalues := v.Values()\n\t\tvalueIndex := len(values) - 1\n\t\tfor i := len(timeStamps) - 1; i > 0; i-- {\n\t\t\tif valueIndex <= 0 {\n\t\t\t\trateMap[k][i-1] = 0\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\telapsed := float64((timeStamps[i] - timeStamps[i-1]) / 1e9)\n\t\t\trateMap[k][i-1] = float64(values[valueIndex]-values[valueIndex-1]) / elapsed\n\t\t\tvalueIndex--\n\t\t}\n\t}\n\treturn\n}", "func GetRate(c *gin.Context) {\n\t// get symbol\n\tsymbol := c.Param(\"symbol\")\n\n\tethrate, err := models.FindEthrateBySymbol(symbol, nil)\n\tif err != nil {\n\t\tfmt.Println(err.Error())\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"msg\": \"Data ethrate find error\", \"error\": err.Error()})\n\t\treturn\n\t}\n\tc.JSON(http.StatusOK, ethrate)\n\treturn\n}", "func (audio *Audio) BitRate() int {\n\treturn audio.bitRate\n}", "func Samplerate(s float64) Option {\n\treturn func(args *Options) {\n\t\targs.Samplerate = s\n\t}\n}", "func (tracer Tracer) IsSampled() bool {\n\tif tracer.SampleRate == 100 {\n\t\treturn true\n\t} else if tracer.SampleRate == 0 {\n\t\treturn false\n\t}\n\n\treturn 100-tracer.SampleRate < rand.Intn(100)\n}", "func newAudioSampler(clockRate uint32, latency time.Duration) samplerFunc {\n\tsamples := uint32(math.Round(float64(clockRate) * latency.Seconds()))\n\treturn samplerFunc(func() uint32 {\n\t\treturn samples\n\t})\n}", "func GenlockSampleRateI3D(hDC unsafe.Pointer, uRate unsafe.Pointer) unsafe.Pointer {\n\tret, _, _ := syscall.Syscall(gpGenlockSampleRateI3D, 2, uintptr(hDC), uintptr(uRate), 0)\n\treturn (unsafe.Pointer)(ret)\n}", "func (info TrackInfo) BitRate() int {\n\t// bit 29 - 1 = 16-bit audio; 0 = 8-bit audio\n\tif info&0x20000000 != 0 {\n\t\t// 16-bit audio.\n\t\treturn 16\n\t}\n\t// 8-bit audio.\n\treturn 8\n}", "func (c *CurrencyServer) GetRate(ctx context.Context, rr *currency.RateRequest) (*currency.RateResponse, error) {\n\tc.log.Info(\"Handle GetRate\", \"base\", rr.GetBase(), \"destination\", rr.GetDestination())\n\n\trate, err := c.rates.GetRate(rr.GetBase().String(), rr.GetDestination().String())\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &currency.RateResponse{Rate: rate}, nil\n}", "func (o *Job) GetRate(ctx context.Context) (rate uint64, err error) {\n\terr = o.object.CallWithContext(ctx, \"org.freedesktop.DBus.Properties.Get\", 0, InterfaceJob, \"Rate\").Store(&rate)\n\treturn\n}", "func (c *Currency) GetRate(ctx context.Context, rr *protos.RateRequest) (*protos.RateResponse, error) {\n\tc.log.Println(\"Handle request for GetRate\", \"base\", rr.GetBase(), \"dest\", rr.GetDestination())\n\treturn &protos.RateResponse{Rate: 0.5}, nil\n}", "func (p *ParkMotorcycle) GetRate(timestamp int64) float64 {\n\treturn 0.5 * float64(ceilUnixTimeHourDiff(p.Timestamp, timestamp))\n}", "func (params *KeyParameters) Sample() (*big.Int, error) {\n\t// Choose a random exponent in [0,Q-1).\n\tR, err := rand.Int(rand.Reader, params.qMinusOne)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// Add 1 so that the exponent is in [1,Q-1].\n\tR.Add(R, params.one)\n\treturn R, nil\n}", "func (t *TimeSeries) GetKey() string {\n\treturn t.key\n}", "func (c Client) Rate(pair string) (float64, error) {\n\tpair = strings.TrimSpace(strings.ToUpper(pair))\n\n\tres, err := c.Live([]string{pair})\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\tif rate, ok := res.Rates[pair]; ok {\n\t\treturn rate.Rate, nil\n\t}\n\n\treturn 0, fmt.Errorf(\"rate for pair '%s' not found\", pair)\n}", "func (s *StatsdClient) Get(key string) int64 {\n\ts.lock.RLock()\n\tdefer s.lock.RUnlock()\n\treturn s.counts[key]\n}", "func rate(read int64, t time.Duration) float64 {\n\treturn float64(read) / (1024 * 1024) / t.Seconds()\n}", "func ResampleKey(t *big.Int, params *Params, precomputed *PreparedAttributeList, key *PrivateKey, delegable bool) (*PrivateKey, error) {\n\tresampled := &PrivateKey{\n\t\tA0: new(bn256.G1),\n\t\tA1: new(bn256.G2),\n\t\tBSig: new(bn256.G1),\n\t}\n\n\t// Randomly choose t in Zp\n\tif t == nil {\n\t\tvar err error\n\t\tt, err = RandomInZp(rand.Reader)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tresampled.A0.ScalarMult((*bn256.G1)(precomputed), t)\n\tresampled.A0.Add(resampled.A0, key.A0)\n\n\tresampled.A1.ScalarMult(params.G, t)\n\tresampled.A1.Add(resampled.A1, key.A1)\n\n\tif params.HSig != nil && key.BSig != nil {\n\t\tresampled.BSig.ScalarMult(params.HSig, t)\n\t\tresampled.BSig.Add(resampled.BSig, key.BSig)\n\t}\n\n\tif delegable {\n\t\tresampled.B = make([]*bn256.G1, len(key.B), cap(key.B))\n\t\tfor attrIndex, bIndex := range key.FreeMap {\n\t\t\th := new(bn256.G1).ScalarMult(params.H[attrIndex], t)\n\t\t\th.Add(h, key.B[bIndex])\n\t\t\tresampled.B[bIndex] = h\n\t\t}\n\t\tresampled.FreeMap = key.FreeMap\n\t}\n\n\treturn resampled, nil\n}", "func (o *TfaCreateMessageRequest) GetSpeechRate() float64 {\n\tif o == nil || o.SpeechRate == nil {\n\t\tvar ret float64\n\t\treturn ret\n\t}\n\treturn *o.SpeechRate\n}", "func (r *RateCounter) Rate() int64 {\n\treturn r.counter.Value()\n}", "func (s *Sampler) GetMaxTPSSampleRate() float64 {\n\t// When above maxTPS, apply an additional sample rate to statistically respect the limit\n\tmaxTPSrate := 1.0\n\tif s.maxTPS > 0 {\n\t\tcurrentTPS := s.Backend.GetUpperSampledScore()\n\t\tif currentTPS > s.maxTPS {\n\t\t\tmaxTPSrate = s.maxTPS / currentTPS\n\t\t}\n\t}\n\n\treturn maxTPSrate\n}", "func (k Keeper) GetRat(ctx sdk.Context, key string) sdk.Rat {\n\tr, err := k.paramsKeeper.Getter().GetRat(ctx, MakeFeeKey(key))\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn r\n}", "func (p *ParkCar) GetRate(timestamp int64) float64 {\n\treturn 1 * float64(ceilUnixTimeHourDiff(p.Timestamp, timestamp))\n}", "func (a *NamespacesApiService) GetSubscribeRate(ctx _context.Context, tenant string, namespace string) (SubscribeRate, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodGet\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue SubscribeRate\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/namespaces/{tenant}/{namespace}/subscribeRate\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"tenant\"+\"}\", _neturl.QueryEscape(fmt.Sprintf(\"%v\", tenant)), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"namespace\"+\"}\", _neturl.QueryEscape(fmt.Sprintf(\"%v\", namespace)), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 200 {\n\t\t\tvar v SubscribeRate\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func (d *DeviceProp) ClockRate() int32 {\n\treturn (int32)(d.clockRate)\n}", "func Get(key Type) time.Duration {\n\tif GlobalStats != nil {\n\t\tmutex.RLock()\n\t\tdefer mutex.RUnlock()\n\t\treturn GlobalStats[key]\n\t}\n\treturn 0\n}", "func MaxSamplingRate(r int) middleware.TraceOption {\n\treturn middleware.MaxSamplingRate(r)\n}", "func (a *Anim) PlayRate() float64 {\n\ta.access.RLock()\n\trate := a.playRate\n\ta.access.RUnlock()\n\treturn rate\n}", "func (o *GstObj) GetControlRate() ClockTime {\n\treturn ClockTime(C.gst_object_get_control_rate(o.g()))\n}", "func (_m *Knapsack) TraceSamplingRate() float64 {\n\tret := _m.Called()\n\n\tvar r0 float64\n\tif rf, ok := ret.Get(0).(func() float64); ok {\n\t\tr0 = rf()\n\t} else {\n\t\tr0 = ret.Get(0).(float64)\n\t}\n\n\treturn r0\n}", "func (r *Rate) Rate() decimal.Decimal {\n\treturn r.rate\n}", "func (r *Registers) GetSoundTimer() uint8 {\n\treturn r.soundTimer\n}", "func (a *Aggregator) Rate(timestamp int64) (Rate, error) {\n\ta.mu.RLock()\n\tdefer a.mu.RUnlock()\n\n\terr := errRateNotFound\n\tvar rate Rate\n\ta.data.Next().Do(func(value interface{}) {\n\t\tif value == nil {\n\t\t\treturn\n\t\t}\n\n\t\tif value.(Rate).Timestamp == timestamp {\n\t\t\trate = value.(Rate)\n\t\t\terr = nil\n\t\t}\n\t})\n\n\treturn rate, err\n}", "func (d *DatastoreSubscription) Get(key string) (*Subscription, error) {\n\tv, err := d.store.Get(d.prefix(key))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif v == nil {\n\t\treturn nil, ErrNotFoundEntry\n\t}\n\treturn decodeRawSubscription(v)\n}", "func (g *ShedderGroup) GetShedder(key string) Shedder {\n\tshedder, _ := g.manager.GetResource(key, func() (closer io.Closer, e error) {\n\t\treturn nopCloser{\n\t\t\tShedder: NewAdaptiveShedder(g.options...),\n\t\t}, nil\n\t})\n\treturn shedder.(Shedder)\n}", "func (a *AudioData) SampleType() SampleType {\n\treturn a.format.SampleType\n}", "func (c *Client) Rate(ctx context.Context, p map[uint32][]string) (err error) {\n\t_, err = c.RateEndpoint(ctx, p)\n\treturn\n}", "func (this *Stats) SendBitRate() float32 { return float32(this.ptr.f_send_bitrate) }", "func (l *GoLimiter) getRateLimiter(keyName string) *rate.Limiter {\n\tl.mu.Lock()\n\tdefer l.mu.Unlock()\n\n\t_, exists := l.visitors[keyName]\n\tif !exists {\n\t\tburst, _ := strconv.Atoi(os.Getenv(environment.BurstLimit))\n\t\tlimit, _ := strconv.Atoi(os.Getenv(environment.RateLimit))\n\t\tlimiter := rate.NewLimiter(rate.Limit(limit), burst)\n\t\tl.visitors[keyName] = &visitor{limiter, time.Now()}\n\t} else {\n\t\tl.visitors[keyName].lastSeen = time.Now()\n\t}\n\n\treturn l.visitors[keyName].limiter\n}", "func (c *Context) GetDirectSampling() (err int) {\n\treturn int(C.rtlsdr_get_direct_sampling((*C.rtlsdr_dev_t)(c.dev)))\n}", "func (a *AudioData) SampleSize() int {\n\treturn a.format.SampleType.Size()\n}", "func (r AKC695X) GetFreq() (kHz uint32) {\n\treturn r.channelToFreq(r.getChannel())\n}", "func (self Source) GetPitch() float32 {\n\treturn self.Getf(AlPitch)\n}", "func getTimecodeRate(t *testing.T, data testdata.TimecodeData) rate.Framerate {\n\tassert := assert.New(t)\n\tntsc := rate.NTSCNone\n\tif data.Ntsc {\n\t\tntsc = rate.NTSCNonDrop\n\t}\n\tif data.DropFrame {\n\t\tntsc = rate.NTSCDrop\n\t}\n\n\tframerate, err := rate.FromInt(data.Timebase, ntsc)\n\tif !assert.NoError(err, \"parse timebase to framerate\") {\n\t\tt.FailNow()\n\t}\n\n\treturn framerate\n}", "func (o *RatesRateResponse) GetRate() int32 {\n\tif o == nil || o.Rate == nil {\n\t\tvar ret int32\n\t\treturn ret\n\t}\n\treturn *o.Rate\n}", "func (this *Stats) InputBitRate() float32 { return float32(this.ptr.f_input_bitrate) }", "func (v Bytes) Rate(seconds float64) string {\n\treturn Bytes(float64(v)/seconds).String() + BYTES_PER_SECOND\n}", "func (c *Chrome) PlaybackRate(rate int) error {\n\tmsg := wsMessage{\n\t\tComponent: \"player\",\n\t\tMethod: \"playbackRate\",\n\t\tArguments: map[string]string{\n\t\t\t\"rate\": strconv.Itoa(rate),\n\t\t},\n\t}\n\n\tsend := c.ConnViewer.getChanSend()\n\tsend <- msg\n\tres := <-c.ConnViewer.getChanReceive()\n\treturn handleRes(res)\n}", "func GetRateBySegment(config *common.Configuration, client *http.Client) {\n\n\tfmt.Println(\"By Segment...\")\n\n\tjsonData := map[string]string{\"schemeId\": config.SchemeId, \"segmentId\": config.SegmentId}\n\tjsonValue, _ := json.Marshal(jsonData)\n\turl := config.ApiUrl + config.RatesUri\n\n\trequest, _ := http.NewRequest(\"POST\", url, bytes.NewBuffer(jsonValue))\n\trequest.Header.Set(\"Content-Type\", \"application/json\")\n\n\tresponse, err := client.Do(request)\n\tif err != nil {\n\t\tfmt.Printf(\"The HTTP request failed with error %s\\n\", err)\n\t} else {\n\t\tdata, _ := ioutil.ReadAll(response.Body)\n\t\tfmt.Println(string(data))\n\t}\n}", "func (tx *Transaction) Sampled() bool {\n\treturn tx.sampled\n}", "func (bucket *Bucket) Rate() float64 {\n\treturn 1e9 / float64(bucket.fillInterval)\n}" ]
[ "0.71399146", "0.7063044", "0.6862884", "0.6834468", "0.6378315", "0.630154", "0.6243745", "0.61906356", "0.60597485", "0.60458696", "0.6014118", "0.6005109", "0.592752", "0.591679", "0.5897877", "0.5785057", "0.5769777", "0.5725354", "0.56908005", "0.5650925", "0.5606956", "0.56064403", "0.56036866", "0.5572537", "0.54828495", "0.54680914", "0.540256", "0.53315455", "0.53237456", "0.5304022", "0.52788407", "0.52770716", "0.5229017", "0.522517", "0.5214259", "0.517575", "0.50661224", "0.50605476", "0.5060197", "0.50412536", "0.49987", "0.49478516", "0.48811838", "0.48552766", "0.48540297", "0.47668973", "0.47346178", "0.4702257", "0.46814206", "0.46656373", "0.46369618", "0.4624614", "0.461521", "0.46021667", "0.45806646", "0.45609066", "0.4549688", "0.45148966", "0.44838658", "0.44818336", "0.44770724", "0.4473369", "0.4452362", "0.44389364", "0.44313365", "0.44206342", "0.4412398", "0.44104862", "0.4401473", "0.43994212", "0.4387548", "0.43857244", "0.43821546", "0.43729413", "0.4364032", "0.4357434", "0.4354892", "0.43369377", "0.4336065", "0.4333614", "0.43239865", "0.43078035", "0.429523", "0.42886826", "0.4280994", "0.42796573", "0.42713878", "0.42629704", "0.42613497", "0.42364737", "0.420119", "0.41933402", "0.41920325", "0.41901046", "0.4185119", "0.41728058", "0.416425", "0.4158769", "0.41579276", "0.4145637" ]
0.82063824
0
GetSampleRateMulti takes a key representing count spans and returns the appropriate sample rate for that key.
func (t *WindowedThroughput) GetSampleRateMulti(key string, count int) int { t.requestCount++ t.eventCount += int64(count) // Insert the new key into the map. current := t.indexGenerator.GetCurrentIndex() err := t.countList.IncrementKey(key, current, count) // We've reached MaxKeys, return 0. if err != nil { return 0 } t.lock.Lock() defer t.lock.Unlock() if rate, found := t.savedSampleRates[key]; found { return rate } return 0 }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (t *WindowedThroughput) GetSampleRate(key string) int {\n\treturn t.GetSampleRateMulti(key, 1)\n}", "func (s *Sampler) GetSampleRate(trace pb.Trace, root *pb.Span, signature Signature) float64 {\n\treturn s.loadRate(s.GetSignatureSampleRate(signature) * s.extraRate)\n}", "func (ss *Rate) GetSingle(key string, nowTs time.Time) (float64, float64) {\n\tnow := nowTs.UnixNano()\n\tvar bucket *bucket\n\tif bucketno, found := ss.keytobucketno[key]; found {\n\t\tbucket = &ss.buckets[bucketno]\n\t\trate := ss.recount(bucket.rate, bucket.lastTs, now)\n\t\terrRate := ss.recount(bucket.errRate, bucket.errLastTs, now)\n\t\treturn rate - errRate, rate\n\t} else {\n\t\tbucketno = uint32(ss.sh.h[0])\n\t\tbucket = &ss.buckets[bucketno]\n\t\terrRate := ss.recount(bucket.rate, bucket.lastTs, now)\n\t\treturn 0, errRate\n\t}\n\n}", "func (s *Source) Samplerate() (n uint) {\n\ts.ifOpen(func() {\n\t\tn = uint(C.aubio_source_get_samplerate(s.s))\n\t})\n\treturn\n}", "func (s *Sink) Samplerate() uint {\n\treturn s.samplerate\n}", "func (f *Format) SampleRate() freq.T {\n\treturn f.freq\n}", "func sampledByRate(n uint64, rate float64) bool {\n\tif rate < 1 {\n\t\treturn n*knuthFactor < uint64(rate*math.MaxUint64)\n\t}\n\treturn true\n}", "func (info TrackInfo) SampleRate() int {\n\t// bits 23-0 - audio sample rate\n\treturn int(info & 0xFFFFFF)\n}", "func (t Track) SampleRate() int {\n\tpanic(\"implement me\")\n}", "func (t *Sampler) SampleRate() Rate {\n\treturn t.sampleRate\n}", "func (c *Context) GetSampleRate() (rate int) {\n\treturn int(C.rtlsdr_get_sample_rate((*C.rtlsdr_dev_t)(c.dev)))\n}", "func (o BackendServiceLogConfigResponseOutput) SampleRate() pulumi.Float64Output {\n\treturn o.ApplyT(func(v BackendServiceLogConfigResponse) float64 { return v.SampleRate }).(pulumi.Float64Output)\n}", "func (rt *Rates) Get() (rateMap map[string][]float64) {\n\trt.mu.Lock()\n\tdefer rt.mu.Unlock()\n\n\trateMap = make(map[string][]float64)\n\ttimeStamps := rt.timeStamps.Values()\n\tif len(timeStamps) <= 1 {\n\t\treturn\n\t}\n\tfor k, v := range rt.counts {\n\t\trateMap[k] = make([]float64, len(timeStamps)-1)\n\t\tvalues := v.Values()\n\t\tvalueIndex := len(values) - 1\n\t\tfor i := len(timeStamps) - 1; i > 0; i-- {\n\t\t\tif valueIndex <= 0 {\n\t\t\t\trateMap[k][i-1] = 0\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\telapsed := float64((timeStamps[i] - timeStamps[i-1]) / 1e9)\n\t\t\trateMap[k][i-1] = float64(values[valueIndex]-values[valueIndex-1]) / elapsed\n\t\t\tvalueIndex--\n\t\t}\n\t}\n\treturn\n}", "func NewRateSampler(rate float64) RateSampler {\n\treturn &rateSampler{rate: rate}\n}", "func (a *AudioData) SampleRate() int {\n\treturn a.format.SampleRate\n}", "func (s *maxEPSSampler) getSampleRate() float64 {\n\trate := 1.0\n\tcurrentEPS := s.rateCounter.GetRate()\n\tif currentEPS > s.maxEPS {\n\t\trate = s.maxEPS / currentEPS\n\t}\n\treturn rate\n}", "func (e *Engine) SampleRate() int { return e.backend.SampleRate() }", "func (cfg Config) GetSampleRate() (sampleRate float64) {\n\treturn cfg.SampleRate\n}", "func Samplerate(s float64) Option {\n\treturn func(args *Options) {\n\t\targs.Samplerate = s\n\t}\n}", "func NewSampler(sampleRate Rate) (*Sampler, error) {\n\tif err := sampleRate.Validate(); err != nil {\n\t\treturn nil, err\n\t}\n\tif sampleRate == 0 {\n\t\treturn &Sampler{\n\t\t\tsampleRate: sampleRate,\n\t\t\tsampleEvery: 0,\n\t\t}, nil\n\t}\n\treturn &Sampler{\n\t\tsampleRate: sampleRate,\n\t\tnumTried: atomic.NewInt32(0),\n\t\tsampleEvery: int32(1.0 / sampleRate),\n\t}, nil\n}", "func (o BackendServiceLogConfigOutput) SampleRate() pulumi.Float64PtrOutput {\n\treturn o.ApplyT(func(v BackendServiceLogConfig) *float64 { return v.SampleRate }).(pulumi.Float64PtrOutput)\n}", "func (r *RecordStream) SampleRate() int {\n\treturn int(r.createReply.Rate)\n}", "func (audio *Audio) SampleRate() int {\n\treturn audio.sampleRate\n}", "func RecordSampleRate(rate int) RecordOption {\n\treturn func(r *RecordStream) {\n\t\tr.createRequest.Rate = uint32(rate)\n\t}\n}", "func GetGenlockSampleRateI3D(hDC unsafe.Pointer, uRate unsafe.Pointer) unsafe.Pointer {\n\tret, _, _ := syscall.Syscall(gpGetGenlockSampleRateI3D, 2, uintptr(hDC), uintptr(uRate), 0)\n\treturn (unsafe.Pointer)(ret)\n}", "func (client *Client) GetMultiRateConfig(request *GetMultiRateConfigRequest) (response *GetMultiRateConfigResponse, err error) {\n\tresponse = CreateGetMultiRateConfigResponse()\n\terr = client.DoAction(request, response)\n\treturn\n}", "func (pipe *pipe) SampleRate() uint {\n\treturn pipe.samplesPerSecond\n}", "func (r *rateSampler) Sample(spn ddtrace.Span) bool {\n\tif r.rate == 1 {\n\t\t// fast path\n\t\treturn true\n\t}\n\ts, ok := spn.(*span)\n\tif !ok {\n\t\treturn false\n\t}\n\tr.RLock()\n\tdefer r.RUnlock()\n\treturn sampledByRate(s.TraceID, r.rate)\n}", "func FrequencySetTimeKeyCount(fset FrequencySet) (statictimeseries.DataSeries, error) {\n\tds := statictimeseries.NewDataSeries()\n\tds.SeriesName = fset.Name\n\tfor rfc3339, fstats := range fset.FrequencyMap {\n\t\tdt, err := time.Parse(time.RFC3339, rfc3339)\n\t\tif err != nil {\n\t\t\treturn ds, err\n\t\t}\n\t\tds.AddItem(statictimeseries.DataItem{\n\t\t\tSeriesName: fset.Name,\n\t\t\tTime: dt,\n\t\t\tValue: int64(len(fstats.Items))})\n\t}\n\treturn ds, nil\n}", "func (ps *prioritySampler) getRate(spn *span) float64 {\n\tkey := \"service:\" + spn.Service + \",env:\" + spn.Meta[ext.Environment]\n\tps.mu.RLock()\n\tdefer ps.mu.RUnlock()\n\tif rate, ok := ps.rates[key]; ok {\n\t\treturn rate\n\t}\n\treturn ps.defaultRate\n}", "func (r *rateSampler) Rate() float64 {\n\tr.RLock()\n\tdefer r.RUnlock()\n\treturn r.rate\n}", "func newAudioSampler(clockRate uint32, latency time.Duration) samplerFunc {\n\tsamples := uint32(math.Round(float64(clockRate) * latency.Seconds()))\n\treturn samplerFunc(func() uint32 {\n\t\treturn samples\n\t})\n}", "func extractSampleRate(data []byte) uint32 {\n\t// Skip 40 bits (28 of the OGG header + 12 of the Opus header) and read the\n\t// next 4 bits to extract the sample rate as an int.\n\tif len(data) < 44 {\n\t\tlog.Panicf(\"unexpected data length: %v\", len(data))\n\t}\n\n\treturn binary.LittleEndian.Uint32(data[40:44])\n}", "func (s *samplerBackendRateCounter) GetRate() float64 {\n\treturn s.backend.getUpperSampledScore()\n}", "func (client *Client) GetMultiRateConfigWithCallback(request *GetMultiRateConfigRequest, callback func(response *GetMultiRateConfigResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *GetMultiRateConfigResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.GetMultiRateConfig(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (c *Client) GetMulti(ctx context.Context, k []*Key) ([]Record, error) {\n\tpbks, err := keys(k).proto()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresp, err := pb.NewStoreClient(c.ClientConn).Get(c.newContext(ctx), &pb.Keys{\n\t\tKeys: pbks,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdocs, err := pbRecords(resp.Records).records()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn docs, multiErrorFromRecordStatusProto(resp.Status)\n}", "func (m *MetricsProvider) SignerGetKey(value time.Duration) {\n}", "func ParseSamplingRate(s string) (float64, error) {\n\tv, err := ParseFloat64(s)\n\tif err != nil {\n\t\treturn 0.0, err\n\t}\n\tif v < 0.0 {\n\t\treturn -1.0 / v, nil\n\t}\n\treturn v, nil\n}", "func newSamplerBackendRateCounter() *samplerBackendRateCounter {\n\treturn &samplerBackendRateCounter{\n\t\tbackend: newMemoryBackend(),\n\t\texit: make(chan struct{}),\n\t\tstopped: make(chan struct{}),\n\t}\n}", "func (m *Model) GetModelSampleRate() int {\n\treturn int(C.GetModelSampleRate(m.w))\n}", "func (o BackendServiceLogConfigPtrOutput) SampleRate() pulumi.Float64PtrOutput {\n\treturn o.ApplyT(func(v *BackendServiceLogConfig) *float64 {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.SampleRate\n\t}).(pulumi.Float64PtrOutput)\n}", "func (o BackendServiceLogConfigResponsePtrOutput) SampleRate() pulumi.Float64PtrOutput {\n\treturn o.ApplyT(func(v *BackendServiceLogConfigResponse) *float64 {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.SampleRate\n\t}).(pulumi.Float64PtrOutput)\n}", "func (o *AudioStreamSample) GetMixRate() gdnative.Int {\n\t//log.Println(\"Calling AudioStreamSample.GetMixRate()\")\n\n\t// Build out the method's arguments\n\tptrArguments := make([]gdnative.Pointer, 0, 0)\n\n\t// Get the method bind\n\tmethodBind := gdnative.NewMethodBind(\"AudioStreamSample\", \"get_mix_rate\")\n\n\t// Call the parent method.\n\t// int\n\tretPtr := gdnative.NewEmptyInt()\n\tgdnative.MethodBindPtrCall(methodBind, o.GetBaseObject(), ptrArguments, retPtr)\n\n\t// If we have a return type, convert it from a pointer into its actual object.\n\tret := gdnative.NewIntFromPointer(retPtr)\n\treturn ret\n}", "func rateCounter(selRange time.Duration) func(samples []promql.Point) float64 {\n\treturn func(samples []promql.Point) float64 {\n\t\treturn extrapolatedRate(samples, selRange, true, true)\n\t}\n}", "func (c *Context) SetSampleRate(rate int) (err int) {\n\treturn int(C.rtlsdr_set_sample_rate((*C.rtlsdr_dev_t)(c.dev),\n\t\tC.uint32_t(rate)))\n}", "func GenlockSampleRateI3D(hDC unsafe.Pointer, uRate unsafe.Pointer) unsafe.Pointer {\n\tret, _, _ := syscall.Syscall(gpGenlockSampleRateI3D, 2, uintptr(hDC), uintptr(uRate), 0)\n\treturn (unsafe.Pointer)(ret)\n}", "func NewAllSampler() RateSampler { return NewRateSampler(1) }", "func (a *NamespacesApiService) GetSubscribeRate(ctx _context.Context, tenant string, namespace string) (SubscribeRate, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodGet\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue SubscribeRate\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/namespaces/{tenant}/{namespace}/subscribeRate\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"tenant\"+\"}\", _neturl.QueryEscape(fmt.Sprintf(\"%v\", tenant)), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"namespace\"+\"}\", _neturl.QueryEscape(fmt.Sprintf(\"%v\", namespace)), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 200 {\n\t\t\tvar v SubscribeRate\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func APISFamily(sampler Sampler, size int) ([]float64, []float64) {\n\tmus, sigmas := make([]float64, size), make([]float64, size)\n\tfor i := range mus {\n\t\tmus[i] = sampler.Sample()\n\t\tsigmas[i] = sampler.Sample()\n\t\tif sigmas[i] < 0 {\n\t\t\tsigmas[i] *= -1\n\t\t}\n\t}\n\treturn mus, sigmas\n}", "func getCounterSamples(name string,\n\tlabels []*prometheus_models.LabelPair,\n\ttimestampMs int64,\n\tc *prometheus_models.Counter,\n\tentity string,\n) []Sample {\n\tsamples := make([]Sample, 1)\n\tsamples[0] = Sample{\n\t\tname: name,\n\t\tlabels: labels,\n\t\ttimestampMs: timestampMs,\n\t\tvalue: strconv.FormatFloat(c.GetValue(), 'f', -1, 64),\n\t\tentity: entity,\n\t}\n\treturn samples\n}", "func (m KeyedSamples) Add(index int64, t time.Duration) {\n\tkey := fmt.Sprintf(\"%010d\", index) // Easiest way to have samples show up in sorted order in the json.\n\tms, found := m[key]\n\tif !found {\n\t\tms = new(Multisample)\n\t\tm[key] = ms\n\t}\n\tms.Add(t)\n}", "func samplesToBytesLPCM24(samples []float64) ([]byte, error) {\n\tconst delta = MAX_INT24 - MIN_INT24\n\tscale := 0.5 * float64(delta)\n\tbuf := createBuffer()\n\n\t/*\n\t * Iterate over the samples and convert them into integer representation.\n\t */\n\tfor _, sample := range samples {\n\n\t\t/*\n\t\t * Make sure that limits are not exceeded.\n\t\t */\n\t\tif sample < -1.0 {\n\t\t\tsample = -1.0\n\t\t} else if sample > 1.0 {\n\t\t\tsample = 1.0\n\t\t}\n\n\t\ttmp := int32(scale * sample)\n\n\t\t/*\n\t\t * Make sure that limits are not exceeded.\n\t\t */\n\t\tif tmp > MAX_INT24 {\n\t\t\ttmp = MAX_INT24\n\t\t} else if tmp < MIN_INT24 {\n\t\t\ttmp = MIN_INT24\n\t\t}\n\n\t\tsampleUint := uint32(tmp)\n\n\t\t/*\n\t\t * Write each byte to the buffer.\n\t\t */\n\t\tfor j := 0; j < SIZE_INT24; j++ {\n\t\t\tshift := BITS_PER_BYTE * uint32(j)\n\t\t\tbyt := byte((sampleUint >> shift) & 0xff)\n\t\t\tbuf.WriteByte(byt)\n\t\t}\n\n\t}\n\n\tdata := buf.Bytes()\n\treturn data, nil\n}", "func (this *fileStruct) SampleRate() uint32 {\n\treturn this.sampleRate\n}", "func (t *WindowedThroughput) updateMaps() {\n\tcurrentIndex := t.indexGenerator.GetCurrentIndex()\n\tlookbackIndexes := t.indexGenerator.DurationToIndexes(t.LookbackFrequencyDuration)\n\taggregateCounts := t.countList.AggregateCounts(currentIndex, lookbackIndexes)\n\n\t// Apply the same aggregation algorithm as total throughput\n\t// Short circuit if no traffic\n\tt.numKeys = len(aggregateCounts)\n\tif t.numKeys == 0 {\n\t\t// no traffic during the last period.\n\t\tt.lock.Lock()\n\t\tdefer t.lock.Unlock()\n\t\tt.savedSampleRates = make(map[string]int)\n\t\treturn\n\t}\n\t// figure out our target throughput per key over the lookback window.\n\ttotalGoalThroughput := t.GoalThroughputPerSec * t.LookbackFrequencyDuration.Seconds()\n\t// floor the throughput but min should be 1 event per bucket per time period\n\tthroughputPerKey := math.Max(1, float64(totalGoalThroughput)/float64(t.numKeys))\n\t// for each key, calculate sample rate by dividing counted events by the\n\t// desired number of events\n\tnewSavedSampleRates := make(map[string]int)\n\tfor k, v := range aggregateCounts {\n\t\trate := int(math.Max(1, (float64(v) / float64(throughputPerKey))))\n\t\tnewSavedSampleRates[k] = rate\n\t}\n\t// save newly calculated sample rates\n\tt.lock.Lock()\n\tdefer t.lock.Unlock()\n\tt.savedSampleRates = newSavedSampleRates\n}", "func rate(read int64, t time.Duration) float64 {\n\treturn float64(read) / (1024 * 1024) / t.Seconds()\n}", "func (s *StatsdClient) Get(key string) int64 {\n\ts.lock.RLock()\n\tdefer s.lock.RUnlock()\n\treturn s.counts[key]\n}", "func GetSamples() ([200]uint64, []byte, plotter.Values) {\n input, err := os.Open(\"/dev/TrueRNG\")\n if err != nil {\n panic(err)\n }\n defer input.Close()\n buffer := make([]byte, 256)\n n, err := input.Read(buffer)\n histogram, sum, count, samples := [200]uint64{}, 0, 0, make([]byte, 0, *Samples)\n v := make(plotter.Values, 0, *Samples)\nOuter:\n for err == nil {\n for _, b := range buffer[:n] {\n for i := 0; i < 8; i++ {\n if b & 1 == 1 {\n sum += 1\n }\n b >>= 1\n count++\n if count == 200 {\n histogram[sum]++\n v = append(v, float64(sum) - 100)\n samples = append(samples, byte(sum))\n if len(samples) == *Samples {\n break Outer\n }\n sum, count = 0, 0\n }\n }\n }\n n, err = input.Read(buffer)\n }\n\n return histogram, samples, v\n}", "func NewRates(name string, countTracker CountTracker, samples int, interval time.Duration) *Rates {\n\tif interval < 1*time.Second && interval != -1*time.Second {\n\t\tpanic(\"interval too small\")\n\t}\n\tctx, cancel := context.WithCancel(context.Background())\n\trt := &Rates{\n\t\ttimeStamps: NewRingInt64(samples + 1),\n\t\tcounts: make(map[string]*RingInt64),\n\t\tcountTracker: countTracker,\n\t\tsamples: samples + 1,\n\t\tinterval: interval,\n\t\ttimestampLastSampling: timeNow(),\n\t\tctx: ctx,\n\t\tcancel: cancel,\n\t}\n\tif name != \"\" {\n\t\tpublish(name, rt)\n\t}\n\tif interval > 0 {\n\t\tgo rt.track()\n\t}\n\treturn rt\n}", "func (c *Client) Rate(ctx context.Context, p map[uint32][]string) (err error) {\n\t_, err = c.RateEndpoint(ctx, p)\n\treturn\n}", "func (rl *limiter) GetRate() int64 {\n\trl.lock.RLock()\n\tqps := rl.qps\n\trl.lock.RUnlock()\n\n\treturn qps\n}", "func (r *rateSampler) SetRate(rate float64) {\n\tr.Lock()\n\tr.rate = rate\n\tr.Unlock()\n}", "func ResampleKey(t *big.Int, params *Params, precomputed *PreparedAttributeList, key *PrivateKey, delegable bool) (*PrivateKey, error) {\n\tresampled := &PrivateKey{\n\t\tA0: new(bn256.G1),\n\t\tA1: new(bn256.G2),\n\t\tBSig: new(bn256.G1),\n\t}\n\n\t// Randomly choose t in Zp\n\tif t == nil {\n\t\tvar err error\n\t\tt, err = RandomInZp(rand.Reader)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tresampled.A0.ScalarMult((*bn256.G1)(precomputed), t)\n\tresampled.A0.Add(resampled.A0, key.A0)\n\n\tresampled.A1.ScalarMult(params.G, t)\n\tresampled.A1.Add(resampled.A1, key.A1)\n\n\tif params.HSig != nil && key.BSig != nil {\n\t\tresampled.BSig.ScalarMult(params.HSig, t)\n\t\tresampled.BSig.Add(resampled.BSig, key.BSig)\n\t}\n\n\tif delegable {\n\t\tresampled.B = make([]*bn256.G1, len(key.B), cap(key.B))\n\t\tfor attrIndex, bIndex := range key.FreeMap {\n\t\t\th := new(bn256.G1).ScalarMult(params.H[attrIndex], t)\n\t\t\th.Add(h, key.B[bIndex])\n\t\t\tresampled.B[bIndex] = h\n\t\t}\n\t\tresampled.FreeMap = key.FreeMap\n\t}\n\n\treturn resampled, nil\n}", "func aggregatePerSecondHandler(perSecondChannel chan *HttpReqResult) {\n\n\tfor {\n\n\t\tvar totalReq int = 0\n\t\tvar totalLatency int = 0\n\t\tuntil := time.Now().UnixNano() + 1000000000\n\t\tfor time.Now().UnixNano() < until {\n\t\t\tselect {\n\t\t\tcase msg := <-perSecondChannel:\n\t\t\t\ttotalReq++\n\t\t\t\ttotalLatency += int(msg.Latency/1000) // measure in microseconds\n\t\t\tdefault:\n\t\t\t// Can be trouble. Uses too much CPU if low, limits throughput if too high\n\t\t\t\ttime.Sleep(100*time.Microsecond)\n\t\t\t}\n\t\t}\n\t\t// concurrently assemble the result and send it off to the websocket.\n\t\tgo assembleAndSendResult(totalReq, totalLatency)\n\t}\n\n}", "func (client *Client) Timing(key string, time int64, sampleRate float32) {\n\tmetricValue := fmt.Sprintf(\"%d|%s\", time, metricTypeTiming)\n\tif sampleRate < 1 {\n\t\tif client.isSendAcceptedBySampleRate(sampleRate) {\n\t\t\tmetricValue = fmt.Sprintf(\"%s|@%g\", metricValue, sampleRate)\n\t\t} else {\n\t\t\treturn\n\t\t}\n\t}\n\n\tclient.addToBuffer(key, metricValue)\n}", "func TestMultipleBuckets(t *testing.T) {\n\tdone := testutils.FreezeTime()\n\tdefer done()\n\n\tfr, err := NewRatioCounter(3, clock.Second)\n\trequire.NoError(t, err)\n\n\tfr.IncB(1)\n\tclock.Advance(clock.Second)\n\tfr.IncA(1)\n\n\tclock.Advance(clock.Second)\n\tfr.IncA(1)\n\n\tassert.Equal(t, true, fr.IsReady())\n\tassert.Equal(t, float64(2)/float64(3), fr.Ratio())\n}", "func (cfg *Config) SetSampleRate(sampleRate float64) {\n\tcfg.SampleRate = sampleRate\n}", "func tickspersecond() int64", "func RateKey_Values() []string {\n\treturn []string{\n\t\tRateKeyIp,\n\t}\n}", "func (c Client) Rate(pair string) (float64, error) {\n\tpair = strings.TrimSpace(strings.ToUpper(pair))\n\n\tres, err := c.Live([]string{pair})\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\tif rate, ok := res.Rates[pair]; ok {\n\t\treturn rate.Rate, nil\n\t}\n\n\treturn 0, fmt.Errorf(\"rate for pair '%s' not found\", pair)\n}", "func instrumentGet(inner func()) {\n\tTotalRequests.Add(1)\n\tPendingRequests.Add(1)\n\tdefer PendingRequests.Add(-1)\n\n\tstart := time.Now()\n\n\tinner()\n\n\t// Capture the histogram over 18 geometric buckets \n\tdelta := time.Since(start)\n\tswitch {\n\tcase delta < time.Millisecond:\n\t\tLatencies.Add(\"0ms\", 1)\n\tcase delta > 32768*time.Millisecond:\n\t\tLatencies.Add(\">32s\", 1)\n\tdefault:\n\t\tfor i := time.Millisecond; i < 32768*time.Millisecond; i *= 2 {\n\t\t\tif delta >= i && delta < i*2 {\n\t\t\t\tLatencies.Add(i.String(), 1)\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n}", "func (s Song) GetSamplesForTick(tickIndex int) (string, []string) {\n var trackNames []string\n var samplePaths []string\n\n for _, currentTrack := range s.TrackList {\n if currentTrack.ShouldSountAt(tickIndex) {\n trackNames = append(trackNames, currentTrack.Name)\n samplePaths = append(samplePaths, currentTrack.PathToSample)\n }\n }\n if len(trackNames) < 1 {\n return \"-\", nil\n }\n //Join the name of all the samples\n return strings.Join(trackNames, \"+\"), samplePaths\n}", "func (t *tracer) sample(span *span) {\n\tif _, ok := span.context.samplingPriority(); ok {\n\t\t// sampling decision was already made\n\t\treturn\n\t}\n\tsampler := t.config.sampler\n\tif !sampler.Sample(span) {\n\t\tspan.context.trace.drop()\n\t\treturn\n\t}\n\tif rs, ok := sampler.(RateSampler); ok && rs.Rate() < 1 {\n\t\tspan.setMetric(sampleRateMetricKey, rs.Rate())\n\t}\n\tif t.rulesSampling.SampleTrace(span) {\n\t\treturn\n\t}\n\tt.prioritySampling.apply(span)\n}", "func GetRateByCurrencyPair(config *common.Configuration, client *http.Client) {\n\n\tfmt.Println(\"Getting rates for USDCAD...\")\n\n\tjsonData := map[string]string{\"schemeId\": config.SchemeId, \"segmentId\": config.SegmentId, \"schemeBuyCurrency\": \"USD\", \"schemeSellCurrency\": \"CAD\"}\n\tjsonValue, _ := json.Marshal(jsonData)\n\turl := config.ApiUrl + config.RatesUri\n\n\trequest, _ := http.NewRequest(\"POST\", url, bytes.NewBuffer(jsonValue))\n\trequest.Header.Set(\"Content-Type\", \"application/json\")\n\n\tresponse, err := client.Do(request)\n\tif err != nil {\n\t\tfmt.Printf(\"The HTTP request failed with error %s\\n\", err)\n\t} else {\n\t\tdata, _ := ioutil.ReadAll(response.Body)\n\t\tfmt.Println(string(data))\n\t}\n}", "func (params *KeyParameters) Sample() (*big.Int, error) {\n\t// Choose a random exponent in [0,Q-1).\n\tR, err := rand.Int(rand.Reader, params.qMinusOne)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// Add 1 so that the exponent is in [1,Q-1].\n\tR.Add(R, params.one)\n\treturn R, nil\n}", "func Format(bytesPerSecond float64) (value, unit string) {\n\t// to make it bits per second\n\tnumPerSecond := bytesPerSecond * 8\n\ti := 0\n\tfor ; i < 3; i++ {\n\t\tif numPerSecond < 1024 {\n\t\t\tbreak\n\t\t}\n\t\tnumPerSecond /= 1024\n\t}\n\treturn fmt.Sprintf(\"%.2f\", numPerSecond), units[i]\n}", "func genSineProfile(samplePeriod float64, frequency float64, ampScale float64) ([]int32) {\n\tvar (\n\t\tsamplesPerSecond\tfloat64 = 1 / float64(samplePeriod) \n\t\tphase \t\t\t\tfloat64\n\t\tradiansPerSample \tfloat64 = float64(frequency * 2 * math.Pi / float64(samplesPerSecond))\n\t\t//Number of samples just enough to generate one period\n\t\tnumberOfSamples\t\tuint32 = uint32(2 * math.Pi / float64(radiansPerSample))\n\t\twaveform\t\t\t[]int32 = make([]int32, numberOfSamples)\n\t)\n\n\tfor sample := uint32(0); sample < numberOfSamples; sample++ {\n\t\tsampleValue := float64(ampScale) * 0.5 * float64( 1 + math.Sin(phase) + 0.1)//0.1 is needed to avoid a divide by zero in boomer\n\t\twaveform[sample] = int32(sampleValue)\n\t\tphase += radiansPerSample\n\t}\n\n\treturn waveform\n}", "func downmixSingleSampleToMono(values ...int16) int16 {\n\tvar sum int = 0\n\tfor _, v := range values {\n\t\tsum += int(v)\n\t}\n\treturn int16(sum / len(values))\n}", "func getMegabitsPerSecond(data *speedTestData) float64 {\n\treturn getMegabits(data) / getSeconds(data)\n}", "func GetRateBySegment(config *common.Configuration, client *http.Client) {\n\n\tfmt.Println(\"By Segment...\")\n\n\tjsonData := map[string]string{\"schemeId\": config.SchemeId, \"segmentId\": config.SegmentId}\n\tjsonValue, _ := json.Marshal(jsonData)\n\turl := config.ApiUrl + config.RatesUri\n\n\trequest, _ := http.NewRequest(\"POST\", url, bytes.NewBuffer(jsonValue))\n\trequest.Header.Set(\"Content-Type\", \"application/json\")\n\n\tresponse, err := client.Do(request)\n\tif err != nil {\n\t\tfmt.Printf(\"The HTTP request failed with error %s\\n\", err)\n\t} else {\n\t\tdata, _ := ioutil.ReadAll(response.Body)\n\t\tfmt.Println(string(data))\n\t}\n}", "func RenderbufferStorageMultisample(target uint32, samples int32, internalformat uint32, width int32, height int32) {\n C.glowRenderbufferStorageMultisample(gpRenderbufferStorageMultisample, (C.GLenum)(target), (C.GLsizei)(samples), (C.GLenum)(internalformat), (C.GLsizei)(width), (C.GLsizei)(height))\n}", "func bytesToSamplesLPCM24(data []byte) ([]float64, error) {\n\tnumBytes := len(data)\n\tnumBytes64 := uint64(numBytes)\n\tnumSamples := numBytes64 / SIZE_INT24\n\tsamplesFloat := make([]float64, numSamples)\n\tscaling := 2.0 / (MAX_INT24 - MIN_INT24)\n\treader := bytes.NewReader(data)\n\tbuf := make([]byte, SIZE_INT24)\n\twords := make([]uint32, SIZE_INT24)\n\n\t/*\n\t * Read samples from input stream.\n\t */\n\tfor idx := range samplesFloat {\n\t\treader.Read(buf)\n\n\t\t/*\n\t\t * Turn the single bytes from the buffer into machine words.\n\t\t */\n\t\tfor i, byt := range buf {\n\t\t\twords[i] = uint32(byt)\n\t\t}\n\n\t\tsampleWord := uint32(0)\n\n\t\t/*\n\t\t * Combine the extracted words into a single machine word.\n\t\t */\n\t\tfor i, word := range words {\n\t\t\tshift := BITS_PER_BYTE * uint32(i)\n\t\t\tsampleWord |= word << shift\n\t\t}\n\n\t\tsampleInt := int32(sampleWord)\n\t\tsignBit := (sampleWord & SIGN_BIT_INT24) != 0\n\n\t\t/*\n\t\t * Handle negative values in two's complement representation.\n\t\t */\n\t\tif signBit {\n\t\t\toffset := sampleInt & MAX_INT24\n\t\t\tsampleInt = MIN_INT24 + offset\n\t\t}\n\n\t\tsamplesFloat[idx] = scaling * float64(sampleInt)\n\t}\n\n\treturn samplesFloat, nil\n}", "func getkey(ciphers []string) []int{\n\tN := len(ciphers[len(ciphers)-1])\n\tvar keys []int\n\tfor i := 0; i < N; i+=2 {\n\t\tvar sample []int\n\t\tfor j := 0; j < len(ciphers); j++ {\n\t\t\tsample = append(sample,formSample(ciphers[j][i:i+2]))\n\t\t}\n\t\tkey := findKey(sample)\n\t\tkeys = append(keys,key)\n\t}\n\treturn keys\n}", "func sampleResponseTimes(xreq, yreq *Request, s *SampleSettings) (\n\txs, ys []time.Duration, err error) {\n\txs = make([]time.Duration, s.SampleSize)\n\tys = make([]time.Duration, s.SampleSize)\n\ttransport := &http.Transport{\n\t\tDial: (&net.Dialer{\n\t\t\tTimeout: s.RequestTimeout,\n\t\t}).Dial}\n\n\t// first reqs will be outliers due to e.g., TCP handshake. discard them\n\tfor i := 0; i < s.NThrowaways; i++ {\n\t\tif _, err = sampleResponseTime(transport, xreq); err != nil {\n\t\t\treturn nil, nil, err\n\t\t}\n\n\t\tif _, err = sampleResponseTime(transport, yreq); err != nil {\n\t\t\treturn nil, nil, err\n\t\t}\n\t}\n\n\tfor i := 0; i < s.SampleSize; i++ {\n\t\tif xs[i], err = sampleResponseTime(transport, xreq); err != nil {\n\t\t\treturn nil, nil, err\n\t\t}\n\n\t\tif ys[i], err = sampleResponseTime(transport, yreq); err != nil {\n\t\t\treturn nil, nil, err\n\t\t}\n\t}\n\n\treturn xs, ys, nil\n}", "func TestGetFrequencyFromData(t *testing.T) {\n\tconst (\n\t\ttolerance = 10\n\t)\n\n\t// Test parameters\n\tvar (\n\t\tsampleRates = []int{16000, 44100, 48000}\n\t\tsamples = 1000\n\t\tamplitudes = []float64{1, 256, 1000, 32768, 2147483647}\n\n\t\tfreqMin = 100\n\t\tfreqMax = 500\n\t\tfreqStep = 5\n\n\t\tshiftsPerFreq = 10\n\t)\n\n\tdata := make([]float64, samples)\n\tfor _, sampleRate := range sampleRates {\n\t\tfor _, amplitude := range amplitudes {\n\t\t\tfor freq := freqMin; freq <= freqMax; freq += freqStep {\n\t\t\t\tfor shift := 0; shift < shiftsPerFreq; shift++ {\n\t\t\t\t\ttShift := float64(sampleRate) / float64(freq) * float64(shift) / float64(shiftsPerFreq)\n\t\t\t\t\t// Generate sine wave\n\t\t\t\t\tfor t := 0; t < samples; t++ {\n\t\t\t\t\t\tdata[t] = float64(amplitude) * math.Sin(2*math.Pi*float64(freq)*(float64(t)+tShift)/float64(sampleRate))\n\t\t\t\t\t}\n\n\t\t\t\t\t// Analyse frequency\n\t\t\t\t\tres := GetFrequencyFromData(data, float64(sampleRate))\n\t\t\t\t\tif math.Abs(float64(freq)-res) > tolerance {\n\t\t\t\t\t\tt.Fatalf(\n\t\t\t\t\t\t\t\"Result differs too much. sampleRate:%d amplitude:%f freq:%d, tShift:%f - got:%.2f, diff:%.2f\",\n\t\t\t\t\t\t\tsampleRate,\n\t\t\t\t\t\t\tamplitude,\n\t\t\t\t\t\t\tfreq,\n\t\t\t\t\t\t\ttShift,\n\t\t\t\t\t\t\tres,\n\t\t\t\t\t\t\tmath.Abs(float64(freq)-res),\n\t\t\t\t\t\t)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n}", "func Get(key Type) time.Duration {\n\tif GlobalStats != nil {\n\t\tmutex.RLock()\n\t\tdefer mutex.RUnlock()\n\t\treturn GlobalStats[key]\n\t}\n\treturn 0\n}", "func (r *InterestRateCalculator) SinglePeriod(value float64) float64 {\n\treturn value * (1 + r.Rate)\n}", "func (s *StatsdClient) Count(name string, value int64, tags []string, rate float64) error {\n\ts.lock.Lock()\n\tdefer s.lock.Unlock()\n\n\tif len(tags) == 0 {\n\t\ts.counts[name] += value\n\t}\n\n\tfor _, tag := range tags {\n\t\ts.counts[name+\":\"+tag] += value\n\t}\n\treturn nil\n}", "func (p *Pump) WavSampleRate() phono.SampleRate {\n\treturn p.wavSampleRate\n}", "func (info TrackInfo) BitRate() int {\n\t// bit 29 - 1 = 16-bit audio; 0 = 8-bit audio\n\tif info&0x20000000 != 0 {\n\t\t// 16-bit audio.\n\t\treturn 16\n\t}\n\t// 8-bit audio.\n\treturn 8\n}", "func (s *EdgeModelStat) SetSamplingDeviceCount(v int64) *EdgeModelStat {\n\ts.SamplingDeviceCount = &v\n\treturn s\n}", "func (c SpanContext) ExtendedSamplingState(key interface{}, initValue func() interface{}) interface{} {\n\treturn c.samplingState.extendedStateForKey(key, initValue)\n}", "func BenchmarkSingleKey_GetSetMultiGet(b *testing.B) {\n\tvar numGets = []int{\n\t\t2,\n\t\t4,\n\t\t8,\n\t}\n\n\tfor _, s := range itemSizes {\n\n\t\tfor _, ng := range numGets {\n\t\t\tb.Run(fmt.Sprintf(\"FreeCache/keySize=%d/reads=%d\", s, ng), func(b *testing.B) {\n\t\t\t\tvar v []byte\n\t\t\t\tfc := freecache.NewCache(s * 1024)\n\t\t\t\tk := []byte(\"key\")\n\n\t\t\t\tb.ResetTimer()\n\n\t\t\t\tfor r := 0; r < b.N; r++ {\n\t\t\t\t\tb.StopTimer()\n\t\t\t\t\tv = randBytes(s)\n\t\t\t\t\tb.StartTimer()\n\n\t\t\t\t\t_, _ = fc.Get(k)\n\t\t\t\t\t_ = fc.Set(k, v, -1)\n\t\t\t\t\tfor i := 0; i < ng; i++ {\n\t\t\t\t\t\t_, _ = fc.Get(k)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t})\n\n\t\t\tb.Run(fmt.Sprintf(\"MangoStandard/keySize=%d/reads=%d\", s, ng), func(b *testing.B) {\n\t\t\t\tvar v []byte\n\t\t\t\tgbc := cache.New(cache.WithMaximumSize(s + 10))\n\t\t\t\tk := \"key\"\n\n\t\t\t\tb.ResetTimer()\n\n\t\t\t\tfor r := 0; r < b.N; r++ {\n\t\t\t\t\tb.StopTimer()\n\t\t\t\t\tv = randBytes(s)\n\t\t\t\t\tb.StartTimer()\n\n\t\t\t\t\tgbc.GetIfPresent(k)\n\t\t\t\t\tgbc.Put(k, v)\n\t\t\t\t\tfor i := 0; i < ng; i++ {\n\t\t\t\t\t\t_, _ = gbc.GetIfPresent(k)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t})\n\n\t\t\tb.Run(fmt.Sprintf(\"MangoLoading/keySize=%d/reads=%d\", s, ng), func(b *testing.B) {\n\t\t\t\tb.SkipNow()\n\t\t\t\tvar v []byte\n\t\t\t\tgbc := cache.NewLoadingCache(func(k cache.Key) (cache.Value, error) { return randBytes(s), nil }, cache.WithMaximumSize(s+10))\n\t\t\t\tk := \"key\"\n\n\t\t\t\tb.ResetTimer()\n\n\t\t\t\tfor r := 0; r < b.N; r++ {\n\t\t\t\t\tb.StopTimer()\n\t\t\t\t\tv = randBytes(s)\n\t\t\t\t\tb.StartTimer()\n\n\t\t\t\t\t_, _ = gbc.Get(k)\n\t\t\t\t\tgbc.Put(k, v)\n\t\t\t\t\tfor i := 0; i < ng; i++ {\n\t\t\t\t\t\t_, _ = gbc.Get(k)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t})\n\n\t\t}\n\t}\n}", "func (m AggregationRecord) GetSamples() []map[string]interface{} {\n\treturn m.Samples.list\n}", "func getSeconds(data *speedTestData) float64 {\n\treturn float64(data.Milliseconds) / 1000\n}", "func (s *Stats) Samples(category string) []time.Duration {\n\ts.access.RLock()\n\tdefer s.access.RUnlock()\n\n\tsamples, ok := s.categories[category]\n\tif !ok {\n\t\treturn nil\n\t}\n\n\tvar samplesCopy []time.Duration\n\tcopy(samplesCopy, samples)\n\treturn samplesCopy\n}", "func (d *Detector) SetSampleRate(x int) error {\n\terrno := C.fvad_set_sample_rate(d.fvad, C.int(x))\n\tif errno != 0 {\n\t\treturn fmt.Errorf(\"invalid sample rate: %v\", x)\n\t}\n\treturn nil\n}", "func (s *samplerBackendRateCounter) Count() {\n\ts.backend.countSample()\n}", "func GetMultisamplefv(pname uint32, index uint32, val *float32) {\n\tsyscall.Syscall(gpGetMultisamplefv, 3, uintptr(pname), uintptr(index), uintptr(unsafe.Pointer(val)))\n}", "func (db *memorydb) MultiGet(key ...[]byte) ([][]byte, error) {\n\tdb.sm.RLock()\n\tdefer db.sm.RUnlock()\n\n\tvalues := [][]byte{}\n\tfor _, k := range key {\n\t\tif value, ok := db.db[string(k)]; ok {\n\t\t\tvalues = append(values, value)\n\t\t} else {\n\t\t\treturn nil, nil\n\t\t}\n\t}\n\treturn values, nil\n}", "func MaxSamplingRate(r int) middleware.TraceOption {\n\treturn middleware.MaxSamplingRate(r)\n}" ]
[ "0.70987207", "0.58055186", "0.52645755", "0.5133258", "0.50206465", "0.50165087", "0.49694243", "0.49418998", "0.4895843", "0.48757643", "0.4787806", "0.478047", "0.4758367", "0.47408044", "0.4729433", "0.47280362", "0.46968856", "0.4675771", "0.46757147", "0.46589676", "0.46544516", "0.46431902", "0.4625479", "0.46078733", "0.4599792", "0.4570054", "0.45433483", "0.45264858", "0.45065242", "0.44525063", "0.44474566", "0.4415283", "0.44143826", "0.44117773", "0.43972632", "0.4385612", "0.4364399", "0.4361767", "0.43266967", "0.42791608", "0.4255085", "0.42542234", "0.42299196", "0.42197597", "0.42147028", "0.4211048", "0.41981107", "0.4188588", "0.4185948", "0.41449052", "0.41379288", "0.41319892", "0.4124787", "0.41157115", "0.4113673", "0.41053888", "0.40884504", "0.4084093", "0.40726423", "0.4053429", "0.40399492", "0.4026956", "0.40268654", "0.40257108", "0.4018695", "0.4012513", "0.39914802", "0.3966254", "0.39453942", "0.39403677", "0.39290687", "0.39227468", "0.3913705", "0.39131433", "0.3868141", "0.38670316", "0.38619712", "0.3859228", "0.38429528", "0.3841629", "0.38376844", "0.38316798", "0.38307452", "0.3827954", "0.38245195", "0.38207245", "0.38205415", "0.38172832", "0.38158965", "0.38150328", "0.38128436", "0.3812336", "0.38022113", "0.3801532", "0.37918335", "0.3785126", "0.37816253", "0.3778355", "0.37742543", "0.37681338" ]
0.82699066
0
SaveState is not implemented
func (t *WindowedThroughput) SaveState() ([]byte, error) { return nil, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (g *Gossiper) SaveState() {\n\tobj, e := json.MarshalIndent(g, \"\", \"\\t\")\n\tutils.HandleError(e)\n\t_ = os.Mkdir(utils.STATE_FOLDER, os.ModePerm)\n\tcwd, _ := os.Getwd()\n\te = ioutil.WriteFile(filepath.Join(cwd, utils.STATE_FOLDER, fmt.Sprint(g.Name, \".json\")), obj, 0644)\n\tutils.HandleError(e)\n}", "func (fb *FileBackend) save(state *storage.State) error {\n\tout, err := proto.Marshal(state)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to encode state: %w\", err)\n\t}\n\ttmp := fmt.Sprintf(fb.path+\".%v\", time.Now())\n\tif err := ioutil.WriteFile(tmp, out, 0600); err != nil {\n\t\treturn fmt.Errorf(\"failed to write state: %w\", err)\n\t}\n\terr = os.Rename(tmp, fb.path)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to move state: %w\", err)\n\t}\n\treturn nil\n}", "func (s *Store) SaveState() (err error) {\n\tif err = s.writePluginIDMap(); err != nil {\n\t\tslog.WithError(err).Error(\"can't write plugin id maps\")\n\t}\n\treturn\n}", "func (c *Action) SaveState(k, v string) {\n\tfmt.Fprintf(c.w, saveStateFmt, k, escapeData(v))\n}", "func (m *stateManager) Save() (err error) {\n\terr = m.storage.Write(m.state)\n\n\tif err == nil {\n\t\tm.stateChanged = false\n\t\tm.stateLoaded = true\n\t}\n\n\treturn\n}", "func (board *Board) saveState(undoInfo UndoInfo) {\n\tboard.gamePly++\n\tboard.undoInfoList[board.gamePly] = undoInfo\n}", "func (app *service) Save(state State) error {\n\tjs, err := app.adapter.ToJSON(state)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tchainHash := state.Chain()\n\tindex := state.Height()\n\tpath := filePath(chainHash, index)\n\treturn app.fileService.Save(path, js)\n}", "func (service *HTTPRestService) saveState() error {\n\tlog.Printf(\"[Azure CNS] saveState\")\n\n\t// Skip if a store is not provided.\n\tif service.store == nil {\n\t\tlog.Printf(\"[Azure CNS] store not initialized.\")\n\t\treturn nil\n\t}\n\n\t// Update time stamp.\n\tservice.state.TimeStamp = time.Now()\n\terr := service.store.Write(storeKey, &service.state)\n\tif err == nil {\n\t\tlog.Printf(\"[Azure CNS] State saved successfully.\\n\")\n\t} else {\n\t\tlog.Errorf(\"[Azure CNS] Failed to save state., err:%v\\n\", err)\n\t}\n\n\treturn err\n}", "func (s *ServiceState) save() {\n\tlog.Lvl3(\"Saving service\")\n\tb, err := network.Marshal(s.Storage)\n\tif err != nil {\n\t\tlog.Error(\"Couldn't marshal service:\", err)\n\t} else {\n\t\terr = ioutil.WriteFile(s.path+\"/prifi.bin\", b, 0660)\n\t\tif err != nil {\n\t\t\tlog.Error(\"Couldn't save file:\", err)\n\t\t}\n\t}\n}", "func SaveState(backupFile string, noteAll *[]*notes.Note) {\n\tdat, err := json.Marshal(noteAll)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\terr = ioutil.WriteFile(backupFile, dat, 0644)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n}", "func (r GopassRepo) saveState(payload []byte) error {\n\tif err := r.prepare(); err != nil {\n\t\treturn err\n\t}\n\tcmd := exec.Command(\"gopass\", \"insert\", \"-f\", r.config.Repo.State)\n\tcmd.Stdout = os.Stdout\n\tcmd.Stderr = os.Stderr\n\tstdin, err := cmd.StdinPipe()\n\tif nil != err {\n\t\treturn fmt.Errorf(\"obtaining stdin: %s\", err.Error())\n\t}\n\tpayload = append(payload, []byte(\"\\n\")...)\n\tencoder := base64.NewEncoder(base64.StdEncoding, stdin)\n\tif _, err := encoder.Write(payload); nil != err {\n\t\treturn fmt.Errorf(\"encoding failed with %s\", err)\n\t}\n\n\tif err := cmd.Start(); nil != err {\n\t\treturn fmt.Errorf(\"gopass failed with %s\", err)\n\t}\n\tstdin.Close()\n\tencoder.Close()\n\tif err := cmd.Wait(); nil != err {\n\t\treturn fmt.Errorf(\"gopass failed with %s\", err)\n\t}\n\n\treturn nil\n}", "func (d *Database) Save() error {\n\tb, err := json.Marshal(d.State)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Save\n\tif err := ioutil.WriteFile(d.FilePath, b, os.ModePerm); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func Save() error {\n\treturn nil\n}", "func (n *mockAgent) save() (s persistapi.AgentState) {\n\treturn\n}", "func (f *File) Save(fileName string, offset int64, xestatus string) error {\n\tif f.file == nil {\n\t\treturn errors.New(\"state file not open\")\n\t}\n\n\terr := writeState(f.file, fileName, offset, xestatus)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"writeStatus\")\n\t}\n\n\treturn nil\n}", "func saveState(ca *cauth.Ca) {\n\terr := ca.SavePrivateKey()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\terr = ca.SaveCertificate()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func Save() {\n\tdata := Savedata{\n\t\tName: GS.current.name,\n\t\tGamestate: GS.current,\n\t}\n\n\tf, err := json.MarshalIndent(data, \"\", \" \")\n\tcheck(err)\n\tioutil.WriteFile(\"data/savegame.json\", f, 0644)\n}", "func (m *Motifini) saveSubDB() {\n\tif err := m.Subs.StateFileSave(); err != nil {\n\t\tm.Error.Printf(\"saving subscribers state file: %v\", err)\n\t\treturn\n\t}\n\tm.Debug.Print(\"Saved state DB file\")\n}", "func (d *Deck) SaveState() {\n\td.saveStateToBucket(phrasesBucket)\n}", "func (s store) Save() {\n\ts.writeToDisk()\n}", "func (manager *basicStateManager) ForceSave() error {\n\tmanager.savingLock.Lock()\n\tdefer manager.savingLock.Unlock()\n\tseelog.Info(\"Saving state!\")\n\ts := manager.state\n\ts.Version = ECSDataVersion\n\n\tdata, err := json.Marshal(s)\n\tif err != nil {\n\t\tseelog.Error(\"Error saving state; could not marshal data; this is odd\", \"err\", err)\n\t\treturn err\n\t}\n\treturn manager.writeFile(data)\n}", "func (m *Machine) Save() error {\n\tm.State = driver.Saved\n\tfmt.Printf(\"Save %s: %s\\n\", m.Name, m.State)\n\treturn nil\n}", "func (aws *ActionWithState) Save() error {\n\tdb, err := sqlconn.GetConnection()\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer db.Close()\n\tupdate, err := db.Query(\"\" +\n\t\t\"UPDATE `action` \" +\n\t\t\"\tSET `state` = ?, \" +\n\t\t\"\t\t`processed` = ? \" +\n\t\t\"\tWHERE `id` = ?\", aws.FSM.Current(), aws.Processed, aws.id)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer update.Close()\n\treturn nil\n}", "func (s *State) Save() error {\n\tf, err := os.Create(getPathForUsername(s.Username))\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer f.Close()\n\n\treturn json.NewEncoder(f).Encode(s)\n}", "func (s *State) Save(db XODB) error {\n\tif s.Exists() {\n\t\treturn s.Update(db)\n\t}\n\n\treturn s.Insert(db)\n}", "func Save() error {\n\treturn instance.Save()\n}", "func (rs *RedisStore) Save(sid SessionID, state interface{}) error {\n\tj, err := json.Marshal(state)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = rs.Client.Set(sid.getRedisKey(), j, rs.SessionDuration).Err()\n\n\treturn err\n}", "func (defaultStorage) Save() error {\n\tpanic(noConfigStorage)\n}", "func (r *RadioStation) Save() error {\n\treturn nil\n}", "func (m Db) Save(n int, p architecture.Version) {\n\tm.Dbm[n] = p\n}", "func (spy *StoreSpy) Save(r *http.Request, w http.ResponseWriter, s *sessions.Session) error {\n\tspy.SaveCalls++\n\treturn nil\n}", "func cmdSave() {\n\tswitch state := status(B2D.VM); state {\n\tcase vmUnregistered:\n\t\tlog.Fatalf(\"%s is not registered.\", B2D.VM)\n\tcase vmRunning:\n\t\tlog.Printf(\"Suspending %s\", B2D.VM)\n\t\tif err := vbm(\"controlvm\", B2D.VM, \"savestate\"); err != nil {\n\t\t\tlog.Fatalf(\"failed to suspend vm: %s\", err)\n\t\t}\n\tdefault:\n\t\tlog.Printf(\"%s is not running.\", B2D.VM)\n\t}\n}", "func (bh browserHistory) Save() error {\n\tbytes, err := json.Marshal(bh.Records)\n\tif err != nil {\n\t\treturn err\n\t}\n\tjs.Global().Get(\"localStorage\").Set(\"history\", string(bytes))\n\n\treturn nil\n}", "func (s *Store) Save() error {\n\tbk, err := os.OpenFile(filepath.Join(s.rwDirPath, storeBkName), os.O_RDWR|os.O_CREATE, 0666)\n\tif err != nil {\n\t\treturn errors.Wrap(err, err.Error())\n\t}\n\tdefer bk.Close()\n\n\tdst, err := os.OpenFile(filepath.Join(s.rwDirPath, storeName), os.O_RDWR|os.O_CREATE, 0666)\n\tif err != nil {\n\t\treturn errors.Wrap(err, err.Error())\n\t}\n\tdefer dst.Close()\n\n\t// backing up current store\n\t_, err = io.Copy(bk, dst)\n\tif err != nil {\n\t\treturn errors.Wrap(err, err.Error())\n\t}\n\n\tenc := gob.NewEncoder(dst)\n\tbook := s.Clone()\n\terr = enc.Encode(book)\n\tif err != nil {\n\t\treturn errors.Wrap(err, err.Error())\n\t}\n\treturn nil\n}", "func (t *StoreCC) save(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n\n\tif len(args) != 2 {\n\t\treturn shim.Error(\"Incorrect number of arguments. Expecting name of the person to query\")\n\t}\n\n\tkey := args[0]\n\tvalue := args[1]\n\terr := stub.PutState(key, []byte(value))\n\tif err != nil {\n\t\treturn shim.Error(\"get failed \")\n\t}\n\tvar rsp saveRsp\n\trsp.Code = 0\n\trsp.TxID = stub.GetTxID()\n\tb, _ := json.Marshal(rsp)\n\n\treturn shim.Success(b)\n}", "func SaveLastRunState(lastRunState LastRunDownServers) {\n\t// write new last run state\n\tb, err := json.MarshalIndent(lastRunState, \"\", \" \")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\terr = ioutil.WriteFile(\"status.json\", b, 0644)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func Save(willSave bool, popSaved uint, delSaved uint) {\n\tos.Exit(0)\n}", "func (rf *Raft) persist() {\n // Your code here (2C).\n // Example:\n // w := new(bytes.Buffer)\n // e := gob.NewEncoder(w)\n // e.Encode(rf.xxx)\n // e.Encode(rf.yyy)\n // data := w.Bytes()\n // rf.persister.SaveRaftState(data)\n}", "func (rs *RedisStore) Save(sid SessionID, sessionState interface{}) error {\n\tlog.Printf(\"!!! I am about to save something to the redis db\")\n\n\tj, err := json.Marshal(sessionState)\n\tif err != nil {\n\t\tlog.Printf(\"!!! There was an error in saving\" + err.Error())\n\t\treturn err\n\t}\n\t//rs.Client.Set(sid.getRedisKey(), j, rs.SessionDuration)\n\trs.Client.Set(sid.getRedisKey(), j, time.Hour)\n\t//JAY MADE THE ABOVE CHANGE TO ATTEMPT A HARDCODED DEBUG FIX\n\treturn nil\n}", "func (c *Container) save() error {\n\tif err := c.runtime.state.SaveContainer(c); err != nil {\n\t\treturn errors.Wrapf(err, \"error saving container %s state\", c.ID())\n\t}\n\treturn nil\n}", "func (s *saver) Save(journey models.Journey) error {\n\ts.Lock()\n\tdefer s.Unlock()\n\n\tif s.state == closed {\n\t\treturn ErrSaverIsClosed\n\t}\n\n\tif len(s.buffer) == cap(s.buffer) {\n\t\treturn ErrInternalBufferIsFull\n\t}\n\n\ts.buffer = append(s.buffer, journey)\n\treturn nil\n}", "func (manager *basicStateManager) Save() error {\n\tdefer metrics.MetricsEngineGlobal.RecordStateManagerMetric(\"SAVE\")()\n\tmanager.saveTimesLock.Lock()\n\tdefer manager.saveTimesLock.Unlock()\n\tif time.Since(manager.lastSave) >= minSaveInterval {\n\t\t// we can just save\n\t\terr := manager.ForceSave()\n\t\tmanager.lastSave = time.Now()\n\t\tmanager.nextPlannedSave = time.Time{} // re-zero it; assume all pending desires to save are fulfilled\n\t\treturn err\n\t} else if manager.nextPlannedSave.IsZero() {\n\t\t// No save planned yet, we should plan one.\n\t\tnext := manager.lastSave.Add(minSaveInterval)\n\t\tmanager.nextPlannedSave = next\n\t\tgo func() {\n\t\t\ttime.Sleep(time.Until(next))\n\t\t\tmanager.Save()\n\t\t}()\n\t}\n\t// else nextPlannedSave wasn't Zero so there's a save planned elsewhere that'll\n\t// fulfill this\n\treturn nil\n}", "func (rf *Raft) persist() {\n // Your code here (2C).\n // Example:\n // w := new(bytes.Buffer)\n // e := labgob.NewEncoder(w)\n // e.Encode(rf.xxx)\n // e.Encode(rf.yyy)\n // data := w.Bytes()\n // rf.persister.SaveRaftState(data)\n}", "func persistStatus() {\n\n}", "func (rf *Raft) persistState() {\n\t// Your code here (2C).\n\t// Example:\n\tstate := rf.serializeState()\n\trf.persister.SaveRaftState(state)\n}", "func savingFile(states AllStates, ID string) {\n\tfile, err := os.Create(\"elevator_states.txt\") //Creates file that will only contain latest data\n\t//checks for errors and saves to file as JSON\n\tcheck(err)\n\te := json.NewEncoder(file).Encode(states) //saves the AllStates struct to file\n\tcheck(e)\n}", "func (s LocalBackend) WriteState(st *State) error {\n\tlog.Debugf(\"Writing state to %s\\n\", s.Path)\n\tdata, err := json.MarshalIndent(st, \"\", \" \")\n\tif err != nil {\n\t\tlog.Fatal(\"Failed to Load State for Writing\")\n\t}\n\terr = ioutil.WriteFile(s.Path, data, 0644)\n\tif err != nil {\n\t\tlog.Fatal(\"Failed to write state to file\")\n\t}\n\treturn nil\n}", "func (ns *NodeStore) save() error {\r\n\tb, err := json.Marshal(ns)\r\n\tif err != nil {\r\n\t\treturn err\r\n\t}\r\n\terr = ioutil.WriteFile(\"nodestore.json\", b, 0660)\r\n\tif err != nil {\r\n\t\treturn err\r\n\t}\r\n\treturn nil\r\n}", "func (self *AccountManager) Save() {\n\n}", "func TestStateSaveLoad(t *testing.T) {\n\ttearDown, stateDB, state := setupTestCase(t)\n\tdefer tearDown(t)\n\t// nolint: vetshadow\n\tassert := assert.New(t)\n\n\tstate.LastBlockHeight++\n\tSaveState(stateDB, state)\n\n\tloadedState := LoadState(stateDB)\n\tassert.True(state.Equals(loadedState),\n\t\tfmt.Sprintf(\"expected state and its copy to be identical.\\ngot: %v\\nexpected: %v\\n\",\n\t\t\tloadedState, state))\n}", "func (t *SimpleChaincode) save_changes(stub shim.ChaincodeStubInterface, sc SalesContractObject) (bool, error) {\n\n\tbytes, err := json.Marshal(sc)\n\n\tif err != nil {\n\t\tfmt.Printf(\"SAVE_CHANGES: Error converting contract : %s\", err)\n\t\treturn false, errors.New(\"Error converting contract \")\n\t}\n\n\terr = stub.PutState(sc.Contractid, bytes)\n\n\tif err != nil {\n\t\tfmt.Printf(\"SAVE_CHANGES: Error storing contract : %s\", err)\n\t\treturn false, errors.New(\"Error storing contract\")\n\t}\n\treturn true, nil\n}", "func (deployment *Deployment) setStateAndSave(state common.DeploymentState) {\n\tdeployment.currentState = state\n\terr := deployment.store.SaveDeployment(deployment)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func (bot *LeagueAnnouncerBot) persist() error {\n\tlog.Debug(\"Writing bot state to disk\")\n\t// Write the current state of the bot to disk, to load for next time\n\tstorageBytes, err := bot.storage.store()\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = ioutil.WriteFile(BOT_STATE, storageBytes, 0600)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *GRPCClient) SaveStateData(ctx context.Context, store, key, etag string, data []byte) error {\n\tif store == \"\" {\n\t\treturn errors.New(\"nil store\")\n\t}\n\tif key == \"\" {\n\t\treturn errors.New(\"nil key\")\n\t}\n\n\treq := &State{\n\t\tStoreName: store,\n\t\tStates: []*StateItem{\n\t\t\t{\n\t\t\t\tKey: key,\n\t\t\t\tValue: data,\n\t\t\t\tEtag: etag,\n\t\t\t},\n\t\t},\n\t}\n\n\treturn c.SaveState(ctx, req)\n}", "func (s *Session) Save() error {\n\treturn s.store.provider.Save(s.ID, s.Data)\n}", "func (ws *WalletStore) Save() {\n\tvar buffer bytes.Buffer\n\tgob.Register(elliptic.P256())\n\tencoder := gob.NewEncoder(&buffer)\n\terr := encoder.Encode(ws.Wallets)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tfile := ws.Config.GetWalletStoreFile(ws.NodeID)\n\terr = ioutil.WriteFile(file, buffer.Bytes(), 0644)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func (this *AccountStore) Save(account *states.AccountState) error {\n\tkey := account.GetKey()\n\tbuff := new(bytes.Buffer)\n\terr := account.Serialize(buff)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = this.store.Put(key, buff.Bytes())\n\tif err != nil {\n\t\treturn err\n\n\t}\n\treturn nil\n}", "func (s *Surface) Save() {\n\ts.Ctx.Call(\"save\")\n}", "func (src *Source) StoreSavepoint() {\n\tsrc.savepointStack = append(src.savepointStack, &savepoint{\n\t\tcurrent: src.current,\n\t})\n}", "func (s Store) Save(param ParameterSet) error {\n\tbytes, err := json.MarshalIndent(param, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn s.backingStore.Save(ItemType, \"\", param.Name, bytes)\n}", "func (c *Component) Save(n string, i interface{}) {\n\tc.GM.DB.Save(n, i)\n}", "func (rf *Raft) persist() {\n\t// Your code here (2C).\n\t// Example:\n\tstate := &StateInfo{\n\t\tCurrentTerm: rf.currentTerm,\n\t\tCommitIndex: rf.commitIndex,\n\t\tLastApplied: rf.lastApplied,\n\t\tLogs: rf.log,\n\t}\n\tw := new(bytes.Buffer)\n\te := labgob.NewEncoder(w)\n\te.Encode(state)\n\tdata := w.Bytes()\n\trf.persister.SaveRaftState(data)\n\t//DPrintf(\"rf [me %v] save stateInfo: %#v\", rf.me, state)\n}", "func save(novel *Novel) {\n\t//if novel exist history\n\ttag := false\n\tfor index, historyNovel := range historyNovels {\n\t\tif historyNovel.Name == novel.Name {\n\t\t\thistoryNovels[index] = novel\n\t\t\ttag = true\n\t\t}\n\t}\n\tif !tag {\n\t\thistoryNovels = append(historyNovels, novel)\n\t}\n\tSaveHistory(historyNovels)\n\tfmt.Println(\"Save complete...\")\n}", "func (flags *Vault) save(log *logging.Logger) (bool, error) {\n\tchanges := 0\n\treturn (changes > 0), nil\n}", "func (g *GamePlayerHand) Save() error {\n\tif g.ID == 0 {\n\t\treturn g.insert()\n\t}\n\n\treturn g.update()\n}", "func (b *stateHistoryBackend) SaveEntry(entry *statestream.StreamEntry) error {\n\tb.flushPendingEntry()\n\tb.pendingEntry = entry\n\tif !b.initialSet {\n\t\tb.flushPendingEntry()\n\t}\n\n\treturn nil\n}", "func (rf *Raft) persist() {\n\t// Your code here (2C).\n\t// Example:\n\t// w := new(bytes.Buffer)\n\t// e := labgob.NewEncoder(w)\n\t// e.Encode(rf.xxx)\n\t// e.Encode(rf.yyy)\n\t// data := w.Bytes()\n\t// rf.persister.SaveRaftState(data)\n\tdata := rf.getPersistState()\n\trf.persister.SaveRaftState(data)\n\t//\trf.persister.SaveRaftState(data)\n}", "func (sess *Session) Save(w http.ResponseWriter) error {\n\treturn sess.store.Save(w, sess)\n}", "func (c *Checkpoint) PersistState(st JournalState) {\n\tc.save <- st\n}", "func save() {\n\tnaksuIniPath := getIniFilePath()\n\n\terr := cfg.SaveTo(naksuIniPath)\n\tif err != nil {\n\t\tlog.Error(\"%s save failed: %v\", naksuIniPath, err)\n\t}\n}", "func (h *Homework) save() (err error) {\n\tbuf, err := json.MarshalIndent(h, \"\", \"\\t\")\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = ioutil.WriteFile(h.path, buf, 0777)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (rf *Raft) persist() {\n\t// Your code here (2C).\n\t// Example:\n\t// w := new(bytes.Buffer)\n\t// e := labgob.NewEncoder(w)\n\t// e.Encode(rf.xxx)\n\t// e.Encode(rf.yyy)\n\t// data := w.Bytes()\n\t// rf.persister.SaveRaftState(data)\n\tdata := rf.getStateBytes()\n\tDPrintf(\"[persist] raft %d persist data=%v\", rf.me, data)\n\trf.persister.SaveRaftState(data)\n}", "func (rf *Raft) persist() {\r\n\t// Your code here.\r\n\t// Example:\r\n\t// w := new(bytes.Buffer)\r\n\t// e := gob.NewEncoder(w)\r\n\t// e.Encode(rf.xxx)\r\n\t// e.Encode(rf.yyy)\r\n\t// data := w.Bytes()\r\n\t// rf.persister.SaveRaftState(data)\r\n}", "func (b *BlockCreator) save() error {\n\treturn persist.SaveJSON(settingsMetadata, b.persist, filepath.Join(b.persistDir, settingsFile))\n}", "func (a *App) save() {\n\ta.TodoStore.Save(a.TodoList.Data)\n\tif a.TodoList.IsSynced {\n\t\ta.EventLogger.ProcessEvents()\n\n\t\tsynchronizer := NewQuietSynchronizer()\n\t\tsynchronizer.ExecSyncInBackground()\n\t}\n}", "func (m *MockManager) Save(arg0 state.State) error {\n\tret := m.ctrl.Call(m, \"Save\", arg0)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func Save() {\n\tenforcer.SavePolicy()\n}", "func (rf *Raft) persist() {\n\t// Your code here (2C).\n\t// Example:\n\t// w := new(bytes.Buffer)\n\t// e := gob.NewEncoder(w)\n\t// e.Encode(rf.xxx)\n\t// e.Encode(rf.yyy)\n\t// data := w.Bytes()\n\t// rf.persister.SaveRaftState(data)\n}", "func (rf *Raft) persist() {\n\t// Your code here (2C).\n\t// Example:\n\t// w := new(bytes.Buffer)\n\t// e := gob.NewEncoder(w)\n\t// e.Encode(rf.xxx)\n\t// e.Encode(rf.yyy)\n\t// data := w.Bytes()\n\t// rf.persister.SaveRaftState(data)\n}", "func (rf *Raft) persist() {\n\t// Your code here (2C).\n\t// Example:\n\t// w := new(bytes.Buffer)\n\t// e := gob.NewEncoder(w)\n\t// e.Encode(rf.xxx)\n\t// e.Encode(rf.yyy)\n\t// data := w.Bytes()\n\t// rf.persister.SaveRaftState(data)\n}", "func (rf *Raft) persist() {\n\t// Your code here (2C).\n\t// Example:\n\t// w := new(bytes.Buffer)\n\t// e := gob.NewEncoder(w)\n\t// e.Encode(rf.xxx)\n\t// e.Encode(rf.yyy)\n\t// data := w.Bytes()\n\t// rf.persister.SaveRaftState(data)\n}", "func (su *StateUpdate) Save(ctx context.Context) (int, error) {\n\tvar (\n\t\terr error\n\t\taffected int\n\t)\n\tif err := su.defaults(); err != nil {\n\t\treturn 0, err\n\t}\n\tif len(su.hooks) == 0 {\n\t\tif err = su.check(); err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t\taffected, err = su.sqlSave(ctx)\n\t} else {\n\t\tvar mut Mutator = MutateFunc(func(ctx context.Context, m Mutation) (Value, error) {\n\t\t\tmutation, ok := m.(*StateMutation)\n\t\t\tif !ok {\n\t\t\t\treturn nil, fmt.Errorf(\"unexpected mutation type %T\", m)\n\t\t\t}\n\t\t\tif err = su.check(); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t\tsu.mutation = mutation\n\t\t\taffected, err = su.sqlSave(ctx)\n\t\t\tmutation.done = true\n\t\t\treturn affected, err\n\t\t})\n\t\tfor i := len(su.hooks) - 1; i >= 0; i-- {\n\t\t\tif su.hooks[i] == nil {\n\t\t\t\treturn 0, fmt.Errorf(\"ent: uninitialized hook (forgotten import ent/runtime?)\")\n\t\t\t}\n\t\t\tmut = su.hooks[i](mut)\n\t\t}\n\t\tif _, err := mut.Mutate(ctx, su.mutation); err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\treturn affected, err\n}", "func (s *SessionStore) Save(r *http.Request, w http.ResponseWriter) {\n\ts.session.Save(r, w)\n}", "func Save() {\n\tgo db.save()\n}", "func (rf *Raft) persist() {\n\t// Your code here (2C).\n\t// Example:\n\t// w := new(bytes.Buffer)\n\t// e := labgob.NewEncoder(w)\n\t// e.Encode(rf.xxx)\n\t// e.Encode(rf.yyy)\n\t// data := w.Bytes()\n\t// rf.persister.SaveRaftState(data)\n}", "func (rf *Raft) persist() {\n\t// Your code here (2C).\n\t// Example:\n\t// w := new(bytes.Buffer)\n\t// e := labgob.NewEncoder(w)\n\t// e.Encode(rf.xxx)\n\t// e.Encode(rf.yyy)\n\t// data := w.Bytes()\n\t// rf.persister.SaveRaftState(data)\n}", "func (rf *Raft) persist() {\n\t// Your code here (2C).\n\t// Example:\n\t// w := new(bytes.Buffer)\n\t// e := labgob.NewEncoder(w)\n\t// e.Encode(rf.xxx)\n\t// e.Encode(rf.yyy)\n\t// data := w.Bytes()\n\t// rf.persister.SaveRaftState(data)\n}", "func (rf *Raft) persist() {\n\t// Your code here (2C).\n\t// Example:\n\t// w := new(bytes.Buffer)\n\t// e := labgob.NewEncoder(w)\n\t// e.Encode(rf.xxx)\n\t// e.Encode(rf.yyy)\n\t// data := w.Bytes()\n\t// rf.persister.SaveRaftState(data)\n}", "func (rf *Raft) persist() {\n\t// Your code here (2C).\n\t// Example:\n\t// w := new(bytes.Buffer)\n\t// e := labgob.NewEncoder(w)\n\t// e.Encode(rf.xxx)\n\t// e.Encode(rf.yyy)\n\t// data := w.Bytes()\n\t// rf.persister.SaveRaftState(data)\n}", "func (rf *Raft) persist() {\n\t// Your code here (2C).\n\t// Example:\n\t// w := new(bytes.Buffer)\n\t// e := labgob.NewEncoder(w)\n\t// e.Encode(rf.xxx)\n\t// e.Encode(rf.yyy)\n\t// data := w.Bytes()\n\t// rf.persister.SaveRaftState(data)\n}", "func (rf *Raft) persist() {\n\t// Your code here (2C).\n\t// Example:\n\t// w := new(bytes.Buffer)\n\t// e := labgob.NewEncoder(w)\n\t// e.Encode(rf.xxx)\n\t// e.Encode(rf.yyy)\n\t// data := w.Bytes()\n\t// rf.persister.SaveRaftState(data)\n}", "func (rf *Raft) persist() {\n\t// Your code here (2C).\n\t// Example:\n\t// w := new(bytes.Buffer)\n\t// e := labgob.NewEncoder(w)\n\t// e.Encode(rf.xxx)\n\t// e.Encode(rf.yyy)\n\t// data := w.Bytes()\n\t// rf.persister.SaveRaftState(data)\n}", "func (rf *Raft) persist() {\n\t// Your code here (2C).\n\t// Example:\n\t// w := new(bytes.Buffer)\n\t// e := labgob.NewEncoder(w)\n\t// e.Encode(rf.xxx)\n\t// e.Encode(rf.yyy)\n\t// data := w.Bytes()\n\t// rf.persister.SaveRaftState(data)\n}", "func (rf *Raft) persist() {\n\t// Your code here (2C).\n\t// Example:\n\t// w := new(bytes.Buffer)\n\t// e := labgob.NewEncoder(w)\n\t// e.Encode(rf.xxx)\n\t// e.Encode(rf.yyy)\n\t// data := w.Bytes()\n\t// rf.persister.SaveRaftState(data)\n}", "func (r *Registry) SaveUnitState(jobName string, unitState *unit.UnitState) {\n\tkey := path.Join(r.keyPrefix, statePrefix, jobName)\n\t//TODO: Handle the error generated by marshal\n\tjson, _ := marshal(unitState)\n\tr.etcd.Set(key, json, 0)\n}", "func (p *PasswordStruct) Save() error {\n\terr := ioutil.WriteFile(path, p.hash, 0700)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (b *bookMark) save(event win_eventlog.EvtHandle) error {\n\tnewBookmark, err := win_eventlog.UpdateBookmark(b.handle, event, b.buf)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif err := b.file.Truncate(0); err != nil {\n\t\treturn err\n\t}\n\tif _, err := b.file.Seek(0, 0); err != nil {\n\t\treturn err\n\t}\n\t_, err = b.file.WriteString(newBookmark)\n\treturn err\n}", "func (r *Reserve) Save(s *Store) error {\n\tdata, _ := json.Marshal(s)\n\tif err := ioutil.WriteFile(r.path, data, 0644); err != nil {\n\t\treturn fmt.Errorf(\"Failed to set %s: %s\", r.name, err)\n\t}\n\treturn nil\n}", "func (m *SynapsesPersist) Save() {\n\tif m.changed {\n\t\tfmt.Println(\"Saving synaptic data...\")\n\t\tindentedJSON, _ := json.MarshalIndent(m.Synapses, \"\", \" \")\n\n\t\tdataPath, err := filepath.Abs(m.relativePath)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\terr = ioutil.WriteFile(dataPath+m.file, indentedJSON, 0644)\n\t\tif err != nil {\n\t\t\tlog.Fatalln(\"ERROR:\", err)\n\t\t}\n\n\t\tm.Clean()\n\t\tfmt.Println(\"Synaptic data saved\")\n\t}\n}", "func (app *service) Save(genesis Genesis) error {\n\t_, err := app.repository.Retrieve()\n\tif err == nil {\n\t\treturn errors.New(\"there is already a Genesis instance\")\n\t}\n\n\tbill := genesis.Bill()\n\terr = app.billService.Save(bill)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttrGenesis, err := app.adapter.ToTransfer(genesis)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn app.trService.Save(trGenesis)\n}" ]
[ "0.7236683", "0.7199711", "0.71685624", "0.7060631", "0.70468223", "0.7015522", "0.6987076", "0.6944727", "0.68399274", "0.6821941", "0.6816623", "0.6716353", "0.665334", "0.642694", "0.6410426", "0.6376332", "0.63302517", "0.6325576", "0.6318987", "0.62886006", "0.62780863", "0.62738", "0.61917365", "0.61706156", "0.6162184", "0.61579883", "0.6138573", "0.61210823", "0.61115867", "0.61050797", "0.60904264", "0.608915", "0.60810995", "0.60350317", "0.6022017", "0.6003001", "0.5985761", "0.596192", "0.59510446", "0.5939137", "0.593168", "0.59304434", "0.59267217", "0.5924021", "0.5921208", "0.5917931", "0.5908505", "0.5894345", "0.58896405", "0.5884412", "0.5854171", "0.58363175", "0.5803075", "0.58015025", "0.5783581", "0.5769925", "0.5764562", "0.57631636", "0.575508", "0.5739827", "0.5737591", "0.5732639", "0.5731883", "0.572984", "0.5729124", "0.5726887", "0.5716998", "0.5714166", "0.5710861", "0.57084674", "0.5702479", "0.5694886", "0.56921095", "0.56799394", "0.56765485", "0.56756115", "0.5660181", "0.56565404", "0.56565404", "0.56565404", "0.56565404", "0.5633462", "0.5632151", "0.56263316", "0.56247246", "0.56247246", "0.56247246", "0.56247246", "0.56247246", "0.56247246", "0.56247246", "0.56247246", "0.56247246", "0.56247246", "0.5623785", "0.56193095", "0.5619151", "0.5618591", "0.56158966", "0.558569" ]
0.6848158
8
LoadState is not implemented
func (t *WindowedThroughput) LoadState(state []byte) error { return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func TestLoadState(t *testing.T) {\n\t// state := loadState()\n\t// fmt.Println(state.FormatVersion)\n}", "func (n *mockAgent) load(s persistapi.AgentState) {}", "func LoadState(id string) *Gossiper {\n\tg := Gossiper{}\n\tcwd, _ := os.Getwd()\n\tf, e := ioutil.ReadFile(filepath.Join(cwd, utils.STATE_FOLDER, fmt.Sprint(id, \".json\")))\n\tutils.HandleError(e)\n\tjson.Unmarshal(f, &g)\n\tfmt.Printf(\"Got this: %+v\\n\", &g)\n\treturn &g\n}", "func (i *inodeFileState) loadLoading(_ struct{}) {\n\ti.loading.Lock()\n}", "func (r *Registrar) LoadState() {\n\tif existing, e := os.Open(r.registryFile); e == nil {\n\t\tdefer existing.Close()\n\t\tfmt.Printf(\"Loading registrar data from %s\\n\", r.registryFile)\n\t\tdecoder := json.NewDecoder(existing)\n\t\tdecoder.Decode(&r.State)\n\t}\n}", "func (m *stateManager) Load() (err error) {\n\tm.state, err = m.storage.Read()\n\n\tif err == nil {\n\t\tm.stateChanged = false\n\t\tm.stateLoaded = true\n\t}\n\n\treturn\n}", "func (m *stateManager) ensureLoaded() (err error) {\n\tif !m.stateLoaded {\n\t\terr = m.Load()\n\t}\n\n\treturn\n}", "func (manager *basicStateManager) Load() error {\n\ts := manager.state\n\tseelog.Info(\"Loading state!\")\n\tdata, err := manager.readFile()\n\tif err != nil {\n\t\tseelog.Error(\"Error reading existing state file\", \"err\", err)\n\t\treturn err\n\t}\n\tif data == nil {\n\t\treturn nil\n\t}\n\t// Dry-run to make sure this is a version we can understand\n\terr = manager.dryRun(data)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// Now load it into the actual state. The reason we do this with the\n\t// intermediate state is that we *must* unmarshal directly into the\n\t// \"saveable\" pointers we were given in AddSaveable; if we unmarshal\n\t// directly into a map with values of pointers, those pointers are lost.\n\t// We *must* unmarshal this way because the existing pointers could have\n\t// semi-initialized data (and are actually expected to)\n\n\tvar intermediate intermediateState\n\terr = json.Unmarshal(data, &intermediate)\n\tif err != nil {\n\t\tseelog.Debug(\"Could not unmarshal into intermediate\")\n\t\treturn err\n\t}\n\n\tfor key, rawJSON := range intermediate.Data {\n\t\tactualPointer, ok := manager.state.Data[key]\n\t\tif !ok {\n\t\t\tseelog.Error(\"Loading state: potentially malformed json key of \" + key)\n\t\t\tcontinue\n\t\t}\n\t\terr = json.Unmarshal(rawJSON, actualPointer)\n\t\tif err != nil {\n\t\t\tseelog.Debug(\"Could not unmarshal into actual\")\n\t\t\treturn err\n\t\t}\n\t}\n\n\tseelog.Debug(\"Loaded state!\", \"state\", s)\n\treturn nil\n}", "func (m Mapper) loadGlobalState() (gs *GlobalState) {\n\tb := m.store.Get(GlobalStateKey)\n\tif b == nil {\n\t\treturn initialGlobalState()\n\t}\n\tgs = new(GlobalState)\n\terr := m.cdc.UnmarshalJSON(b, gs)\n\tif err != nil {\n\t\tpanic(err) // This error should never occur big problem if does\n\t}\n\treturn\n}", "func (me TxsdActuate) IsOnLoad() bool { return me.String() == \"onLoad\" }", "func (p *BaseProvider) Load() error {\n\treturn nil\n}", "func (hdSt *HeaderState) LoadHeaderState() error {\n\tsqlStr := \"select content from state where id=1\"\n\tvar text string\n\terr := hdSt.db.QueryRowx(sqlStr).Scan(&text)\n\tif err == sql.ErrNoRows {\n\t\treturn nil\n\t}\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn json.Unmarshal([]byte(text), hdSt)\n}", "func LoadState(username string) (*State, error) {\n\tf, err := os.Open(getPathForUsername(username))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer f.Close()\n\n\tvar s State\n\tif err := json.NewDecoder(f).Decode(&s); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &s, nil\n}", "func (ds *RegularStateMachineWrapper) Loaded(from rsm.From) {\n\tds.mu.Lock()\n\tdefer ds.mu.Unlock()\n\tds.SetLoaded(from)\n}", "func loadGlobalState(store types.KVStore) (gs *GlobalState) {\n\tb := store.Get(GlobalStateKey)\n\tif b == nil {\n\t\treturn initialGlobalState()\n\t}\n\tgs = new(GlobalState)\n\terr := cdc.UnmarshalJSON(b, gs)\n\tif err != nil {\n\t\tpanic(err) // This error should never occur big problem if does\n\t}\n\treturn\n}", "func (p *ProjectList) BeforeLoad() {\n}", "func (me TxsdActuate) IsOnLoad() bool { return me == \"onLoad\" }", "func (m *stateManager) Loaded() bool {\n\treturn m.stateLoaded\n}", "func (bot *LeagueAnnouncerBot) loadState() error {\n\t// Look for a state file and read it in\n\tbytes, err := ioutil.ReadFile(BOT_STATE)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Load the state into memory, loading summoner details and discord channels\n\tbotState := botStorage{MonitoredSummoners: make(map[string]*SummonerDetails)}\n\terr = botState.load(bytes, bot.riot, bot.discord)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Overwrite the bots current state with what we loaded from disk\n\tbot.storage = botState\n\tlog.Debugf(\"Loaded bot state from disk, Loaded %d users\", len(botState.MonitoredSummoners))\n\treturn nil\n}", "func Load() error {\n\treturn def.Load()\n}", "func (j *JobLog) BeforeLoad() {\n}", "func TestStateSaveLoad(t *testing.T) {\n\ttearDown, stateDB, state := setupTestCase(t)\n\tdefer tearDown(t)\n\t// nolint: vetshadow\n\tassert := assert.New(t)\n\n\tstate.LastBlockHeight++\n\tSaveState(stateDB, state)\n\n\tloadedState := LoadState(stateDB)\n\tassert.True(state.Equals(loadedState),\n\t\tfmt.Sprintf(\"expected state and its copy to be identical.\\ngot: %v\\nexpected: %v\\n\",\n\t\t\tloadedState, state))\n}", "func (d *mlDelegate) LocalState(bool) []byte { return nil }", "func (sm *StateMachine) LoadState(in *Msg) {\n\ttmp, err := json.Marshal(sm.state)\n\tif err != nil {\n\t\tsm.plugin.Log.Info(\"failed to marshal state for db.\", err)\n\t\treturn\n\t}\n\n\t// Using upsert to either insert and return a value or on conflict to\n\t// update and return a value doesn't work, leading to this longer form.\n\t// Could it be a Postgres bug? This can and should be optimized.\n\tif in.User.ID > 0 {\n\t\tq := `INSERT INTO states\n\t\t (key, userid, value, pluginname) VALUES ($1, $2, $3, $4)`\n\t\t_, err = sm.plugin.DB.Exec(q, StateKey, in.User.ID, tmp,\n\t\t\tsm.plugin.Config.Name)\n\t} else {\n\t\tq := `INSERT INTO states\n\t\t (key, flexid, flexidtype, value, pluginname) VALUES ($1, $2, $3, $4, $5)`\n\t\t_, err = sm.plugin.DB.Exec(q, StateKey, in.User.FlexID,\n\t\t\tin.User.FlexIDType, tmp, sm.plugin.Config.Name)\n\t}\n\tif err != nil {\n\t\tif err.Error() != `pq: duplicate key value violates unique constraint \"states_userid_pkgname_key_key\"` &&\n\t\t\terr.Error() != `pq: duplicate key value violates unique constraint \"states_flexid_flexidtype_pluginname_key_key\"` {\n\t\t\tsm.plugin.Log.Info(\"could not insert value into states.\", err)\n\t\t\tsm.state = 0\n\t\t\treturn\n\t\t}\n\t\tif in.User.ID > 0 {\n\t\t\tq := `SELECT value FROM states\n\t\t\t WHERE userid=$1 AND key=$2 AND pluginname=$3`\n\t\t\terr = sm.plugin.DB.Get(&tmp, q, in.User.ID, StateKey,\n\t\t\t\tsm.plugin.Config.Name)\n\t\t} else {\n\t\t\tq := `SELECT value FROM states\n\t\t\t WHERE flexid=$1 AND flexidtype=$2 AND key=$3 AND pluginname=$4`\n\t\t\terr = sm.plugin.DB.Get(&tmp, q, in.User.FlexID,\n\t\t\t\tin.User.FlexIDType, StateKey, sm.plugin.Config.Name)\n\t\t}\n\t\tif err != nil {\n\t\t\tsm.plugin.Log.Info(\"failed to get value from state.\", err)\n\t\t\treturn\n\t\t}\n\t}\n\tvar val int\n\tif err = json.Unmarshal(tmp, &val); err != nil {\n\t\tsm.plugin.Log.Info(\"failed unmarshaling state from db.\", err)\n\t\treturn\n\t}\n\tsm.state = val\n\n\t// Have we already entered a state?\n\tsm.stateEntered = sm.plugin.GetMemory(in, stateEnteredKey).Bool()\n\treturn\n}", "func (d delegate) LocalState(join bool) []byte { return nil }", "func (r *Registrar) loadStates() error {\n\tf, err := os.Open(r.registryFile)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdefer f.Close()\n\n\tlogp.Info(\"Loading registrar data from %s\", r.registryFile)\n\n\tdecoder := json.NewDecoder(f)\n\tstates := []models.State{}\n\terr = decoder.Decode(&states)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Error decoding states: %s\", err)\n\t}\n\n\n\tlogp.Info(\"States Loaded from registrar: %+v\", len(states))\n\n\treturn nil\n}", "func LoadFromState(board Board, tiles Tiles, players []Player, turn Turn) Game {\n\n\tdict, err := LoadDictionary(dictPath)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn Game{\n\t\tboard: board,\n\t\tplayers: players,\n\t\tTiles: tiles,\n\t\tTurn: turn,\n\t\tDictionary: dict,\n\t}\n}", "func Load() {\n\tjsonFile, err := os.Open(\"data/savegame.json\")\n\tcheck(err)\n\tfmt.Println(\"Loaded the file.\")\n\tdefer jsonFile.Close()\n\n\tbyteValue, err := ioutil.ReadAll(jsonFile)\n\tcheck(err)\n\tvar data Savedata\n\tjson.Unmarshal(byteValue, &data)\n\n\tGS.paused = false\n\tswitch loadedGamestate := data.Name; loadedGamestate {\n\tcase \"Flying\":\n\t\tGS.current = flying\n\t\tUI76.currentPanel = MenuFlying\n\tdefault:\n\t\tGS.current = landed\n\t\tUI76.currentPanel = MenuLanded\n\t}\n}", "func LoadState(backupFile string, noteAll *[]*notes.Note) {\n\tif _, err := os.Stat(backupFile); os.IsNotExist(err) {\n\t\tfmt.Println(\"No backup file\")\n\t\treturn\n\t}\n\tcontent, err := ioutil.ReadFile(backupFile)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\tvar initNote []*notes.Note\n\terr = json.Unmarshal(content, &initNote)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\t*noteAll = initNote\n\tfmt.Println(initNote)\n}", "func (env *Environment) Load(source, code string) error {\n\treturn nil\n}", "func (env *Environment) Load(source, code string) error {\n\treturn nil\n}", "func (this *Manager) load() error{\n bytes, err := ioutil.ReadFile(this.filename())\n if err != nil {\n return err\n }\n mp := dynmap.NewDynMap()\n err = mp.UnmarshalJSON(bytes)\n if err != nil {\n return err\n }\n table,err := ToRouterTable(mp)\n if err != nil {\n return err\n }\n this.SetRouterTable(table) \n return nil\n}", "func LoadIdentity() {\n C.glowLoadIdentity(gpLoadIdentity)\n}", "func LoadState(ctx context.Context, key int) (*sql.State, error) {\n\tvar r sql.State\n\n\tdata, err := Load(ctx, stateLoaderKey, strconv.Itoa(key))\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tr, ok := data.(sql.State)\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"wrong type: the expected type is %T but got %T\", r, data)\n\t}\n\n\treturn &r, nil\n}", "func (defaultStorage) Load() error {\n\tpanic(noConfigStorage)\n}", "func (h *handler) Load(state []byte) (core.Template, error) {\n\tdec := scale.NewDecoder(bytes.NewBuffer(state))\n\tvault := &Vault{}\n\tif _, err := vault.DecodeScale(dec); err != nil {\n\t\treturn nil, fmt.Errorf(\"%w: %w\", core.ErrInternal, err)\n\t}\n\treturn vault, nil\n}", "func Load() (string, test.TestStepFactory, []event.Name) {\n\treturn Name, New, Events\n}", "func (s *VCStore) LoadState() (types.SandboxState, error) {\n\tvar state types.SandboxState\n\n\tif err := s.state.Load(State, &state); err != nil {\n\t\treturn types.SandboxState{}, err\n\t}\n\n\treturn state, nil\n}", "func Load(cfg interface{}) error { return c.Load(cfg) }", "func IsPrePopulateStateCache() bool {\n\treturn false\n}", "func ActivateLoad(r *Root, e *Editor) {\n\tr.a = &SaveAndLoadEditor{\n\t\te: e,\n\t\tt: &Typer{\n\t\t\ttyp: true,\n\t\t\tPlaceholder: \"\",\n\t\t\tC: &e.c,\n\t\t},\n\t\tload: true,\n\t}\n}", "func (d *Database) Load() error {\n\tb, err := ioutil.ReadFile(d.FilePath)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err := json.Unmarshal(b, &d.State); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (d *Drawer) load() error {\n\tdata, err := ioutil.ReadFile(d.path)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif len(data) > 0 {\n\t\tvar payload interface{}\n\t\tpayload, err = d.serializer.Deserialize(data)\n\t\tif !isPointer(payload) {\n\t\t\tpanic(NonPointerErr)\n\t\t}\n\t\td.payload = payload\n\t} else {\n\t\td.payload = nil\n\t}\n\treturn err\n}", "func ShowLoader() {\n\tatomic.CompareAndSwapInt32(&show_loader, 0, 1)\n}", "func (r *Raft) loadInitialState() error {\n\titerator, err := r.logManager.ReplayEntries()\n\tif err == persister.ErrDatabaseEmpty {\n\t\treturn nil\n\t} else if err != nil {\n\t\treturn err\n\t}\n\n\terr = r.stateMachine.LoadState(iterator)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tr.stateManager.SetCommitIndex(r.logManager.GetLastLogIndex())\n\n\treturn nil\n}", "func LoadIdentity() {\n\tC.glowLoadIdentity(gpLoadIdentity)\n}", "func MustLoad(cfg interface{}) { c.MustLoad(cfg) }", "func (sys *IAMSys) Load(objAPI ObjectLayer) error {\n\tif globalEtcdClient != nil {\n\t\treturn sys.refreshEtcd()\n\t}\n\treturn sys.refresh(objAPI)\n}", "func (ctx *appContext) loadData() {\n\tctx.Lock()\n\tdefer ctx.Unlock()\n}", "func (*GameScene) Preload() {\n\tengo.Files.Load(\"textures/city.png\")\n}", "func (o *Storestate) Reload(exec boil.Executor) error {\n\tret, err := FindStorestate(exec, o.Statename)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (bh *browserHistory) Load() error {\n\thist := js.Global().Get(\"localStorage\").Get(\"history\")\n\tif hist.Type() == js.TypeUndefined {\n\t\treturn nil // nothing to unmarashal\n\t}\n\tvar records []string\n\tif err := json.Unmarshal([]byte(hist.String()), &records); err != nil {\n\t\treturn err\n\t}\n\tbh.Records = records\n\n\treturn nil\n}", "func (a *App) load() error {\n\ttodos, err := a.TodoStore.Load()\n\tif err != nil {\n\t\treturn err\n\t}\n\ta.TodoList.Load(todos)\n\ta.EventLogger = NewEventLogger(a.TodoList, a.TodoStore)\n\treturn nil\n}", "func (p *Periph) LoadSTATE() State {\n\treturn State(p.state.Bits(0xf))\n}", "func (i *Interpreter) Load(v string) (*big.Int, bool) {\n\tx, ok := i.state[v]\n\treturn x, ok\n}", "func loadStateFromJSON(fileName string) (*resultState, error) {\n\tret := &resultState{}\n\texists, err := loadJSONFile(fileName, ret)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif !exists {\n\t\treturn nil, skerr.Fmt(\"The state file %q doesn't exist.\", fileName)\n\t}\n\treturn ret, nil\n}", "func (i *inodeFileState) saveLoading() struct{} {\n\treturn struct{}{}\n}", "func (a *API) loadOAuthState(w http.ResponseWriter, r *http.Request) (context.Context, error) {\n\tvar state string\n\tif r.Method == http.MethodPost {\n\t\tstate = r.FormValue(\"state\")\n\t} else {\n\t\tstate = r.URL.Query().Get(\"state\")\n\t}\n\n\tif state == \"\" {\n\t\treturn nil, badRequestError(\"OAuth state parameter missing\")\n\t}\n\n\tctx := r.Context()\n\toauthToken := r.URL.Query().Get(\"oauth_token\")\n\tif oauthToken != \"\" {\n\t\tctx = withRequestToken(ctx, oauthToken)\n\t}\n\toauthVerifier := r.URL.Query().Get(\"oauth_verifier\")\n\tif oauthVerifier != \"\" {\n\t\tctx = withOAuthVerifier(ctx, oauthVerifier)\n\t}\n\treturn a.loadExternalState(ctx, state)\n}", "func (s *ShardFamily) Load(slice *Slice) (err error) {\n\n}", "func _iload(frame *runtime.Frame, index uint) {\n\tval := frame.GetInt(index)\n\tframe.PushInt(val)\n}", "func LoadFlag() {\n\tredis.LoadFlag()\n}", "func (ds *RegularStateMachineWrapper) Offloaded(from rsm.From) {\n\tds.mu.Lock()\n\tdefer ds.mu.Unlock()\n\tds.SetOffloaded(from)\n\tif ds.ReadyToDestroy() && !ds.Destroyed() {\n\t\tds.destroy()\n\t\tds.SetDestroyed()\n\t}\n}", "func (this *DatastoreOperations) LoadIfNeeded(increment bool) (*DatastoreState, error) {\n\t// Lock the state mutex\n\tthis.stateLock.Lock()\n\n\t// Unlock the state mutex whenever the function exits\n\tdefer this.stateLock.Unlock()\n\n\t// If the file was previously known not to be found and that result was cached, immediately\n\t// return with that error.\n\tif this.lastPathError != nil {\n\t\treturn nil, this.lastPathError\n\t}\n\n\t// Get existing state object\n\texistingState := this.State\n\n\t// If it exists (I.e. the datastore is already open)\n\tif existingState != nil {\n\t\t// If the returned state's reference count should be incremented\n\t\tif increment == true {\n\t\t\t// Increment the current state's file reference count\n\t\t\texistingState.Increment()\n\t\t}\n\n\t\t// Return the existing state object\n\t\treturn existingState, nil\n\t}\n\t// Otherwise\n\n\t// Start measuring the operation time\n\tstartTime := MonoUnixTimeMilliFloat()\n\n\t// Load the datastore\n\tstate, err := this.Load()\n\n\t// If an error occurred while loading the datastore\n\tif err != nil {\n\t\t// If the file wasn't found, cache that outcome to prevent unnecessary future calls to the file system.\n\t\t// This gives a significant performance boost, especially when trying to load dedicated configuration\n\t\t// datastores, which are quite common to be missing.\n\t\t//\n\t\t// However a disadvantage is that files that are copied to the storage directory after the server has\n\t\t// been started are not guaranteed to be seen by the server. One way to counteract this would be to reset\n\t\t// the cached value to nil every time interval (say every a second or two)\n\t\tif pathError, ok := err.(*os.PathError); ok {\n\t\t\tthis.lastPathError = pathError\n\t\t}\n\n\t\t// Return the error\n\t\treturn nil, err\n\t}\n\n\t// If the returned state's reference count should be incremented\n\tif increment == true {\n\t\t// Increment the current file's reference count\n\t\tstate.Increment()\n\t}\n\n\t// Set the resulting state object to the current state object\n\tthis.State = state\n\n\t// Log a completion message\n\tthis.ParentServer.Logf(1, \"Loaded datastore '%s' in %fms\", this.Name, MonoUnixTimeMilliFloat()-startTime)\n\n\treturn state, nil\n}", "func Load() {\n\trouter.Get(uri, Index)\n}", "func (T *Task) Load() bool {\n\n\tif Gdb.IsSetup() != true {\n\t\treturn false\n\t}\n\tT.Name = T.GetDBKey(\"Instance\")\n\tT.ID = T.GetDBKey(\"ID\")\n\tT.State = T.GetDBKey(\"State\")\n\tT.SlaveOf = T.GetDBKey(\"SlaveOf\")\n\tT.EID = T.GetDBKey(\"EID\")\n\tT.SID = T.GetDBKey(\"SID\")\n\tT.Type = T.GetDBKey(\"Type\")\n\tT.Stats.FromJson(T.GetDBKey(\"Stats\"))\n\n\treturn true\n}", "func (i *Inventoree) Load() error {\n\tvar err error\n\tif i.cacheExpired() {\n\t\treturn i.Reload()\n\t}\n\t// trying to use cache\n\terr = i.loadLocal()\n\tif err != nil {\n\t\t// if it failed, trying to get data from remote\n\t\treturn i.loadRemote()\n\t}\n\treturn nil\n}", "func (self *LSHforest) Load(path string) error {\n\tb, err := ioutil.ReadFile(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn msgpack.Unmarshal(b, self)\n}", "func (a *Agent) loadCheckState(check *structs.HealthCheck) error {\n\t// Try to read the persisted state for this check\n\tfile := filepath.Join(a.config.DataDir, checkStateDir, checkIDHash(check.CheckID))\n\tbuf, err := ioutil.ReadFile(file)\n\tif err != nil {\n\t\tif os.IsNotExist(err) {\n\t\t\treturn nil\n\t\t}\n\t\treturn fmt.Errorf(\"failed reading file %q: %s\", file, err)\n\t}\n\n\t// Decode the state data\n\tvar p persistedCheckState\n\tif err := json.Unmarshal(buf, &p); err != nil {\n\t\ta.logger.Printf(\"[ERR] agent: failed decoding check state: %s\", err)\n\t\treturn a.purgeCheckState(check.CheckID)\n\t}\n\n\t// Check if the state has expired\n\tif time.Now().Unix() >= p.Expires {\n\t\ta.logger.Printf(\"[DEBUG] agent: check state expired for %q, not restoring\", check.CheckID)\n\t\treturn a.purgeCheckState(check.CheckID)\n\t}\n\n\t// Restore the fields from the state\n\tcheck.Output = p.Output\n\tcheck.Status = p.Status\n\treturn nil\n}", "func Load() {\n\tloadNS()\n\tloadRoot()\n}", "func (c *configLoader) Loaded() {\n\tc.o.GetStruct(\"size\", &Fcfg.Size)\n\tc.o.GetStruct(\"file\", &Fcfg.File)\n}", "func (s *BaselimboListener) EnterLoad_expression(ctx *Load_expressionContext) {}", "func (page *Page) Load() (e error) {\n\tfmt.Print(\".\")\n\tif body, err := page.Fetch(); err != nil {\n\t\te = err\n\t} else {\n\t\tif body != nil {\n\t\t\tpage.Parse(body)\n\t\t}\n\t}\n\tpage.Site.WaitGroup.Done()\n\treturn\n}", "func load(interpreter *Interpreter) {\n\tname := interpreter.PopName()\n\tvalue, _ := interpreter.FindValueInDictionaries(name)\n\tif value == nil {\n\t\tlog.Printf(\"Can't find value %s\\n\", name)\n\t}\n\tinterpreter.Push(value)\n}", "func (*Scene) Preload() {\n\tengo.Files.Load(\"textures/citySheet.png\", \"tilemap/TrafficMap.tmx\")\n\tengo.Files.LoadReaderData(\"go.ttf\", bytes.NewReader(gosmallcaps.TTF))\n}", "func (db *LocalDb) load() error {\n\n\tdb.mutex.Lock()\n\tdefer db.mutex.Unlock()\n\n\t// read file\n\tdata, err := ioutil.ReadFile(constLocalDbFn)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t\treturn err\n\t}\n\n\t// decode json\n\tdb.users = make(map[string]*UserInfo)\n\tif err := json.Unmarshal(data, &db.users); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func loadTestCase(tc end2endTest) (*types.State, error) {\n\tf, err := os.Open(fmt.Sprintf(\"./testdata/%s.out\", tc.name))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer f.Close()\n\n\tp, err := serializer.Load(f)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn types.NewState(p), nil\n}", "func (t *StoreCC) load(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n\n\tvar A string // Entities\n\tvar err error\n\n\tif len(args) != 1 {\n\t\treturn shim.Error(\"Incorrect number of arguments. Expecting name of the person to query\")\n\t}\n\n\tA = args[0]\n\n\t// Get the state from the ledger\n\tAvalbytes, err := stub.GetState(A)\n\tif err != nil {\n\t\tjsonResp := \"{\\\"Error\\\":\\\"Failed to get state for \" + A + \"\\\"}\"\n\t\treturn shim.Error(jsonResp)\n\t}\n\n\tif Avalbytes == nil {\n\t\tjsonResp := \"{\\\"Error\\\":\\\"Nil amount for \" + A + \"\\\"}\"\n\t\treturn shim.Error(jsonResp)\n\t}\n\n\tjsonResp := string(Avalbytes)\n\tlogger.Infof(\"Query Response:%s\\n\", jsonResp)\n\treturn shim.Success(Avalbytes)\n}", "func loadData() {\r\n\tloadImages()\r\n\tsortPredictDetails()\t \r\n}", "func (p *GenericPlugin) Load() error {\n\n\tp.mu.Lock()\n\tdefer p.mu.Unlock()\n\n\tswitch p.state {\n\tcase stateNotLoaded:\n\tcase stateLoaded:\n\t\tp.agent.Logger().Printf(\"Cannot Load() module [ %s ], symbols already loaded\", p.Name())\n\t\treturn nil\n\tcase stateActive:\n\t\tp.agent.Logger().Printf(\"Cannot Load() module [ %s ], already running\", p.Name())\n\t\treturn nil\n\t}\n\n\t// Load the default symbols.\n\tif err := p.loadDefaultSymbols(); err != nil {\n\t\tp.agent.Logger().Printf(\"plugin load error: Failed to load default symbols from file [ %s ]\", p.filename)\n\t\treturn err\n\t}\n\n\tif p.init == nil {\n\t\tp.agent.Logger().Printf(\"Loaded all symbols except Init() for module [ %s ]\", p.Name())\n\t\treturn nil\n\t}\n\n\tp.state = stateLoaded\n\n\tp.agent.Logger().Printf(\"Loaded all symbols for module [ %s ]\", p.Name())\n\n\treturn nil\n}", "func Load() {\n\tpostgres.Load()\n}", "func (tfl tiltfileLoader) Load(ctx context.Context, tf *corev1alpha1.Tiltfile, prevResult *TiltfileLoadResult) TiltfileLoadResult {\n\tstart := time.Now()\n\tfilename := tf.Spec.Path\n\tabsFilename, err := ospath.RealAbs(tf.Spec.Path)\n\tif err != nil {\n\t\tif os.IsNotExist(err) {\n\t\t\treturn TiltfileLoadResult{\n\t\t\t\tConfigFiles: []string{filename},\n\t\t\t\tError: fmt.Errorf(\"No Tiltfile found at paths '%s'. Check out https://docs.tilt.dev/tutorial.html\", filename),\n\t\t\t}\n\t\t}\n\t\tabsFilename, _ = filepath.Abs(filename)\n\t\treturn TiltfileLoadResult{\n\t\t\tConfigFiles: []string{absFilename},\n\t\t\tError: err,\n\t\t}\n\t}\n\n\ttiltignorePath := watch.TiltignorePath(absFilename)\n\ttlr := TiltfileLoadResult{\n\t\tConfigFiles: []string{absFilename, tiltignorePath},\n\t}\n\n\ttiltignore, err := watch.ReadTiltignore(tiltignorePath)\n\n\t// missing tiltignore is fine, but a filesystem error is not\n\tif err != nil {\n\t\ttlr.Error = err\n\t\treturn tlr\n\t}\n\n\ttlr.Tiltignore = tiltignore\n\n\ts := newTiltfileState(ctx, tfl.dcCli, tfl.webHost, tfl.execer, tfl.k8sContextPlugin, tfl.versionPlugin,\n\t\ttfl.configPlugin, tfl.extensionPlugin, tfl.ciSettingsPlugin, feature.FromDefaults(tfl.fDefaults))\n\n\tmanifests, result, err := s.loadManifests(tf)\n\n\ttlr.BuiltinCalls = result.BuiltinCalls\n\ttlr.DefaultRegistry = s.defaultReg\n\n\t// All data models are loaded with GetState. We ignore the error if the state\n\t// isn't properly loaded. This is necessary for handling partial Tiltfile\n\t// execution correctly, where some state is correctly assembled but other\n\t// state is not (and should be assumed empty).\n\tws, _ := watch.GetState(result)\n\ttlr.WatchSettings = ws\n\n\t// NOTE(maia): if/when add secret settings that affect the engine, add them to tlr here\n\tss, _ := secretsettings.GetState(result)\n\ts.secretSettings = ss\n\n\tioState, _ := io.GetState(result)\n\n\ttlr.ConfigFiles = append(tlr.ConfigFiles, ioState.Paths...)\n\ttlr.ConfigFiles = append(tlr.ConfigFiles, s.postExecReadFiles...)\n\ttlr.ConfigFiles = sliceutils.DedupedAndSorted(tlr.ConfigFiles)\n\n\tdps, _ := dockerprune.GetState(result)\n\ttlr.DockerPruneSettings = dps\n\n\taSettings, _ := tiltfileanalytics.GetState(result)\n\ttlr.AnalyticsOpt = aSettings.Opt\n\n\ttlr.Secrets = s.extractSecrets()\n\ttlr.FeatureFlags = s.features.ToEnabled()\n\ttlr.Error = err\n\ttlr.Manifests = manifests\n\ttlr.TeamID = s.teamID\n\n\tobjectSet, _ := v1alpha1.GetState(result)\n\ttlr.ObjectSet = objectSet\n\n\tvs, _ := version.GetState(result)\n\ttlr.VersionSettings = vs\n\n\ttelemetrySettings, _ := telemetry.GetState(result)\n\ttlr.TelemetrySettings = telemetrySettings\n\n\tus, _ := updatesettings.GetState(result)\n\ttlr.UpdateSettings = us\n\n\tci, _ := cisettings.GetState(result)\n\ttlr.CISettings = ci\n\n\tconfigSettings, _ := config.GetState(result)\n\tif tlr.Error == nil {\n\t\ttlr.EnabledManifests, tlr.Error = configSettings.EnabledResources(tf, manifests)\n\t}\n\n\tduration := time.Since(start)\n\tif tlr.Error == nil {\n\t\ts.logger.Infof(\"Successfully loaded Tiltfile (%s)\", duration)\n\t}\n\textState, _ := tiltextension.GetState(result)\n\thashState, _ := hasher.GetState(result)\n\n\tvar prevHashes hasher.Hashes\n\tif prevResult != nil {\n\t\tprevHashes = prevResult.Hashes\n\t}\n\ttlr.Hashes = hashState.GetHashes()\n\n\ttfl.reportTiltfileLoaded(s.builtinCallCounts, s.builtinArgCounts, duration,\n\t\textState.ExtsLoaded, prevHashes, tlr.Hashes)\n\n\tif len(aSettings.CustomTagsToReport) > 0 {\n\t\treportCustomTags(tfl.analytics, aSettings.CustomTagsToReport)\n\t}\n\n\treturn tlr\n}", "func (t *qmlfrontend) onLoad(v *backend.View) {\n\tw2 := t.windows[v.Window()]\n\ti := 0\n\tfor i = range w2.views {\n\t\tif w2.views[i].bv == v {\n\t\t\tbreak\n\t\t}\n\t}\n\tv2 := w2.views[i]\n\tv2.Title.Text = v.FileName()\n\ttabs := w2.window.ObjectByName(\"tabs\")\n\ttabs.Set(\"currentIndex\", w2.ActiveViewIndex())\n\ttab := tabs.Call(\"getTab\", i).(qml.Object)\n\ttab.Set(\"title\", v2.Title.Text)\n}", "func (self *Iload1) Execute(frame *runtime.Frame) {\n\t_iload(frame, 1)\n}", "func loadNonResource(m ResourceMetas) {\n\tloadK9s(m)\n\tloadRBAC(m)\n\tloadHelm(m)\n\t// BOZO!! Revamp with latest...\n\t// if IsOpenFaasEnabled() {\n\t// \tloadOpenFaas(m)\n\t// }\n}", "func (s *LoaderSuite) TestLoadRubbish() {\n\tif testing.Short() {\n\t\ts.T().Skip()\n\t}\n\t_, err := s.loader.Load(context.TODO(), \"ajklfjkjva\")\n\ts.NotNil(err)\n}", "func (a *App) updateState() {\n\t// a.state.Parts1 = a.parseLine(viewPart1)\n\t// a.state.Parts2 = a.parseLine(viewPart2)\n\t// a.state.Tlds = a.parseLine(viewTLD)\n\t// a.state.Domains = a.parseLine(viewDomain)\n}", "func Loaded(source string) bool {\n\treturn source != \"\"\n}", "func (g *Gonf) Load() error {\n\tb, err := ioutil.ReadFile(g.path)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tcfg := map[string]interface{}{}\n\tif err := json.Unmarshal(b, &cfg); err != nil {\n\t\treturn err\n\t}\n\n\t// forbids access to Get/Set local and HTTP while reloading\n\tg.confLock.Lock()\n\tdefer g.confLock.Unlock()\n\tg.conf = cfg\n\n\treturn g.load(cfg, []string{}, []otoFlag{})\n}", "func (l *Loader) Init(ctx context.Context) (err error) {\n\trollbackHolder := fr.NewRollbackHolder(\"loader\")\n\tdefer func() {\n\t\tif err != nil {\n\t\t\trollbackHolder.RollbackReverseOrder()\n\t\t}\n\t}()\n\n\ttctx := tcontext.NewContext(ctx, l.logger)\n\n\tcheckpoint, err := newRemoteCheckPoint(tctx, l.cfg, l.checkpointID())\n\tfailpoint.Inject(\"ignoreLoadCheckpointErr\", func(_ failpoint.Value) {\n\t\tl.logger.Info(\"\", zap.String(\"failpoint\", \"ignoreLoadCheckpointErr\"))\n\t\terr = nil\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\tl.checkPoint = checkpoint\n\trollbackHolder.Add(fr.FuncRollback{Name: \"close-checkpoint\", Fn: l.checkPoint.Close})\n\n\tl.baList, err = filter.New(l.cfg.CaseSensitive, l.cfg.BAList)\n\tif err != nil {\n\t\treturn terror.ErrLoadUnitGenBAList.Delegate(err)\n\t}\n\n\terr = l.genRouter(l.cfg.RouteRules)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif len(l.cfg.ColumnMappingRules) > 0 {\n\t\tl.columnMapping, err = cm.NewMapping(l.cfg.CaseSensitive, l.cfg.ColumnMappingRules)\n\t\tif err != nil {\n\t\t\treturn terror.ErrLoadUnitGenColumnMapping.Delegate(err)\n\t\t}\n\t}\n\n\tdbCfg := l.cfg.To\n\tdbCfg.RawDBCfg = config.DefaultRawDBConfig().\n\t\tSetMaxIdleConns(l.cfg.PoolSize)\n\n\t// used to change loader's specified DB settings, currently SQL Mode\n\tlcfg, err := l.cfg.Clone()\n\tif err != nil {\n\t\treturn err\n\t}\n\t// fix nil map after clone, which we will use below\n\t// TODO: we may develop `SafeClone` in future\n\tif lcfg.To.Session == nil {\n\t\tlcfg.To.Session = make(map[string]string)\n\t}\n\tlcfg.To.Session[\"time_zone\"] = \"+00:00\"\n\n\thasSQLMode := false\n\tfor k := range l.cfg.To.Session {\n\t\tif strings.ToLower(k) == \"sql_mode\" {\n\t\t\thasSQLMode = true\n\t\t\tbreak\n\t\t}\n\t}\n\n\tif !hasSQLMode {\n\t\tsqlModes, err3 := utils.AdjustSQLModeCompatible(l.cfg.LoaderConfig.SQLMode)\n\t\tif err3 != nil {\n\t\t\tl.logger.Warn(\"cannot adjust sql_mode compatible, the sql_mode will stay the same\", log.ShortError(err3))\n\t\t}\n\t\tlcfg.To.Session[\"sql_mode\"] = sqlModes\n\t}\n\n\tl.logger.Info(\"loader's sql_mode is\", zap.String(\"sqlmode\", lcfg.To.Session[\"sql_mode\"]))\n\n\tl.toDB, l.toDBConns, err = createConns(tctx, lcfg, l.cfg.PoolSize)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (s *Index) loadFromDS() error {\n\tvar index IndexSnapshot\n\tif _, err := s.store.GetLastCheckpoint(&index); err != nil {\n\t\treturn err\n\t}\n\ts.index = index\n\treturn nil\n}", "func (this *DatastoreOperations) Load() (*DatastoreState, error) {\n\tvar err error\n\n\t// Initialize a blank state object\n\tstate := &DatastoreState{\n\t\tFlushScheduler: NewDatastoreFlushScheduler(),\n\t}\n\n\t// Open the datastore file\n\tstate.File, err = FileDescriptors.OpenAndIncrement(this.FilePath, os.O_RDWR, 0666)\n\n\t// If an error occurred while opening the file\n\tif err != nil {\n\t\t// Return the error\n\t\treturn nil, err\n\t}\n\n\t// Get file size\n\tfileSize, err := state.GetFileSize()\n\n\t// If an error occurred while getting the file size\n\tif err != nil {\n\t\t// Close the file\n\t\tstate.Decrement()\n\n\t\t// Return the error\n\t\treturn nil, err\n\t}\n\n\t// Create a new index object\n\tstate.Index = NewDatastoreIndex()\n\n\t// Add new entries to the index by scanning the datastore file\n\terr = state.Index.AppendFromEntryStream(NewPrefetchingReaderAt(state.File), 0, fileSize, func(iteratorResult *EntryStreamIteratorResult) error {\n\t\tchecksumErr := iteratorResult.VerifyAllChecksums()\n\t\tif checksumErr != nil {\n\t\t\treturn checksumErr\n\t\t}\n\n\t\t// If the entry is the first one\n\t\tif iteratorResult.Offset == 0 {\n\t\t\t// Verify it is a valid head entry\n\t\t\theadEntryErr := iteratorResult.VerifyValidHeadEntry()\n\t\t\tif headEntryErr != nil {\n\t\t\t\treturn headEntryErr\n\t\t\t}\n\t\t}\n\n\t\t// If the current entry is the last one but it doesn't have a transaction end flag\n\t\tif iteratorResult.EndOffset() == fileSize && !iteratorResult.HasTransactionEndFlag() {\n\t\t\t// Return an unexpected end of stream error\n\t\t\treturn io.ErrUnexpectedEOF\n\t\t}\n\n\t\treturn nil\n\t})\n\n\t// If an error occurred while appending to the index, or the file was empty\n\tif err != nil || fileSize == 0 {\n\t\t// If file ended unexpectedly, was corrupted or last entry didn't include a transaction end marker\n\t\tif fileSize == 0 || err == io.ErrUnexpectedEOF || err == ErrCorruptedEntry || err == ErrInvalidHeadEntry {\n\t\t\t// Log message\n\t\t\tthis.ParentServer.Logf(1, \"An incomplete or corrupted transcacion found in datastore '%s'. Attempting repair..\", this.Name)\n\n\t\t\t// Attempt to roll back to last succesful transaction\n\t\t\trepairedState, err := this.Repair(state)\n\n\t\t\t// Release original file (the repaired file would have a different descriptor)\n\t\t\tstate.Decrement()\n\n\t\t\t// If an error occurred while repairing the datastore file\n\t\t\tif err != nil {\n\t\t\t\t// Return the error\n\t\t\t\treturn nil, err\n\t\t\t}\n\n\t\t\t// Set the repaired datastore as the loaded datastore\n\t\t\tstate = repairedState\n\t\t} else { // Otherwise, index creation failed for some other reason\n\t\t\t// Release file\n\t\t\tstate.Decrement()\n\n\t\t\t// Return the error\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// Load head entry\n\terr = state.LoadHeadEntry()\n\n\t// If some error occurred while loading the head entry\n\tif err != nil {\n\t\t// An error here would be highly unexpected since the datastore has just been checked for corruption\n\t\t// which included validation of its head entry.\n\t\tthis.ParentServer.Logf(1, \"Datastore '%s' cannot be opened due to an unexpected error while trying to load its head entry: %s\", this.Name, err.Error())\n\n\t\t// Release file\n\t\tstate.Decrement()\n\n\t\t// Return the error\n\t\treturn nil, err\n\t}\n\n\t// If this is a cached datastore, load its content to memory\n\tif this.IsCached {\n\t\t// Load and deserialize the file's content\n\t\terr = state.UpdateDataCache(state.File, 0, state.Size())\n\n\t\t// If some error occured while trying load the file's content\n\t\tif err != nil {\n\t\t\t// Release file\n\t\t\tstate.Decrement()\n\n\t\t\t// Return the error\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn state, nil\n}", "func (p *Processor) Load() {\n\tpath := p.URLPath\n\tpathWithID := p.URLPath + \"/{id}\"\n\tpathWithTrigger := p.URLPath + \"/__trigger\"\n\tRegister(\"POST\", path, p.PostHandler)\n\tRegister(\"PUT\", pathWithID, p.PutHandler)\n\tRegister(\"PATCH\", pathWithID, p.PatchHandler)\n\tRegister(\"GET\", pathWithID, p.GetHandler)\n\tRegister(\"GET\", path, p.GetPageHandler)\n\tRegister(\"DELETE\", pathWithID, p.DeleteHandler)\n\t// TriggerHandler do something internal\n\tRegister(\"POST\", pathWithTrigger, p.TriggerHandler)\n}", "func (i *inodeFileState) afterLoad() {\n\tload := func() (err error) {\n\t\t// See comment on i.loading().\n\t\tdefer func() {\n\t\t\tif err == nil {\n\t\t\t\ti.loading.Unlock()\n\t\t\t}\n\t\t}()\n\n\t\t// Manually restore the p9.File.\n\t\tname, ok := i.s.inodeMappings[i.sattr.InodeID]\n\t\tif !ok {\n\t\t\t// This should be impossible, see assertion in\n\t\t\t// beforeSave.\n\t\t\treturn fmt.Errorf(\"failed to find path for inode number %d. Device %s contains %s\", i.sattr.InodeID, i.s.connID, fs.InodeMappings(i.s.inodeMappings))\n\t\t}\n\t\t// TODO(b/38173783): Context is not plumbed to save/restore.\n\t\tctx := &dummyClockContext{context.Background()}\n\n\t\t_, i.file, err = i.s.attach.walk(ctx, splitAbsolutePath(name))\n\t\tif err != nil {\n\t\t\treturn fs.ErrCorruption{fmt.Errorf(\"failed to walk to %q: %v\", name, err)}\n\t\t}\n\n\t\t// Remap the saved inode number into the gofer device using the\n\t\t// actual device and actual inode that exists in our new\n\t\t// environment.\n\t\tqid, mask, attrs, err := i.file.getAttr(ctx, p9.AttrMaskAll())\n\t\tif err != nil {\n\t\t\treturn fs.ErrCorruption{fmt.Errorf(\"failed to get file attributes of %s: %v\", name, err)}\n\t\t}\n\t\tif !mask.RDev {\n\t\t\treturn fs.ErrCorruption{fmt.Errorf(\"file %s lacks device\", name)}\n\t\t}\n\t\ti.key = device.MultiDeviceKey{\n\t\t\tDevice: attrs.RDev,\n\t\t\tSecondaryDevice: i.s.connID,\n\t\t\tInode: qid.Path,\n\t\t}\n\t\tif !goferDevice.Load(i.key, i.sattr.InodeID) {\n\t\t\treturn fs.ErrCorruption{fmt.Errorf(\"gofer device %s -> %d conflict in gofer device mappings: %s\", i.key, i.sattr.InodeID, goferDevice)}\n\t\t}\n\n\t\tif i.sattr.Type == fs.RegularFile {\n\t\t\tenv, ok := fs.CurrentRestoreEnvironment()\n\t\t\tif !ok {\n\t\t\t\treturn errors.New(\"missing restore environment\")\n\t\t\t}\n\t\t\tuattr := unstable(ctx, mask, attrs, i.s.mounter, i.s.client)\n\t\t\tif env.ValidateFileSize && uattr.Size != i.savedUAttr.Size {\n\t\t\t\treturn fs.ErrCorruption{fmt.Errorf(\"file size has changed for %s: previously %d, now %d\", i.s.inodeMappings[i.sattr.InodeID], i.savedUAttr.Size, uattr.Size)}\n\t\t\t}\n\t\t\tif env.ValidateFileTimestamp && uattr.ModificationTime != i.savedUAttr.ModificationTime {\n\t\t\t\treturn fs.ErrCorruption{fmt.Errorf(\"file modification time has changed for %s: previously %v, now %v\", i.s.inodeMappings[i.sattr.InodeID], i.savedUAttr.ModificationTime, uattr.ModificationTime)}\n\t\t\t}\n\t\t\ti.savedUAttr = nil\n\t\t}\n\n\t\treturn nil\n\t}\n\n\tfs.Async(fs.CatchError(load))\n}", "func Loader(state *lua.LState) int {\n\tmod := state.SetFuncs(state.NewTable(), map[string]lua.LGFunction{\n\t\t\"compile\": func(L *lua.LState) int {\n\t\t\tcode := L.CheckString(1)\n\n\t\t\tluaCode, err := Compile(L, code)\n\t\t\tif err != nil {\n\t\t\t\tstate.Push(lua.LNil)\n\t\t\t\tstate.Push(lua.LString(err.Error()))\n\n\t\t\t\treturn 2\n\t\t\t}\n\n\t\t\tL.Push(lua.LString(luaCode))\n\t\t\treturn 1\n\t\t},\n\t})\n\n\t// returns the module\n\tstate.Push(mod)\n\treturn 1\n}", "func IsLoaded() bool {\n\treturn len(mapping) > 0\n}", "func GetCurrentState() {\n\n}", "func (e *engine) loadSession(ctx *Context) {\n\tif AppSessionManager().IsStateful() {\n\t\tctx.subject.Session = AppSessionManager().GetSession(ctx.Req.Unwrap())\n\t}\n}", "func (c *Component) DefaultOnLoad(...interface{}) bool {\n\treturn true\n}", "func Preload(L *lua.LState) {\n\tL.PreloadModule(\"xmlpath\", Loader)\n}", "func (b *Backend) init() {\n\tif b.state == \"\" {\n\t\tb.state = IDLE\n\t}\n}" ]
[ "0.65661424", "0.6459875", "0.6434492", "0.6406391", "0.6354136", "0.6343778", "0.6215261", "0.6191638", "0.59591895", "0.59208554", "0.5880978", "0.5877874", "0.5870722", "0.58493215", "0.58443236", "0.58175236", "0.5790054", "0.57744104", "0.5737251", "0.57220036", "0.567688", "0.56726086", "0.56669056", "0.5589726", "0.55884737", "0.5553848", "0.5506028", "0.54944277", "0.54330546", "0.54072344", "0.54072344", "0.54044497", "0.53931785", "0.53929347", "0.5390703", "0.53692013", "0.53620505", "0.5358968", "0.5350293", "0.5349969", "0.5335187", "0.53279173", "0.53127766", "0.5311007", "0.52932954", "0.52888644", "0.52785623", "0.5270503", "0.5267349", "0.52460563", "0.52247596", "0.5224643", "0.5218111", "0.5216513", "0.5214777", "0.5184952", "0.5171409", "0.5164288", "0.5163565", "0.5156954", "0.5147343", "0.51448107", "0.5142407", "0.5140728", "0.5132974", "0.5128452", "0.51244867", "0.50973713", "0.50952584", "0.5093856", "0.509354", "0.5081833", "0.50805676", "0.5076588", "0.50753874", "0.5074925", "0.5071441", "0.5069392", "0.50674784", "0.50674194", "0.5063053", "0.5059793", "0.50517565", "0.50516164", "0.50473386", "0.5046096", "0.5038818", "0.5031773", "0.502472", "0.5021852", "0.50196093", "0.50141823", "0.5013727", "0.50049216", "0.4998677", "0.4996466", "0.49949762", "0.4991625", "0.49913386", "0.49911126" ]
0.6319585
6
NewTestClient returns a TestClient with a replacement http handler function. Methods on the new TestClient are overrideable as well.
func NewTestClient(handleFunc http.HandlerFunc) (*httptest.Server, *TestClient, error) { ts := httptest.NewServer(handleFunc) opts := []option.ClientOption{ option.WithEndpoint(ts.URL), option.WithHTTPClient(http.DefaultClient), } c, err := NewClient(context.Background(), opts...) if err != nil { return nil, nil, err } tc := &TestClient{} tc.client = *c.(*client) tc.client.i = tc return ts, tc, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func newTestClient(fn roundTripFunc) *http.Client {\n\treturn &http.Client{\n\t\tTransport: fn,\n\t}\n}", "func newTestClient(fn RoundTripFunc) *http.Client {\n\treturn &http.Client{\n\t\tTransport: fn,\n\t}\n}", "func NewTestClient(fn http.RoundTripper) *http.Client {\n\treturn &http.Client{\n\t\tTransport: fn,\n\t}\n}", "func NewTestClient(fn http.RoundTripper) *http.Client {\n\treturn &http.Client{\n\t\tTransport: fn,\n\t}\n}", "func NewTestClient(fn roundTripFunc) *http.Client {\n\treturn &http.Client{\n\t\tTransport: roundTripFunc(fn),\n\t}\n}", "func NewTestClient(fn RoundTripFunc) *http.Client {\n\treturn &http.Client{\n\t\tTransport: RoundTripFunc(fn),\n\t}\n}", "func NewTestClient(fn RoundTripFunc) *http.Client {\n\treturn &http.Client{\n\t\tTransport: RoundTripFunc(fn),\n\t}\n}", "func NewTestClient(fn RoundTripFunc) *http.Client {\n\treturn &http.Client{\n\t\tTransport: RoundTripFunc(fn),\n\t}\n}", "func NewTestClient(fn RoundTripFunc) *http.Client {\n\treturn &http.Client{\n\t\tTransport: RoundTripFunc(fn),\n\t}\n}", "func newMockClient(doer func(*http.Request) (*http.Response, error)) *http.Client {\n\treturn &http.Client{\n\t\tTransport: transportFunc(doer),\n\t}\n}", "func NewTestClient(fn RoundTripFunc) *http.Client {\n\treturn &http.Client {\n\t\tTransport: RoundTripFunc(fn),\n\t}\n}", "func NewTestClient(c Client, handler http.Handler) (Client, *httptest.Server) {\n\tserver := httptest.NewServer(handler)\n\tcw := testClient{c, server.URL, &http.Client{}, log.Writer()}\n\n\treturn cw, server\n}", "func NewTestClient(fn RoundTripFunc) *httputil.HTTPClient {\n\ttestClient := httputil.NewHTTPClient()\n\ttestClient.HTTPClient.Transport = RoundTripFunc(fn)\n\treturn testClient\n}", "func newMockClient(doer func(*http.Request) (*http.Response, error)) *http.Client {\n\tv := &http.Transport{\n\t\tProxy: http.ProxyFromEnvironment,\n\t\tDialContext: (&net.Dialer{\n\t\t\tTimeout: 30 * time.Second,\n\t\t\tKeepAlive: 30 * time.Second,\n\t\t\tDualStack: true,\n\t\t}).DialContext,\n\t\tMaxIdleConns: 100,\n\t\tIdleConnTimeout: 90 * time.Second,\n\t\tTLSHandshakeTimeout: 10 * time.Second,\n\t\tExpectContinueTimeout: 1 * time.Second,\n\t}\n\tv.RegisterProtocol(\"http\", transportFunc(doer))\n\treturn &http.Client{\n\t\tTransport: http.RoundTripper(v),\n\t}\n}", "func NewTestClient(cfg *app.Configuration) (*TestClient, error) {\n\tvar cli TestClient\n\n\tbase := url.URL{}\n\tcli.baseURL = &base\n\n\tif cfg == nil {\n\t\tcfg = &app.Configuration{}\n\t}\n\n\tif len(cfg.UserSecret) == 0 {\n\t\tcfg.UserSecret = apptest.TestSecret\n\t}\n\tif cfg.IPPerMinute == 0 {\n\t\tcfg.IPPerMinute = 100000\n\t}\n\tif cfg.IPRateBurst == 0 {\n\t\tcfg.IPRateBurst = 100000\n\t}\n\n\tif cfg.DBDSN == \"\" {\n\t\ttdb, db, err := dbutil.NewTestDB()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\t// We don't need the db handle\n\t\tif err := db.Close(); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tcli.closers = append(cli.closers, tdb)\n\n\t\tcfg.DBDSN = dbutil.DefaultDataSource + \" dbname=\" + tdb.Name()\n\t}\n\n\ta, err := app.New(cfg)\n\tif err != nil {\n\t\tcli.Close()\n\t\treturn nil, err\n\t}\n\tcli.closers = append(cli.closers, a)\n\n\tcli.do = func(req *http.Request) (*http.Response, error) {\n\t\trr := httptest.NewRecorder()\n\t\ta.ServeHTTP(rr, req)\n\n\t\tresp := http.Response{\n\t\t\tStatus: fmt.Sprintf(\"%d %s\", rr.Code, http.StatusText(rr.Code)),\n\t\t\tStatusCode: rr.Code,\n\t\t\tBody: ioutil.NopCloser(rr.Body),\n\t\t\tHeader: rr.HeaderMap,\n\t\t\tContentLength: int64(rr.Body.Len()),\n\t\t\tRequest: req,\n\t\t}\n\n\t\treturn &resp, nil\n\t}\n\n\treturn &cli, nil\n}", "func wrappedClient(t *testing.T, testID string) (*Client, error) {\n\tctx := context.Background()\n\tbase := http.DefaultTransport\n\n\ttrans, err := htransport.NewTransport(ctx, base, option.WithoutAuthentication(), option.WithUserAgent(\"custom-user-agent\"))\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to create http client: %v\", err)\n\t}\n\n\tc := http.Client{Transport: trans}\n\n\t// Add RoundTripper to the created HTTP client\n\twrappedTrans := &retryTestRoundTripper{rt: c.Transport, testID: testID, T: t}\n\tc.Transport = wrappedTrans\n\n\t// Supply this client to storage.NewClient\n\t// STORAGE_EMULATOR_HOST takes care of setting the correct endpoint\n\tclient, err := NewClient(ctx, option.WithHTTPClient(&c))\n\treturn client, err\n}", "func newTestClient(cfg *TestClientConfig) *TestClient {\n\tdialer := &websocket.Dialer{\n\t\tReadBufferSize: cfg.BufferSize,\n\t\tWriteBufferSize: cfg.BufferSize,\n\t\tHandshakeTimeout: cfg.HandshakeTimeout,\n\t}\n\tif len(cfg.Subprotocol) > 0 {\n\t\tdialer.Subprotocols = cfg.Subprotocol\n\t}\n\treturn &TestClient{\n\t\tEndpoint: cfg.Endpoint,\n\t\tDialer: dialer,\n\t}\n}", "func newHTTPClient() *http.Client {\n\tclient := &http.Client{\n\t\tTimeout: defaultTimeout,\n\t}\n\treturn client\n}", "func testingHTTPClient(handler http.Handler) (*http.Client, func()) {\n\ts := httptest.NewServer(handler)\n\n\tcli := &http.Client{\n\t\tTransport: &http.Transport{\n\t\t\tDialContext: func(_ context.Context, network, _ string) (net.Conn, error) {\n\t\t\t\treturn net.Dial(network, s.Listener.Addr().String())\n\t\t\t},\n\t\t},\n\t}\n\n\treturn cli, s.Close\n}", "func NewTestClient(t *testing.T, fn RoundTripFunc, opts ...ClientOptionsFunc) *Client {\n\tc, err := NewClient(nil, defaultHost, nil, initLogger(t))\n\tif err != nil {\n\t\tt.Fatalf(\"unexpected error while declaring a client: %v\", err)\n\t}\n\n\tc.Client.HTTPClient.Transport = RoundTripFunc(fn)\n\n\tfor _, opt := range opts {\n\t\topt(c)\n\t}\n\n\treturn c\n}", "func testingHTTPClient(handler http.Handler) (*http.Client, func()) {\n\tserver := httptest.NewServer(handler)\n\n\tclient := &http.Client{\n\t\tTransport: &http.Transport{\n\t\t\tDialContext: func(_ context.Context, network, _ string) (net.Conn, error) {\n\t\t\t\treturn net.Dial(network, server.Listener.Addr().String())\n\t\t\t},\n\t\t},\n\t}\n\n\treturn client, server.Close\n}", "func newFakeClient() client.Client {\n\treturn fakeclient.NewFakeClient()\n}", "func NewClient(t mockConstructorTestingTNewClient) *Client {\n\tmock := &Client{}\n\tmock.Mock.Test(t)\n\n\tt.Cleanup(func() { mock.AssertExpectations(t) })\n\n\treturn mock\n}", "func NewClient(t mockConstructorTestingTNewClient) *Client {\n\tmock := &Client{}\n\tmock.Mock.Test(t)\n\n\tt.Cleanup(func() { mock.AssertExpectations(t) })\n\n\treturn mock\n}", "func NewClient(t mockConstructorTestingTNewClient) *Client {\n\tmock := &Client{}\n\tmock.Mock.Test(t)\n\n\tt.Cleanup(func() { mock.AssertExpectations(t) })\n\n\treturn mock\n}", "func NewClient(t mockConstructorTestingTNewClient) *Client {\n\tmock := &Client{}\n\tmock.Mock.Test(t)\n\n\tt.Cleanup(func() { mock.AssertExpectations(t) })\n\n\treturn mock\n}", "func NewHTTPClient(tc *trace.Client, orig *http.Client) *HTTPClient {\n\tif orig == nil {\n\t\torig = http.DefaultClient\n\t}\n\trt := orig.Transport\n\tif rt == nil {\n\t\trt = http.DefaultTransport\n\t}\n\tclient := http.Client{\n\t\tTransport: &tracerTransport{base: rt},\n\t\tCheckRedirect: orig.CheckRedirect,\n\t\tJar: orig.Jar,\n\t\tTimeout: orig.Timeout,\n\t}\n\treturn &HTTPClient{\n\t\tClient: client,\n\t\ttc: tc,\n\t}\n}", "func NewTestClient() *Client {\n\tbaseURL, _ := url.Parse(testNetClientURL)\n\tc := &Client{client: http.DefaultClient, BaseURL: baseURL}\n\treturn c\n}", "func NewClient(baseClient *httpclient.Client) *Client {\n\treturn &Client{\n\t\thttp: baseClient,\n\t}\n}", "func newHTTPClient() *http.Client {\n\treturn &http.Client{\n\t\tTransport: &http.Transport{\n\t\t\tProxy: http.ProxyFromEnvironment,\n\t\t\tDialContext: (&net.Dialer{\n\t\t\t\tTimeout: timeout,\n\t\t\t\tKeepAlive: 30 * time.Second,\n\t\t\t\tDualStack: true,\n\t\t\t}).DialContext,\n\n\t\t\tTLSHandshakeTimeout: timeout,\n\t\t\tResponseHeaderTimeout: timeout,\n\t\t\tExpectContinueTimeout: 1 * time.Second,\n\t\t\tMaxIdleConns: 5,\n\t\t\tIdleConnTimeout: 90 * time.Second,\n\t\t},\n\t}\n}", "func NewTestClientNew() *Client {\n\tif testClient == nil {\n\t\ttestClient = NewClientNew(TestAccessKeyId, TestAccessKeySecret)\n\t}\n\treturn testClient\n}", "func New(url string, httpClient *http.Client, customHeaders http.Header) *Client {\n\tif httpClient == nil {\n\t\thttpClient = &http.Client{\n\t\t\tTimeout: defaultHTTPTimeout,\n\t\t}\n\t}\n\n\treturn &Client{\n\t\turl: url,\n\t\thttpClient: httpClient,\n\t\tcustomHeaders: customHeaders,\n\t}\n}", "func NewMockClient(fn RoundTripFunc) *http.Client {\n\treturn &http.Client{\n\t\tTransport: RoundTripFunc(fn),\n\t}\n}", "func newCloudlyckeClient() *http.Client {\n\treturn &http.Client{}\n}", "func (rpc *RpcClient) newHTTPClient() (*http.Client, error) {\n\t// Configure proxy if needed.\n\tvar dial func(network, addr string) (net.Conn, error)\n\n\t// Configure TLS if needed.\n\tvar tlsConfig *tls.Config\n\n\t// Create and return the new HTTP client potentially configured with a\n\t// proxy and TLS.\n\tclient := http.Client{\n\t\tTransport: &http.Transport{\n\t\t\tDial: dial,\n\t\t\tTLSClientConfig: tlsConfig,\n\t\t\tDialContext: (&net.Dialer{\n\t\t\t\tTimeout: 5 * time.Second,\n\t\t\t\tKeepAlive: 5 * time.Second,\n\t\t\t\tDualStack: true,\n\t\t\t}).DialContext,\n\t\t},\n\t}\n\treturn &client, nil\n}", "func NewClient(config *Config) *Client {\n\ttr := config.Transport()\n\n\treturn &Client{\n\t\tconfig: config.Clone(),\n\t\ttr: tr,\n\t\tclient: &http.Client{Transport: tr},\n\t}\n}", "func (r *Request) newClient() *http.Client {\n\treturn &http.Client{Timeout: r.timeout}\n}", "func NewClient(options *ClientOptions, customHTTPClient *http.Client,\r\n\tcustomEnvironment string) (c *Client) {\r\n\r\n\t// Create a client\r\n\tc = new(Client)\r\n\r\n\t// Set options (either default or user modified)\r\n\tif options == nil {\r\n\t\toptions = DefaultClientOptions()\r\n\t}\r\n\r\n\t// Set the options\r\n\tc.Options = options\r\n\r\n\t// Set the environment\r\n\tvar found bool\r\n\tif c.Environment, found = environments[customEnvironment]; !found {\r\n\t\tc.Environment = environments[EnvironmentProduction]\r\n\t}\r\n\r\n\t// Is there a custom HTTP client to use?\r\n\tif customHTTPClient != nil {\r\n\t\tc.httpClient = customHTTPClient\r\n\t\treturn\r\n\t}\r\n\r\n\t// dial is the net dialer for clientDefaultTransport\r\n\tdial := &net.Dialer{KeepAlive: options.DialerKeepAlive, Timeout: options.DialerTimeout}\r\n\r\n\t// clientDefaultTransport is the default transport struct for the HTTP client\r\n\tclientDefaultTransport := &http.Transport{\r\n\t\tDialContext: dial.DialContext,\r\n\t\tExpectContinueTimeout: options.TransportExpectContinueTimeout,\r\n\t\tIdleConnTimeout: options.TransportIdleTimeout,\r\n\t\tMaxIdleConns: options.TransportMaxIdleConnections,\r\n\t\tProxy: http.ProxyFromEnvironment,\r\n\t\tTLSHandshakeTimeout: options.TransportTLSHandshakeTimeout,\r\n\t}\r\n\r\n\t// Determine the strategy for the http client\r\n\tif options.RequestRetryCount <= 0 {\r\n\r\n\t\t// no retry enabled\r\n\t\tc.httpClient = httpclient.NewClient(\r\n\t\t\thttpclient.WithHTTPTimeout(options.RequestTimeout),\r\n\t\t\thttpclient.WithHTTPClient(&http.Client{\r\n\t\t\t\tTransport: clientDefaultTransport,\r\n\t\t\t\tTimeout: options.RequestTimeout,\r\n\t\t\t}),\r\n\t\t)\r\n\t\treturn\r\n\t}\r\n\r\n\t// Retry enabled - create exponential back-off\r\n\tc.httpClient = httpclient.NewClient(\r\n\t\thttpclient.WithHTTPTimeout(options.RequestTimeout),\r\n\t\thttpclient.WithRetrier(heimdall.NewRetrier(\r\n\t\t\theimdall.NewExponentialBackoff(\r\n\t\t\t\toptions.BackOffInitialTimeout,\r\n\t\t\t\toptions.BackOffMaxTimeout,\r\n\t\t\t\toptions.BackOffExponentFactor,\r\n\t\t\t\toptions.BackOffMaximumJitterInterval,\r\n\t\t\t))),\r\n\t\thttpclient.WithRetryCount(options.RequestRetryCount),\r\n\t\thttpclient.WithHTTPClient(&http.Client{\r\n\t\t\tTransport: clientDefaultTransport,\r\n\t\t\tTimeout: options.RequestTimeout,\r\n\t\t}),\r\n\t)\r\n\r\n\treturn\r\n}", "func NewClient() *Client {\n\treturn &Client{\n\t\tisMocked: false,\n\t\tbaseURL: \"http://jsonplaceholder.typicode.com\",\n\t\tclient: &http.Client{},\n\t}\n}", "func newBaseClient() *baseClient {\n\treturn &baseClient{\n\t\thttpClient: http.DefaultClient,\n\t\tmethod: \"GET\",\n\t\theader: make(http.Header),\n\t}\n}", "func newInputService18ProtocolTestClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegion string) *InputService18ProtocolTest {\n\tsvc := &InputService18ProtocolTest{\n\t\tClient: client.New(\n\t\t\tcfg,\n\t\t\tmetadata.ClientInfo{\n\t\t\t\tServiceName: \"inputservice18protocoltest\",\n\t\t\t\tSigningRegion: signingRegion,\n\t\t\t\tEndpoint: endpoint,\n\t\t\t\tAPIVersion: \"2014-01-01\",\n\t\t\t},\n\t\t\thandlers,\n\t\t),\n\t}\n\n\t// Handlers\n\tsvc.Handlers.Sign.PushBack(v4.Sign)\n\tsvc.Handlers.Build.PushBackNamed(restxml.BuildHandler)\n\tsvc.Handlers.Unmarshal.PushBackNamed(restxml.UnmarshalHandler)\n\tsvc.Handlers.UnmarshalMeta.PushBackNamed(restxml.UnmarshalMetaHandler)\n\tsvc.Handlers.UnmarshalError.PushBackNamed(restxml.UnmarshalErrorHandler)\n\n\treturn svc\n}", "func NewClient(meta *metadata.Client, acc string) *http.Client {\n\treturn &http.Client{\n\t\tTransport: newRoundTripper(meta, acc),\n\t}\n}", "func NewClient(client *http.Client) *Client {\n\tif client == nil {\n\t\tclient = http.DefaultClient\n\t}\n\n\t// Set a custom default transport instead of the default transport.\n\tif client.Transport == nil {\n\t\tclient.Transport = DefaultTransport\n\t}\n\n\treturn &Client{client}\n}", "func newClient(httpClient *http.Client) (c *Client) {\n\tc = &Client{httpClient: httpClient}\n\tc.service.client = c\n\tc.Auth = (*AuthService)(&c.service)\n\tc.Providers = (*ProvidersService)(&c.service)\n\tc.Projects = (*ProjectsService)(&c.service)\n\tc.Releases = (*ReleasesService)(&c.service)\n\tc.SlackChannels = (*SlackChannelsService)(&c.service)\n\tc.TelegramChats = (*TelegramChatsService)(&c.service)\n\tc.DiscordChannels = (*DiscordChannelsService)(&c.service)\n\tc.HangoutsChatWebhooks = (*HangoutsChatWebhooksService)(&c.service)\n\tc.MicrosoftTeamsWebhooks = (*MicrosoftTeamsWebhooksService)(&c.service)\n\tc.MattermostWebhooks = (*MattermostWebhooksService)(&c.service)\n\tc.RocketchatWebhooks = (*RocketchatWebhooksService)(&c.service)\n\tc.MatrixRooms = (*MatrixRoomsService)(&c.service)\n\tc.Webhooks = (*WebhooksService)(&c.service)\n\tc.Tags = (*TagsService)(&c.service)\n\treturn c\n}", "func NewClient() *http.Client {\n\treturn &http.Client{\n\t\tTimeout: 10 * time.Second,\n\t}\n}", "func NewClient(httpClient *http.Client) *Client {\n\tif httpClient == nil {\n\t\tcloned := *http.DefaultClient\n\t\thttpClient = &cloned\n\t}\n\n\tbaseURL, _ := url.Parse(defaultBaseURL)\n\n\tc := &Client{\n\t\tclient: httpClient,\n\t\tBaseURL: baseURL,\n\t}\n\n\tc.common.client = c\n\tc.Question = (*QuestionService)(&c.common)\n\tc.Token = (*TokenService)(&c.common)\n\n\treturn c\n}", "func NewClient(\n\tmethod string,\n\ttgt *url.URL,\n\tenc kithttp.EncodeRequestFunc,\n\tdec kithttp.DecodeResponseFunc,\n\toptions ...kithttp.ClientOption,\n) *Client {\n\treturn &Client{Client: kithttp.NewClient(method, tgt, enc, makeDecodeResponseFunc(dec), options...)}\n}", "func newHTTPClient(cfg *Config) (*http.Client, error) {\n\t// Configure proxy if needed.\n\tvar dial func(network, addr string) (net.Conn, error)\n\tif cfg.Proxy != \"\" {\n\t\tproxy := &socks.Proxy{\n\t\t\tAddr: cfg.Proxy,\n\t\t\tUsername: cfg.ProxyUser,\n\t\t\tPassword: cfg.ProxyPass,\n\t\t}\n\t\tdial = func(network, addr string) (net.Conn, error) {\n\t\t\tc, err := proxy.Dial(network, addr)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\treturn c, nil\n\t\t}\n\t}\n\n\t// Configure TLS if needed.\n\tvar tlsConfig *tls.Config\n\tif !cfg.NoTLS {\n\t\ttlsConfig = &tls.Config{\n\t\t\tInsecureSkipVerify: cfg.TLSSkipVerify,\n\t\t}\n\t\tif !cfg.TLSSkipVerify && cfg.RPCCert != \"\" {\n\t\t\tpem, err := ioutil.ReadFile(cfg.RPCCert)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\n\t\t\tpool := x509.NewCertPool()\n\t\t\tif ok := pool.AppendCertsFromPEM(pem); !ok {\n\t\t\t\treturn nil, fmt.Errorf(\"invalid certificate file: %v\",\n\t\t\t\t\tcfg.RPCCert)\n\t\t\t}\n\t\t\ttlsConfig.RootCAs = pool\n\t\t}\n\t}\n\n\ttimeout, _ := time.ParseDuration(\"30s\")\n\n\t// Create and return the new HTTP client potentially configured with a\n\t// proxy and TLS.\n\tclient := http.Client{\n\t\tTransport: &http.Transport{\n\t\t\tDial: dial,\n\t\t\tTLSClientConfig: tlsConfig,\n\t\t},\n\t\tTimeout: timeout,\n\t}\n\treturn &client, nil\n}", "func NewClient() *http.Client {\n\treturn &http.Client{}\n}", "func CreateHTTPClient(handler http.Handler) (*http.Client, func()) {\n\ts := httptest.NewServer(handler)\n\n\tcli := &http.Client{\n\t\tTransport: &http.Transport{\n\t\t\tDialContext: func(_ context.Context, network, _ string) (net.Conn, error) {\n\t\t\t\treturn net.Dial(network, s.Listener.Addr().String())\n\t\t\t},\n\t\t},\n\t}\n\n\treturn cli, s.Close\n}", "func NewClient(config *config.Config, httpClient *http.Client) *Client {\n\treturn &Client{\n\t\tGetter: NewGetter(config, httpClient),\n\t}\n}", "func NewClient(c *http.Client, baseURL *url.URL) *client {\n\treturn &client{\n\t\tbaseURL: baseURL,\n\t\tclient: c,\n\t}\n}", "func NewMock(r ...MockResponse) *http.Client {\n\treturn &http.Client{\n\t\tTransport: newRoundTripper(r...),\n\t}\n}", "func (rpc *RpcClient) newHTTPClient() (*http.Client, error) {\n\t// Configure proxy if needed.\n\tvar dial func(network, addr string) (net.Conn, error)\n\tif rpc.Cfg.OptionConfig.Proxy != \"\" {\n\t\tproxy := &socks.Proxy{\n\t\t\tAddr: rpc.Cfg.OptionConfig.Proxy,\n\t\t\tUsername: rpc.Cfg.OptionConfig.ProxyUser,\n\t\t\tPassword: rpc.Cfg.OptionConfig.ProxyPass,\n\t\t}\n\t\tdial = func(network, addr string) (net.Conn, error) {\n\t\t\tc, err := proxy.Dial(network, addr)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\treturn c, nil\n\t\t}\n\t}\n\n\t// Configure TLS if needed.\n\tvar tlsConfig *tls.Config\n\tif !rpc.Cfg.SoloConfig.NoTLS && rpc.Cfg.SoloConfig.RPCCert != \"\" {\n\t\tpem, err := ioutil.ReadFile(rpc.Cfg.SoloConfig.RPCCert)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tpool := x509.NewCertPool()\n\t\tpool.AppendCertsFromPEM(pem)\n\t\ttlsConfig = &tls.Config{\n\t\t\tRootCAs: pool,\n\t\t\tInsecureSkipVerify: rpc.Cfg.SoloConfig.NoTLS,\n\t\t}\n\t} else {\n\t\ttlsConfig = &tls.Config{\n\t\t\tInsecureSkipVerify: rpc.Cfg.SoloConfig.NoTLS,\n\t\t}\n\t}\n\n\t// Create and return the new HTTP client potentially configured with a\n\t// proxy and TLS.\n\tclient := http.Client{\n\t\tTransport: &http.Transport{\n\t\t\tDial: dial,\n\t\t\tTLSClientConfig: tlsConfig,\n\t\t\tDialContext: (&net.Dialer{\n\t\t\t\tTimeout: time.Duration(rpc.Cfg.OptionConfig.Timeout) * time.Second,\n\t\t\t\tKeepAlive: time.Duration(rpc.Cfg.OptionConfig.Timeout) * time.Second,\n\t\t\t\tDualStack: true,\n\t\t\t}).DialContext,\n\t\t},\n\t}\n\treturn &client, nil\n}", "func NewTestClient(dialString string, pingPeriodic int, reopenConnection, debug bool, tcpKeepAlive int, tlsConfig *tls.Config, logger *Logging.Logger) *TestClient {\n\tclient := Pinger.NewClient(dialString, reopenConnection, tlsConfig, tcpKeepAlive, debug, logger)\n\tif client == nil {\n\t\tlogger.Error(\"Could not get Client\")\n\t\treturn nil\n\t}\n\tstopCh := make(chan int)\n\treturn &TestClient{\n\t\tclient: client,\n\t\tpingPeriodicity: pingPeriodic,\n\t\tdebug: debug,\n\t\tlogger: logger,\n\t\tstopCh: client.StopCh,\n\t\tstats: Utils.NewStatLogger(stopCh, logger, true),\n\t}\n}", "func NewClient(ctx context.Context, src TicketSource) *http.Client {\n\tif src == nil {\n\t\tc, err := internal.ContextClient(ctx)\n\t\tif err != nil {\n\t\t\treturn &http.Client{Transport: internal.ErrorTransport{Err: err}}\n\t\t}\n\t\treturn c\n\t}\n\treturn &http.Client{\n\t\tTransport: &Transport{\n\t\t\tBase: internal.ContextTransport(ctx),\n\t\t\tSource: ReuseTicketSource(nil, src),\n\t\t},\n\t}\n}", "func newClient(configuration *Configuration, options ...ClientOption) (Client, error) {\n\tclientCfg, err := newTLSClientConfig(configuration)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error reading and/or parsing the certification files. Cause: %w\", err)\n\t}\n\n\tnetClient := http.Client{\n\t\tTransport: &http.Transport{\n\t\t\tTLSClientConfig: clientCfg,\n\t\t},\n\t}\n\n\tinstance := &client{client: &netClient, configuration: configuration, encoder: newJSONEncoder(), decoder: newJSONDecoder()}\n\n\t// Apply options if there are any, can overwrite default\n\tfor _, option := range options {\n\t\toption(instance)\n\t}\n\n\treturn instance, nil\n}", "func NewTestClient(t *testing.T) *Client {\n\tconf := &Config{\n\t\tUsername: \"user\",\n\t\tPassword: \"pass\",\n\t}\n\n\tcl, err := NewClient(\"service\", \"hostname\", conf)\n\tif err != nil {\n\t\tt.Fatalf(\"could not create client\\n%v\", err)\n\t}\n\n\treturn cl\n}", "func NewClient() *Client {\n\tvar client Client\n\ttr := &http.Transport{\n\t\tDisableCompression: true,\n\t\tProxy: http.ProxyFromEnvironment,\n\t}\n\n\tclient.client = &http.Client{Transport: tr}\n\tclient.UserAgent = fmt.Sprintf(\"KUTTL/%s\", strings.TrimPrefix(version.Get().GitVersion, \"v\"))\n\treturn &client\n}", "func NewTestClient(stopChan chan os.Signal) (*TestClient, error) {\n\tbaseTestClientConfig, err := utils.CreateBaseTestClientConfig(stopChan)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttestClient := &TestClient{BaseTestClientConfig: baseTestClientConfig}\n\n\terr = testClient.initialize()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn testClient, nil\n}", "func New(mws ...MiddlewareFunc) *http.Client {\n\tclient := &http.Client{\n\t\tTimeout: 5 * time.Second,\n\t\tTransport: http.DefaultTransport,\n\t}\n\twithMiddleware(client, mws...)\n\treturn client\n}", "func NewClient(baseURL string, apiKey string) Client {\n\treturn &httpClient{\n\t\tapiKey: apiKey,\n\t\tbaseURL: baseURL,\n\t\tinst: &http.Client{},\n\t}\n}", "func NewClient(endpointURL, soapActionBase string, cl *http.Client) Caller {\n\tif cl == nil {\n\t\tcl = http.DefaultClient\n\t}\n\tif cl.Transport == nil {\n\t\tcl.Transport = http.DefaultTransport\n\t}\n\tcl.Transport = soaptrip.New(cl.Transport)\n\treturn &soapClient{\n\t\tClient: cl,\n\t\tURL: endpointURL,\n\t\tSOAPActionBase: soapActionBase,\n\t\tbufpool: bp.New(1024),\n\t}\n}", "func New(url string) *Client {\n\treturn &Client{&http.Client{}, url, func(r *http.Request) *http.Request { return r }}\n}", "func NewClient(httpClient *http.Client, c ClientConfig) Client {\n\treturn Client{\n\t\thttpClient: *httpClient,\n\t\tconfig: c,\n\t}\n}", "func newClient(certFile, keyFile string) (*http.Client, error) {\n\tcaCert, err := ioutil.ReadFile(\"/etc/insights-client/cert-api.access.redhat.com.pem\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tcaCertPool, err := x509.SystemCertPool()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tcaCertPool.AppendCertsFromPEM(caCert)\n\n\tcert, err := tls.LoadX509KeyPair(certFile, keyFile)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttlsConfig := tls.Config{\n\t\tRootCAs: caCertPool,\n\t\tCertificates: []tls.Certificate{cert},\n\t\tMaxVersion: tls.VersionTLS12, // cloud.redhat.com appears to exhibit this openssl bug https://github.com/openssl/openssl/issues/9767\n\t}\n\n\ttlsConfig.BuildNameToCertificate()\n\ttransport := http.Transport{\n\t\tTLSClientConfig: &tlsConfig,\n\t}\n\tclient := http.Client{\n\t\tTransport: &transport,\n\t}\n\treturn &client, nil\n}", "func NewClient(httpClient *http.Client, URL string, Token string, Source string, SourceType string, Index string) (*Client) {\n\t// Create a new client\n\tif httpClient == nil {\n\t\ttr := &http.Transport{TLSClientConfig: &tls.Config{InsecureSkipVerify: true}} // turn off certificate checking\n\t\thttpClient = &http.Client{Timeout: time.Second * 20, Transport: tr}\n\t}\n\n\tc := &Client{HTTPClient: httpClient, URL: URL, Token: Token, Source: Source, SourceType: SourceType, Index: Index}\n\n\treturn c\n}", "func NewClient() *client.Client {\n\tclnt := newMockClient(mockDoer)\n\tc, err := client.NewClient(\"tcp://127.0.0.1:2333\", mockAPIVersion, clnt, nil)\n\tif err != nil {\n\t\ttestlogF(err)\n\t\tpanic(err)\n\t}\n\treturn c\n}", "func NewClient(service, apiID, apiKey string, client *http.Client) *Client {\n\tc := &Client{\n\t\tservice: service,\n\t\tapiID: apiID,\n\t\tapiKey: apiKey,\n\t\tclient: client,\n\t}\n\tif c.client == nil {\n\t\tc.client = http.DefaultClient\n\t}\n\treturn c\n}", "func NewClient(httpClient *http.Client) *Client {\n\tvar c *http.Client\n\n\tif httpClient == nil {\n\t\tc = http.DefaultClient\n\t} else {\n\t\tc = httpClient\n\t}\n\n\treturn &Client{\n\t\tclient: c,\n\t}\n}", "func NewClient(url string) *Client {\n\treturn &Client{&http.Client{}, url}\n}", "func NewClient(httpClient *http.Client) *Client {\n\tif httpClient == nil {\n\t\thttpClient = http.DefaultClient\n\t}\n\tbaseURL, err := url.Parse(baseURL)\n\tc := &Client{client: httpClient, BaseURL: baseURL, err: err}\n\tc.common.client = c\n\tc.Teams = (*TeamsService)(&c.common)\n\tc.Invitations = (*InvitationsService)(&c.common)\n\treturn c\n}", "func testHTTPClient() *http.Client {\n\treturn &http.Client{\n\t\tCheckRedirect: func(req *http.Request, via []*http.Request) error {\n\t\t\treturn http.ErrUseLastResponse\n\t\t},\n\t}\n}", "func NewClient() *http.Client {\n\treturn &http.Client{\n\t\tTransport: roundTripper,\n\t\tTimeout: TCPConnectionTimeout,\n\t}\n}", "func NewClient(httpClient *http.Client) *Client {\n\tif httpClient == nil {\n\t\thttpClient = http.DefaultClient\n\t}\n\n\treturn &Client{\n\t\thttpClient: httpClient,\n\t}\n}", "func newInputService14ProtocolTestClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegion string) *InputService14ProtocolTest {\n\tsvc := &InputService14ProtocolTest{\n\t\tClient: client.New(\n\t\t\tcfg,\n\t\t\tmetadata.ClientInfo{\n\t\t\t\tServiceName: \"inputservice14protocoltest\",\n\t\t\t\tSigningRegion: signingRegion,\n\t\t\t\tEndpoint: endpoint,\n\t\t\t\tAPIVersion: \"2014-01-01\",\n\t\t\t},\n\t\t\thandlers,\n\t\t),\n\t}\n\n\t// Handlers\n\tsvc.Handlers.Sign.PushBack(v4.Sign)\n\tsvc.Handlers.Build.PushBackNamed(restxml.BuildHandler)\n\tsvc.Handlers.Unmarshal.PushBackNamed(restxml.UnmarshalHandler)\n\tsvc.Handlers.UnmarshalMeta.PushBackNamed(restxml.UnmarshalMetaHandler)\n\tsvc.Handlers.UnmarshalError.PushBackNamed(restxml.UnmarshalErrorHandler)\n\n\treturn svc\n}", "func NewClient(ctx context.Context, host string, port uint16, ctype ClientType,\n\tqueueSize int, flushInterval time.Duration) (hrpc.RegionClient, error) {\n\treturn &testClient{\n\t\thost: host,\n\t\tport: port,\n\t}, nil\n}", "func NewClient() *Client {\n baseURL, _ := url.Parse(defaultBaseURL)\n return &Client{client: http.DefaultClient, BaseURL: baseURL, UserAgent: userAgent}\n}", "func testHttpClient() *http.Client {\n\treturn &http.Client{\n\t\tCheckRedirect: func(req *http.Request, via []*http.Request) error {\n\t\t\treturn http.ErrUseLastResponse\n\t\t},\n\t}\n}", "func NewTest() (*client, error) {\n\tconfig := os.Getenv(\"VELA_QUEUE_CONFIG\")\n\tif len(config) == 0 {\n\t\tconfig = \"localhost:6379\"\n\t}\n\n\t// parse the url provided\n\toptions, err := redis.ParseURL(config)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// create the Redis client from the parsed url\n\tqueue := redis.NewClient(options)\n\n\t// create the client object\n\tclient := &client{\n\t\tQueue: queue,\n\t\tOptions: options,\n\t}\n\n\treturn client, nil\n}", "func NewClient(httpClient *http.Client) *Client {\n\tif httpClient == nil {\n\t\thttpClient = &http.Client{}\n\t}\n\tbaseURL, _ := url.Parse(baseURL)\n\tc := &Client{\n\t\tclient: httpClient,\n\t\tBaseURL: baseURL,\n\t}\n\tc.common.client = c\n\tc.Tags = (*TagsService)(&c.common)\n\tc.Manifests = (*ManifestsService)(&c.common)\n\treturn c\n}", "func NewClient(httpClient *http.Client) *Client {\n\tif httpClient == nil {\n\t\thttpClient = http.DefaultClient\n\t}\n\tbaseURL, _ := url.Parse(baseURL)\n\n\tc := &Client{client: httpClient, BaseURL: baseURL, UserAgent: userAgent}\n\tc.common.client = c\n\tc.RRSet = (*RRSetService)(&c.common)\n\tc.RData = (*RDataService)(&c.common)\n\n\treturn c\n}", "func NewHTTPClient(uri string) HTTPClient {\n\treturn HTTPClient{\n\t\tBackendURI: uri,\n\t\tclient: &http.Client{},\n\t}\n}", "func NewClient(c *http.Client) *Client {\n\tapi := &Client{\n\t\tc: c,\n\t\terrorHandler: defaultErrorHandler,\n\t\theaders: make(map[string]string),\n\t}\n\treturn api\n}", "func NewHTTPClient() *http.Client {\n\ttr := &http.Transport{\n\t\tTLSClientConfig: &tls.Config{\n\t\t\tInsecureSkipVerify: true, //nolint:gosec // Needs to be enabled in suites. Not used in production.\n\t\t},\n\t}\n\n\treturn &http.Client{\n\t\tTransport: tr,\n\t\tCheckRedirect: func(req *http.Request, via []*http.Request) error {\n\t\t\treturn http.ErrUseLastResponse\n\t\t},\n\t}\n}", "func NewClient(httpClient *http.Client) *Client {\n\tif httpClient == nil {\n\t\thttpClient = http.DefaultClient\n\t}\n\tbaseURL, _ := url.Parse(defaultBaseURL)\n\tc := Client{\n\t\tBaseURL: baseURL,\n\t\tclient: httpClient,\n\t\tUserAgent: userAgent,\n\t}\n\treturn &c\n}", "func NewClient(token string, client *http.Client) *Client {\n\tif client == nil {\n\t\thttpClient = &http.Client{Timeout: time.Second * 10}\n\t} else {\n\t\thttpClient = client\n\t}\n\treturn &Client{token}\n}", "func NewClient() *http.Client {\n\tt := &http.Transport{\n\t\tMaxIdleConns: 10,\n\t\tIdleConnTimeout: 30 * time.Second,\n\t}\n\n\treturn &http.Client{Transport: t}\n}", "func newClient(uri string, hc *http.Client, opts jsonclient.Options, log *entitylist.LogInfo) (*LogClient, error) {\n\tlogClient, err := jsonclient.New(uri, hc, opts)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &LogClient{*logClient, *log}, err\n}", "func NewClient(config vipps.ClientConfig) *Client {\n\tvar baseURL string\n\tvar logger logging.Logger\n\n\tif config.HTTPClient == nil {\n\t\tpanic(\"config.HTTPClient cannot be nil\")\n\t}\n\n\tif config.Environment == vipps.EnvironmentTesting {\n\t\tbaseURL = vipps.BaseURLTesting\n\t} else {\n\t\tbaseURL = vipps.BaseURL\n\t}\n\n\tif config.Logger == nil {\n\t\tlogger = logging.NewNopLogger()\n\t} else {\n\t\tlogger = config.Logger\n\t}\n\n\treturn &Client{\n\t\tBaseURL: baseURL,\n\t\tAPIClient: &internal.APIClient{\n\t\t\tL: logger,\n\t\t\tC: config.HTTPClient,\n\t\t},\n\t}\n}", "func NewClient(apiURL string, logger lager.Logger) Client {\n\treturn &client{\n\t\trequestGenerator: rata.NewRequestGenerator(apiURL, api.Routes),\n\t\tgivenHTTPClient: &http.Client{\n\t\t\tTransport: &http.Transport{\n\t\t\t\tDisableKeepAlives: false,\n\t\t\t\tResponseHeaderTimeout: 20 * time.Second,\n\t\t\t\tMaxIdleConns: 200,\n\t\t\t},\n\t\t},\n\t\tlogger: logger,\n\t}\n}", "func newNetClient() *http.Client {\n\tonce.Do(func() {\n\t\tvar netTransport = &http.Transport{\n\t\t\tDial: (&net.Dialer{\n\t\t\t\tTimeout: 15 * time.Second,\n\t\t\t}).Dial,\n\t\t\tTLSHandshakeTimeout: 8 * time.Second,\n\t\t\tExpectContinueTimeout: 8 * time.Second,\n\t\t\tResponseHeaderTimeout: 8 * time.Second,\n\t\t\tReadBufferSize: 8,\n\t\t}\n\t\tnetClient = &http.Client{\n\t\t\tTimeout: time.Second * 15,\n\t\t\tTransport: netTransport,\n\t\t}\n\t})\n\n\treturn netClient\n}", "func NewWithClient(chainID string, client SignStatusClient) provider.Provider {\n\treturn &http{\n\t\tchainID: chainID,\n\t\tclient: client,\n\t}\n}", "func NewClient(httpClient *http.Client) *Client {\n\tif httpClient == nil {\n\t\thttpClient = http.DefaultClient\n\t}\n\treturn &Client{httpClient: httpClient}\n}", "func TestNewClient_CustomHttpClient(t *testing.T) {\n\tt.Parallel()\n\n\tclient := NewClient(nil, http.DefaultClient, ProviderPreev)\n\n\tif client == nil {\n\t\tt.Fatal(\"failed to load client\")\n\t}\n\n\t// Test providers\n\tif client.Providers[0] != ProviderPreev {\n\t\tt.Fatalf(\"expected the first provider to be %d, not %d\", ProviderPreev, client.Providers[0])\n\t}\n}", "func NewClient() http.Client {\n\treturn http.Client{\n\t\tTransport: NewLogRoundTripper(),\n\t}\n}", "func NewClient(httpClient *http.Client) *Client {\n\tu, _ := url.Parse(BaseURL)\n\treturn &Client{\n\t\tBaseURL: u,\n\t\tHTTPClient: httpClient,\n\t}\n}", "func WithTestHTTPClient(ctx context.Context, client *http.Client) context.Context {\n\treturn context.WithValue(ctx, &httpClientCtxKey, client)\n}", "func NewTestClient() *TestClient {\n\treturn &TestClient{gpbrpc.NewRPCBaseClient()}\n}", "func newInputService20ProtocolTestClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegion string) *InputService20ProtocolTest {\n\tsvc := &InputService20ProtocolTest{\n\t\tClient: client.New(\n\t\t\tcfg,\n\t\t\tmetadata.ClientInfo{\n\t\t\t\tServiceName: \"inputservice20protocoltest\",\n\t\t\t\tSigningRegion: signingRegion,\n\t\t\t\tEndpoint: endpoint,\n\t\t\t\tAPIVersion: \"2014-01-01\",\n\t\t\t},\n\t\t\thandlers,\n\t\t),\n\t}\n\n\t// Handlers\n\tsvc.Handlers.Sign.PushBack(v4.Sign)\n\tsvc.Handlers.Build.PushBackNamed(restxml.BuildHandler)\n\tsvc.Handlers.Unmarshal.PushBackNamed(restxml.UnmarshalHandler)\n\tsvc.Handlers.UnmarshalMeta.PushBackNamed(restxml.UnmarshalMetaHandler)\n\tsvc.Handlers.UnmarshalError.PushBackNamed(restxml.UnmarshalErrorHandler)\n\n\treturn svc\n}" ]
[ "0.8164903", "0.815504", "0.73907304", "0.73907304", "0.7359928", "0.7343221", "0.7343221", "0.7343221", "0.7343221", "0.7332404", "0.7283051", "0.72369397", "0.72026974", "0.69602436", "0.679415", "0.6767714", "0.6720823", "0.6649837", "0.66306263", "0.66198933", "0.6580387", "0.6571907", "0.65621835", "0.65621835", "0.65621835", "0.65621835", "0.6546608", "0.65434396", "0.64968693", "0.6478145", "0.6466766", "0.6413288", "0.64101595", "0.64032465", "0.6371717", "0.6358649", "0.63425034", "0.632978", "0.6317538", "0.6310335", "0.6299954", "0.62876534", "0.6286836", "0.62688154", "0.6264032", "0.6251884", "0.62447923", "0.62426335", "0.623834", "0.6228055", "0.62276584", "0.62254167", "0.6221386", "0.62154454", "0.6191551", "0.6182038", "0.61813927", "0.6176404", "0.61690426", "0.61665004", "0.61615855", "0.61580676", "0.61535865", "0.61502236", "0.614552", "0.6142615", "0.61398727", "0.61370105", "0.61354756", "0.6126015", "0.61165476", "0.6112537", "0.6103921", "0.6103775", "0.6103763", "0.6103454", "0.61007345", "0.6100733", "0.6087226", "0.6084832", "0.60809875", "0.6080822", "0.6080065", "0.60771585", "0.60735893", "0.60732263", "0.6071198", "0.60637915", "0.60490006", "0.6045484", "0.60453695", "0.60439676", "0.6043152", "0.6032657", "0.6032314", "0.60319966", "0.601773", "0.6012998", "0.60105735", "0.60011905" ]
0.7024743
13
Retry uses the override method RetryFn or the real implementation.
func (c *TestClient) Retry(f func(opts ...googleapi.CallOption) (*compute.Operation, error), opts ...googleapi.CallOption) (op *compute.Operation, err error) { if c.RetryFn != nil { return c.RetryFn(f, opts...) } return c.client.Retry(f, opts...) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Retry(\n\tctx context.Context,\n\tf func(ctx context.Context) error,\n\tretryable func(err error) bool,\n\tnap time.Duration,\n) error {\n\treturn RetryN(ctx, f, retryable, nap, 100_000_000)\n}", "func Retry(totalRetryCount int, retryPause time.Duration, retriableFunction RetriableFunction) (err error) {\n\tretryCounter := 0\n\tretry := Retriable(true)\n\tfor {\n\t\tretry, err = retriableFunction()\n\t\tif err == nil || !retry {\n\t\t\tbreak\n\t\t}\n\n\t\tretryCounter++\n\t\tif totalRetryCount != -1 && retryCounter >= totalRetryCount {\n\t\t\tbreak\n\t\t}\n\n\t\tklog.Infof(\"Retrying in %s\", retryPause)\n\t\ttime.Sleep(retryPause)\n\t}\n\treturn\n}", "func (r *CustomRetrier) Retry(\n\tctx context.Context,\n\tretry int,\n\treq *http.Request,\n\tresp *http.Response,\n\terr error) (time.Duration, bool, error) {\n\t// Fail hard on a specific error\n\tif err == syscall.ECONNREFUSED {\n\t\treturn 0, false, errors.New(\"Elasticsearch or network down\")\n\t}\n\n\t// Stop after 5 retries\n\tif retry >= 5 {\n\t\treturn 0, false, nil\n\t}\n\n\t// Let the backoff strategy decide how long to wait and whether to stop\n\twait, stop := r.backoff.Next(retry)\n\treturn wait, stop, nil\n}", "func (r *CustomRetrier) Retry(\n\tctx context.Context,\n\tretry int,\n\treq *http.Request,\n\tresp *http.Response,\n\terr error) (time.Duration, bool, error) {\n\t// Fail hard on a specific error\n\tif err == syscall.ECONNREFUSED {\n\t\treturn 0, false, errors.New(\"elasticsearch or network down\")\n\t}\n\n\t// Stop after 5 retries\n\tif retry >= 5 {\n\t\treturn 0, false, nil\n\t}\n\n\t// Let the backoff strategy decide how long to wait and whether to stop\n\twait, stop := r.backoff.Next(retry)\n\n\treturn wait, stop, nil\n}", "func Retry(backoff Backoff, condition func() error) error {\n\treturn retry(backoff, condition, nil, nil)\n}", "func Retry(fn func() error) error {\n\tbo := backoff.WithMaxRetries(backoff.NewExponentialBackOff(), maxRetries)\n\tvar i int\n\treturn backoff.Retry(func() error {\n\t\terr := fn()\n\t\tif err != nil {\n\t\t\tif shouldRetry(err) {\n\t\t\t\tfmt.Printf(\"Retrying after error: %s\\n\", err)\n\t\t\t\ttime.Sleep(500 * time.Millisecond)\n\t\t\t\ti++\n\t\t\t\treturn fmt.Errorf(\"attempt #%d failed: %w\", i, err)\n\t\t\t}\n\t\t\treturn backoff.Permanent(err)\n\t\t}\n\t\treturn nil\n\t}, bo)\n}", "func (client *BaseClient) Retry() int {\n\treturn client.retry\n}", "func Retry(retryFunc RetryFunc, opts ...Option) error {\n\tconfig := &RetryConfig{\n\t\tretryTimes: DefaultRetryTimes,\n\t\tretryDuration: DefaultRetryDuration,\n\t\tcontext: context.TODO(),\n\t}\n\n\tfor _, opt := range opts {\n\t\topt(config)\n\t}\n\n\tvar i uint\n\tfor i < config.retryTimes {\n\t\terr := retryFunc()\n\t\tif err != nil {\n\t\t\tselect {\n\t\t\tcase <-time.After(config.retryDuration):\n\t\t\tcase <-config.context.Done():\n\t\t\t\treturn errors.New(\"retry is cancelled\")\n\t\t\t}\n\t\t} else {\n\t\t\treturn nil\n\t\t}\n\t\ti++\n\t}\n\n\tfuncPath := runtime.FuncForPC(reflect.ValueOf(retryFunc).Pointer()).Name()\n\tlastSlash := strings.LastIndex(funcPath, \"/\")\n\tfuncName := funcPath[lastSlash+1:]\n\n\treturn fmt.Errorf(\"function %s run failed after %d times retry\", funcName, i)\n}", "func (r *Retrier) Retry(fn func() error) error {\n\tstart := time.Now()\n\tretries := 0\n\tvar err error\n\tfor retry := true; retry; retry = time.Since(start) < r.timeout {\n\t\terr = fn()\n\t\tretries += 1\n\t\tif err == nil {\n\t\t\tfmt.Printf(\"Retry execution successful with %d retries in duration %v\", retries, time.Since(start))\n\t\t\treturn nil\n\t\t}\n\t\tfmt.Printf(\"Error happened during retry after %d retries: %v\", retries, err)\n\n\t\tretry, wait := r.retryPolicy(retries, err)\n\t\tif !retry {\n\t\t\tfmt.Println(\"Execution aborted by retry policy\")\n\t\t\treturn err\n\t\t}\n\n\t\tfmt.Printf(\"Sleeping before next retry: duration - %v\", wait)\n\t\ttime.Sleep(wait)\n\t}\n\n\tfmt.Printf(\"Timeout reached after %d retries in duration %v. Returning error: %v\", retries, time.Since(start), err)\n\n\treturn err\n}", "func Retry(n uint, interval time.Duration, fn func() error) (err error) {\n\treturn WithContext(context.Background(), n, interval, fn)\n}", "func Retry(attempts int, waitTime time.Duration, gf GenFunc) error {\n\tfor i := 0; i < attempts; i++ {\n\t\tok, err := gf()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif ok {\n\t\t\treturn nil\n\t\t}\n\n\t\ttime.Sleep(waitTime)\n\t}\n\n\treturn fmt.Errorf(\"Retriable function did not reach the expected outcome. Retry attempts: %v. Wait time: %v\", attempts, waitTime)\n}", "func Retry(function func() error, backoffPolicy backoff.Policy) (err error) {\n\tb, cancel := backoffPolicy.Start(context.Background())\n\n\tdefer cancel()\n\tfor {\n\t\tselect {\n\t\tcase <-b.Done():\n\t\t\treturn errors.WrapIf(err, \"all attempts failed\")\n\t\tcase <-b.Next():\n\t\t\terr = function()\n\t\t\tif err == nil {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\tif backoff.IsPermanentError(err) {\n\t\t\t\treturn errors.WrapIf(err, \"permanent error happened during retrying\")\n\t\t\t}\n\t\t}\n\t}\n}", "func Retry(o Operation, b Backoff) error { return RetryNotify(o, b, nil) }", "func Retry[T any](ctx context.Context, log logrus.FieldLogger, f func() (T, error)) (T, error) {\n\tconst idempotent = false\n\tv, err := retry(ctx, log, idempotent, f)\n\treturn v, trace.Wrap(err)\n}", "func Retry(attempts int, interval time.Duration, fn func() error) error {\n\t// @step: give it a go once before jumping in\n\tfor i := 0; i < attempts; i++ {\n\t\tif err := fn(); err == nil {\n\t\t\treturn nil\n\t\t}\n\n\t\tif interval > 0 {\n\t\t\ttime.Sleep(interval)\n\t\t}\n\t}\n\n\treturn errors.New(\"operation failed\")\n}", "func retry(g getter, retries int, delay time.Duration) getter {\n\treturn func(client http.Client, url string) (*http.Response, error) {\n\t\tfor r := 0; ; r++ {\n\t\t\tresponse, err := g(client, url)\n\t\t\tif err == nil || r >= retries {\n\t\t\t\treturn response, err\n\t\t\t}\n\t\t\ttime.Sleep(delay)\n\t\t}\n\t}\n}", "func Retry(\n\tf func() error,\n\tnumberOfRetries int,\n\tonError func(error),\n\tperiod ...time.Duration) {\n\tp := time.Second * 5\n\tif len(period) > 0 && period[0] > 0 {\n\t\tp = period[0]\n\t}\n\tfor numberOfRetries != 0 {\n\t\tif numberOfRetries > 0 {\n\t\t\tnumberOfRetries--\n\t\t}\n\t\tif err := Try(f); err != nil {\n\t\t\tif onError != nil {\n\t\t\t\tonError(err)\n\t\t\t}\n\t\t\tif numberOfRetries != 0 {\n\t\t\t\ttime.Sleep(p)\n\t\t\t}\n\t\t} else {\n\t\t\tbreak\n\t\t}\n\t}\n}", "func Retry(attempts int, delay time.Duration, f func() error) error {\n\tvar err error\n\n\tfor i := 0; i < attempts; i++ {\n\t\terr = f()\n\t\tif err == nil {\n\t\t\tbreak\n\t\t}\n\n\t\tif i < attempts-1 {\n\t\t\ttime.Sleep(delay)\n\t\t}\n\t}\n\n\treturn err\n}", "func (s *Service) Retry(id int64, apiretryrequestmessage *ApiRetryRequestMessage) *RetryCall {\n\tc := &RetryCall{s: s, urlParams_: make(gensupport.URLParams)}\n\tc.id = id\n\tc.apiretryrequestmessage = apiretryrequestmessage\n\treturn c\n}", "func Retry(interval time.Duration, maxRetries int, f ConditionFunc) error {\n\tif maxRetries <= 0 {\n\t\treturn fmt.Errorf(\"maxRetries (%d) should be > 0\", maxRetries)\n\t}\n\ttick := time.NewTicker(interval)\n\tdefer tick.Stop()\n\n\tfor i := 0; ; i++ {\n\t\tok, err := f()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif ok {\n\t\t\treturn nil\n\t\t}\n\t\tif i+1 == maxRetries {\n\t\t\tbreak\n\t\t}\n\t\t<-tick.C\n\t}\n\treturn &RetryError{maxRetries}\n}", "func (r *Retrier) Run(funcToRetry func() error) error {\n\treturn r.RunContext(context.Background(), func(_ context.Context) error {\n\t\treturn funcToRetry()\n\t})\n}", "func Retry(ctx context.Context, times int, delay time.Duration, callback func() error) error {\n\t// Retries n-1 times and ignores failures. Only return early on success.\n\tfor i := 1; (i < times || times < 0) && ctx.Err() == nil; i++ {\n\t\tif err := callback(); err == nil {\n\t\t\treturn nil\n\t\t}\n\n\t\ttime.Sleep(delay)\n\t}\n\n\t// The last error (if any) is the one to keep.\n\treturn callback()\n}", "func (b *ExponentialBackoff) Retry(fn Func) error {\n\twait := time.Duration(b.InitialDelayInterval)\n\tctx := context.Background()\n\tif b.Ctx != nil {\n\t\tctx = b.Ctx\n\t}\n\n\tfor i := 0; i < b.MaxRetryAttempts || b.MaxRetryAttempts == 0; i++ {\n\t\tif i != 0 {\n\t\t\t// Verify if we reached the MaxElapsedTime\n\t\t\tif b.MaxElapsedTime != 0 && time.Since(b.start) > time.Duration(b.MaxElapsedTime) {\n\t\t\t\treturn ErrMaxElapsedTime\n\t\t\t}\n\n\t\t\t// Sleep for the determined duration\n\t\t\tif b.MaxDelayInterval > 0 && wait > time.Duration(b.MaxDelayInterval) {\n\t\t\t\twait = time.Duration(b.MaxDelayInterval)\n\t\t\t}\n\n\t\t\tselect {\n\t\t\tcase <-ctx.Done():\n\t\t\t\treturn ctx.Err()\n\t\t\tcase <-time.After(time.Duration(wait)):\n\t\t\t}\n\n\t\t\t// Exponentially increase that sleep duration\n\t\t\tmultiplier := b.Multiplier\n\t\t\tif multiplier == 0 {\n\t\t\t\tmultiplier = DefaultMultiplier\n\t\t\t}\n\t\t\twait = time.Duration(float64(wait) * multiplier)\n\n\t\t\t// Add a jitter (randomized delay) for the next attempt, to prevent\n\t\t\t// potential collisions\n\t\t\twait = wait + time.Duration(rand.Float64()*float64(wait))\n\t\t} else {\n\t\t\t// Save the current time, in order to measure the total execution time\n\t\t\tb.start = time.Now()\n\t\t}\n\n\t\tif ok, err := fn(i); err != nil || ok {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn ErrMaxRetryAttempts\n}", "func (b *Backoff) Retry(ctx context.Context, retry RetryFunc) error {\n\tif err := b.validate(); err != nil {\n\t\treturn err\n\t}\n\n\tfor attempt := 0; attempt < b.Iterations; attempt++ {\n\t\tif retry(attempt) {\n\t\t\treturn contextDoneOr(ctx, nil)\n\t\t}\n\n\t\tif attempt == 0 {\n\t\t\tcontinue\n\t\t}\n\n\t\tmultiple := float64(uint(1) << (uint(attempt) - 1))\n\n\t\tif b.Jitter != 0 {\n\t\t\tj := (((b.Jitter * rand.Float64()) - (b.Jitter / 2)) / 100)\n\t\t\tmultiple += multiple * j\n\t\t}\n\n\t\tselect {\n\t\tcase <-time.After(b.Coefficient * time.Duration(multiple)):\n\t\t\tcontinue\n\t\tcase <-ctx.Done():\n\t\t\treturn ctx.Err()\n\t\t}\n\t}\n\n\treturn contextDoneOr(ctx, ErrRetriesExhausted)\n}", "func Retry(main func() error, retries int, afterTryFailure func(error) error, beforeRetry func() error) error {\r\n\tvar mainErr error\r\n\r\n\tif main == nil {\r\n\t\treturn fmt.Errorf(\"the main function to try can't be nil\")\r\n\t}\r\n\r\n\tfor i := 0; i <= retries; i++ {\r\n\t\tif i != 0 && beforeRetry != nil {\r\n\t\t\terr := beforeRetry()\r\n\t\t\tif err != nil {\r\n\t\t\t\treturn fmt.Errorf(\"retry before function: %s\", err)\r\n\t\t\t}\r\n\t\t}\r\n\r\n\t\tmainErr = main()\r\n\t\tif mainErr == nil {\r\n\t\t\tbreak\r\n\t\t} else if _, ok := mainErr.(*NoFail); ok {\r\n\t\t\ti--\r\n\t\t\tmainErr = nil\r\n\t\t\tcontinue\r\n\t\t} else if re, ok := mainErr.(*PermFail); ok {\r\n\t\t\tmainErr = re.Err\r\n\t\t\tbreak\r\n\t\t}\r\n\r\n\t\tif afterTryFailure != nil {\r\n\t\t\terr := afterTryFailure(mainErr)\r\n\t\t\tif err != nil {\r\n\t\t\t\treturn fmt.Errorf(\"retry after function: %s\", err)\r\n\t\t\t}\r\n\t\t}\r\n\t\tif i == retries {\r\n\t\t\tbreak\r\n\t\t}\r\n\t\tn, err := RandInt(-500, 1000)\r\n\t\tif err != nil {\r\n\t\t\treturn fmt.Errorf(\"retry rand: %s\", err)\r\n\t\t}\r\n\t\twait := Min(15000, i*750) + n\r\n\t\ttime.Sleep(time.Duration(wait) * time.Millisecond)\r\n\t}\r\n\r\n\treturn mainErr\r\n}", "func (m *Message) Retry() {\n\tm.Trial++\n}", "func Retry(callback func() error, retry int, sleep time.Duration) (err error) {\n\tfor i := 0; i < retry; i++ {\n\t\tif err = callback(); err == nil {\n\t\t\treturn\n\t\t}\n\t\ttime.Sleep(sleep)\n\t}\n\treturn\n}", "func Retry(log logrus.Entry, timeout time.Duration, f func() (interface{}, interface{}, error)) (interface{}, interface{}, error) {\n\n\tvar ret interface{}\n\tvar res interface{}\n\n\top := func() error {\n\t\tvar err error\n\t\tret, res, err = f()\n\t\treturn err\n\t}\n\n\tb := backoff.NewExponentialBackOff()\n\tb.MaxElapsedTime = timeout\n\n\tbackoffErr := backoff.RetryNotify(op, b, func(err error, duration time.Duration) {\n\t\t// Round to a whole number of milliseconds\n\t\tduration /= retryBackoffRoundRatio // Convert nanoseconds to milliseconds\n\t\tduration *= retryBackoffRoundRatio // Convert back so it appears correct\n\n\t\tlog.Errorf(\"error performing operation; retrying in %v: %v\", duration, err)\n\t})\n\n\treturn ret, res, backoffErr\n}", "func Retry(ctx context.Context, retryable func() error, backoffDuration time.Duration) error {\n\tfor {\n\t\tif err := retryable(); err != nil {\n\t\t\tlog.Errorf(\"Failed to execute the retryable function with: %v\", err)\n\t\t\tselect {\n\t\t\tcase <-ctx.Done():\n\t\t\t\treturn fmt.Errorf(\"failed to complete retry function within the timeout\")\n\t\t\tdefault:\n\t\t\t\ttime.Sleep(backoffDuration)\n\t\t\t}\n\t\t} else {\n\t\t\tlog.Infof(\"successfully completed the retry function\")\n\t\t\treturn nil\n\t\t}\n\t}\n}", "func (tx *Tx) Retry() {\n\tpanic(Retry)\n}", "func retry(fn executeJob, attempts int, logger *zap.SugaredLogger) (*common.CrawlResult, error) {\n\tvar (\n\t\tresult *common.CrawlResult\n\t\terr error\n\t)\n\n\tfor i := 0; i < attempts; i++ {\n\t\tresult, err = fn(logger)\n\t\tif err == nil {\n\t\t\treturn result, nil\n\t\t}\n\n\t\tlogger.Debugw(\"failed to execute job\", \"attempt\", i)\n\t\ttime.Sleep(time.Second)\n\t}\n\n\treturn result, err\n}", "func RetryOp(\n\top func() error,\n\tisNonRetryableError func(error) bool,\n) error {\n\tvar err error\n\tfor retryCount := 0; retryCount < DefaultMaxRetryCount; retryCount++ {\n\t\tif err = op(); err == nil {\n\t\t\treturn nil\n\t\t}\n\n\t\tif isNonRetryableError != nil && isNonRetryableError(err) {\n\t\t\treturn err\n\t\t}\n\n\t\tif retryCount < DefaultMaxRetryCount-1 {\n\t\t\ttime.Sleep(DefaultRetryBackoffDuration)\n\t\t}\n\t}\n\n\treturn err\n}", "func (f *Find) Retry(retry driver.RetryMode) *Find {\n\tif f == nil {\n\t\tf = new(Find)\n\t}\n\n\tf.retry = &retry\n\treturn f\n}", "func (n *Sub) retry(uri *model.NotifyURL, msg string, source int) (err error) {\n\tlog.Info(\"Notify.retry do callback url(%v), msg(%s), source(%d)\", uri, msg, source)\n\tfor i := 0; i < _retry; i++ {\n\t\terr = n.clients.Post(context.TODO(), uri, msg)\n\t\tif err != nil {\n\t\t\ttime.Sleep(n.backoff.Backoff(i))\n\t\t\tcontinue\n\t\t} else {\n\t\t\tlog.Info(\"Notify.retry callback success group(%s), topic(%s), retry(%d), msg(%s), source(%d)\",\n\t\t\t\tn.w.Group, n.w.Topic, i, msg, source)\n\t\t\treturn\n\t\t}\n\t}\n\tif err != nil {\n\t\tlog.Error(\"Notify.retry callback error(%v), uri(%s), msg(%s), source(%d)\",\n\t\t\terr, uri, msg, source)\n\t}\n\treturn\n}", "func WithRetry(fn func() (interface{}, error), retries int, delay time.Duration) (interface{}, error) {\n\tvar err error\n\tvar res interface{}\n\tfor count := 0; true; count++ {\n\t\tif delay > 0 {\n\t\t\ttime.Sleep(delay)\n\t\t}\n\t\tif count >= retries {\n\t\t\tbreak\n\t\t}\n\t\tres, err = fn()\n\t\tif err == nil {\n\t\t\treturn res, nil\n\t\t}\n\t\tcontinue\n\t}\n\treturn nil, err\n}", "func retry(ctx context.Context, op backoff.Operation, maxInterval time.Duration, maxElapsedTime time.Duration) error {\n\teb := backoff.NewExponentialBackOff()\n\teb.MaxInterval = maxInterval\n\teb.MaxElapsedTime = maxElapsedTime\n\tif ctx == nil {\n\t\treturn backoff.Retry(op, eb)\n\t}\n\treturn backoff.Retry(op, backoff.WithContext(eb, ctx))\n}", "func retry(backoff Backoff, condition func() error, ignored, continued func(error) bool) error {\n\tvar lastErr error\n\terr := wait.ExponentialBackoff(backoff, func() (bool, error) {\n\t\terr := condition()\n\t\tif err == nil {\n\t\t\treturn true, nil\n\t\t}\n\t\tif ignored != nil && ignored(err) {\n\t\t\treturn true, nil\n\t\t}\n\t\t// record lastErr\n\t\tlastErr = err\n\t\tif continued != nil && continued(err) {\n\t\t\treturn false, nil\n\t\t}\n\t\treturn false, err\n\t})\n\tif err == wait.ErrWaitTimeout {\n\t\terr = lastErr\n\t}\n\treturn err\n}", "func (r *Retrier) RunRetry() error {\n\t// Start signal handler.\n\tsigHandler := signals.NewSignalHandler(10)\n\tgo sigHandler.Register()\n\n\tfinish := make(chan bool, 1)\n\tgo func() {\n\t\tselect {\n\t\tcase <-finish:\n\t\t\treturn\n\t\tcase <-time.After(10 * time.Second):\n\t\t\treturn\n\t\tdefault:\n\t\t\tfor {\n\t\t\t\tif sigHandler.GetState() != 0 {\n\t\t\t\t\tlogger.Critical(\"detected signal. retry failed.\")\n\t\t\t\t\tos.Exit(1)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\n\tfor i := 0; i < r.retries; i++ {\n\t\terr := r.retryable.Try()\n\t\tif err != nil {\n\t\t\tlogger.Info(\"Retryable error: %v\", err)\n\t\t\ttime.Sleep(time.Duration(r.sleepSeconds) * time.Second)\n\t\t\tcontinue\n\t\t}\n\t\tfinish <- true\n\t\treturn nil\n\t}\n\n\tfinish <- true\n\treturn fmt.Errorf(\"unable to succeed at retry after %d attempts at %d seconds\", r.retries, r.sleepSeconds)\n}", "func Retry(attempts int, timeout time.Duration, callback func() error) (err error) {\n\tfor i := 0; i <= attempts-1; i++ {\n\t\terr = callback()\n\t\tif err == nil {\n\t\t\treturn nil\n\t\t}\n\n\t\t//time.Sleep(timeout)\n\t\t<-time.After(timeout)\n\t\tlog.Println(\"retrying...\")\n\t}\n\treturn fmt.Errorf(\"after %d attempts, last error: %s\", attempts, err)\n}", "func (b *Backoff) Retry(ctx context.Context, f func() error) error {\n\t// If the context is already done, don't make any attempts to call f.\n\tif ctx.Err() != nil {\n\t\treturn ctx.Err()\n\t}\n\n\t// Try calling f while the error is retryable and ctx is not done.\n\tfor {\n\t\tif err := f(); IsRetryable(err) {\n\t\t\tselect {\n\t\t\tcase <-time.After(b.Duration()):\n\t\t\t\tcontinue\n\t\t\tcase <-ctx.Done():\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t\treturn nil\n\t}\n}", "func Retry(attempts int, sleep time.Duration, fn func() bool) bool {\n\tif err := fn(); err != true {\n\n\t\tif attempts--; attempts > 0 {\n\t\t\tlog.Warnf(\"retry func error: %s. attemps #%d after %s.\", \"failed\", attempts, sleep)\n\t\t\ttime.Sleep(60 * time.Second)\n\t\t\treturn Retry(attempts, 2*sleep, fn)\n\t\t}\n\n\t\treturn err\n\t}\n\t//IsConnected = true\n\treturn true\n}", "func (c *Client) retry(action string, call func() (*github.Response, error)) (*github.Response, error) {\n\tvar err error\n\tvar resp *github.Response\n\n\tfor retryCount := 0; retryCount <= c.retries; retryCount++ {\n\t\tif resp, err = call(); err == nil {\n\t\t\tc.limitRate(&resp.Rate)\n\t\t\treturn resp, nil\n\t\t}\n\t\tswitch err := err.(type) {\n\t\tcase *github.RateLimitError:\n\t\t\tc.limitRate(&err.Rate)\n\t\tcase *github.TwoFactorAuthError:\n\t\t\treturn resp, err\n\t\tcase *retryAbort:\n\t\t\treturn resp, err\n\t\t}\n\n\t\tif retryCount == c.retries {\n\t\t\treturn resp, err\n\t\t}\n\t\tglog.Errorf(\"error %s: %v. Will retry.\\n\", action, err)\n\t\tc.sleepForAttempt(retryCount)\n\t}\n\treturn resp, err\n}", "func (c Client) doRetry(ctx context.Context, method, uri string, body []byte, result any) error {\n\toperation := func() error {\n\t\treturn c.do(ctx, method, uri, body, result)\n\t}\n\n\tnotify := func(err error, duration time.Duration) {\n\t\tlog.Printf(\"client retries because of %v\", err)\n\t}\n\n\tbo := backoff.NewExponentialBackOff()\n\tbo.InitialInterval = 1 * time.Second\n\n\terr := backoff.RetryNotify(operation, bo, notify)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func Retry(logger *logrusx.Logger, maxWait time.Duration, failAfter time.Duration, f func() error) (err error) {\n\tvar lastStart time.Time\n\terr = errors.New(\"did not connect\")\n\tloopWait := time.Millisecond * 100\n\tretryStart := time.Now().UTC()\n\tfor retryStart.Add(failAfter).After(time.Now().UTC()) {\n\t\tlastStart = time.Now().UTC()\n\t\tif err = f(); err == nil {\n\t\t\treturn nil\n\t\t}\n\n\t\tif lastStart.Add(maxWait * 2).Before(time.Now().UTC()) {\n\t\t\tretryStart = time.Now().UTC()\n\t\t}\n\n\t\tlogger.WithError(err).Infof(\"Retrying in %f seconds...\", loopWait.Seconds())\n\t\ttime.Sleep(loopWait)\n\t\tloopWait = loopWait * time.Duration(int64(2))\n\t\tif loopWait > maxWait {\n\t\t\tloopWait = maxWait\n\t\t}\n\t}\n\treturn err\n}", "func Retrying(operation func() error, retryTimeoutSeconds int, logger zap.Logger) error {\n\tnotify := func(err error, duration time.Duration) {\n\t\tlogger.Warn(\n\t\t\t\"Encountered error during run\",\n\t\t\tzap.Int64(\"duration\", duration.Nanoseconds()/int64(time.Millisecond)),\n\t\t\tzap.Error(err),\n\t\t)\n\t}\n\n\tcustomBackoff := backoff.NewExponentialBackOff()\n\tcustomBackoff.MaxElapsedTime = time.Duration(retryTimeoutSeconds) * time.Second\n\tcustomBackoff.MaxInterval = defaultMaxInterval\n\n\treturn backoff.RetryNotify(operation, customBackoff, notify)\n}", "func retry(attempts int, sleep time.Duration, fn func() error) error {\n\tvar err error\n\tfor i := attempts; i > 0; i-- {\n\t\terr = fn()\n\t\tif err == nil {\n\t\t\tbreak\n\t\t}\n\t\ttime.Sleep(sleep)\n\t}\n\treturn err\n}", "func (s stdlib) RetryInterval(time.Duration) {}", "func Retry(f func() error) error {\n\t// TODO: the retry loop should be configurable.\n\tvar err error\n\tfor i := 0; i < maxRetries; i++ {\n\t\terr = f()\n\t\tif err != nil {\n\t\t\t// No point in re-trying or logging a no-row or not found error.\n\t\t\tif errors.Is(err, sql.ErrNoRows) || api.StatusErrorCheck(err, http.StatusNotFound) {\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\t// Process actual errors.\n\t\t\tif IsRetriableError(err) {\n\t\t\t\tif i == maxRetries {\n\t\t\t\t\tlogger.Warn(\"Database error, giving up\", logger.Ctx{\"attempt\": i, \"err\": err})\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tlogger.Debug(\"Database error, retrying\", logger.Ctx{\"attempt\": i, \"err\": err})\n\t\t\t\ttime.Sleep(jitter.Deviation(nil, 0.8)(100 * time.Millisecond))\n\t\t\t\tcontinue\n\t\t\t} else {\n\t\t\t\tlogger.Debug(\"Database error\", logger.Ctx{\"err\": err})\n\t\t\t}\n\t\t}\n\t\tbreak\n\t}\n\n\treturn err\n}", "func (c *SearchCall) RetryOf(retryOf int64) *SearchCall {\n\tc.urlParams_.Set(\"retry_of\", fmt.Sprint(retryOf))\n\treturn c\n}", "func RetryWithBackoff(fn func() bool) error {\n\tdefaultOpts := Options{\n\t\tBackoff: time.Millisecond * 10,\n\t\tMaxBackoff: time.Minute,\n\t\tConstant: 2,\n\t\tMaxAttempts: 10,\n\t}\n\treturn RetryWithBackoffOptions(defaultOpts, fn)\n}", "func Retries(retries int, maxRepeat int, nonIdempotent bool) RetryPolicy {\n\treturn func(req *http.Request, err error, ctx RoundTripContext) bool {\n\t\tif ctx.Retries() >= retries {\n\t\t\treturn false\n\t\t}\n\t\tif ctx.Exhausted() > maxRepeat {\n\t\t\treturn false\n\t\t}\n\t\tif nonIdempotent || isIdempotent(req.Method) {\n\t\t\treturn true\n\t\t}\n\t\tif ne, ok := err.(*net.OpError); ok {\n\t\t\tif ne.Op == \"dial\" {\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t\treturn false\n\t}\n}", "func Retry(f Func, opts ...Option) error {\n\tif f == nil {\n\t\tpanic(\"f is nil\")\n\t}\n\n\tmaxTries := 3\n\timplies := func(error) bool { return true }\n\n\tfor _, opt := range opts {\n\t\tswitch v := opt.(type) {\n\t\tcase MaxTries:\n\t\t\tif v > 1 {\n\t\t\t\tmaxTries = int(v)\n\t\t\t} else if v >= 0 {\n\t\t\t\tmaxTries = 1\n\t\t\t}\n\t\tcase If:\n\t\t\tif v != nil {\n\t\t\t\timplies = v\n\t\t\t}\n\t\t}\n\t}\n\n\tvar err error\n\n\tfor try := 0; try < maxTries; try++ {\n\t\tif err = f(); err == nil || !implies(err) {\n\t\t\tbreak\n\t\t}\n\t}\n\n\treturn err\n}", "func Retry(ctx context.Context, times int, period time.Duration, fn func() error) error {\n\tvar err error\n\tfor i := 0; i < times; i += 1 {\n\t\terr = fn()\n\t\tif err == nil {\n\t\t\treturn nil\n\t\t}\n\t\tlog.Debugf(\"Attempt %v, result: %v, retry in %v\", i+1, err, period)\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\tlog.Debug(\"Context is closing, return.\")\n\t\t\treturn err\n\t\tcase <-time.After(period):\n\t\t}\n\t}\n\treturn trace.Wrap(err)\n}", "func (message *Message) Retry(err error) {\n\tif message == nil {\n\t\treturn\n\t}\n\n\tmessage.resolve(err)\n}", "func Retry(timeoutSeconds int, sleepSeconds int, try func() error) error {\n\tstart, err := time.Now(), try() // gurantees that it tries once regardless of timeout\n\ttime.Sleep(time.Duration(sleepSeconds) * time.Second) // all try calls are seperated by the sleep period\n\tfor err != nil && time.Since(start) < time.Duration(timeoutSeconds)*time.Second {\n\t\terr = try()\n\t\ttime.Sleep(time.Duration(sleepSeconds) * time.Second)\n\t}\n\treturn err\n}", "func callWithRetries(ctx context.Context, f func() error) error {\n\ttaggedFunc := func() error {\n\t\treturn tagErrIfTransient(f())\n\t}\n\treturn retry.Retry(ctx, transient.Only(retryParams), taggedFunc, nil)\n}", "func (r *Retry) Run(funcToRetry func() (interface{}, error)) (interface{}, error) {\n\treturn r.RunWithContext(context.Background(), func(_ context.Context) (result interface{}, err error) {\n\t\tresult, err = funcToRetry()\n\t\treturn\n\t})\n}", "func retryCall(times int, sleep int64, fn func() (interface{}, error)) (intf interface{}, err error) {\n\tfor i := 0; i < times; i++ {\n\t\tintf, err = fn()\n\t\tif err == nil {\n\t\t\treturn intf, nil\n\t\t}\n\t\ttime.Sleep(time.Millisecond * time.Duration(sleep))\n\t}\n\treturn nil, err\n}", "func Retry(fn RetryFunc) Option {\n\treturn func(o *Options) {\n\t\to.CallOptions.Retry = fn\n\t}\n}", "func Retry(c context.Context, text string, f func(context.Context) error, durations ...time.Duration) error {\n\tdelay := defaultRetryDelay\n\tmaxDelay := defaultMaxDelay\n\n\tswitch len(durations) {\n\tcase 2:\n\t\tmaxDelay = durations[1]\n\t\tif maxDelay == 0 {\n\t\t\tmaxDelay = defaultMaxDelay\n\t\t}\n\t\tfallthrough\n\tcase 1:\n\t\tdelay = durations[0]\n\t\tif delay == 0 {\n\t\t\tdelay = defaultRetryDelay\n\t\t}\n\t}\n\n\tif maxDelay < delay {\n\t\tmaxDelay = delay\n\t}\n\n\tfor {\n\t\terr := f(c)\n\t\tif err == nil {\n\t\t\t// success\n\t\t\treturn nil\n\t\t}\n\n\t\t// Logging at higher log levels should be done in the called function\n\t\tdlog.Debugf(c, \"%s waiting %s before retrying after error: %v\", text, delay.String(), err)\n\n\t\tselect {\n\t\tcase <-c.Done():\n\t\t\treturn err\n\t\tcase <-time.After(delay):\n\t\t}\n\t\tdelay *= 2\n\t\tif delay > maxDelay {\n\t\t\tdelay = maxDelay\n\t\t}\n\t}\n}", "func retry(fn func() error) error {\n\tfor count := 0; count < 10; count++ {\n\t\terr := fn()\n\t\tif err == nil {\n\t\t\treturn nil\n\t\t}\n\n\t\tif retryable(err) {\n\t\t\tcontinue\n\t\t}\n\t\treturn err\n\t}\n\treturn errors.New(\"retry limit exhausted\")\n}", "func RetryN(\n\tctx context.Context,\n\tf func(ctx context.Context) error,\n\tretryable func(err error) bool,\n\tnap time.Duration,\n\tn int,\n) error {\n\tif retryable == nil {\n\t\tretryable = func(_ error) bool { return false }\n\t}\n\n\tif n < 1 {\n\t\tn = 1\n\t}\n\n\tvar err error\n\tfor i := 0; i < n; i++ {\n\t\tif err = f(ctx); err == nil {\n\t\t\treturn nil\n\t\t} else if !errors.Is(err, ErrRetryable) && !retryable(err) {\n\t\t\treturn err\n\t\t}\n\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\treturn err\n\t\tcase <-time.After(nap):\n\t\t}\n\t}\n\n\treturn err\n}", "func (in *Retry) DeepCopy() *Retry {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(Retry)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (c *client) requestRetry(method, path, accept, org string, body interface{}) (*http.Response, error) {\n\treturn c.requestRetryWithContext(context.Background(), method, path, accept, org, body)\n}", "func retryWithExponentialBackoff(opts wait.Backoff, operation func() error) error {\n\n\ti := 0\n\terr := wait.ExponentialBackoff(opts, func() (bool, error) {\n\t\ti++\n\t\tif err := operation(); err != nil {\n\t\t\tif i < opts.Steps {\n\t\t\t\treturn false, nil\n\t\t\t}\n\t\t\treturn false, err\n\t\t}\n\t\treturn true, nil\n\t})\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"action failed after %d attempts\", i)\n\t}\n\treturn nil\n}", "func (rl *Loop) Retry(desc string, err error) error {\n\tif rl.DoRetries {\n\t\treturn rl.RetryForAll(desc, err)\n\t}\n\treturn rl.Exit(desc, err)\n}", "func Retry(ctx context.Context, f func() (*http.Response, error), backoff BackoffStrategy) (*http.Response, error) {\n\tfor {\n\t\tresp, err := f()\n\n\t\tvar status int\n\t\tif resp != nil {\n\t\t\tstatus = resp.StatusCode\n\t\t}\n\n\t\t// Return if we shouldn't retry.\n\t\tpause, retry := backoff.Pause()\n\t\tif !shouldRetry(status, err) || !retry {\n\t\t\treturn resp, err\n\t\t}\n\n\t\t// Ensure the response body is closed, if any.\n\t\tif resp != nil && resp.Body != nil {\n\t\t\tresp.Body.Close()\n\t\t}\n\n\t\t// Pause, but still listen to ctx.Done if context is not nil.\n\t\tvar done <-chan struct{}\n\t\tif ctx != nil {\n\t\t\tdone = ctx.Done()\n\t\t}\n\t\tselect {\n\t\tcase <-done:\n\t\t\treturn nil, ctx.Err()\n\t\tcase <-time.After(pause):\n\t\t}\n\t}\n}", "func retry(attempts int, sleep time.Duration, do func() error) error {\n\tvar err error\n\tfor i := 0; i < attempts; i++ {\n\t\tif err = do(); err == nil {\n\t\t\treturn nil\n\t\t}\n\t\ttime.Sleep(sleep)\n\t\tlog.Infof(\"retrying after error: %v\", err)\n\t}\n\treturn fmt.Errorf(\"failed after %d attempts; last error: %v\", attempts, err)\n}", "func RetryOnErr(\n\ti time.Duration,\n\tn int,\n\tfn func() (interface{}, error),\n) (r interface{}, err error) {\n\ttry := 1\n\tfor range time.Tick(i) {\n\t\tif try > n {\n\t\t\tbreak\n\t\t}\n\n\t\tif r, err = fn(); err == nil {\n\t\t\tbreak\n\t\t}\n\n\t\ttry++\n\t}\n\treturn r, err\n}", "func NewRetry(maxTries int, initialDelay, maxDelay time.Duration) *Retry {\n\tif maxTries <= 0 {\n\t\tmaxTries = DefaultMaxTries\n\t}\n\tif initialDelay <= 0 {\n\t\tinitialDelay = DefaultInitialDelay\n\t}\n\tif maxDelay <= 0 {\n\t\tmaxDelay = DefaultMaxDelay\n\t}\n\treturn &Retry{maxTries, initialDelay, maxDelay}\n}", "func (plm *PLM) retry(packet *Packet, retries int) (ack *Packet, err error) {\n\tplm.Lock()\n\tdefer plm.Unlock()\n\n\tfor err == ErrNak || retries > 0 {\n\t\tack, err = plm.tx(packet, time.Second)\n\t\tif err == nil {\n\t\t\tbreak\n\t\t}\n\t\tretries--\n\t}\n\n\tif err == ErrNak {\n\t\tinsteon.Log.Debugf(\"Retry count exceeded\")\n\t\terr = ErrRetryCountExceeded\n\t}\n\treturn ack, err\n}", "func toRetryFn(retry RetryFn, delay DelayFn) retryFn {\n\treturn func(attempt Attempt) (bool, time.Duration) {\n\t\tif ok := retry(attempt); !ok {\n\t\t\treturn false, 0\n\t\t}\n\t\treturn true, delay(attempt)\n\t}\n}", "func retryFunc(timeout time.Duration, f func() error) error {\n\tfinish := time.After(timeout)\n\tfor {\n\t\terr := f()\n\t\tif err == nil {\n\t\t\treturn nil\n\t\t}\n\t\tlog.Printf(\"Retryable error: %v\", err)\n\n\t\tselect {\n\t\tcase <-finish:\n\t\t\treturn err\n\t\tcase <-time.After(3 * time.Second):\n\t\t}\n\t}\n}", "func (m *ClientWrapper) getFuncRetry(servKey, funcName string) int {\n\tvar configCenter xconfig.ConfigCenter\n\tif m.configCenter != nil {\n\t\tconfigCenter = m.configCenter\n\t} else {\n\t\t// 兼容之前的行为\n\t\tconfigCenter = GetConfigCenter()\n\t}\n\treturn GetFuncRetryInner(configCenter, servKey, funcName)\n}", "func TestRetryDoInteralFunc(t *testing.T) {\n\texec := func() (error, interface{}){\n\t\tr := number.Random(0, 100000000)\n\t\tfmt.Println(\"r is\", r)\n\t\tif r < 100000000 / 2{\n\t\t\treturn errors.New(\"xx\"), nil\n\t\t}\n\t\treturn nil, map[string]string{\"abc\":\"wocao\"}\n\t}\n\terr, res, count := RetryDoInteralTime(exec, 5, 100)\n\tfmt.Printf(\"TestRetryDoInteralFunc error is %s res is %v count is %d\", err, res, count)\n}", "func (request ListFsuDiscoveriesRequest) RetryPolicy() *common.RetryPolicy {\n\treturn request.RequestMetadata.RetryPolicy\n}", "func Retry(ctx context.Context, b Backoff, f func() error, c chan<- error) error {\n\tb.Reset()\n\tvar timer *time.Timer\n\tfor {\n\t\terr := f()\n\t\tif err == nil {\n\t\t\treturn nil\n\t\t}\n\n\t\tnext := b.Next()\n\t\tif next == Stop {\n\t\t\treturn err\n\t\t}\n\n\t\tif timer == nil {\n\t\t\ttimer = time.NewTimer(next)\n\t\t\tdefer timer.Stop()\n\t\t} else {\n\t\t\t// Reset() is only safe to call after the timer has fired and its\n\t\t\t// channel has been drained. At this point in the loop we know that\n\t\t\t// the timer's channel was drained on the last iteration (otherwise\n\t\t\t// we would have exited the loop), so it's safe to call Reset().\n\t\t\ttimer.Reset(next)\n\t\t}\n\n\t\t// If the context has already been canceled, exit immediately rather\n\t\t// than entering the select statement to ensure determinism when the\n\t\t// backoff is zero. If multiple channels in a select statement are ready\n\t\t// when entering the statement, Go will randomly choose one, which is\n\t\t// not the desired behavior here - we always want to stop once the\n\t\t// context gets canceled.\n\t\tif ctx.Err() != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\treturn err\n\t\tcase <-timer.C:\n\t\t\tif c != nil {\n\t\t\t\tc <- err\n\t\t\t}\n\t\t}\n\t}\n}", "func RetryContined(backoff Backoff, condition func() error, continued func(error) bool) error {\n\treturn retry(backoff, condition, nil, continued)\n}", "func (s *SaveRetries) Retried(req *http.Request, attempt int) {\n\t// it is the request attempt x, but the retry attempt is x-1\n\tif attempt > 0 {\n\t\tattempt--\n\t}\n\n\ttable := GetLogDataTable(req)\n\ttable.Core[RetryAttempts] = attempt\n}", "func InitRetryClient() *RetryHTTPClient {\n\trc := rhttp.NewClient()\n\t// Replace default timeout \"0\" for http.client\n\trc.HTTPClient.Timeout = timeout\n\trc.Logger = log.New()\n\trc.RetryMax = defaultRetryMax\n\trc.CheckRetry = defaultRetryPolicy\n\t//rc.Backoff = rhttp.LinearJitterBackoff\n\n\t// Replace default config for http.Transport\n\tt := rc.HTTPClient.Transport.(*http.Transport)\n\tt.MaxConnsPerHost = defaultMaxConnsPerHost\n\tt.MaxIdleConns = defaultMaxIdleConns\n\tt.MaxIdleConnsPerHost = defaultMaxIdleConnsPerHost\n\tt.IdleConnTimeout = defaultIdleConnTimeout\n\t//t.ExpectContinueTimeout = defaultExpectContinueTimeout\n\n\treturn &RetryHTTPClient{rc}\n}", "func Retries(retries int) func(*Config) error {\n\treturn func(c *Config) error {\n\t\tc.Retries = retries\n\t\treturn nil\n\t}\n}", "func (m *MetricsRetryListener) Retried(req *http.Request, attempt int) {\n\tm.retryMetrics.RetriesCounter().With(\"backend\", m.backendName).Add(1)\n}", "func Retry(action Action, strategies ...Strategy) (uint, error) {\n\tfor i := uint(1); ; i++ {\n\t\terr := action()\n\t\tif err == nil {\n\t\t\treturn i, nil\n\t\t}\n\n\t\tfor _, s := range strategies {\n\t\t\tif shouldRetry := s(i, err); !shouldRetry {\n\t\t\t\treturn i, err\n\t\t\t}\n\t\t}\n\t}\n}", "func RetryWithStrategy(strategy RetryStrategy) InterceptorFn {\n\tif strategy.Backoff == nil {\n\t\tstrategy.Backoff = func(r *http.Request, i int) time.Duration { return 0 }\n\t}\n\tif strategy.MaxRetries == nil {\n\t\tstrategy.MaxRetries = func(r *http.Request) int {\n\t\t\tif retryStr := r.Header.Get(\"X-Max-Retries\"); retryStr != \"\" {\n\t\t\t\tif retries, err := strconv.ParseInt(retryStr, 10, 64); err == nil {\n\t\t\t\t\treturn int(retries)\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn 3\n\t\t}\n\t}\n\tif strategy.ShouldRetry == nil {\n\t\tstrategy.ShouldRetry = func(r *http.Request, resp *http.Response, err error) bool {\n\t\t\treturn err != nil || xslice.ContainsInt([]int{500}, resp.StatusCode)\n\t\t}\n\t}\n\treturn func(rt http.RoundTripper) http.RoundTripper {\n\t\treturn RoundTripperFn(func(request *http.Request) (response *http.Response, err error) {\n\t\t\tmaxRetries := strategy.MaxRetries(request)\n\t\t\tfor i := 0; i < maxRetries; i++ {\n\t\t\t\tresponse, err = rt.RoundTrip(request)\n\t\t\t\tif strategy.ShouldRetry(request, response, err) {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\treturn\n\t\t\t}\n\t\t\treturn\n\t\t})\n\t}\n}", "func NewRetryDialer() *RetryDialer {\n\treturn &RetryDialer{\n\t\tDialer: net.Dialer{\n\t\t\tTimeout: DefaultTimeout,\n\t\t\tKeepAlive: DefaultKeepAlive,\n\t\t},\n\t\tRetries: DefaultRetries,\n\t}\n}", "func RetryDuring(attempts int, timeout time.Duration, sleep time.Duration, myfunc func() error) error {\n\tt0 := time.Now()\n\n\tvar err error\n\tfor i := 0; ; i++ {\n\t\tif err = myfunc(); err == nil {\n\t\t\tglog.V(4).Infof(\"[retry-%d/%d] success\", i+1, attempts)\n\t\t\treturn nil\n\t\t}\n\n\t\tif _, skipRetry := err.(*ErrorSkipRetry); skipRetry {\n\t\t\terr = fmt.Errorf(\"failing without retries: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\tglog.V(4).Infof(\"[retry-%d/%d] Warning %v\", i+1, attempts, err)\n\t\tif i >= (attempts - 1) {\n\t\t\tbreak\n\t\t}\n\n\t\tif timeout > 0 {\n\t\t\tif delta := time.Now().Sub(t0); delta > timeout {\n\t\t\t\terr = fmt.Errorf(\"failed after %d attepmts (during %v) last error: %v\", i+1, delta, err)\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t\tif sleep > 0 {\n\t\t\ttime.Sleep(sleep)\n\t\t}\n\t}\n\n\terr = fmt.Errorf(\"failed after %d attepmts, last error: %v\", attempts, err)\n\treturn err\n}", "func RetryWithBackoff(ctx context.Context, iters int, retry RetryFunc) error {\n\tif iters < 2 {\n\t\treturn ErrTooFewIterations\n\t}\n\n\tbos, err := newBackoffSession(ctx, iters)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor i := 0; i < iters; i++ {\n\t\tif err := bos.sleep(ctx, i); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif retry(i) {\n\t\t\treturn contextDoneOr(ctx, nil)\n\t\t}\n\t}\n\n\treturn contextDoneOr(ctx, ErrRetriesExhausted)\n}", "func catchRetry(fn func(*Tx), tx *Tx) (retry bool) {\n\tdefer func() {\n\t\tif r := recover(); r == Retry {\n\t\t\tretry = true\n\t\t} else if r != nil {\n\t\t\tpanic(r)\n\t\t}\n\t}()\n\tfn(tx)\n\treturn\n}", "func (o ApplicationOperationOutput) Retry() ApplicationOperationRetryPtrOutput {\n\treturn o.ApplyT(func(v ApplicationOperation) *ApplicationOperationRetry { return v.Retry }).(ApplicationOperationRetryPtrOutput)\n}", "func NewRetryable(conf RetryableConfig, c *http.Client) *http.Client {\n\tretrier := retryablehttp.NewClient()\n\n\t// default http client\n\tretrier.HTTPClient = &http.Client{Timeout: 30 * time.Second}\n\n\t// overriding http client if exists\n\tif c != nil {\n\t\tretrier.HTTPClient = c\n\t}\n\n\t// no logger applied\n\tretrier.Logger = nil\n\n\tretrier.Backoff = retryablehttp.LinearJitterBackoff\n\tretrier.RetryMax = conf.RetryMax\n\tretrier.RetryWaitMin = conf.RetryWaitMin\n\tretrier.RetryWaitMax = conf.RetryWaitMax\n\tretrier.ErrorHandler = retryablehttp.PassthroughErrorHandler\n\n\tstd := retrier.StandardClient()\n\n\treturn std\n}", "func (c *Authorized) applyRetryWait(retryCount int) {\n\ttime.Sleep(time.Duration(c.retrySettings.MinMsBetweenRetries) * time.Millisecond)\n}", "func retryNTimes(targetFunc retryableFunc, numOfRetries int, delay time.Duration) error {\n\tretryNo := 0\n\tvar err error\n\tvar continueRetrying bool\n\tfor retryNo <= numOfRetries {\n\t\tcontinueRetrying, err = targetFunc()\n\t\tif !continueRetrying {\n\t\t\treturn err\n\t\t}\n\t\t//delay between retries.\n\t\tretryNo++\n\t\ttime.Sleep(delay * time.Duration(retryNo))\n\t}\n\tif err != nil {\n\t\treqErr, ok := err.(RequestError)\n\t\tif ok {\n\t\t\tif reqErr.Description == \"\" {\n\t\t\t\treqErr.Description = \"no error message received from server\"\n\t\t\t}\n\t\t\treqErr.Description = fmt.Sprintf(\"Maximum number of re-tries has been exhausted with error: %s\", reqErr.Description)\n\t\t\treturn reqErr\n\t\t}\n\t\treturn fmt.Errorf(\"maximum number of tries has been exhausted with error: %v\", err)\n\t}\n\treturn errors.New(\"maximum number of tries has been exhausted\")\n}", "func (d *sqlDB) retry(operation retryFunc) (err error) {\n\tfor retries := 0; retries < maxRetries; retries++ {\n\t\tif err = operation(); err != nil {\n\t\t\t// cleanup\n\t\t\td.conn.Close()\n\n\t\t\t// reconnect and retry\n\t\t\ttime.Sleep(1000 * time.Millisecond)\n\t\t\td.connect()\n\t\t\tcontinue\n\t\t}\n\t}\n\treturn\n}", "func LimitRetry(s RetryStrategy, max int) RetryStrategy {\n\treturn &limitedRetry{s: s, max: max}\n}", "func LoaderWithRetry(l LoaderFunc, b BackoffFunc, attempts int) LoaderFunc {\n\treturn func(p string) (io.ReadCloser, error) {\n\t\tvar attempt int\n\t\tfor {\n\t\t\tattempt += 1\n\t\t\ttime.Sleep(b(attempt))\n\t\t\tres, err := l(p)\n\t\t\tif err != nil {\n\t\t\t\tif attempt >= attempts {\n\t\t\t\t\treturn nil, err\n\t\t\t\t}\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\treturn res, err\n\t\t}\n\t}\n}", "func ConnRetryStrategy(s func(nth int) time.Duration) ConnOption {\n\treturn func(c *Conn) error {\n\t\tc.retry = s\n\t\treturn nil\n\t}\n}", "func WithRetry(fn func() Retryer) CallOption {\n\treturn retryerOption(fn)\n}", "func RetryBackoff(retry int, minBackoff, maxBackoff time.Duration) time.Duration {\n\tif retry < 0 {\n\t\tretry = 0\n\t}\n\n\tbackoff := minBackoff << uint(retry)\n\tif backoff > maxBackoff || backoff < minBackoff {\n\t\tbackoff = maxBackoff\n\t}\n\n\tif backoff == 0 {\n\t\treturn 0\n\t}\n\treturn time.Duration(rand.Int63n(int64(backoff)))\n}", "func (b *MaxRetryBackOff) Reset() { b.tries = 0 }", "func (a *ADMResponse) Retry() int {\n\treturn a.RetryAfter\n}" ]
[ "0.75524664", "0.7430092", "0.7286703", "0.7266477", "0.71488553", "0.7112779", "0.7085762", "0.7074702", "0.70640314", "0.7056989", "0.70281065", "0.7000672", "0.6979822", "0.6955388", "0.6935681", "0.6915945", "0.6909987", "0.68883574", "0.6858563", "0.6847784", "0.68230885", "0.6819417", "0.681068", "0.6791429", "0.67777634", "0.67319584", "0.67245024", "0.67160046", "0.6702951", "0.6688751", "0.6665131", "0.66497225", "0.66455346", "0.6631435", "0.6622597", "0.66222334", "0.6613393", "0.6605955", "0.6605415", "0.66001153", "0.65910155", "0.658238", "0.6580703", "0.65430105", "0.6531018", "0.6529569", "0.6504709", "0.6462653", "0.64124274", "0.63991094", "0.6398026", "0.6396025", "0.6389184", "0.63770413", "0.637484", "0.63412696", "0.63183886", "0.6298642", "0.62921184", "0.6245259", "0.624372", "0.62429595", "0.62268966", "0.61965406", "0.61879224", "0.61858165", "0.6181173", "0.6180586", "0.6178125", "0.61754835", "0.6171227", "0.6131831", "0.6103505", "0.60872614", "0.60842806", "0.6079281", "0.6076423", "0.60644287", "0.60587156", "0.6048851", "0.60451174", "0.6032965", "0.6025082", "0.6010889", "0.60091823", "0.60009384", "0.59978795", "0.59805393", "0.597617", "0.59744185", "0.5969995", "0.5967195", "0.59571666", "0.5954853", "0.59484327", "0.59246683", "0.59239453", "0.5921554", "0.5914482", "0.5904675" ]
0.73036975
2
AttachDisk uses the override method AttachDiskFn or the real implementation.
func (c *TestClient) AttachDisk(project, zone, instance string, ad *compute.AttachedDisk) error { if c.AttachDiskFn != nil { return c.AttachDiskFn(project, zone, instance, ad) } return c.client.AttachDisk(project, zone, instance, ad) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (attacher *azureDiskAttacher) Attach(spec *volume.Spec, nodeName types.NodeName) (string, error) {\n\tvolumeSource, err := getVolumeSource(spec)\n\tif err != nil {\n\t\tglog.Warningf(\"failed to get azure disk spec\")\n\t\treturn \"\", err\n\t}\n\tinstanceid, err := attacher.azureProvider.InstanceID(nodeName)\n\tif err != nil {\n\t\tglog.Warningf(\"failed to get azure instance id\")\n\t\treturn \"\", fmt.Errorf(\"failed to get azure instance id for node %q\", nodeName)\n\t}\n\tif ind := strings.LastIndex(instanceid, \"/\"); ind >= 0 {\n\t\tinstanceid = instanceid[(ind + 1):]\n\t}\n\n\tlun, err := attacher.azureProvider.GetDiskLun(volumeSource.DiskName, volumeSource.DataDiskURI, nodeName)\n\tif err == cloudprovider.InstanceNotFound {\n\t\t// Log error and continue with attach\n\t\tglog.Warningf(\n\t\t\t\"Error checking if volume is already attached to current node (%q). Will continue and try attach anyway. err=%v\",\n\t\t\tinstanceid, err)\n\t}\n\n\tif err == nil {\n\t\t// Volume is already attached to node.\n\t\tglog.V(4).Infof(\"Attach operation is successful. volume %q is already attached to node %q at lun %d.\", volumeSource.DiskName, instanceid, lun)\n\t} else {\n\t\tglog.V(4).Infof(\"GetDiskLun returned: %v. Initiating attaching volume %q to node %q.\", err, volumeSource.DataDiskURI, nodeName)\n\t\tgetLunMutex.LockKey(instanceid)\n\t\tdefer getLunMutex.UnlockKey(instanceid)\n\n\t\tlun, err = attacher.azureProvider.GetNextDiskLun(nodeName)\n\t\tif err != nil {\n\t\t\tglog.Warningf(\"no LUN available for instance %q\", nodeName)\n\t\t\treturn \"\", fmt.Errorf(\"all LUNs are used, cannot attach volume %q to instance %q\", volumeSource.DiskName, instanceid)\n\t\t}\n\t\tglog.V(4).Infof(\"Trying to attach volume %q lun %d to node %q.\", volumeSource.DataDiskURI, lun, nodeName)\n\t\terr = attacher.azureProvider.AttachDisk(volumeSource.DiskName, volumeSource.DataDiskURI, nodeName, lun, compute.CachingTypes(*volumeSource.CachingMode))\n\t\tif err == nil {\n\t\t\tglog.V(4).Infof(\"Attach operation successful: volume %q attached to node %q.\", volumeSource.DataDiskURI, nodeName)\n\t\t} else {\n\t\t\tglog.V(2).Infof(\"Attach volume %q to instance %q failed with %v\", volumeSource.DataDiskURI, instanceid, err)\n\t\t\treturn \"\", fmt.Errorf(\"Attach volume %q to instance %q failed with %v\", volumeSource.DiskName, instanceid, err)\n\t\t}\n\t}\n\n\treturn strconv.Itoa(int(lun)), err\n}", "func (d *DefaultDriver) AttachVolume(volumeID string) (string, error) {\n\treturn \"\", &errors.ErrNotSupported{\n\t\tType: \"Function\",\n\t\tOperation: \"CreateVolume()\",\n\t}\n}", "func (r Virtual_Guest) AttachDiskImage(imageId *int) (resp datatypes.Provisioning_Version1_Transaction, err error) {\n\tparams := []interface{}{\n\t\timageId,\n\t}\n\terr = r.Session.DoRequest(\"SoftLayer_Virtual_Guest\", \"attachDiskImage\", params, &r.Options, &resp)\n\treturn\n}", "func (m *Machine) attachDrive(ctx context.Context, dev models.Drive) error {\n\thostPath := StringValue(dev.PathOnHost)\n\tm.logger.Infof(\"Attaching drive %s, slot %s, root %t.\", hostPath, StringValue(dev.DriveID), BoolValue(dev.IsRootDevice))\n\trespNoContent, err := m.client.PutGuestDriveByID(ctx, StringValue(dev.DriveID), &dev)\n\tif err == nil {\n\t\tm.logger.Printf(\"Attached drive %s: %s\", hostPath, respNoContent.Error())\n\t} else {\n\t\tm.logger.Errorf(\"Attach drive failed: %s: %s\", hostPath, err)\n\t}\n\treturn err\n}", "func (p *Tmpfs) Attach(ctx driver.Context, v *types.Volume) error {\n\tctx.Log.Debugf(\"Tmpfs attach volume: %s\", v.Name)\n\tmountPath := v.Path()\n\tsize := v.Size()\n\treqID := v.Option(\"reqID\")\n\tids := v.Option(\"ids\")\n\n\tif ids != \"\" {\n\t\tif !strings.Contains(ids, reqID) {\n\t\t\tids = ids + \",\" + reqID\n\t\t}\n\t} else {\n\t\tids = reqID\n\t}\n\n\tif err := os.MkdirAll(mountPath, 0755); err != nil && !os.IsExist(err) {\n\t\treturn fmt.Errorf(\"error creating %q directory: %v\", mountPath, err)\n\t}\n\n\tif !utils.IsMountpoint(mountPath) {\n\t\terr := syscall.Mount(\"shm\", mountPath, \"tmpfs\",\n\t\t\tuintptr(syscall.MS_NOEXEC|syscall.MS_NOSUID|syscall.MS_NODEV),\n\t\t\tfmt.Sprintf(\"mode=1777,size=%s\", size))\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"mounting shm tmpfs: %s %v\", mountPath, err)\n\t\t}\n\t}\n\n\tv.SetOption(\"ids\", ids)\n\tv.SetOption(\"freeTime\", \"\")\n\n\treturn nil\n}", "func NewCmdDiskAttach() *cobra.Command {\n\tvar async *bool\n\tvar udiskIDs *[]string\n\n\treq := base.BizClient.NewAttachUDiskRequest()\n\tcmd := &cobra.Command{\n\t\tUse: \"attach\",\n\t\tShort: \"Attach udisk instances to an uhost\",\n\t\tLong: \"Attach udisk instances to an uhost\",\n\t\tExample: \"ucloud udisk attach --uhost-id uhost-xxxx --udisk-id bs-xxx1,bs-xxx2\",\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\tfor _, id := range *udiskIDs {\n\t\t\t\tid = base.PickResourceID(id)\n\t\t\t\treq.UDiskId = &id\n\t\t\t\t*req.UHostId = base.PickResourceID(*req.UHostId)\n\t\t\t\tresp, err := base.BizClient.AttachUDisk(req)\n\t\t\t\tif err != nil {\n\t\t\t\t\tbase.HandleError(err)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\ttext := fmt.Sprintf(\"udisk[%s] is attaching to uhost uhost[%s]\", *req.UDiskId, *req.UHostId)\n\t\t\t\tif *async {\n\t\t\t\t\tbase.Cxt.Println(text)\n\t\t\t\t} else {\n\t\t\t\t\tpollDisk(resp.UDiskId, *req.ProjectId, *req.Region, *req.Zone, text, []string{status.DISK_INUSE, status.DISK_FAILED})\n\t\t\t\t}\n\t\t\t}\n\t\t},\n\t}\n\tflags := cmd.Flags()\n\tflags.SortFlags = false\n\treq.UHostId = flags.String(\"uhost-id\", \"\", \"Required. Resource ID of the uhost instance which you want to attach the disk\")\n\tudiskIDs = flags.StringSlice(\"udisk-id\", nil, \"Required. Resource ID of the udisk instances to attach\")\n\treq.ProjectId = flags.String(\"project-id\", base.ConfigInstance.ProjectID, \"Optional. Assign project-id\")\n\treq.Region = flags.String(\"region\", base.ConfigInstance.Region, \"Optional. Assign region\")\n\treq.Zone = flags.String(\"zone\", base.ConfigInstance.Zone, \"Optional. Assign availability zone\")\n\tasync = flags.Bool(\"async\", false, \"Optional. Do not wait for the long-running operation to finish.\")\n\n\tflags.SetFlagValuesFunc(\"udisk-id\", func() []string {\n\t\treturn getDiskList([]string{status.DISK_AVAILABLE}, *req.ProjectId, *req.Region, *req.Zone)\n\t})\n\tflags.SetFlagValuesFunc(\"uhost-id\", func() []string {\n\t\treturn getUhostList([]string{status.HOST_RUNNING, status.HOST_STOPPED}, *req.ProjectId, *req.Region, *req.Zone)\n\t})\n\n\tcmd.MarkFlagRequired(\"uhost-id\")\n\tcmd.MarkFlagRequired(\"udisk-id\")\n\n\treturn cmd\n}", "func (srv *VolumeService) Attach(volumename string, vmname string, path string, format string) error {\n\t// Get volume ID\n\tvolume, err := srv.Get(volumename)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"No volume found with name or id '%s'\", volumename)\n\t}\n\n\t// Get VM ID\n\tvmService := NewVMService(srv.provider)\n\tvm, err := vmService.Get(vmname)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"No VM found with name or id '%s'\", vmname)\n\t}\n\n\tvolatt, err := srv.provider.CreateVolumeAttachment(api.VolumeAttachmentRequest{\n\t\tName: fmt.Sprintf(\"%s-%s\", volume.Name, vm.Name),\n\t\tServerID: vm.ID,\n\t\tVolumeID: volume.ID,\n\t})\n\tif err != nil {\n\t\t// TODO Use more explicit error\n\t\treturn err\n\t}\n\n\t// Create mount point\n\tmountPoint := path\n\tif path == api.DefaultVolumeMountPoint {\n\t\tmountPoint = api.DefaultVolumeMountPoint + volume.Name\n\t}\n\n\tsshConfig, err := srv.provider.GetSSHConfig(vm.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tserver, err := nfs.NewServer(sshConfig)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = server.MountBlockDevice(volatt.Device, mountPoint)\n\n\tif err != nil {\n\t\tsrv.Detach(volumename, vmname)\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (util *PortworxVolumeUtil) AttachVolume(m *portworxVolumeMounter) (string, error) {\n\tdriver, err := util.getPortworxDriver(m.plugin.host, true /*localOnly*/)\n\tif err != nil || driver == nil {\n\t\tglog.Errorf(\"Failed to get portworx driver. Err: %v\", err)\n\t\treturn \"\", err\n\t}\n\n\tdevicePath, err := driver.Attach(m.volName)\n\tif err != nil {\n\t\tglog.Errorf(\"Error attaching Portworx Volume (%v): %v\", m.volName, err)\n\t\treturn \"\", err\n\t}\n\treturn devicePath, nil\n}", "func (a *AWS) AttachVolume(ctx *lepton.Context, instanceName, name string, attachID int) error {\n\tvol, err := a.findVolumeByName(name)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tinstance, err := a.findInstanceByName(instanceName)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdevice := \"\"\n\tif attachID >= 0 {\n\t\tif attachID >= 1 && attachID <= 25 {\n\t\t\tdevice = \"/dev/sd\" + string(rune('a'+attachID))\n\t\t} else {\n\t\t\treturn fmt.Errorf(\"invalid attachment point %d; allowed values: 1-25\", attachID)\n\t\t}\n\t} else {\n\t\t// Look for an unused device name to be assigned to the volume, starting from \"/dev/sdb\"\n\t\tfor deviceLetter := 'b'; deviceLetter <= 'z'; deviceLetter++ {\n\t\t\tname := \"/dev/sd\" + string(deviceLetter)\n\t\t\tnameUsed := false\n\t\t\tfor _, mapping := range instance.BlockDeviceMappings {\n\t\t\t\tif *mapping.DeviceName == name {\n\t\t\t\t\tnameUsed = true\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\tif !nameUsed {\n\t\t\t\tdevice = name\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif device == \"\" {\n\t\t\treturn errors.New(\"no available device names\")\n\t\t}\n\t}\n\n\tinput := &ec2.AttachVolumeInput{\n\t\tDevice: aws.String(device),\n\t\tInstanceId: aws.String(*instance.InstanceId),\n\t\tVolumeId: aws.String(*vol.VolumeId),\n\t}\n\t_, err = a.ec2.AttachVolume(input)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (g *GCEVolumes) AttachVolume(volume *volumes.Volume) error {\n\tvolumeURL := volume.ProviderID\n\n\tvolumeName := lastComponent(volumeURL)\n\n\tattachedDisk := &compute.AttachedDisk{\n\t\tDeviceName: volumeName,\n\t\t// TODO: The k8s GCE provider sets Kind, but this seems wrong. Open an issue?\n\t\t//Kind: disk.Kind,\n\t\tMode: \"READ_WRITE\",\n\t\tSource: volumeURL,\n\t\tType: \"PERSISTENT\",\n\t}\n\n\tattachOp, err := g.compute.Instances.AttachDisk(g.project, g.zone, g.instanceName, attachedDisk).Do()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error attaching disk %q: %v\", volumeName, err)\n\t}\n\n\terr = WaitForOp(g.compute, attachOp)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error waiting for disk attach to complete %q: %v\", volumeName, err)\n\t}\n\n\tdevicePath := \"/dev/disk/by-id/google-\" + volumeName\n\n\t// TODO: Wait for device to appear?\n\n\tvolume.LocalDevice = devicePath\n\n\treturn nil\n}", "func (vpcIks *IksVpcSession) AttachVolume(volumeAttachmentRequest provider.VolumeAttachmentRequest) (*provider.VolumeAttachmentResponse, error) {\n\tvpcIks.Logger.Debug(\"Entry of IksVpcSession.AttachVolume method...\")\n\tdefer vpcIks.Logger.Debug(\"Exit from IksVpcSession.AttachVolume method...\")\n\treturn vpcIks.IksSession.AttachVolume(volumeAttachmentRequest)\n}", "func (s *VolumeListener) Attach(inctx context.Context, in *protocol.VolumeAttachmentRequest) (_ *googleprotobuf.Empty, err error) {\n\tdefer fail.OnExitConvertToGRPCStatus(inctx, &err)\n\tdefer fail.OnExitWrapError(inctx, &err, \"cannot attach volume\")\n\n\tempty := &googleprotobuf.Empty{}\n\tif s == nil {\n\t\treturn empty, fail.InvalidInstanceError()\n\t}\n\tif in == nil {\n\t\treturn empty, fail.InvalidParameterCannotBeNilError(\"in\")\n\t}\n\tif inctx == nil {\n\t\treturn nil, fail.InvalidParameterCannotBeNilError(\"inctx\")\n\t}\n\n\tvolumeRef, _ := srvutils.GetReference(in.GetVolume())\n\tif volumeRef == \"\" {\n\t\treturn empty, fail.InvalidRequestError(\"neither name nor id given as reference for volume\")\n\t}\n\thostRef, _ := srvutils.GetReference(in.GetHost())\n\tif hostRef == \"\" {\n\t\treturn empty, fail.InvalidRequestError(\"neither name nor id given as reference for host\")\n\t}\n\tmountPath := in.GetMountPath()\n\n\tfilesystem := in.GetFormat()\n\tdoNotFormat := in.DoNotFormat\n\tdoNotMount := in.DoNotMount\n\n\tjob, xerr := PrepareJob(inctx, in.GetVolume().GetTenantId(), fmt.Sprintf(\"/volume/%s/host/%s/attach\", volumeRef, hostRef))\n\tif xerr != nil {\n\t\treturn nil, xerr\n\t}\n\tdefer job.Close()\n\n\thandler := VolumeHandler(job)\n\tif xerr = handler.Attach(volumeRef, hostRef, mountPath, filesystem, doNotFormat, doNotMount); xerr != nil {\n\t\treturn empty, xerr\n\t}\n\n\treturn empty, nil\n}", "func (m *Machine) AttachStorage(ctlName string, medium driver.StorageMedium) error {\n\tfmt.Println(\"Attach storage\")\n\treturn nil\n}", "func (c *Controller) Attach(attachRequest k8sresources.FlexVolumeAttachRequest) k8sresources.FlexVolumeResponse {\n\tc.logger.Println(\"controller-attach-start\")\n\tdefer c.logger.Println(\"controller-attach-end\")\n\n\tif attachRequest.Version == k8sresources.KubernetesVersion_1_5 {\n\t\tc.logger.Printf(\"k8s 1.5 attach just returning Success\")\n\t\treturn k8sresources.FlexVolumeResponse{\n\t\t\tStatus: \"Success\",\n\t\t}\n\t}\n\tc.logger.Printf(\"For k8s version 1.6 and higher, Ubiquity just returns NOT supported for Attach API. This might change in the future\")\n\treturn k8sresources.FlexVolumeResponse{\n\t\tStatus: \"Not supported\",\n\t}\n}", "func (a *HetznerVolumes) AttachVolume(volume *volumes.Volume) error {\n\tctx, cancel := context.WithTimeout(context.Background(), time.Minute)\n\tdefer cancel()\n\n\tvolumeID, err := strconv.Atoi(volume.ProviderID)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to convert volume id %q to int: %w\", volume.ProviderID, err)\n\t}\n\n\thetznerVolume, _, err := a.hcloudClient.Volume.GetByID(ctx, volumeID)\n\tif err != nil || hetznerVolume == nil {\n\t\treturn fmt.Errorf(\"failed to get info for volume id %q: %w\", volume.ProviderID, err)\n\t}\n\n\tif hetznerVolume.Server != nil {\n\t\tif hetznerVolume.Server.ID != a.server.ID {\n\t\t\treturn fmt.Errorf(\"found volume %s(%d) attached to a different server: %d\", hetznerVolume.Name, hetznerVolume.ID, hetznerVolume.Server.ID)\n\t\t}\n\n\t\tklog.V(2).Infof(\"Attached volume %s(%d) to the running server\", hetznerVolume.Name, hetznerVolume.ID)\n\n\t\tvolume.LocalDevice = fmt.Sprintf(\"%s%d\", localDevicePrefix, hetznerVolume.ID)\n\t\treturn nil\n\t}\n\n\tklog.V(2).Infof(\"Attaching volume %s(%d) to the running server\", hetznerVolume.Name, hetznerVolume.ID)\n\n\taction, _, err := a.hcloudClient.Volume.Attach(ctx, hetznerVolume, a.server)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to attach volume %s(%d): %w\", hetznerVolume.Name, hetznerVolume.ID, err)\n\t}\n\n\t_, errCh := a.hcloudClient.Action.WatchProgress(ctx, action)\n\tif err := <-errCh; err != nil {\n\t\treturn fmt.Errorf(\"failed to wait for volume to %s(%d) be ready: %w\", hetznerVolume.Name, hetznerVolume.ID, err)\n\t}\n\n\treturn nil\n}", "func (stack *OpenstackVolumes) AttachVolume(volume *volumes.Volume) error {\n\topts := volumeattach.CreateOpts{\n\t\tVolumeID: volume.ProviderID,\n\t}\n\tmc := NewMetricContext(\"volume\", \"attach\")\n\tvolumeAttachment, err := volumeattach.Create(stack.computeClient, stack.meta.ServerID, opts).Extract()\n\tif mc.ObserveRequest(err) != nil {\n\t\treturn fmt.Errorf(\"error attaching volume %s to server %s: %v\", opts.VolumeID, stack.meta.ServerID, err)\n\t}\n\tvolume.LocalDevice = volumeAttachment.Device\n\treturn nil\n}", "func (r *vdm) Attach(volumeName, instanceID string, force bool) (string, error) {\n\tfor _, d := range r.drivers {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"moduleName\": r.rexray.Context,\n\t\t\t\"driverName\": d.Name(),\n\t\t\t\"volumeName\": volumeName,\n\t\t\t\"instanceID\": instanceID,\n\t\t\t\"force\": force}).Info(\"vdm.Attach\")\n\n\t\treturn d.Attach(volumeName, instanceID, force)\n\t}\n\treturn \"\", errors.ErrNoVolumesDetected\n}", "func (driver *Driver) Attach(volumeName, instanceID string) (string, error) {\n\tvolumes, err := driver.sdm.GetVolume(\"\", volumeName)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tswitch {\n\tcase len(volumes) == 0:\n\t\treturn \"\", errors.New(\"No volumes returned by name\")\n\tcase len(volumes) > 1:\n\t\treturn \"\", errors.New(\"Multiple volumes returned by name\")\n\t}\n\n\t_, err = driver.sdm.AttachVolume(true, volumes[0].VolumeID, instanceID)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tvolumes, err = driver.sdm.GetVolume(\"\", volumeName)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn volumes[0].NetworkName, nil\n}", "func (endpoint *MacvtapEndpoint) Attach(h hypervisor) error {\n\tnetworkLogger().WithField(\"endpoint-type\", \"macvtap\").Info(\"Attaching endpoint\")\n\tvar err error\n\n\tendpoint.VMFds, err = createMacvtapFds(endpoint.EndpointProperties.Iface.Index, int(h.hypervisorConfig().NumVCPUs))\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Could not setup macvtap fds %s: %s\", endpoint.EndpointProperties.Iface.Name, err)\n\t}\n\n\tif !h.hypervisorConfig().DisableVhostNet {\n\t\tvhostFds, err := createVhostFds(int(h.hypervisorConfig().NumVCPUs))\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Could not setup vhost fds %s : %s\", endpoint.EndpointProperties.Iface.Name, err)\n\t\t}\n\t\tendpoint.VhostFds = vhostFds\n\t}\n\n\treturn h.addDevice(endpoint, netDev)\n}", "func Mount(disk string, out io.Writer, entry log.FieldLogger) (err error) {\n\tconfig := &config{\n\t\tFieldLogger: entry,\n\t\tdisk: disk,\n\t\tout: out,\n\t}\n\tif err = config.createPhysicalVolume(); err != nil {\n\t\treturn trace.Wrap(err)\n\t}\n\tif err = config.createVolumeGroup(); err != nil {\n\t\treturn trace.Wrap(err)\n\t}\n\tif err = config.setProfileConfiguration(&profileContext{\n\t\tThreshold: defaults.DevicemapperAutoextendThreshold,\n\t\tPercent: defaults.DevicemapperAutoextendStep,\n\t}); err != nil {\n\t\treturn trace.Wrap(err)\n\t}\n\tif err = config.createThinPool(); err != nil {\n\t\treturn trace.Wrap(err)\n\t}\n\tif err = config.applyVolumeProfile(); err != nil {\n\t\treturn trace.Wrap(err)\n\t}\n\treturn nil\n}", "func (endpoint *PhysicalEndpoint) Attach(h hypervisor) error {\n\tnetworkLogger().WithField(\"endpoint-type\", \"physical\").Info(\"Attaching endpoint\")\n\n\t// Unbind physical interface from host driver and bind to vfio\n\t// so that it can be passed to qemu.\n\tif err := bindNICToVFIO(endpoint); err != nil {\n\t\treturn err\n\t}\n\n\t// TODO: use device manager as general device management entrance\n\td := config.VFIODev{\n\t\tBDF: endpoint.BDF,\n\t}\n\n\treturn h.addDevice(d, vfioDev)\n}", "func Attach(c *golangsdk.ServiceClient, opts AttachOpts) (*jobs.Job, error) {\n\tb, err := golangsdk.BuildRequestBody(opts, \"volumeAttachment\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar r jobs.Job\n\t_, err = c.Post(attachURL(c, opts.ServerId), b, &r, &golangsdk.RequestOpts{\n\t\tMoreHeaders: requestOpts.MoreHeaders,\n\t})\n\treturn &r, err\n}", "func (c *TestClient) CreateDisk(project, zone string, d *compute.Disk) error {\n\tif c.CreateDiskFn != nil {\n\t\treturn c.CreateDiskFn(project, zone, d)\n\t}\n\treturn c.client.CreateDisk(project, zone, d)\n}", "func (cl *Client) VolumeAttach(ctx context.Context, vaa *csp.VolumeAttachArgs) (*csp.Volume, error) {\n\tsvc, vid, _ := VolumeIdentifierParse(vaa.VolumeIdentifier)\n\tswitch svc {\n\tcase ServiceGCE:\n\t\treturn cl.gceVolumeAttach(ctx, vaa, vid)\n\t}\n\treturn nil, fmt.Errorf(\"storage type currently unsupported\")\n}", "func (_m *OpenStackMock) WaitDiskAttached(instanceID string, volumeID string) error {\n\tret := _m.Called(instanceID, volumeID)\n\n\tvar r0 error\n\tif rf, ok := ret.Get(0).(func(string, string) error); ok {\n\t\tr0 = rf(instanceID, volumeID)\n\t} else {\n\t\tr0 = ret.Error(0)\n\t}\n\n\treturn r0\n}", "func (c *Core) AttachVolume(id types.VolumeID, extra map[string]string) (*types.Volume, error) {\n\tc.lock.Lock(id.Name)\n\tdefer c.lock.Unlock(id.Name)\n\n\tv, dv, err := c.getVolumeDriver(id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tctx := driver.Contexts()\n\n\t// merge extra to volume spec extra.\n\tfor key, value := range extra {\n\t\tv.Spec.Extra[key] = value\n\t}\n\n\tif d, ok := dv.(driver.AttachDetach); ok {\n\t\tif err := d.Attach(ctx, v); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// update meta info.\n\tif err := c.store.Put(v); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn v, nil\n}", "func (s *StorageActionsServiceOp) Attach(driveID string, dropletID int) (*Action, *Response, error) {\n\trequest := &ActionRequest{\n\t\t\"type\": \"attach\",\n\t\t\"droplet_id\": dropletID,\n\t}\n\treturn s.doAction(driveID, request)\n}", "func (t *Template) AddDisk() *shared.Disk {\n\tdisk := shared.NewDisk()\n\tt.Elements = append(t.Elements, disk)\n\treturn disk\n}", "func AttachVolume(ip, volName, containerName string) (string, error) {\n\tlog.Printf(\"Attaching volume [%s] on VM [%s]\\n\", volName, ip)\n\treturn ssh.InvokeCommand(ip, dockercli.RunContainer+\" -d -v \"+volName+\n\t\t\":/vol1 --name \"+containerName+dockercli.TestContainer)\n}", "func (c *Container) Attach(vAttach, hAttach AttachPoint) *Container {\n\tc.vAttach, c.hAttach = vAttach, hAttach\n\treturn c\n}", "func (c *Compute) Disk(name string) (string, error) {\n\tdisk, err := c.Disks.Get(c.Project, c.Zone, name).Do()\n\tif err == nil {\n\t\tlog.Printf(\"found existing root disk: %q\", disk.SelfLink)\n\t\treturn disk.SelfLink, nil\n\t}\n\tlog.Printf(\"not found, creating new root disk: %q\", name)\n\top, err := c.Disks.Insert(c.Project, c.Zone, &compute.Disk{\n\t\tName: name,\n\t}).SourceImage(*image).Do()\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"disk insert api call failed: %v\", err)\n\t}\n\tif err := c.wait(op); err != nil {\n\t\treturn \"\", fmt.Errorf(\"disk insert operation failed: %v\", err)\n\t}\n\tlog.Printf(\"root disk created: %q\", op.TargetLink)\n\treturn op.TargetLink, nil\n}", "func (s *VarlinkInterface) Attach(ctx context.Context, c VarlinkCall, name_ string, detachKeys_ string, start_ bool) error {\n\treturn c.ReplyMethodNotImplemented(ctx, \"io.podman.Attach\")\n}", "func (endpoint *VirtualEndpoint) Attach(h hypervisor) error {\n\tnetworkLogger().WithField(\"endpoint-type\", \"virtual\").Info(\"Attaching endpoint\")\n\n\tif err := xconnectVMNetwork(endpoint, true, h.hypervisorConfig().NumVCPUs, h.hypervisorConfig().DisableVhostNet); err != nil {\n\t\tnetworkLogger().WithError(err).Error(\"Error bridging virtual endpoint\")\n\t\treturn err\n\t}\n\n\treturn h.addDevice(endpoint, netDev)\n}", "func (c *Client) AttachVolume(ctx context.Context, params *AttachVolumeInput, optFns ...func(*Options)) (*AttachVolumeOutput, error) {\n\tif params == nil {\n\t\tparams = &AttachVolumeInput{}\n\t}\n\n\tresult, metadata, err := c.invokeOperation(ctx, \"AttachVolume\", params, optFns, c.addOperationAttachVolumeMiddlewares)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tout := result.(*AttachVolumeOutput)\n\tout.ResultMetadata = metadata\n\treturn out, nil\n}", "func (a *Amazon) AttachVolume(volumeID, instanceID, devicePath string) error {\n\tif err := a.Client.AttachVolume(volumeID, instanceID, devicePath); err != nil {\n\t\treturn err\n\t}\n\n\tcheckAttaching := func(currentPercentage int) (machinestate.State, error) {\n\t\tvol, err := a.Client.VolumeByID(volumeID)\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\n\t\tif len(vol.Attachments) == 0 {\n\t\t\treturn machinestate.Pending, nil\n\t\t}\n\n\t\tif aws.StringValue(vol.Attachments[0].State) != \"attached\" {\n\t\t\treturn machinestate.Pending, nil\n\t\t}\n\n\t\treturn machinestate.Stopped, nil\n\t}\n\n\tws := waitstate.WaitState{\n\t\tStateFunc: checkAttaching,\n\t\tDesiredState: machinestate.Stopped,\n\t}\n\treturn ws.Wait()\n}", "func (endpoint *BridgedMacvlanEndpoint) Attach(h hypervisor) error {\n\tnetworkLogger().Info(\"Attaching macvlan endpoint\")\n\tif err := xconnectVMNetwork(endpoint, true, h.hypervisorConfig().NumVCPUs, h.hypervisorConfig().DisableVhostNet); err != nil {\n\t\tnetworkLogger().WithError(err).Error(\"Error bridging virtual ep\")\n\t\treturn err\n\t}\n\n\treturn h.addDevice(endpoint, netDev)\n}", "func (r *VmwareMapper) MapDisk(vmSpec *kubevirtv1.VirtualMachine, dv cdiv1.DataVolume) {\n\tname := fmt.Sprintf(\"dv-%v\", dv.Name)\n\tname = utils.EnsureLabelValueLength(name)\n\tvolume := kubevirtv1.Volume{\n\t\tName: name,\n\t\tVolumeSource: kubevirtv1.VolumeSource{\n\t\t\tDataVolume: &kubevirtv1.DataVolumeSource{\n\t\t\t\tName: dv.Name,\n\t\t\t},\n\t\t},\n\t}\n\n\tkubevirtDisk := kubevirtv1.Disk{\n\t\tName: name,\n\t\tDiskDevice: kubevirtv1.DiskDevice{\n\t\t\tDisk: &kubevirtv1.DiskTarget{\n\t\t\t\tBus: busTypeVirtio,\n\t\t\t},\n\t\t},\n\t}\n\n\tvmSpec.Spec.Template.Spec.Volumes = append(vmSpec.Spec.Template.Spec.Volumes, volume)\n\tdisks := append(vmSpec.Spec.Template.Spec.Domain.Devices.Disks, kubevirtDisk)\n\n\t// Since the import controller is iterating over a map of DVs,\n\t// MapDisk gets called for each DV in a nondeterministic order which results\n\t// in the disks being in an arbitrary order. This sort ensure the disks are\n\t// attached in the same order as the devices on the source VM.\n\tsort.Slice(disks, func(i, j int) bool {\n\t\treturn disks[i].Name < disks[j].Name\n\t})\n\tvmSpec.Spec.Template.Spec.Domain.Devices.Disks = disks\n}", "func DiskMount(srcPath string, dstPath string, readonly bool, recursive bool, propagation string, mountOptions []string, fsName string) error {\n\tvar err error\n\n\t// Prepare the mount flags\n\tflags := 0\n\tif readonly {\n\t\tflags |= unix.MS_RDONLY\n\t}\n\n\t// Detect the filesystem\n\tif fsName == \"none\" {\n\t\tflags |= unix.MS_BIND\n\t}\n\n\tif propagation != \"\" {\n\t\tswitch propagation {\n\t\tcase \"private\":\n\t\t\tflags |= unix.MS_PRIVATE\n\t\tcase \"shared\":\n\t\t\tflags |= unix.MS_SHARED\n\t\tcase \"slave\":\n\t\t\tflags |= unix.MS_SLAVE\n\t\tcase \"unbindable\":\n\t\t\tflags |= unix.MS_UNBINDABLE\n\t\tcase \"rprivate\":\n\t\t\tflags |= unix.MS_PRIVATE | unix.MS_REC\n\t\tcase \"rshared\":\n\t\t\tflags |= unix.MS_SHARED | unix.MS_REC\n\t\tcase \"rslave\":\n\t\t\tflags |= unix.MS_SLAVE | unix.MS_REC\n\t\tcase \"runbindable\":\n\t\t\tflags |= unix.MS_UNBINDABLE | unix.MS_REC\n\t\tdefault:\n\t\t\treturn fmt.Errorf(\"Invalid propagation mode %q\", propagation)\n\t\t}\n\t}\n\n\tif recursive {\n\t\tflags |= unix.MS_REC\n\t}\n\n\t// Mount the filesystem\n\terr = unix.Mount(srcPath, dstPath, fsName, uintptr(flags), strings.Join(mountOptions, \",\"))\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Unable to mount %q at %q with filesystem %q: %w\", srcPath, dstPath, fsName, err)\n\t}\n\n\t// Remount bind mounts in readonly mode if requested\n\tif readonly == true && flags&unix.MS_BIND == unix.MS_BIND {\n\t\tflags = unix.MS_RDONLY | unix.MS_BIND | unix.MS_REMOUNT\n\t\terr = unix.Mount(\"\", dstPath, fsName, uintptr(flags), \"\")\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Unable to mount %q in readonly mode: %w\", dstPath, err)\n\t\t}\n\t}\n\n\tflags = unix.MS_REC | unix.MS_SLAVE\n\terr = unix.Mount(\"\", dstPath, \"\", uintptr(flags), \"\")\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Unable to make mount %q private: %w\", dstPath, err)\n\t}\n\n\treturn nil\n}", "func (plugin *azureDataDiskPlugin) NewAttacher() (volume.Attacher, error) {\n\tazure, err := getAzureCloudProvider(plugin.host.GetCloudProvider())\n\tif err != nil {\n\t\tglog.V(4).Infof(\"failed to get azure provider\")\n\t\treturn nil, err\n\t}\n\n\treturn &azureDiskAttacher{\n\t\thost: plugin.host,\n\t\tazureProvider: azure,\n\t}, nil\n}", "func diskAttachedToVM(vmCID VMCID, diskUsers []string) bool {\n\tfor _, v := range diskUsers {\n\t\tif strings.HasSuffix(v, string(vmCID)) {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (f *FakeInstance) AttachPrivateNetwork(_ context.Context, _, _ string) error {\n\tpanic(\"implement me\")\n}", "func (j *AuroraJob) Disk(disk int64) Job {\n\t*j.resources[DISK].DiskMb = disk\n\treturn j\n}", "func attach(c *lxc.Container, o *lxc.AttachOptions) {\n\terr := c.AttachShell(*o)\n\tif err != nil {\n\t\terrorExit(2, err)\n\t}\n}", "func (cl *Client) gceVolumeAttach(ctx context.Context, vaa *csp.VolumeAttachArgs, vid string) (*csp.Volume, error) {\n\tcomputeService, err := cl.getComputeService(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdeviceName := vid\n\tif !strings.HasPrefix(deviceName, nuvoNamePrefix) {\n\t\tdeviceName = nuvoNamePrefix + vid\n\t}\n\tdisk := &compute.AttachedDisk{\n\t\tDeviceName: deviceName,\n\t\tSource: fmt.Sprintf(diskSourceURL, cl.projectID, cl.attrs[AttrZone].Value, deviceName),\n\t}\n\top, err := computeService.Instances().AttachDisk(cl.projectID, cl.attrs[AttrZone].Value, vaa.NodeIdentifier, disk).Context(ctx).Do()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif err = cl.waitForOperation(ctx, op); err != nil {\n\t\treturn nil, err\n\t}\n\tvol, err := cl.vr.gceVolumeGet(ctx, vid)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn vol, nil\n}", "func (f *Fixture) EnableDisk(nSlots int64) {\n\tvar e error\n\tf.Bdev, e = bdev.NewMalloc((1 + nSlots) * 16)\n\tf.noError(e)\n\tf.t.Cleanup(func() { f.Bdev.Close() })\n\n\tf.SpdkTh, e = spdkenv.NewThread()\n\tf.noError(e)\n\tf.t.Cleanup(func() { f.SpdkTh.Close() })\n\tf.noError(ealthread.AllocLaunch(f.SpdkTh))\n\n\tf.DiskStore, e = disk.NewStore(f.Bdev, f.SpdkTh, 16, disk.StoreGetDataGo)\n\tf.noError(e)\n\n\tmin, max := f.DiskStore.SlotRange()\n\tf.DiskAlloc = disk.NewAlloc(min, max, eal.NumaSocket{})\n\tf.t.Cleanup(func() {\n\t\tf.DiskAlloc.Close()\n\t\tf.DiskStore.Close()\n\t})\n\n\te = f.Cs.SetDisk(f.DiskStore, f.DiskAlloc)\n\tf.noError(e)\n}", "func (endpoint *HNSEndpoint) VirtualMachineNICAttach(virtualMachineNICName string) error {\n\toperation := \"VirtualMachineNicAttach\"\n\ttitle := \"hcsshim::HNSEndpoint::\" + operation\n\tlogrus.Debugf(title+\" id=%s\", endpoint.Id)\n\trequestMessage := &EndpointAttachDetachRequest{\n\t\tVirtualNICName: virtualMachineNICName,\n\t\tSystemType: VirtualMachineType,\n\t}\n\tresponse := &EndpointResquestResponse{}\n\n\tjsonString, err := json.Marshal(requestMessage)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn hnsCall(\"POST\", \"/endpoints/\"+endpoint.Id+\"/attach\", string(jsonString), &response)\n}", "func (z *zpoolctl) Attach(ctx context.Context, name string, force bool, properties map[string]string, dev, newDev string) *execute {\n\targs := []string{\"attach\"}\n\tif force {\n\t\targs = append(args, \"-f\")\n\t}\n\tif properties != nil {\n\t\tkv := \"-o \"\n\t\tfor k, v := range properties {\n\t\t\tkv += fmt.Sprintf(\"%s=%s,\", k, v)\n\t\t}\n\t\tkv = strings.TrimSuffix(kv, \",\")\n\t\targs = append(args, kv)\n\t}\n\targs = append(args, name, dev, newDev)\n\treturn &execute{ctx: ctx, name: z.cmd, args: args}\n}", "func (c *TestClient) DetachDisk(project, zone, instance, disk string) error {\n\tif c.DetachDiskFn != nil {\n\t\treturn c.DetachDiskFn(project, zone, instance, disk)\n\t}\n\treturn c.client.DetachDisk(project, zone, instance, disk)\n}", "func (d DobsClient) AttachVolume(ctx Context, volumeID string, dropletID string) (error) {\n\tdropletIDI, err := strconv.Atoi(dropletID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvol, _, err := d.GodoClient.Storage.GetVolume(ctx, volumeID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif len(vol.DropletIDs) > 0 {\n\t\totherDropletID := vol.DropletIDs[0]\n\t\tif otherDropletID == dropletIDI {\n\t\t\tlog.Printf(\"Volume %s already attached to this droplet, skipping attach\\n\", volumeID)\n\t\t\treturn nil\n\t\t}\n\t\t\n\t\treturn fmt.Errorf(\"Volume %s already attached to different droplet %d\", volumeID, otherDropletID)\n\t}\n\n\taction, _, err := d.GodoClient.StorageActions.Attach(ctx, volumeID, dropletIDI)\t\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = d.waitForAction(ctx, volumeID, action)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (s *gceOps) waitForAttach(\n\tdisk *compute.Disk,\n\ttimeout time.Duration,\n) (string, error) {\n\tdevicePath, err := task.DoRetryWithTimeout(\n\t\tfunc() (interface{}, bool, error) {\n\t\t\tdevicePath, err := s.DevicePath(disk.Name)\n\t\t\tif se, ok := err.(*cloudops.StorageError); ok &&\n\t\t\t\tse.Code == cloudops.ErrVolAttachedOnRemoteNode {\n\t\t\t\treturn \"\", false, err\n\t\t\t} else if err != nil {\n\t\t\t\treturn \"\", true, err\n\t\t\t}\n\n\t\t\treturn devicePath, false, nil\n\t\t},\n\t\tcloudops.ProviderOpsTimeout,\n\t\tcloudops.ProviderOpsRetryInterval)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn devicePath.(string), nil\n}", "func (s *CPIO9P) Attach() (p9.File, error) {\n\treturn &CPIO9PFID{fs: s, path: 0}, nil\n}", "func (c *Container) Attach(noStdin bool, keys string, attached chan<- bool) error {\n\tif err := c.syncContainer(); err != nil {\n\t\treturn err\n\t}\n\n\tif c.state.State != ContainerStateCreated &&\n\t\tc.state.State != ContainerStateRunning {\n\t\treturn errors.Wrapf(ErrCtrStateInvalid, \"can only attach to created or running containers\")\n\t}\n\n\t// Check the validity of the provided keys first\n\tvar err error\n\tdetachKeys := []byte{}\n\tif len(keys) > 0 {\n\t\tdetachKeys, err = term.ToBytes(keys)\n\t\tif err != nil {\n\t\t\treturn errors.Wrapf(err, \"invalid detach keys\")\n\t\t}\n\t}\n\n\tresize := make(chan remotecommand.TerminalSize)\n\tdefer close(resize)\n\n\terr = c.attachContainerSocket(resize, noStdin, detachKeys, attached)\n\treturn err\n}", "func (device *BlockDevice) Attach(ctx context.Context, devReceiver api.DeviceReceiver) (err error) {\n\tskip, err := device.bumpAttachCount(true)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif skip {\n\t\treturn nil\n\t}\n\n\t// Increment the block index for the sandbox. This is used to determine the name\n\t// for the block device in the case where the block device is used as container\n\t// rootfs and the predicted block device name needs to be provided to the agent.\n\tindex, err := devReceiver.GetAndSetSandboxBlockIndex()\n\n\tdefer func() {\n\t\tif err != nil {\n\t\t\tdevReceiver.UnsetSandboxBlockIndex(index)\n\t\t\tdevice.bumpAttachCount(false)\n\t\t}\n\t}()\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\thypervisorType := devReceiver.GetHypervisorType()\n\tif hypervisorType == \"acrn\" {\n\t\tdeviceLogger().Debug(\"Special casing for ACRN to increment BlockIndex\")\n\t\tindex = index + 1\n\t}\n\n\tdrive := &config.BlockDrive{\n\t\tFile: device.DeviceInfo.HostPath,\n\t\tFormat: \"raw\",\n\t\tID: utils.MakeNameID(\"drive\", device.DeviceInfo.ID, maxDevIDSize),\n\t\tIndex: index,\n\t\tPmem: device.DeviceInfo.Pmem,\n\t\tReadOnly: device.DeviceInfo.ReadOnly,\n\t}\n\n\tif fs, ok := device.DeviceInfo.DriverOptions[config.FsTypeOpt]; ok {\n\t\tdrive.Format = fs\n\t}\n\n\tcustomOptions := device.DeviceInfo.DriverOptions\n\tif customOptions == nil ||\n\t\tcustomOptions[config.BlockDriverOpt] == config.VirtioSCSI {\n\t\t// User has not chosen a specific block device type\n\t\t// Default to SCSI\n\t\tscsiAddr, err := utils.GetSCSIAddress(index)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tdrive.SCSIAddr = scsiAddr\n\t} else if customOptions[config.BlockDriverOpt] != config.Nvdimm {\n\t\tvar globalIdx int\n\n\t\tswitch customOptions[config.BlockDriverOpt] {\n\t\tcase config.VirtioBlock:\n\t\t\tglobalIdx = index\n\t\tcase config.VirtioBlockCCW:\n\t\t\tglobalIdx = index\n\t\tcase config.VirtioMmio:\n\t\t\t//With firecracker the rootfs for the VM itself\n\t\t\t//sits at /dev/vda and consumes the first index.\n\t\t\t//Longer term block based VM rootfs should be added\n\t\t\t//as a regular block device which eliminates the\n\t\t\t//offset.\n\t\t\t//https://github.com/kata-containers/runtime/issues/1061\n\t\t\tglobalIdx = index + 1\n\t\t}\n\n\t\tdriveName, err := utils.GetVirtDriveName(globalIdx)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tdrive.VirtPath = filepath.Join(\"/dev\", driveName)\n\t}\n\n\tdeviceLogger().WithField(\"device\", device.DeviceInfo.HostPath).WithField(\"VirtPath\", drive.VirtPath).Infof(\"Attaching %s device\", customOptions[config.BlockDriverOpt])\n\tdevice.BlockDrive = drive\n\tif err = devReceiver.HotplugAddDevice(ctx, device, config.DeviceBlock); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (o FioSpecVolumeVolumeSourceOutput) AzureDisk() FioSpecVolumeVolumeSourceAzureDiskPtrOutput {\n\treturn o.ApplyT(func(v FioSpecVolumeVolumeSource) *FioSpecVolumeVolumeSourceAzureDisk { return v.AzureDisk }).(FioSpecVolumeVolumeSourceAzureDiskPtrOutput)\n}", "func (c *Container) Attach(e Entity) error {\n\tif err := c.validAttach(e); err != nil {\n\t\treturn err\n\t}\n\n\tif c.image != nil {\n\t\treturn ErrorMaxAttached\n\t}\n\n\t// entity must be image\n\tc.image = e.(*Image)\n\treturn nil\n}", "func (attacher *azureDiskAttacher) WaitForAttach(spec *volume.Spec, lunStr string, timeout time.Duration) (string, error) {\n\tvolumeSource, err := getVolumeSource(spec)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif len(lunStr) == 0 {\n\t\treturn \"\", fmt.Errorf(\"WaitForAttach failed for Azure disk %q: lun is empty.\", volumeSource.DiskName)\n\t}\n\n\tlun, err := strconv.Atoi(lunStr)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"WaitForAttach: wrong lun %q, err: %v\", lunStr, err)\n\t}\n\tscsiHostRescan(&osIOHandler{})\n\texe := exec.New()\n\tdevicePath := \"\"\n\n\terr = wait.Poll(checkSleepDuration, timeout, func() (bool, error) {\n\t\tglog.V(4).Infof(\"Checking Azure disk %q(lun %s) is attached.\", volumeSource.DiskName, lunStr)\n\t\tif devicePath, err = findDiskByLun(lun, &osIOHandler{}, exe); err == nil {\n\t\t\tif len(devicePath) == 0 {\n\t\t\t\tglog.Warningf(\"cannot find attached Azure disk %q(lun %s) locally.\", volumeSource.DiskName, lunStr)\n\t\t\t\treturn false, fmt.Errorf(\"cannot find attached Azure disk %q(lun %s) locally.\", volumeSource.DiskName, lunStr)\n\t\t\t}\n\t\t\tglog.V(4).Infof(\"Successfully found attached Azure disk %q(lun %s, device path %s).\", volumeSource.DiskName, lunStr, devicePath)\n\t\t\treturn true, nil\n\t\t} else {\n\t\t\t//Log error, if any, and continue checking periodically\n\t\t\tglog.V(4).Infof(\"Error Stat Azure disk (%q) is attached: %v\", volumeSource.DiskName, err)\n\t\t\treturn false, nil\n\t\t}\n\t})\n\treturn devicePath, err\n}", "func (o FioSpecVolumeVolumeSourcePtrOutput) AzureDisk() FioSpecVolumeVolumeSourceAzureDiskPtrOutput {\n\treturn o.ApplyT(func(v *FioSpecVolumeVolumeSource) *FioSpecVolumeVolumeSourceAzureDisk {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.AzureDisk\n\t}).(FioSpecVolumeVolumeSourceAzureDiskPtrOutput)\n}", "func (r *remoteRuntimeService) Attach(ctx context.Context, req *runtimeapi.AttachRequest) (*runtimeapi.AttachResponse, error) {\n\tklog.V(10).InfoS(\"[RemoteRuntimeService] Attach\", \"containerID\", req.ContainerId, \"timeout\", r.timeout)\n\tctx, cancel := context.WithTimeout(ctx, r.timeout)\n\tdefer cancel()\n\n\treturn r.attachV1(ctx, req)\n}", "func (s *Module) DiskWrite(name string, image string) error {\n\tpath, err := s.findDisk(name)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"couldn't find disk with id: %s\", name)\n\t}\n\n\tif !s.isEmptyDisk(path) {\n\t\tlog.Debug().Str(\"disk\", path).Msg(\"disk already has a filesystem. no write\")\n\t\treturn nil\n\t}\n\n\tsource, err := os.Open(image)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to open image\")\n\t}\n\tdefer source.Close()\n\tfile, err := os.OpenFile(path, os.O_WRONLY, 0644)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer file.Close()\n\timgStat, err := source.Stat()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to stat image\")\n\t}\n\tfileStat, err := file.Stat()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to state disk\")\n\t}\n\n\tif imgStat.Size() > fileStat.Size() {\n\t\treturn fmt.Errorf(\"image size is bigger than disk\")\n\t}\n\n\t_, err = io.Copy(file, source)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to write disk image\")\n\t}\n\n\treturn nil\n}", "func (c *Client) Attach(afile *Remote, user, aname string) (*Remote, error) {\n\tfid := c.nextFID()\n\n\tafid := NoFID\n\tif afile != nil {\n\t\tafid = afile.fid\n\t}\n\n\trsp, err := c.Send(&Tattach{\n\t\tFID: fid,\n\t\tAFID: afid,\n\t\tUname: user,\n\t\tAname: aname,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tattach := rsp.(*Rattach)\n\n\treturn &Remote{\n\t\tclient: c,\n\t\tfid: fid,\n\t\tqid: attach.QID,\n\t}, nil\n}", "func (driver *linodeVolumeDriver) ensureVolumeAttached(volumeID int) error {\n\t// TODO: validate whether a volume is in use in a local container\n\n\tapi, err := driver.linodeAPI()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Wait for detachment if already detaching\n\tif err := waitForVolumeNotBusy(api, volumeID); err != nil {\n\t\treturn err\n\t}\n\n\t// Fetch volume\n\tvol, err := api.GetVolume(context.Background(), volumeID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// If volume is already attached, do nothing\n\tif vol.LinodeID != nil && *vol.LinodeID == driver.instanceID {\n\t\treturn nil\n\t}\n\n\t// Forcibly attach the volume if forceAttach is enabled\n\tif forceAttach && vol.LinodeID != nil && *vol.LinodeID != driver.instanceID {\n\t\tif err := detachAndWait(api, volumeID); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn attachAndWait(api, volumeID, driver.instanceID)\n\t}\n\n\t// Throw an error if the instance is not in an attachable state\n\tif vol.LinodeID != nil && *vol.LinodeID != driver.instanceID {\n\t\treturn fmt.Errorf(\"failed to attach volume: volume is currently attached to linode %d\", *vol.LinodeID)\n\t}\n\n\treturn attachAndWait(api, volumeID, driver.instanceID)\n}", "func (d *Delete) AttachCmd(cmd *cobra.Command) {\n\td.AdminGlobal = &AdminGlobal{}\n\td.AdminGlobal.AttachCmd(cmd)\n\n\td.AdminFileInput = &AdminFileInput{}\n\td.AdminFileInput.AttachCmd(cmd)\n}", "func (s *Module) DiskCreate(name string, size gridtypes.Unit) (disk pkg.VDisk, err error) {\n\tpath, err := s.findDisk(name)\n\tif err == nil {\n\t\treturn disk, errors.Wrapf(os.ErrExist, \"disk with id '%s' already exists\", name)\n\t}\n\n\tbase, err := s.diskFindCandidate(size)\n\tif err != nil {\n\t\treturn disk, errors.Wrapf(err, \"failed to find a candidate to host vdisk of size '%d'\", size)\n\t}\n\n\tpath, err = s.safePath(base, name)\n\tif err != nil {\n\t\treturn disk, err\n\t}\n\n\tdefer func() {\n\t\t// clean up disk file if error\n\t\tif err != nil {\n\t\t\tos.RemoveAll(path)\n\t\t}\n\t}()\n\n\tdefer syscall.Sync()\n\n\tvar file *os.File\n\tfile, err = os.Create(path)\n\tif err != nil {\n\t\treturn disk, err\n\t}\n\n\tdefer file.Close()\n\tif err = chattr.SetAttr(file, chattr.FS_NOCOW_FL); err != nil {\n\t\treturn disk, err\n\t}\n\n\tif err = syscall.Fallocate(int(file.Fd()), 0, 0, int64(size)); err != nil {\n\t\treturn disk, errors.Wrap(err, \"failed to truncate disk to size\")\n\t}\n\n\treturn pkg.VDisk{Path: path, Size: int64(size)}, nil\n}", "func (m *Model) Attach(vAttach, hAttach AttachPoint) *Model {\n\tm.vAttach, m.hAttach = vAttach, hAttach\n\treturn m\n}", "func StartAttachCtr(ctx context.Context, ctr *libpod.Container, stdout, stderr, stdin *os.File, detachKeys string, sigProxy bool, startContainer bool) error { //nolint: interfacer\n\tresize := make(chan resize.TerminalSize)\n\n\thaveTerminal := term.IsTerminal(int(os.Stdin.Fd()))\n\n\t// Check if we are attached to a terminal. If we are, generate resize\n\t// events, and set the terminal to raw mode\n\n\tif haveTerminal && ctr.Terminal() {\n\t\tcancel, oldTermState, err := handleTerminalAttach(ctx, resize)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer func() {\n\t\t\tif err := restoreTerminal(oldTermState); err != nil {\n\t\t\t\tlogrus.Errorf(\"Unable to restore terminal: %q\", err)\n\t\t\t}\n\t\t}()\n\t\tdefer cancel()\n\t}\n\n\tstreams := new(define.AttachStreams)\n\tstreams.OutputStream = stdout\n\tstreams.ErrorStream = stderr\n\tstreams.InputStream = bufio.NewReader(stdin)\n\tstreams.AttachOutput = true\n\tstreams.AttachError = true\n\tstreams.AttachInput = true\n\n\tif stdout == nil {\n\t\tlogrus.Debugf(\"Not attaching to stdout\")\n\t\tstreams.AttachOutput = false\n\t}\n\tif stderr == nil {\n\t\tlogrus.Debugf(\"Not attaching to stderr\")\n\t\tstreams.AttachError = false\n\t}\n\tif stdin == nil {\n\t\tlogrus.Debugf(\"Not attaching to stdin\")\n\t\tstreams.AttachInput = false\n\t}\n\n\tif sigProxy {\n\t\t// To prevent a race condition, install the signal handler\n\t\t// before starting/attaching to the container.\n\t\tProxySignals(ctr)\n\t}\n\n\tif !startContainer {\n\t\treturn ctr.Attach(streams, detachKeys, resize)\n\t}\n\n\tattachChan, err := ctr.StartAndAttach(ctx, streams, detachKeys, resize, true)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif stdout == nil && stderr == nil {\n\t\tfmt.Printf(\"%s\\n\", ctr.ID())\n\t}\n\n\terr = <-attachChan\n\tif err != nil {\n\t\treturn fmt.Errorf(\"attaching to container %s: %w\", ctr.ID(), err)\n\t}\n\n\treturn nil\n}", "func (mr *MockStoreProviderMockRecorder) Disk(arg0 interface{}) *gomock.Call {\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"Disk\", reflect.TypeOf((*MockStoreProvider)(nil).Disk), arg0)\n}", "func CreateDisk(form resources.DiskForm) {\n\tdisk := model.Disk{\n\t\tSize: form.Size,\n\t\tLeft: form.Left,\n\t\tComputerID: form.ComputerID,\n\t}\n\tnow := time.Now()\n\tdisk.UpdatedAt = now\n\tdisk.CreatedAt = now\n\tdb.Session().Create(&disk)\n}", "func (a *Client) Attach(params *AttachParams, opts ...ClientOption) (*AttachCreated, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewAttachParams()\n\t}\n\top := &runtime.ClientOperation{\n\t\tID: \"attach\",\n\t\tMethod: \"POST\",\n\t\tPathPattern: \"/attach\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &AttachReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t}\n\tfor _, opt := range opts {\n\t\topt(op)\n\t}\n\n\tresult, err := a.transport.Submit(op)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*AttachCreated)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\tunexpectedSuccess := result.(*AttachDefault)\n\treturn nil, runtime.NewAPIError(\"unexpected success response: content available as default response in error\", unexpectedSuccess, unexpectedSuccess.Code())\n}", "func (c *UHostClient) NewResizeAttachedDiskRequest() *ResizeAttachedDiskRequest {\n\treq := &ResizeAttachedDiskRequest{}\n\n\t// setup request with client config\n\tc.Client.SetupRequest(req)\n\n\t// setup retryable with default retry policy (retry for non-create action and common error)\n\treq.SetRetryable(true)\n\treturn req\n}", "func AttachStorage(storageID, deviceID string) error {\n\tclient, err := NewExtPacketClient()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\trequest := &extpackngo.AttachStorageRequest{\n\t\tDeviceID: deviceID,\n\t}\n\n\t_, e := client.Storages.Attach(storageID, request)\n\treturn e\n}", "func (o AttachedDiskInitializeParamsOutput) DiskName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v AttachedDiskInitializeParams) *string { return v.DiskName }).(pulumi.StringPtrOutput)\n}", "func (endpoint *VhostUserEndpoint) Attach(h hypervisor) error {\n\tnetworkLogger().WithField(\"endpoint-type\", \"vhostuser\").Info(\"Attaching endpoint\")\n\n\t// Generate a unique ID to be used for hypervisor commandline fields\n\trandBytes, err := utils.GenerateRandomBytes(8)\n\tif err != nil {\n\t\treturn err\n\t}\n\tid := hex.EncodeToString(randBytes)\n\n\td := config.VhostUserDeviceAttrs{\n\t\tDevID: id,\n\t\tSocketPath: endpoint.SocketPath,\n\t\tMacAddress: endpoint.HardAddr,\n\t\tType: config.VhostUserNet,\n\t}\n\n\treturn h.addDevice(d, vhostuserDev)\n}", "func (o IopingSpecVolumeVolumeSourceOutput) AzureDisk() IopingSpecVolumeVolumeSourceAzureDiskPtrOutput {\n\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSource) *IopingSpecVolumeVolumeSourceAzureDisk { return v.AzureDisk }).(IopingSpecVolumeVolumeSourceAzureDiskPtrOutput)\n}", "func Attach(ctx context.Context, client *containerd.Client, req string, options types.ContainerAttachOptions) error {\n\t// Find the container.\n\tvar container containerd.Container\n\twalker := &containerwalker.ContainerWalker{\n\t\tClient: client,\n\t\tOnFound: func(ctx context.Context, found containerwalker.Found) error {\n\t\t\tcontainer = found.Container\n\t\t\treturn nil\n\t\t},\n\t}\n\tn, err := walker.Walk(ctx, req)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error when trying to find the container: %w\", err)\n\t}\n\tif n == 0 {\n\t\treturn fmt.Errorf(\"no container is found given the string: %s\", req)\n\t} else if n > 1 {\n\t\treturn fmt.Errorf(\"more than one containers are found given the string: %s\", req)\n\t}\n\n\t// Attach to the container.\n\tvar task containerd.Task\n\tdetachC := make(chan struct{})\n\tspec, err := container.Spec(ctx)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to get the OCI runtime spec for the container: %w\", err)\n\t}\n\tvar (\n\t\topt cio.Opt\n\t\tcon console.Console\n\t)\n\tif spec.Process.Terminal {\n\t\tcon = console.Current()\n\t\tdefer con.Reset()\n\t\tif err := con.SetRaw(); err != nil {\n\t\t\treturn fmt.Errorf(\"failed to set the console to raw mode: %w\", err)\n\t\t}\n\t\tcloser := func() {\n\t\t\tdetachC <- struct{}{}\n\t\t\t// task will be set by container.Task later.\n\t\t\t//\n\t\t\t// We cannot use container.Task(ctx, cio.Load) to get the IO here\n\t\t\t// because the `cancel` field of the returned `*cio` is nil. [1]\n\t\t\t//\n\t\t\t// [1] https://github.com/containerd/containerd/blob/8f756bc8c26465bd93e78d9cd42082b66f276e10/cio/io.go#L358-L359\n\t\t\tio := task.IO()\n\t\t\tif io == nil {\n\t\t\t\tlogrus.Errorf(\"got a nil io\")\n\t\t\t\treturn\n\t\t\t}\n\t\t\tio.Cancel()\n\t\t}\n\t\tin, err := consoleutil.NewDetachableStdin(con, options.DetachKeys, closer)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\topt = cio.WithStreams(in, con, nil)\n\t} else {\n\t\topt = cio.WithStreams(options.Stdin, options.Stdout, options.Stderr)\n\t}\n\ttask, err = container.Task(ctx, cio.NewAttach(opt))\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to attach to the container: %w\", err)\n\t}\n\tif spec.Process.Terminal {\n\t\tif err := consoleutil.HandleConsoleResize(ctx, task, con); err != nil {\n\t\t\tlogrus.WithError(err).Error(\"console resize\")\n\t\t}\n\t}\n\tsigC := signalutil.ForwardAllSignals(ctx, task)\n\tdefer signalutil.StopCatch(sigC)\n\n\t// Wait for the container to exit.\n\tstatusC, err := task.Wait(ctx)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to init an async wait for the container to exit: %w\", err)\n\t}\n\tselect {\n\t// io.Wait() would return when either 1) the user detaches from the container OR 2) the container is about to exit.\n\t//\n\t// If we replace the `select` block with io.Wait() and\n\t// directly use task.Status() to check the status of the container after io.Wait() returns,\n\t// it can still be running even though the container is about to exit (somehow especially for Windows).\n\t//\n\t// As a result, we need a separate detachC to distinguish from the 2 cases mentioned above.\n\tcase <-detachC:\n\t\tio := task.IO()\n\t\tif io == nil {\n\t\t\treturn errors.New(\"got a nil IO from the task\")\n\t\t}\n\t\tio.Wait()\n\tcase status := <-statusC:\n\t\tcode, _, err := status.Result()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif code != 0 {\n\t\t\treturn errutil.NewExitCoderErr(int(code))\n\t\t}\n\t}\n\treturn nil\n}", "func (f *FakeInstance) AttachVPC(_ context.Context, _, _ string) error {\n\tpanic(\"implement me\")\n}", "func addFeatureToDisk(workflow *daisy.Workflow, feature string, diskIndex int) {\n\tdisk := getDisk(workflow, diskIndex)\n\tdisk.GuestOsFeatures = append(disk.GuestOsFeatures, &compute.GuestOsFeature{\n\t\tType: feature,\n\t})\n}", "func (o AttachedDiskInitializeParamsPtrOutput) DiskName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *AttachedDiskInitializeParams) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.DiskName\n\t}).(pulumi.StringPtrOutput)\n}", "func (a *Apply) AttachCmd(cmd *cobra.Command) {\n\ta.AdminGlobal = &AdminGlobal{}\n\ta.AdminGlobal.AttachCmd(cmd)\n\n\ta.AdminFileInput = &AdminFileInput{}\n\ta.AdminFileInput.AttachCmd(cmd)\n}", "func (c *TestClient) ResizeDisk(project, zone, disk string, drr *compute.DisksResizeRequest) error {\n\tif c.ResizeDiskFn != nil {\n\t\treturn c.ResizeDiskFn(project, zone, disk, drr)\n\t}\n\treturn c.client.ResizeDisk(project, zone, disk, drr)\n}", "func (o IopingSpecVolumeVolumeSourcePtrOutput) AzureDisk() IopingSpecVolumeVolumeSourceAzureDiskPtrOutput {\n\treturn o.ApplyT(func(v *IopingSpecVolumeVolumeSource) *IopingSpecVolumeVolumeSourceAzureDisk {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.AzureDisk\n\t}).(IopingSpecVolumeVolumeSourceAzureDiskPtrOutput)\n}", "func (i *InstanceServiceHandler) AttachPrivateNetwork(ctx context.Context, instanceID, networkID string) error {\n\turi := fmt.Sprintf(\"%s/%s/private-networks/attach\", instancePath, instanceID)\n\tbody := RequestBody{\"network_id\": networkID}\n\n\treq, err := i.client.NewRequest(ctx, http.MethodPost, uri, body)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn i.client.DoWithContext(ctx, req, nil)\n}", "func (mr *MockServiceMockRecorder) AttachStorage(ctx, r interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"AttachStorage\", reflect.TypeOf((*MockService)(nil).AttachStorage), ctx, r)\n}", "func (s *SceneUtils) Attach(child, scene, parent float64) *SceneUtils {\n\ts.p.Call(\"attach\", child, scene, parent)\n\treturn s\n}", "func (attacher *azureDiskAttacher) MountDevice(spec *volume.Spec, devicePath string, deviceMountPath string) error {\n\tmounter := attacher.host.GetMounter()\n\tnotMnt, err := mounter.IsLikelyNotMountPoint(deviceMountPath)\n\tif err != nil {\n\t\tif os.IsNotExist(err) {\n\t\t\tif err := os.MkdirAll(deviceMountPath, 0750); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tnotMnt = true\n\t\t} else {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvolumeSource, err := getVolumeSource(spec)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\toptions := []string{}\n\tif spec.ReadOnly {\n\t\toptions = append(options, \"ro\")\n\t}\n\tif notMnt {\n\t\tdiskMounter := &mount.SafeFormatAndMount{Interface: mounter, Runner: exec.New()}\n\t\tmountOptions := volume.MountOptionFromSpec(spec, options...)\n\t\terr = diskMounter.FormatAndMount(devicePath, deviceMountPath, *volumeSource.FSType, mountOptions)\n\t\tif err != nil {\n\t\t\tos.Remove(deviceMountPath)\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (endpoint *VirtualEndpoint) HotAttach(h hypervisor) error {\n\tnetworkLogger().Info(\"Hot attaching virtual endpoint\")\n\tif err := xconnectVMNetwork(endpoint, true, h.hypervisorConfig().NumVCPUs, h.hypervisorConfig().DisableVhostNet); err != nil {\n\t\tnetworkLogger().WithError(err).Error(\"Error bridging virtual ep\")\n\t\treturn err\n\t}\n\n\tif _, err := h.hotplugAddDevice(endpoint, netDev); err != nil {\n\t\tnetworkLogger().WithError(err).Error(\"Error attach virtual ep\")\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *Controller) Mount(mountRequest k8sresources.FlexVolumeMountRequest) k8sresources.FlexVolumeResponse {\n\tc.logger.Println(\"controller-mount-start\")\n\tdefer c.logger.Println(\"controller-mount-end\")\n\tc.logger.Println(fmt.Sprintf(\"mountRequest [%#v]\", mountRequest))\n\tvar lnPath string\n\tattachRequest := resources.AttachRequest{Name: mountRequest.MountDevice, Host: getHost()}\n\tmountedPath, err := c.Client.Attach(attachRequest)\n\n\tif err != nil {\n\t\tmsg := fmt.Sprintf(\"Failed to mount volume [%s], Error: %#v\", mountRequest.MountDevice, err)\n\t\tc.logger.Println(msg)\n\t\treturn k8sresources.FlexVolumeResponse{\n\t\t\tStatus: \"Failure\",\n\t\t\tMessage: msg,\n\t\t}\n\t}\n\tif mountRequest.Version == k8sresources.KubernetesVersion_1_5 {\n\t\t//For k8s 1.5, by the time we do the attach/mount, the mountDir (MountPath) is not created trying to do mount and ln will fail because the dir is not found, so we need to create the directory before continuing\n\t\tdir := filepath.Dir(mountRequest.MountPath)\n\t\tc.logger.Printf(\"mountrequest.MountPath %s\", mountRequest.MountPath)\n\t\tlnPath = mountRequest.MountPath\n\t\tk8sRequiredMountPoint := path.Join(mountRequest.MountPath, mountRequest.MountDevice)\n\t\tif _, err = os.Stat(k8sRequiredMountPoint); err != nil {\n\t\t\tif os.IsNotExist(err) {\n\n\t\t\t\tc.logger.Printf(\"creating volume directory %s\", dir)\n\t\t\t\terr = os.MkdirAll(dir, 0777)\n\t\t\t\tif err != nil && !os.IsExist(err) {\n\t\t\t\t\tmsg := fmt.Sprintf(\"Failed creating volume directory %#v\", err)\n\t\t\t\t\tc.logger.Println(msg)\n\n\t\t\t\t\treturn k8sresources.FlexVolumeResponse{\n\t\t\t\t\t\tStatus: \"Failure\",\n\t\t\t\t\t\tMessage: msg,\n\t\t\t\t\t}\n\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\t// For k8s 1.6 and later kubelet creates a folder as the MountPath, including the volume name, whenwe try to create the symlink this will fail because the same name exists. This is why we need to remove it before continuing.\n\t} else {\n\t\tubiquityMountPrefix := fmt.Sprintf(resources.PathToMountUbiquityBlockDevices, \"\")\n\t\tif strings.HasPrefix(mountedPath, ubiquityMountPrefix) {\n\t\t\tlnPath = mountRequest.MountPath\n\t\t} else {\n\t\t\tlnPath, _ = path.Split(mountRequest.MountPath)\n\t\t}\n\t\tc.logger.Printf(\"removing folder %s\", mountRequest.MountPath)\n\n\t\terr = os.Remove(mountRequest.MountPath)\n\t\tif err != nil && !os.IsExist(err) {\n\t\t\tmsg := fmt.Sprintf(\"Failed removing existing volume directory %#v\", err)\n\t\t\tc.logger.Println(msg)\n\n\t\t\treturn k8sresources.FlexVolumeResponse{\n\t\t\t\tStatus: \"Failure\",\n\t\t\t\tMessage: msg,\n\t\t\t}\n\n\t\t}\n\n\t}\n\tsymLinkCommand := \"/bin/ln\"\n\targs := []string{\"-s\", mountedPath, lnPath}\n\tc.logger.Printf(fmt.Sprintf(\"creating slink from %s -> %s\", mountedPath, lnPath))\n\n\tvar stderr bytes.Buffer\n\tcmd := exec.Command(symLinkCommand, args...)\n\tcmd.Stderr = &stderr\n\n\terr = cmd.Run()\n\tif err != nil {\n\t\tmsg := fmt.Sprintf(\"Controller: mount failed to symlink %#v\", stderr.String())\n\t\tc.logger.Println(msg)\n\t\treturn k8sresources.FlexVolumeResponse{\n\t\t\tStatus: \"Failure\",\n\t\t\tMessage: msg,\n\t\t}\n\n\t}\n\tmsg := fmt.Sprintf(\"Volume mounted successfully to %s\", mountedPath)\n\tc.logger.Println(msg)\n\n\treturn k8sresources.FlexVolumeResponse{\n\t\tStatus: \"Success\",\n\t\tMessage: msg,\n\t}\n}", "func (f *Find) AllowDiskUse(allowDiskUse bool) *Find {\n\tif f == nil {\n\t\tf = new(Find)\n\t}\n\n\tf.allowDiskUse = &allowDiskUse\n\treturn f\n}", "func doAttachVolumeCommand(t *testing.T, fail bool) (client *testutil.SsntpTestClient, tenant string, volume string) {\n\tvar reason payloads.StartFailureReason\n\n\tclient, instances := testStartWorkload(t, 1, false, reason)\n\n\ttenantID := instances[0].TenantID\n\n\tsendStatsCmd(client, t)\n\n\tdata := addTestBlockDevice(t, tenantID)\n\n\tserverCh := server.AddCmdChan(ssntp.AttachVolume)\n\tagentCh := client.AddCmdChan(ssntp.AttachVolume)\n\tvar serverErrorCh chan testutil.Result\n\tvar controllerCh chan struct{}\n\n\tif fail == true {\n\t\tserverErrorCh = server.AddErrorChan(ssntp.AttachVolumeFailure)\n\t\tcontrollerCh = wrappedClient.addErrorChan(ssntp.AttachVolumeFailure)\n\t\tclient.AttachFail = true\n\t\tclient.AttachVolumeFailReason = payloads.AttachVolumeAlreadyAttached\n\n\t\tdefer func() {\n\t\t\tclient.AttachFail = false\n\t\t\tclient.AttachVolumeFailReason = \"\"\n\t\t}()\n\t}\n\n\terr := ctl.AttachVolume(tenantID, data.ID, instances[0].ID, \"\")\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tresult, err := server.GetCmdChanResult(serverCh, ssntp.AttachVolume)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tif result.InstanceUUID != instances[0].ID ||\n\t\tresult.NodeUUID != client.UUID ||\n\t\tresult.VolumeUUID != data.ID {\n\t\tt.Fatalf(\"expected %s %s %s, got %s %s %s\", instances[0].ID, client.UUID, data.ID, result.InstanceUUID, result.NodeUUID, result.VolumeUUID)\n\t}\n\n\tif fail == true {\n\t\t_, err = client.GetCmdChanResult(agentCh, ssntp.AttachVolume)\n\t\tif err == nil {\n\t\t\tt.Fatal(\"Success when Failure expected\")\n\t\t}\n\n\t\t_, err = server.GetErrorChanResult(serverErrorCh, ssntp.AttachVolumeFailure)\n\t\tif err != nil {\n\t\t\tt.Fatal(err)\n\t\t}\n\n\t\terr = wrappedClient.getErrorChan(controllerCh, ssntp.AttachVolumeFailure)\n\t\tif err != nil {\n\t\t\tt.Fatal(err)\n\t\t}\n\n\t\t// at this point, the state of the block device should\n\t\t// be set back to available.\n\t\tdata2, err := ctl.ds.GetBlockDevice(data.ID)\n\t\tif err != nil {\n\t\t\tt.Fatal(err)\n\t\t}\n\n\t\tif data2.State != types.Available {\n\t\t\tt.Fatalf(\"block device state not updated\")\n\t\t}\n\t} else {\n\t\t_, err = client.GetCmdChanResult(agentCh, ssntp.AttachVolume)\n\t\tif err != nil {\n\t\t\tt.Fatal(err)\n\t\t}\n\t}\n\n\treturn client, tenantID, data.ID\n}", "func (a *DefaultApiService) VmAddDiskPut(ctx _context.Context, diskConfig DiskConfig) (PciDeviceInfo, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodPut\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue PciDeviceInfo\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/vm.add-disk\"\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{\"application/json\"}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &diskConfig\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func (c *clustermgrClient) AddMigratingDisk(ctx context.Context, value *MigratingDiskMeta) (err error) {\n\tvalue.Ctime = time.Now().String()\n\treturn c.setTask(ctx, value.ID(), value)\n}", "func (s *StackEbrc) CreateVolumeAttachment(request abstract.VolumeAttachmentRequest) (string, fail.Error) {\n\tlogrus.Debugf(\">>> stacks.ebrc::CreateVolumeAttachment(%s)\", request.Name)\n\tdefer logrus.Debugf(\"<<< stacks.ebrc::CreateVolumeAttachment(%s)\", request.Name)\n\n\tvm, err := s.findVMByID(request.HostID)\n\tif err != nil || utils.IsEmpty(vm) {\n\t\treturn \"\", fail.Wrap(err, fmt.Sprintf(\"Error creating attachment, vm empty\"))\n\t}\n\n\tdisk, err := s.findDiskByID(request.VolumeID)\n\tif err != nil || utils.IsEmpty(disk) {\n\t\treturn \"\", fail.Wrap(err, fmt.Sprintf(\"Error creating attachment, disk empty\"))\n\t}\n\n\tattask, err := vm.AttachDisk(&types.DiskAttachOrDetachParams{Disk: &types.Reference{HREF: disk.Disk.HREF}})\n\tif err != nil {\n\t\treturn \"\", fail.Wrap(err, fmt.Sprintf(\"Error creating attachment\"))\n\t}\n\n\terr = attask.WaitTaskCompletion()\n\tif err != nil {\n\t\treturn \"\", fail.Wrap(err, fmt.Sprintf(\"Error creating attachment\"))\n\t}\n\n\treturn getAttachmentID(request.HostID, request.VolumeID), nil\n}", "func (ds *dockerService) Attach(req *runtimeapi.AttachRequest) (*runtimeapi.AttachResponse, error) {\n\tif ds.streamingServer == nil {\n\t\treturn nil, streaming.ErrorStreamingDisabled(\"attach\")\n\t}\n\t_, err := checkContainerStatus(ds.client, req.GetContainerId())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn ds.streamingServer.GetAttach(req)\n}", "func (c *UPHostClient) NewResizePHostAttachedDiskRequest() *ResizePHostAttachedDiskRequest {\n\treq := &ResizePHostAttachedDiskRequest{}\n\n\t// setup request with client config\n\tc.Client.SetupRequest(req)\n\n\t// setup retryable with default retry policy (retry for non-create action and common error)\n\treq.SetRetryable(true)\n\treturn req\n}", "func (o *snapshotter) attachAndMountBlockDevice(ctx context.Context, snID string, snKey string, writable bool) (retErr error) {\n\tif err := lookup(o.tgtTargetMountpoint(snID)); err == nil {\n\t\treturn nil\n\t}\n\n\t// If the target already exists, it won't be processed, see man TGT-ADMIN(8)\n\ttargetConfPath := o.tgtTargetConfPath(snID, snKey)\n\tout, err := exec.CommandContext(ctx, \"tgt-admin\", \"-e\", \"-c\", targetConfPath).CombinedOutput()\n\tif err != nil {\n\t\t// read the init-debug.log for readable\n\t\tdebugLogPath := o.tgtOverlayBDInitDebuglogPath(snID)\n\t\tif data, derr := ioutil.ReadFile(debugLogPath); derr == nil {\n\t\t\treturn errors.Wrapf(err, \"failed to create target by tgt-admin: %s, more detail in %s\", out, data)\n\t\t}\n\t\treturn errors.Wrapf(err, \"failed to create target by tgt-admin: %s\", out)\n\t}\n\n\ttargetIqn := o.tgtTargetIqn(snID, snKey)\n\tdefer func() {\n\t\tif retErr != nil {\n\t\t\tdeferCtx, deferCancel := rollbackContext()\n\t\t\tdefer deferCancel()\n\n\t\t\tout, err = exec.CommandContext(ctx, \"tgt-admin\", \"--delete\", targetIqn).CombinedOutput()\n\t\t\tif err != nil {\n\t\t\t\tlog.G(deferCtx).WithError(err).Warnf(\"failed to rollback target by tgt-admin: %s\", out)\n\t\t\t}\n\t\t}\n\t}()\n\n\t// Add a portal on a target\n\tout, err = exec.CommandContext(ctx, \"iscsiadm\", \"-m\", \"node\", \"-p\", defaultPortal, \"-T\", targetIqn, \"-o\", \"new\").CombinedOutput()\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"failed to add a portal on a target %s: %s\", targetIqn, out)\n\t}\n\tdefer func() {\n\t\t// rollback the portal\n\t\tif retErr != nil {\n\t\t\tdeferCtx, deferCancel := rollbackContext()\n\t\t\tdefer deferCancel()\n\n\t\t\tout, err = exec.CommandContext(deferCtx, \"iscsiadm\", \"-m\", \"node\", \"-p\", defaultPortal, \"-T\", targetIqn, \"-o\", \"delete\").CombinedOutput()\n\t\t\tif err != nil {\n\t\t\t\tlog.G(deferCtx).WithError(err).Warnf(\"failed to rollback a portal on a target %s: %s\", targetIqn, out)\n\t\t\t}\n\t\t}\n\t}()\n\n\t// Login a portal on a target\n\tout, err = exec.CommandContext(ctx, \"iscsiadm\", \"-m\", \"node\", \"-p\", defaultPortal, \"-T\", targetIqn, \"--login\").CombinedOutput()\n\tif err != nil {\n\t\texiterr, ok := err.(*exec.ExitError)\n\t\tif !ok || iscsi.Errno(exiterr.ExitCode()) != iscsi.ESESSEXISTS {\n\t\t\treturn errors.Wrapf(err, \"failed to login a portal on a target %s: %s\", targetIqn, out)\n\t\t}\n\t}\n\tdefer func() {\n\t\t// NOTE(fuweid): Basically, do login only once. The rollback doesn't impact other running portal.\n\t\tif retErr != nil {\n\t\t\tdeferCtx, deferCancel := rollbackContext()\n\t\t\tdefer deferCancel()\n\n\t\t\tout, err = exec.CommandContext(deferCtx, \"iscsiadm\", \"-m\", \"node\", \"-p\", defaultPortal, \"-T\", targetIqn, \"--logout\").CombinedOutput()\n\t\t\tif err != nil {\n\t\t\t\tlog.G(deferCtx).WithError(err).Warnf(\"failed to rollback to logout on a target %s: %s\", targetIqn, out)\n\t\t\t}\n\t\t}\n\t}()\n\n\t// Find the session and hostNumber mapping\n\thostToSessionID, err := iscsi.GetISCSIHostSessionMapForTarget(targetIqn, defaultPortal)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"failed to get hostNumber->SessionID mapping for %s\", targetIqn)\n\t}\n\n\tif len(hostToSessionID) != 1 {\n\t\treturn errors.Errorf(\"unexpected hostNumber->SessionID mapping result %v for %s\", hostToSessionID, targetIqn)\n\t}\n\n\t// The device doesn't show up instantly. Need retry here.\n\tvar lastErr error = nil\n\tvar mountPoint = o.tgtTargetMountpoint(snID)\n\tfor i := 1; i <= maxAttachAttempts; i++ {\n\t\tfor hostNumber, sessionIDs := range hostToSessionID {\n\t\t\tif len(sessionIDs) != 1 {\n\t\t\t\treturn errors.Errorf(\"unexpected hostNumber->SessionID mapping result %v for %s\", hostToSessionID, targetIqn)\n\t\t\t}\n\n\t\t\t// Assume that both channelID and targetID are zero.\n\t\t\tdevices, err := iscsi.GetDevicesForTarget(targetIqn, hostNumber, sessionIDs[0], 0, 0)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tif len(devices) != 1 {\n\t\t\t\tlastErr = errors.Errorf(\"unexpected devices %v for %s\", devices, targetIqn)\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tvar mflag uintptr = unix.MS_RDONLY\n\t\t\tif writable {\n\t\t\t\tmflag = 0\n\t\t\t}\n\n\t\t\t// TODO(fuweid): how to support multiple filesystem?\n\t\t\tif err := unix.Mount(devices[0], mountPoint, \"ext4\", mflag, \"\"); err != nil {\n\t\t\t\treturn errors.Wrapf(err, \"failed to mount the device %s on %s\", devices[0], mountPoint)\n\t\t\t}\n\t\t\tlastErr = nil\n\t\t}\n\n\t\tif lastErr == nil {\n\t\t\tbreak\n\t\t}\n\t\ttime.Sleep(1 * time.Second)\n\t}\n\treturn lastErr\n}", "func (*container) AttachTask(context.Context, libcontainerdtypes.StdioCallback) (libcontainerdtypes.Task, error) {\n\treturn nil, errdefs.NotFound(cerrdefs.ErrNotImplemented)\n}", "func (s *SelectableAttribute) Attach(cons Constraint) {\n\ts.AttributeImpl.Attach(cons)\n}", "func (d *Driver) writeDiskImage() error { // TODO\n\tlog.Debugf(\"Creating hard disk image...\")\n\n\tmagicString := \"boot2docker, this is xhyve speaking\"\n\n\tbuf := new(bytes.Buffer)\n\ttw := tar.NewWriter(buf)\n\n\t// magicString first so the automount script knows to format the disk\n\tfile := &tar.Header{Name: magicString, Size: int64(len(magicString))}\n\tif err := tw.WriteHeader(file); err != nil {\n\t\treturn err\n\t}\n\tif _, err := tw.Write([]byte(magicString)); err != nil {\n\t\treturn err\n\t}\n\t// .ssh/key.pub => authorized_keys\n\tfile = &tar.Header{Name: \".ssh\", Typeflag: tar.TypeDir, Mode: 0700}\n\tif err := tw.WriteHeader(file); err != nil {\n\t\treturn err\n\t}\n\tpubKey, err := ioutil.ReadFile(d.publicSSHKeyPath())\n\tif err != nil {\n\t\treturn err\n\t}\n\tfile = &tar.Header{Name: \".ssh/authorized_keys\", Size: int64(len(pubKey)), Mode: 0644}\n\tif err := tw.WriteHeader(file); err != nil {\n\t\treturn err\n\t}\n\tif _, err := tw.Write([]byte(pubKey)); err != nil {\n\t\treturn err\n\t}\n\tfile = &tar.Header{Name: \".ssh/authorized_keys2\", Size: int64(len(pubKey)), Mode: 0644}\n\tif err := tw.WriteHeader(file); err != nil {\n\t\treturn err\n\t}\n\tif _, err := tw.Write([]byte(pubKey)); err != nil {\n\t\treturn err\n\t}\n\tif err := tw.Close(); err != nil {\n\t\treturn err\n\t}\n\traw := buf.Bytes()\n\n\tif err := ioutil.WriteFile(d.imgPath(), raw, 0644); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (s *System) Attach(w *world.World) error {\n\ts.save = w.Save\n\treturn nil\n}", "func (c *MockAzureCloud) Disk() azure.DisksClient {\n\treturn c.DisksClient\n}", "func (c *CrosDisks) Mount(ctx context.Context, devicePath, fsType string, options []string) error {\n\treturn c.call(ctx, \"Mount\", devicePath, fsType, options).Err\n}" ]
[ "0.6433191", "0.6408497", "0.6383675", "0.6364639", "0.6311051", "0.62781525", "0.6211151", "0.6208306", "0.61464", "0.6119266", "0.5994655", "0.5925808", "0.58510363", "0.5824707", "0.5704301", "0.56972575", "0.56700426", "0.56573075", "0.56385034", "0.56340134", "0.56317014", "0.56194186", "0.56184995", "0.5617897", "0.5606501", "0.5584837", "0.5583585", "0.5579462", "0.55521256", "0.554135", "0.5516188", "0.5514965", "0.5503668", "0.5490587", "0.54857624", "0.5474303", "0.5472326", "0.54720604", "0.5470478", "0.5470465", "0.5442267", "0.54394025", "0.5434959", "0.54318345", "0.54065645", "0.5361119", "0.53552777", "0.5336166", "0.5332793", "0.53113395", "0.52886236", "0.52800643", "0.5256239", "0.52538353", "0.5208196", "0.5200313", "0.51853585", "0.51638025", "0.5147564", "0.5139147", "0.5125345", "0.51251256", "0.51203954", "0.5116642", "0.5116428", "0.51112145", "0.5106366", "0.5092506", "0.5088701", "0.5080933", "0.5071461", "0.50659347", "0.50408566", "0.50322163", "0.50100976", "0.5005472", "0.4996595", "0.49859104", "0.49812403", "0.49786496", "0.49592525", "0.49534816", "0.49406758", "0.49354333", "0.49270263", "0.49250415", "0.49036455", "0.49016425", "0.48979706", "0.4890172", "0.48708323", "0.48535094", "0.48517135", "0.48506606", "0.48450106", "0.484114", "0.48401996", "0.48394033", "0.48389784", "0.4838703" ]
0.76755553
0
DetachDisk uses the override method DetachDiskFn or the real implementation.
func (c *TestClient) DetachDisk(project, zone, instance, disk string) error { if c.DetachDiskFn != nil { return c.DetachDiskFn(project, zone, instance, disk) } return c.client.DetachDisk(project, zone, instance, disk) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (detacher *azureDiskDetacher) Detach(diskName string, nodeName types.NodeName) error {\n\tif diskName == \"\" {\n\t\treturn fmt.Errorf(\"invalid disk to detach: %q\", diskName)\n\t}\n\tinstanceid, err := detacher.azureProvider.InstanceID(nodeName)\n\tif err != nil {\n\t\tglog.Warningf(\"no instance id for node %q, skip detaching\", nodeName)\n\t\treturn nil\n\t}\n\tif ind := strings.LastIndex(instanceid, \"/\"); ind >= 0 {\n\t\tinstanceid = instanceid[(ind + 1):]\n\t}\n\n\tglog.V(4).Infof(\"detach %v from node %q\", diskName, nodeName)\n\terr = detacher.azureProvider.DetachDiskByName(diskName, \"\" /* diskURI */, nodeName)\n\tif err != nil {\n\t\tglog.Errorf(\"failed to detach azure disk %q, err %v\", diskName, err)\n\t}\n\n\treturn err\n}", "func (r Virtual_Guest) DetachDiskImage(imageId *int) (resp datatypes.Provisioning_Version1_Transaction, err error) {\n\tparams := []interface{}{\n\t\timageId,\n\t}\n\terr = r.Session.DoRequest(\"SoftLayer_Virtual_Guest\", \"detachDiskImage\", params, &r.Options, &resp)\n\treturn\n}", "func (p *Tmpfs) Detach(ctx driver.Context, v *types.Volume) error {\n\tctx.Log.Debugf(\"Tmpfs detach volume: %s\", v.Name)\n\tmountPath := v.Path()\n\treqID := v.Option(\"reqID\")\n\tids := v.Option(\"ids\")\n\n\tarr := strings.Split(ids, \",\")\n\tnewIDs := []string{}\n\tfor _, id := range arr {\n\t\tif id != reqID {\n\t\t\tnewIDs = append(newIDs, reqID)\n\t\t}\n\t}\n\n\tif len(newIDs) == 0 && utils.IsMountpoint(mountPath) {\n\t\tif err := syscall.Unmount(mountPath, 0); err != nil {\n\t\t\treturn fmt.Errorf(\"failed to umount %q, err: %v\", mountPath, err)\n\t\t}\n\n\t\tif err := os.Remove(mountPath); err != nil && !os.IsNotExist(err) {\n\t\t\treturn fmt.Errorf(\"remove %q directory failed, err: %v\", mountPath, err)\n\t\t}\n\n\t\tv.SetOption(\"freeTime\", strconv.FormatInt(time.Now().Unix(), 10))\n\t}\n\n\tv.SetOption(\"ids\", strings.Join(newIDs, \",\"))\n\n\treturn nil\n}", "func NewCmdDiskDetach() *cobra.Command {\n\tvar async, yes *bool\n\tvar udiskIDs *[]string\n\treq := base.BizClient.NewDetachUDiskRequest()\n\tcmd := &cobra.Command{\n\t\tUse: \"detach\",\n\t\tShort: \"Detach udisk instances from an uhost\",\n\t\tLong: \"Detach udisk instances from an uhost\",\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\ttext := `Please confirm that you have already unmounted file system corresponding to this hard drive,(See \"https://docs.ucloud.cn/storage_cdn/udisk/userguide/umount\" for help), otherwise it will cause file system damage and UHost cannot be normally shut down. Sure to detach?`\n\t\t\tif !*yes {\n\t\t\t\tsure, err := ux.Prompt(text)\n\t\t\t\tif err != nil {\n\t\t\t\t\tbase.Cxt.PrintErr(err)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tif !sure {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t\tfor _, id := range *udiskIDs {\n\t\t\t\tid = base.PickResourceID(id)\n\t\t\t\tany, err := describeUdiskByID(id, *req.ProjectId, *req.Region, *req.Zone)\n\t\t\t\tif err != nil {\n\t\t\t\t\tbase.HandleError(err)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tif any == nil {\n\t\t\t\t\tbase.Cxt.PrintErr(fmt.Errorf(\"udisk[%v] is not exist\", any))\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tins, ok := any.(*udisk.UDiskDataSet)\n\t\t\t\tif !ok {\n\t\t\t\t\tbase.Cxt.PrintErr(fmt.Errorf(\"%#v convert to udisk failed\", any))\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\treq.UHostId = &ins.UHostId\n\t\t\t\treq.UDiskId = &id\n\t\t\t\t*req.UHostId = base.PickResourceID(*req.UHostId)\n\t\t\t\tresp, err := base.BizClient.DetachUDisk(req)\n\t\t\t\tif err != nil {\n\t\t\t\t\tbase.HandleError(err)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\ttext := fmt.Sprintf(\"udisk[%s] is detaching from uhost[%s]\", resp.UDiskId, resp.UHostId)\n\t\t\t\tif *async {\n\t\t\t\t\tbase.Cxt.Println(text)\n\t\t\t\t} else {\n\t\t\t\t\tpollDisk(resp.UDiskId, *req.ProjectId, *req.Region, *req.Zone, text, []string{status.DISK_AVAILABLE, status.DISK_FAILED})\n\t\t\t\t}\n\t\t\t}\n\t\t},\n\t}\n\tflags := cmd.Flags()\n\tflags.SortFlags = false\n\tudiskIDs = flags.StringSlice(\"udisk-id\", nil, \"Required. Resource ID of the udisk instances to detach\")\n\treq.ProjectId = flags.String(\"project-id\", base.ConfigInstance.ProjectID, \"Optional. Assign project-id\")\n\treq.Region = flags.String(\"region\", base.ConfigInstance.Region, \"Optional. Assign region\")\n\treq.Zone = flags.String(\"zone\", base.ConfigInstance.Zone, \"Optional. Assign availability zone\")\n\tasync = flags.Bool(\"async\", false, \"Optional. Do not wait for the long-running operation to finish.\")\n\tyes = flags.BoolP(\"yes\", \"y\", false, \"Optional. Do not prompt for confirmation.\")\n\n\tflags.SetFlagValuesFunc(\"udisk-id\", func() []string {\n\t\treturn getDiskList([]string{status.DISK_INUSE}, *req.ProjectId, *req.Region, *req.Zone)\n\t})\n\n\tcmd.MarkFlagRequired(\"udisk-id\")\n\treturn cmd\n}", "func (srv *VolumeService) Detach(volumename string, vmname string) error {\n\tvol, err := srv.Get(volumename)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"No volume found with name or id '%s'\", volumename)\n\t}\n\n\t// Get VM ID\n\tvmService := NewVMService(srv.provider)\n\tvm, err := vmService.Get(vmname)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"No VM found with name or id '%s'\", vmname)\n\t}\n\n\tvolatt, err := srv.provider.GetVolumeAttachment(vm.ID, vol.ID)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Error getting volume attachment: %s\", err)\n\t}\n\n\tsshConfig, err := srv.provider.GetSSHConfig(vm.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tserver, err := nfs.NewServer(sshConfig)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = server.UnmountBlockDevice(volatt.Device)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Finaly delete the attachment\n\treturn srv.provider.DeleteVolumeAttachment(vm.ID, vol.ID)\n}", "func (driver *Driver) Detach(volumeName, instanceID string) error {\n\tvolume, err := driver.sdm.GetVolume(\"\", volumeName)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn driver.sdm.DetachVolume(true, volume[0].VolumeID, instanceID)\n}", "func (endpoint *VirtualEndpoint) Detach(netNsCreated bool, netNsPath string) error {\n\t// The network namespace would have been deleted at this point\n\t// if it has not been created by virtcontainers.\n\tif !netNsCreated {\n\t\treturn nil\n\t}\n\n\tnetworkLogger().WithField(\"endpoint-type\", \"virtual\").Info(\"Detaching endpoint\")\n\n\treturn doNetNS(netNsPath, func(_ ns.NetNS) error {\n\t\treturn xconnectVMNetwork(endpoint, false, 0, false)\n\t})\n}", "func (d *DefaultDriver) DetachVolume(volumeID string) error {\n\treturn &errors.ErrNotSupported{\n\t\tType: \"Function\",\n\t\tOperation: \"DetachVolume()\",\n\t}\n}", "func (d *detacherDefaults) Detach(volumeName string, hostName types.NodeName) error {\n\tklog.Warning(logPrefix(d.plugin.flexVolumePlugin), \"using default Detach for volume \", volumeName, \", host \", hostName)\n\treturn nil\n}", "func (endpoint *BridgedMacvlanEndpoint) Detach(netNsCreated bool, netNsPath string) error {\n\t// The network namespace would have been deleted at this point\n\t// if it has not been created by virtcontainers.\n\tif !netNsCreated {\n\t\treturn nil\n\t}\n\n\tnetworkLogger().Info(\"Detaching virtual endpoint\")\n\n\treturn doNetNS(netNsPath, func(_ ns.NetNS) error {\n\t\t//return xconnectVMNetwork(&(endpoint.NetPair), false, 0, false, endpoint.EndpointType)\n\t\treturn xconnectVMNetwork(endpoint, false, 0, false)\n\t})\n}", "func (c *Controller) Detach(detachRequest k8sresources.FlexVolumeDetachRequest) k8sresources.FlexVolumeResponse {\n\tc.logger.Println(\"controller-detach-start\")\n\tdefer c.logger.Println(\"controller-detach-end\")\n\tif detachRequest.Version == k8sresources.KubernetesVersion_1_5 {\n\t\treturn k8sresources.FlexVolumeResponse{\n\t\t\tStatus: \"Success\",\n\t\t}\n\t}\n\treturn k8sresources.FlexVolumeResponse{\n\t\tStatus: \"Not supported\",\n\t}\n}", "func (z *zpoolctl) Detach(ctx context.Context, name, dev string) *execute {\n\targs := []string{\"attach\", name, dev}\n\treturn &execute{ctx: ctx, name: z.cmd, args: args}\n}", "func (s *StorageActionsServiceOp) Detach(driveID string) (*Action, *Response, error) {\n\trequest := &ActionRequest{\n\t\t\"type\": \"detach\",\n\t}\n\treturn s.doAction(driveID, request)\n}", "func (m *ClusterService) decommissionDisk(ctx context.Context, args struct {\n\tOffLineAddr string\n\tDiskPath string\n}) (*proto.GeneralResp, error) {\n\n\tnode, err := m.cluster.dataNode(args.OffLineAddr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbadPartitions := node.badPartitions(args.DiskPath, m.cluster)\n\tif len(badPartitions) == 0 {\n\t\terr = fmt.Errorf(\"node[%v] disk[%v] does not have any data partition\", node.Addr, args.DiskPath)\n\t\treturn nil, err\n\t}\n\n\tvar badPartitionIds []uint64\n\tfor _, bdp := range badPartitions {\n\t\tbadPartitionIds = append(badPartitionIds, bdp.PartitionID)\n\t}\n\trstMsg := fmt.Sprintf(\"receive decommissionDisk node[%v] disk[%v], badPartitionIds[%v] has offline successfully\",\n\t\tnode.Addr, args.DiskPath, badPartitionIds)\n\tif err = m.cluster.decommissionDisk(node, args.DiskPath, badPartitions); err != nil {\n\t\treturn nil, err\n\t}\n\tWarn(m.cluster.Name, rstMsg)\n\n\treturn proto.Success(\"success\"), nil\n\n}", "func (cl *Client) gceVolumeDetach(ctx context.Context, vda *csp.VolumeDetachArgs, vid string) (*csp.Volume, error) {\n\tif vda.Force {\n\t\tcl.csp.dbgF(\"ignoring unsupported force on detach [%s]\", vid)\n\t}\n\tcomputeService, err := cl.getComputeService(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\top, err := computeService.Instances().DetachDisk(cl.projectID, cl.attrs[AttrZone].Value, vda.NodeIdentifier, vid).Context(ctx).Do()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif err = cl.waitForOperation(ctx, op); err != nil {\n\t\treturn nil, err\n\t}\n\tvol, err := cl.vr.gceVolumeGet(ctx, vid)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn vol, nil\n}", "func (s *VolumeListener) Detach(inctx context.Context, in *protocol.VolumeDetachmentRequest) (empty *googleprotobuf.Empty, err error) {\n\tdefer fail.OnExitConvertToGRPCStatus(inctx, &err)\n\tdefer fail.OnExitWrapError(inctx, &err, \"cannot detach volume\")\n\n\tempty = &googleprotobuf.Empty{}\n\tif s == nil {\n\t\treturn empty, fail.InvalidInstanceError()\n\t}\n\tif in == nil {\n\t\treturn empty, fail.InvalidParameterCannotBeNilError(\"in\")\n\t}\n\tif inctx == nil {\n\t\treturn empty, fail.InvalidParameterCannotBeNilError(\"inctx\")\n\t}\n\n\tvolumeRef, volumeRefLabel := srvutils.GetReference(in.GetVolume())\n\tif volumeRef == \"\" {\n\t\treturn empty, fail.InvalidRequestError(\"neither name nor id given as reference for volume\")\n\t}\n\thostRef, hostRefLabel := srvutils.GetReference(in.GetHost())\n\tif hostRef == \"\" {\n\t\treturn empty, fail.InvalidRequestError(\"neither name nor id given as reference for host\")\n\t}\n\n\tjob, xerr := PrepareJob(inctx, in.GetVolume().GetTenantId(), fmt.Sprintf(\"/volume/%s/host/%s/detach\", volumeRef, hostRef))\n\tif xerr != nil {\n\t\treturn nil, xerr\n\t}\n\tdefer job.Close()\n\n\thandler := VolumeHandler(job)\n\tif xerr = handler.Detach(volumeRef, hostRef); xerr != nil {\n\t\treturn empty, xerr\n\t}\n\n\tlogrus.WithContext(job.Context()).Infof(\"Volume %s successfully detached from %s.\", volumeRefLabel, hostRefLabel)\n\treturn empty, nil\n}", "func (s *SharedMemorySegment) Detach() error {\n\tdata := (*reflect.SliceHeader)(unsafe.Pointer(&s.data))\n\t_, _, errno := syscall.Syscall(syscall.SYS_SHMDT, data.Data, 0, 0)\n\tif errno != 0 {\n\t\treturn errors.New(errno.Error())\n\t}\n\treturn nil\n}", "func (endpoint *PhysicalEndpoint) Detach(netNsCreated bool, netNsPath string) error {\n\t// Bind back the physical network interface to host.\n\t// We need to do this even if a new network namespace has not\n\t// been created by virtcontainers.\n\tnetworkLogger().WithField(\"endpoint-type\", \"physical\").Info(\"Detaching endpoint\")\n\n\t// We do not need to enter the network namespace to bind back the\n\t// physical interface to host driver.\n\treturn bindNICToHost(endpoint)\n}", "func (digitalocean DigitalOcean) DetachVolume(volumeID string) error {\n\tdoc, err := DigitalOceanClient()\n\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\t_, err = doc.client.Storage.DeleteVolume(doc.context, volumeID)\n\n\treturn err\n}", "func (d DobsClient) DetachVolume(ctx Context, volumeID string, dropletID string) (error) {\n\tdropletIDI, err := strconv.Atoi(dropletID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\taction, _, err := d.GodoClient.StorageActions.DetachByDropletID(ctx, volumeID, dropletIDI)\t\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = d.waitForAction(ctx, volumeID, action)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (_m *OpenStackMock) WaitDiskDetached(instanceID string, volumeID string) error {\n\tret := _m.Called(instanceID, volumeID)\n\n\tvar r0 error\n\tif rf, ok := ret.Get(0).(func(string, string) error); ok {\n\t\tr0 = rf(instanceID, volumeID)\n\t} else {\n\t\tr0 = ret.Error(0)\n\t}\n\n\treturn r0\n}", "func (endpoint *MacvtapEndpoint) Detach(netNsCreated bool, netNsPath string) error {\n\tnetworkLogger().WithField(\"endpoint-type\", \"macvtap\").Info(\"Detaching endpoint\")\n\treturn nil\n}", "func (endpoint *VirtualEndpoint) HotDetach(h hypervisor, netNsCreated bool, netNsPath string) error {\n\tif !netNsCreated {\n\t\treturn nil\n\t}\n\tnetworkLogger().Info(\"Hot detaching virtual endpoint\")\n\tif err := doNetNS(netNsPath, func(_ ns.NetNS) error {\n\t\treturn xconnectVMNetwork(endpoint, false, 0, h.hypervisorConfig().DisableVhostNet)\n\t}); err != nil {\n\t\tnetworkLogger().WithError(err).Warn(\"Error un-bridging virtual ep\")\n\t}\n\n\tif _, err := h.hotplugRemoveDevice(endpoint, netDev); err != nil {\n\t\tnetworkLogger().WithError(err).Error(\"Error detach virtual ep\")\n\t\treturn err\n\t}\n\treturn nil\n}", "func (util *PortworxVolumeUtil) DetachVolume(u *portworxVolumeUnmounter) error {\n\tdriver, err := util.getPortworxDriver(u.plugin.host, true /*localOnly*/)\n\tif err != nil || driver == nil {\n\t\tglog.Errorf(\"Failed to get portworx driver. Err: %v\", err)\n\t\treturn err\n\t}\n\n\terr = driver.Detach(u.volName)\n\tif err != nil {\n\t\tglog.Errorf(\"Error detaching Portworx Volume (%v): %v\", u.volName, err)\n\t\treturn err\n\t}\n\treturn nil\n}", "func Detach(deviceNode string, flags ...detachFlag) error {\n\tcmd := exec.Command(hdiutilPath, \"detach\", deviceNode)\n\tif len(flags) > 0 {\n\t\tfor _, flag := range flags {\n\t\t\tcmd.Args = append(cmd.Args, flag.detachFlag()...)\n\t\t}\n\t}\n\n\terr := cmd.Run()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (endpoint *VhostUserEndpoint) Detach(netNsCreated bool, netNsPath string) error {\n\tnetworkLogger().WithField(\"endpoint-type\", \"vhostuser\").Info(\"Detaching endpoint\")\n\treturn nil\n}", "func (cl *Client) VolumeDetach(ctx context.Context, vda *csp.VolumeDetachArgs) (*csp.Volume, error) {\n\tsvc, vid, _ := VolumeIdentifierParse(vda.VolumeIdentifier)\n\tswitch svc {\n\tcase ServiceGCE:\n\t\treturn cl.gceVolumeDetach(ctx, vda, vid)\n\t}\n\treturn nil, fmt.Errorf(\"storage type currently unsupported\")\n}", "func (c *Core) DetachVolume(id types.VolumeID, extra map[string]string) (*types.Volume, error) {\n\tc.lock.Lock(id.Name)\n\tdefer c.lock.Unlock(id.Name)\n\n\tv, dv, err := c.getVolumeDriver(id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tctx := driver.Contexts()\n\n\t// merge extra to volume spec extra.\n\tfor key, value := range extra {\n\t\tv.Spec.Extra[key] = value\n\t}\n\n\t// if volume has referance, skip to detach volume.\n\tref := v.Option(types.OptionRef)\n\tif d, ok := dv.(driver.AttachDetach); ok && ref == \"\" {\n\t\tif err := d.Detach(ctx, v); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// update meta info.\n\tif err := c.store.Put(v); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn v, nil\n}", "func (f *FakeInstance) DetachPrivateNetwork(_ context.Context, _, _ string) error {\n\tpanic(\"implement me\")\n}", "func (vpcIks *IksVpcSession) DetachVolume(volumeAttachmentRequest provider.VolumeAttachmentRequest) (*http.Response, error) {\n\tvpcIks.IksSession.Logger.Debug(\"Entry of IksVpcSession.DetachVolume method...\")\n\tdefer vpcIks.Logger.Debug(\"Exit from IksVpcSession.DetachVolume method...\")\n\treturn vpcIks.IksSession.DetachVolume(volumeAttachmentRequest)\n}", "func DiskVMVirtiofsdStop(socketPath string, pidPath string) error {\n\tif shared.PathExists(pidPath) {\n\t\tproc, err := subprocess.ImportProcess(pidPath)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\terr = proc.Stop()\n\t\t// The virtiofsd process will terminate automatically once the VM has stopped.\n\t\t// We therefore should only return an error if it's still running and fails to stop.\n\t\tif err != nil && err != subprocess.ErrNotRunning {\n\t\t\treturn err\n\t\t}\n\n\t\t// Remove PID file if needed.\n\t\tos.Remove(pidPath)\n\t}\n\n\t// Remove socket file if needed.\n\tos.Remove(socketPath)\n\n\treturn nil\n}", "func (endpoint *HNSEndpoint) VirtualMachineNICDetach() error {\n\toperation := \"VirtualMachineNicDetach\"\n\ttitle := \"hcsshim::HNSEndpoint::\" + operation\n\tlogrus.Debugf(title+\" id=%s\", endpoint.Id)\n\n\trequestMessage := &EndpointAttachDetachRequest{\n\t\tSystemType: VirtualMachineType,\n\t}\n\tresponse := &EndpointResquestResponse{}\n\n\tjsonString, err := json.Marshal(requestMessage)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn hnsCall(\"POST\", \"/endpoints/\"+endpoint.Id+\"/detach\", string(jsonString), &response)\n}", "func DetachStorage(attachmentID string) error {\n\tclient, err := NewExtPacketClient()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, e := client.Storages.Detach(attachmentID)\n\treturn e\n}", "func (s *Module) DiskDelete(name string) error {\n\tpath, err := s.findDisk(name)\n\tif os.IsNotExist(err) {\n\t\treturn nil\n\t} else if err != nil {\n\t\treturn err\n\t}\n\n\tif err := os.Remove(path); err != nil && !os.IsNotExist(err) {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (dm *FloatingIPService) DetachFloatingIP(ID string) (err error) {\n\tlog.Debug(\"DetachFloatingIP\")\n\n\tdata, status, err := dm.concertoService.Delete(fmt.Sprintf(\"/network/floating_ips/%s/attached_server\", ID))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err = utils.CheckStandardStatus(status, data); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *Computation) Detach(ctx context.Context) error {\n\treturn c.DetachWithReason(ctx, \"\")\n}", "func DiskDelete(w rest.ResponseWriter, r *rest.Request) {\n\treq := DiskDeleteRequest{}\n\terr := r.DecodeJsonPayload(&req)\n\tif err != nil {\n\t\tlogit.Error.Println(err.Error())\n\t\trest.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\tlogit.Info.Println(\"DiskDelete called \" + req.Path)\n\n\tvar cmd *exec.Cmd\n\tcmd = exec.Command(\"deletevolume\", req.Path)\n\tvar out bytes.Buffer\n\tvar stderr bytes.Buffer\n\tcmd.Stdout = &out\n\tcmd.Stderr = &stderr\n\terr = cmd.Run()\n\tif err != nil {\n\t\tlogit.Error.Println(err.Error())\n\t\trest.Error(w, err.Error(), 400)\n\t\treturn\n\t}\n\n\tvar response DiskDeleteResponse\n\tresponse.Output = out.String()\n\tresponse.Status = \"OK\"\n\tw.WriteJson(&response)\n}", "func (detacher *azureDiskDetacher) UnmountDevice(deviceMountPath string) error {\n\tvolume := path.Base(deviceMountPath)\n\tif err := util.UnmountPath(deviceMountPath, detacher.mounter); err != nil {\n\t\tglog.Errorf(\"Error unmounting %q: %v\", volume, err)\n\t\treturn err\n\t} else {\n\t\treturn nil\n\t}\n}", "func Detach(c *deis.Client, name string, domain string) error {\n\turl := fmt.Sprintf(\"/v2/certs/%s/domain/%s\", name, domain)\n\tres, err := c.Request(\"DELETE\", url, nil)\n\tif err == nil {\n\t\tres.Body.Close()\n\t}\n\treturn err\n}", "func (r *vdm) Detach(volumeName, instanceID string, force bool) error {\n\tfor _, d := range r.drivers {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"moduleName\": r.rexray.Context,\n\t\t\t\"driverName\": d.Name(),\n\t\t\t\"volumeName\": volumeName,\n\t\t\t\"instanceID\": instanceID,\n\t\t\t\"force\": force}).Info(\"vdm.Detach\")\n\t\treturn d.Detach(volumeName, instanceID, force)\n\t}\n\treturn errors.ErrNoVolumesDetected\n}", "func DeleteDisk(id int) {\n\tdb.Session().Delete(&model.Disk{}, \"id = ?\", id)\n}", "func (m *Module) detach(recv []byte) ([]byte, error) {\n\ttype paramsProto struct {\n\t\tUUID llib.UUID\n\t\tMAC [17]byte\n\t\tVLAN int16\n\t\tType netType\n\t}\n\tparams := paramsProto{}\n\tbinary.Read(bytes.NewReader(recv), m.Config.Endianness_, &params)\n\tuuid := lib.TrimBuf(params.UUID[:])\n\tmac := lib.TrimBuf(params.MAC[:])\n\n\tconn, err := llib.Connect()\n\tif err != nil {\n\t\treturn lib.RespError(err)\n\t}\n\tdefer func() {\n\t\tconn.Close()\n\t}()\n\tdom, err := conn.LookupDomainByUUIDString(string(uuid))\n\tif err != nil {\n\t\treturn lib.RespError(err)\n\t}\n\tdomXML, err := dom.GetXMLDesc(libvirt.DOMAIN_XML_SECURE)\n\tif err != nil {\n\t\treturn lib.RespError(err)\n\t}\n\tdoc := etree.NewDocument()\n\tif err = doc.ReadFromString(domXML); err != nil {\n\t\treturn lib.RespError(err)\n\t}\n\tdomain := doc.SelectElement(\"domain\")\n\tdevices := domain.SelectElement(\"devices\")\n\tifaces := devices.SelectElements(\"interface\")\n\tvar iface *etree.Element\n\tfor _, iface = range ifaces {\n\t\tmacElem := iface.SelectElement(\"mac\")\n\t\tif macElem.SelectAttrValue(\"address\", \"\") == string(mac) {\n\t\t\tbreak\n\t\t}\n\t}\n\tif iface == nil {\n\t\treturn lib.RespError(errors.New(\"can't find iface\"))\n\t}\n\tnewDoc := etree.NewDocument()\n\tnewDoc.SetRoot(iface)\n\tifaceXML, _ := newDoc.WriteToString()\n\t// ifaceXML := iface.Text()\n\tlog.Println(\"ifacexml\", ifaceXML)\n\tflags := libvirt.DOMAIN_DEVICE_MODIFY_CONFIG |\n\t\tlibvirt.DOMAIN_DEVICE_MODIFY_LIVE\n\tif err = dom.DetachDeviceFlags(ifaceXML, flags); err != nil {\n\t\treturn lib.RespError(err)\n\t}\n\treturn lib.RespOk(\"\")\n}", "func (device *BlockDevice) Detach(ctx context.Context, devReceiver api.DeviceReceiver) error {\n\tskip, err := device.bumpAttachCount(false)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif skip {\n\t\treturn nil\n\t}\n\n\tdefer func() {\n\t\tif err != nil {\n\t\t\tdevice.bumpAttachCount(true)\n\t\t} else {\n\t\t\tdevReceiver.UnsetSandboxBlockIndex(device.BlockDrive.Index)\n\t\t}\n\t}()\n\n\tdeviceLogger().WithField(\"device\", device.DeviceInfo.HostPath).Info(\"Unplugging block device\")\n\n\tif err = devReceiver.HotplugRemoveDevice(ctx, device, config.DeviceBlock); err != nil {\n\t\tdeviceLogger().WithError(err).Error(\"Failed to unplug block device\")\n\t\treturn err\n\t}\n\treturn nil\n}", "func (s *Module) DiskCreate(name string, size gridtypes.Unit) (disk pkg.VDisk, err error) {\n\tpath, err := s.findDisk(name)\n\tif err == nil {\n\t\treturn disk, errors.Wrapf(os.ErrExist, \"disk with id '%s' already exists\", name)\n\t}\n\n\tbase, err := s.diskFindCandidate(size)\n\tif err != nil {\n\t\treturn disk, errors.Wrapf(err, \"failed to find a candidate to host vdisk of size '%d'\", size)\n\t}\n\n\tpath, err = s.safePath(base, name)\n\tif err != nil {\n\t\treturn disk, err\n\t}\n\n\tdefer func() {\n\t\t// clean up disk file if error\n\t\tif err != nil {\n\t\t\tos.RemoveAll(path)\n\t\t}\n\t}()\n\n\tdefer syscall.Sync()\n\n\tvar file *os.File\n\tfile, err = os.Create(path)\n\tif err != nil {\n\t\treturn disk, err\n\t}\n\n\tdefer file.Close()\n\tif err = chattr.SetAttr(file, chattr.FS_NOCOW_FL); err != nil {\n\t\treturn disk, err\n\t}\n\n\tif err = syscall.Fallocate(int(file.Fd()), 0, 0, int64(size)); err != nil {\n\t\treturn disk, errors.Wrap(err, \"failed to truncate disk to size\")\n\t}\n\n\treturn pkg.VDisk{Path: path, Size: int64(size)}, nil\n}", "func (c *TestClient) DeleteDisk(project, zone, name string) error {\n\tif c.DeleteDiskFn != nil {\n\t\treturn c.DeleteDiskFn(project, zone, name)\n\t}\n\treturn c.client.DeleteDisk(project, zone, name)\n}", "func (c *Compute) Disk(name string) (string, error) {\n\tdisk, err := c.Disks.Get(c.Project, c.Zone, name).Do()\n\tif err == nil {\n\t\tlog.Printf(\"found existing root disk: %q\", disk.SelfLink)\n\t\treturn disk.SelfLink, nil\n\t}\n\tlog.Printf(\"not found, creating new root disk: %q\", name)\n\top, err := c.Disks.Insert(c.Project, c.Zone, &compute.Disk{\n\t\tName: name,\n\t}).SourceImage(*image).Do()\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"disk insert api call failed: %v\", err)\n\t}\n\tif err := c.wait(op); err != nil {\n\t\treturn \"\", fmt.Errorf(\"disk insert operation failed: %v\", err)\n\t}\n\tlog.Printf(\"root disk created: %q\", op.TargetLink)\n\treturn op.TargetLink, nil\n}", "func (endpoint *HNSEndpoint) HostDetach() error {\n\toperation := \"HostDetach\"\n\ttitle := \"hcsshim::HNSEndpoint::\" + operation\n\tlogrus.Debugf(title+\" id=%s\", endpoint.Id)\n\trequestMessage := &EndpointAttachDetachRequest{\n\t\tSystemType: HostType,\n\t}\n\tresponse := &EndpointResquestResponse{}\n\n\tjsonString, err := json.Marshal(requestMessage)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn hnsCall(\"POST\", \"/endpoints/\"+endpoint.Id+\"/detach\", string(jsonString), &response)\n}", "func (endpoint *VhostUserEndpoint) HotDetach(h hypervisor, netNsCreated bool, netNsPath string) error {\n\treturn fmt.Errorf(\"VhostUserEndpoint does not support Hot detach\")\n}", "func TestCreateAndDetach(t *testing.T) {\n\tlog.SetLevel(log.DebugLevel)\n\n\tclient := Session(context.Background(), t)\n\tif client == nil {\n\t\treturn\n\t}\n\n\timagestore := client.Datastore.Path(\"imagestore\")\n\n\tfm := object.NewFileManager(client.Vim25())\n\n\t// create a directory in the datastore\n\t// eat the error because we dont care if it exists\n\tfm.MakeDirectory(context.TODO(), imagestore, nil, true)\n\n\tvdm, err := NewDiskManager(context.TODO(), client)\n\tif err != nil && err.Error() == \"can't find the hosting vm\" {\n\t\tt.Skip(\"Skipping: test must be run in a VM\")\n\t}\n\n\tif !assert.NoError(t, err) || !assert.NotNil(t, vdm) {\n\t\treturn\n\t}\n\n\tdiskSize := int64(1 << 10)\n\tparent, err := vdm.Create(context.TODO(), client.Datastore.Path(\"imagestore/scratch.vmdk\"), diskSize)\n\tif !assert.NoError(t, err) {\n\t\treturn\n\t}\n\n\tnumChildren := 3\n\tchildren := make([]*VirtualDisk, numChildren)\n\n\ttestString := \"Ground control to Major Tom\"\n\twriteSize := len(testString) / numChildren\n\t// Create children which inherit from eachother\n\tfor i := 0; i < numChildren; i++ {\n\n\t\tp := client.Datastore.Path(fmt.Sprintf(\"imagestore/child%d.vmdk\", i))\n\t\tchild, cerr := vdm.CreateAndAttach(context.TODO(), p, parent.DatastoreURI, 0, os.O_RDWR)\n\t\tif !assert.NoError(t, cerr) {\n\t\t\treturn\n\t\t}\n\n\t\tchildren[i] = child\n\n\t\t// Write directly to the disk\n\t\tf, cerr := os.OpenFile(child.DevicePath, os.O_RDWR, os.FileMode(0777))\n\t\tif !assert.NoError(t, cerr) {\n\t\t\treturn\n\t\t}\n\n\t\tstart := i * writeSize\n\t\tend := start + writeSize\n\n\t\tif i == numChildren-1 {\n\t\t\t// last chunk, write to the end.\n\t\t\t_, cerr = f.WriteAt([]byte(testString[start:]), int64(start))\n\t\t\tif !assert.NoError(t, cerr) {\n\t\t\t\treturn\n\t\t\t}\n\t\t\t// Try to read the whole string\n\t\t\tb := make([]byte, len(testString))\n\t\t\tf.Seek(0, 0)\n\t\t\t_, cerr = f.Read(b)\n\t\t\tif !assert.NoError(t, cerr) {\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\t//check against the test string\n\t\t\tif !assert.True(t, strings.Compare(testString, string(b)) == 0) {\n\t\t\t\treturn\n\t\t\t}\n\n\t\t} else {\n\t\t\t_, cerr = f.WriteAt([]byte(testString[start:end]), int64(start))\n\t\t\tif !assert.NoError(t, cerr) {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t\tf.Close()\n\n\t\tcerr = vdm.Detach(context.TODO(), child)\n\t\tif !assert.NoError(t, cerr) {\n\t\t\treturn\n\t\t}\n\n\t\t// use this image as the next parent\n\t\tparent = child\n\t}\n\n\t//\t// Nuke the images\n\t//\tfor i := len(children) - 1; i >= 0; i-- {\n\t//\t\terr = vdm.Delete(context.TODO(), children[i])\n\t//\t\tif !assert.NoError(t, err) {\n\t//\t\t\treturn\n\t//\t\t}\n\t//\t}\n\n\t// Nuke the image store\n\t_, err = tasks.WaitForResult(context.TODO(), func(ctx context.Context) (tasks.ResultWaiter, error) {\n\t\treturn fm.DeleteDatastoreFile(ctx, imagestore, nil)\n\t})\n\n\tif !assert.NoError(t, err) {\n\t\treturn\n\t}\n}", "func (m *Mixer) Detach() error {\n\tif !bool(C.al_detach_mixer((*C.ALLEGRO_MIXER)(m))) {\n\t\treturn errors.New(\"failed to detach mixer\")\n\t}\n\treturn nil\n}", "func (endpoint *BridgedMacvlanEndpoint) HotDetach(h hypervisor, netNsCreated bool, netNsPath string) error {\n\treturn fmt.Errorf(\"BridgedMacvlanEndpoint does not support Hot detach\")\n}", "func DiskMount(srcPath string, dstPath string, readonly bool, recursive bool, propagation string, mountOptions []string, fsName string) error {\n\tvar err error\n\n\t// Prepare the mount flags\n\tflags := 0\n\tif readonly {\n\t\tflags |= unix.MS_RDONLY\n\t}\n\n\t// Detect the filesystem\n\tif fsName == \"none\" {\n\t\tflags |= unix.MS_BIND\n\t}\n\n\tif propagation != \"\" {\n\t\tswitch propagation {\n\t\tcase \"private\":\n\t\t\tflags |= unix.MS_PRIVATE\n\t\tcase \"shared\":\n\t\t\tflags |= unix.MS_SHARED\n\t\tcase \"slave\":\n\t\t\tflags |= unix.MS_SLAVE\n\t\tcase \"unbindable\":\n\t\t\tflags |= unix.MS_UNBINDABLE\n\t\tcase \"rprivate\":\n\t\t\tflags |= unix.MS_PRIVATE | unix.MS_REC\n\t\tcase \"rshared\":\n\t\t\tflags |= unix.MS_SHARED | unix.MS_REC\n\t\tcase \"rslave\":\n\t\t\tflags |= unix.MS_SLAVE | unix.MS_REC\n\t\tcase \"runbindable\":\n\t\t\tflags |= unix.MS_UNBINDABLE | unix.MS_REC\n\t\tdefault:\n\t\t\treturn fmt.Errorf(\"Invalid propagation mode %q\", propagation)\n\t\t}\n\t}\n\n\tif recursive {\n\t\tflags |= unix.MS_REC\n\t}\n\n\t// Mount the filesystem\n\terr = unix.Mount(srcPath, dstPath, fsName, uintptr(flags), strings.Join(mountOptions, \",\"))\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Unable to mount %q at %q with filesystem %q: %w\", srcPath, dstPath, fsName, err)\n\t}\n\n\t// Remount bind mounts in readonly mode if requested\n\tif readonly == true && flags&unix.MS_BIND == unix.MS_BIND {\n\t\tflags = unix.MS_RDONLY | unix.MS_BIND | unix.MS_REMOUNT\n\t\terr = unix.Mount(\"\", dstPath, fsName, uintptr(flags), \"\")\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Unable to mount %q in readonly mode: %w\", dstPath, err)\n\t\t}\n\t}\n\n\tflags = unix.MS_REC | unix.MS_SLAVE\n\terr = unix.Mount(\"\", dstPath, \"\", uintptr(flags), \"\")\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Unable to make mount %q private: %w\", dstPath, err)\n\t}\n\n\treturn nil\n}", "func (endpoint *MacvtapEndpoint) HotDetach(h hypervisor, netNsCreated bool, netNsPath string) error {\n\treturn fmt.Errorf(\"MacvtapEndpoint does not support Hot detach\")\n}", "func (endpoint *PhysicalEndpoint) HotDetach(h hypervisor, netNsCreated bool, netNsPath string) error {\n\treturn fmt.Errorf(\"PhysicalEndpoint does not support Hot detach\")\n}", "func (s stage) partitionDisk(dev types.Disk, devAlias string) error {\n\tif cutil.IsTrue(dev.WipeTable) {\n\t\top := sgdisk.Begin(s.Logger, devAlias)\n\t\ts.Logger.Info(\"wiping partition table requested on %q\", devAlias)\n\t\top.WipeTable(true)\n\t\tif err := op.Commit(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Ensure all partitions with number 0 are last\n\tsort.Stable(PartitionList(dev.Partitions))\n\n\top := sgdisk.Begin(s.Logger, devAlias)\n\n\tdiskInfo, err := s.getPartitionMap(devAlias)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// get a list of parititions that have size and start 0 replaced with the real sizes\n\t// that would be used if all specified partitions were to be created anew.\n\t// Also calculate sectors for all of the start/size values.\n\tresolvedPartitions, err := s.getRealStartAndSize(dev, devAlias, diskInfo)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor _, part := range resolvedPartitions {\n\t\tshouldExist := partitionShouldExist(part)\n\t\tinfo, exists := diskInfo.GetPartition(part.Number)\n\t\tvar matchErr error\n\t\tif exists {\n\t\t\tmatchErr = partitionMatches(info, part)\n\t\t}\n\t\tmatches := exists && matchErr == nil\n\t\twipeEntry := cutil.IsTrue(part.WipePartitionEntry)\n\n\t\t// This is a translation of the matrix in the operator notes.\n\t\tswitch {\n\t\tcase !exists && !shouldExist:\n\t\t\ts.Logger.Info(\"partition %d specified as nonexistant and no partition was found. Success.\", part.Number)\n\t\tcase !exists && shouldExist:\n\t\t\top.CreatePartition(part)\n\t\tcase exists && !shouldExist && !wipeEntry:\n\t\t\treturn fmt.Errorf(\"partition %d exists but is specified as nonexistant and wipePartitionEntry is false\", part.Number)\n\t\tcase exists && !shouldExist && wipeEntry:\n\t\t\top.DeletePartition(part.Number)\n\t\tcase exists && shouldExist && matches:\n\t\t\ts.Logger.Info(\"partition %d found with correct specifications\", part.Number)\n\t\tcase exists && shouldExist && !wipeEntry && !matches:\n\t\t\tif partitionMatchesResize(info, part) {\n\t\t\t\ts.Logger.Info(\"resizing partition %d\", part.Number)\n\t\t\t\top.DeletePartition(part.Number)\n\t\t\t\tpart.Number = info.Number\n\t\t\t\tpart.GUID = &info.GUID\n\t\t\t\tpart.TypeGUID = &info.TypeGUID\n\t\t\t\tpart.Label = &info.Label\n\t\t\t\tpart.StartSector = &info.StartSector\n\t\t\t\top.CreatePartition(part)\n\t\t\t} else {\n\t\t\t\treturn fmt.Errorf(\"Partition %d didn't match: %v\", part.Number, matchErr)\n\t\t\t}\n\t\tcase exists && shouldExist && wipeEntry && !matches:\n\t\t\ts.Logger.Info(\"partition %d did not meet specifications, wiping partition entry and recreating\", part.Number)\n\t\t\top.DeletePartition(part.Number)\n\t\t\top.CreatePartition(part)\n\t\tdefault:\n\t\t\t// unfortunatey, golang doesn't check that all cases are handled exhaustively\n\t\t\treturn fmt.Errorf(\"Unreachable code reached when processing partition %d. golang--\", part.Number)\n\t\t}\n\t}\n\n\tif err := op.Commit(); err != nil {\n\t\treturn fmt.Errorf(\"commit failure: %v\", err)\n\t}\n\treturn nil\n}", "func (m *KMount) Close() error {\n\tvar umntErr error\n\tfor i := 0; i < 4; i++ {\n\t\tif umntErr = syscall.Unmount(m.mntPoint, syscall.MNT_DETACH); umntErr != nil {\n\t\t\ttime.Sleep(250 * time.Millisecond)\n\t\t\tcontinue\n\t\t}\n\t\tbreak\n\t}\n\tif umntErr != nil {\n\t\treturn fmt.Errorf(\"unmount failed: %v\", umntErr)\n\t}\n\n\tif err := m.loop.Detach(); err != nil {\n\t\treturn fmt.Errorf(\"loopback detach failed: %v\", err)\n\t}\n\tif m.mntPoint != \"\" {\n\t\treturn os.Remove(m.mntPoint)\n\t}\n\treturn nil\n}", "func (*DcsDisk) Descriptor() ([]byte, []int) {\n\treturn file_dcs_model_proto_rawDescGZIP(), []int{1}\n}", "func (d *detacherDefaults) WaitForDetach(devicePath string, timeout time.Duration) error {\n\tklog.Warning(logPrefix(d.plugin.flexVolumePlugin), \"using default WaitForDetach for device \", devicePath)\n\treturn nil\n}", "func (d *Driver) Unmount(mountDir string) {\n\tDebug(\"findmnt: \" + mountDir)\n\t_, err := RunCommand(\"findmnt\", \"-n\", \"-o\", \"SOURCE\", \"--target\", mountDir)\n\tif err != nil {\n\t\tDebug(err.Error())\n\t}\n\n\tDebug(\"syscall.Unmount: \" + mountDir)\n\tif err := syscall.Unmount(mountDir, 0); err != nil {\n\t\tFailure(err)\n\t}\n\n\tDebug(\"Detach hetzner volume from server\")\n\tvolume := GetVolume(d.client, d.options.PVOrVolumeName)\n\t_, _, errDetach := d.client.Volume.Detach(context.Background(), volume)\n\n\tif errDetach != nil {\n\t\tFailure(errDetach)\n\t}\n\n\t// Delete json file with token in it\n\t//Debug(\"os.Remove\")\n\t//if err := os.Remove(jsonOptionsFile); err != nil {\n\t//\tfailure(err)\n\t//}\n\n\tSuccess()\n}", "func UnmountPartition(cmd CommandS) {\n\tpmounted := Mounted{}\n\tflgfound := false\n\tindex := 0\n\tfmt.Println(\"\\n===== DESMONTAR PARTICION =======================================\")\n\tfor _, param := range cmd.Params {\n\t\tidp := param.Value\n\t\tfor i, mp := range sliceMP {\n\t\t\tidm := \"vd\" + string(mp.Letter) + strconv.FormatInt(mp.Number, 10)\n\t\t\tif idp == idm {\n\t\t\t\tflgfound = true\n\t\t\t\tindex = i\n\t\t\t\tpmounted = mp\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\t//Validaciones\n\t\tif flgfound {\n\t\t\tmbr := readMBR(pmounted.Path)\n\t\t\tvar bname [16]byte\n\t\t\tcopy(bname[:], pmounted.Name)\n\t\t\tfor i, p := range mbr.MbrPartitions {\n\t\t\t\tif bname == p.PartName {\n\t\t\t\t\tmbr.MbrPartitions[i] = pmounted.Part\n\t\t\t\t\twriteMBR(pmounted.Path, mbr)\n\t\t\t\t\tif index == 0 && len(sliceMP) > 1 {\n\t\t\t\t\t\tsliceMP = sliceMP[1:len(sliceMP)]\n\t\t\t\t\t} else if index == 0 && len(sliceMP) == 0 {\n\t\t\t\t\t\tsliceMP = make([]Mounted, 0)\n\t\t\t\t\t} else {\n\t\t\t\t\t\tsliceMP = RemoveMountedPartition(sliceMP, index)\n\t\t\t\t\t}\n\t\t\t\t\tfmt.Println(\"*** Particion\", idp, \" Desmontada ***\")\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\tif pmounted.Part.PartType == 'l' {\n\t\t\t\tif index == 0 && len(sliceMP) > 1 {\n\t\t\t\t\tsliceMP = sliceMP[1:len(sliceMP)]\n\t\t\t\t} else if index == 0 && len(sliceMP) == 0 {\n\t\t\t\t\tsliceMP = make([]Mounted, 0)\n\t\t\t\t} else {\n\t\t\t\t\tsliceMP = RemoveMountedPartition(sliceMP, index)\n\t\t\t\t}\n\t\t\t\tfmt.Println(\"*** Particion\", idp, \" Desmontada ***\")\n\t\t\t}\n\t\t} else {\n\t\t\tfmt.Println(\"[!] No se ha montado la particion con el id\", idp, \"...\")\n\t\t}\n\t\t//Reinicio de valores...\n\t\tpmounted = Mounted{}\n\t\tflgfound = false\n\t\tindex = 0\n\t}\n\tfmt.Println(\"================================================================\")\n}", "func (o LocalCopyPtrOutput) Disk() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *LocalCopy) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Disk\n\t}).(pulumi.StringPtrOutput)\n}", "func getVirtualHardDisk(c *wssdcloudstorage.VirtualHardDisk, group string) *storage.VirtualHardDisk {\n\treturn &storage.VirtualHardDisk{\n\t\tName: &c.Name,\n\t\tID: &c.Id,\n\t\tVersion: &c.Status.Version.Number,\n\t\tVirtualHardDiskProperties: &storage.VirtualHardDiskProperties{\n\t\t\tStatuses: status.GetStatuses(c.GetStatus()),\n\t\t\tDiskSizeBytes: &c.Size,\n\t\t\tDynamic: &c.Dynamic,\n\t\t\tBlocksizebytes: &c.Blocksizebytes,\n\t\t\tLogicalsectorbytes: &c.Logicalsectorbytes,\n\t\t\tPhysicalsectorbytes: &c.Physicalsectorbytes,\n\t\t\tControllernumber: &c.Controllernumber,\n\t\t\tControllerlocation: &c.Controllerlocation,\n\t\t\tDisknumber: &c.Disknumber,\n\t\t\tVirtualMachineName: &c.VirtualmachineName,\n\t\t\tScsipath: &c.Scsipath,\n\t\t\tHyperVGeneration: c.HyperVGeneration,\n\t\t\tDiskFileFormat: c.DiskFileFormat,\n\t\t},\n\t\tTags: tags.ProtoToMap(c.Tags),\n\t}\n}", "func (f *Pub) Detach(rm Publisher) {\n\tf.enders.Disable(rm)\n\tf.branches.Disable(rm)\n}", "func (s *gceOps) waitForDetach(\n\tdiskURL string,\n\ttimeout time.Duration,\n) error {\n\n\t_, err := task.DoRetryWithTimeout(\n\t\tfunc() (interface{}, bool, error) {\n\t\t\tinst, err := s.describeinstance()\n\t\t\tif err != nil {\n\t\t\t\treturn nil, true, err\n\t\t\t}\n\n\t\t\tfor _, d := range inst.Disks {\n\t\t\t\tif d.Source == diskURL {\n\t\t\t\t\treturn nil, true,\n\t\t\t\t\t\tfmt.Errorf(\"disk: %s is still attached to instance: %s\",\n\t\t\t\t\t\t\tdiskURL, s.inst.name)\n\t\t\t\t}\n\t\t\t}\n\n\t\t\treturn nil, false, nil\n\n\t\t},\n\t\tcloudops.ProviderOpsTimeout,\n\t\tcloudops.ProviderOpsRetryInterval)\n\n\treturn err\n}", "func Detach(ctx context.Context) context.Context {\n\treturn detachedContext{context.Background(), ctx}\n}", "func Deduped(vdiskID string, blockSize, lbaCacheLimit int64, cluster, templateCluster ardb.StorageCluster) (BlockStorage, error) {\n\t// define the LBA cache limit\n\tcacheLimit := lbaCacheLimit\n\tif cacheLimit < lba.BytesPerSector {\n\t\tlog.Infof(\n\t\t\t\"LBACacheLimit (%d) will be defaulted to %d (min-capped)\",\n\t\t\tcacheLimit, lba.BytesPerSector)\n\t\tcacheLimit = ardb.DefaultLBACacheLimit\n\t}\n\n\t// create the LBA (used to store deduped metadata)\n\tlbaStorage := newLBASectorStorage(vdiskID, cluster)\n\tvlba, err := lba.NewLBA(cacheLimit, lbaStorage)\n\tif err != nil {\n\t\tlog.Errorf(\"couldn't create the LBA: %s\", err.Error())\n\t\treturn nil, err\n\t}\n\n\tdedupedStorage := &dedupedStorage{\n\t\tblockSize: blockSize,\n\t\tvdiskID: vdiskID,\n\t\tzeroContentHash: zerodisk.HashBytes(make([]byte, blockSize)),\n\t\tcluster: cluster,\n\t\tlba: vlba,\n\t}\n\n\t// getContent is ALWAYS defined,\n\t// but the actual function used depends on\n\t// whether or not this storage has template support.\n\tif isInterfaceValueNil(templateCluster) {\n\t\tdedupedStorage.getContent = dedupedStorage.getPrimaryContent\n\t} else {\n\t\tdedupedStorage.getContent = dedupedStorage.getPrimaryOrTemplateContent\n\t\tdedupedStorage.templateCluster = templateCluster\n\t}\n\n\treturn dedupedStorage, nil\n}", "func (driver *Driver) Unmount(volumeName, volumeID string) error {\n\tif volumeName == \"\" && volumeID == \"\" {\n\t\treturn errors.New(\"Missing volume name or ID\")\n\t}\n\n\tinstances, err := driver.sdm.GetInstance()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tswitch {\n\tcase len(instances) == 0:\n\t\treturn errors.New(\"No instances\")\n\tcase len(instances) > 1:\n\t\treturn errors.New(\"Too many instances returned, limit the storagedrivers\")\n\t}\n\n\tvolumes, err := driver.sdm.GetVolume(volumeID, volumeName)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tswitch {\n\tcase len(volumes) == 0:\n\t\treturn errors.New(\"No volumes returned by name\")\n\tcase len(volumes) > 1:\n\t\treturn errors.New(\"Multiple volumes returned by name\")\n\t}\n\n\tvolumeAttachment, err := driver.sdm.GetVolumeAttach(volumes[0].VolumeID, instances[0].InstanceID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif len(volumeAttachment) == 0 {\n\t\treturn nil\n\t}\n\n\tmounts, err := driver.osdm.GetMounts(volumeAttachment[0].DeviceName, \"\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif len(mounts) == 0 {\n\t\treturn nil\n\t}\n\n\terr = driver.osdm.Unmount(mounts[0].Mountpoint)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = driver.sdm.DetachVolume(false, volumes[0].VolumeID, \"\")\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n\n}", "func (d *Domain) DetachDevice(x string, flags libvirt.DomainDeviceModifyFlags) error {\n\treq := libvirt.RemoteDomainDetachDeviceFlagsReq{\n\t\tDomain: d.RemoteDomain,\n\t\tXml: x,\n\t\tFlags: uint32(flags)}\n\n\tbuf, err := encode(&req)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tresp, err := d.l.send(libvirt.RemoteProcDomainDetachDeviceFlags, 0, libvirt.MessageTypeCall, libvirt.RemoteProgram, libvirt.MessageStatusOK, &buf)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tr := <-resp\n\tif r.Header.Status != libvirt.MessageStatusOK {\n\t\treturn decodeError(r.Payload)\n\t}\n\n\treturn nil\n}", "func (l *Libvirt) DomainDetachDevice(Dom Domain, XML string) (err error) {\n\tvar buf []byte\n\n\targs := DomainDetachDeviceArgs {\n\t\tDom: Dom,\n\t\tXML: XML,\n\t}\n\n\tbuf, err = encode(&args)\n\tif err != nil {\n\t\treturn\n\t}\n\n\n\t_, err = l.requestStream(13, constants.Program, buf, nil, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func getVirtualDeviceByDiskID(ctx context.Context, vm *object.VirtualMachine, diskID string) (vim25types.BaseVirtualDevice, error) {\n\tvmname, err := vm.Common.ObjectName(ctx)\n\tgomega.Expect(err).NotTo(gomega.HaveOccurred())\n\tvmDevices, err := vm.Device(ctx)\n\tif err != nil {\n\t\tframework.Logf(\"Failed to get the devices for VM: %q. err: %+v\", vmname, err)\n\t\treturn nil, err\n\t}\n\tfor _, device := range vmDevices {\n\t\tif vmDevices.TypeName(device) == \"VirtualDisk\" {\n\t\t\tif virtualDisk, ok := device.(*vim25types.VirtualDisk); ok {\n\t\t\t\tif virtualDisk.VDiskId != nil && virtualDisk.VDiskId.Id == diskID {\n\t\t\t\t\tframework.Logf(\"Found FCDID %q attached to VM %q\", diskID, vmname)\n\t\t\t\t\treturn device, nil\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tframework.Logf(\"Failed to find FCDID %q attached to VM %q\", diskID, vmname)\n\treturn nil, nil\n}", "func (a *AWS) DetachVolume(ctx *lepton.Context, instanceName, name string) error {\n\tvol, err := a.findVolumeByName(name)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tinstance, err := a.findInstanceByName(instanceName)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tinput := &ec2.DetachVolumeInput{\n\t\tInstanceId: aws.String(*instance.InstanceId),\n\t\tVolumeId: aws.String(*vol.VolumeId),\n\t}\n\n\t_, err = a.ec2.DetachVolume(input)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *DefaultChecker) DiskPerf() (info DiskPerfInfo, err error) {\n\tvar data []byte\n\n\texe := execCommand(\"dd\", \"bs=1M\", \"count=256\", \"if=/dev/zero\", fmt.Sprintf(\"of=%v\", path.Join(c.WritePath, \"test\")), \"conv=fdatasync\")\n\tdata, err = exe.CombinedOutput()\n\tgohtypes.PanicIfError(\"Not possible to execute the 'dd' command.\", 500, err)\n\n\tout := string(data)\n\tlogrus.Info(out)\n\n\trate := strings.Split(strings.Split(strings.Replace(out, \"\\n\", \"\", -1), \" s, \")[1], \" \")\n\n\tinfo = DiskPerfInfo{\n\t\tWriteSpeed: rate[0],\n\t\tUnit: rate[1],\n\t}\n\treturn\n}", "func (plugin *azureDataDiskPlugin) NewDetacher() (volume.Detacher, error) {\n\tazure, err := getAzureCloudProvider(plugin.host.GetCloudProvider())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &azureDiskDetacher{\n\t\tmounter: plugin.host.GetMounter(),\n\t\tazureProvider: azure,\n\t}, nil\n}", "func (d *detacherDefaults) UnmountDevice(deviceMountPath string) error {\n\tklog.Warning(logPrefix(d.plugin.flexVolumePlugin), \"using default UnmountDevice for device mount path \", deviceMountPath)\n\treturn mount.CleanupMountPoint(deviceMountPath, d.plugin.host.GetMounter(d.plugin.GetPluginName()), false)\n}", "func (s *SceneUtils) Detach(child, parent, scene float64) *SceneUtils {\n\ts.p.Call(\"detach\", child, parent, scene)\n\treturn s\n}", "func GoDetached(ctx context.Context, fun func(ctx context.Context)) {\n\tgo func() {\n\t\tfun(Detach(ctx))\n\t}()\n}", "func (na *cnmNetworkAllocator) DeallocateAttachment(node *api.Node, networkAttachment *api.NetworkAttachment) error {\n\n\tdelete(na.nodes[node.ID], networkAttachment.Network.ID)\n\tif len(na.nodes[node.ID]) == 0 {\n\t\tdelete(na.nodes, node.ID)\n\t}\n\n\treturn na.releaseEndpoints([]*api.NetworkAttachment{networkAttachment})\n}", "func (o AttachedDiskInitializeParamsOutput) DiskName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v AttachedDiskInitializeParams) *string { return v.DiskName }).(pulumi.StringPtrOutput)\n}", "func (f *FakeInstance) DetachVPC(_ context.Context, _, _ string) error {\n\tpanic(\"implement me\")\n}", "func DeleteVM(c config.Cpi, extInput bosh.MethodArguments) error {\n\tvar cid string\n\tif reflect.TypeOf(extInput[0]) != reflect.TypeOf(cid) {\n\t\treturn errors.New(\"Received unexpected type for vm cid\")\n\t}\n\n\tcid = extInput[0].(string)\n\tnode, err := rackhdapi.GetNodeByVMCID(c, cid)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif node.PersistentDisk.IsAttached {\n\t\terr = rackhdapi.MakeDiskRequest(c, node, false)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tworkflowName, err := workflows.PublishDeprovisionNodeWorkflow(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = workflows.RunDeprovisionNodeWorkflow(c, node.ID, workflowName, cid)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor _, tag := range node.Tags {\n\t\tif strings.HasPrefix(tag, DiskCIDTagPrefix) {\n\t\t\treturn nil\n\t\t}\n\t}\n\n\terr = rackhdapi.ReleaseNode(c, node.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func filterOutVmOsDisk(attachedDisks []attachedDisk, machine *machinev1.Machine) []attachedDisk {\n\tvar disks []attachedDisk\n\n\tfor _, disk := range attachedDisks {\n\t\tif strings.HasSuffix(disk.fileName, fmt.Sprintf(\"/%s.vmdk\", machine.GetName())) {\n\t\t\tcontinue\n\t\t}\n\t\tdisks = append(disks, disk)\n\t}\n\treturn disks\n}", "func detachENI(ec2Client *ec2.EC2, attachment *ec2.AttachNetworkInterfaceOutput) error {\n\t_, err := ec2Client.DetachNetworkInterface(&ec2.DetachNetworkInterfaceInput{\n\t\tAttachmentId: attachment.AttachmentId,\n\t\tForce: aws.Bool(true),\n\t})\n\n\tif err != nil {\n\t\terrors.Wrapf(err, \"unable to detach ENI\")\n\t}\n\treturn nil\n}", "func (_m *OpenStackMock) DetachVolume(instanceID string, volumeID string) error {\n\tret := _m.Called(instanceID, volumeID)\n\n\tvar r0 error\n\tif rf, ok := ret.Get(0).(func(string, string) error); ok {\n\t\tr0 = rf(instanceID, volumeID)\n\t} else {\n\t\tr0 = ret.Error(0)\n\t}\n\n\treturn r0\n}", "func (o AttachedDiskInitializeParamsPtrOutput) DiskName() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *AttachedDiskInitializeParams) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.DiskName\n\t}).(pulumi.StringPtrOutput)\n}", "func DiskMountClear(mntPath string) error {\n\tif shared.PathExists(mntPath) {\n\t\tif filesystem.IsMountPoint(mntPath) {\n\t\t\terr := unix.Unmount(mntPath, unix.MNT_DETACH)\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"Failed unmounting %q: %w\", mntPath, err)\n\t\t\t}\n\t\t}\n\n\t\terr := os.Remove(mntPath)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Failed removing %q: %w\", mntPath, err)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o InstanceFromTemplateOutput) BootDisk() InstanceFromTemplateBootDiskOutput {\n\treturn o.ApplyT(func(v *InstanceFromTemplate) InstanceFromTemplateBootDiskOutput { return v.BootDisk }).(InstanceFromTemplateBootDiskOutput)\n}", "func Detach(c *golangsdk.ServiceClient, volumeId string, opts DetachOpts) (*jobs.Job, error) {\n\turl := detachURL(c, opts.ServerId, volumeId)\n\tquery, err := golangsdk.BuildQueryString(opts)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\turl += query.String()\n\n\tvar r jobs.Job\n\t_, err = c.DeleteWithBody(url, &r, &golangsdk.RequestOpts{\n\t\tMoreHeaders: requestOpts.MoreHeaders,\n\t})\n\treturn &r, err\n}", "func (c *conn) Detach() DetachedConn {\n\tc.detached = true\n\tcmds := c.cmds\n\tc.cmds = nil\n\treturn &detachedConn{conn: c, cmds: cmds}\n}", "func NewCmdDiskAttach() *cobra.Command {\n\tvar async *bool\n\tvar udiskIDs *[]string\n\n\treq := base.BizClient.NewAttachUDiskRequest()\n\tcmd := &cobra.Command{\n\t\tUse: \"attach\",\n\t\tShort: \"Attach udisk instances to an uhost\",\n\t\tLong: \"Attach udisk instances to an uhost\",\n\t\tExample: \"ucloud udisk attach --uhost-id uhost-xxxx --udisk-id bs-xxx1,bs-xxx2\",\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\tfor _, id := range *udiskIDs {\n\t\t\t\tid = base.PickResourceID(id)\n\t\t\t\treq.UDiskId = &id\n\t\t\t\t*req.UHostId = base.PickResourceID(*req.UHostId)\n\t\t\t\tresp, err := base.BizClient.AttachUDisk(req)\n\t\t\t\tif err != nil {\n\t\t\t\t\tbase.HandleError(err)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\ttext := fmt.Sprintf(\"udisk[%s] is attaching to uhost uhost[%s]\", *req.UDiskId, *req.UHostId)\n\t\t\t\tif *async {\n\t\t\t\t\tbase.Cxt.Println(text)\n\t\t\t\t} else {\n\t\t\t\t\tpollDisk(resp.UDiskId, *req.ProjectId, *req.Region, *req.Zone, text, []string{status.DISK_INUSE, status.DISK_FAILED})\n\t\t\t\t}\n\t\t\t}\n\t\t},\n\t}\n\tflags := cmd.Flags()\n\tflags.SortFlags = false\n\treq.UHostId = flags.String(\"uhost-id\", \"\", \"Required. Resource ID of the uhost instance which you want to attach the disk\")\n\tudiskIDs = flags.StringSlice(\"udisk-id\", nil, \"Required. Resource ID of the udisk instances to attach\")\n\treq.ProjectId = flags.String(\"project-id\", base.ConfigInstance.ProjectID, \"Optional. Assign project-id\")\n\treq.Region = flags.String(\"region\", base.ConfigInstance.Region, \"Optional. Assign region\")\n\treq.Zone = flags.String(\"zone\", base.ConfigInstance.Zone, \"Optional. Assign availability zone\")\n\tasync = flags.Bool(\"async\", false, \"Optional. Do not wait for the long-running operation to finish.\")\n\n\tflags.SetFlagValuesFunc(\"udisk-id\", func() []string {\n\t\treturn getDiskList([]string{status.DISK_AVAILABLE}, *req.ProjectId, *req.Region, *req.Zone)\n\t})\n\tflags.SetFlagValuesFunc(\"uhost-id\", func() []string {\n\t\treturn getUhostList([]string{status.HOST_RUNNING, status.HOST_STOPPED}, *req.ProjectId, *req.Region, *req.Zone)\n\t})\n\n\tcmd.MarkFlagRequired(\"uhost-id\")\n\tcmd.MarkFlagRequired(\"udisk-id\")\n\n\treturn cmd\n}", "func (us *UserStorage) DetachDeviceToken(token string) error {\n\treturn nil\n}", "func (r *VmwareMapper) MapDisk(vmSpec *kubevirtv1.VirtualMachine, dv cdiv1.DataVolume) {\n\tname := fmt.Sprintf(\"dv-%v\", dv.Name)\n\tname = utils.EnsureLabelValueLength(name)\n\tvolume := kubevirtv1.Volume{\n\t\tName: name,\n\t\tVolumeSource: kubevirtv1.VolumeSource{\n\t\t\tDataVolume: &kubevirtv1.DataVolumeSource{\n\t\t\t\tName: dv.Name,\n\t\t\t},\n\t\t},\n\t}\n\n\tkubevirtDisk := kubevirtv1.Disk{\n\t\tName: name,\n\t\tDiskDevice: kubevirtv1.DiskDevice{\n\t\t\tDisk: &kubevirtv1.DiskTarget{\n\t\t\t\tBus: busTypeVirtio,\n\t\t\t},\n\t\t},\n\t}\n\n\tvmSpec.Spec.Template.Spec.Volumes = append(vmSpec.Spec.Template.Spec.Volumes, volume)\n\tdisks := append(vmSpec.Spec.Template.Spec.Domain.Devices.Disks, kubevirtDisk)\n\n\t// Since the import controller is iterating over a map of DVs,\n\t// MapDisk gets called for each DV in a nondeterministic order which results\n\t// in the disks being in an arbitrary order. This sort ensure the disks are\n\t// attached in the same order as the devices on the source VM.\n\tsort.Slice(disks, func(i, j int) bool {\n\t\treturn disks[i].Name < disks[j].Name\n\t})\n\tvmSpec.Spec.Template.Spec.Domain.Devices.Disks = disks\n}", "func DetachContextSegment(ctx context.Context) context.Context {\n\treturn &detachedContext{Context: ctx}\n}", "func (s *SelectableAttribute) Detach() {\n\ts.AttributeImpl.Detach()\n}", "func (us *UserStorage) DetachDeviceToken(token string) error {\n\treturn model.ErrorNotImplemented\n}", "func (us *UserStorage) DetachDeviceToken(token string) error {\n\treturn model.ErrorNotImplemented\n}", "func (in *Disk) DeepCopy() *Disk {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(Disk)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (in *Disk) DeepCopy() *Disk {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(Disk)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func setupNullDisk(t *testing.T, s *DiskIO, devName string) func() {\n\ttd, err := os.CreateTemp(\"\", \".telegraf.DiskInfoTest\")\n\trequire.NoError(t, err)\n\n\tif s.infoCache == nil {\n\t\ts.infoCache = make(map[string]diskInfoCache)\n\t}\n\tic, ok := s.infoCache[devName]\n\tif !ok {\n\t\t// No previous calls for the device were done, easy to poison the cache\n\t\ts.infoCache[devName] = diskInfoCache{\n\t\t\tmodifiedAt: 0,\n\t\t\tudevDataPath: td.Name(),\n\t\t\tvalues: map[string]string{},\n\t\t}\n\t}\n\torigUdevPath := ic.udevDataPath\n\n\tcleanFunc := func() {\n\t\tic.udevDataPath = origUdevPath\n\t\tos.Remove(td.Name())\n\t}\n\n\tic.udevDataPath = td.Name()\n\t_, err = td.Write(nullDiskInfo)\n\tif err != nil {\n\t\tcleanFunc()\n\t\tt.Fatal(err)\n\t}\n\n\treturn cleanFunc\n}", "func (g *Gandi) DetachDomain(fqdn string) (err error) {\n\t_, err = g.askGandi(mDELETE, \"domains/\"+fqdn, nil, nil)\n\treturn\n}", "func (d *Driver) internalUnmount(ctx context.Context, vol *smbVolume) error {\n\ttargetPath := getInternalMountPath(d.workingMountDir, vol)\n\n\t// Unmount smb server at base-dir\n\tklog.V(4).Infof(\"internally unmounting %v\", targetPath)\n\t_, err := d.NodeUnstageVolume(ctx, &csi.NodeUnstageVolumeRequest{\n\t\tVolumeId: vol.id,\n\t\tStagingTargetPath: targetPath,\n\t})\n\treturn err\n}" ]
[ "0.7060961", "0.64664584", "0.6421784", "0.63693166", "0.62287784", "0.6098612", "0.6074481", "0.59831476", "0.5969405", "0.5891906", "0.58823645", "0.58606005", "0.58583224", "0.58261454", "0.5794392", "0.57870233", "0.5748925", "0.5640547", "0.56379455", "0.5554582", "0.5533992", "0.55265564", "0.5521123", "0.5517761", "0.5509181", "0.54994786", "0.54939634", "0.5431497", "0.54205024", "0.53895056", "0.5366841", "0.53633225", "0.5357051", "0.5312432", "0.5279882", "0.5260372", "0.52459127", "0.5238477", "0.5183793", "0.5173291", "0.51304543", "0.5117953", "0.5085589", "0.5053479", "0.50056237", "0.5005146", "0.5000599", "0.4991671", "0.49720037", "0.49717802", "0.49641457", "0.4954671", "0.4953672", "0.49530247", "0.49153545", "0.4888618", "0.48806623", "0.4872151", "0.48542097", "0.48516706", "0.48365104", "0.48313832", "0.48276475", "0.482683", "0.48241922", "0.48168173", "0.48158804", "0.47979605", "0.47807312", "0.47740775", "0.47734427", "0.47691476", "0.47557747", "0.47548115", "0.4745524", "0.47434235", "0.47397742", "0.47370243", "0.4736644", "0.47297195", "0.47275126", "0.47274464", "0.4690591", "0.46900117", "0.46793282", "0.4678496", "0.46760905", "0.46658555", "0.46551734", "0.4638788", "0.46321237", "0.4627902", "0.46239215", "0.46194437", "0.46194437", "0.45828035", "0.45828035", "0.4576409", "0.4571763", "0.45711118" ]
0.7506682
0
CreateDisk uses the override method CreateDiskFn or the real implementation.
func (c *TestClient) CreateDisk(project, zone string, d *compute.Disk) error { if c.CreateDiskFn != nil { return c.CreateDiskFn(project, zone, d) } return c.client.CreateDisk(project, zone, d) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *Module) DiskCreate(name string, size gridtypes.Unit) (disk pkg.VDisk, err error) {\n\tpath, err := s.findDisk(name)\n\tif err == nil {\n\t\treturn disk, errors.Wrapf(os.ErrExist, \"disk with id '%s' already exists\", name)\n\t}\n\n\tbase, err := s.diskFindCandidate(size)\n\tif err != nil {\n\t\treturn disk, errors.Wrapf(err, \"failed to find a candidate to host vdisk of size '%d'\", size)\n\t}\n\n\tpath, err = s.safePath(base, name)\n\tif err != nil {\n\t\treturn disk, err\n\t}\n\n\tdefer func() {\n\t\t// clean up disk file if error\n\t\tif err != nil {\n\t\t\tos.RemoveAll(path)\n\t\t}\n\t}()\n\n\tdefer syscall.Sync()\n\n\tvar file *os.File\n\tfile, err = os.Create(path)\n\tif err != nil {\n\t\treturn disk, err\n\t}\n\n\tdefer file.Close()\n\tif err = chattr.SetAttr(file, chattr.FS_NOCOW_FL); err != nil {\n\t\treturn disk, err\n\t}\n\n\tif err = syscall.Fallocate(int(file.Fd()), 0, 0, int64(size)); err != nil {\n\t\treturn disk, errors.Wrap(err, \"failed to truncate disk to size\")\n\t}\n\n\treturn pkg.VDisk{Path: path, Size: int64(size)}, nil\n}", "func CreateDisk(form resources.DiskForm) {\n\tdisk := model.Disk{\n\t\tSize: form.Size,\n\t\tLeft: form.Left,\n\t\tComputerID: form.ComputerID,\n\t}\n\tnow := time.Now()\n\tdisk.UpdatedAt = now\n\tdisk.CreatedAt = now\n\tdb.Session().Create(&disk)\n}", "func DiskFactory() worker.Worker {\n\treturn &Disk{}\n}", "func NewCmdDiskCreate() *cobra.Command {\n\tvar async *bool\n\tvar count *int\n\treq := base.BizClient.NewCreateUDiskRequest()\n\tenableDataArk := sdk.String(\"false\")\n\tcmd := &cobra.Command{\n\t\tUse: \"create\",\n\t\tShort: \"Create udisk instance\",\n\t\tLong: \"Create udisk instance\",\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\tif *count > 10 || *count < 1 {\n\t\t\t\tbase.Cxt.Printf(\"Error, count should be between 1 and 10\\n\")\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif *enableDataArk == \"true\" {\n\t\t\t\treq.UDataArkMode = sdk.String(\"Yes\")\n\t\t\t} else {\n\t\t\t\treq.UDataArkMode = sdk.String(\"No\")\n\t\t\t}\n\n\t\t\tif *req.DiskType == \"Oridinary\" {\n\t\t\t\t*req.DiskType = \"DataDisk\"\n\t\t\t} else if *req.DiskType == \"SSD\" {\n\t\t\t\t*req.DiskType = \"SSDDataDisk\"\n\t\t\t}\n\t\t\tfor i := 0; i < *count; i++ {\n\t\t\t\tresp, err := base.BizClient.CreateUDisk(req)\n\t\t\t\tif err != nil {\n\t\t\t\t\tbase.HandleError(err)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tif count := len(resp.UDiskId); count == 1 {\n\t\t\t\t\ttext := fmt.Sprintf(\"udisk:%v is initializing\", resp.UDiskId)\n\t\t\t\t\tif *async {\n\t\t\t\t\t\tbase.Cxt.Println(text)\n\t\t\t\t\t} else {\n\t\t\t\t\t\tpollDisk(resp.UDiskId[0], *req.ProjectId, *req.Region, *req.Zone, text, []string{status.DISK_AVAILABLE, status.DISK_FAILED})\n\t\t\t\t\t}\n\t\t\t\t} else if count > 1 {\n\t\t\t\t\tbase.Cxt.Printf(\"udisk:%v created\\n\", resp.UDiskId)\n\t\t\t\t} else {\n\t\t\t\t\tbase.Cxt.PrintErr(fmt.Errorf(\"none udisk created\"))\n\t\t\t\t}\n\t\t\t}\n\t\t},\n\t}\n\tflags := cmd.Flags()\n\tflags.SortFlags = false\n\treq.Name = flags.String(\"name\", \"\", \"Required. Name of the udisk to create\")\n\treq.Size = flags.Int(\"size-gb\", 10, \"Required. Size of the udisk to create. Unit:GB. Normal udisk [1,8000]; SSD udisk [1,4000] \")\n\treq.ProjectId = flags.String(\"project-id\", base.ConfigInstance.ProjectID, \"Optional. Assign project-id\")\n\treq.Region = flags.String(\"region\", base.ConfigInstance.Region, \"Optional. Assign region\")\n\treq.Zone = flags.String(\"zone\", base.ConfigInstance.Zone, \"Optional. Assign availability zone\")\n\treq.ChargeType = flags.String(\"charge-type\", \"Dynamic\", \"Optional.'Year',pay yearly;'Month',pay monthly;'Dynamic', pay hourly\")\n\treq.Quantity = flags.Int(\"quantity\", 1, \"Optional. The duration of the instance. N years/months.\")\n\tenableDataArk = flags.String(\"enable-data-ark\", \"false\", \"Optional. DataArk supports real-time backup, which can restore the udisk back to any moment within the last 12 hours.\")\n\treq.Tag = flags.String(\"group\", \"Default\", \"Optional. Business group\")\n\treq.DiskType = flags.String(\"udisk-type\", \"Oridinary\", \"Optional. 'Ordinary' or 'SSD'\")\n\treq.CouponId = flags.String(\"coupon-id\", \"\", \"Optional. Coupon ID, The Coupon can deduct part of the payment.See https://accountv2.ucloud.cn\")\n\tasync = flags.Bool(\"async\", false, \"Optional. Do not wait for the long-running operation to finish.\")\n\tcount = flags.Int(\"count\", 1, \"Optional. The count of udisk to create. Range [1,10]\")\n\n\tflags.SetFlagValues(\"charge-type\", \"Month\", \"Year\", \"Dynamic\", \"Trial\")\n\tflags.SetFlagValues(\"enable-data-ark\", \"true\", \"false\")\n\tflags.SetFlagValues(\"udisk-type\", \"Oridinary\", \"SSD\")\n\n\tcmd.MarkFlagRequired(\"size-gb\")\n\tcmd.MarkFlagRequired(\"name\")\n\n\treturn cmd\n}", "func (s *StackEbrc) CreateVolume(request abstract.VolumeRequest) (*abstract.Volume, fail.Error) {\n\tdiskCreateParams := &types.DiskCreateParams{\n\t\tDisk: &types.Disk{\n\t\t\tName: request.Name,\n\t\t\tSize: int(request.Size * 1024 * 1024 * 1024),\n\t\t\tBusType: \"6\",\n\t\t\tBusSubType: \"lsilogicsas\",\n\t\t},\n\t}\n\n\t_, vdc, err := s.getOrgVdc()\n\tif err != nil {\n\t\treturn nil, fail.Wrap(err, fmt.Sprintf(\"Error creating volume\"))\n\t}\n\n\tstorageProfileValue := \"\"\n\tfor _, sps := range vdc.Vdc.VdcStorageProfiles {\n\t\tfor _, sp := range sps.VdcStorageProfile {\n\t\t\tstorageProfileValue = sp.Name\n\t\t}\n\t}\n\n\tvar storageReference types.Reference\n\tif storageProfileValue != \"\" {\n\t\tstorageReference, err = vdc.FindStorageProfileReference(storageProfileValue)\n\t\tif err != nil {\n\t\t\treturn nil, fail.Errorf(fmt.Sprintf(\"error finding storage profile %s\", storageProfileValue), err)\n\t\t}\n\t\tdiskCreateParams.Disk.StorageProfile = &types.Reference{HREF: storageReference.HREF}\n\t}\n\n\ttask, err := vdc.CreateDisk(diskCreateParams)\n\tif err != nil {\n\t\treturn nil, fail.Errorf(fmt.Sprintf(\"error creating independent disk: %s\", err), err)\n\t}\n\n\terr = task.WaitTaskCompletion()\n\tif err != nil {\n\t\treturn nil, fail.Errorf(fmt.Sprintf(\"error waiting to finish creation of independent disk: %s\", err), err)\n\t}\n\n\tdrec, err := vdc.QueryDisk(request.Name)\n\tif err != nil {\n\t\treturn nil, fail.Errorf(fmt.Sprintf(\"error creating independent disk: %s\", err), err)\n\t}\n\tdisk, err := vdc.FindDiskByHREF(drec.Disk.HREF)\n\tif err != nil {\n\t\treturn nil, fail.Errorf(fmt.Sprintf(\"unable to find disk by reference: %s\", err), err)\n\t}\n\n\trevol := &abstract.Volume{\n\t\tID: disk.Disk.Id,\n\t\tName: disk.Disk.Name,\n\t\tSize: disk.Disk.Size,\n\t}\n\n\treturn revol, nil\n}", "func (r opResult) createDiskAccount(\n\tctx context.Context, flowCtx *execinfra.FlowCtx, opName string, processorID int32,\n) *mon.BoundAccount {\n\tmonitorName := r.getMemMonitorName(opName, processorID, \"disk\" /* suffix */)\n\topDiskMonitor := execinfra.NewMonitor(ctx, flowCtx.DiskMonitor, monitorName)\n\tr.OpMonitors = append(r.OpMonitors, opDiskMonitor)\n\topDiskAccount := opDiskMonitor.MakeBoundAccount()\n\tr.OpAccounts = append(r.OpAccounts, &opDiskAccount)\n\treturn &opDiskAccount\n}", "func NewDisk(radius float64, segments int) *Geometry {\n\treturn NewDiskSector(radius, segments, 0, 2*math.Pi)\n}", "func NewDisk(period uint, level uint8, opts interface{}) *Disk {\n\n\tallowedDisks := map[string]struct{}{}\n\tif opts != nil {\n\t\tif options, ok := opts.(map[string]interface{}); ok {\n\t\t\tif val, ok := options[\"names\"]; ok {\n\t\t\t\tif diskNames, ok := val.([]interface{}); ok {\n\t\t\t\t\tfor _, v := range diskNames {\n\t\t\t\t\t\tif diskName, ok := v.(string); ok {\n\t\t\t\t\t\t\tallowedDisks[diskName] = struct{}{}\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\tc := &Disk{\n\t\tlevel: level,\n\t\tperiod: period,\n\t\tallowedDisks: allowedDisks,\n\t}\n\n\tif level > 0 {\n\t\ttick := time.Tick(time.Duration(period) * time.Millisecond)\n\t\tgo func() {\n\t\t\tfor range tick {\n\t\t\t\tif err := c.scrape(); err != nil {\n\t\t\t\t\tlog.Error(err)\n\t\t\t\t}\n\t\t\t}\n\t\t}()\n\t}\n\n\treturn c\n}", "func (fsOnDisk) Create(name string) (File, error) { return os.Create(name) }", "func (d *Driver) Create() error {\n\n\tvolume := NodesNodeStorageStorageContentPostParameter{\n\t\tFilename: d.StorageFilename,\n\t\tSize: d.DiskSize + \"G\",\n\t\tVMID: d.VMID,\n\t}\n\n\td.debugf(\"Creating disk volume '%s' with size '%s'\", volume.Filename, volume.Size)\n\tdiskname, err := d.driver.NodesNodeStorageStorageContentPost(d.Node, d.Storage, &volume)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif !strings.HasSuffix(diskname, d.StorageFilename) {\n\t\treturn fmt.Errorf(\"returned diskname is not correct: should be '%s' but was '%s'\", d.StorageFilename, diskname)\n\t}\n\n\tnpp := NodesNodeQemuPostParameter{\n\t\tVMID: d.VMID,\n\t\tAgent: \"1\",\n\t\tAutostart: \"1\",\n\t\tMemory: d.Memory,\n\t\tCores: d.Cores,\n\t\tNet0: fmt.Sprintf(\"virtio,bridge=%s\", d.NetBridge),\n\t\tSCSI0: d.StorageFilename,\n\t\tOstype: \"l26\",\n\t\tName: d.BaseDriver.MachineName,\n\t\tKVM: \"1\", // if you test in a nested environment, you may have to change this to 0 if you do not have nested virtualization\n\t\tCdrom: d.ImageFile,\n\t\tPool: d.Pool,\n\t}\n\n\tif d.NetVlanTag != 0 {\n\t\tnpp.Net0 = fmt.Sprintf(\"virtio,bridge=%s,tag=%d\", d.NetBridge, d.NetVlanTag)\n\t}\n\n\tif d.StorageType == \"qcow2\" {\n\t\tnpp.SCSI0 = d.Storage + \":\" + d.VMID + \"/\" + volume.Filename\n\t} else if d.StorageType == \"raw\" {\n\t\tif strings.HasSuffix(volume.Filename, \".raw\") {\n\t\t\t// raw files (having .raw) should have the VMID in the path\n\t\t\tnpp.SCSI0 = d.Storage + \":\" + d.VMID + \"/\" + volume.Filename\n\t\t} else {\n\t\t\tnpp.SCSI0 = d.Storage + \":\" + volume.Filename\n\t\t}\n\t}\n\td.debugf(\"Creating VM '%s' with '%d' of memory\", npp.VMID, npp.Memory)\n\ttaskid, err := d.driver.NodesNodeQemuPost(d.Node, &npp)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = d.driver.WaitForTaskToComplete(d.Node, taskid)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = d.Start()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn d.waitAndPrepareSSH()\n}", "func New(diskPath string) (Disk, *probe.Error) {\n\tif diskPath == \"\" {\n\t\treturn Disk{}, probe.NewError(InvalidArgument{})\n\t}\n\tst, err := os.Stat(diskPath)\n\tif err != nil {\n\t\treturn Disk{}, probe.NewError(err)\n\t}\n\n\tif !st.IsDir() {\n\t\treturn Disk{}, probe.NewError(syscall.ENOTDIR)\n\t}\n\ts := syscall.Statfs_t{}\n\terr = syscall.Statfs(diskPath, &s)\n\tif err != nil {\n\t\treturn Disk{}, probe.NewError(err)\n\t}\n\tdisk := Disk{\n\t\tlock: &sync.Mutex{},\n\t\tpath: diskPath,\n\t\tfsInfo: make(map[string]string),\n\t}\n\tif fsType := getFSType(s.Type); fsType != \"UNKNOWN\" {\n\t\tdisk.fsInfo[\"FSType\"] = fsType\n\t\tdisk.fsInfo[\"MountPoint\"] = disk.path\n\t\treturn disk, nil\n\t}\n\treturn Disk{}, probe.NewError(UnsupportedFilesystem{Type: strconv.FormatInt(int64(s.Type), 10)})\n}", "func diskSpecForCreate(dss *object.HostDatastoreSystem, name string) (*types.VmfsDatastoreCreateSpec, error) {\n\tdisk, err := availableScsiDisk(dss, name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tctx, cancel := context.WithTimeout(context.Background(), defaultAPITimeout)\n\tdefer cancel()\n\toptions, err := dss.QueryVmfsDatastoreCreateOptions(ctx, disk.DevicePath)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"could not get disk creation options for %q: %s\", name, err)\n\t}\n\tvar option *types.VmfsDatastoreOption\n\tfor _, o := range options {\n\t\tif _, ok := o.Info.(*types.VmfsDatastoreAllExtentOption); ok {\n\t\t\toption = &o\n\t\t\tbreak\n\t\t}\n\t}\n\tif option == nil {\n\t\treturn nil, fmt.Errorf(\"device %q is not available as a new whole-disk device for datastore\", name)\n\t}\n\treturn option.Spec.(*types.VmfsDatastoreCreateSpec), nil\n}", "func makeDiskImage(dest string, size uint, initialBytes []byte) error {\n\t// Create the dest dir.\n\tif err := os.MkdirAll(filepath.Dir(dest), 0755); err != nil {\n\t\treturn err\n\t}\n\t// Fill in the magic string so boot2docker VM will detect this and format\n\t// the disk upon first boot.\n\traw := bytes.NewReader(initialBytes)\n\treturn MakeDiskImage(dest, size, raw)\n}", "func (c *Compute) Disk(name string) (string, error) {\n\tdisk, err := c.Disks.Get(c.Project, c.Zone, name).Do()\n\tif err == nil {\n\t\tlog.Printf(\"found existing root disk: %q\", disk.SelfLink)\n\t\treturn disk.SelfLink, nil\n\t}\n\tlog.Printf(\"not found, creating new root disk: %q\", name)\n\top, err := c.Disks.Insert(c.Project, c.Zone, &compute.Disk{\n\t\tName: name,\n\t}).SourceImage(*image).Do()\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"disk insert api call failed: %v\", err)\n\t}\n\tif err := c.wait(op); err != nil {\n\t\treturn \"\", fmt.Errorf(\"disk insert operation failed: %v\", err)\n\t}\n\tlog.Printf(\"root disk created: %q\", op.TargetLink)\n\treturn op.TargetLink, nil\n}", "func (m *DisksClientMock) CreateOrUpdate(ctx context.Context, resourceGroupName string, diskName string, diskParameter compute.Disk) *retry.Error {\n\treturn nil\n}", "func (d *Driver) Create() error {\n\tif err := d.verifyRootPermissions(); err != nil {\n\t\treturn err\n\t}\n\n\tif err := copyFile(d.ImageSourcePath, d.getDiskPath()); err != nil {\n\t\treturn err\n\t}\n\n\tif _, err := d.resizeDiskImageIfNeeded(d.DiskCapacity); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (d *Disk) CreateSinglePartition() error {\n\tfd, err := diskfs.Open(d.DevPath)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error opening disk fd for disk %s: %v\", d.DevPath, err)\n\t}\n\td.disk = fd\n\n\t// check for any existing partition table on the disk\n\tif _, err := d.disk.GetPartitionTable(); err == nil {\n\t\tklog.Errorf(\"aborting partition creation, disk %s already contains a known partition table\", d.DevPath)\n\t\treturn fmt.Errorf(\"disk %s contains a partition table, cannot create a single partition\", d.DevPath)\n\t}\n\n\t// check for any existing filesystem on the disk\n\tdeviceIdentifier := blkid.DeviceIdentifier{\n\t\tDevPath: d.DevPath,\n\t}\n\tif fs := deviceIdentifier.GetOnDiskFileSystem(); len(fs) != 0 {\n\t\tklog.Errorf(\"aborting partition creation, disk %s contains a known filesystem: %s\", d.DevPath, fs)\n\t\treturn fmt.Errorf(\"disk %s contains a known filesyste: %s, cannot create a single partition\", d.DevPath, fs)\n\t}\n\n\terr = d.createPartitionTable()\n\tif err != nil {\n\t\tklog.Error(\"partition table initialization failed\")\n\t\treturn err\n\t}\n\n\terr = d.addPartition()\n\tif err != nil {\n\t\tklog.Error(\"could not add a partition to partition table\")\n\t\treturn err\n\t}\n\n\terr = d.applyPartitionTable()\n\tif err != nil {\n\t\tklog.Error(\"writing partition table to disk failed\")\n\t\treturn err\n\t}\n\tklog.Infof(\"created a single partition on disk %s\", d.DevPath)\n\treturn nil\n}", "func NewDiskCookieJar(name, path string) CookieJar {\n\tjar := &DiskCookieJar{name: name, path: path}\n\n\treturn jar\n}", "func (of OperatorFactory) DiskOrder() types.DiskOrder {\n\treturn Operator{}.DiskOrder()\n}", "func (of OperatorFactory) DiskOrder() types.DiskOrder {\n\treturn operator{}.DiskOrder()\n}", "func New(options Options) *Diskv {\n\tif options.BasePath == \"\" {\n\t\toptions.BasePath = defaultBasePath\n\t}\n\tif options.Transform == nil {\n\t\toptions.Transform = defaultTransform\n\t}\n\tif options.PathPerm == 0 {\n\t\toptions.PathPerm = defaultPathPerm\n\t}\n\tif options.FilePerm == 0 {\n\t\toptions.FilePerm = defaultFilePerm\n\t}\n\n\td := &Diskv{\n\t\tOptions: options,\n\t\tcache: map[string][]byte{},\n\t\tcacheSize: 0,\n\t}\n\n\tif d.Index != nil && d.IndexLess != nil {\n\t\td.Index.Initialize(d.IndexLess, d.Keys())\n\t}\n\n\treturn d\n}", "func OpenDisk(filename string) (SectorDisk, error) {\n\text := strings.ToLower(path.Ext(filename))\n\tswitch ext {\n\tcase \".dsk\":\n\t\treturn LoadDSK(filename)\n\t}\n\treturn nil, fmt.Errorf(\"Unimplemented/unknown disk file extension %q\", ext)\n}", "func (cs *ControllerServer) CreateVolume(ctx context.Context, req *csi.CreateVolumeRequest) (*csi.CreateVolumeResponse, error) {\n\tname := req.GetName()\n\tif len(name) == 0 {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"CreateVolume name must be provided\")\n\t}\n\tif err := cs.validateVolumeCapabilities(req.GetVolumeCapabilities()); err != nil {\n\t\treturn nil, status.Error(codes.InvalidArgument, err.Error())\n\t}\n\n\treqCapacity := req.GetCapacityRange().GetRequiredBytes()\n\tnfsVol, err := cs.newNFSVolume(name, reqCapacity, req.GetParameters())\n\tif err != nil {\n\t\treturn nil, status.Error(codes.InvalidArgument, err.Error())\n\t}\n\n\tvar volCap *csi.VolumeCapability\n\tif len(req.GetVolumeCapabilities()) > 0 {\n\t\tvolCap = req.GetVolumeCapabilities()[0]\n\t} // 执行挂载 命令\n\t// Mount nfs base share so we can create a subdirectory\n\tif err = cs.internalMount(ctx, nfsVol, volCap); err != nil {\n\t\treturn nil, status.Errorf(codes.Internal, \"failed to mount nfs server: %v\", err.Error())\n\t}\n\tdefer func() {\n\t\tif err = cs.internalUnmount(ctx, nfsVol); err != nil {\n\t\t\tklog.Warningf(\"failed to unmount nfs server: %v\", err.Error())\n\t\t}\n\t}()\n\n\t// Create subdirectory under base-dir\n\t// TODO: revisit permissions\n\tinternalVolumePath := cs.getInternalVolumePath(nfsVol)\n\tif err = os.Mkdir(internalVolumePath, 0777); err != nil && !os.IsExist(err) {\n\t\treturn nil, status.Errorf(codes.Internal, \"failed to make subdirectory: %v\", err.Error())\n\t}\n\t// Remove capacity setting when provisioner 1.4.0 is available with fix for\n\t// https://github.com/kubernetes-csi/external-provisioner/pull/271\n\treturn &csi.CreateVolumeResponse{Volume: cs.nfsVolToCSI(nfsVol, reqCapacity)}, nil\n}", "func FileSystemCreate(f types.Filesystem) error {\n\tvar cmd *exec.Cmd\n\tvar debugCMD string\n\n\tswitch f.Mount.Format {\n\tcase \"swap\":\n\t\tcmd = exec.Command(\"/sbin/mkswap\", f.Mount.Device)\n\t\tdebugCMD = fmt.Sprintf(\"%s %s\", \"/sbin/mkswap\", f.Mount.Device)\n\tcase \"ext4\", \"ext3\", \"ext2\":\n\t\t// Add filesystem flags\n\t\tf.Mount.Create.Options = append(f.Mount.Create.Options, \"-t\")\n\t\tf.Mount.Create.Options = append(f.Mount.Create.Options, f.Mount.Format)\n\n\t\t// Add force\n\t\tf.Mount.Create.Options = append(f.Mount.Create.Options, \"-F\")\n\n\t\t// Add Device to formate\n\t\tf.Mount.Create.Options = append(f.Mount.Create.Options, f.Mount.Device)\n\n\t\t// Format disk\n\t\tcmd = exec.Command(\"/sbin/mke2fs\", f.Mount.Create.Options...)\n\t\tfor i := range f.Mount.Create.Options {\n\t\t\tdebugCMD = fmt.Sprintf(\"%s %s\", debugCMD, f.Mount.Create.Options[i])\n\t\t}\n\tcase \"vfat\":\n\t\tcmd = exec.Command(\"/sbin/mkfs.fat\", f.Mount.Device)\n\t\tdebugCMD = fmt.Sprintf(\"%s %s\", \"/sbin/mkfs.fat\", f.Mount.Device)\n\tdefault:\n\t\tlog.Warnf(\"Unknown filesystem type [%s]\", f.Mount.Format)\n\t}\n\tcmd.Stdin, cmd.Stdout, cmd.Stderr = os.Stdin, os.Stdout, os.Stderr\n\n\terr := cmd.Start()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Command [%s] Filesystem [%v]\", debugCMD, err)\n\t}\n\terr = cmd.Wait()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Command [%s] Filesystem [%v]\", debugCMD, err)\n\t}\n\n\treturn nil\n}", "func (c *MockDisksClient) CreateOrUpdate(ctx context.Context, resourceGroupName, diskName string, parameters compute.Disk) error {\n\t// Ignore resourceGroupName for simplicity.\n\tif _, ok := c.Disks[diskName]; ok {\n\t\treturn fmt.Errorf(\"update not supported\")\n\t}\n\tparameters.Name = &diskName\n\tc.Disks[diskName] = parameters\n\treturn nil\n}", "func NewHardDisk(size uint64, throughput uint64, prng *rand.Rand) Drive {\n\treturn &hardDisk{\n\t\tsize: size,\n\t\tthroughput: throughput,\n\t\tage: 0,\n\t\tstatus: OK,\n\t\tprng: prng,\n\t}\n}", "func (d *Disk) CreatePartitionTable() error {\n\tfd, err := diskfs.Open(d.DevPath)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error opening disk fd for disk %s: %v\", d.DevPath, err)\n\t}\n\td.disk = fd\n\n\t// check for any existing partition table on the disk\n\tif _, err := d.disk.GetPartitionTable(); err == nil {\n\t\tklog.Errorf(\"aborting partition creation, disk %s already contains a known partition table\", d.DevPath)\n\t\treturn fmt.Errorf(\"disk %s contains a partition table, cannot create a new partition table\", d.DevPath)\n\t}\n\n\t// check for any existing filesystem on the disk\n\tdeviceIdentifier := blkid.DeviceIdentifier{\n\t\tDevPath: d.DevPath,\n\t}\n\tif fs := deviceIdentifier.GetOnDiskFileSystem(); len(fs) != 0 {\n\t\tklog.Errorf(\"aborting partition creation, disk %s contains a known filesystem: %s\", d.DevPath, fs)\n\t\treturn fmt.Errorf(\"disk %s contains a known filesyste: %s, cannot create a partition table\", d.DevPath, fs)\n\t}\n\n\terr = d.createPartitionTable()\n\tif err != nil {\n\t\tklog.Error(\"partition table initialization failed\")\n\t\treturn err\n\t}\n\n\terr = d.disk.Partition(d.table)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to create/write partition table. %v\", err)\n\t}\n\tklog.Infof(\"created partition table on disk %s\", d.DevPath)\n\treturn nil\n}", "func formatDisk(device, fstype string) error {\n\tcmd := exec.Command(fmt.Sprintf(\"mkfs.%s\", fstype), device)\n\tif b, err := cmd.CombinedOutput(); err != nil {\n\t\treturn fmt.Errorf(\"failed to format %s: %s\", device, string(b))\n\t}\n\treturn nil\n}", "func NewMappedDisk(sd SectorDisk, logicalToPhysical []byte) (MappedDisk, error) {\n\tif logicalToPhysical != nil && len(logicalToPhysical) != int(sd.Sectors()) {\n\t\treturn MappedDisk{}, fmt.Errorf(\"NewMappedDisk called on a disk image with %d sectors per track, but a mapping of length %d\", sd.Sectors(), len(logicalToPhysical))\n\t}\n\tif logicalToPhysical == nil {\n\t\tlogicalToPhysical = make([]byte, int(sd.Sectors()))\n\t\tfor i := range logicalToPhysical {\n\t\t\tlogicalToPhysical[i] = byte(i)\n\t\t}\n\t}\n\treturn MappedDisk{\n\t\tsectorDisk: sd,\n\t\tlogicalToPhysical: logicalToPhysical,\n\t}, nil\n}", "func NewDiskMaker(configLocation, symLinkLocation string) *DiskMaker {\n\tt := &DiskMaker{}\n\tt.configLocation = configLocation\n\tt.symlinkLocation = symLinkLocation\n\tt.apiClient = newAPIUpdater()\n\tt.eventSync = newEventReporter(t.apiClient)\n\treturn t\n}", "func makeDiskImage(dest string, size int) error {\n\tlog.Printf(\"Creating %d MB hard disk image...\", size)\n\tcmd := exec.Command(B2D.Vbm, \"convertfromraw\", \"stdin\", dest, fmt.Sprintf(\"%d\", size*1024*1024), \"--format\", \"VMDK\")\n\tcmd.Stdout = os.Stdout\n\tcmd.Stderr = os.Stderr\n\tw, err := cmd.StdinPipe()\n\tif err != nil {\n\t\treturn err\n\t}\n\t// Write the magic string so the VM auto-formats the disk upon first boot.\n\tif _, err := w.Write([]byte(\"boot2docker, please format-me\")); err != nil {\n\t\treturn err\n\t}\n\tif err := w.Close(); err != nil {\n\t\treturn err\n\t}\n\n\treturn cmd.Run()\n}", "func (d *DefaultDriver) CreateVolume(volName string, size uint64, haLevel int64) (string, error) {\n\treturn \"\", &errors.ErrNotSupported{\n\t\tType: \"Function\",\n\t\tOperation: \"CreateVolume()\",\n\t}\n}", "func comandoMkdisk(comando string) {\n\tfmt.Println(\"\\nEJECUTANDO: \" + comando)\n\t//Descomponiendo en atributos\n\tatributos := strings.Split(comando, \" -\")\n\t//verificando parametros\n\tif len(atributos) > 3 {\n\t\tsize := 0\n\t\tpath := \"\"\n\t\tname := \"\"\n\t\tunit := \"\"\n\t\tfor i := 1; i < len(atributos); i++ {\n\t\t\tatributoActual := strings.Split(atributos[i], \"->\")\n\t\t\tswitch strings.ToLower(atributoActual[0]) {\n\t\t\tcase \"size\":\n\t\t\t\tsize = atributoSize(atributoActual[1])\n\t\t\tcase \"path\":\n\t\t\t\tpath = strings.ToLower(strings.TrimSpace(strings.ReplaceAll(atributoActual[1], \"\\\"\", \"\")))\n\t\t\tcase \"name\":\n\t\t\t\tname = strings.ToLower(strings.TrimSpace(strings.ReplaceAll(atributoActual[1], \"\\\"\", \"\")))\n\t\t\tcase \"unit\":\n\t\t\t\tunit = atributoUnit(atributoActual[1])\n\t\t\tdefault:\n\t\t\t\tfmt.Println(\"RESULTADO: El atributo \" + atributoActual[0] + \" no se reconoce\")\n\t\t\t}\n\t\t}\n\t\t//verificando tamano\n\t\tif size < 1 {\n\t\t\tfmt.Println(\"RESULTADO: Error en el atributo SIZE\")\n\t\t} else {\n\t\t\t//verificando unidad\n\t\t\tif strings.Compare(unit, \"error\") == 0 {\n\t\t\t\tfmt.Println(\"RESULTADO: Error en el atributo UNIT\")\n\t\t\t} else if strings.Compare(unit, \"\") == 0 {\n\t\t\t\tunit = \"m\"\n\t\t\t} else {\n\t\t\t\t//verificando path\n\t\t\t\tif strings.Compare(path, \"\") == 1 {\n\t\t\t\t\t_, err := os.Stat(strings.ReplaceAll(path, \"\\\"\", \"\"))\n\t\t\t\t\tif err == nil {\n\t\t\t\t\t} else {\n\t\t\t\t\t\tcrearDirectorioSiNoExiste(path)\n\t\t\t\t\t}\n\t\t\t\t\t//En este punto ya tiene que estar creado el directorio si es que no existia\n\t\t\t\t}\n\t\t\t\t//verificando nombre\n\t\t\t\tif strings.Compare(name, \"\") == 1 {\n\t\t\t\t\t_, err := os.Stat(strings.ReplaceAll(path+\"/\"+name, \"\\\"\", \"\"))\n\t\t\t\t\tif err == nil {\n\t\t\t\t\t\tfmt.Println(\"*********** ATENCION! ***********\")\n\t\t\t\t\t\tfmt.Print(\"¿Desea sobre-escribir el disco?\\n1) SI\\n2) NO\\nIngrese una opcion: \")\n\t\t\t\t\t\tlector := bufio.NewReader(os.Stdin)\n\t\t\t\t\t\tcomando, _ := lector.ReadString('\\n')\n\t\t\t\t\t\tif strings.Compare(strings.TrimSpace(comando), \"1\") == 0 {\n\t\t\t\t\t\t\tsName := strings.Split(name, \".\")\n\t\t\t\t\t\t\tif strings.Compare(strings.ToLower(strings.TrimSpace(sName[1])), \"dsk\") == 0 {\n\t\t\t\t\t\t\t\t//Aqui mando a crear el archivo\n\t\t\t\t\t\t\t\tcrearDisco(size, unit, path+\"/\"+name)\n\t\t\t\t\t\t\t\tfmt.Println(\"RESULTADO: Disco creado\")\n\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\tfmt.Println(\"RESULTADO: Solo se pueden crear discos con extension .DSK\")\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t} else if strings.Compare(strings.TrimSpace(comando), \"2\") == 0 {\n\t\t\t\t\t\t\tfmt.Println(\"RESULTADO: Se ha cancelado la creacion del disco\")\n\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\tfmt.Println(\"RESULTADO: Opcion incorrecta, el disco no se creara\")\n\t\t\t\t\t\t}\n\t\t\t\t\t} else {\n\t\t\t\t\t\t//verificar extension\n\t\t\t\t\t\tsName := strings.Split(name, \".\")\n\t\t\t\t\t\tif len(sName) > 1 {\n\t\t\t\t\t\t\tif strings.Compare(strings.ToLower(strings.TrimSpace(sName[1])), \"dsk\") == 0 {\n\t\t\t\t\t\t\t\t//Aqui mando a crear el archivo\n\t\t\t\t\t\t\t\tcrearDisco(size, unit, path+\"/\"+name)\n\t\t\t\t\t\t\t\tfmt.Println(\"RESULTADO: Disco creado\")\n\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\tfmt.Println(\"RESULTADO: Solo se pueden crear discos con extension .DSK\")\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\tfmt.Println(\"RESULTADO: Error en el nombre del disco a crear\")\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t} else {\n\t\tfmt.Println(\"RESULTADO: Faltan atributos obligatorios para el comando MKDISK\")\n\t}\n}", "func (c *MockFileStorageClient) CreateFileSystem(ctx context.Context, details filestorage.CreateFileSystemDetails) (*filestorage.FileSystem, error) {\n\treturn &filestorage.FileSystem{Id: &fileSystemID}, nil\n}", "func (g Graph) CreateDrive(w http.ResponseWriter, r *http.Request) {\n\tus, ok := ctxpkg.ContextGetUser(r.Context())\n\tif !ok {\n\t\terrorcode.GeneralException.Render(w, r, http.StatusUnauthorized, \"invalid user\")\n\t\treturn\n\t}\n\n\t// TODO determine if the user tries to create his own personal space and pass that as a boolean\n\tif !canCreateSpace(r.Context(), false) {\n\t\t// if the permission is not existing for the user in context we can assume we don't have it. Return 401.\n\t\terrorcode.GeneralException.Render(w, r, http.StatusUnauthorized, \"insufficient permissions to create a space.\")\n\t\treturn\n\t}\n\n\tclient := g.GetGatewayClient()\n\tdrive := libregraph.Drive{}\n\tif err := json.NewDecoder(r.Body).Decode(&drive); err != nil {\n\t\terrorcode.GeneralException.Render(w, r, http.StatusBadRequest, \"invalid schema definition\")\n\t\treturn\n\t}\n\tspaceName := *drive.Name\n\tif spaceName == \"\" {\n\t\terrorcode.GeneralException.Render(w, r, http.StatusInternalServerError, \"invalid name\")\n\t\treturn\n\t}\n\n\tvar driveType string\n\tif drive.DriveType != nil {\n\t\tdriveType = *drive.DriveType\n\t}\n\tswitch driveType {\n\tcase \"\", \"project\":\n\t\tdriveType = \"project\"\n\tdefault:\n\t\terrorcode.GeneralException.Render(w, r, http.StatusBadRequest, fmt.Sprintf(\"drives of type %s cannot be created via this api\", driveType))\n\t\treturn\n\t}\n\n\tcsr := storageprovider.CreateStorageSpaceRequest{\n\t\tOwner: us,\n\t\tType: driveType,\n\t\tName: spaceName,\n\t\tQuota: getQuota(drive.Quota, g.config.Spaces.DefaultQuota),\n\t}\n\n\tif drive.Description != nil {\n\t\tcsr.Opaque = utils.AppendPlainToOpaque(csr.Opaque, \"description\", *drive.Description)\n\t}\n\n\tif drive.DriveAlias != nil {\n\t\tcsr.Opaque = utils.AppendPlainToOpaque(csr.Opaque, \"spaceAlias\", *drive.DriveAlias)\n\t}\n\n\tresp, err := client.CreateStorageSpace(r.Context(), &csr)\n\tif err != nil {\n\t\terrorcode.GeneralException.Render(w, r, http.StatusInternalServerError, err.Error())\n\t\treturn\n\t}\n\n\tif resp.GetStatus().GetCode() != cs3rpc.Code_CODE_OK {\n\t\terrorcode.GeneralException.Render(w, r, http.StatusInternalServerError, \"\")\n\t\treturn\n\t}\n\n\twdu, err := url.Parse(g.config.Spaces.WebDavBase + g.config.Spaces.WebDavPath)\n\tif err != nil {\n\t\tg.logger.Error().Err(err).Msg(\"error parsing url\")\n\t\terrorcode.GeneralException.Render(w, r, http.StatusInternalServerError, err.Error())\n\t\treturn\n\t}\n\n\tnewDrive, err := g.cs3StorageSpaceToDrive(r.Context(), wdu, resp.StorageSpace)\n\tif err != nil {\n\t\tg.logger.Error().Err(err).Msg(\"error parsing space\")\n\t\terrorcode.GeneralException.Render(w, r, http.StatusInternalServerError, err.Error())\n\t\treturn\n\t}\n\trender.Status(r, http.StatusCreated)\n\trender.JSON(w, r, newDrive)\n}", "func (s *Module) DiskWrite(name string, image string) error {\n\tpath, err := s.findDisk(name)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"couldn't find disk with id: %s\", name)\n\t}\n\n\tif !s.isEmptyDisk(path) {\n\t\tlog.Debug().Str(\"disk\", path).Msg(\"disk already has a filesystem. no write\")\n\t\treturn nil\n\t}\n\n\tsource, err := os.Open(image)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to open image\")\n\t}\n\tdefer source.Close()\n\tfile, err := os.OpenFile(path, os.O_WRONLY, 0644)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer file.Close()\n\timgStat, err := source.Stat()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to stat image\")\n\t}\n\tfileStat, err := file.Stat()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to state disk\")\n\t}\n\n\tif imgStat.Size() > fileStat.Size() {\n\t\treturn fmt.Errorf(\"image size is bigger than disk\")\n\t}\n\n\t_, err = io.Copy(file, source)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to write disk image\")\n\t}\n\n\treturn nil\n}", "func NewDiskSyncer(am *AccountManager) *DiskSyncer {\n\treturn &DiskSyncer{\n\t\tflushAccount: make(chan *flushAccountRequest),\n\t\tscheduleWallet: make(chan *Account),\n\t\tscheduleTxStore: make(chan *Account),\n\t\twriteBatch: make(chan *writeBatchRequest),\n\t\texportAccount: make(chan *exportRequest),\n\t\tam: am,\n\t\tquit: make(chan struct{}),\n\t\tshutdown: make(chan struct{}),\n\t}\n}", "func (c *TestClient) GetDisk(project, zone, name string) (*compute.Disk, error) {\n\tif c.GetDiskFn != nil {\n\t\treturn c.GetDiskFn(project, zone, name)\n\t}\n\treturn c.client.GetDisk(project, zone, name)\n}", "func (d *Driver) CreateVolume(ctx context.Context, req *csi.CreateVolumeRequest) (*csi.CreateVolumeResponse, error) {\n\tif err := d.ValidateControllerServiceRequest(csi.ControllerServiceCapability_RPC_CREATE_DELETE_VOLUME); err != nil {\n\t\tklog.Errorf(\"invalid create volume req: %v\", req)\n\t\treturn nil, err\n\t}\n\n\tvolumeCapabilities := req.GetVolumeCapabilities()\n\tname := req.GetName()\n\tif len(name) == 0 {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"CreateVolume Name must be provided\")\n\t}\n\tif len(volumeCapabilities) == 0 {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"CreateVolume Volume capabilities must be provided\")\n\t}\n\n\tvolSizeBytes := int64(req.GetCapacityRange().GetRequiredBytes())\n\trequestGiB := int(util.RoundUpGiB(volSizeBytes))\n\n\tparameters := req.GetParameters()\n\tvar storageAccountType, resourceGroup, location, accountName, containerName string\n\n\t// Apply ProvisionerParameters (case-insensitive). We leave validation of\n\t// the values to the cloud provider.\n\tfor k, v := range parameters {\n\t\tswitch strings.ToLower(k) {\n\t\tcase \"skuname\":\n\t\t\tstorageAccountType = v\n\t\tcase \"storageaccounttype\":\n\t\t\tstorageAccountType = v\n\t\tcase \"location\":\n\t\t\tlocation = v\n\t\tcase \"storageaccount\":\n\t\t\taccountName = v\n\t\tcase \"resourcegroup\":\n\t\t\tresourceGroup = v\n\t\tcase \"containername\":\n\t\t\tcontainerName = v\n\t\tdefault:\n\t\t\treturn nil, fmt.Errorf(\"invalid option %q\", k)\n\t\t}\n\t}\n\n\tif resourceGroup == \"\" {\n\t\tresourceGroup = d.cloud.ResourceGroup\n\t}\n\n\taccount, accountKey, err := d.cloud.EnsureStorageAccount(accountName, storageAccountType, string(storage.BlockBlobStorage), resourceGroup, location, goofysAccountNamePrefix)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"could not get storage key for storage account %s: %v\", accountName, err)\n\t}\n\taccountName = account\n\n\tif containerName == \"\" {\n\t\tcontainerName = getValidContainerName(name)\n\t}\n\n\tklog.V(2).Infof(\"begin to create container(%s) on account(%s) type(%s) rg(%s) location(%s) size(%d)\", containerName, accountName, storageAccountType, resourceGroup, location, requestGiB)\n\tclient, err := azstorage.NewBasicClientOnSovereignCloud(accountName, accountKey, d.cloud.Environment)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tblobClient := client.GetBlobService()\n\tcontainer := blobClient.GetContainerReference(containerName)\n\t_, err = container.CreateIfNotExists(&azstorage.CreateContainerOptions{Access: azstorage.ContainerAccessTypePrivate})\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to create container(%s) on account(%s) type(%s) rg(%s) location(%s) size(%d), error: %v\", containerName, accountName, storageAccountType, resourceGroup, location, requestGiB, err)\n\t}\n\n\tvolumeID := fmt.Sprintf(volumeIDTemplate, resourceGroup, accountName, containerName)\n\n\t/* todo: snapshot support\n\tif req.GetVolumeContentSource() != nil {\n\t\tcontentSource := req.GetVolumeContentSource()\n\t\tif contentSource.GetSnapshot() != nil {\n\t\t}\n\t}\n\t*/\n\tklog.V(2).Infof(\"create container %s on storage account %s successfully\", containerName, accountName)\n\n\treturn &csi.CreateVolumeResponse{\n\t\tVolume: &csi.Volume{\n\t\t\tVolumeId: volumeID,\n\t\t\tCapacityBytes: req.GetCapacityRange().GetRequiredBytes(),\n\t\t\tVolumeContext: parameters,\n\t\t},\n\t}, nil\n}", "func (s stage) partitionDisk(dev types.Disk, devAlias string) error {\n\tif cutil.IsTrue(dev.WipeTable) {\n\t\top := sgdisk.Begin(s.Logger, devAlias)\n\t\ts.Logger.Info(\"wiping partition table requested on %q\", devAlias)\n\t\top.WipeTable(true)\n\t\tif err := op.Commit(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Ensure all partitions with number 0 are last\n\tsort.Stable(PartitionList(dev.Partitions))\n\n\top := sgdisk.Begin(s.Logger, devAlias)\n\n\tdiskInfo, err := s.getPartitionMap(devAlias)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// get a list of parititions that have size and start 0 replaced with the real sizes\n\t// that would be used if all specified partitions were to be created anew.\n\t// Also calculate sectors for all of the start/size values.\n\tresolvedPartitions, err := s.getRealStartAndSize(dev, devAlias, diskInfo)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor _, part := range resolvedPartitions {\n\t\tshouldExist := partitionShouldExist(part)\n\t\tinfo, exists := diskInfo.GetPartition(part.Number)\n\t\tvar matchErr error\n\t\tif exists {\n\t\t\tmatchErr = partitionMatches(info, part)\n\t\t}\n\t\tmatches := exists && matchErr == nil\n\t\twipeEntry := cutil.IsTrue(part.WipePartitionEntry)\n\n\t\t// This is a translation of the matrix in the operator notes.\n\t\tswitch {\n\t\tcase !exists && !shouldExist:\n\t\t\ts.Logger.Info(\"partition %d specified as nonexistant and no partition was found. Success.\", part.Number)\n\t\tcase !exists && shouldExist:\n\t\t\top.CreatePartition(part)\n\t\tcase exists && !shouldExist && !wipeEntry:\n\t\t\treturn fmt.Errorf(\"partition %d exists but is specified as nonexistant and wipePartitionEntry is false\", part.Number)\n\t\tcase exists && !shouldExist && wipeEntry:\n\t\t\top.DeletePartition(part.Number)\n\t\tcase exists && shouldExist && matches:\n\t\t\ts.Logger.Info(\"partition %d found with correct specifications\", part.Number)\n\t\tcase exists && shouldExist && !wipeEntry && !matches:\n\t\t\tif partitionMatchesResize(info, part) {\n\t\t\t\ts.Logger.Info(\"resizing partition %d\", part.Number)\n\t\t\t\top.DeletePartition(part.Number)\n\t\t\t\tpart.Number = info.Number\n\t\t\t\tpart.GUID = &info.GUID\n\t\t\t\tpart.TypeGUID = &info.TypeGUID\n\t\t\t\tpart.Label = &info.Label\n\t\t\t\tpart.StartSector = &info.StartSector\n\t\t\t\top.CreatePartition(part)\n\t\t\t} else {\n\t\t\t\treturn fmt.Errorf(\"Partition %d didn't match: %v\", part.Number, matchErr)\n\t\t\t}\n\t\tcase exists && shouldExist && wipeEntry && !matches:\n\t\t\ts.Logger.Info(\"partition %d did not meet specifications, wiping partition entry and recreating\", part.Number)\n\t\t\top.DeletePartition(part.Number)\n\t\t\top.CreatePartition(part)\n\t\tdefault:\n\t\t\t// unfortunatey, golang doesn't check that all cases are handled exhaustively\n\t\t\treturn fmt.Errorf(\"Unreachable code reached when processing partition %d. golang--\", part.Number)\n\t\t}\n\t}\n\n\tif err := op.Commit(); err != nil {\n\t\treturn fmt.Errorf(\"commit failure: %v\", err)\n\t}\n\treturn nil\n}", "func DiskMount(srcPath string, dstPath string, readonly bool, recursive bool, propagation string, mountOptions []string, fsName string) error {\n\tvar err error\n\n\t// Prepare the mount flags\n\tflags := 0\n\tif readonly {\n\t\tflags |= unix.MS_RDONLY\n\t}\n\n\t// Detect the filesystem\n\tif fsName == \"none\" {\n\t\tflags |= unix.MS_BIND\n\t}\n\n\tif propagation != \"\" {\n\t\tswitch propagation {\n\t\tcase \"private\":\n\t\t\tflags |= unix.MS_PRIVATE\n\t\tcase \"shared\":\n\t\t\tflags |= unix.MS_SHARED\n\t\tcase \"slave\":\n\t\t\tflags |= unix.MS_SLAVE\n\t\tcase \"unbindable\":\n\t\t\tflags |= unix.MS_UNBINDABLE\n\t\tcase \"rprivate\":\n\t\t\tflags |= unix.MS_PRIVATE | unix.MS_REC\n\t\tcase \"rshared\":\n\t\t\tflags |= unix.MS_SHARED | unix.MS_REC\n\t\tcase \"rslave\":\n\t\t\tflags |= unix.MS_SLAVE | unix.MS_REC\n\t\tcase \"runbindable\":\n\t\t\tflags |= unix.MS_UNBINDABLE | unix.MS_REC\n\t\tdefault:\n\t\t\treturn fmt.Errorf(\"Invalid propagation mode %q\", propagation)\n\t\t}\n\t}\n\n\tif recursive {\n\t\tflags |= unix.MS_REC\n\t}\n\n\t// Mount the filesystem\n\terr = unix.Mount(srcPath, dstPath, fsName, uintptr(flags), strings.Join(mountOptions, \",\"))\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Unable to mount %q at %q with filesystem %q: %w\", srcPath, dstPath, fsName, err)\n\t}\n\n\t// Remount bind mounts in readonly mode if requested\n\tif readonly == true && flags&unix.MS_BIND == unix.MS_BIND {\n\t\tflags = unix.MS_RDONLY | unix.MS_BIND | unix.MS_REMOUNT\n\t\terr = unix.Mount(\"\", dstPath, fsName, uintptr(flags), \"\")\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Unable to mount %q in readonly mode: %w\", dstPath, err)\n\t\t}\n\t}\n\n\tflags = unix.MS_REC | unix.MS_SLAVE\n\terr = unix.Mount(\"\", dstPath, \"\", uintptr(flags), \"\")\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Unable to make mount %q private: %w\", dstPath, err)\n\t}\n\n\treturn nil\n}", "func (op *Operation) CreatePartition(p Partition) error {\n\t// TODO(vc): sanity check p against op.parts\n\t// XXX(vc): I don't think sgdisk likes zero-based partition numbers, TODO: verify this! sgdisk _feels_ poorly made when using it, consider alternatives.\n\top.parts = append(op.parts, p)\n\treturn nil\n}", "func GuestCreateDisks(endpoint string, guestid string, body GuestCreateDiskList) (int, []byte) {\n\n\tcreateReq, _ := json.Marshal(body)\n\n\tbuffer := getEndpointwithGuests(endpoint)\n\tbuffer.WriteString(\"/\")\n\tbuffer.WriteString(guestid)\n\tbuffer.WriteString(\"/disks\")\n\n\tstatus, data := hq.Post(buffer.String(), createReq)\n\n\treturn status, data\n}", "func setupNullDisk(t *testing.T, s *DiskIO, devName string) func() {\n\ttd, err := os.CreateTemp(\"\", \".telegraf.DiskInfoTest\")\n\trequire.NoError(t, err)\n\n\tif s.infoCache == nil {\n\t\ts.infoCache = make(map[string]diskInfoCache)\n\t}\n\tic, ok := s.infoCache[devName]\n\tif !ok {\n\t\t// No previous calls for the device were done, easy to poison the cache\n\t\ts.infoCache[devName] = diskInfoCache{\n\t\t\tmodifiedAt: 0,\n\t\t\tudevDataPath: td.Name(),\n\t\t\tvalues: map[string]string{},\n\t\t}\n\t}\n\torigUdevPath := ic.udevDataPath\n\n\tcleanFunc := func() {\n\t\tic.udevDataPath = origUdevPath\n\t\tos.Remove(td.Name())\n\t}\n\n\tic.udevDataPath = td.Name()\n\t_, err = td.Write(nullDiskInfo)\n\tif err != nil {\n\t\tcleanFunc()\n\t\tt.Fatal(err)\n\t}\n\n\treturn cleanFunc\n}", "func (*FileSystemBase) Create(path string, flags int, mode uint32) (int, uint64) {\n\treturn -ENOSYS, ^uint64(0)\n}", "func createDiskTable(memTable *memTable, dbDir string, index, sparseKeyDistance int) error {\n\tprefix := strconv.Itoa(index) + \"-\"\n\n\tw, err := newDiskTableWriter(dbDir, prefix, sparseKeyDistance)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to create disk table writer: %w\", err)\n\t}\n\n\tfor it := memTable.iterator(); it.hasNext(); {\n\t\tkey, value := it.next()\n\t\tif err := w.write(key, value); err != nil {\n\t\t\treturn fmt.Errorf(\"failed to write to disk table %d: %w\", index, err)\n\t\t}\n\t}\n\n\tif err := w.sync(); err != nil {\n\t\treturn fmt.Errorf(\"failed to sync disk table: %w\", err)\n\t}\n\n\tif err := w.close(); err != nil {\n\t\treturn fmt.Errorf(\"failed to close disk table: %w\", err)\n\t}\n\n\treturn nil\n}", "func (d *Driver) writeDiskImage() error { // TODO\n\tlog.Debugf(\"Creating hard disk image...\")\n\n\tmagicString := \"boot2docker, this is xhyve speaking\"\n\n\tbuf := new(bytes.Buffer)\n\ttw := tar.NewWriter(buf)\n\n\t// magicString first so the automount script knows to format the disk\n\tfile := &tar.Header{Name: magicString, Size: int64(len(magicString))}\n\tif err := tw.WriteHeader(file); err != nil {\n\t\treturn err\n\t}\n\tif _, err := tw.Write([]byte(magicString)); err != nil {\n\t\treturn err\n\t}\n\t// .ssh/key.pub => authorized_keys\n\tfile = &tar.Header{Name: \".ssh\", Typeflag: tar.TypeDir, Mode: 0700}\n\tif err := tw.WriteHeader(file); err != nil {\n\t\treturn err\n\t}\n\tpubKey, err := ioutil.ReadFile(d.publicSSHKeyPath())\n\tif err != nil {\n\t\treturn err\n\t}\n\tfile = &tar.Header{Name: \".ssh/authorized_keys\", Size: int64(len(pubKey)), Mode: 0644}\n\tif err := tw.WriteHeader(file); err != nil {\n\t\treturn err\n\t}\n\tif _, err := tw.Write([]byte(pubKey)); err != nil {\n\t\treturn err\n\t}\n\tfile = &tar.Header{Name: \".ssh/authorized_keys2\", Size: int64(len(pubKey)), Mode: 0644}\n\tif err := tw.WriteHeader(file); err != nil {\n\t\treturn err\n\t}\n\tif _, err := tw.Write([]byte(pubKey)); err != nil {\n\t\treturn err\n\t}\n\tif err := tw.Close(); err != nil {\n\t\treturn err\n\t}\n\traw := buf.Bytes()\n\n\tif err := ioutil.WriteFile(d.imgPath(), raw, 0644); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *TestClient) DeleteDisk(project, zone, name string) error {\n\tif c.DeleteDiskFn != nil {\n\t\treturn c.DeleteDiskFn(project, zone, name)\n\t}\n\treturn c.client.DeleteDisk(project, zone, name)\n}", "func (mzk *MockZK) Create(path string, data []byte, flags int32, acl []zk.ACL) (string, error) {\n\tmzk.Args = append(mzk.Args, []interface{}{\n\t\t\"create\",\n\t\tpath,\n\t\tdata,\n\t\tflags,\n\t\tacl,\n\t})\n\treturn mzk.CreateFn(path, data, flags, acl)\n}", "func (p *Tmpfs) Create(ctx driver.Context, id types.VolumeID) (*types.Volume, error) {\n\tctx.Log.Debugf(\"Tmpfs create volume: %s\", id.Name)\n\n\t// parse the mount path\n\tmountPath := path.Join(dataDir, id.Name)\n\n\treturn types.NewVolumeFromID(mountPath, \"\", id), nil\n}", "func (cs *DefaultControllerServer) CreateVolume(ctx context.Context, req *csi.CreateVolumeRequest) (*csi.CreateVolumeResponse, error) {\n\treturn nil, status.Error(codes.Unimplemented, \"\")\n}", "func (z *zfsctl) CreateFileSystem(ctx context.Context, name string, properties map[string]string) *execute {\n\targs := []string{\"create\", \"-p\"}\n\tif properties != nil {\n\t\tkv := \"-o \"\n\t\tfor k, v := range properties {\n\t\t\tkv += fmt.Sprintf(\"%s=%s \", k, v)\n\t\t}\n\t\targs = append(args, kv)\n\t}\n\targs = append(args, name)\n\treturn &execute{ctx: ctx, name: z.cmd, args: args}\n}", "func (s stack) CreateVolume(ctx context.Context, request abstract.VolumeRequest) (_ *abstract.Volume, ferr fail.Error) {\n\tif valid.IsNil(s) {\n\t\treturn nil, fail.InvalidInstanceError()\n\t}\n\treturn nil, fail.NotImplementedError(\"implement me\")\n}", "func NewDiskWriter(ctx context.Context, logger log.Logger, bDir string) (_ *DiskWriter, err error) {\n\tbTmp := bDir + tmpForCreationBlockDirSuffix\n\n\td := &DiskWriter{\n\t\tbTmp: bTmp,\n\t\tbDir: bDir,\n\t\tlogger: logger,\n\t}\n\tdefer func() {\n\t\tif err != nil {\n\t\t\terr = tsdb_errors.NewMulti(err, tsdb_errors.CloseAll(d.closers)).Err()\n\t\t\tif err := os.RemoveAll(bTmp); err != nil {\n\t\t\t\tlevel.Error(logger).Log(\"msg\", \"removed tmp folder after failed compaction\", \"err\", err.Error())\n\t\t\t}\n\t\t}\n\t}()\n\n\tif err = os.RemoveAll(bTmp); err != nil {\n\t\treturn nil, err\n\t}\n\tif err = os.MkdirAll(bTmp, 0750); err != nil {\n\t\treturn nil, err\n\t}\n\n\tchunkw, err := chunks.NewWriter(filepath.Join(bTmp, ChunksDirname))\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"open chunk writer\")\n\t}\n\td.closers = append(d.closers, chunkw)\n\n\t// TODO(bwplotka): Setup instrumentedChunkWriter if we want to upstream this code.\n\n\tindexw, err := index.NewWriter(ctx, filepath.Join(bTmp, IndexFilename))\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"open index writer\")\n\t}\n\td.closers = append(d.closers, indexw)\n\td.statsGatheringSeriesWriter = statsGatheringSeriesWriter{iw: indexw, cw: chunkw}\n\treturn d, nil\n}", "func CreateCreateFileSystemRequest() (request *CreateFileSystemRequest) {\n\trequest = &CreateFileSystemRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"DFS\", \"2018-06-20\", \"CreateFileSystem\", \"alidfs\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func NewDiskCache() *DiskCache {\n\treturn &DiskCache{\n\t\tDir: os.TempDir(),\n\t\tMaxBytes: 1 << 20, // 1mb\n\t\tMaxFiles: 256,\n\t\tCleanupSleep: 60 * time.Second,\n\t}\n}", "func NewDiskStorage(name string, path string, serializer Serializer, deserializer Deserializer) (Storage, error) {\n\tif err := os.MkdirAll(filepath.Dir(path), 0700); err != nil {\n\t\treturn nil, err\n\t}\n\n\tdb, err := bolt.Open(path, 0600, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdiskstorage := &DiskStorage{\n\t\tdb: db,\n\t\tname: name,\n\t\tmemory: NewMemoryStorage(),\n\t\tserializer: serializer,\n\t\tdeserializer: deserializer,\n\t}\n\n\terr = diskstorage.load()\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn diskstorage, nil\n}", "func (r *VmwareMapper) MapDisk(vmSpec *kubevirtv1.VirtualMachine, dv cdiv1.DataVolume) {\n\tname := fmt.Sprintf(\"dv-%v\", dv.Name)\n\tname = utils.EnsureLabelValueLength(name)\n\tvolume := kubevirtv1.Volume{\n\t\tName: name,\n\t\tVolumeSource: kubevirtv1.VolumeSource{\n\t\t\tDataVolume: &kubevirtv1.DataVolumeSource{\n\t\t\t\tName: dv.Name,\n\t\t\t},\n\t\t},\n\t}\n\n\tkubevirtDisk := kubevirtv1.Disk{\n\t\tName: name,\n\t\tDiskDevice: kubevirtv1.DiskDevice{\n\t\t\tDisk: &kubevirtv1.DiskTarget{\n\t\t\t\tBus: busTypeVirtio,\n\t\t\t},\n\t\t},\n\t}\n\n\tvmSpec.Spec.Template.Spec.Volumes = append(vmSpec.Spec.Template.Spec.Volumes, volume)\n\tdisks := append(vmSpec.Spec.Template.Spec.Domain.Devices.Disks, kubevirtDisk)\n\n\t// Since the import controller is iterating over a map of DVs,\n\t// MapDisk gets called for each DV in a nondeterministic order which results\n\t// in the disks being in an arbitrary order. This sort ensure the disks are\n\t// attached in the same order as the devices on the source VM.\n\tsort.Slice(disks, func(i, j int) bool {\n\t\treturn disks[i].Name < disks[j].Name\n\t})\n\tvmSpec.Spec.Template.Spec.Domain.Devices.Disks = disks\n}", "func (d *Driver) Create() error {\n\tb2dutils := mcnutils.NewB2dUtils(d.StorePath)\n\tif err := b2dutils.CopyIsoToMachineDir(d.Boot2DockerURL, d.MachineName); err != nil {\n\t\treturn err\n\t}\n\n\tlog.Infof(\"Generating SSH Keypair...\")\n\tif err := ssh.GenerateSSHKey(d.GetSSHKeyPath()); err != nil {\n\t\treturn err\n\t}\n\n\t// Create context\n\tctx, cancel := context.WithCancel(context.Background())\n\tdefer cancel()\n\n\tc, err := d.vsphereLogin(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer c.Logout(ctx)\n\n\t// Create a new finder\n\tf := find.NewFinder(c.Client, true)\n\n\tdc, err := f.DatacenterOrDefault(ctx, d.Datacenter)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tf.SetDatacenter(dc)\n\n\tdss, err := f.DatastoreOrDefault(ctx, d.Datastore)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tnetworks := make(map[string]object.NetworkReference)\n\tfor _, netName := range d.Networks {\n\t\tnet, err := f.NetworkOrDefault(ctx, netName)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tnetworks[netName] = net\n\t}\n\n\tvar hs *object.HostSystem\n\tif d.HostSystem != \"\" {\n\t\tvar err error\n\t\ths, err = f.HostSystemOrDefault(ctx, d.HostSystem)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvar rp *object.ResourcePool\n\tif d.Pool != \"\" {\n\t\t// Find specified Resource Pool\n\t\trp, err = f.ResourcePool(ctx, d.Pool)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t} else if d.HostSystem != \"\" {\n\t\t// Pick default Resource Pool for Host System\n\t\trp, err = hs.ResourcePool(ctx)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\t// Pick the default Resource Pool for the Datacenter.\n\t\trp, err = f.DefaultResourcePool(ctx)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tspec := types.VirtualMachineConfigSpec{\n\t\tName: d.MachineName,\n\t\tGuestId: \"otherLinux64Guest\",\n\t\tFiles: &types.VirtualMachineFileInfo{VmPathName: fmt.Sprintf(\"[%s]\", dss.Name())},\n\t\tNumCPUs: int32(d.CPU),\n\t\tMemoryMB: int64(d.Memory),\n\t}\n\n\tscsi, err := object.SCSIControllerTypes().CreateSCSIController(\"pvscsi\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tspec.DeviceChange = append(spec.DeviceChange, &types.VirtualDeviceConfigSpec{\n\t\tOperation: types.VirtualDeviceConfigSpecOperationAdd,\n\t\tDevice: scsi,\n\t})\n\n\tlog.Infof(\"Creating VM...\")\n\tfolders, err := dc.Folders(ctx)\n\ttask, err := folders.VmFolder.CreateVM(ctx, spec, rp, hs)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tinfo, err := task.WaitForResult(ctx, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tlog.Infof(\"Uploading Boot2docker ISO ...\")\n\tdsurl, err := dss.URL(ctx, dc, fmt.Sprintf(\"%s/%s\", d.MachineName, isoFilename))\n\tif err != nil {\n\t\treturn err\n\t}\n\tp := soap.DefaultUpload\n\tif err = c.Client.UploadFile(d.ISO, dsurl, &p); err != nil {\n\t\treturn err\n\t}\n\n\t// Retrieve the new VM\n\tvm := object.NewVirtualMachine(c.Client, info.Result.(types.ManagedObjectReference))\n\n\tdevices, err := vm.Device(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar add []types.BaseVirtualDevice\n\n\tcontroller, err := devices.FindDiskController(\"scsi\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdisk := devices.CreateDisk(controller, dss.Reference(),\n\t\tdss.Path(fmt.Sprintf(\"%s/%s.vmdk\", d.MachineName, d.MachineName)))\n\n\t// Convert MB to KB\n\tdisk.CapacityInKB = int64(d.DiskSize) * 1024\n\n\tadd = append(add, disk)\n\tide, err := devices.FindIDEController(\"\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tcdrom, err := devices.CreateCdrom(ide)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tadd = append(add, devices.InsertIso(cdrom, dss.Path(fmt.Sprintf(\"%s/%s\", d.MachineName, isoFilename))))\n\n\tfor _, netName := range d.Networks {\n\t\tbacking, err := networks[netName].EthernetCardBackingInfo(ctx)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tnetdev, err := object.EthernetCardTypes().CreateEthernetCard(\"vmxnet3\", backing)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tlog.Infof(\"adding network: %s\", netName)\n\t\tadd = append(add, netdev)\n\t}\n\n\tlog.Infof(\"Reconfiguring VM\")\n\tif vm.AddDevice(ctx, add...); err != nil {\n\t\treturn err\n\t}\n\n\t// Adding some guestinfo data\n\tvar opts []types.BaseOptionValue\n\tfor _, param := range d.CfgParams {\n\t\tv := strings.SplitN(param, \"=\", 2)\n\t\tkey := v[0]\n\t\tvalue := \"\"\n\t\tif len(v) > 1 {\n\t\t\tvalue = v[1]\n\t\t}\n\t\tfmt.Printf(\"Setting %s to %s\\n\", key, value)\n\t\topts = append(opts, &types.OptionValue{\n\t\t\tKey: key,\n\t\t\tValue: value,\n\t\t})\n\t}\n\tif d.CloudInit != \"\" {\n\t\tif _, err := url.ParseRequestURI(d.CloudInit); err == nil {\n\t\t\tlog.Infof(\"setting guestinfo.cloud-init.data.url to %s\\n\", d.CloudInit)\n\t\t\topts = append(opts, &types.OptionValue{\n\t\t\t\tKey: \"guestinfo.cloud-init.config.url\",\n\t\t\t\tValue: d.CloudInit,\n\t\t\t})\n\t\t} else {\n\t\t\tif _, err := os.Stat(d.CloudInit); err == nil {\n\t\t\t\tif value, err := ioutil.ReadFile(d.CloudInit); err == nil {\n\t\t\t\t\tlog.Infof(\"setting guestinfo.cloud-init.data to encoded content of %s\\n\", d.CloudInit)\n\t\t\t\t\tencoded := base64.StdEncoding.EncodeToString(value)\n\t\t\t\t\topts = append(opts, &types.OptionValue{\n\t\t\t\t\t\tKey: \"guestinfo.cloud-init.config.data\",\n\t\t\t\t\t\tValue: encoded,\n\t\t\t\t\t})\n\t\t\t\t\topts = append(opts, &types.OptionValue{\n\t\t\t\t\t\tKey: \"guestinfo.cloud-init.data.encoding\",\n\t\t\t\t\t\tValue: \"base64\",\n\t\t\t\t\t})\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\ttask, err = vm.Reconfigure(ctx, types.VirtualMachineConfigSpec{\n\t\tExtraConfig: opts,\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\ttask.Wait(ctx)\n\n\tif err := d.Start(); err != nil {\n\t\treturn err\n\t}\n\n\tlog.Infof(\"Provisioning certs and ssh keys...\")\n\t// Generate a tar keys bundle\n\tif err := d.generateKeyBundle(); err != nil {\n\t\treturn err\n\t}\n\n\topman := guest.NewOperationsManager(c.Client, vm.Reference())\n\n\tfileman, err := opman.FileManager(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tsrc := d.ResolveStorePath(\"userdata.tar\")\n\ts, err := os.Stat(src)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tauth := AuthFlag{}\n\tflag := FileAttrFlag{}\n\tauth.auth.Username = B2DUser\n\tauth.auth.Password = B2DPass\n\tflag.SetPerms(0, 0, 660)\n\turl, err := fileman.InitiateFileTransferToGuest(ctx, auth.Auth(), \"/home/docker/userdata.tar\", flag.Attr(), s.Size(), true)\n\tif err != nil {\n\t\treturn err\n\t}\n\tu, err := c.Client.ParseURL(url)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif err = c.Client.UploadFile(src, u, nil); err != nil {\n\t\treturn err\n\t}\n\n\tprocman, err := opman.ProcessManager(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// first, untar - only boot2docker has /var/lib/boot2docker\n\t// TODO: don't hard-code to docker & staff - they are also just b2d\n\tvar env []string\n\tguestspec := types.GuestProgramSpec{\n\t\tProgramPath: \"/usr/bin/sudo\",\n\t\tArguments: \"/usr/bin/sudo /bin/sh -c \\\"tar xvf /home/docker/userdata.tar -C /home/docker > /var/log/userdata.log 2>&1 && chown -R docker:staff /home/docker\\\"\",\n\t\tWorkingDirectory: \"\",\n\t\tEnvVariables: env,\n\t}\n\n\t_, err = procman.StartProgram(ctx, auth.Auth(), &guestspec)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// now move to /var/lib/boot2docker if its there\n\tguestspec = types.GuestProgramSpec{\n\t\tProgramPath: \"/usr/bin/sudo\",\n\t\tArguments: \"/bin/mv /home/docker/userdata.tar /var/lib/boot2docker/userdata.tar\",\n\t\tWorkingDirectory: \"\",\n\t\tEnvVariables: env,\n\t}\n\n\t_, err = procman.StartProgram(ctx, auth.Auth(), &guestspec)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (f *Fixture) EnableDisk(nSlots int64) {\n\tvar e error\n\tf.Bdev, e = bdev.NewMalloc((1 + nSlots) * 16)\n\tf.noError(e)\n\tf.t.Cleanup(func() { f.Bdev.Close() })\n\n\tf.SpdkTh, e = spdkenv.NewThread()\n\tf.noError(e)\n\tf.t.Cleanup(func() { f.SpdkTh.Close() })\n\tf.noError(ealthread.AllocLaunch(f.SpdkTh))\n\n\tf.DiskStore, e = disk.NewStore(f.Bdev, f.SpdkTh, 16, disk.StoreGetDataGo)\n\tf.noError(e)\n\n\tmin, max := f.DiskStore.SlotRange()\n\tf.DiskAlloc = disk.NewAlloc(min, max, eal.NumaSocket{})\n\tf.t.Cleanup(func() {\n\t\tf.DiskAlloc.Close()\n\t\tf.DiskStore.Close()\n\t})\n\n\te = f.Cs.SetDisk(f.DiskStore, f.DiskAlloc)\n\tf.noError(e)\n}", "func (t *Template) AddDisk() *shared.Disk {\n\tdisk := shared.NewDisk()\n\tt.Elements = append(t.Elements, disk)\n\treturn disk\n}", "func (c *MockBlockStorageClient) CreateVolume(ctx context.Context, details core.CreateVolumeDetails) (*core.Volume, error) {\n\tid := \"oc1.volume1.xxxx\"\n\tad := \"zkJl:US-ASHBURN-AD-1\"\n\treturn &core.Volume{\n\t\tId: &id,\n\t\tAvailabilityDomain: &ad,\n\t}, nil\n}", "func NewMapDisk(ctx context.Context, wg *sync.WaitGroup) *MapDisk {\n\tmd := &MapDisk{\n\t\tstore: make(map[string][]byte),\n\t\tinChan: make(chan Request),\n\t}\n\twg.Add(1)\n\tgo md.serve(ctx, wg)\n\treturn md\n}", "func createVmdk(host string, size string, objType string, diskFormat string) (string, error) {\n\tdsName := GetAndExpectStringEnvVar(envSharedDatastoreName)\n\tdir := \"/vmfs/volumes/\" + dsName + \"/e2e\"\n\terr := createDir(dir, host)\n\tgomega.Expect(err).NotTo(gomega.HaveOccurred())\n\tif diskFormat == \"\" {\n\t\tdiskFormat = \"thin\"\n\t}\n\tif objType == \"\" {\n\t\tobjType = \"vsan\"\n\t}\n\tif size == \"\" {\n\t\tsize = \"2g\"\n\t}\n\trand.Seed(time.Now().UnixNano())\n\tvmdkPath := fmt.Sprintf(\"%s/test-%v-%v.vmdk\", dir, time.Now().UnixNano(), rand.Intn(1000))\n\tsshCmd := fmt.Sprintf(\"vmkfstools -c %s -d %s -W %s %s\", size, diskFormat, objType, vmdkPath)\n\tframework.Logf(\"Invoking command '%v' on ESX host %v\", sshCmd, host)\n\tresult, err := fssh.SSH(sshCmd, host+\":22\", framework.TestContext.Provider)\n\tif err != nil || result.Code != 0 {\n\t\tfssh.LogResult(result)\n\t\treturn vmdkPath, fmt.Errorf(\"couldn't execute command: '%s' on ESX host: %v\", sshCmd, err)\n\t}\n\treturn vmdkPath, nil\n}", "func (c *MockAzureCloud) Disk() azure.DisksClient {\n\treturn c.DisksClient\n}", "func (driver *Driver) CreateVolume(ctx context.Context, req *csi.CreateVolumeRequest) (*csi.CreateVolumeResponse, error) {\n\t//klog.V(4).Infof(\"CreateVolume: called with args %#v\", req)\n\n\t// volume name is created by CO for idempotency\n\tvolName := req.GetName()\n\tif len(volName) == 0 {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"Volume name not provided\")\n\t}\n\tvolID := generateVolumeID(volName)\n\n\tklog.V(4).Infof(\"CreateVolume: volumeName(%#v)\", volName)\n\n\tvolCaps := req.GetVolumeCapabilities()\n\tif len(volCaps) == 0 {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"Volume capabilities not provided\")\n\t}\n\n\tif !driver.isValidVolumeCapabilities(volCaps) {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"Volume capabilities not supported\")\n\t}\n\n\tcapRange := req.GetCapacityRange()\n\tvolCapacity := DefaultVolumeSize\n\tif capRange != nil {\n\t\tvolCapacity = capRange.GetRequiredBytes()\n\t}\n\n\t// create a new volume\n\t// need to provide idempotency\n\tvolParams := req.GetParameters()\n\tvolSecrets := req.GetSecrets()\n\tvolRootPath := \"\"\n\n\tsecrets := make(map[string]string)\n\tfor k, v := range driver.secrets {\n\t\tsecrets[k] = v\n\t}\n\n\tfor k, v := range volSecrets {\n\t\tsecrets[k] = v\n\t}\n\n\tirodsClient := ExtractIRODSClientType(volParams, secrets, FuseType)\n\tif irodsClient != FuseType {\n\t\treturn nil, status.Errorf(codes.InvalidArgument, \"unsupported driver type - %v\", irodsClient)\n\t}\n\n\t// check security flags\n\tenforceProxyAccess := driver.getDriverConfigEnforceProxyAccess()\n\tproxyUser := driver.getDriverConfigUser()\n\n\tirodsConn, err := ExtractIRODSConnectionInfo(volParams, secrets)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif enforceProxyAccess {\n\t\tif proxyUser == irodsConn.User {\n\t\t\t// same proxy user\n\t\t\t// enforce clientUser\n\t\t\tif len(irodsConn.ClientUser) == 0 {\n\t\t\t\treturn nil, status.Error(codes.InvalidArgument, \"Argument clientUser must be given\")\n\t\t\t}\n\n\t\t\tif irodsConn.User == irodsConn.ClientUser {\n\t\t\t\treturn nil, status.Errorf(codes.InvalidArgument, \"Argument clientUser cannot be the same as user - user %s, clientUser %s\", irodsConn.User, irodsConn.ClientUser)\n\t\t\t}\n\t\t} else {\n\t\t\t// replaced user\n\t\t\t// static volume provisioning takes user argument from pv\n\t\t\t// this is okay\n\t\t}\n\t}\n\n\t// do not allow anonymous access\n\tif irodsConn.User == \"anonymous\" {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"Argument user must be a non-anonymous user\")\n\t}\n\n\tif irodsConn.ClientUser == \"anonymous\" {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"Argument clientUser must be a non-anonymous user\")\n\t}\n\n\tvolContext := make(map[string]string)\n\tvolRetain := false\n\tvolCreate := true\n\tvolPath := \"\"\n\tfor k, v := range secrets {\n\t\tswitch strings.ToLower(k) {\n\t\tcase \"volumerootpath\", \"volume_root_path\":\n\t\t\tif !filepath.IsAbs(v) {\n\t\t\t\treturn nil, status.Errorf(codes.InvalidArgument, \"Argument %q must be an absolute path\", k)\n\t\t\t}\n\t\t\tif v == \"/\" {\n\t\t\t\tvolRootPath = v\n\t\t\t} else {\n\t\t\t\tvolRootPath = strings.TrimRight(v, \"/\")\n\t\t\t}\n\t\tcase \"retaindata\", \"retain_data\":\n\t\t\tretain, err := strconv.ParseBool(v)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, status.Errorf(codes.InvalidArgument, \"Argument %q must be a boolean value - %s\", k, err)\n\t\t\t}\n\t\t\tvolRetain = retain\n\t\tcase \"novolumedir\", \"no_volume_dir\":\n\t\t\tnovolumedir, err := strconv.ParseBool(v)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, status.Errorf(codes.InvalidArgument, \"Argument %q must be a boolean value - %s\", k, err)\n\t\t\t}\n\t\t\tvolCreate = !novolumedir\n\t\t}\n\t\t// do not copy secret params\n\t}\n\n\tfor k, v := range volParams {\n\t\tswitch strings.ToLower(k) {\n\t\tcase \"volumerootpath\", \"volume_root_path\":\n\t\t\tif !filepath.IsAbs(v) {\n\t\t\t\treturn nil, status.Errorf(codes.InvalidArgument, \"Argument %q must be an absolute path\", k)\n\t\t\t}\n\t\t\tif v == \"/\" {\n\t\t\t\tvolRootPath = v\n\t\t\t} else {\n\t\t\t\tvolRootPath = strings.TrimRight(v, \"/\")\n\t\t\t}\n\t\tcase \"retaindata\", \"retain_data\":\n\t\t\tretain, err := strconv.ParseBool(v)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, status.Errorf(codes.InvalidArgument, \"Argument %q must be a boolean value - %s\", k, err)\n\t\t\t}\n\t\t\tvolRetain = retain\n\t\tcase \"novolumedir\", \"no_volume_dir\":\n\t\t\tnovolumedir, err := strconv.ParseBool(v)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, status.Errorf(codes.InvalidArgument, \"Argument %q must be a boolean value - %s\", k, err)\n\t\t\t}\n\t\t\tvolCreate = !novolumedir\n\t\t}\n\t\t// copy all params\n\t\tvolContext[k] = v\n\t}\n\n\tif len(volRootPath) == 0 {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"Argument volumeRootPath is not provided\")\n\t}\n\n\t// need to check if mount path is in whitelist\n\tif !driver.isMountPathAllowed(volRootPath) {\n\t\treturn nil, status.Errorf(codes.InvalidArgument, \"Argument volumeRootPath %s is not allowed to mount\", volRootPath)\n\t}\n\n\t// generate path\n\tif volCreate {\n\t\tvolPath = fmt.Sprintf(\"%s/%s\", volRootPath, volName)\n\n\t\tklog.V(5).Infof(\"Creating a volume dir %s\", volPath)\n\t\terr = IRODSMkdir(irodsConn, volPath)\n\t\tif err != nil {\n\t\t\treturn nil, status.Errorf(codes.Internal, \"Could not create a volume dir %s : %v\", volPath, err)\n\t\t}\n\t} else {\n\t\tvolPath = volRootPath\n\t\t// in this case, we should retain data because the mounted path may have files\n\t\t// we should not delete these old files when the pvc is deleted\n\t\tvolRetain = true\n\t}\n\n\tvolContext[\"path\"] = volPath\n\n\t// create a irods volume and put it to manager\n\tirodsVolume := NewIRODSVolume(volID, volName, volRootPath, volPath, irodsConn, volRetain)\n\tPutIRODSVolume(irodsVolume)\n\n\tvolume := &csi.Volume{\n\t\tVolumeId: volID,\n\t\tCapacityBytes: volCapacity,\n\t\tVolumeContext: volContext,\n\t}\n\n\treturn &csi.CreateVolumeResponse{Volume: volume}, nil\n}", "func (client StorageGatewayClient) CreateFileSystem(ctx context.Context, request CreateFileSystemRequest) (response CreateFileSystemResponse, err error) {\n\tvar ociResponse common.OCIResponse\n\tpolicy := common.NoRetryPolicy()\n\tif client.RetryPolicy() != nil {\n\t\tpolicy = *client.RetryPolicy()\n\t}\n\tif request.RetryPolicy() != nil {\n\t\tpolicy = *request.RetryPolicy()\n\t}\n\n\tif !(request.OpcRetryToken != nil && *request.OpcRetryToken != \"\") {\n\t\trequest.OpcRetryToken = common.String(common.RetryToken())\n\t}\n\n\tociResponse, err = common.Retry(ctx, request, client.createFileSystem, policy)\n\tif err != nil {\n\t\tif ociResponse != nil {\n\t\t\tif httpResponse := ociResponse.HTTPResponse(); httpResponse != nil {\n\t\t\t\topcRequestId := httpResponse.Header.Get(\"opc-request-id\")\n\t\t\t\tresponse = CreateFileSystemResponse{RawResponse: httpResponse, OpcRequestId: &opcRequestId}\n\t\t\t} else {\n\t\t\t\tresponse = CreateFileSystemResponse{}\n\t\t\t}\n\t\t}\n\t\treturn\n\t}\n\tif convertedResponse, ok := ociResponse.(CreateFileSystemResponse); ok {\n\t\tresponse = convertedResponse\n\t} else {\n\t\terr = fmt.Errorf(\"failed to convert OCIResponse into CreateFileSystemResponse\")\n\t}\n\treturn\n}", "func (manager *mockFileSystemManager) CreateFileSystem(ctx context.Context, groupName string, filesystemName string, timeout *int32, xMsDate string, datalakeName string) (*autorest.Response, error) {\n\tfs := fileSystemResource{\n\t\tresourceGroupName: groupName,\n\t\tstorageAccountName: datalakeName,\n\t\tfilesystemName: filesystemName,\n\t}\n\n\tmanager.fileSystemResource = append(manager.fileSystemResource, fs)\n\tmockresponse := helpers.GetRestResponse(http.StatusOK)\n\n\treturn &mockresponse, nil\n}", "func (c *MockBlockStorageClient) CreateVolume(ctx context.Context, details core.CreateVolumeDetails) (*core.Volume, error) {\n\treturn &core.Volume{Id: &VolumeBackupID}, nil\n}", "func (z *ZkPlus) Create(path string, data []byte, flags int32, acl []zk.ACL) (string, error) {\n\tz.forPath(path).Log(logkey.ZkMethod, \"Create\")\n\tp, err := z.blockOnConn().Create(z.realPath(path), data, flags, acl)\n\tif strings.HasPrefix(p, z.pathPrefix) && z.pathPrefix != \"\" {\n\t\tp = p[len(z.pathPrefix)+1:]\n\t}\n\treturn p, errors.Annotatef(err, \"cannot create zk path %s\", path)\n}", "func (e *ExternalInterface) CreateVolume(ctx context.Context, req *systemsproto.VolumeRequest, pc *PluginContact, taskID string) {\n\tvar resp response.RPC\n\tvar targetURI = \"/redfish/v1/Systems/\" + req.SystemID + \"/Storage/\" + req.StorageInstance + \"/Volumes\"\n\t//create task\n\ttaskInfo := &common.TaskUpdateInfo{Context: ctx, TaskID: taskID, TargetURI: targetURI,\n\t\tUpdateTask: pc.UpdateTask, TaskRequest: string(req.RequestBody)}\n\t// spliting the uuid and system id\n\trequestData := strings.SplitN(req.SystemID, \".\", 2)\n\tif len(requestData) <= 1 {\n\t\terrorMessage := \"error: SystemUUID not found\"\n\t\tcommon.GeneralError(http.StatusNotFound, response.ResourceNotFound, errorMessage, []interface{}{\"System\", req.SystemID}, taskInfo)\n\t\treturn\n\t}\n\tuuid := requestData[0]\n\ttarget, gerr := e.DB.GetTarget(uuid)\n\tif gerr != nil {\n\t\tcommon.GeneralError(http.StatusNotFound, response.ResourceNotFound, gerr.Error(), []interface{}{\"System\", uuid}, taskInfo)\n\t\treturn\n\t}\n\t// Validating the storage instance\n\tif strings.TrimSpace(req.StorageInstance) == \"\" {\n\t\terrorMessage := \"error: Storage instance is not found\"\n\t\tcommon.GeneralError(http.StatusBadRequest, response.ResourceNotFound, errorMessage, []interface{}{\"Storage\", req.StorageInstance}, taskInfo)\n\t\treturn\n\t}\n\n\tvar volume smodel.Volume\n\t// unmarshalling the volume\n\terr := json.Unmarshal(req.RequestBody, &volume)\n\tif err != nil {\n\t\terrorMessage := \"Error while unmarshaling the create volume request: \" + err.Error()\n\t\tif StringContain(err.Error(), \"smodel.OdataIDLink\") {\n\t\t\terrorMessage = \"Error processing create volume request: @odata.id key(s) is missing in Drives list\"\n\t\t}\n\t\tl.LogWithFields(ctx).Error(errorMessage)\n\t\tcommon.GeneralError(http.StatusBadRequest, response.MalformedJSON, errorMessage, []interface{}{}, taskInfo)\n\t\treturn\n\t}\n\n\t// Validating the request JSON properties for case sensitive\n\tinvalidProperties, err := RequestParamsCaseValidatorFunc(req.RequestBody, volume)\n\tif err != nil {\n\t\terrMsg := \"error while validating request parameters for volume creation: \" + err.Error()\n\t\tl.LogWithFields(ctx).Error(errMsg)\n\t\tcommon.GeneralError(http.StatusInternalServerError, response.InternalError, errMsg, nil, taskInfo)\n\t\treturn\n\t} else if invalidProperties != \"\" {\n\t\terrorMessage := \"error: one or more properties given in the request body are not valid, ensure properties are listed in uppercamelcase \"\n\t\tl.LogWithFields(ctx).Error(errorMessage)\n\t\tcommon.GeneralError(http.StatusBadRequest, response.PropertyUnknown, errorMessage, []interface{}{invalidProperties}, taskInfo)\n\t\treturn\n\t}\n\t//fields validation\n\tstatuscode, statusMessage, messageArgs, err := e.validateProperties(ctx, &volume, req.SystemID)\n\tif err != nil {\n\t\terrorMessage := \"error: request payload validation failed: \" + err.Error()\n\t\tl.LogWithFields(ctx).Error(errorMessage)\n\t\tcommon.GeneralError(statuscode, statusMessage, errorMessage, messageArgs, taskInfo)\n\t\treturn\n\t}\n\tdecryptedPasswordByte, err := e.DevicePassword(target.Password)\n\tif err != nil {\n\t\terrorMessage := \"error while trying to decrypt device password: \" + err.Error()\n\t\tcommon.GeneralError(http.StatusInternalServerError, response.InternalError, errorMessage, nil, taskInfo)\n\t\treturn\n\t}\n\ttarget.Password = decryptedPasswordByte\n\t// Get the Plugin info\n\tplugin, gerr := e.DB.GetPluginData(target.PluginID)\n\tif gerr != nil {\n\t\terrorMessage := \"error while trying to get plugin details\"\n\t\tcommon.GeneralError(http.StatusInternalServerError, response.InternalError, errorMessage, nil, taskInfo)\n\t\treturn\n\t}\n\tvar contactRequest scommon.PluginContactRequest\n\tcontactRequest.ContactClient = e.ContactClient\n\tcontactRequest.Plugin = plugin\n\tcontactRequest.GetPluginStatus = e.GetPluginStatus\n\n\tif StringsEqualFold(plugin.PreferredAuthType, \"XAuthToken\") {\n\t\tvar err error\n\t\tcontactRequest.HTTPMethodType = http.MethodPost\n\t\tcontactRequest.DeviceInfo = map[string]interface{}{\n\t\t\t\"UserName\": plugin.Username,\n\t\t\t\"Password\": string(plugin.Password),\n\t\t}\n\t\tcontactRequest.OID = \"/ODIM/v1/Sessions\"\n\t\t_, token, _, getResponse, err := scommon.ContactPlugin(ctx, contactRequest, \"error while creating session with the plugin: \")\n\n\t\tif err != nil {\n\t\t\tcommon.GeneralError(getResponse.StatusCode, getResponse.StatusMessage, err.Error(), nil, taskInfo)\n\t\t\treturn\n\t\t}\n\t\tcontactRequest.Token = token\n\t} else {\n\t\tcontactRequest.BasicAuth = map[string]string{\n\t\t\t\"UserName\": plugin.Username,\n\t\t\t\"Password\": string(plugin.Password),\n\t\t}\n\n\t}\n\ttarget.PostBody = req.RequestBody\n\n\tcontactRequest.HTTPMethodType = http.MethodPost\n\tcontactRequest.DeviceInfo = target\n\tcontactRequest.OID = fmt.Sprintf(\"/ODIM/v1/Systems/%s/Storage/%s/Volumes\", requestData[1], req.StorageInstance)\n\n\tbody, location, pluginIP, getResponse, err := ContactPluginFunc(ctx, contactRequest, \"error while creating a volume: \")\n\tif err != nil {\n\t\tresp.StatusCode = getResponse.StatusCode\n\t\tjson.Unmarshal(body, &resp.Body)\n\t\tl.LogWithFields(ctx).Error(err)\n\t\ttask := fillTaskData(taskID, targetURI, string(req.RequestBody), resp,\n\t\t\tcommon.Completed, common.Warning, 100, http.MethodPost)\n\t\tpc.UpdateTask(ctx, task)\n\t\treturn\n\t}\n\tif getResponse.StatusCode == http.StatusAccepted {\n\t\terr = pc.SavePluginTaskInfo(ctx, pluginIP, plugin.IP, taskID, location)\n\t\tif err != nil {\n\t\t\tl.LogWithFields(ctx).Error(err)\n\t\t}\n\t\treturn\n\t}\n\tresp.StatusCode = http.StatusOK\n\tresp.StatusMessage = response.Success\n\terr = JSONUnmarshalFunc(body, &resp.Body)\n\tif err != nil {\n\t\tcommon.GeneralError(http.StatusInternalServerError, response.InternalError, err.Error(), nil, taskInfo)\n\t\treturn\n\t}\n\ttask := fillTaskData(taskID, targetURI, string(req.RequestBody), resp,\n\t\tcommon.Completed, common.OK, 100, http.MethodPost)\n\tpc.UpdateTask(ctx, task)\n}", "func (z *ZfsH) CreateFilesystem(name string, properties map[string]string) (*Dataset, error) {\n\targs := make([]string, 1, 4)\n\targs[0] = \"create\"\n\n\tif properties != nil {\n\t\targs = append(args, propsSlice(properties)...)\n\t}\n\n\targs = append(args, name)\n\t_, err := z.zfs(args...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn z.GetDataset(name)\n}", "func (c *UpCloudControllerServer) CreateVolume(ctx context.Context, req *csi.CreateVolumeRequest) (*csi.CreateVolumeResponse, error) {\n\tvolName := req.Name\n\tif volName == \"\" {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"CreateVolume Name is missing\")\n\t}\n\n\tif req.VolumeCapabilities == nil || len(req.VolumeCapabilities) == 0 {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"CreateVolume Volume Capabilities is missing\")\n\t}\n\n\t// Validate\n\tif !isValidCapability(req.VolumeCapabilities) {\n\t\treturn nil, status.Errorf(codes.InvalidArgument, \"CreateVolume Volume capability is not compatible: %v\", req)\n\t}\n\n\tc.Driver.log.WithFields(logrus.Fields{\n\t\t\"volume-name\": volName,\n\t\t\"capabilities\": req.VolumeCapabilities,\n\t}).Info(\"Create Volume: called\")\n\n\t// check that the volume doesnt already exist\n\tstorages, err := c.Driver.upCloudClient.ListStorages()\n\tif err != nil {\n\t\treturn nil, status.Errorf(codes.Internal, \"CreateVolume Volume: List volume error %s\", err.Error())\n\t}\n\tfor _, storage := range storages.Storages.Storage {\n\t\tif storage.Title == volName {\n\t\t\treturn &csi.CreateVolumeResponse{\n\t\t\t\tVolume: &csi.Volume{\n\t\t\t\t\tVolumeId: storage.UUID,\n\t\t\t\t\tCapacityBytes: int64(storage.Size) * giB,\n\t\t\t\t},\n\t\t\t}, nil\n\t\t}\n\t}\n\n\t// if applicable, create volume\n\tsize, err := getStorageBytes(req.CapacityRange)\n\tif err != nil {\n\t\treturn nil, status.Errorf(codes.OutOfRange, \"invalid volume capacity range: %v\", err)\n\t}\n\tnewStorage, err := c.Driver.upCloudClient.CreateStorage(int(size/giB), req.Parameters[\"tier\"], volName, c.Driver.region)\n\tif err != nil {\n\t\treturn nil, status.Errorf(codes.Internal, \"error create volume: %s\", err.Error())\n\t}\n\n\tres := &csi.CreateVolumeResponse{\n\t\tVolume: &csi.Volume{\n\t\t\tVolumeId: newStorage.Storage.UUID,\n\t\t\tCapacityBytes: size,\n\t\t\tAccessibleTopology: []*csi.Topology{\n\t\t\t\t{\n\t\t\t\t\tSegments: map[string]string{\n\t\t\t\t\t\t\"region\": c.Driver.region,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\n\tc.Driver.log.WithFields(logrus.Fields{\n\t\t\"size\": size,\n\t\t\"volume-id\": newStorage.Storage.UUID,\n\t\t\"volume-name\": volName,\n\t\t\"volume-size\": int(size / giB),\n\t}).Info(\"Create Volume: created volume\")\n\n\treturn res, nil\n}", "func (r opResult) makeDiskBackedSorterConstructor(\n\tctx context.Context,\n\tflowCtx *execinfra.FlowCtx,\n\targs *colexecargs.NewColOperatorArgs,\n\topNamePrefix string,\n\tfactory coldata.ColumnFactory,\n) colexec.DiskBackedSorterConstructor {\n\treturn func(input colexecop.Operator, inputTypes []*types.T, orderingCols []execinfrapb.Ordering_Column, maxNumberPartitions int) colexecop.Operator {\n\t\tif maxNumberPartitions < colexecop.ExternalSorterMinPartitions {\n\t\t\tcolexecerror.InternalError(errors.AssertionFailedf(\n\t\t\t\t\"external sorter is attempted to be created with %d partitions, minimum %d required\",\n\t\t\t\tmaxNumberPartitions, colexecop.ExternalSorterMinPartitions,\n\t\t\t))\n\t\t}\n\t\tsortArgs := *args\n\t\tif !args.TestingKnobs.DelegateFDAcquisitions {\n\t\t\t// Set the FDSemaphore to nil. This indicates that no FDs should be\n\t\t\t// acquired. The hash-based partitioner will do this up front.\n\t\t\tsortArgs.FDSemaphore = nil\n\t\t}\n\t\tsorter, err := r.createDiskBackedSort(\n\t\t\tctx, flowCtx, &sortArgs, input, inputTypes,\n\t\t\texecinfrapb.Ordering{Columns: orderingCols},\n\t\t\t0 /* matchLen */, maxNumberPartitions, args.Spec.ProcessorID,\n\t\t\t&execinfrapb.PostProcessSpec{}, opNamePrefix+\"-\", factory,\n\t\t)\n\t\tif err != nil {\n\t\t\tcolexecerror.InternalError(err)\n\t\t}\n\t\treturn sorter\n\t}\n}", "func (s *OsdCsiServer) CreateVolume(\n\tctx context.Context,\n\treq *csi.CreateVolumeRequest,\n) (*csi.CreateVolumeResponse, error) {\n\n\t// Log request\n\tlogrus.Debugf(\"CreateVolume req[%#v]\", *req)\n\n\tif len(req.GetName()) == 0 {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"Name must be provided\")\n\t}\n\tif req.GetVolumeCapabilities() == nil || len(req.GetVolumeCapabilities()) == 0 {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"Volume capabilities must be provided\")\n\t}\n\n\t// Get parameters\n\tspec, locator, source, err := s.specHandler.SpecFromOpts(req.GetParameters())\n\tif err != nil {\n\t\te := fmt.Sprintf(\"Unable to get parameters: %s\\n\", err.Error())\n\t\tlogrus.Errorln(e)\n\t\treturn nil, status.Error(codes.InvalidArgument, e)\n\t}\n\n\t// Get Size\n\tif req.GetCapacityRange() != nil && req.GetCapacityRange().GetRequiredBytes() != 0 {\n\t\tspec.Size = uint64(req.GetCapacityRange().GetRequiredBytes())\n\t} else {\n\t\tspec.Size = defaultCSIVolumeSize\n\t}\n\n\t// Create response\n\tvolume := &csi.Volume{}\n\tresp := &csi.CreateVolumeResponse{\n\t\tVolume: volume,\n\t}\n\n\t// Check if the volume has already been created or is in process of creation\n\tv, err := util.VolumeFromName(s.driver, req.GetName())\n\tif err == nil {\n\t\t// Check the requested arguments match that of the existing volume\n\t\tif spec.Size != v.GetSpec().GetSize() {\n\t\t\treturn nil, status.Errorf(\n\t\t\t\tcodes.AlreadyExists,\n\t\t\t\t\"Existing volume has a size of %v which differs from requested size of %v\",\n\t\t\t\tv.GetSpec().GetSize(),\n\t\t\t\tspec.Size)\n\t\t}\n\t\tif v.GetSpec().GetShared() != csiRequestsSharedVolume(req) {\n\t\t\treturn nil, status.Errorf(\n\t\t\t\tcodes.AlreadyExists,\n\t\t\t\t\"Existing volume has shared=%v while request is asking for shared=%v\",\n\t\t\t\tv.GetSpec().GetShared(),\n\t\t\t\tcsiRequestsSharedVolume(req))\n\t\t}\n\t\tif v.GetSource().GetParent() != source.GetParent() {\n\t\t\treturn nil, status.Error(codes.AlreadyExists, \"Existing volume has conflicting parent value\")\n\t\t}\n\n\t\t// Return information on existing volume\n\t\tosdToCsiVolumeInfo(volume, v)\n\t\treturn resp, nil\n\t}\n\n\t// Check if this is a cloning request to create a volume from a snapshot\n\tif req.GetVolumeContentSource().GetSnapshot() != nil {\n\t\tsource.Parent = req.GetVolumeContentSource().GetSnapshot().GetId()\n\t}\n\n\t// Check if the caller is asking to create a snapshot or for a new volume\n\tvar id string\n\tif source != nil && len(source.GetParent()) != 0 {\n\t\t// Get parent volume information\n\t\tparent, err := util.VolumeFromName(s.driver, source.Parent)\n\t\tif err != nil {\n\t\t\te := fmt.Sprintf(\"unable to get parent volume information: %s\\n\", err.Error())\n\t\t\tlogrus.Errorln(e)\n\t\t\treturn nil, status.Error(codes.InvalidArgument, e)\n\t\t}\n\n\t\t// Create a snapshot from the parent\n\t\tid, err = s.driver.Snapshot(parent.GetId(), false, &api.VolumeLocator{\n\t\t\tName: req.GetName(),\n\t\t},\n\t\t\tfalse)\n\t\tif err != nil {\n\t\t\te := fmt.Sprintf(\"unable to create snapshot: %s\\n\", err.Error())\n\t\t\tlogrus.Errorln(e)\n\t\t\treturn nil, status.Error(codes.Internal, e)\n\t\t}\n\t} else {\n\t\t// Get Capabilities and Size\n\t\tspec.Shared = csiRequestsSharedVolume(req)\n\n\t\t// Create the volume\n\t\tlocator.Name = req.GetName()\n\n\t\t// get enforced policy specs\n\t\t// 0.3 Does not support user context or auth\n\t\tspec, err = sdkVol.GetDefaultVolSpecs(context.Background(), spec, false /* not an update */)\n\t\tif err != nil {\n\t\t\treturn nil, status.Errorf(codes.Internal, err.Error())\n\t\t}\n\t\tid, err = s.driver.Create(context.TODO(), locator, source, spec)\n\t\tif err != nil {\n\t\t\treturn nil, status.Error(codes.Internal, err.Error())\n\t\t}\n\t}\n\n\t// id must have been set\n\tv, err = util.VolumeFromName(s.driver, id)\n\tif err != nil {\n\t\te := fmt.Sprintf(\"Unable to find newly created volume: %s\", err.Error())\n\t\tlogrus.Errorln(e)\n\t\treturn nil, status.Error(codes.Internal, e)\n\t}\n\tosdToCsiVolumeInfo(volume, v)\n\treturn resp, nil\n}", "func (c *TestClient) AttachDisk(project, zone, instance string, ad *compute.AttachedDisk) error {\n\tif c.AttachDiskFn != nil {\n\t\treturn c.AttachDiskFn(project, zone, instance, ad)\n\t}\n\treturn c.client.AttachDisk(project, zone, instance, ad)\n}", "func (d *VolumeDriver) Create(r volume.Request) volume.Response {\n\tlog.Errorf(\"VolumeDriver Create to be implemented\")\n\treturn volume.Response{Err: \"\"}\n}", "func (in *Disk) DeepCopy() *Disk {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(Disk)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (in *Disk) DeepCopy() *Disk {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(Disk)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (d *Driver) Create(id, parent string, opts *graphdriver.CreateOpts) (retErr error) {\n\tlogrus.Debugf(\"secureoverlay2: Create called w. id: %s, parent: %s, opts: %s\", id, parent, opts)\n\timgCryptOpts := &secureImgCryptOptions{}\n\timgCryptOpts.init(d.options.defaultSecOpts)\n\tdriver := &Driver{}\n\terr := errors.New(\"\")\n\n\tif opts != nil && len(opts.ImgCryptOpt) != 0 {\n\t\terr = d.parseImgCryptOpt(opts.ImgCryptOpt, imgCryptOpts, driver)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"--storage-opt parsing error: %s\", err.Error())\n\t\t}\n\t}\n\n\t// create all directories\n\t// - standard ones\n\tdir := d.dir(id)\n\trootUID, rootGID, err := idtools.GetRootUIDGID(d.uidMaps, d.gidMaps)\n\tif err != nil {\n\t\treturn err\n\t}\n\troot := idtools.Identity{UID: rootUID, GID: rootGID}\n\tif err := idtools.MkdirAllAndChown(path.Dir(dir), 0700, root); err != nil {\n\t\treturn err\n\t}\n\tif err := idtools.MkdirAndChown(dir, 0700, root); err != nil {\n\t\treturn err\n\t}\n\t// unclear why couldn't just have done MkDirAllAs of dir in one go but that's what overlay2 did, so left it ..\n\n\tdefer func() {\n\t\t// Clean up on failure\n\t\tif retErr != nil {\n\t\t\tos.RemoveAll(dir)\n\t\t}\n\t}()\n\tif opts != nil && len(opts.ImgCryptOpt) > 0 && projectQuotaSupported {\n\t\tif driver.options.quota.Size > 0 {\n\t\t\t// Set container disk quota limit\n\t\t\tif err := d.quotaCtl.SetQuota(dir, driver.options.quota); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif err := idtools.MkdirAndChown(path.Join(dir, \"diff\"), 0755, root); err != nil {\n\t\treturn err\n\t}\n\tlid := generateID(idLength)\n\tif err := os.Symlink(path.Join(\"..\", id, \"diff\"), path.Join(d.home, linkDir, lid)); err != nil {\n\t\treturn err\n\t}\n\n\t// Write link id to link file\n\tif err := ioutil.WriteFile(path.Join(dir, \"link\"), []byte(lid), 0644); err != nil {\n\t\treturn err\n\t}\n\t// if no parent directory, done\n\n\tif parent != \"\" {\n\t\tif err := idtools.MkdirAndChown(path.Join(dir, \"work\"), 0700, root); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif err := idtools.MkdirAndChown(path.Join(dir, \"merged\"), 0700, root); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tlower, err := d.getLower(parent)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif lower != \"\" {\n\t\t\tif err := ioutil.WriteFile(path.Join(dir, lowerFile), []byte(lower), 0666); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\t// create secure dirs\n\tsecureDir := path.Join(dir, constSecureBaseDirName)\n\n\tif err := idtools.MkdirAndChown(secureDir, 0755, root); err != nil {\n\t\treturn err\n\t}\n\tif err := d.createSecureDiffDir(id, \"\"); err != nil {\n\t\treturn err\n\t}\n\n\tif err := idtools.MkdirAndChown(path.Join(secureDir, constSecureCryptMntDirName), 0755, root); err != nil {\n\t\treturn err\n\t}\n\t// initialize secure storage space\n\tif err := d.initSecureStorage(id, *imgCryptOpts); err != nil {\n\t\tlogrus.Debugf(\"secureoverlay2: Create w. id: %s, failed to initalize secure storage %s\", id, err.Error())\n\t\treturn err\n\t}\n\n\tlogrus.Debug(\"secureoverlay2: Create returns\")\n\n\treturn nil\n}", "func (d *Disk) createPartitionTable() error {\n\tif d.DiskSize == 0 {\n\t\tklog.Errorf(\"disk %s has size zero\", d.DevPath)\n\t\treturn fmt.Errorf(\"disk size is zero, unable to initialize partition table\")\n\t}\n\tif d.LogicalBlockSize == 0 {\n\t\tklog.Warningf(\"logical block size of %s not set, falling back to 512 bytes\", d.DevPath)\n\t\tklog.Warning(\"partitioning may fail.\")\n\t\td.LogicalBlockSize = 512\n\t}\n\t// set protective MBR to true.\n\t// https://en.wikipedia.org/wiki/GUID_Partition_Table#Protective_MBR_(LBA_0)\n\td.table = &gpt.Table{\n\t\tLogicalSectorSize: int(d.LogicalBlockSize),\n\t\tProtectiveMBR: true,\n\t}\n\treturn nil\n}", "func (h *csiService) CreateVolume(ctx context.Context, params *csi.CreateVolumeRequest) (*csi.CreateVolumeResponse, error) {\n\th.Log.Debug(\"CSI: CreateVolume\", h.dump(params))\n\tvolID, err := h.ControllerOps.CreateVolumeID(ctx)\n\tif err != nil {\n\t\th.Log.Errorf(\"CSI: CreateVolume: %s\", err.Error())\n\t\treturn nil, err\n\t}\n\tspID, ok := params.Parameters[com.K8sVolCtxKeyServicePlanID]\n\tif !ok {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"CSI: CreateVolume storageClass missing nuvoloso-service-plan parameter\")\n\t}\n\tsizeBytesStr := strconv.FormatInt(params.CapacityRange.RequiredBytes, 10)\n\treturn &csi.CreateVolumeResponse{\n\t\tVolume: &csi.Volume{\n\t\t\tVolumeId: volID,\n\t\t\tCapacityBytes: params.CapacityRange.RequiredBytes, // 1GB\n\t\t\tVolumeContext: map[string]string{\n\t\t\t\tcom.K8sVolCtxKeyDynamic: \"dynamic\",\n\t\t\t\tcom.K8sVolCtxKeySizeBytes: sizeBytesStr,\n\t\t\t\tcom.K8sVolCtxKeyServicePlanID: spID,\n\t\t\t\tcom.K8sVolCtxKeyVolumeSeriesName: params.Name,\n\t\t\t},\n\t\t},\n\t}, nil\n}", "func NewDiskArchiver(cfg *config.WildHaiku) (*DiskArchiver, error) {\n\tarchiveChan := make(chan *haiku.Output, 10000)\n\tnow := time.Now().UTC()\n\tfileName := fmt.Sprintf(\"haiku_%s.json\", now.Format(time.RFC3339))\n\tabsOutPath, err := filepath.Abs(cfg.OutputPath)\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"Could not determine absolute path for %s\", absOutPath)\n\t}\n\tif _, err := os.Stat(absOutPath); err != nil {\n\t\treturn nil, errors.Wrapf(err, \"Error accessing output file path %s\", absOutPath)\n\t}\n\n\tfilePath := filepath.Join(absOutPath, fileName)\n\tsymLink := filepath.Join(absOutPath, \"current.json\")\n\treturn &DiskArchiver{Config: cfg, ArchiveChannel: archiveChan, outFilePath: filePath, symlinkPath: symLink}, nil\n}", "func New(cfg config.Cfg, locator storage.Locator, opts ...Option) *DiskCache {\n\tvar cacheConfig cacheConfig\n\tfor _, opt := range opts {\n\t\topt(&cacheConfig)\n\t}\n\n\tcache := &DiskCache{\n\t\tlocator: locator,\n\t\tstorages: cfg.Storages,\n\t\taf: activeFiles{\n\t\t\tMutex: &sync.Mutex{},\n\t\t\tm: map[string]int{},\n\t\t},\n\t\tcacheConfig: cacheConfig,\n\t\twalkersDone: make(chan struct{}),\n\n\t\trequestTotals: prometheus.NewCounter(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"gitaly_diskcache_requests_total\",\n\t\t\t\tHelp: \"Total number of disk cache requests\",\n\t\t\t},\n\t\t),\n\t\tmissTotals: prometheus.NewCounter(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"gitaly_diskcache_miss_total\",\n\t\t\t\tHelp: \"Total number of disk cache misses\",\n\t\t\t},\n\t\t),\n\t\tbytesStoredtotals: prometheus.NewCounter(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"gitaly_diskcache_bytes_stored_total\",\n\t\t\t\tHelp: \"Total number of disk cache bytes stored\",\n\t\t\t},\n\t\t),\n\t\tbytesFetchedtotals: prometheus.NewCounter(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"gitaly_diskcache_bytes_fetched_total\",\n\t\t\t\tHelp: \"Total number of disk cache bytes fetched\",\n\t\t\t},\n\t\t),\n\t\tbytesLoserTotals: prometheus.NewCounter(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"gitaly_diskcache_bytes_loser_total\",\n\t\t\t\tHelp: \"Total number of disk cache bytes from losing writes\",\n\t\t\t},\n\t\t),\n\t\terrTotal: prometheus.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"gitaly_diskcache_errors_total\",\n\t\t\t\tHelp: \"Total number of errors encountered by disk cache\",\n\t\t\t},\n\t\t\t[]string{\"error\"},\n\t\t),\n\t\twalkerCheckTotal: prometheus.NewCounter(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"gitaly_diskcache_walker_check_total\",\n\t\t\t\tHelp: \"Total number of events during diskcache filesystem walks\",\n\t\t\t},\n\t\t),\n\t\twalkerRemovalTotal: prometheus.NewCounter(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"gitaly_diskcache_walker_removal_total\",\n\t\t\t\tHelp: \"Total number of events during diskcache filesystem walks\",\n\t\t\t},\n\t\t),\n\t\twalkerErrorTotal: prometheus.NewCounter(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"gitaly_diskcache_walker_error_total\",\n\t\t\t\tHelp: \"Total number of errors during diskcache filesystem walks\",\n\t\t\t},\n\t\t),\n\t\twalkerEmptyDirTotal: prometheus.NewCounter(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"gitaly_diskcache_walker_empty_dir_total\",\n\t\t\t\tHelp: \"Total number of empty directories encountered\",\n\t\t\t},\n\t\t),\n\t\twalkerEmptyDirRemovalTotal: prometheus.NewCounter(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"gitaly_diskcache_walker_empty_dir_removal_total\",\n\t\t\t\tHelp: \"Total number of empty directories removed\",\n\t\t\t},\n\t\t),\n\t}\n\tcache.keyer = newLeaseKeyer(locator, cache.countErr)\n\n\treturn cache\n}", "func (r *DiskDirectoryResource) Create(item DiskDirectoryConfig) error {\n\tif err := r.c.ModQuery(\"POST\", BasePath+DiskDirectoryEndpoint, item); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (s *OsdCsiServer) CreateVolume(\n\tctx context.Context,\n\treq *csi.CreateVolumeRequest,\n) (*csi.CreateVolumeResponse, error) {\n\t// Log request\n\tclogger.WithContext(ctx).Infof(\"csi.CreateVolume request received. Volume: %s\", req.GetName())\n\n\tif len(req.GetName()) == 0 {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"Name must be provided\")\n\t}\n\n\t// Get parameters\n\tspec, locator, source, err := s.specHandler.SpecFromOpts(req.GetParameters())\n\tif err != nil {\n\t\te := fmt.Sprintf(\"Unable to get parameters: %s\\n\", err.Error())\n\t\tclogger.WithContext(ctx).Errorln(e)\n\t\treturn nil, status.Error(codes.InvalidArgument, e)\n\t}\n\n\tif spec.IsPureVolume() {\n\t\terr = validateCreateVolumeCapabilitiesPure(req.GetVolumeCapabilities(), spec.GetProxySpec())\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t} else {\n\t\terr = validateCreateVolumeCapabilities(req.GetVolumeCapabilities())\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// Get PVC Metadata and add to locator.VolumeLabels\n\t// This will override any storage class secrets added above.\n\tpvcMetadata, err := getPVCMetadata(req.GetParameters())\n\tif err != nil {\n\t\treturn nil, status.Errorf(\n\t\t\tcodes.Internal,\n\t\t\t\"Unable to get PVC Metadata: %v\", err)\n\t}\n\tfor k, v := range pvcMetadata {\n\t\tlocator.VolumeLabels[k] = v\n\t}\n\n\t// Copy all SC Parameters (from req.Parameters) to locator.VolumeLabels.\n\t// This explicit copy matches the equivalent behavior in the in-tree driver\n\tif len(locator.VolumeLabels) == 0 {\n\t\tlocator.VolumeLabels = make(map[string]string)\n\t}\n\tfor k, v := range req.Parameters {\n\t\tlocator.VolumeLabels[k] = v\n\t}\n\n\t// Add encryption secret information to VolumeLabels\n\tlocator.VolumeLabels = s.addEncryptionInfoToLabels(locator.VolumeLabels, req.GetSecrets())\n\n\t// Get parent ID from request: snapshot or volume\n\tif req.GetVolumeContentSource() != nil {\n\t\tif sourceSnap := req.GetVolumeContentSource().GetSnapshot(); sourceSnap != nil {\n\t\t\tsource.Parent = sourceSnap.SnapshotId\n\t\t}\n\n\t\tif sourceVol := req.GetVolumeContentSource().GetVolume(); sourceVol != nil {\n\t\t\tsource.Parent = sourceVol.VolumeId\n\t\t}\n\t}\n\n\t// Get Size\n\tif req.GetCapacityRange() != nil && req.GetCapacityRange().GetRequiredBytes() != 0 {\n\t\tsize, err := roundUpToNearestGiB(req.GetCapacityRange().GetRequiredBytes())\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"failed to round volume size up to nearest GiB: %v\", err.Error())\n\t\t}\n\t\tspec.Size = size\n\n\t} else {\n\t\tspec.Size = defaultCSIVolumeSize\n\t}\n\n\t// cleanup duplicate information after pulling from req.GetParameters\n\tlocator.VolumeLabels = cleanupVolumeLabels(locator.VolumeLabels)\n\n\t// Get grpc connection\n\tconn, err := s.getRemoteConn(ctx)\n\tif err != nil {\n\t\tlogrus.Errorf(\"failed to get remote connection: %v, continuing with local node instead\", err)\n\t\tconn, err = s.getConn()\n\t\tif err != nil {\n\t\t\treturn nil, status.Errorf(\n\t\t\t\tcodes.Unavailable,\n\t\t\t\t\"Unable to connect to SDK server: %v\", err)\n\t\t}\n\t}\n\n\t// Get secret if any was passed\n\tctx = s.setupContext(ctx, req.GetSecrets())\n\tctx, cancel := grpcutil.WithDefaultTimeout(ctx)\n\tdefer cancel()\n\n\t// Check ID is valid with the specified volume capabilities\n\tvolumes := api.NewOpenStorageVolumeClient(conn)\n\n\t// Create volume\n\tvar newVolumeId string\n\tif source.Parent == \"\" {\n\t\tspec, err := resolveSpecFromCSI(spec, req)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tvar createResp *api.SdkVolumeCreateResponse\n\t\tvar createErr error\n\t\ttopologies := getAllTopologies(req.AccessibilityRequirements)\n\t\tif spec.IsPureVolume() && len(topologies) > 0 {\n\t\t\tfor _, topo := range topologies {\n\t\t\t\tspec.TopologyRequirement = &api.TopologyRequirement{\n\t\t\t\t\tLabels: topo.Segments,\n\t\t\t\t}\n\t\t\t\tcreateResp, createErr = volumes.Create(ctx, &api.SdkVolumeCreateRequest{\n\t\t\t\t\tName: req.GetName(),\n\t\t\t\t\tSpec: spec,\n\t\t\t\t\tLabels: locator.GetVolumeLabels(),\n\t\t\t\t})\n\t\t\t\tif s, ok := status.FromError(createErr); createErr == nil || (ok && s.Code() != codes.ResourceExhausted) {\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t} else {\n\t\t\tcreateResp, createErr = volumes.Create(ctx, &api.SdkVolumeCreateRequest{\n\t\t\t\tName: req.GetName(),\n\t\t\t\tSpec: spec,\n\t\t\t\tLabels: locator.GetVolumeLabels(),\n\t\t\t})\n\t\t}\n\t\tif createErr != nil {\n\t\t\treturn nil, createErr\n\t\t}\n\t\tnewVolumeId = createResp.VolumeId\n\t} else {\n\t\tcloneResp, err := volumes.Clone(ctx, &api.SdkVolumeCloneRequest{\n\t\t\tName: req.GetName(),\n\t\t\tParentId: source.Parent,\n\t\t\tAdditionalLabels: getClonedPVCMetadata(locator),\n\t\t})\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tnewVolumeId = cloneResp.VolumeId\n\t}\n\n\t// Get volume information\n\tinspectResp, err := volumes.Inspect(ctx, &api.SdkVolumeInspectRequest{\n\t\tVolumeId: newVolumeId,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Create response\n\tvolume := &csi.Volume{}\n\tosdToCsiVolumeInfo(volume, inspectResp.GetVolume(), req)\n\treturn &csi.CreateVolumeResponse{\n\t\tVolume: volume,\n\t}, nil\n}", "func NewDynamicDiskFactory(params *FactoryParams) *DynamicDiskBlockFactory {\n\tblockFactory := &DynamicDiskBlockFactory{params: params}\n\n\tblockFactory.bitmapFactory = bitmap.NewFactory(blockFactory.params.VhdReader,\n\t\tblockFactory.params.BlockAllocationTable)\n\n\tblockFactory.sectorFactory = NewSectorFactory(blockFactory.params.VhdReader,\n\t\tblockFactory.params.BlockAllocationTable.HasData,\n\t\tblockFactory.params.BlockAllocationTable.GetBlockDataAddress)\n\n\tblockFactory.blockDataReader = NewDynamicDiskBlockReader(blockFactory.params.VhdReader,\n\t\tblockFactory.params.BlockAllocationTable,\n\t\tblockFactory.params.VhdHeader.BlockSize)\n\treturn blockFactory\n}", "func (d *MinioDriver) Create(r volume.Request) volume.Response {\n\td.m.Lock()\n\tdefer d.m.Unlock()\n\n\tglog.V(1).Infof(\"Create request is: %#v\", r)\n\tif err := d.createClient(r.Options); err != nil {\n\t\treturn volumeResp(\"\",\n\t\t\t\"\",\n\t\t\tnil,\n\t\t\tcapability,\n\t\t\tfmt.Errorf(\"error creating client: %s\", err).Error(),\n\t\t)\n\t}\n\n\tvolPath := createName(volumePrefix)\n\tvolMount := filepath.Join(\"/mnt\", volPath)\n\tif err := d.createVolumeMount(volMount); err != nil {\n\t\treturn volumeResp(\"\", \"\", nil, capability, err.Error())\n\t}\n\n\tvolName := createName(volumePrefix)\n\td.volumes[r.Name] = newVolume(volName, volMount, d.c.BucketName)\n\tglog.V(1).Infof(\"this is the d.volumes: %#v\", d.volumes)\n\treturn volumeResp(\"\", \"\", nil, capability, \"\")\n}", "func (z *ZfsH) CreateVolume(name string, size uint64, properties map[string]string) (*Dataset, error) {\n\targs := make([]string, 4, 5)\n\targs[0] = \"create\"\n\targs[1] = \"-p\"\n\targs[2] = \"-V\"\n\targs[3] = strconv.FormatUint(size, 10)\n\tif properties != nil {\n\t\targs = append(args, propsSlice(properties)...)\n\t}\n\targs = append(args, name)\n\t_, err := z.zfs(args...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn z.GetDataset(name)\n}", "func (realFS) Create(name string) (File, error) { return os.Create(name) }", "func NewDiskSector(radius float64, segments int, thetaStart, thetaLength float64) *Geometry {\n\n\td := NewGeometry()\n\n\t// Validate arguments\n\tif segments < 3 {\n\t\tpanic(\"Invalid argument: segments. The number of segments needs to be greater or equal to 3.\")\n\t}\n\n\t// Create buffers\n\tpositions := math32.NewArrayF32(0, 16)\n\tnormals := math32.NewArrayF32(0, 16)\n\tuvs := math32.NewArrayF32(0, 16)\n\tindices := math32.NewArrayU32(0, 16)\n\n\t// Append circle center position\n\tcenter := math32.NewVector3(0, 0, 0)\n\tpositions.AppendVector3(center)\n\n\t// Append circle center normal\n\tvar normal math32.Vector3\n\tnormal.Z = 1\n\tnormals.AppendVector3(&normal)\n\n\t// Append circle center uv coordinate\n\tcenterUV := math32.NewVector2(0.5, 0.5)\n\tuvs.AppendVector2(centerUV)\n\n\t// Generate the segments\n\tfor i := 0; i <= segments; i++ {\n\t\tsegment := thetaStart + float64(i)/float64(segments)*thetaLength\n\n\t\tvx := float32(radius * math.Cos(segment))\n\t\tvy := float32(radius * math.Sin(segment))\n\n\t\t// Appends vertex position, normal and uv coordinates\n\t\tpositions.Append(vx, vy, 0)\n\t\tnormals.AppendVector3(&normal)\n\t\tuvs.Append((vx/float32(radius)+1)/2, (vy/float32(radius)+1)/2)\n\t}\n\n\tfor i := 1; i <= segments; i++ {\n\t\tindices.Append(uint32(i), uint32(i)+1, 0)\n\t}\n\n\td.SetIndices(indices)\n\td.AddVBO(gls.NewVBO(positions).AddAttrib(gls.VertexPosition))\n\td.AddVBO(gls.NewVBO(normals).AddAttrib(gls.VertexNormal))\n\td.AddVBO(gls.NewVBO(uvs).AddAttrib(gls.VertexTexcoord))\n\n\t// Update volume\n\td.volume = 0\n\td.volumeValid = true\n\n\treturn d\n}", "func (o ClusterNodeGroupOutput) SystemDisk() ClusterNodeGroupSystemDiskOutput {\n\treturn o.ApplyT(func(v ClusterNodeGroup) ClusterNodeGroupSystemDisk { return v.SystemDisk }).(ClusterNodeGroupSystemDiskOutput)\n}", "func NewCmdDiskAttach() *cobra.Command {\n\tvar async *bool\n\tvar udiskIDs *[]string\n\n\treq := base.BizClient.NewAttachUDiskRequest()\n\tcmd := &cobra.Command{\n\t\tUse: \"attach\",\n\t\tShort: \"Attach udisk instances to an uhost\",\n\t\tLong: \"Attach udisk instances to an uhost\",\n\t\tExample: \"ucloud udisk attach --uhost-id uhost-xxxx --udisk-id bs-xxx1,bs-xxx2\",\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\tfor _, id := range *udiskIDs {\n\t\t\t\tid = base.PickResourceID(id)\n\t\t\t\treq.UDiskId = &id\n\t\t\t\t*req.UHostId = base.PickResourceID(*req.UHostId)\n\t\t\t\tresp, err := base.BizClient.AttachUDisk(req)\n\t\t\t\tif err != nil {\n\t\t\t\t\tbase.HandleError(err)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\ttext := fmt.Sprintf(\"udisk[%s] is attaching to uhost uhost[%s]\", *req.UDiskId, *req.UHostId)\n\t\t\t\tif *async {\n\t\t\t\t\tbase.Cxt.Println(text)\n\t\t\t\t} else {\n\t\t\t\t\tpollDisk(resp.UDiskId, *req.ProjectId, *req.Region, *req.Zone, text, []string{status.DISK_INUSE, status.DISK_FAILED})\n\t\t\t\t}\n\t\t\t}\n\t\t},\n\t}\n\tflags := cmd.Flags()\n\tflags.SortFlags = false\n\treq.UHostId = flags.String(\"uhost-id\", \"\", \"Required. Resource ID of the uhost instance which you want to attach the disk\")\n\tudiskIDs = flags.StringSlice(\"udisk-id\", nil, \"Required. Resource ID of the udisk instances to attach\")\n\treq.ProjectId = flags.String(\"project-id\", base.ConfigInstance.ProjectID, \"Optional. Assign project-id\")\n\treq.Region = flags.String(\"region\", base.ConfigInstance.Region, \"Optional. Assign region\")\n\treq.Zone = flags.String(\"zone\", base.ConfigInstance.Zone, \"Optional. Assign availability zone\")\n\tasync = flags.Bool(\"async\", false, \"Optional. Do not wait for the long-running operation to finish.\")\n\n\tflags.SetFlagValuesFunc(\"udisk-id\", func() []string {\n\t\treturn getDiskList([]string{status.DISK_AVAILABLE}, *req.ProjectId, *req.Region, *req.Zone)\n\t})\n\tflags.SetFlagValuesFunc(\"uhost-id\", func() []string {\n\t\treturn getUhostList([]string{status.HOST_RUNNING, status.HOST_STOPPED}, *req.ProjectId, *req.Region, *req.Zone)\n\t})\n\n\tcmd.MarkFlagRequired(\"uhost-id\")\n\tcmd.MarkFlagRequired(\"udisk-id\")\n\n\treturn cmd\n}", "func CreateVolume(req systemsproto.VolumeRequest) (*systemsproto.SystemsResponse, error) {\n\tconn, err := services.ODIMService.Client(services.Systems)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to create client connection: %v\", err)\n\t}\n\tdefer conn.Close()\n\tasService := systemsproto.NewSystemsClient(conn)\n\tresp, err := asService.CreateVolume(context.TODO(), &req)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error: RPC error: %v\", err)\n\t}\n\treturn resp, nil\n}", "func (d ImagefsDriver) Create(r *volume.CreateRequest) error {\n\tfmt.Printf(\"-> Create %+v\\n\", r)\n\tsource, ok := r.Options[\"source\"]\n\tif !ok {\n\t\treturn fmt.Errorf(\"no source volume specified\")\n\t}\n\n\t// pull the image\n\t/*readCloser, err := d.cli.ImagePull(context.Background(), source, types.ImagePullOptions{\n\t\t// HACK assume the registry ignores the auth header\n\t\tRegistryAuth: \"null\",\n\t})\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unexpected error: %s\", err)\n\t}\n\tscanner := bufio.NewScanner(readCloser)\n\tfor scanner.Scan() {\n\t}*/\n\n\tcontainerConfig := &container.Config{\n\t\tImage: source,\n\t\tEntrypoint: []string{\"/runtime/loop\"},\n\t\tLabels: map[string]string{\n\t\t\t\"com.docker.imagefs.version\": version,\n\t\t},\n\t\tNetworkDisabled: true,\n\t}\n\n\tif target, ok := r.Options[\"target\"]; ok {\n\t\tcontainerConfig.Labels[\"com.docker.imagefs.target\"] = target\n\t}\n\t// TODO handle error\n\thostConfig := &container.HostConfig{\n\t\tBinds: []string{\"/tmp/runtime:/runtime\"},\n\t\t//AutoRemove: true,\n\t}\n\n\tvar platform *specs.Platform\n\tif platformStr, ok := r.Options[\"platform\"]; ok {\n\t\tif versions.GreaterThanOrEqualTo(d.cli.ClientVersion(), \"1.41\") {\n\t\t\tp, err := platforms.Parse(platformStr)\n\t\t\tif err != nil {\n\t\t\t\treturn errors.Wrap(err, \"error parsing specified platform\")\n\t\t\t}\n\t\t\tplatform = &p\n\t\t}\n\t}\n\n\tnetworkConfig := &network.NetworkingConfig{}\n\tcont, err := d.cli.ContainerCreate(\n\t\tcontext.Background(),\n\t\tcontainerConfig,\n\t\thostConfig,\n\t\tnetworkConfig,\n\t\tplatform,\n\t\t// TODO(rabrams) namespace\n\t\tr.Name,\n\t)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unexpected error: %s\", err)\n\t}\n\tfmt.Printf(\"Temp container ID: %s\", cont.ID)\n\td.cli.ContainerStart(\n\t\tcontext.Background(),\n\t\tcont.ID,\n\t\ttypes.ContainerStartOptions{},\n\t)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unexpected error: %s\", err)\n\t}\n\treturn nil\n}", "func (z *ZFS) Create(req *acomm.Request) (interface{}, *url.URL, error) {\n\tvar args CreateArgs\n\tif err := req.UnmarshalArgs(&args); err != nil {\n\t\treturn nil, nil, err\n\t}\n\tif args.Name == \"\" {\n\t\treturn nil, nil, errors.New(\"missing arg: name\")\n\t}\n\n\tif err := fixPropertyTypesFromJSON(args.Properties); err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tvar ds *gozfs.Dataset\n\tvar err error\n\tswitch args.Type {\n\tcase gozfs.DatasetFilesystem:\n\t\tds, err = gozfs.CreateFilesystem(args.Name, args.Properties)\n\tcase gozfs.DatasetVolume:\n\t\tif args.Volsize <= 0 {\n\t\t\terr = errors.New(\"missing or invalid arg: volsize\")\n\t\t\tbreak\n\t\t}\n\t\tds, err = gozfs.CreateVolume(args.Name, args.Volsize, args.Properties)\n\tdefault:\n\t\terr = errors.New(\"missing or invalid arg: type\")\n\t}\n\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treturn &DatasetResult{newDataset(ds)}, nil, nil\n}", "func CreateVolume(\n\tctx context.Context,\n\tc csi.ControllerClient,\n\tversion *csi.Version,\n\tname string,\n\trequiredBytes, limitBytes uint64,\n\tcapabilities []*csi.VolumeCapability,\n\tparams map[string]string,\n\tcallOpts ...grpc.CallOption) (volume *csi.VolumeInfo, err error) {\n\n\treq := &csi.CreateVolumeRequest{\n\t\tName: name,\n\t\tVersion: version,\n\t\tParameters: params,\n\t\tVolumeCapabilities: capabilities,\n\t}\n\n\tif requiredBytes > 0 || limitBytes > 0 {\n\t\treq.CapacityRange = &csi.CapacityRange{\n\t\t\tLimitBytes: limitBytes,\n\t\t\tRequiredBytes: requiredBytes,\n\t\t}\n\t}\n\n\tres, err := c.CreateVolume(ctx, req, callOpts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn res.GetResult().VolumeInfo, nil\n}", "func New(drv Driver) FileSystem {\n\treturn FileSystem{drv}\n}", "func (op *OnPrem) CreateVolume(ctx *lepton.Context, name, data, provider string) (lepton.NanosVolume, error) {\n\tc := ctx.Config()\n\tif c.BaseVolumeSz == \"\" {\n\t\tc.BaseVolumeSz = strconv.Itoa(MinimumVolumeSize)\n\t}\n\treturn lepton.CreateLocalVolume(c, name, data, provider)\n}", "func newDiskTableWriter(dbDir, prefix string, sparseKeyDistance int) (*diskTableWriter, error) {\n\tdataPath := path.Join(dbDir, prefix+diskTableDataFileName)\n\tdataFile, err := os.OpenFile(dataPath, newDiskTableFlag, 0600)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to open data file %s: %w\", dataPath, err)\n\t}\n\n\tindexPath := path.Join(dbDir, prefix+diskTableIndexFileName)\n\tindexFile, err := os.OpenFile(indexPath, newDiskTableFlag, 0600)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to open index file %s: %w\", indexPath, err)\n\t}\n\n\tsparseIndexPath := path.Join(dbDir, prefix+diskTableSparseIndexFileName)\n\tsparseIndexFile, err := os.OpenFile(sparseIndexPath, newDiskTableFlag, 0600)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to open sparse index file %s: %w\", sparseIndexPath, err)\n\t}\n\n\treturn &diskTableWriter{\n\t\tdataFile: dataFile,\n\t\tindexFile: indexFile,\n\t\tsparseIndexFile: sparseIndexFile,\n\t\tsparseKeyDistance: sparseKeyDistance,\n\t\tkeyNum: 0,\n\t\tdataPos: 0,\n\t\tindexPos: 0,\n\t}, nil\n}" ]
[ "0.71595836", "0.6695677", "0.66890204", "0.6411042", "0.6318817", "0.6231678", "0.6222331", "0.62199277", "0.6124369", "0.6048241", "0.60400635", "0.59416604", "0.58752865", "0.58658576", "0.5824718", "0.5819471", "0.5782254", "0.57550234", "0.5740949", "0.5681647", "0.56515664", "0.5637585", "0.56193465", "0.5598222", "0.55791575", "0.5574086", "0.55548555", "0.5494645", "0.54934657", "0.5451317", "0.5425547", "0.5421569", "0.5419066", "0.54080737", "0.54009277", "0.5395952", "0.5394723", "0.53810406", "0.5373645", "0.5373122", "0.534524", "0.53386074", "0.5334909", "0.53310955", "0.53268266", "0.5317018", "0.5306171", "0.53042847", "0.5301194", "0.52733076", "0.52601385", "0.5252788", "0.52264845", "0.5219853", "0.5216084", "0.5214889", "0.52123165", "0.5210496", "0.5208203", "0.51836824", "0.5166609", "0.5153632", "0.51249003", "0.5113295", "0.5111559", "0.5108631", "0.5093176", "0.50910485", "0.50860673", "0.50835896", "0.50822854", "0.50666356", "0.5063049", "0.5058324", "0.5049902", "0.50349003", "0.502693", "0.5023586", "0.5023586", "0.5016788", "0.5012452", "0.49940476", "0.49929422", "0.4991167", "0.49897614", "0.4971469", "0.49709994", "0.49686533", "0.49600592", "0.4958432", "0.49509302", "0.49505162", "0.4944856", "0.49445704", "0.49220222", "0.4919616", "0.4917587", "0.49171862", "0.49164766", "0.49153593" ]
0.70917547
1
CreateForwardingRule uses the override method CreateForwardingRuleFn or the real implementation.
func (c *TestClient) CreateForwardingRule(project, region string, fr *compute.ForwardingRule) error { if c.CreateForwardingRuleFn != nil { return c.CreateForwardingRuleFn(project, region, fr) } return c.client.CreateForwardingRule(project, region, fr) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (f *fakeLB) CreateForwardingRule(_ context.Context, _ string, _ *govultr.ForwardingRule) (*govultr.ForwardingRule, *http.Response, error) {\n\tpanic(\"implement me\")\n}", "func NewForwardingRule(ctx *pulumi.Context,\n\tname string, args *ForwardingRuleArgs, opts ...pulumi.ResourceOption) (*ForwardingRule, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.AcceleratorId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'AcceleratorId'\")\n\t}\n\tif args.ListenerId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'ListenerId'\")\n\t}\n\tif args.RuleActions == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'RuleActions'\")\n\t}\n\tif args.RuleConditions == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'RuleConditions'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource ForwardingRule\n\terr := ctx.RegisterResource(\"alicloud:ga/forwardingRule:ForwardingRule\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (g *Google) createForwardingRule(targetLink string) (string, error) {\n\tif rule, err := g.getForwardingRule(); err == nil {\n\t\tlog.Infof(\"found ForwardingRule %s: %s\", forwardingRuleName, rule.SelfLink)\n\t\treturn rule.SelfLink, nil\n\t}\n\n\top, err := g.computeService.GlobalForwardingRules.Insert(g.project,\n\t\t&compute.ForwardingRule{\n\t\t\tName: forwardingRuleName,\n\t\t\tIPProtocol: cockroachProtocol,\n\t\t\tPortRange: fmt.Sprintf(\"%d\", g.context.Port),\n\t\t\tTarget: targetLink,\n\t\t}).Do()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif err := g.waitForOperation(op); err != nil {\n\t\treturn \"\", err\n\t}\n\n\tlog.Infof(\"created ForwardingRule %s: %s\", forwardingRuleName, op.TargetLink)\n\treturn op.TargetLink, nil\n}", "func (f *fakeLB) GetForwardingRule(_ context.Context, _, _ string) (*govultr.ForwardingRule, *http.Response, error) {\n\tpanic(\"implement me\")\n}", "func (f *fakeLB) DeleteForwardingRule(_ context.Context, _, _ string) error {\n\tpanic(\"implement me\")\n}", "func (m *MockFirewallServiceIface) CreatePortForwardingRule(p *CreatePortForwardingRuleParams) (*CreatePortForwardingRuleResponse, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"CreatePortForwardingRule\", p)\n\tret0, _ := ret[0].(*CreatePortForwardingRuleResponse)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func CreatePortForwarding(t *testing.T, client *gophercloud.ServiceClient, fipID string, portID string, portFixedIPs []ports.IP) (*portforwarding.PortForwarding, error) {\n\tt.Logf(\"Attempting to create Port forwarding for floating IP with ID: %s\", fipID)\n\n\tfixedIP := portFixedIPs[0]\n\tinternalIP := fixedIP.IPAddress\n\tcreateOpts := &portforwarding.CreateOpts{\n\t\tProtocol: \"tcp\",\n\t\tInternalPort: 25,\n\t\tExternalPort: 2230,\n\t\tInternalIPAddress: internalIP,\n\t\tInternalPortID: portID,\n\t}\n\n\tpf, err := portforwarding.Create(client, fipID, createOpts).Extract()\n\tif err != nil {\n\t\treturn pf, err\n\t}\n\n\tt.Logf(\"Created Port Forwarding.\")\n\n\tth.AssertEquals(t, pf.Protocol, \"tcp\")\n\n\treturn pf, err\n}", "func (c *TestClient) GetForwardingRule(project, region, name string) (*compute.ForwardingRule, error) {\n\tif c.GetForwardingRuleFn != nil {\n\t\treturn c.GetForwardingRuleFn(project, region, name)\n\t}\n\treturn c.client.GetForwardingRule(project, region, name)\n}", "func NewForwardingHook(logger Logger) Hook {\n\treturn &forwardingHook{logger}\n}", "func (r *oauthProxy) createForwardingProxy() error {\n\tr.log.Info(\"enabling forward signing mode, listening on\", zap.String(\"interface\", r.config.Listen))\n\n\tif r.config.SkipUpstreamTLSVerify {\n\t\tr.log.Warn(\"tls verification switched off. In forward signing mode it's recommended you verify! (--skip-upstream-tls-verify=false)\")\n\t}\n\tif err := r.createUpstreamProxy(nil); err != nil {\n\t\treturn err\n\t}\n\tforwardingHandler := r.forwardProxyHandler()\n\n\t// set the http handler\n\tproxy := r.upstream.(*goproxy.ProxyHttpServer)\n\tr.router = proxy\n\n\t// setup the tls configuration\n\tif r.config.TLSCaCertificate != \"\" && r.config.TLSCaPrivateKey != \"\" {\n\t\tca, err := loadCA(r.config.TLSCaCertificate, r.config.TLSCaPrivateKey)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unable to load certificate authority, error: %s\", err)\n\t\t}\n\n\t\t// implement the goproxy connect method\n\t\tproxy.OnRequest().HandleConnectFunc(\n\t\t\tfunc(host string, ctx *goproxy.ProxyCtx) (*goproxy.ConnectAction, string) {\n\t\t\t\treturn &goproxy.ConnectAction{\n\t\t\t\t\tAction: goproxy.ConnectMitm,\n\t\t\t\t\tTLSConfig: goproxy.TLSConfigFromCA(ca),\n\t\t\t\t}, host\n\t\t\t},\n\t\t)\n\t} else {\n\t\t// use the default certificate provided by goproxy\n\t\tproxy.OnRequest().HandleConnect(goproxy.AlwaysMitm)\n\t}\n\n\tproxy.OnResponse().DoFunc(func(resp *http.Response, ctx *goproxy.ProxyCtx) *http.Response {\n\t\t// @NOTES, somewhat annoying but goproxy hands back a nil response on proxy client errors\n\t\tif resp != nil && r.config.EnableLogging {\n\t\t\tstart := ctx.UserData.(time.Time)\n\t\t\tlatency := time.Since(start)\n\t\t\tlatencyMetric.Observe(latency.Seconds())\n\t\t\tr.log.Info(\"client request\",\n\t\t\t\tzap.String(\"method\", resp.Request.Method),\n\t\t\t\tzap.String(\"path\", resp.Request.URL.Path),\n\t\t\t\tzap.Int(\"status\", resp.StatusCode),\n\t\t\t\tzap.Int64(\"bytes\", resp.ContentLength),\n\t\t\t\tzap.String(\"host\", resp.Request.Host),\n\t\t\t\tzap.String(\"path\", resp.Request.URL.Path),\n\t\t\t\tzap.String(\"latency\", latency.String()))\n\t\t}\n\n\t\treturn resp\n\t})\n\tproxy.OnRequest().DoFunc(func(req *http.Request, ctx *goproxy.ProxyCtx) (*http.Request, *http.Response) {\n\t\tctx.UserData = time.Now()\n\t\tforwardingHandler(req, ctx.Resp)\n\t\treturn req, ctx.Resp\n\t})\n\n\treturn nil\n}", "func (mr *MockFirewallServiceIfaceMockRecorder) CreatePortForwardingRule(p interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"CreatePortForwardingRule\", reflect.TypeOf((*MockFirewallServiceIface)(nil).CreatePortForwardingRule), p)\n}", "func (f *fakeLB) ListForwardingRules(_ context.Context, _ string, _ *govultr.ListOptions) ([]govultr.ForwardingRule, *govultr.Meta, *http.Response, error) {\n\treturn []govultr.ForwardingRule{{\n\t\t\tRuleID: \"1234\",\n\t\t\tFrontendProtocol: \"tcp\",\n\t\t\tFrontendPort: 80,\n\t\t\tBackendProtocol: \"tcp\",\n\t\t\tBackendPort: 80,\n\t\t}}, &govultr.Meta{\n\t\t\tTotal: 0,\n\t\t\tLinks: &govultr.Links{\n\t\t\t\tNext: \"\",\n\t\t\t\tPrev: \"\",\n\t\t\t},\n\t\t}, nil, nil\n}", "func (g *Google) getForwardingRule() (*compute.ForwardingRule, error) {\n\treturn g.computeService.GlobalForwardingRules.Get(g.project, forwardingRuleName).Do()\n}", "func (g *Google) createFirewallRule() (string, error) {\n\tif rule, err := g.getFirewallRule(); err == nil {\n\t\tlog.Infof(\"found FirewallRule %s: %s\", firewallRuleName, rule.SelfLink)\n\t\treturn rule.SelfLink, nil\n\t}\n\n\top, err := g.computeService.Firewalls.Insert(g.project,\n\t\t&compute.Firewall{\n\t\t\tName: firewallRuleName,\n\t\t\tAllowed: []*compute.FirewallAllowed{\n\t\t\t\t{\n\t\t\t\t\tIPProtocol: cockroachProtocol,\n\t\t\t\t\tPorts: []string{\n\t\t\t\t\t\tfmt.Sprintf(\"%d\", g.context.Port),\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t\tSourceRanges: []string{\n\t\t\t\tallIPAddresses,\n\t\t\t},\n\t\t}).Do()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif err := g.waitForOperation(op); err != nil {\n\t\treturn \"\", err\n\t}\n\tlog.Infof(\"created FirewallRule %s: %s\", firewallRuleName, op.TargetLink)\n\treturn op.TargetLink, nil\n}", "func (s *ForwardingServiceOp) Create(portForwardingConfig *PortForwardingConfig) error {\n\tportForwardingJSON, err := json.Marshal(*portForwardingConfig)\n\tif err != nil {\n\t\treturn err\n\t}\n\treq, err := http.NewRequest(\"POST\", s.client.ServerURL+\"/cloudapi/portforwarding/create\", bytes.NewBuffer(portForwardingJSON))\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = s.client.Do(req)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (ruleset *DnsForwardingRuleset) ValidateCreate() (admission.Warnings, error) {\n\tvalidations := ruleset.createValidations()\n\tvar temp any = ruleset\n\tif runtimeValidator, ok := temp.(genruntime.Validator); ok {\n\t\tvalidations = append(validations, runtimeValidator.CreateValidations()...)\n\t}\n\treturn genruntime.ValidateCreate(validations)\n}", "func NewComputeForwardingRuleConverter() *ComputeForwardingRuleConverter {\n\treturn &ComputeForwardingRuleConverter{\n\t\tname: \"google_compute_forwarding_rule\",\n\t\tschema: schemaProvider.ResourcesMap[\"google_compute_forwarding_rule\"].Schema,\n\t}\n}", "func (m *MockFirewallServiceIface) NewCreatePortForwardingRuleParams(ipaddressid string, privateport int, protocol string, publicport int, virtualmachineid string) *CreatePortForwardingRuleParams {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"NewCreatePortForwardingRuleParams\", ipaddressid, privateport, protocol, publicport, virtualmachineid)\n\tret0, _ := ret[0].(*CreatePortForwardingRuleParams)\n\treturn ret0\n}", "func (r *rulesRPCHandler) CreateRule(ctx context.Context, rule *api.Rule) (*api.Rule, error) {\n\tr.Lock()\n\tclient := clients[clientID%numClients]\n\tclientID++\n\tr.Unlock()\n\tif err := client.Create(ctx, path.Join(\"/rules\", rule.Name), rule); err != nil {\n\t\treturn nil, err\n\t}\n\treturn rule, nil\n}", "func (cli *OpsGenieUserV2Client) ListForwardingRules(req userv2.ListUserForwardingRulesRequest) (*userv2.ListUserForwardingRulesResponse, error) {\n\tvar response userv2.ListUserForwardingRulesResponse\n\terr := cli.sendGetRequest(&req, &response)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &response, nil\n}", "func (c *TestClient) DeleteForwardingRule(project, region, name string) error {\n\tif c.DeleteForwardingRuleFn != nil {\n\t\treturn c.DeleteForwardingRuleFn(project, region, name)\n\t}\n\treturn c.client.DeleteForwardingRule(project, region, name)\n}", "func Create(c *gophercloud.ServiceClient, floatingIpId string, opts CreateOptsBuilder) (r CreateResult) {\n\tb, err := opts.ToPortForwardingCreateMap()\n\tif err != nil {\n\t\tr.Err = err\n\t\treturn\n\t}\n\t_, r.Err = c.Post(portForwardingUrl(c, floatingIpId), b, &r.Body, nil)\n\treturn\n}", "func (c *TestClient) ListForwardingRules(project, region string, opts ...ListCallOption) ([]*compute.ForwardingRule, error) {\n\tif c.ListForwardingRulesFn != nil {\n\t\treturn c.ListForwardingRulesFn(project, region, opts...)\n\t}\n\treturn c.client.ListForwardingRules(project, region, opts...)\n}", "func (r *V1Beta1ExtensionsIngressRule) createRule(ingress *v1beta1Extensions.Ingress, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(ingress)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(ingress)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(ingress)\n\t\t},\n\t}\n\treturn rule\n}", "func CreateRule(fileName string, namingRules []NamingRule, defaultRuleKind string) *bazel.Rule {\n\tkind := defaultRuleKind\n\tfor _, r := range namingRules {\n\t\tm := r.FileNameMatcher.FindStringSubmatch(fileName)\n\t\tif m != nil {\n\t\t\tkind = r.RuleKind\n\t\t\tbreak\n\t\t}\n\t}\n\tpkgName := filepath.Dir(fileName)\n\tif pkgName == \".\" {\n\t\tpkgName = \"\"\n\t}\n\tsrc := filepath.Base(fileName)\n\tname := strings.TrimSuffix(filepath.Base(fileName), filepath.Ext(fileName))\n\treturn bazel.NewRule(kind, pkgName, name, map[string]interface{}{\"srcs\": []string{src}})\n}", "func (c *TestClient) CreateFirewallRule(project string, i *compute.Firewall) error {\n\tif c.CreateFirewallRuleFn != nil {\n\t\treturn c.CreateFirewallRuleFn(project, i)\n\t}\n\treturn c.client.CreateFirewallRule(project, i)\n}", "func (r *GenericRule) createRule(resource *Resource, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(resource)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(resource)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(resource)\n\t\t},\n\t}\n\treturn rule\n}", "func NewNetworkLoadBalancerForwardingRuleTarget(ip string, port int32, weight int32) *NetworkLoadBalancerForwardingRuleTarget {\n\tthis := NetworkLoadBalancerForwardingRuleTarget{}\n\n\tthis.Ip = &ip\n\tthis.Port = &port\n\tthis.Weight = &weight\n\n\treturn &this\n}", "func NewSecurityLogForwardingCreateDefault(code int) *SecurityLogForwardingCreateDefault {\n\treturn &SecurityLogForwardingCreateDefault{\n\t\t_statusCode: code,\n\t}\n}", "func (o *AggregatedDomain) CreateForwardingPathList(child *ForwardingPathList) *bambou.Error {\n\n\treturn bambou.CurrentSession().CreateChild(o, child)\n}", "func (a *BackendOptionsApiService) CreateHTTPRequestRule(ctx _context.Context, parentName string, parentType string, httpRequestRule HttpRequestRule, localVarOptionals *CreateHTTPRequestRuleOpts) (HttpRequestRule, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodPost\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue HttpRequestRule\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/services/haproxy/configuration/http_request_rules\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tlocalVarQueryParams.Add(\"parent_name\", parameterToString(parentName, \"\"))\n\tlocalVarQueryParams.Add(\"parent_type\", parameterToString(parentType, \"\"))\n\tif localVarOptionals != nil && localVarOptionals.TransactionId.IsSet() {\n\t\tlocalVarQueryParams.Add(\"transaction_id\", parameterToString(localVarOptionals.TransactionId.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Version.IsSet() {\n\t\tlocalVarQueryParams.Add(\"version\", parameterToString(localVarOptionals.Version.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.ForceReload.IsSet() {\n\t\tlocalVarQueryParams.Add(\"force_reload\", parameterToString(localVarOptionals.ForceReload.Value(), \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{\"application/json\"}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &httpRequestRule\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 201 {\n\t\t\tvar v HttpRequestRule\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 202 {\n\t\t\tvar v HttpRequestRule\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 400 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 409 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tvar v ModelError\n\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\tif err != nil {\n\t\t\tnewErr.error = err.Error()\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tnewErr.model = v\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func (s *LocalTests) createFirewallRule(c *gc.C) *cloudapi.FirewallRule {\n\tfwRule, err := s.testClient.CreateFirewallRule(cloudapi.CreateFwRuleOpts{Enabled: false, Rule: testFwRule})\n\tc.Assert(err, gc.IsNil)\n\tc.Assert(fwRule, gc.NotNil)\n\tc.Assert(fwRule.Rule, gc.Equals, testFwRule)\n\tc.Assert(fwRule.Enabled, gc.Equals, false)\n\ttime.Sleep(10 * time.Second)\n\n\treturn fwRule\n}", "func Forward(in, out Link) rules.Rule {\n\treturn rules.Rule(fmt.Sprintf(\n\t\t\"-t filter -A fw-interfaces -j ACCEPT -i %v -o %v\",\n\t\tin.Name(), out.Name()))\n}", "func (t *OpenconfigQos_Qos_ForwardingGroups) NewForwardingGroup(Name string) (*OpenconfigQos_Qos_ForwardingGroups_ForwardingGroup, error){\n\n\t// Initialise the list within the receiver struct if it has not already been\n\t// created.\n\tif t.ForwardingGroup == nil {\n\t\tt.ForwardingGroup = make(map[string]*OpenconfigQos_Qos_ForwardingGroups_ForwardingGroup)\n\t}\n\n\tkey := Name\n\n\t// Ensure that this key has not already been used in the\n\t// list. Keyed YANG lists do not allow duplicate keys to\n\t// be created.\n\tif _, ok := t.ForwardingGroup[key]; ok {\n\t\treturn nil, fmt.Errorf(\"duplicate key %v for list ForwardingGroup\", key)\n\t}\n\n\tt.ForwardingGroup[key] = &OpenconfigQos_Qos_ForwardingGroups_ForwardingGroup{\n\t\tName: &Name,\n\t}\n\n\treturn t.ForwardingGroup[key], nil\n}", "func (fw *IPtables) CreateDefaultRule(chain int, target string) error {\n\tlog.Infof(\"In CreateDefaultRule() %s rules for chain %d\", target, chain)\n\tchainName := fw.chains[chain].ChainName\n\tbody := fmt.Sprintf(\"%s %s %s\", chainName, \"-j\", target)\n\trule := &IPtablesRule{\n\t\tBody: body,\n\t\tState: setRuleInactive.String(),\n\t}\n\n\t// First create rule record in database.\n\terr0 := fw.addIPtablesRule(rule)\n\tif err0 != nil {\n\t\tlog.Error(\"In CreateDefaultRule() create db record for iptables rule \", rule.GetBody())\n\t\treturn err0\n\t}\n\n\terr1 := fw.EnsureRule(rule, EnsureLast)\n\tif err1 != nil {\n\t\tlog.Errorf(\"In CreateDefaultRule() %s rules failed\", target)\n\t\treturn err1\n\t}\n\n\t// Finally, set 'active' flag in database record.\n\tif err2 := fw.Store.switchIPtablesRule(rule, setRuleActive); err2 != nil {\n\t\tlog.Error(\"In CreateDefaultRule() iptables rule created but activation failed \", rule.GetBody())\n\t\treturn err2\n\t}\n\n\tlog.Info(\"In CreateDefaultRule() success\")\n\treturn nil\n}", "func GetForwardingRule(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *ForwardingRuleState, opts ...pulumi.ResourceOption) (*ForwardingRule, error) {\n\tvar resource ForwardingRule\n\terr := ctx.ReadResource(\"alicloud:ga/forwardingRule:ForwardingRule\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (egw *NsxtEdgeGateway) CreateNatRule(natRuleConfig *types.NsxtNatRule) (*NsxtNatRule, error) {\n\tclient := egw.client\n\tendpoint := types.OpenApiPathVersion1_0_0 + types.OpenApiEndpointNsxtNatRules\n\tapiVersion, err := client.getOpenApiHighestElevatedVersion(endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Insert Edge Gateway ID into endpoint path edgeGateways/%s/nat/rules\n\turlRef, err := client.OpenApiBuildEndpoint(fmt.Sprintf(endpoint, egw.EdgeGateway.ID))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Creating NAT rule must follow different way than usual OpenAPI one because this item has an API bug and\n\t// NAT rule ID is not returned after this object is created. The only way to find its ID afterwards is to GET all\n\t// items, and manually match it based on rule name, etc.\n\ttask, err := client.OpenApiPostItemAsync(apiVersion, urlRef, nil, natRuleConfig)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error creating NSX-T NAT rule: %s\", err)\n\t}\n\n\terr = task.WaitTaskCompletion()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"task failed while creating NSX-T NAT rule: %s\", err)\n\t}\n\n\t// queryParameters (API side filtering) are not used because pretty much nothing is accepted as filter (such fields as\n\t// name, description, ruleType and even ID are not allowed\n\tallNatRules, err := egw.GetAllNatRules(nil)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error fetching all NAT rules: %s\", err)\n\t}\n\n\tfor index, singleRule := range allNatRules {\n\t\t// Look for a matching rule\n\t\tif singleRule.IsEqualTo(natRuleConfig) {\n\t\t\treturn allNatRules[index], nil\n\n\t\t}\n\t}\n\treturn nil, fmt.Errorf(\"rule '%s' of type '%s' not found after creation\", natRuleConfig.Name, natRuleConfig.RuleType)\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) PutFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRulesRule(ctx context.Context, name string, policyId string, sequenceId int64, frinxOpenconfigPolicyForwardingPfforwardingpolicystructuralPoliciesPolicyRulesRuleBodyParam FrinxOpenconfigPolicyForwardingPfforwardingpolicystructuralPoliciesPolicyRulesRuleRequest, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Put\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/frinx-openconfig-network-instance:rule/{sequence-id}/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"sequence-id\"+\"}\", fmt.Sprintf(\"%v\", sequenceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &frinxOpenconfigPolicyForwardingPfforwardingpolicystructuralPoliciesPolicyRulesRuleBodyParam\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func CreateForwarder(entry config.AmazonEntry, snsClient ...snsiface.SNSAPI) forwarder.Client {\n\tvar client snsiface.SNSAPI\n\tif len(snsClient) > 0 {\n\t\tclient = snsClient[0]\n\t} else {\n\t\tclient = sns.New(session.Must(session.NewSession()))\n\t}\n\tforwarder := Forwarder{entry.Name, client, entry.Target}\n\tlog.WithField(\"forwarderName\", forwarder.Name()).Info(\"Created forwarder\")\n\treturn forwarder\n}", "func (p *fleetUpdate) CreateRules() ([]string, error) {\n\t// No rules here\n\treturn nil, nil\n}", "func NewForwarder(cc *ClusterConfig) proto.Forwarder {\n\t// new Forwarder\n\tif _, ok := defaultForwardCacheTypes[cc.CacheType]; ok {\n\t\treturn newDefaultForwarder(cc)\n\t}\n\tif cc.CacheType == types.CacheTypeRedisCluster {\n\t\tdto := time.Duration(cc.DialTimeout) * time.Millisecond\n\t\trto := time.Duration(cc.ReadTimeout) * time.Millisecond\n\t\twto := time.Duration(cc.WriteTimeout) * time.Millisecond\n\t\treturn rclstr.NewForwarder(cc.Name, cc.ListenAddr, cc.Servers, cc.NodeConnections, cc.NodePipeCount, dto, rto, wto, []byte(cc.HashTag))\n\t}\n\tpanic(\"unsupported protocol\")\n}", "func (opts CreateOpts) ToPortForwardingCreateMap() (map[string]interface{}, error) {\n\treturn gophercloud.BuildRequestBody(opts, \"port_forwarding\")\n}", "func (r *V1ServiceRule) createRule(service *v1.Service, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(service)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(service)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(service)\n\t\t},\n\t}\n\treturn rule\n}", "func (r *V1NamespaceRule) createRule(namespace *v1.Namespace, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(namespace)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(namespace)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(namespace)\n\t\t},\n\t}\n\treturn rule\n}", "func NewProxyForward(name, hostname, proxyHostname, localPort, forwardPort string) *ProxyForward {\n\tproxyForward := &ProxyForward{\n\t\tName: name,\n\t\tHostname: hostname,\n\t\tProxyHostname: proxyHostname,\n\t\tLocalPort: localPort,\n\t\tForwardPort: forwardPort,\n\t}\n\n\t// In case of a forward type 'proxy', just set the proxy port with\n\t// the given forward port (proxy component will not generate one)\n\tif proxyHostname != \"\" {\n\t\tproxyForward.ProxyPort = forwardPort\n\t}\n\n\treturn proxyForward\n}", "func (r *V1Beta1ExtensionsNetworkPolicyRule) createRule(networkpolicy *v1beta1Extensions.NetworkPolicy, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(networkpolicy)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(networkpolicy)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(networkpolicy)\n\t\t},\n\t}\n\treturn rule\n}", "func (r *NetworkingV1NetworkPolicyRule) createRule(networkpolicy *networkingV1.NetworkPolicy, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(networkpolicy)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(networkpolicy)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(networkpolicy)\n\t\t},\n\t}\n\treturn rule\n}", "func (r *oauthProxy) createStdProxy(upstream *url.URL) error {\n\tdialer := (&net.Dialer{\n\t\tKeepAlive: r.config.UpstreamKeepaliveTimeout,\n\t\tTimeout: r.config.UpstreamTimeout, // NOTE(http2): in order to properly receive response headers, this have to be less than ServerWriteTimeout\n\t}).DialContext\n\n\t// are we using a unix socket?\n\t// TODO(fredbi): this does not work with multiple upstream configuration\n\t// TODO(fredbi): create as many upstreams as different upstream schemes\n\tif upstream != nil && upstream.Scheme == \"unix\" {\n\t\tr.log.Info(\"using unix socket for upstream\", zap.String(\"socket\", fmt.Sprintf(\"%s%s\", upstream.Host, upstream.Path)))\n\n\t\tsocketPath := fmt.Sprintf(\"%s%s\", upstream.Host, upstream.Path)\n\t\tdialer = func(_ context.Context, network, address string) (net.Conn, error) {\n\t\t\treturn net.Dial(\"unix\", socketPath)\n\t\t}\n\t\tupstream.Path = \"\"\n\t\tupstream.Host = \"domain-sock\"\n\t\tupstream.Scheme = unsecureScheme\n\t}\n\n\t// create the upstream tls configuration\n\ttlsConfig, err := r.buildProxyTLSConfig()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttransport := &http.Transport{\n\t\tForceAttemptHTTP2: true,\n\t\tDialContext: dialer,\n\t\tTLSClientConfig: tlsConfig,\n\t\tTLSHandshakeTimeout: r.config.UpstreamTLSHandshakeTimeout,\n\t\tMaxIdleConns: r.config.MaxIdleConns,\n\t\tMaxIdleConnsPerHost: r.config.MaxIdleConnsPerHost,\n\t\tDisableKeepAlives: !r.config.UpstreamKeepalives,\n\t\tExpectContinueTimeout: r.config.UpstreamExpectContinueTimeout,\n\t\tResponseHeaderTimeout: r.config.UpstreamResponseHeaderTimeout,\n\t}\n\tif err = http2.ConfigureTransport(transport); err != nil {\n\t\treturn err\n\t}\n\tr.upstream = &httputil.ReverseProxy{\n\t\tDirector: func(*http.Request) {}, // most of the work is already done by middleware above. Some of this could be done by Director just as well\n\t\tTransport: transport,\n\t\tErrorHandler: func(w http.ResponseWriter, req *http.Request, err error) {\n\t\t\t_, span, logger := r.traceSpan(req.Context(), \"reverse proxy middleware\")\n\t\t\tif span != nil {\n\t\t\t\tdefer span.End()\n\t\t\t\tspan.SetStatus(trace.Status{Code: trace.StatusCodeInternal, Message: err.Error()})\n\t\t\t}\n\n\t\t\tlogger.Warn(\"reverse proxy error\", zap.Error(err))\n\t\t\tr.errorResponse(w, req, \"\", http.StatusBadGateway, err)\n\t\t},\n\t\tModifyResponse: func(res *http.Response) error {\n\t\t\tif r.config.Verbose {\n\t\t\t\t// debug response headers\n\t\t\t\tr.log.Debug(\"response from upstream\",\n\t\t\t\t\tzap.Int(\"status code\", res.StatusCode),\n\t\t\t\t\tzap.String(\"proto\", res.Proto),\n\t\t\t\t\tzap.Int64(\"content-length\", res.ContentLength),\n\t\t\t\t\tzap.Any(\"headers\", res.Header))\n\t\t\t}\n\t\t\t// filter out possible conflicting headers from upstream (i.e. gatekeeper value override)\n\t\t\tif r.config.EnableSecurityFilter {\n\t\t\t\tif r.config.EnableBrowserXSSFilter {\n\t\t\t\t\tres.Header.Del(headerXXSSProtection)\n\t\t\t\t}\n\t\t\t\tif r.config.ContentSecurityPolicy != \"\" {\n\t\t\t\t\tres.Header.Del(headerXPolicy)\n\t\t\t\t}\n\t\t\t\tif r.config.EnableContentNoSniff {\n\t\t\t\t\tres.Header.Del(headerXContentTypeOptions)\n\t\t\t\t}\n\t\t\t\tif r.config.EnableFrameDeny {\n\t\t\t\t\tres.Header.Del(headerXFrameOptions)\n\t\t\t\t}\n\t\t\t\tif r.config.EnableSTS || r.config.EnableSTSPreload {\n\t\t\t\t\tres.Header.Del(headerXSTS)\n\t\t\t\t}\n\t\t\t}\n\t\t\tfor hdr := range r.config.Headers {\n\t\t\t\tres.Header.Del(hdr)\n\t\t\t}\n\n\t\t\tif len(r.config.CorsOrigins) > 0 {\n\t\t\t\t// remove cors headers from upstream\n\t\t\t\t// This avoids the concatenation of multiple headers whenever\n\t\t\t\t// upstreams response provides some CORS headers.\n\t\t\t\tres.Header.Del(\"Access-Control-Allow-Origin\")\n\t\t\t\tres.Header.Del(\"Access-Control-Allow-Credentials\")\n\t\t\t\tres.Header.Del(\"Access-Control-Allow-Headers\")\n\t\t\t\tres.Header.Del(\"Access-Control-Allow-Methods\")\n\t\t\t\tres.Header.Del(\"Access-Control-Max-Age\")\n\t\t\t}\n\t\t\treturn nil\n\t\t},\n\t}\n\n\treturn nil\n}", "func checkCreateRule(t *testing.T, expError bool, tenant, policy, ruleID, dir, fnet, fepg, fip, tnet, tepg, tip, proto, act string, prio, port int) {\n\tpol := client.Rule{\n\t\tTenantName: tenant,\n\t\tPolicyName: policy,\n\t\tRuleID: ruleID,\n\t\tDirection: dir,\n\t\tPriority: prio,\n\t\tFromNetwork: fnet,\n\t\tFromEndpointGroup: fepg,\n\t\tFromIpAddress: fip,\n\t\tToNetwork: tnet,\n\t\tToEndpointGroup: tepg,\n\t\tToIpAddress: tip,\n\t\tProtocol: proto,\n\t\tPort: port,\n\t\tAction: act,\n\t}\n\terr := contivClient.RulePost(&pol)\n\tif err != nil && !expError {\n\t\tt.Fatalf(\"Error creating rule {%+v}. Err: %v\", pol, err)\n\t} else if err == nil && expError {\n\t\tt.Fatalf(\"Create rule {%+v} succeded while expecing error\", pol)\n\t} else if err == nil {\n\t\t// verify rule is created\n\t\t_, err := contivClient.RuleGet(tenant, policy, ruleID)\n\t\tif err != nil {\n\t\t\tt.Fatalf(\"Error getting rule %s/%s/%s. Err: %v\", tenant, policy, ruleID, err)\n\t\t}\n\t}\n}", "func (r *RbacV1Beta1RoleBindingRule) createRule(rolebinding *rbacV1beta1.RoleBinding, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(rolebinding)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(rolebinding)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(rolebinding)\n\t\t},\n\t}\n\treturn rule\n}", "func NewPortForward(logger log.Logger, objectStore store.Store, portForwarder portforward.PortForwarder) *PortForward {\n\treturn &PortForward{\n\t\tlogger: logger,\n\t\tobjectStore: objectStore,\n\t\tportForwarder: portForwarder,\n\t}\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) PutFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRulesRuleAction(ctx context.Context, name string, policyId string, sequenceId int64, frinxOpenconfigPolicyForwardingPfforwardingpolicystructuralPoliciesPolicyRulesRuleActionBodyParam FrinxOpenconfigPolicyForwardingPfforwardingpolicystructuralPoliciesPolicyRulesRuleActionRequest, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Put\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/frinx-openconfig-network-instance:rule/{sequence-id}/frinx-openconfig-network-instance:action/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"sequence-id\"+\"}\", fmt.Sprintf(\"%v\", sequenceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &frinxOpenconfigPolicyForwardingPfforwardingpolicystructuralPoliciesPolicyRulesRuleActionBodyParam\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (r *V1Beta1ExtensionsDeploymentRule) createRule(deployment *v1beta1Extensions.Deployment, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(deployment)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(deployment)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(deployment)\n\t\t},\n\t}\n\treturn rule\n}", "func (m *MockFirewallServiceIface) ListPortForwardingRules(p *ListPortForwardingRulesParams) (*ListPortForwardingRulesResponse, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"ListPortForwardingRules\", p)\n\tret0, _ := ret[0].(*ListPortForwardingRulesResponse)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (f *FakeInstance) CreateReverseIPv6(_ context.Context, _ string, _ *govultr.ReverseIP) error {\n\tpanic(\"implement me\")\n}", "func (r *V1ServiceAccountRule) createRule(serviceaccount *v1.ServiceAccount, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(serviceaccount)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(serviceaccount)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(serviceaccount)\n\t\t},\n\t}\n\treturn rule\n}", "func (*PageRuleApiV1) NewPageRulesBodyActionsItemActionsForwardingURL(id string) (model *PageRulesBodyActionsItemActionsForwardingURL, err error) {\n\tmodel = &PageRulesBodyActionsItemActionsForwardingURL{\n\t\tID: core.StringPtr(id),\n\t}\n\terr = core.ValidateStruct(model, \"required parameters\")\n\treturn\n}", "func (self *PolicyAgent) AddRule(rule *OfnetPolicyRule, ret *bool) error {\n\tvar ipDa *net.IP = nil\n\tvar ipDaMask *net.IP = nil\n\tvar ipSa *net.IP = nil\n\tvar ipSaMask *net.IP = nil\n\tvar md *uint64 = nil\n\tvar mdm *uint64 = nil\n\tvar flag, flagMask uint16\n\tvar flagPtr, flagMaskPtr *uint16\n\tvar err error\n\n\t// make sure switch is connected\n\tif !self.agent.IsSwitchConnected() {\n\t\tself.agent.WaitForSwitchConnection()\n\t}\n\n\t// check if we already have the rule\n\tself.mutex.RLock()\n\tif self.Rules[rule.RuleId] != nil {\n\t\toldRule := self.Rules[rule.RuleId].Rule\n\n\t\tif ruleIsSame(oldRule, rule) {\n\t\t\tself.mutex.RUnlock()\n\t\t\treturn nil\n\t\t} else {\n\t\t\tself.mutex.RUnlock()\n\t\t\tlog.Errorf(\"Rule already exists. new rule: {%+v}, old rule: {%+v}\", rule, oldRule)\n\t\t\treturn errors.New(\"Rule already exists\")\n\t\t}\n\t}\n\tself.mutex.RUnlock()\n\n\tlog.Infof(\"Received AddRule: %+v\", rule)\n\n\t// Parse dst ip\n\tif rule.DstIpAddr != \"\" {\n\t\tipDa, ipDaMask, err = ParseIPAddrMaskString(rule.DstIpAddr)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error parsing dst ip %s. Err: %v\", rule.DstIpAddr, err)\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// parse src ip\n\tif rule.SrcIpAddr != \"\" {\n\t\tipSa, ipSaMask, err = ParseIPAddrMaskString(rule.SrcIpAddr)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error parsing src ip %s. Err: %v\", rule.SrcIpAddr, err)\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// parse source/dst endpoint groups\n\tif rule.SrcEndpointGroup != 0 && rule.DstEndpointGroup != 0 {\n\t\tsrcMetadata, srcMetadataMask := SrcGroupMetadata(rule.SrcEndpointGroup)\n\t\tdstMetadata, dstMetadataMask := DstGroupMetadata(rule.DstEndpointGroup)\n\t\tmetadata := srcMetadata | dstMetadata\n\t\tmetadataMask := srcMetadataMask | dstMetadataMask\n\t\tmd = &metadata\n\t\tmdm = &metadataMask\n\t} else if rule.SrcEndpointGroup != 0 {\n\t\tsrcMetadata, srcMetadataMask := SrcGroupMetadata(rule.SrcEndpointGroup)\n\t\tmd = &srcMetadata\n\t\tmdm = &srcMetadataMask\n\t} else if rule.DstEndpointGroup != 0 {\n\t\tdstMetadata, dstMetadataMask := DstGroupMetadata(rule.DstEndpointGroup)\n\t\tmd = &dstMetadata\n\t\tmdm = &dstMetadataMask\n\t}\n\n\t// Setup TCP flags\n\tif rule.IpProtocol == 6 && rule.TcpFlags != \"\" {\n\t\tswitch rule.TcpFlags {\n\t\tcase \"syn\":\n\t\t\tflag = TCP_FLAG_SYN\n\t\t\tflagMask = TCP_FLAG_SYN\n\t\tcase \"syn,ack\":\n\t\t\tflag = TCP_FLAG_ACK | TCP_FLAG_SYN\n\t\t\tflagMask = TCP_FLAG_ACK | TCP_FLAG_SYN\n\t\tcase \"ack\":\n\t\t\tflag = TCP_FLAG_ACK\n\t\t\tflagMask = TCP_FLAG_ACK\n\t\tcase \"syn,!ack\":\n\t\t\tflag = TCP_FLAG_SYN\n\t\t\tflagMask = TCP_FLAG_ACK | TCP_FLAG_SYN\n\t\tcase \"!syn,ack\":\n\t\t\tflag = TCP_FLAG_ACK\n\t\t\tflagMask = TCP_FLAG_ACK | TCP_FLAG_SYN\n\t\tdefault:\n\t\t\tlog.Errorf(\"Unknown TCP flags: %s, in rule: %+v\", rule.TcpFlags, rule)\n\t\t\treturn errors.New(\"Unknown TCP flag\")\n\t\t}\n\n\t\tflagPtr = &flag\n\t\tflagMaskPtr = &flagMask\n\t}\n\t// Install the rule in policy table\n\truleFlow, err := self.policyTable.NewFlow(ofctrl.FlowMatch{\n\t\tPriority: uint16(FLOW_POLICY_PRIORITY_OFFSET + rule.Priority),\n\t\tEthertype: 0x0800,\n\t\tIpDa: ipDa,\n\t\tIpDaMask: ipDaMask,\n\t\tIpSa: ipSa,\n\t\tIpSaMask: ipSaMask,\n\t\tIpProto: rule.IpProtocol,\n\t\tTcpSrcPort: rule.SrcPort,\n\t\tTcpDstPort: rule.DstPort,\n\t\tUdpSrcPort: rule.SrcPort,\n\t\tUdpDstPort: rule.DstPort,\n\t\tMetadata: md,\n\t\tMetadataMask: mdm,\n\t\tTcpFlags: flagPtr,\n\t\tTcpFlagsMask: flagMaskPtr,\n\t})\n\tif err != nil {\n\t\tlog.Errorf(\"Error adding flow for rule {%v}. Err: %v\", rule, err)\n\t\treturn err\n\t}\n\n\t// Point it to next table\n\tif rule.Action == \"allow\" {\n\t\terr = ruleFlow.Next(self.nextTable)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error installing flow {%+v}. Err: %v\", ruleFlow, err)\n\t\t\treturn err\n\t\t}\n\t} else if rule.Action == \"deny\" {\n\t\terr = ruleFlow.Next(self.ofSwitch.DropAction())\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error installing flow {%+v}. Err: %v\", ruleFlow, err)\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\tlog.Errorf(\"Unknown action in rule {%+v}\", rule)\n\t\treturn errors.New(\"Unknown action in rule\")\n\t}\n\n\t// save the rule\n\tpRule := PolicyRule{\n\t\tRule: rule,\n\t\tflow: ruleFlow,\n\t}\n\tself.mutex.Lock()\n\tself.Rules[rule.RuleId] = &pRule\n\tself.mutex.Unlock()\n\n\treturn nil\n}", "func NewForwarder() *Forwarder {\n\treturn &Forwarder{}\n}", "func (r *Router) AddRule(w http.ResponseWriter, req *http.Request, _ httprouter.Params) {\n\ts, rule, err := r.readRule(req)\n\tif err != nil {\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\tw.Write([]byte(err.Error()))\n\t\treturn\n\t}\n\t// Put the rule into the proxy that belongs to the correct service\n\ts.Proxy.AddRule(*rule)\n\t// Everything went well\n\tw.Write([]byte(config.OK))\n\tlog.Debug(\"Added rule\")\n}", "func (r *BatchV1JobRule) createRule(job *batchV1.Job, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(job)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(job)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(job)\n\t\t},\n\t}\n\treturn rule\n}", "func (d *AppsV1DeploymentRule) createRule(deployment *appsv1.Deployment, ydr *YamlDerivedResource) *rule {\n\tr := &rule{\n\t\tID: d.ID,\n\t\tPrereqs: d.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif d.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn d.Condition(deployment)\n\t\t},\n\t\tMessage: d.Message,\n\t\tLevel: d.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif d.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn d.Fix(deployment)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif d.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn d.FixDescription(deployment)\n\t\t},\n\t}\n\treturn r\n}", "func newRuleHandler(d discovery.Discoverer, cfg RunConfig) discovery.RuleHandler {\n\trh := &ruleHandler{\n\t\td: d.(*discoverer),\n\t\tph: cfg.Handler,\n\t\tdaemon: cfg.Daemon,\n\t\tkubeClient: cfg.KubeClient,\n\t\tlister: cfg.Lister,\n\t\trulesCount: gm.GetOrRegisterGauge(\"discovery.rules.count\", gm.DefaultRegistry),\n\t}\n\tcount := int64(len(rh.d.delegates))\n\trh.rulesCount.Update(count)\n\treturn rh\n}", "func (c *Client) CreateRule(args *CreateRuleArgs) (*CreateRuleResult, error) {\n\tjsonBytes, jsonErr := json.Marshal(args)\n\tif jsonErr != nil {\n\t\treturn nil, jsonErr\n\t}\n\tbody, err := bce.NewBodyFromBytes(jsonBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn CreateRule(c, body)\n}", "func TestFunnelCreate(t *testing.T) {\n\twithDB(func(db *DB) {\n\t\tdb.Do(func(tx *Tx) error {\n\t\t\t// Create an account and funnel.\n\t\t\ta := &Account{}\n\t\t\tassert.NoError(t, tx.CreateAccount(a))\n\t\t\tf := &Funnel{Name: \"Funnel Y\", Steps: []*FunnelStep{{Condition: \"action == 'foo'\"}}}\n\t\t\tassert.NoError(t, a.CreateFunnel(f))\n\t\t\tassert.Equal(t, f.ID(), 1)\n\n\t\t\t// Retrieve the funnel.\n\t\t\tf2, err := tx.Funnel(1)\n\t\t\tif assert.NoError(t, err) && assert.NotNil(t, f2) {\n\t\t\t\tassert.Equal(t, f2.Tx, tx)\n\t\t\t\tassert.Equal(t, f2.ID(), 1)\n\t\t\t\tassert.Equal(t, f2.AccountID, 1)\n\t\t\t\tassert.Equal(t, f2.Name, \"Funnel Y\")\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\t})\n}", "func forwardFunc(\n\tname string,\n\tforwardProcessors []*RunningProcessor,\n\tforwardSinks []*RunningSink,\n) func(optic.Event) {\n\n\tlenFilters := len(forwardProcessors)\n\tlenSinks := len(forwardSinks)\n\tlenForwards := lenFilters + lenSinks\n\n\tswitch {\n\tcase lenForwards == 1:\n\t\tif lenFilters == 1 {\n\t\t\treturn func(event optic.Event) {\n\t\t\t\tforwardProcessors[0].ForwardEvent(event)\n\t\t\t}\n\t\t}\n\t\treturn func(event optic.Event) {\n\t\t\tforwardSinks[0].WriteEvent(event)\n\t\t}\n\tcase lenForwards > 1:\n\t\tswitch {\n\t\tcase lenFilters == 1 && lenSinks == 1:\n\t\t\treturn func(event optic.Event) {\n\t\t\t\tforwardProcessors[0].ForwardEvent(event.Copy())\n\t\t\t\tforwardSinks[0].WriteEvent(event)\n\t\t\t}\n\t\tcase lenFilters > 0 && lenSinks > 0:\n\t\t\treturn func(event optic.Event) {\n\t\t\t\tfor _, ff := range forwardProcessors {\n\t\t\t\t\tff.ForwardEvent(event.Copy())\n\t\t\t\t}\n\t\t\t\tfor i, fs := range forwardSinks {\n\t\t\t\t\tif i == lenSinks-1 {\n\t\t\t\t\t\tfs.WriteEvent(event)\n\t\t\t\t\t} else {\n\t\t\t\t\t\tfs.WriteEvent(event.Copy())\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\tcase lenFilters > 0:\n\t\t\treturn func(event optic.Event) {\n\t\t\t\tfor i, ff := range forwardProcessors {\n\t\t\t\t\tif i == lenFilters-1 {\n\t\t\t\t\t\tff.ForwardEvent(event)\n\t\t\t\t\t} else {\n\t\t\t\t\t\tff.ForwardEvent(event.Copy())\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\tcase lenSinks > 0:\n\t\t\treturn func(event optic.Event) {\n\t\t\t\tfor i, fs := range forwardSinks {\n\t\t\t\t\tif i == lenSinks-1 {\n\t\t\t\t\t\tfs.WriteEvent(event)\n\t\t\t\t\t} else {\n\t\t\t\t\t\tfs.WriteEvent(event.Copy())\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\tlog.Printf(\"INFO [%s] will not forward events anywhere.\", name)\n\n\treturn func(event optic.Event) {\n\t\t// no-op\n\t}\n}", "func (a *BackendOptionsApiService) CreateHTTPResponseRule(ctx _context.Context, parentName string, parentType string, httpResponseRule HttpResponseRule, localVarOptionals *CreateHTTPResponseRuleOpts) (HttpResponseRule, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodPost\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue HttpResponseRule\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/services/haproxy/configuration/http_response_rules\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tlocalVarQueryParams.Add(\"parent_name\", parameterToString(parentName, \"\"))\n\tlocalVarQueryParams.Add(\"parent_type\", parameterToString(parentType, \"\"))\n\tif localVarOptionals != nil && localVarOptionals.TransactionId.IsSet() {\n\t\tlocalVarQueryParams.Add(\"transaction_id\", parameterToString(localVarOptionals.TransactionId.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Version.IsSet() {\n\t\tlocalVarQueryParams.Add(\"version\", parameterToString(localVarOptionals.Version.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.ForceReload.IsSet() {\n\t\tlocalVarQueryParams.Add(\"force_reload\", parameterToString(localVarOptionals.ForceReload.Value(), \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{\"application/json\"}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &httpResponseRule\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 201 {\n\t\t\tvar v HttpResponseRule\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 202 {\n\t\t\tvar v HttpResponseRule\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 400 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 409 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tvar v ModelError\n\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\tif err != nil {\n\t\t\tnewErr.error = err.Error()\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tnewErr.model = v\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func NewApplicationLoadBalancerForwardingRuleProperties(name string, protocol string, listenerIp string, listenerPort int32) *ApplicationLoadBalancerForwardingRuleProperties {\n\tthis := ApplicationLoadBalancerForwardingRuleProperties{}\n\n\tthis.Name = &name\n\tthis.Protocol = &protocol\n\tthis.ListenerIp = &listenerIp\n\tthis.ListenerPort = &listenerPort\n\n\treturn &this\n}", "func (m *ItemMailFoldersItemMessagesMessageItemRequestBuilder) CreateForward()(*ItemMailFoldersItemMessagesItemCreateForwardRequestBuilder) {\n return NewItemMailFoldersItemMessagesItemCreateForwardRequestBuilderInternal(m.BaseRequestBuilder.PathParameters, m.BaseRequestBuilder.RequestAdapter)\n}", "func newRuleImpl(impl starlark.Callable, defaults *starlark.Dict) (*ruleImpl, error) {\n\tpairs := defaults.Items()\n\tfor _, pair := range pairs {\n\t\tk, v := pair[0], pair[1]\n\t\tif _, ok := k.(starlark.String); !ok {\n\t\t\treturn nil, fmt.Errorf(\"lucicfg.rule: keys in \\\"defaults\\\" must be strings\")\n\t\t}\n\t\tif !isNamedStruct(v, \"lucicfg.var\") {\n\t\t\treturn nil, fmt.Errorf(\"lucicfg.rule: values in \\\"defaults\\\" must be lucicfg.var\")\n\t\t}\n\t}\n\treturn &ruleImpl{\n\t\tCallable: impl,\n\t\tdefaults: starlarkstruct.FromKeywords(starlark.String(\"lucicfg.rule.defaults\"), pairs),\n\t}, nil\n}", "func (r *V1PodSpecRule) createRule(podSpec *v1.PodSpec, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(podSpec)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(podSpec)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(podSpec)\n\t\t},\n\t}\n\treturn rule\n}", "func (m *MockIPTService) CreateRule(ruleType int, ruleData interface{}) error {\n\tre, _, err := m.s.CreateRuleEntryString(ruleType, ruleData)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, ok := m.rules[re.ID]\n\tif ok {\n\t\treturn errors.New(\"Rule already exists\")\n\t}\n\n\tm.rules[re.ID] = re\n\n\treturn nil\n}", "func New(setters ...optSetter) (*Forwarder, error) {\n\tf := &Forwarder{\n\t\thttpForwarder: &httpForwarder{\n\t\t\tlog: &internalLogger{Logger: log.StandardLogger()},\n\t\t\twebsocketPongWait: 30 * time.Second,\n\t\t},\n\t\thandlerContext: &handlerContext{},\n\t}\n\n\tf.websocketDialer = &websocket.Dialer{\n\t\tProxy: http.ProxyFromEnvironment,\n\t\tHandshakeTimeout: 45 * time.Second,\n\t}\n\n\tfor _, s := range setters {\n\t\tif err := s(f); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tif !f.stream {\n\t\tf.flushInterval = 0\n\t} else if f.flushInterval == 0 {\n\t\tf.flushInterval = defaultFlushInterval\n\t}\n\n\tif f.httpForwarder.rewriter == nil {\n\t\th, err := os.Hostname()\n\t\tif err != nil {\n\t\t\th = \"localhost\"\n\t\t}\n\t\tf.httpForwarder.rewriter = &HeaderRewriter{TrustForwardHeader: true, Hostname: h}\n\t}\n\n\tif f.httpForwarder.roundTripper == nil {\n\t\tf.httpForwarder.roundTripper = http.DefaultTransport\n\t}\n\n\tif f.errHandler == nil {\n\t\tf.errHandler = utils.DefaultHandler\n\t}\n\n\tif f.tlsClientConfig == nil {\n\t\tif ht, ok := f.httpForwarder.roundTripper.(*http.Transport); ok {\n\t\t\tf.tlsClientConfig = ht.TLSClientConfig\n\t\t\tif f.websocketDialer.TLSClientConfig == nil && ht.TLSClientConfig != nil {\n\t\t\t\t_ = WebsocketTLSClientConfig(ht.TLSClientConfig)(f)\n\t\t\t}\n\t\t}\n\t}\n\n\tf.postConfig()\n\n\treturn f, nil\n}", "func NewForwarder(s *stack.Stack, rcvWnd, maxInFlight int, handler func(*ForwarderRequest)) *Forwarder {\n\tif rcvWnd == 0 {\n\t\trcvWnd = DefaultReceiveBufferSize\n\t}\n\treturn &Forwarder{\n\t\tstack: s,\n\t\tmaxInFlight: maxInFlight,\n\t\thandler: handler,\n\t\tinFlight: make(map[stack.TransportEndpointID]struct{}),\n\t\tlisten: newListenContext(s, protocolFromStack(s), nil /* listenEP */, seqnum.Size(rcvWnd), true, 0),\n\t}\n}", "func (mr *MockFirewallServiceIfaceMockRecorder) NewCreatePortForwardingRuleParams(ipaddressid, privateport, protocol, publicport, virtualmachineid interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"NewCreatePortForwardingRuleParams\", reflect.TypeOf((*MockFirewallServiceIface)(nil).NewCreatePortForwardingRuleParams), ipaddressid, privateport, protocol, publicport, virtualmachineid)\n}", "func ForwardToSubnet(in, out Link, dst Addr) rules.Rule {\n\treturn rules.Rule(fmt.Sprintf(\n\t\t\"-t filter -A fw-interfaces -j ACCEPT -d %v -i %v -o %v\",\n\t\tdst, in.Name(), out.Name()))\n}", "func (s *Service) CreateNotificationRule(ctx context.Context, nr influxdb.NotificationRuleCreate, userID influxdb.ID) error {\n\treturn s.kv.Update(ctx, func(tx Tx) error {\n\t\treturn s.createNotificationRule(ctx, tx, nr, userID)\n\t})\n}", "func (ruleset *DnsForwardingRuleset) createValidations() []func() (admission.Warnings, error) {\n\treturn []func() (admission.Warnings, error){ruleset.validateResourceReferences}\n}", "func (fw *IPtables) CreateRules(chain int) error {\n\tlog.Info(\"In CreateRules() for chain\", chain)\n\tfor _, rule := range fw.chains[chain].Rules {\n\t\t// First create rule record in database.\n\t\terr0 := fw.addIPtablesRule(rule)\n\t\tif err0 != nil {\n\t\t\tlog.Error(\"In CreateRules() create db record for iptables rule \", rule.GetBody())\n\t\t\treturn err0\n\t\t}\n\n\t\terr1 := fw.EnsureRule(rule, EnsureFirst)\n\t\tif err1 != nil {\n\t\t\tlog.Error(\"In CreateRules() failed to create install firewall rule \", rule.GetBody())\n\t\t\treturn err1\n\t\t}\n\n\t\t// Finally, set 'active' flag in database record.\n\t\tif err2 := fw.Store.switchIPtablesRule(rule, setRuleActive); err2 != nil {\n\t\t\tlog.Error(\"In CreateRules() iptables rule created, but activation failed \", rule.GetBody())\n\t\t\treturn err2\n\t\t}\n\t}\n\tlog.Info(\"Creating firewall rules success\")\n\treturn nil\n}", "func NewPortForwarder(namespace string, labels metav1.LabelSelector, port int, opts ...ResourceForwardOption) (*PortForward, error) {\n\tpf := &PortForward{\n\t\tNamespace: namespace,\n\t\tLabels: labels,\n\t\tDestinationPort: port,\n\t\tresType: serviceType,\n\t}\n\n\tfor _, o := range opts {\n\t\tpf = o(pf)\n\t}\n\n\tvar err error\n\tpf.Config, err = clientcmd.NewNonInteractiveDeferredLoadingClientConfig(\n\t\tclientcmd.NewDefaultClientConfigLoadingRules(),\n\t\t&clientcmd.ConfigOverrides{},\n\t).ClientConfig()\n\tif err != nil {\n\t\treturn pf, fmt.Errorf(\"could not load kubernetes configuration file: %w\", err)\n\t}\n\n\tpf.Clientset, err = kubernetes.NewForConfig(pf.Config)\n\tif err != nil {\n\t\treturn pf, fmt.Errorf(\"could not create kubernetes client: %w\", err)\n\t}\n\n\treturn pf, nil\n}", "func (m *MockFirewallServiceIface) NewListPortForwardingRulesParams() *ListPortForwardingRulesParams {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"NewListPortForwardingRulesParams\")\n\tret0, _ := ret[0].(*ListPortForwardingRulesParams)\n\treturn ret0\n}", "func (a *BackendOptionsApiService) CreateTCPResponseRule(ctx _context.Context, backend string, tcpResponseRule TcpResponseRule, localVarOptionals *CreateTCPResponseRuleOpts) (TcpResponseRule, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodPost\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue TcpResponseRule\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/services/haproxy/configuration/tcp_response_rules\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tlocalVarQueryParams.Add(\"backend\", parameterToString(backend, \"\"))\n\tif localVarOptionals != nil && localVarOptionals.TransactionId.IsSet() {\n\t\tlocalVarQueryParams.Add(\"transaction_id\", parameterToString(localVarOptionals.TransactionId.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Version.IsSet() {\n\t\tlocalVarQueryParams.Add(\"version\", parameterToString(localVarOptionals.Version.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.ForceReload.IsSet() {\n\t\tlocalVarQueryParams.Add(\"force_reload\", parameterToString(localVarOptionals.ForceReload.Value(), \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{\"application/json\"}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &tcpResponseRule\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 201 {\n\t\t\tvar v TcpResponseRule\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 202 {\n\t\t\tvar v TcpResponseRule\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 400 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 409 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tvar v ModelError\n\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\tif err != nil {\n\t\t\tnewErr.error = err.Error()\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tnewErr.model = v\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func (pf *PortForwarder) ReverseForwardPort(ctx context.Context, req *waterfall_grpc_pb.PortForwardRequest) (*empty_pb.Empty, error) {\n\tpf.reverseSessionsMutex.Lock()\n\tdefer pf.reverseSessionsMutex.Unlock()\n\n\tlog.Printf(\"Reverse forwarding %s -> %s ...\\n\", req.Session.Src, req.Session.Dst)\n\tif rs, ok := pf.reverseSessions[req.Session.Src]; ok {\n\t\tif !req.Rebind {\n\t\t\treturn nil, status.Errorf(codes.AlreadyExists, \"no-rebind specified, can't forward address: %s\", req.Session.Src)\n\t\t}\n\t\tif err := pf.stopReverseForwarding(ctx, rs); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tsrcKind, srcAddr, err := parseAddr(req.Session.Src)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdstKind, dstAddr, err := parseAddr(req.Session.Dst)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tsrcNtwk, err := networkKind(srcKind)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// The context we create in this case is scoped to the duration of the forwarding\n\t// session, which outlives this request, therefore we can't propagate the request\n\t// context and are forced to create a new one.\n\tfCtx, cancel := context.WithCancel(context.Background())\n\n\t// 1) Ask the server to listen for connections on src\n\tncs, err := pf.client.StartReverseForward(\n\t\tfCtx,\n\t\t&waterfall_grpc_pb.ForwardMessage{\n\t\t\tOp: waterfall_grpc_pb.ForwardMessage_OPEN,\n\t\t\tKind: srcNtwk,\n\t\t\tAddr: srcAddr})\n\tif err != nil {\n\t\tlog.Printf(\"Failed to start reverse forwarding session (%s -> %s): %v\", req.Session.Src, req.Session.Dst, err)\n\t\treturn nil, err\n\t}\n\n\tss := &forwardSession{src: req.Session.Src, dst: req.Session.Dst, cancel: cancel}\n\tpf.reverseSessions[req.Session.Src] = ss\n\n\t// Listen for new connections on a different goroutine so we can return to the client\n\tgo func() {\n\t\tdefer pf.stopReverseForwarding(fCtx, ss)\n\t\tfor {\n\t\t\tlog.Print(\"Waiting for new connection to forward...\")\n\t\t\tfwd, err := ncs.Recv()\n\t\t\tif err != nil {\n\t\t\t\tlog.Printf(\"Waterfall server error when listening for reverse forwarding connections: %v\", err)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tif fwd.Op != waterfall_grpc_pb.ForwardMessage_OPEN {\n\t\t\t\t// The only type of message the server can reply is with an OPEN message\n\t\t\t\tlog.Printf(\"Requested OP %v but only open is supported ...\\n\", fwd.Op)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\t// 2) Hand the server a stream to start forwarding the connection\n\t\t\tfs, err := pf.client.ReverseForward(fCtx)\n\t\t\tif err != nil {\n\t\t\t\tlog.Printf(\"Failed to create new forwarding session: %v\", err)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tntwk := srcNtwk\n\t\t\taddr := srcAddr\n\t\t\tif fwd.GetKind() != waterfall_grpc_pb.ForwardMessage_UNSET && len(fwd.Addr) > 0 {\n\t\t\t\tntwk = fwd.Kind\n\t\t\t\taddr = fwd.Addr\n\t\t\t}\n\t\t\tif err := fs.Send(&waterfall_grpc_pb.ForwardMessage{\n\t\t\t\tOp: waterfall_grpc_pb.ForwardMessage_OPEN,\n\t\t\t\tKind: ntwk,\n\t\t\t\tAddr: addr,\n\t\t\t}); err != nil {\n\t\t\t\tlog.Printf(\"Failed to create new forwarding request: %v\", err)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tconn, err := net.Dial(dstKind, dstAddr)\n\t\t\tif err != nil {\n\t\t\t\t// Ignore this error. The socket might not be open initially\n\t\t\t\t// but can be created after the forwarding session.\n\t\t\t\tlog.Printf(\"Failed to connect %s:%s: %v\", dstKind, dstAddr, err)\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// 3) Forward the stream to the connection\n\t\t\tgo forward.NewStreamForwarder(fs, conn.(forward.HalfReadWriteCloser)).Forward()\n\t\t}\n\t}()\n\treturn &empty_pb.Empty{}, nil\n}", "func (p *Proxy) CreateOrUpdateRedirect(l4 *policy.L4Filter, id string, source ProxySource, kind string) (Redirect, error) {\n\tgcOnce.Do(func() {\n\t\tif lf := viper.GetString(\"access-log\"); lf != \"\" {\n\t\t\tif err := accesslog.OpenLogfile(lf); err != nil {\n\t\t\t\tlog.WithError(err).WithField(accesslog.FieldFilePath, lf).\n\t\t\t\t\tWarn(\"Cannot open L7 access log\")\n\t\t\t}\n\t\t}\n\n\t\tif labels := viper.GetStringSlice(\"agent-labels\"); len(labels) != 0 {\n\t\t\taccesslog.SetMetadata(labels)\n\t\t}\n\n\t\tgo func() {\n\t\t\tfor {\n\t\t\t\ttime.Sleep(time.Duration(10) * time.Second)\n\t\t\t\tif deleted := GC(); deleted > 0 {\n\t\t\t\t\tlog.WithField(\"count\", deleted).\n\t\t\t\t\t\tDebug(\"Evicted entries from proxy table\")\n\t\t\t\t}\n\t\t\t}\n\t\t}()\n\t})\n\n\tp.mutex.Lock()\n\tdefer p.mutex.Unlock()\n\n\tscopedLog := log.WithField(fieldProxyRedirectID, id)\n\n\tif r, ok := p.redirects[id]; ok {\n\t\terr := r.UpdateRules(l4)\n\t\tif err != nil {\n\t\t\tscopedLog.WithError(err).Error(\"Unable to update \", l4.L7Parser, \" proxy\")\n\t\t\treturn nil, err\n\t\t}\n\t\tscopedLog.WithField(logfields.Object, logfields.Repr(r)).\n\t\t\tDebug(\"updated existing \", l4.L7Parser, \" proxy instance\")\n\t\treturn r, nil\n\t}\n\n\tscopedLog = scopedLog.WithField(fieldProxyKind, kind)\n\n\tto, err := p.allocatePort()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar redir Redirect\n\n\tswitch l4.L7Parser {\n\tcase policy.ParserTypeKafka:\n\t\tredir, err = createKafkaRedirect(kafkaConfiguration{\n\t\t\tpolicy: l4,\n\t\t\tid: id,\n\t\t\tsource: source,\n\t\t\tlistenPort: to})\n\tcase policy.ParserTypeHTTP:\n\t\tswitch kind {\n\t\tcase ProxyKindOxy:\n\t\t\tredir, err = createOxyRedirect(l4, id, source, to)\n\t\tcase ProxyKindEnvoy:\n\t\t\tredir, err = createEnvoyRedirect(l4, id, source, to)\n\t\tdefault:\n\t\t\treturn nil, fmt.Errorf(\"Unknown proxy kind: %s\", kind)\n\t\t}\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"Unsupported L7 parser type: %s\", l4.L7Parser)\n\t}\n\tif err != nil {\n\t\tscopedLog.WithError(err).Error(\"Unable to create \", l4.L7Parser, \" proxy\")\n\t\treturn nil, err\n\t}\n\tscopedLog.WithField(logfields.Object, logfields.Repr(redir)).\n\t\tDebug(\"Created new \", l4.L7Parser, \" proxy instance\")\n\n\tp.allocatedPorts[to] = redir\n\tp.redirects[id] = redir\n\n\treturn redir, nil\n}", "func NewPortForwarder(f Factory) *PortForwarder {\n\treturn &PortForwarder{\n\t\tFactory: f,\n\t\tstopChan: make(chan struct{}),\n\t\treadyChan: make(chan struct{}),\n\t}\n}", "func grokRuleDefinition(fset *token.FileSet, astFile *ast.File, newAstFile *ast.File,\n\tdecl ast.Decl, info *types.Info) *RuleSpec {\n\t// body.\n\tfd, ok := decl.(*ast.FuncDecl)\n\tif !ok {\n\t\treturn (*RuleSpec)(nil)\n\t}\n\tif !strings.HasPrefix(fd.Name.Name, ruleNamePrefix) {\n\t\treturn (*RuleSpec)(nil)\n\t}\n\t// Verify that the first parameter is type rete.Node:\n\t// *** MAYBE WE NEED A BETTER TEST HERE\n\ttype_string := typeString(info.TypeOf(fd.Type.Params.List[0].Names[0]))\n\tif type_string != NODE_TYPE_STRING {\n\t\treturn (*RuleSpec)(nil)\n\t}\n\trule_name := ruleBaseName(fd.Name.Name)\n\tspec := &RuleSpec{\n\t\tPackage: astFile.Name.Name,\n\t\tRuleName: rule_name,\n\t\tRuleFunctionName: fd.Name.Name,\n\t\tRuleCallerName: \"call_\" + rule_name,\n\t\tRuleInstallerName: \"install_\" + rule_name,\n\t\tRuleParameters: []*ruleParameter{},\n\t\tRuleEmits: []string{},\n\t}\n\t// Determine parameter types.\n\t// We assume here that the zeroth element of the params list, since\n\t// it's a rete.Node, has only one identifier.\n\tfor _, field := range fd.Type.Params.List[1:] {\n\t\tfor _, nameId := range field.Names {\n\n\t\t\t// types.TypeString with a nil Qualifier will always write\n\t\t\t// the package path.\n\t\t\tpt := types.TypeString(info.TypeOf(nameId), package_qualifier(astFile))\n\t\t\tname := nameId.Name\n\t\t\tspec.RuleParameters = append(spec.RuleParameters, &ruleParameter{\n\t\t\t\tName: name,\n\t\t\t\tParamType: pt,\n\t\t\t})\n\t\t}\n\t}\n\t// All rule parameters have been determined. Fill in\n\t// ruleParameter.Remaining for the convenience of the\n\t// code generation template.\n\tspec.fillRemaining()\n\t// Determine rule output types by finding calls to rete.Node.Emit\n\tast.Inspect(decl, func(n ast.Node) bool {\n\t\tif n == nil {\n\t\t\treturn false\n\t\t}\n\t\tc, ok := n.(*ast.CallExpr)\n\t\tif !ok {\n\t\t\treturn true\n\t\t}\n\t\tsel, ok := c.Fun.(*ast.SelectorExpr)\n\t\tif !ok {\n\t\t\treturn true\n\t\t}\n\t\tif receiver, ok := sel.X.(*ast.Ident); !ok ||\n\t\t\ttypeString(info.TypeOf(receiver)) != NODE_TYPE_STRING {\n\t\t\treturn true\n\t\t}\n\t\tif sel.Sel.Name == \"Emit\" {\n\t\t\tspec.RuleEmits = adjoin(spec.RuleEmits,\n\t\t\t\ttypes.TypeString(info.TypeOf(c.Args[0]),\n\t\t\t\t\tpackage_qualifier(astFile)))\n\t\t\treturn false\n\t\t}\n\t\treturn true\n\t})\n\tif len(spec.RuleEmits) == 0 {\n\t\tfmt.Fprintf(os.Stderr, \" Rule %s doesn't call Emit.\\n\", rule_name)\n\t}\n\treturn spec\n}", "func NewPortForwarder(client Client, pod *corev1.Pod, ports []string, addresses []string, stopChan chan struct{}, readyChan chan struct{}, errorChan chan error) (*portforward.PortForwarder, error) {\n\texecRequest := client.KubeClient().CoreV1().RESTClient().Post().\n\t\tResource(\"pods\").\n\t\tName(pod.Name).\n\t\tNamespace(pod.Namespace).\n\t\tSubResource(\"portforward\")\n\n\ttransport, upgrader, err := GetUpgraderWrapper(client)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlogFile := log.GetFileLogger(\"portforwarding\")\n\tdialer := spdy.NewDialer(upgrader, &http.Client{Transport: transport}, \"POST\", execRequest.URL())\n\n\tfw, err := portforward.NewOnAddresses(dialer, addresses, ports, stopChan, readyChan, errorChan, logFile.Writer(logrus.InfoLevel, false), logFile.Writer(logrus.WarnLevel, false))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn fw, nil\n}", "func (g *Grammar) CreateRule(name string, terms ...Term) {\n\trule := g.Lookup(name)\n\tprod := Production{}\n\tfor _, t := range terms {\n\t\tprod = append(prod, t)\n\t}\n\trule.Productions = append(rule.Productions, prod)\n\tg.rules[name] = rule\n}", "func (fw IPtables) EnsureRule(rule FirewallRule, opType RuleState) error {\n\truleExists := fw.isRuleExist(rule)\n\n\targs := []string{}\n\tif ruleExists {\n\n\t\tswitch opType {\n\t\tcase EnsureAbsent:\n\t\t\targs = append(args, \"-D\")\n\t\tdefault:\n\t\t\tlog.Info(\"In EnsureRule - nothing to do \", rule.GetBody())\n\t\t\treturn nil\n\t\t}\n\t} else {\n\n\t\tswitch opType {\n\t\tcase EnsureLast:\n\t\t\targs = append(args, \"-A\")\n\t\tcase EnsureFirst:\n\t\t\targs = append(args, \"-I\")\n\t\tdefault:\n\t\t\tlog.Info(\"In EnsureRule - nothing to do \", rule.GetBody())\n\t\t\treturn nil\n\t\t}\n\t}\n\n\targs = append(args, strings.Split(rule.GetBody(), \" \")...)\n\tcmdStr := iptablesCmd + \" \" + strings.Join(args, \" \")\n\tout, err := fw.os.Exec(iptablesCmd, args)\n\tif err != nil {\n\t\tlog.Errorf(\"[%s]: %s failed for rule %s with error %v, saying [%s]\", cmdStr, opType, rule.GetBody(), err, string(out))\n\t} else {\n\t\tif out != nil && len(out) > 0 {\n\t\t\tlog.Infof(\"%s success %s: [%s]\", opType, rule.GetBody(), string(out))\n\t\t} else {\n\t\t\tlog.Infof(\"%s success %s\", opType, rule.GetBody())\n\t\t}\n\t}\n\n\treturn err\n}", "func createRemoteProxy(proxy *NTLMProxy, pURL, protocol, origin string) http.Handler {\r\n\tgenConn := func() (net.Conn, error) {\r\n\t\t//conn, err := ProxyDial(pURL, \"\", origin)\r\n\t\tconn, err := proxy.Websocket(pURL, \"\", origin)\r\n\t\tif err != nil {\r\n\t\t\treturn nil, err\r\n\t\t}\r\n\r\n\t\treturn fetch.NewClientConn(conn.UnderlyingConn(), 0x56), nil\r\n\t}\r\n\tgenConn = logConnect(genConn)\r\n\treturn Tunnel(genConn)\r\n}", "func CreateJumpRule(v, tableName, srcChainName, dstChainName string) error {\n\tif err := isSupportedIPVersion(v); err != nil {\n\t\treturn err\n\t}\n\n\tconn, err := initNftConn()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttb := &nftables.Table{\n\t\tName: tableName,\n\t}\n\tif v == \"4\" {\n\t\ttb.Family = nftables.TableFamilyIPv4\n\t} else {\n\t\ttb.Family = nftables.TableFamilyIPv6\n\t}\n\n\tch := &nftables.Chain{\n\t\tName: srcChainName,\n\t\tTable: tb,\n\t}\n\n\tchainProps, err := GetChainProps(v, tableName, srcChainName)\n\tif err != nil {\n\t\treturn fmt.Errorf(\n\t\t\t\"failed adding jump rule from chain %s in ipv%s %s table to chain %s due to failure to list chains: %s\",\n\t\t\tsrcChainName, v, tableName, dstChainName, err,\n\t\t)\n\t}\n\n\tr := &nftables.Rule{\n\t\tTable: tb,\n\t\tChain: ch,\n\t\tExprs: []expr.Any{},\n\t}\n\n\tif chainProps.RuleCount > 0 {\n\t\tr.Position = chainProps.Positions[0]\n\t}\n\n\tr.Exprs = append(r.Exprs, &expr.Verdict{\n\t\tKind: expr.VerdictJump,\n\t\tChain: dstChainName,\n\t})\n\n\tif chainProps.RuleCount == 0 {\n\t\tconn.AddRule(r)\n\t} else {\n\t\tconn.InsertRule(r)\n\t}\n\tif err := conn.Flush(); err != nil {\n\t\treturn fmt.Errorf(\n\t\t\t\"failed adding jump rule from chain %s in ipv%s table to chain %s: %s\",\n\t\t\tsrcChainName, v, tableName, dstChainName, err,\n\t\t)\n\t}\n\n\treturn nil\n}", "func (m *MockFirewallServiceIface) DeletePortForwardingRule(p *DeletePortForwardingRuleParams) (*DeletePortForwardingRuleResponse, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"DeletePortForwardingRule\", p)\n\tret0, _ := ret[0].(*DeletePortForwardingRuleResponse)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func StartPortForwarding(ctx devspacecontext.Context, devPod *latest.DevPod, selector targetselector.TargetSelector, parent *tomb.Tomb) (retErr error) {\n\tif ctx == nil || ctx.Config() == nil || ctx.Config().Config() == nil {\n\t\treturn fmt.Errorf(\"DevSpace config is not set\")\n\t}\n\n\t// forward\n\tinitDoneArray := []chan struct{}{}\n\tif len(devPod.Ports) > 0 {\n\t\tinitDoneArray = append(initDoneArray, parent.NotifyGo(func() error {\n\t\t\treturn startPortForwardingWithHooks(ctx, devPod.Name, devPod.Ports, selector, parent)\n\t\t}))\n\t}\n\n\t// reverse\n\tloader.EachDevContainer(devPod, func(devContainer *latest.DevContainer) bool {\n\t\tif len(devContainer.ReversePorts) > 0 {\n\t\t\tinitDoneArray = append(initDoneArray, parent.NotifyGo(func() error {\n\t\t\t\treturn startReversePortForwardingWithHooks(ctx, devPod.Name, string(devContainer.Arch), devContainer.ReversePorts, selector.WithContainer(devContainer.Container), parent)\n\t\t\t}))\n\t\t}\n\t\treturn true\n\t})\n\n\t// wait until everything is initialized\n\tfor _, initDone := range initDoneArray {\n\t\t<-initDone\n\t}\n\treturn nil\n}", "func (f *Framework) CreateProxierOperator(namespace string, operatorImage string, namespacesToWatch []string) error {\n\tcrd, err := MakeCRD(\"../../deploy/crds/maegus_v1beta1_proxier_crd.yaml\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err := CreateCRD(f.ApiextensionsClientV1, namespace, crd); err != nil && !apierrors.IsAlreadyExists(err) {\n\t\treturn errors.Wrap(err, \"failed to create proxier crd\")\n\t}\n\n\t_, err = CreateServiceAccount(f.KubeClient, namespace, \"../../deploy/service_account.yaml\")\n\tif err != nil && !apierrors.IsAlreadyExists(err) {\n\t\treturn errors.Wrap(err, \"failed to create proxier operator service account\")\n\t}\n\n\tif err := CreateClusterRole(f.KubeClient, \"../../deploy/cluster_role.yaml\"); err != nil && !apierrors.IsAlreadyExists(err) {\n\t\treturn errors.Wrap(err, \"failed to create proxier cluster role\")\n\t}\n\n\tif _, err := CreateClusterRoleBinding(f.KubeClient, namespace, \"../../deploy/cluster_role_binding.yaml\"); err != nil && !apierrors.IsAlreadyExists(err) {\n\t\treturn errors.Wrap(err, \"failed to create prometheus cluster role binding\")\n\t}\n\n\tdeployment, err := MakeDeployment(\"../../deploy/operator.yaml\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif operatorImage != \"\" {\n\t\t// Override operator image used, if specified when running tests.\n\t\tdeployment.Spec.Template.Spec.Containers[0].Image = operatorImage\n\t\trepoAndTag := strings.Split(operatorImage, \":\")\n\t\tif len(repoAndTag) != 2 {\n\t\t\treturn errors.Errorf(\n\t\t\t\t\"expected operator image '%v' split by colon to result in two substrings but got '%v'\",\n\t\t\t\toperatorImage,\n\t\t\t\trepoAndTag,\n\t\t\t)\n\t\t}\n\t}\n\n\terr = CreateDeployment(f.KubeClient, namespace, deployment)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\topts := metav1.ListOptions{LabelSelector: fields.SelectorFromSet(fields.Set(deployment.Spec.Template.ObjectMeta.Labels)).String()}\n\terr = WaitForPodsReady(f.KubeClient, namespace, f.DefaultTimeout, 1, opts)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to wait for prometheus operator to become ready\")\n\t}\n\n\terr = WaitForCRDReady(func(opts metav1.ListOptions) (runtime.Object, error) {\n\t\treturn f.MaegusClientV1.Proxiers(v1.NamespaceAll).List(opts)\n\t})\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"Proxier CRD not ready: %v\\n\")\n\t}\n\n\treturn nil\n}", "func (a *BackendOptionsApiService) CreateTCPRequestRule(ctx _context.Context, parentName string, parentType string, tcpRequestRule TcpRequestRule, localVarOptionals *CreateTCPRequestRuleOpts) (TcpRequestRule, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodPost\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue TcpRequestRule\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/services/haproxy/configuration/tcp_request_rules\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tlocalVarQueryParams.Add(\"parent_name\", parameterToString(parentName, \"\"))\n\tlocalVarQueryParams.Add(\"parent_type\", parameterToString(parentType, \"\"))\n\tif localVarOptionals != nil && localVarOptionals.TransactionId.IsSet() {\n\t\tlocalVarQueryParams.Add(\"transaction_id\", parameterToString(localVarOptionals.TransactionId.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Version.IsSet() {\n\t\tlocalVarQueryParams.Add(\"version\", parameterToString(localVarOptionals.Version.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.ForceReload.IsSet() {\n\t\tlocalVarQueryParams.Add(\"force_reload\", parameterToString(localVarOptionals.ForceReload.Value(), \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{\"application/json\"}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &tcpRequestRule\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 201 {\n\t\t\tvar v TcpRequestRule\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 202 {\n\t\t\tvar v TcpRequestRule\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 400 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 409 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tvar v ModelError\n\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\tif err != nil {\n\t\t\tnewErr.error = err.Error()\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tnewErr.model = v\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) PutFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRules(ctx context.Context, name string, policyId string, frinxOpenconfigPolicyForwardingPfforwardingpolicystructuralPoliciesPolicyRulesBodyParam FrinxOpenconfigPolicyForwardingPfforwardingpolicystructuralPoliciesPolicyRulesRequest, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Put\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &frinxOpenconfigPolicyForwardingPfforwardingpolicystructuralPoliciesPolicyRulesBodyParam\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func NewPortForwarder(namespace string, labels metav1.LabelSelector, port int) (*PortForward, error) {\n\tpf := &PortForward{\n\t\tNamespace: namespace,\n\t\tLabels: labels,\n\t\tDestinationPort: port,\n\t}\n\n\tvar err error\n\tpf.Config, err = clientcmd.NewNonInteractiveDeferredLoadingClientConfig(\n\t\tclientcmd.NewDefaultClientConfigLoadingRules(),\n\t\t&clientcmd.ConfigOverrides{},\n\t).ClientConfig()\n\tif err != nil {\n\t\treturn pf, errors.Wrap(err, \"Could not load kubernetes configuration file\")\n\t}\n\n\tpf.Clientset, err = kubernetes.NewForConfig(pf.Config)\n\tif err != nil {\n\t\treturn pf, errors.Wrap(err, \"Could not create kubernetes client\")\n\t}\n\n\treturn pf, nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) PutFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRulesRuleTransport(ctx context.Context, name string, policyId string, sequenceId int64, frinxOpenconfigPacketMatchTransportfieldstopTransportBodyParam FrinxOpenconfigPacketMatchTransportfieldstopTransportRequest, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Put\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/frinx-openconfig-network-instance:rule/{sequence-id}/frinx-openconfig-network-instance:transport/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"sequence-id\"+\"}\", fmt.Sprintf(\"%v\", sequenceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &frinxOpenconfigPacketMatchTransportfieldstopTransportBodyParam\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (r *FirewallGlobalRulesStagedPolicyRulesResource) Create(item FirewallGlobalRulesStagedPolicyRulesConfig) error {\n\tif err := r.c.ModQuery(\"POST\", BasePath+FirewallGlobalRulesStagedPolicyRulesEndpoint, item); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (hd *Datapath) CreateTCPProxyPolicy(tcp *netproto.TCPProxyPolicy, vrf *netproto.Vrf) error {\n\t// This will ensure that only one datapath config will be active at a time. This is a temporary restriction\n\t// to ensure that HAL will use a single config thread , this will be removed prior to FCS to allow parallel configs to go through.\n\t// TODO Remove Global Locking\n\thd.Lock()\n\tdefer hd.Unlock()\n\tvrfKey := &halproto.VrfKeyHandle{\n\t\tKeyOrHandle: &halproto.VrfKeyHandle_VrfId{\n\t\t\tVrfId: vrf.Status.VrfID,\n\t\t},\n\t}\n\n\tvar tcpProxyRules []*halproto.TcpProxyRuleMatchSpec\n\n\tfor _, r := range tcp.Spec.Rules {\n\t\t// Match source and dest attributes\n\t\truleMatches, err := hd.buildHALRuleMatches(r.Src, r.Dst, nil, nil)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Could not convert match criteria Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\tfor _, match := range ruleMatches {\n\t\t\trule := &halproto.TcpProxyRuleMatchSpec{\n\t\t\t\tRuleId: r.ID,\n\t\t\t\tMatch: match,\n\t\t\t\tTcpProxyAction: convertProxyAction(r.Action),\n\t\t\t}\n\t\t\ttcpProxyRules = append(tcpProxyRules, rule)\n\t\t}\n\n\t}\n\n\ttcpProxyReqMsg := &halproto.TcpProxyRuleRequestMsg{\n\t\tRequest: []*halproto.TcpProxyRuleSpec{\n\t\t\t{\n\t\t\t\tKeyOrHandle: &halproto.TcpProxyRuleKeyHandle{\n\t\t\t\t\tKeyOrHandle: &halproto.TcpProxyRuleKeyHandle_RuleKey{\n\t\t\t\t\t\tRuleKey: &halproto.TcpProxyRuleKey{\n\t\t\t\t\t\t\tTcpProxyRuleId: tcp.Status.TCPProxyPolicyID,\n\t\t\t\t\t\t\tVrfKeyOrHandle: vrfKey,\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tVrfKeyHandle: vrfKey,\n\t\t\t\tRules: tcpProxyRules,\n\t\t\t},\n\t\t},\n\t}\n\n\tif hd.Kind == \"hal\" {\n\t\tresp, err := hd.Hal.TCPProxyPolicyClient.TcpProxyRuleCreate(context.Background(), tcpProxyReqMsg)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error creating TCPProxy Policy. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\t\tif !(resp.Response[0].ApiStatus == halproto.ApiStatus_API_STATUS_OK || resp.Response[0].ApiStatus == halproto.ApiStatus_API_STATUS_EXISTS_ALREADY) {\n\t\t\tlog.Errorf(\"HAL returned non OK status. %v\", resp.Response[0].ApiStatus.String())\n\t\t\treturn fmt.Errorf(\"HAL returned non OK status. %v\", resp.Response[0].ApiStatus.String())\n\t\t}\n\t} else {\n\t\t_, err := hd.Hal.TCPProxyPolicyClient.TcpProxyRuleCreate(context.Background(), tcpProxyReqMsg)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error creating TCPProxy Policy. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}" ]
[ "0.81244224", "0.7319604", "0.7065577", "0.6473114", "0.61948085", "0.617615", "0.5928932", "0.5825406", "0.57282436", "0.57121503", "0.56377894", "0.5556974", "0.5553113", "0.5518824", "0.54471165", "0.5443175", "0.5431345", "0.53823966", "0.5333005", "0.5328435", "0.5290705", "0.5248748", "0.5147813", "0.51416165", "0.51194793", "0.51031995", "0.5093856", "0.5067378", "0.5036041", "0.5012663", "0.49915242", "0.49861267", "0.49771655", "0.4943755", "0.4936129", "0.4929772", "0.49179572", "0.49057478", "0.48787457", "0.48623508", "0.48573673", "0.48372042", "0.4814461", "0.48052227", "0.47939208", "0.4775449", "0.47713968", "0.47580513", "0.4752748", "0.4746018", "0.47407877", "0.47348765", "0.47270492", "0.47257066", "0.47207016", "0.47160625", "0.4687468", "0.46870315", "0.467783", "0.46735027", "0.4670133", "0.46658802", "0.465848", "0.46576044", "0.46567735", "0.46566013", "0.4628921", "0.462705", "0.46065944", "0.45992336", "0.45973215", "0.45969346", "0.45963272", "0.45960242", "0.45950294", "0.459064", "0.45863876", "0.45842817", "0.45822072", "0.455494", "0.45335135", "0.4533056", "0.45283526", "0.45278656", "0.4520914", "0.45119607", "0.4507224", "0.4504502", "0.45034492", "0.44918013", "0.4490312", "0.448828", "0.4481875", "0.44812277", "0.4481032", "0.44803923", "0.446126", "0.44584534", "0.44563767", "0.44553405" ]
0.72475964
2
CreateFirewallRule uses the override method CreateFirewallRuleFn or the real implementation.
func (c *TestClient) CreateFirewallRule(project string, i *compute.Firewall) error { if c.CreateFirewallRuleFn != nil { return c.CreateFirewallRuleFn(project, i) } return c.client.CreateFirewallRule(project, i) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (g *Google) createFirewallRule() (string, error) {\n\tif rule, err := g.getFirewallRule(); err == nil {\n\t\tlog.Infof(\"found FirewallRule %s: %s\", firewallRuleName, rule.SelfLink)\n\t\treturn rule.SelfLink, nil\n\t}\n\n\top, err := g.computeService.Firewalls.Insert(g.project,\n\t\t&compute.Firewall{\n\t\t\tName: firewallRuleName,\n\t\t\tAllowed: []*compute.FirewallAllowed{\n\t\t\t\t{\n\t\t\t\t\tIPProtocol: cockroachProtocol,\n\t\t\t\t\tPorts: []string{\n\t\t\t\t\t\tfmt.Sprintf(\"%d\", g.context.Port),\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t\tSourceRanges: []string{\n\t\t\t\tallIPAddresses,\n\t\t\t},\n\t\t}).Do()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif err := g.waitForOperation(op); err != nil {\n\t\treturn \"\", err\n\t}\n\tlog.Infof(\"created FirewallRule %s: %s\", firewallRuleName, op.TargetLink)\n\treturn op.TargetLink, nil\n}", "func (f *fakeLB) CreateForwardingRule(_ context.Context, _ string, _ *govultr.ForwardingRule) (*govultr.ForwardingRule, *http.Response, error) {\n\tpanic(\"implement me\")\n}", "func NewFirewallRule(ctx *pulumi.Context,\n\tname string, args *FirewallRuleArgs, opts ...pulumi.ResourceOption) (*FirewallRule, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.EndIpAddress == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'EndIpAddress'\")\n\t}\n\tif args.ServerId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'ServerId'\")\n\t}\n\tif args.StartIpAddress == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'StartIpAddress'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource FirewallRule\n\terr := ctx.RegisterResource(\"azure:mssql/firewallRule:FirewallRule\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (s *LocalTests) createFirewallRule(c *gc.C) *cloudapi.FirewallRule {\n\tfwRule, err := s.testClient.CreateFirewallRule(cloudapi.CreateFwRuleOpts{Enabled: false, Rule: testFwRule})\n\tc.Assert(err, gc.IsNil)\n\tc.Assert(fwRule, gc.NotNil)\n\tc.Assert(fwRule.Rule, gc.Equals, testFwRule)\n\tc.Assert(fwRule.Enabled, gc.Equals, false)\n\ttime.Sleep(10 * time.Second)\n\n\treturn fwRule\n}", "func (g *Cloud) CreateFirewall(f *compute.Firewall) error {\n\tctx, cancel := cloud.ContextWithCallTimeout()\n\tdefer cancel()\n\n\tmc := newFirewallMetricContext(\"create\")\n\treturn mc.Observe(g.c.Firewalls().Insert(ctx, meta.GlobalKey(f.Name), f))\n}", "func (c *TestClient) CreateForwardingRule(project, region string, fr *compute.ForwardingRule) error {\n\tif c.CreateForwardingRuleFn != nil {\n\t\treturn c.CreateForwardingRuleFn(project, region, fr)\n\t}\n\treturn c.client.CreateForwardingRule(project, region, fr)\n}", "func CreateNetworkSecurityGroupRule() {}", "func (r *rulesRPCHandler) CreateRule(ctx context.Context, rule *api.Rule) (*api.Rule, error) {\n\tr.Lock()\n\tclient := clients[clientID%numClients]\n\tclientID++\n\tr.Unlock()\n\tif err := client.Create(ctx, path.Join(\"/rules\", rule.Name), rule); err != nil {\n\t\treturn nil, err\n\t}\n\treturn rule, nil\n}", "func (r *GenericRule) createRule(resource *Resource, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(resource)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(resource)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(resource)\n\t\t},\n\t}\n\treturn rule\n}", "func (r *V1Beta1ExtensionsIngressRule) createRule(ingress *v1beta1Extensions.Ingress, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(ingress)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(ingress)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(ingress)\n\t\t},\n\t}\n\treturn rule\n}", "func (f *fakeLB) GetFirewallRule(_ context.Context, _, _ string) (*govultr.LBFirewallRule, *http.Response, error) {\n\treturn nil, nil, nil\n}", "func (r *V1Beta1ExtensionsDeploymentRule) createRule(deployment *v1beta1Extensions.Deployment, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(deployment)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(deployment)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(deployment)\n\t\t},\n\t}\n\treturn rule\n}", "func (d *AppsV1DeploymentRule) createRule(deployment *appsv1.Deployment, ydr *YamlDerivedResource) *rule {\n\tr := &rule{\n\t\tID: d.ID,\n\t\tPrereqs: d.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif d.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn d.Condition(deployment)\n\t\t},\n\t\tMessage: d.Message,\n\t\tLevel: d.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif d.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn d.Fix(deployment)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif d.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn d.FixDescription(deployment)\n\t\t},\n\t}\n\treturn r\n}", "func (r *V1ServiceRule) createRule(service *v1.Service, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(service)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(service)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(service)\n\t\t},\n\t}\n\treturn rule\n}", "func CreateRule(fileName string, namingRules []NamingRule, defaultRuleKind string) *bazel.Rule {\n\tkind := defaultRuleKind\n\tfor _, r := range namingRules {\n\t\tm := r.FileNameMatcher.FindStringSubmatch(fileName)\n\t\tif m != nil {\n\t\t\tkind = r.RuleKind\n\t\t\tbreak\n\t\t}\n\t}\n\tpkgName := filepath.Dir(fileName)\n\tif pkgName == \".\" {\n\t\tpkgName = \"\"\n\t}\n\tsrc := filepath.Base(fileName)\n\tname := strings.TrimSuffix(filepath.Base(fileName), filepath.Ext(fileName))\n\treturn bazel.NewRule(kind, pkgName, name, map[string]interface{}{\"srcs\": []string{src}})\n}", "func (r *V1NamespaceRule) createRule(namespace *v1.Namespace, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(namespace)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(namespace)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(namespace)\n\t\t},\n\t}\n\treturn rule\n}", "func (g *Google) createForwardingRule(targetLink string) (string, error) {\n\tif rule, err := g.getForwardingRule(); err == nil {\n\t\tlog.Infof(\"found ForwardingRule %s: %s\", forwardingRuleName, rule.SelfLink)\n\t\treturn rule.SelfLink, nil\n\t}\n\n\top, err := g.computeService.GlobalForwardingRules.Insert(g.project,\n\t\t&compute.ForwardingRule{\n\t\t\tName: forwardingRuleName,\n\t\t\tIPProtocol: cockroachProtocol,\n\t\t\tPortRange: fmt.Sprintf(\"%d\", g.context.Port),\n\t\t\tTarget: targetLink,\n\t\t}).Do()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif err := g.waitForOperation(op); err != nil {\n\t\treturn \"\", err\n\t}\n\n\tlog.Infof(\"created ForwardingRule %s: %s\", forwardingRuleName, op.TargetLink)\n\treturn op.TargetLink, nil\n}", "func (r *V1ServiceAccountRule) createRule(serviceaccount *v1.ServiceAccount, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(serviceaccount)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(serviceaccount)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(serviceaccount)\n\t\t},\n\t}\n\treturn rule\n}", "func (p *fleetUpdate) CreateRules() ([]string, error) {\n\t// No rules here\n\treturn nil, nil\n}", "func (r *NetworkingV1NetworkPolicyRule) createRule(networkpolicy *networkingV1.NetworkPolicy, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(networkpolicy)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(networkpolicy)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(networkpolicy)\n\t\t},\n\t}\n\treturn rule\n}", "func (m *MockIPTService) CreateRule(ruleType int, ruleData interface{}) error {\n\tre, _, err := m.s.CreateRuleEntryString(ruleType, ruleData)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, ok := m.rules[re.ID]\n\tif ok {\n\t\treturn errors.New(\"Rule already exists\")\n\t}\n\n\tm.rules[re.ID] = re\n\n\treturn nil\n}", "func (s *Service) CreateFirewallRulePort(req CreateFirewallRulePortRequest) (TaskReference, error) {\n\tbody, err := s.createFirewallRulePortResponseBody(req)\n\n\treturn body.Data, err\n}", "func (r *V1Beta1ExtensionsNetworkPolicyRule) createRule(networkpolicy *v1beta1Extensions.NetworkPolicy, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(networkpolicy)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(networkpolicy)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(networkpolicy)\n\t\t},\n\t}\n\treturn rule\n}", "func NewForwardingRule(ctx *pulumi.Context,\n\tname string, args *ForwardingRuleArgs, opts ...pulumi.ResourceOption) (*ForwardingRule, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.AcceleratorId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'AcceleratorId'\")\n\t}\n\tif args.ListenerId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'ListenerId'\")\n\t}\n\tif args.RuleActions == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'RuleActions'\")\n\t}\n\tif args.RuleConditions == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'RuleConditions'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource ForwardingRule\n\terr := ctx.RegisterResource(\"alicloud:ga/forwardingRule:ForwardingRule\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (r *FirewallManagementIPRulesResource) Create(item FirewallManagementIPRulesConfig) error {\n\tif err := r.c.ModQuery(\"POST\", BasePath+FirewallManagementIPRulesEndpoint, item); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (m *MockFirewallServiceIface) CreateFirewallRule(p *CreateFirewallRuleParams) (*CreateFirewallRuleResponse, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"CreateFirewallRule\", p)\n\tret0, _ := ret[0].(*CreateFirewallRuleResponse)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (egw *NsxtEdgeGateway) CreateNatRule(natRuleConfig *types.NsxtNatRule) (*NsxtNatRule, error) {\n\tclient := egw.client\n\tendpoint := types.OpenApiPathVersion1_0_0 + types.OpenApiEndpointNsxtNatRules\n\tapiVersion, err := client.getOpenApiHighestElevatedVersion(endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Insert Edge Gateway ID into endpoint path edgeGateways/%s/nat/rules\n\turlRef, err := client.OpenApiBuildEndpoint(fmt.Sprintf(endpoint, egw.EdgeGateway.ID))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Creating NAT rule must follow different way than usual OpenAPI one because this item has an API bug and\n\t// NAT rule ID is not returned after this object is created. The only way to find its ID afterwards is to GET all\n\t// items, and manually match it based on rule name, etc.\n\ttask, err := client.OpenApiPostItemAsync(apiVersion, urlRef, nil, natRuleConfig)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error creating NSX-T NAT rule: %s\", err)\n\t}\n\n\terr = task.WaitTaskCompletion()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"task failed while creating NSX-T NAT rule: %s\", err)\n\t}\n\n\t// queryParameters (API side filtering) are not used because pretty much nothing is accepted as filter (such fields as\n\t// name, description, ruleType and even ID are not allowed\n\tallNatRules, err := egw.GetAllNatRules(nil)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error fetching all NAT rules: %s\", err)\n\t}\n\n\tfor index, singleRule := range allNatRules {\n\t\t// Look for a matching rule\n\t\tif singleRule.IsEqualTo(natRuleConfig) {\n\t\t\treturn allNatRules[index], nil\n\n\t\t}\n\t}\n\treturn nil, fmt.Errorf(\"rule '%s' of type '%s' not found after creation\", natRuleConfig.Name, natRuleConfig.RuleType)\n}", "func (r *V1PodSpecRule) createRule(podSpec *v1.PodSpec, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(podSpec)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(podSpec)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(podSpec)\n\t\t},\n\t}\n\treturn rule\n}", "func (mr *MockFirewallServiceIfaceMockRecorder) CreateFirewallRule(p interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"CreateFirewallRule\", reflect.TypeOf((*MockFirewallServiceIface)(nil).CreateFirewallRule), p)\n}", "func (r *V1ContainerRule) createRule(container *v1.Container, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(container)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(container)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(container)\n\t\t},\n\t}\n\treturn rule\n}", "func (c *Client) CreateFirewall(ctx context.Context, createOpts FirewallCreateOptions) (*Firewall, error) {\n\tvar body string\n\te, err := c.Firewalls.Endpoint()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq := c.R(ctx).SetResult(&Firewall{})\n\n\tif bodyData, err := json.Marshal(createOpts); err == nil {\n\t\tbody = string(bodyData)\n\t} else {\n\t\treturn nil, NewError(err)\n\t}\n\n\tr, err := coupleAPIErrors(req.\n\t\tSetBody(body).\n\t\tPost(e))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn r.Result().(*Firewall), nil\n}", "func NewCreateRule(edgeID string, natRule Rule) *CreateNatRuleAPI {\n\tthis := new(CreateNatRuleAPI)\n\tvar requestPayload Rules\n\trequestPayload.Rules = append(requestPayload.Rules, natRule)\n\n\tthis.BaseAPI = api.NewBaseAPI(http.MethodPost, \"/api/4.0/edges/\"+edgeID+\"/nat/config/rules\", requestPayload, new(string))\n\treturn this\n}", "func (r *RbacV1Beta1RoleBindingRule) createRule(rolebinding *rbacV1beta1.RoleBinding, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(rolebinding)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(rolebinding)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(rolebinding)\n\t\t},\n\t}\n\treturn rule\n}", "func (c *Client) CreateRule(args *CreateRuleArgs) (*CreateRuleResult, error) {\n\tjsonBytes, jsonErr := json.Marshal(args)\n\tif jsonErr != nil {\n\t\treturn nil, jsonErr\n\t}\n\tbody, err := bce.NewBodyFromBytes(jsonBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn CreateRule(c, body)\n}", "func NewRule(ctx *pulumi.Context,\n\tname string, args *RuleArgs, opts ...pulumi.ResourceOpt) (*Rule, error) {\n\tif args == nil || args.MetricName == nil {\n\t\treturn nil, errors.New(\"missing required argument 'MetricName'\")\n\t}\n\tinputs := make(map[string]interface{})\n\tif args == nil {\n\t\tinputs[\"metricName\"] = nil\n\t\tinputs[\"name\"] = nil\n\t\tinputs[\"predicates\"] = nil\n\t\tinputs[\"tags\"] = nil\n\t} else {\n\t\tinputs[\"metricName\"] = args.MetricName\n\t\tinputs[\"name\"] = args.Name\n\t\tinputs[\"predicates\"] = args.Predicates\n\t\tinputs[\"tags\"] = args.Tags\n\t}\n\tinputs[\"arn\"] = nil\n\ts, err := ctx.RegisterResource(\"aws:waf/rule:Rule\", name, true, inputs, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Rule{s: s}, nil\n}", "func (c *Client) CreateFirewall(ctx context.Context, params *CreateFirewallInput, optFns ...func(*Options)) (*CreateFirewallOutput, error) {\n\tif params == nil {\n\t\tparams = &CreateFirewallInput{}\n\t}\n\n\tresult, metadata, err := c.invokeOperation(ctx, \"CreateFirewall\", params, optFns, addOperationCreateFirewallMiddlewares)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tout := result.(*CreateFirewallOutput)\n\tout.ResultMetadata = metadata\n\treturn out, nil\n}", "func (a *BackendOptionsApiService) CreateHTTPRequestRule(ctx _context.Context, parentName string, parentType string, httpRequestRule HttpRequestRule, localVarOptionals *CreateHTTPRequestRuleOpts) (HttpRequestRule, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodPost\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue HttpRequestRule\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/services/haproxy/configuration/http_request_rules\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tlocalVarQueryParams.Add(\"parent_name\", parameterToString(parentName, \"\"))\n\tlocalVarQueryParams.Add(\"parent_type\", parameterToString(parentType, \"\"))\n\tif localVarOptionals != nil && localVarOptionals.TransactionId.IsSet() {\n\t\tlocalVarQueryParams.Add(\"transaction_id\", parameterToString(localVarOptionals.TransactionId.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Version.IsSet() {\n\t\tlocalVarQueryParams.Add(\"version\", parameterToString(localVarOptionals.Version.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.ForceReload.IsSet() {\n\t\tlocalVarQueryParams.Add(\"force_reload\", parameterToString(localVarOptionals.ForceReload.Value(), \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{\"application/json\"}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &httpRequestRule\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 201 {\n\t\t\tvar v HttpRequestRule\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 202 {\n\t\t\tvar v HttpRequestRule\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 400 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 409 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tvar v ModelError\n\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\tif err != nil {\n\t\t\tnewErr.error = err.Error()\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tnewErr.model = v\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func (r *BatchV1JobRule) createRule(job *batchV1.Job, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(job)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(job)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(job)\n\t\t},\n\t}\n\treturn rule\n}", "func checkCreateRule(t *testing.T, expError bool, tenant, policy, ruleID, dir, fnet, fepg, fip, tnet, tepg, tip, proto, act string, prio, port int) {\n\tpol := client.Rule{\n\t\tTenantName: tenant,\n\t\tPolicyName: policy,\n\t\tRuleID: ruleID,\n\t\tDirection: dir,\n\t\tPriority: prio,\n\t\tFromNetwork: fnet,\n\t\tFromEndpointGroup: fepg,\n\t\tFromIpAddress: fip,\n\t\tToNetwork: tnet,\n\t\tToEndpointGroup: tepg,\n\t\tToIpAddress: tip,\n\t\tProtocol: proto,\n\t\tPort: port,\n\t\tAction: act,\n\t}\n\terr := contivClient.RulePost(&pol)\n\tif err != nil && !expError {\n\t\tt.Fatalf(\"Error creating rule {%+v}. Err: %v\", pol, err)\n\t} else if err == nil && expError {\n\t\tt.Fatalf(\"Create rule {%+v} succeded while expecing error\", pol)\n\t} else if err == nil {\n\t\t// verify rule is created\n\t\t_, err := contivClient.RuleGet(tenant, policy, ruleID)\n\t\tif err != nil {\n\t\t\tt.Fatalf(\"Error getting rule %s/%s/%s. Err: %v\", tenant, policy, ruleID, err)\n\t\t}\n\t}\n}", "func (r *BatchV1Beta1CronJobRule) createRule(cronjob *batchV1beta1.CronJob, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(cronjob)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(cronjob)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(cronjob)\n\t\t},\n\t}\n\treturn rule\n}", "func (_m *ComputeAPI) DeleteFirewallRule(project string, firewall string) {\n\t_m.Called(project, firewall)\n}", "func (a *Client) CreateScopeRule(params *CreateScopeRuleParams, authInfo runtime.ClientAuthInfoWriter) (*CreateScopeRuleOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewCreateScopeRuleParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"CreateScopeRule\",\n\t\tMethod: \"POST\",\n\t\tPathPattern: \"/cdn/v1/stacks/{stack_id}/sites/{site_id}/scopes/{scope_id}/rules\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &CreateScopeRuleReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*CreateScopeRuleOK), nil\n\n}", "func RuleCreateHandler(w http.ResponseWriter, req *http.Request) {\n\tdecoder := json.NewDecoder(req.Body)\n\tdefer req.Body.Close()\n\n\trule := api.Rule{}\n\tif err := decoder.Decode(&rule); err != nil {\n\t\terrors.SendBadRequest(w, err.Error())\n\t\treturn\n\t}\n\n\tencoder := json.NewEncoder(w)\n\tmutex.Lock()\n\tclient := clients[clientID%numClients]\n\tclientID++\n\tmutex.Unlock()\n\tif err := client.Create(context.Background(), path.Join(\"/rules\", rule.Name), &rule); err != nil {\n\t\tencoder.Encode(err)\n\t} else {\n\t\tencoder.Encode(&rule)\n\t}\n}", "func LookupFirewallRule(ctx *pulumi.Context, args *LookupFirewallRuleArgs, opts ...pulumi.InvokeOption) (*LookupFirewallRuleResult, error) {\n\tvar rv LookupFirewallRuleResult\n\terr := ctx.Invoke(\"azure-native:datalakeanalytics/v20151001preview:getFirewallRule\", args, &rv, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &rv, nil\n}", "func (m *MockFirewallServiceIface) CreateIpv6FirewallRule(p *CreateIpv6FirewallRuleParams) (*CreateIpv6FirewallRuleResponse, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"CreateIpv6FirewallRule\", p)\n\tret0, _ := ret[0].(*CreateIpv6FirewallRuleResponse)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func createDistributedFirewallDefinitions(check *C, vcd *TestVCD, vdcGroupId string, vcdClient *VCDClient, vdc *Vdc) ([]*types.DistributedFirewallRule, *NsxtFirewallGroup, *NsxtFirewallGroup) {\n\t// This number does not impact performance because all rules are created at once in the API\n\tnumberOfRules := 40\n\n\t// Pre-Create Firewall Groups (IP Set and Security Group to randomly configure them)\n\tipSet := preCreateVdcGroupIpSet(check, vcd, vdcGroupId, vdc)\n\tsecGroup := preCreateVdcGroupSecurityGroup(check, vcd, vdcGroupId, vdc)\n\tfwGroupIds := []string{ipSet.NsxtFirewallGroup.ID, secGroup.NsxtFirewallGroup.ID}\n\tfwGroupRefs := convertSliceOfStringsToOpenApiReferenceIds(fwGroupIds)\n\tappPortProfileReferences := getRandomListOfAppPortProfiles(check, vcd)\n\tnetworkContextProfiles := getRandomListOfNetworkContextProfiles(check, vcd, vcdClient)\n\n\tfirewallRules := make([]*types.DistributedFirewallRule, numberOfRules)\n\tfor a := 0; a < numberOfRules; a++ {\n\n\t\t// Feed in empty value for source and destination or a firewall group\n\t\tsrc := pickRandomOpenApiRefOrEmpty(fwGroupRefs)\n\t\tvar srcValue []types.OpenApiReference\n\t\tdst := pickRandomOpenApiRefOrEmpty(fwGroupRefs)\n\t\tvar dstValue []types.OpenApiReference\n\t\tif src != (types.OpenApiReference{}) {\n\t\t\tsrcValue = []types.OpenApiReference{src}\n\t\t}\n\t\tif dst != (types.OpenApiReference{}) {\n\t\t\tdstValue = []types.OpenApiReference{dst}\n\t\t}\n\n\t\tfirewallRules[a] = &types.DistributedFirewallRule{\n\t\t\tName: check.TestName() + strconv.Itoa(a),\n\t\t\tActionValue: pickRandomString([]string{\"ALLOW\", \"DROP\", \"REJECT\"}),\n\t\t\tEnabled: a%2 == 0,\n\t\t\tSourceFirewallGroups: srcValue,\n\t\t\tDestinationFirewallGroups: dstValue,\n\t\t\tApplicationPortProfiles: appPortProfileReferences[0:a],\n\t\t\tIpProtocol: pickRandomString([]string{\"IPV6\", \"IPV4\", \"IPV4_IPV6\"}),\n\t\t\tLogging: a%2 == 1,\n\t\t\tDirection: pickRandomString([]string{\"IN\", \"OUT\", \"IN_OUT\"}),\n\t\t}\n\n\t\t// Network Context Profile can usually work with up to one Application Profile therefore this\n\t\t// needs to be explicitly preset\n\t\tif a%5 == 1 { // Every fifth rule\n\t\t\tnetCtxProfile := networkContextProfiles[0:a]\n\t\t\tnetworkContextProfile := make([]types.OpenApiReference, 0)\n\t\t\tfor _, netCtxProf := range netCtxProfile {\n\t\t\t\tif netCtxProf.ID != \"\" {\n\t\t\t\t\tnetworkContextProfile = append(networkContextProfile, types.OpenApiReference{ID: netCtxProf.ID, Name: netCtxProf.Name})\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tfirewallRules[a].NetworkContextProfiles = networkContextProfile\n\t\t\t// firewallRules[a].ApplicationPortProfiles = appPortProfileReferences[0:1]\n\t\t\tfirewallRules[a].ApplicationPortProfiles = nil\n\n\t\t}\n\n\t\t// API V36.2 introduced new field Comment which is shown in UI\n\t\tif vcd.client.Client.APIVCDMaxVersionIs(\">= 36.2\") {\n\t\t\tfirewallRules[a].Comments = \"Comment Rule\"\n\t\t}\n\n\t}\n\n\treturn firewallRules, ipSet, secGroup\n}", "func (manager *NetworkPolicyManager) getOrCreateFirewallManager(pod, prev *core_v1.Pod) (pcn_firewall.PcnFirewall, bool) {\n\tl := log.New().WithFields(log.Fields{\"by\": PM, \"method\": \"getOrCreateFirewallManager(\" + pod.Name + \")\"})\n\tfwKey := pod.Namespace + \"|\" + utils.ImplodeLabels(pod.Labels, \",\", true)\n\n\t//-------------------------------------\n\t// Already linked?\n\t//-------------------------------------\n\tlinkedFw, wasLinked := manager.linkedPods[pod.UID]\n\tif wasLinked && linkedFw != fwKey {\n\t\tprevFwKey := prev.Namespace + \"|\" + utils.ImplodeLabels(prev.Labels, \",\", true)\n\n\t\t// This pod was already linked to a firewall manager,\n\t\t// but it's not the one we expected. This means that someone\n\t\t// (user or plugin) changed this pod's labels,\n\t\t// so we now need to unlink the pod from its current fw manager.\n\t\tprevFw, exists := manager.localFirewalls[prevFwKey]\n\t\tif exists {\n\t\t\tunlinked, remaining := prevFw.Unlink(pod, pcn_firewall.CleanFirewall)\n\t\t\tif !unlinked {\n\t\t\t\tl.Warningf(\"%s was not linked in previous firewall manager!\", pod.UID)\n\t\t\t} else {\n\t\t\t\tif remaining == 0 {\n\t\t\t\t\tmanager.flagForDeletion(prevFw.Name())\n\t\t\t\t}\n\t\t\t\tdelete(manager.linkedPods, pod.UID)\n\t\t\t}\n\t\t} else {\n\t\t\tl.Warningf(\"Could not find %s previous firewall manager!\", pod.UID)\n\t\t}\n\t}\n\n\t//-------------------------------------\n\t// Create and link it\n\t//-------------------------------------\n\tfw, exists := manager.localFirewalls[fwKey]\n\tif !exists {\n\t\tmanager.localFirewalls[fwKey] = startFirewall(manager.fwAPI, manager.podController, manager.vPodsRange, fwKey, pod.Namespace, pod.Labels)\n\t\tfw = manager.localFirewalls[fwKey]\n\t\treturn fw, true\n\t}\n\treturn fw, false\n}", "func (self *PolicyAgent) AddRule(rule *OfnetPolicyRule, ret *bool) error {\n\tvar ipDa *net.IP = nil\n\tvar ipDaMask *net.IP = nil\n\tvar ipSa *net.IP = nil\n\tvar ipSaMask *net.IP = nil\n\tvar md *uint64 = nil\n\tvar mdm *uint64 = nil\n\tvar flag, flagMask uint16\n\tvar flagPtr, flagMaskPtr *uint16\n\tvar err error\n\n\t// make sure switch is connected\n\tif !self.agent.IsSwitchConnected() {\n\t\tself.agent.WaitForSwitchConnection()\n\t}\n\n\t// check if we already have the rule\n\tself.mutex.RLock()\n\tif self.Rules[rule.RuleId] != nil {\n\t\toldRule := self.Rules[rule.RuleId].Rule\n\n\t\tif ruleIsSame(oldRule, rule) {\n\t\t\tself.mutex.RUnlock()\n\t\t\treturn nil\n\t\t} else {\n\t\t\tself.mutex.RUnlock()\n\t\t\tlog.Errorf(\"Rule already exists. new rule: {%+v}, old rule: {%+v}\", rule, oldRule)\n\t\t\treturn errors.New(\"Rule already exists\")\n\t\t}\n\t}\n\tself.mutex.RUnlock()\n\n\tlog.Infof(\"Received AddRule: %+v\", rule)\n\n\t// Parse dst ip\n\tif rule.DstIpAddr != \"\" {\n\t\tipDa, ipDaMask, err = ParseIPAddrMaskString(rule.DstIpAddr)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error parsing dst ip %s. Err: %v\", rule.DstIpAddr, err)\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// parse src ip\n\tif rule.SrcIpAddr != \"\" {\n\t\tipSa, ipSaMask, err = ParseIPAddrMaskString(rule.SrcIpAddr)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error parsing src ip %s. Err: %v\", rule.SrcIpAddr, err)\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// parse source/dst endpoint groups\n\tif rule.SrcEndpointGroup != 0 && rule.DstEndpointGroup != 0 {\n\t\tsrcMetadata, srcMetadataMask := SrcGroupMetadata(rule.SrcEndpointGroup)\n\t\tdstMetadata, dstMetadataMask := DstGroupMetadata(rule.DstEndpointGroup)\n\t\tmetadata := srcMetadata | dstMetadata\n\t\tmetadataMask := srcMetadataMask | dstMetadataMask\n\t\tmd = &metadata\n\t\tmdm = &metadataMask\n\t} else if rule.SrcEndpointGroup != 0 {\n\t\tsrcMetadata, srcMetadataMask := SrcGroupMetadata(rule.SrcEndpointGroup)\n\t\tmd = &srcMetadata\n\t\tmdm = &srcMetadataMask\n\t} else if rule.DstEndpointGroup != 0 {\n\t\tdstMetadata, dstMetadataMask := DstGroupMetadata(rule.DstEndpointGroup)\n\t\tmd = &dstMetadata\n\t\tmdm = &dstMetadataMask\n\t}\n\n\t// Setup TCP flags\n\tif rule.IpProtocol == 6 && rule.TcpFlags != \"\" {\n\t\tswitch rule.TcpFlags {\n\t\tcase \"syn\":\n\t\t\tflag = TCP_FLAG_SYN\n\t\t\tflagMask = TCP_FLAG_SYN\n\t\tcase \"syn,ack\":\n\t\t\tflag = TCP_FLAG_ACK | TCP_FLAG_SYN\n\t\t\tflagMask = TCP_FLAG_ACK | TCP_FLAG_SYN\n\t\tcase \"ack\":\n\t\t\tflag = TCP_FLAG_ACK\n\t\t\tflagMask = TCP_FLAG_ACK\n\t\tcase \"syn,!ack\":\n\t\t\tflag = TCP_FLAG_SYN\n\t\t\tflagMask = TCP_FLAG_ACK | TCP_FLAG_SYN\n\t\tcase \"!syn,ack\":\n\t\t\tflag = TCP_FLAG_ACK\n\t\t\tflagMask = TCP_FLAG_ACK | TCP_FLAG_SYN\n\t\tdefault:\n\t\t\tlog.Errorf(\"Unknown TCP flags: %s, in rule: %+v\", rule.TcpFlags, rule)\n\t\t\treturn errors.New(\"Unknown TCP flag\")\n\t\t}\n\n\t\tflagPtr = &flag\n\t\tflagMaskPtr = &flagMask\n\t}\n\t// Install the rule in policy table\n\truleFlow, err := self.policyTable.NewFlow(ofctrl.FlowMatch{\n\t\tPriority: uint16(FLOW_POLICY_PRIORITY_OFFSET + rule.Priority),\n\t\tEthertype: 0x0800,\n\t\tIpDa: ipDa,\n\t\tIpDaMask: ipDaMask,\n\t\tIpSa: ipSa,\n\t\tIpSaMask: ipSaMask,\n\t\tIpProto: rule.IpProtocol,\n\t\tTcpSrcPort: rule.SrcPort,\n\t\tTcpDstPort: rule.DstPort,\n\t\tUdpSrcPort: rule.SrcPort,\n\t\tUdpDstPort: rule.DstPort,\n\t\tMetadata: md,\n\t\tMetadataMask: mdm,\n\t\tTcpFlags: flagPtr,\n\t\tTcpFlagsMask: flagMaskPtr,\n\t})\n\tif err != nil {\n\t\tlog.Errorf(\"Error adding flow for rule {%v}. Err: %v\", rule, err)\n\t\treturn err\n\t}\n\n\t// Point it to next table\n\tif rule.Action == \"allow\" {\n\t\terr = ruleFlow.Next(self.nextTable)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error installing flow {%+v}. Err: %v\", ruleFlow, err)\n\t\t\treturn err\n\t\t}\n\t} else if rule.Action == \"deny\" {\n\t\terr = ruleFlow.Next(self.ofSwitch.DropAction())\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error installing flow {%+v}. Err: %v\", ruleFlow, err)\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\tlog.Errorf(\"Unknown action in rule {%+v}\", rule)\n\t\treturn errors.New(\"Unknown action in rule\")\n\t}\n\n\t// save the rule\n\tpRule := PolicyRule{\n\t\tRule: rule,\n\t\tflow: ruleFlow,\n\t}\n\tself.mutex.Lock()\n\tself.Rules[rule.RuleId] = &pRule\n\tself.mutex.Unlock()\n\n\treturn nil\n}", "func (m *MockFirewallServiceIface) CreatePortForwardingRule(p *CreatePortForwardingRuleParams) (*CreatePortForwardingRuleResponse, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"CreatePortForwardingRule\", p)\n\tret0, _ := ret[0].(*CreatePortForwardingRuleResponse)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func newRuleHandler(d discovery.Discoverer, cfg RunConfig) discovery.RuleHandler {\n\trh := &ruleHandler{\n\t\td: d.(*discoverer),\n\t\tph: cfg.Handler,\n\t\tdaemon: cfg.Daemon,\n\t\tkubeClient: cfg.KubeClient,\n\t\tlister: cfg.Lister,\n\t\trulesCount: gm.GetOrRegisterGauge(\"discovery.rules.count\", gm.DefaultRegistry),\n\t}\n\tcount := int64(len(rh.d.delegates))\n\trh.rulesCount.Update(count)\n\treturn rh\n}", "func (fw *IPtables) CreateRules(chain int) error {\n\tlog.Info(\"In CreateRules() for chain\", chain)\n\tfor _, rule := range fw.chains[chain].Rules {\n\t\t// First create rule record in database.\n\t\terr0 := fw.addIPtablesRule(rule)\n\t\tif err0 != nil {\n\t\t\tlog.Error(\"In CreateRules() create db record for iptables rule \", rule.GetBody())\n\t\t\treturn err0\n\t\t}\n\n\t\terr1 := fw.EnsureRule(rule, EnsureFirst)\n\t\tif err1 != nil {\n\t\t\tlog.Error(\"In CreateRules() failed to create install firewall rule \", rule.GetBody())\n\t\t\treturn err1\n\t\t}\n\n\t\t// Finally, set 'active' flag in database record.\n\t\tif err2 := fw.Store.switchIPtablesRule(rule, setRuleActive); err2 != nil {\n\t\t\tlog.Error(\"In CreateRules() iptables rule created, but activation failed \", rule.GetBody())\n\t\t\treturn err2\n\t\t}\n\t}\n\tlog.Info(\"Creating firewall rules success\")\n\treturn nil\n}", "func (f *fakeLB) ListFirewallRules(_ context.Context, _ string, _ *govultr.ListOptions) ([]govultr.LBFirewallRule, *govultr.Meta, *http.Response, error) {\n\treturn nil, nil, nil, nil\n}", "func CreateL7Rule(t *testing.T, client *gophercloud.ServiceClient, policyID string, lb *loadbalancers.LoadBalancer) (*l7policies.Rule, error) {\n\tt.Logf(\"Attempting to create l7 rule for policy %s\", policyID)\n\n\tcreateOpts := l7policies.CreateRuleOpts{\n\t\tRuleType: l7policies.TypePath,\n\t\tCompareType: l7policies.CompareTypeStartWith,\n\t\tValue: \"/api\",\n\t}\n\n\trule, err := l7policies.CreateRule(client, policyID, createOpts).Extract()\n\tif err != nil {\n\t\treturn rule, err\n\t}\n\n\tt.Logf(\"Successfully created l7 rule for policy %s\", policyID)\n\n\tif err := WaitForLoadBalancerState(client, lb.ID, \"ACTIVE\"); err != nil {\n\t\treturn rule, fmt.Errorf(\"Timed out waiting for loadbalancer to become active: %s\", err)\n\t}\n\n\tth.AssertEquals(t, rule.RuleType, string(l7policies.TypePath))\n\tth.AssertEquals(t, rule.CompareType, string(l7policies.CompareTypeStartWith))\n\tth.AssertEquals(t, rule.Value, \"/api\")\n\n\treturn rule, nil\n}", "func newRuleImpl(impl starlark.Callable, defaults *starlark.Dict) (*ruleImpl, error) {\n\tpairs := defaults.Items()\n\tfor _, pair := range pairs {\n\t\tk, v := pair[0], pair[1]\n\t\tif _, ok := k.(starlark.String); !ok {\n\t\t\treturn nil, fmt.Errorf(\"lucicfg.rule: keys in \\\"defaults\\\" must be strings\")\n\t\t}\n\t\tif !isNamedStruct(v, \"lucicfg.var\") {\n\t\t\treturn nil, fmt.Errorf(\"lucicfg.rule: values in \\\"defaults\\\" must be lucicfg.var\")\n\t\t}\n\t}\n\treturn &ruleImpl{\n\t\tCallable: impl,\n\t\tdefaults: starlarkstruct.FromKeywords(starlark.String(\"lucicfg.rule.defaults\"), pairs),\n\t}, nil\n}", "func NewFirewallRuleListResultPage (getNextPage func(context.Context, FirewallRuleListResult) (FirewallRuleListResult, error)) FirewallRuleListResultPage {\n return FirewallRuleListResultPage{fn: getNextPage}\n }", "func (api *API) TeamsCreateRule(ctx context.Context, accountID string, rule TeamsRule) (TeamsRule, error) {\n\turi := fmt.Sprintf(\"/accounts/%s/gateway/rules\", accountID)\n\n\tres, err := api.makeRequestContext(ctx, http.MethodPost, uri, rule)\n\tif err != nil {\n\t\treturn TeamsRule{}, err\n\t}\n\n\tvar teamsRuleResponse TeamsRuleResponse\n\terr = json.Unmarshal(res, &teamsRuleResponse)\n\tif err != nil {\n\t\treturn TeamsRule{}, fmt.Errorf(\"%s: %w\", errUnmarshalError, err)\n\t}\n\n\treturn teamsRuleResponse.Result, nil\n}", "func (a *BackendOptionsApiService) CreateTCPRequestRule(ctx _context.Context, parentName string, parentType string, tcpRequestRule TcpRequestRule, localVarOptionals *CreateTCPRequestRuleOpts) (TcpRequestRule, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodPost\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue TcpRequestRule\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/services/haproxy/configuration/tcp_request_rules\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tlocalVarQueryParams.Add(\"parent_name\", parameterToString(parentName, \"\"))\n\tlocalVarQueryParams.Add(\"parent_type\", parameterToString(parentType, \"\"))\n\tif localVarOptionals != nil && localVarOptionals.TransactionId.IsSet() {\n\t\tlocalVarQueryParams.Add(\"transaction_id\", parameterToString(localVarOptionals.TransactionId.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Version.IsSet() {\n\t\tlocalVarQueryParams.Add(\"version\", parameterToString(localVarOptionals.Version.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.ForceReload.IsSet() {\n\t\tlocalVarQueryParams.Add(\"force_reload\", parameterToString(localVarOptionals.ForceReload.Value(), \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{\"application/json\"}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &tcpRequestRule\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 201 {\n\t\t\tvar v TcpRequestRule\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 202 {\n\t\t\tvar v TcpRequestRule\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 400 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 409 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tvar v ModelError\n\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\tif err != nil {\n\t\t\tnewErr.error = err.Error()\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tnewErr.model = v\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func (ruleset *DnsForwardingRuleset) ValidateCreate() (admission.Warnings, error) {\n\tvalidations := ruleset.createValidations()\n\tvar temp any = ruleset\n\tif runtimeValidator, ok := temp.(genruntime.Validator); ok {\n\t\tvalidations = append(validations, runtimeValidator.CreateValidations()...)\n\t}\n\treturn genruntime.ValidateCreate(validations)\n}", "func (c *TestClient) GetFirewallRule(project, name string) (*compute.Firewall, error) {\n\tif c.GetFirewallRuleFn != nil {\n\t\treturn c.GetFirewallRuleFn(project, name)\n\t}\n\treturn c.client.GetFirewallRule(project, name)\n}", "func (client *Client) CreateRule(request *CreateRuleRequest) (_result *CreateRuleResponse, _err error) {\n\truntime := &util.RuntimeOptions{}\n\t_result = &CreateRuleResponse{}\n\t_body, _err := client.CreateRuleWithOptions(request, runtime)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\t_result = _body\n\treturn _result, _err\n}", "func (r *FirewallGlobalRulesStagedPolicyRulesResource) Create(item FirewallGlobalRulesStagedPolicyRulesConfig) error {\n\tif err := r.c.ModQuery(\"POST\", BasePath+FirewallGlobalRulesStagedPolicyRulesEndpoint, item); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func NewFirewall() *Firewall {\n\treturn &Firewall{\n\t\tBlacklisting: true,\n\t\tErrorHandler: router.ErrorHandler(func(w http.ResponseWriter, r *http.Request, code int) {\n\t\t\tswitch code {\n\t\t\tcase 403:\n\t\t\t\thttp.Error(w, \"Forbidden\", 403)\n\t\t\t}\n\t\t}),\n\t}\n}", "func (r *RbacV1RoleRule) createRule(role *rbacV1.Role, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(role)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(role)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(role)\n\t\t},\n\t}\n\treturn rule\n}", "func (a *BackendOptionsApiService) CreateServerSwitchingRule(ctx _context.Context, backend string, serverSwitchingRule ServerSwitchingRule, localVarOptionals *CreateServerSwitchingRuleOpts) (ServerSwitchingRule, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodPost\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue ServerSwitchingRule\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/services/haproxy/configuration/server_switching_rules\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tlocalVarQueryParams.Add(\"backend\", parameterToString(backend, \"\"))\n\tif localVarOptionals != nil && localVarOptionals.TransactionId.IsSet() {\n\t\tlocalVarQueryParams.Add(\"transaction_id\", parameterToString(localVarOptionals.TransactionId.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Version.IsSet() {\n\t\tlocalVarQueryParams.Add(\"version\", parameterToString(localVarOptionals.Version.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.ForceReload.IsSet() {\n\t\tlocalVarQueryParams.Add(\"force_reload\", parameterToString(localVarOptionals.ForceReload.Value(), \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{\"application/json\"}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &serverSwitchingRule\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 201 {\n\t\t\tvar v ServerSwitchingRule\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 202 {\n\t\t\tvar v ServerSwitchingRule\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 400 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 409 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tvar v ModelError\n\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\tif err != nil {\n\t\t\tnewErr.error = err.Error()\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tnewErr.model = v\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func (ks *kuiperService) CreateRules(ctx context.Context, token string, rules ...Rule) ([]Rule, error) {\n\tres, err := ks.auth.Identify(ctx, &mainflux.Token{Value: token})\n\tif err != nil {\n\t\treturn []Rule{}, ErrUnauthorizedAccess\n\t}\n\tfor i := range rules {\n\t\trules[i].ID, err = ks.idp.ID()\n\t\tif err != nil {\n\t\t\treturn []Rule{}, err\n\t\t}\n\t\trules[i].Owner = res.GetValue()\n\t}\n\n\t// Save the created rules into repository at first\n\tif _, err := ks.rules.Save(ctx, rules...); err != nil {\n\t\treturn []Rule{}, err\n\t}\n\tfor i := range rules {\n\t\tr, err := ks.ruleManager.getRuleByJson(rules[i].Name, rules[i].SQL)\n\t\tif err != nil {\n\t\t\treturn []Rule{}, err\n\t\t}\n\t\t//Start the rule\n\t\trs, err := ks.ruleManager.createRuleState(r)\n\t\tif err != nil {\n\t\t\treturn []Rule{}, err\n\t\t} else {\n\t\t\terr = ks.ruleManager.doStartRule(rs)\n\t\t\tif err != nil {\n\t\t\t\treturn []Rule{}, err\n\t\t\t}\n\t\t}\n\t}\n\treturn rules, nil\n}", "func NewRule() *Rule {\n\tthis := Rule{}\n\treturn &this\n}", "func newRule(defaultExpiration time.Duration, allowed int, estimated ...int) *singleRule {\n\tif allowed <= 0 {\n\t\tallowed = 1\n\t}\n\tuserEstimated := 0\n\tif len(estimated) > 0 {\n\t\tuserEstimated = estimated[0]\n\t}\n\tif userEstimated <= 0 {\n\t\tuserEstimated = allowed\n\t}\n\tcleanupInterval := defaultExpiration / 100\n\tif cleanupInterval < time.Second*1 {\n\t\tcleanupInterval = time.Second * 1\n\t}\n\tif cleanupInterval > time.Second*60 {\n\t\tcleanupInterval = time.Second * 60\n\t}\n\tvc := createRule(defaultExpiration, cleanupInterval, allowed, userEstimated)\n\tgo vc.deleteExpired()\n\treturn vc\n}", "func (r *V1PersistentVolumeClaimRule) createRule(pvc *v1.PersistentVolumeClaim, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(pvc)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(pvc)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(pvc)\n\t\t},\n\t}\n\treturn rule\n}", "func (s *PolicySets) NewRule(isInbound bool, priority uint16) *hns.ACLPolicy {\n\tdirection := hns.Out\n\tif isInbound {\n\t\tdirection = hns.In\n\t}\n\n\treturn &hns.ACLPolicy{\n\t\tType: hns.ACL,\n\t\tRuleType: hns.Switch,\n\t\tAction: hns.Block,\n\t\tDirection: direction,\n\t\tProtocol: 256, // Any, only required for RS3\n\t\tPriority: priority,\n\t}\n}", "func (g *Grammar) CreateRule(name string, terms ...Term) {\n\trule := g.Lookup(name)\n\tprod := Production{}\n\tfor _, t := range terms {\n\t\tprod = append(prod, t)\n\t}\n\trule.Productions = append(rule.Productions, prod)\n\tg.rules[name] = rule\n}", "func AddRule(r AuditRule) (err error) {\n\tard, _, _, err := r.toKernelAuditRule()\n\tif err != nil {\n\t\treturn\n\t}\n\tclient, err := libaudit.NewAuditClient(nil)\n\tdefer client.Close()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"Failed to initialize client\")\n\t}\n\terr = client.AddRule(ard.toWireFormat())\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (a *SyncApiService) CreateSyncRule(ctx context.Context, syncRule SyncRuleCreateParams) (CreateResponse, *http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Post\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t \tsuccessPayload CreateResponse\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/platform/3/sync/rules\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{ \"application/json\", }\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\n\t\t\"application/json\",\n\t\t}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &syncRule\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn successPayload, nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn successPayload, localVarHttpResponse, err\n\t}\n\tdefer localVarHttpResponse.Body.Close()\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tbodyBytes, _ := ioutil.ReadAll(localVarHttpResponse.Body)\n\t\treturn successPayload, localVarHttpResponse, reportError(\"Status: %v, Body: %s\", localVarHttpResponse.Status, bodyBytes)\n\t}\n\n\tif err = json.NewDecoder(localVarHttpResponse.Body).Decode(&successPayload); err != nil {\n\t\treturn successPayload, localVarHttpResponse, err\n\t}\n\n\n\treturn successPayload, localVarHttpResponse, err\n}", "func (fw *IPtables) injectRule(rule *IPtablesRule) error {\n\trule2arr := strings.Split(rule.GetBody(), \" \")\n\tif len(rule2arr) < 3 {\n\t\treturn fmt.Errorf(\"In injectRule() not enough elements in rule %s\", rule.GetBody())\n\t}\n\n\truleChain := rule2arr[0]\n\n\tfor i, chain := range fw.chains {\n\t\tif chain.ChainName == ruleChain {\n\t\t\tfw.chains[i].Rules = append(fw.chains[i].Rules, rule)\n\t\t\tlog.Infof(\"In injectRule() adding rule %s into chain %s\", rule.GetBody(), chain.ChainName)\n\t\t\treturn nil\n\t\t}\n\t}\n\n\t// TODO should we create new chain instead of throwing error?\n\treturn fmt.Errorf(\"In injectRule() firewall doesn't manage chain for rule %s\", rule.GetBody())\n}", "func (manager *NetworkPolicyManager) getOrCreateFirewallManager(pod, prev *core_v1.Pod) (pcn_firewall.PcnFirewallManager, bool) {\n\tf := \"[Network Policies Manager](getOrCreateFirewallManager) \"\n\tfwKey := pod.Namespace + \"|\" + utils.ImplodeLabels(pod.Labels, \",\", true)\n\n\t//-------------------------------------\n\t// Already linked?\n\t//-------------------------------------\n\tlinkedFw, wasLinked := manager.linkedPods[pod.UID]\n\tif wasLinked && linkedFw != fwKey {\n\t\tprevFwKey := prev.Namespace + \"|\" + utils.ImplodeLabels(prev.Labels, \",\", true)\n\n\t\t// This pod was already linked to a firewall manager,\n\t\t// but it's not the one we expected. This means that someone\n\t\t// (user or plugin) changed this pod's labels,\n\t\t// so we now need to unlink the pod from its current fw manager.\n\t\tprevFw, exists := manager.localFirewalls[prevFwKey]\n\t\tif exists {\n\t\t\tunlinked, remaining := prevFw.Unlink(pod, pcn_firewall.CleanFirewall)\n\t\t\tif !unlinked {\n\t\t\t\tlogger.Warningf(f+\"%s was not linked in previous firewall manager!\", pod.UID)\n\t\t\t} else {\n\t\t\t\tif remaining == 0 {\n\t\t\t\t\tmanager.flagForDeletion(prevFw.Name())\n\t\t\t\t}\n\t\t\t\tdelete(manager.linkedPods, pod.UID)\n\t\t\t}\n\t\t} else {\n\t\t\tlogger.Warningf(f+\"Could not find %s previous firewall manager!\", pod.UID)\n\t\t}\n\t}\n\n\t//-------------------------------------\n\t// Create and link it\n\t//-------------------------------------\n\tfw, exists := manager.localFirewalls[fwKey]\n\tif !exists {\n\t\tmanager.localFirewalls[fwKey] = startFirewall(fwKey, pod.Namespace, pod.Labels)\n\t\tfw = manager.localFirewalls[fwKey]\n\t\treturn fw, true\n\t}\n\treturn fw, false\n}", "func (s *Service) CreateNetworkRule(req CreateNetworkRuleRequest) (TaskReference, error) {\n\tbody, err := s.createNetworkRuleResponseBody(req)\n\n\treturn body.Data, err\n}", "func RegisterRule(rule string, f RuleFunc) error {\n\tcustomRuleFuncMap[rule] = f\n\treturn nil\n}", "func CreateBandwidthLimitRule(client *gophercloud.ServiceClient, policyID string, opts CreateBandwidthLimitRuleOptsBuilder) (r CreateBandwidthLimitRuleResult) {\n\tb, err := opts.ToBandwidthLimitRuleCreateMap()\n\tif err != nil {\n\t\tr.Err = err\n\t\treturn\n\t}\n\t_, r.Err = client.Post(createBandwidthLimitRuleURL(client, policyID), b, &r.Body, &gophercloud.RequestOpts{\n\t\tOkCodes: []int{201},\n\t})\n\treturn\n}", "func (fw *IPtables) CreateDefaultRule(chain int, target string) error {\n\tlog.Infof(\"In CreateDefaultRule() %s rules for chain %d\", target, chain)\n\tchainName := fw.chains[chain].ChainName\n\tbody := fmt.Sprintf(\"%s %s %s\", chainName, \"-j\", target)\n\trule := &IPtablesRule{\n\t\tBody: body,\n\t\tState: setRuleInactive.String(),\n\t}\n\n\t// First create rule record in database.\n\terr0 := fw.addIPtablesRule(rule)\n\tif err0 != nil {\n\t\tlog.Error(\"In CreateDefaultRule() create db record for iptables rule \", rule.GetBody())\n\t\treturn err0\n\t}\n\n\terr1 := fw.EnsureRule(rule, EnsureLast)\n\tif err1 != nil {\n\t\tlog.Errorf(\"In CreateDefaultRule() %s rules failed\", target)\n\t\treturn err1\n\t}\n\n\t// Finally, set 'active' flag in database record.\n\tif err2 := fw.Store.switchIPtablesRule(rule, setRuleActive); err2 != nil {\n\t\tlog.Error(\"In CreateDefaultRule() iptables rule created but activation failed \", rule.GetBody())\n\t\treturn err2\n\t}\n\n\tlog.Info(\"In CreateDefaultRule() success\")\n\treturn nil\n}", "func (a *BackendOptionsApiService) CreateTCPResponseRule(ctx _context.Context, backend string, tcpResponseRule TcpResponseRule, localVarOptionals *CreateTCPResponseRuleOpts) (TcpResponseRule, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodPost\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue TcpResponseRule\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/services/haproxy/configuration/tcp_response_rules\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tlocalVarQueryParams.Add(\"backend\", parameterToString(backend, \"\"))\n\tif localVarOptionals != nil && localVarOptionals.TransactionId.IsSet() {\n\t\tlocalVarQueryParams.Add(\"transaction_id\", parameterToString(localVarOptionals.TransactionId.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Version.IsSet() {\n\t\tlocalVarQueryParams.Add(\"version\", parameterToString(localVarOptionals.Version.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.ForceReload.IsSet() {\n\t\tlocalVarQueryParams.Add(\"force_reload\", parameterToString(localVarOptionals.ForceReload.Value(), \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{\"application/json\"}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &tcpResponseRule\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 201 {\n\t\t\tvar v TcpResponseRule\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 202 {\n\t\t\tvar v TcpResponseRule\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 400 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 409 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tvar v ModelError\n\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\tif err != nil {\n\t\t\tnewErr.error = err.Error()\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tnewErr.model = v\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func (fr *FirewallRule) UnmarshalJSON(body []byte) error {\n\tvar m map[string]*json.RawMessage\n\terr := json.Unmarshal(body, &m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor k, v := range m {\n\t\tswitch k {\n\t\tcase \"properties\":\n\t\t\tif v != nil {\n\t\t\t\tvar firewallRuleProperties FirewallRuleProperties\n\t\t\t\terr = json.Unmarshal(*v, &firewallRuleProperties)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tfr.FirewallRuleProperties = &firewallRuleProperties\n\t\t\t}\n\t\tcase \"id\":\n\t\t\tif v != nil {\n\t\t\t\tvar ID string\n\t\t\t\terr = json.Unmarshal(*v, &ID)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tfr.ID = &ID\n\t\t\t}\n\t\tcase \"name\":\n\t\t\tif v != nil {\n\t\t\t\tvar name string\n\t\t\t\terr = json.Unmarshal(*v, &name)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tfr.Name = &name\n\t\t\t}\n\t\tcase \"type\":\n\t\t\tif v != nil {\n\t\t\t\tvar typeVar string\n\t\t\t\terr = json.Unmarshal(*v, &typeVar)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tfr.Type = &typeVar\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (fw *AzureSqlFirewallRuleManager) Ensure(ctx context.Context, obj runtime.Object, opts ...resourcemanager.ConfigOption) (bool, error) {\n\tinstance, err := fw.convert(obj)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\tsubscriptionID := instance.Spec.SubscriptionID\n\tgroupName := instance.Spec.ResourceGroup\n\tserver := instance.Spec.Server\n\truleName := instance.ObjectMeta.Name\n\tstartIP := instance.Spec.StartIPAddress\n\tendIP := instance.Spec.EndIPAddress\n\n\tfwr, err := fw.GetSQLFirewallRule(ctx, subscriptionID, groupName, server, ruleName)\n\tif err == nil {\n\t\tinstance.Status.Provisioning = false\n\t\tinstance.Status.Provisioned = true\n\t\tinstance.Status.Message = resourcemanager.SuccessMsg\n\t\tinstance.Status.ResourceId = *fwr.ID\n\t\treturn true, nil\n\t}\n\tinstance.Status.Message = fmt.Sprintf(\"AzureSqlFirewallRule Get error %s\", err.Error())\n\trequeuErrors := []string{\n\t\terrhelp.ResourceGroupNotFoundErrorCode,\n\t\terrhelp.ParentNotFoundErrorCode,\n\t}\n\tazerr := errhelp.NewAzureError(err)\n\tif helpers.ContainsString(requeuErrors, azerr.Type) {\n\t\tinstance.Status.Provisioning = false\n\t\treturn false, nil\n\t}\n\n\t_, err = fw.CreateOrUpdateSQLFirewallRule(ctx, subscriptionID, groupName, server, ruleName, startIP, endIP)\n\tif err != nil {\n\t\tinstance.Status.Message = err.Error()\n\t\tcatch := []string{\n\t\t\terrhelp.AsyncOpIncompleteError,\n\t\t\terrhelp.ResourceGroupNotFoundErrorCode,\n\t\t\terrhelp.ParentNotFoundErrorCode,\n\t\t\terrhelp.AlreadyExists,\n\t\t\terrhelp.ResourceNotFound,\n\t\t}\n\t\tazerr := errhelp.NewAzureError(err)\n\t\tif helpers.ContainsString(catch, azerr.Type) {\n\t\t\treturn false, nil\n\t\t}\n\t\treturn false, err\n\t}\n\treturn false, nil\n}", "func (s *SecurityGroup) AddRule(ctx context.Context, rule *SecurityGroupRule) (*SecurityGroupRule, error) {\n\tvar icmp *struct {\n\t\tCode *int64 `json:\"code,omitempty\"`\n\t\tType *int64 `json:\"type,omitempty\"`\n\t}\n\n\tif rule.Protocol == \"icmp\" || rule.Protocol == \"icmpv6\" {\n\t\ticmpCode := int64(rule.ICMPCode)\n\t\ticmpType := int64(rule.ICMPType)\n\n\t\ticmp = &struct {\n\t\t\tCode *int64 `json:\"code,omitempty\"`\n\t\t\tType *int64 `json:\"type,omitempty\"`\n\t\t}{\n\t\t\tCode: &icmpCode,\n\t\t\tType: &icmpType,\n\t\t}\n\t}\n\n\t// The API doesn't return the Security Group rule created directly, so in order to\n\t// return a *SecurityGroupRule corresponding to the new rule we have to manually\n\t// compare the list of rules in the SG before and after the rule creation, and\n\t// identify the rule that wasn't there before.\n\t// Note: in case of multiple rules creation in parallel this technique is subject\n\t// to race condition as we could return an unrelated rule. To prevent this, we\n\t// also compare the protocol/start port/end port parameters of the new rule to the\n\t// ones specified in the input rule parameter.\n\trules := make(map[string]struct{})\n\tfor _, r := range s.Rules {\n\t\trules[r.ID] = struct{}{}\n\t}\n\n\tstartPort := int64(rule.StartPort)\n\tendPort := int64(rule.EndPort)\n\n\tresp, err := s.c.AddRuleToSecurityGroupWithResponse(\n\t\tapiv2.WithZone(ctx, s.zone),\n\t\ts.ID,\n\t\tpapi.AddRuleToSecurityGroupJSONRequestBody{\n\t\t\tDescription: &rule.Description,\n\t\t\tEndPort: &endPort,\n\t\t\tFlowDirection: rule.FlowDirection,\n\t\t\tIcmp: icmp,\n\t\t\tNetwork: func() (v *string) {\n\t\t\t\tif rule.Network != nil {\n\t\t\t\t\tip := rule.Network.String()\n\t\t\t\t\tv = &ip\n\t\t\t\t}\n\t\t\t\treturn\n\t\t\t}(),\n\t\t\tProtocol: &rule.Protocol,\n\t\t\tSecurityGroup: func() (v *papi.SecurityGroupResource) {\n\t\t\t\tif rule.SecurityGroupID != \"\" {\n\t\t\t\t\tv = &papi.SecurityGroupResource{Id: &rule.SecurityGroupID}\n\t\t\t\t}\n\t\t\t\treturn\n\t\t\t}(),\n\t\t\tStartPort: &startPort,\n\t\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tres, err := papi.NewPoller().\n\t\tWithTimeout(s.c.timeout).\n\t\tPoll(ctx, s.c.OperationPoller(s.zone, *resp.JSON200.Id))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tsgUpdated, err := s.c.GetSecurityGroup(ctx, s.zone, *res.(*papi.Reference).Id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Look for an unknown rule: if we find one we hope it's the one we've just created.\n\tfor _, r := range sgUpdated.Rules {\n\t\tif _, ok := rules[r.ID]; !ok && (r.Protocol == rule.Protocol &&\n\t\t\tr.StartPort == rule.StartPort &&\n\t\t\tr.EndPort == rule.EndPort) {\n\t\t\treturn r, nil\n\t\t}\n\t}\n\n\treturn nil, errors.New(\"unable to identify the rule created\")\n}", "func (mr *MockFirewallServiceIfaceMockRecorder) CreatePortForwardingRule(p interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"CreatePortForwardingRule\", reflect.TypeOf((*MockFirewallServiceIface)(nil).CreatePortForwardingRule), p)\n}", "func (r *Router) AddRule(w http.ResponseWriter, req *http.Request, _ httprouter.Params) {\n\ts, rule, err := r.readRule(req)\n\tif err != nil {\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\tw.Write([]byte(err.Error()))\n\t\treturn\n\t}\n\t// Put the rule into the proxy that belongs to the correct service\n\ts.Proxy.AddRule(*rule)\n\t// Everything went well\n\tw.Write([]byte(config.OK))\n\tlog.Debug(\"Added rule\")\n}", "func (f *fakeLB) GetForwardingRule(_ context.Context, _, _ string) (*govultr.ForwardingRule, *http.Response, error) {\n\tpanic(\"implement me\")\n}", "func NewRule(call *CallExpr) *Rule {\n\treturn &Rule{call, \"\"}\n}", "func (client *FirewallRulesClient) createOrUpdateCreateRequest(ctx context.Context, resourceGroupName string, serverName string, firewallRuleName string, parameters FirewallRule, options *FirewallRulesBeginCreateOrUpdateOptions) (*azcore.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DBforPostgreSQL/servers/{serverName}/firewallRules/{firewallRuleName}\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif serverName == \"\" {\n\t\treturn nil, errors.New(\"parameter serverName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{serverName}\", url.PathEscape(serverName))\n\tif firewallRuleName == \"\" {\n\t\treturn nil, errors.New(\"parameter firewallRuleName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{firewallRuleName}\", url.PathEscape(firewallRuleName))\n\treq, err := azcore.NewRequest(ctx, http.MethodPut, azcore.JoinPaths(client.con.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq.Telemetry(telemetryInfo)\n\treqQP := req.URL.Query()\n\treqQP.Set(\"api-version\", \"2017-12-01\")\n\treq.URL.RawQuery = reqQP.Encode()\n\treq.Header.Set(\"Accept\", \"application/json\")\n\treturn req, req.MarshalAsJSON(parameters)\n}", "func (sdk *MockGoSDKClient) GetSQLFirewallRule(ctx context.Context, resourceGroupName string, serverName string, ruleName string) (result sql.FirewallRule, err error) {\n\n\tvar sqlFirewallRule = sql.FirewallRule{\n\t\tResponse: helpers.GetRestResponse(http.StatusCreated),\n\t}\n\n\tsdk.FirewallRule = sqlFirewallRule\n\n\treturn sqlFirewallRule, nil\n}", "func (f *fakeLB) DeleteForwardingRule(_ context.Context, _, _ string) error {\n\tpanic(\"implement me\")\n}", "func NewFirewallRuleListResultIterator (page FirewallRuleListResultPage) FirewallRuleListResultIterator {\n return FirewallRuleListResultIterator{page: page}\n }", "func MakeFirewallRuleSlice() []*FirewallRule {\n\treturn []*FirewallRule{}\n}", "func (fw *IPtables) makeRules(netif FirewallEndpoint) error {\n\tlog.Infof(\"In makeRules() with %s\", netif.GetName())\n\n\tvar err error\n\tfw.u32filter, fw.chainPrefix, err = fw.prepareU32Rules(netif.GetIP())\n\tif err != nil {\n\t\t// TODO need personalized error here, or even panic\n\t\treturn err\n\t}\n\tfw.interfaceName = netif.GetName()\n\n\tfw.chains = append(fw.chains, IPtablesChain{\n\t\tBaseChain: \"INPUT\",\n\t\tDirections: []string{\"i\"},\n\t\tChainName: \"ROMANA-INPUT\",\n\t})\n\tfw.chains = append(fw.chains, IPtablesChain{\n\t\tBaseChain: \"OUTPUT\",\n\t\tDirections: []string{\"o\"},\n\t\tChainName: \"ROMANA-FORWARD-IN\",\n\t})\n\tfw.chains = append(fw.chains, IPtablesChain{\n\t\tBaseChain: \"FORWARD\",\n\t\tDirections: []string{\"i\"},\n\t\tChainName: \"ROMANA-FORWARD-OUT\",\n\t})\n\tfw.chains = append(fw.chains, IPtablesChain{\n\t\tBaseChain: \"FORWARD\",\n\t\tDirections: []string{\"o\"},\n\t\t// Using ROMANA-FORWARD-IN second time to capture both\n\t\t// traffic from host to endpoint and\n\t\t// traffic from endpoint to another endpoint.\n\t\tChainName: \"ROMANA-FORWARD-IN\",\n\t})\n\n\tlog.Infof(\"In makeRules() created chains %v\", fw.chains)\n\treturn nil\n}", "func (mr *MockFirewallServiceIfaceMockRecorder) CreateIpv6FirewallRule(p interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"CreateIpv6FirewallRule\", reflect.TypeOf((*MockFirewallServiceIface)(nil).CreateIpv6FirewallRule), p)\n}", "func (g *Google) getFirewallRule() (*compute.Firewall, error) {\n\treturn g.computeService.Firewalls.Get(g.project, firewallRuleName).Do()\n}", "func (f *FakeInstance) CreateReverseIPv6(_ context.Context, _ string, _ *govultr.ReverseIP) error {\n\tpanic(\"implement me\")\n}", "func (ruleset *DnsForwardingRuleset) createValidations() []func() (admission.Warnings, error) {\n\treturn []func() (admission.Warnings, error){ruleset.validateResourceReferences}\n}", "func (c *Client) CreateAclRule(args *CreateAclRuleArgs) error {\n\treturn bce.NewRequestBuilder(c).\n\t\tWithURL(getURLForAclRule()).\n\t\tWithMethod(http.POST).\n\t\tWithBody(args).\n\t\tWithQueryParamFilter(\"clientToken\", args.ClientToken).\n\t\tDo()\n}", "func NewTopicRule(ctx *pulumi.Context,\n\tname string, args *TopicRuleArgs, opts ...pulumi.ResourceOpt) (*TopicRule, error) {\n\tif args == nil || args.Enabled == nil {\n\t\treturn nil, errors.New(\"missing required argument 'Enabled'\")\n\t}\n\tif args == nil || args.Sql == nil {\n\t\treturn nil, errors.New(\"missing required argument 'Sql'\")\n\t}\n\tif args == nil || args.SqlVersion == nil {\n\t\treturn nil, errors.New(\"missing required argument 'SqlVersion'\")\n\t}\n\tinputs := make(map[string]interface{})\n\tif args == nil {\n\t\tinputs[\"cloudwatchAlarm\"] = nil\n\t\tinputs[\"cloudwatchMetric\"] = nil\n\t\tinputs[\"description\"] = nil\n\t\tinputs[\"dynamodb\"] = nil\n\t\tinputs[\"elasticsearch\"] = nil\n\t\tinputs[\"enabled\"] = nil\n\t\tinputs[\"firehose\"] = nil\n\t\tinputs[\"kinesis\"] = nil\n\t\tinputs[\"lambda\"] = nil\n\t\tinputs[\"name\"] = nil\n\t\tinputs[\"republish\"] = nil\n\t\tinputs[\"s3\"] = nil\n\t\tinputs[\"sns\"] = nil\n\t\tinputs[\"sql\"] = nil\n\t\tinputs[\"sqlVersion\"] = nil\n\t\tinputs[\"sqs\"] = nil\n\t} else {\n\t\tinputs[\"cloudwatchAlarm\"] = args.CloudwatchAlarm\n\t\tinputs[\"cloudwatchMetric\"] = args.CloudwatchMetric\n\t\tinputs[\"description\"] = args.Description\n\t\tinputs[\"dynamodb\"] = args.Dynamodb\n\t\tinputs[\"elasticsearch\"] = args.Elasticsearch\n\t\tinputs[\"enabled\"] = args.Enabled\n\t\tinputs[\"firehose\"] = args.Firehose\n\t\tinputs[\"kinesis\"] = args.Kinesis\n\t\tinputs[\"lambda\"] = args.Lambda\n\t\tinputs[\"name\"] = args.Name\n\t\tinputs[\"republish\"] = args.Republish\n\t\tinputs[\"s3\"] = args.S3\n\t\tinputs[\"sns\"] = args.Sns\n\t\tinputs[\"sql\"] = args.Sql\n\t\tinputs[\"sqlVersion\"] = args.SqlVersion\n\t\tinputs[\"sqs\"] = args.Sqs\n\t}\n\tinputs[\"arn\"] = nil\n\ts, err := ctx.RegisterResource(\"aws:iot/topicRule:TopicRule\", name, true, inputs, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &TopicRule{s: s}, nil\n}", "func (f *FortiWebClient) CreateServerPoolRule(serverPoolName string,\n\tip string,\n\tport int32,\n\tstatus int,\n\tconnectionLimit int) error {\n\n\tbody := map[string]interface{}{\n\t\t\"ip\": ip,\n\t\t\"status\": status,\n\t\t\"port\": port,\n\t\t\"connectLimit\": connectionLimit,\n\t\t\"inHeritHCheck\": true}\n\n\tjsonByte, err := json.Marshal(body)\n\n\tif err != nil {\n\t\tfmt.Printf(\"Error in json data: %s\\n\", err)\n\t\treturn err\n\t}\n\n\tresponse, error := f.DoPost(\n\t\tstrings.Join([]string{\"api/v1.0/ServerObjects/Server/ServerPool/\",\n\t\t\tf.SafeName(serverPoolName),\n\t\t\t\"/EditServerPoolRule\"}, \"\"),\n\t\tstring(jsonByte))\n\n\tif error != nil {\n\t\tfmt.Printf(\"The HTTP request failed with error %s\\n\", error)\n\t\treturn error\n\t}\n\tif response != nil && response.StatusCode != 200 {\n\t\tfmt.Printf(\"The HTTP request failed with HTTP code: %d, %s\\n\", response.StatusCode, response.Status)\n\t}\n\n\treturn nil\n}", "func CreateWebhookListener(config,\n\tremoteConfig *rest.Config,\n\tscheme *runtime.Scheme,\n\ttlsKeyFile, tlsCrtFile string,\n\tcreateService bool) (*WebhookListener, error) {\n\tif klog.V(utils.QuiteLogLel) {\n\t\tfnName := utils.GetFnName()\n\t\tklog.Infof(\"Entering: %v()\", fnName)\n\n\t\tdefer klog.Infof(\"Exiting: %v()\", fnName)\n\t}\n\n\tvar err error\n\n\tdynamicClient := dynamic.NewForConfigOrDie(config)\n\n\tl := &WebhookListener{\n\t\tDynamicClient: dynamicClient,\n\t\tlocalConfig: config,\n\t}\n\n\t// The user-provided key and cert files take precedence over the default provided files if both sets exist.\n\tif _, err := os.Stat(defaultKeyFile); err == nil {\n\t\tl.TLSKeyFile = defaultKeyFile\n\t}\n\n\tif _, err := os.Stat(defaultCrtFile); err == nil {\n\t\tl.TLSCrtFile = defaultCrtFile\n\t}\n\n\tif _, err := os.Stat(tlsKeyFile); err == nil {\n\t\tl.TLSKeyFile = tlsKeyFile\n\t}\n\n\tif _, err := os.Stat(tlsCrtFile); err == nil {\n\t\tl.TLSCrtFile = tlsCrtFile\n\t}\n\n\tl.LocalClient, err = client.New(config, client.Options{})\n\n\tif err != nil {\n\t\tklog.Error(\"Failed to initialize client to update local status. error: \", err)\n\t\treturn nil, err\n\t}\n\n\tl.RemoteClient = l.LocalClient\n\tif remoteConfig != nil {\n\t\tl.RemoteClient, err = client.New(remoteConfig, client.Options{})\n\n\t\tif err != nil {\n\t\t\tklog.Error(\"Failed to initialize client to update remote status. error: \", err)\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tif createService {\n\t\tnamespace, err := getOperatorNamespace()\n\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\t// Create the webhook listener service only when the subscription controller runs in hub mode.\n\t\terr = createWebhookListnerService(l.LocalClient, namespace)\n\n\t\tif err != nil {\n\t\t\tklog.Error(\"Failed to create a service for Git webhook listener. error: \", err)\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn l, err\n}" ]
[ "0.7080871", "0.69296676", "0.68744314", "0.6725503", "0.6292887", "0.61700433", "0.6086539", "0.5966617", "0.59360117", "0.590821", "0.5884807", "0.5869949", "0.5863809", "0.579888", "0.5756061", "0.57383156", "0.5729393", "0.5726847", "0.5677176", "0.5675465", "0.56125", "0.5604363", "0.5580383", "0.5549001", "0.54931605", "0.5489614", "0.547709", "0.54657435", "0.5464037", "0.54541016", "0.54391634", "0.54342556", "0.54108495", "0.540092", "0.5379361", "0.5365434", "0.53347594", "0.5317378", "0.53146976", "0.5306949", "0.5290145", "0.52712166", "0.52701765", "0.52584875", "0.52536786", "0.52444744", "0.5219601", "0.52144057", "0.5206177", "0.5190869", "0.5179683", "0.51742953", "0.51627", "0.51594937", "0.51589686", "0.5150411", "0.5134197", "0.5133311", "0.5130034", "0.5121891", "0.509893", "0.50833684", "0.5073997", "0.50659406", "0.50502974", "0.5044635", "0.50164986", "0.50164026", "0.5013872", "0.50123984", "0.5009241", "0.50066173", "0.5003547", "0.49986768", "0.49972352", "0.49806523", "0.4980328", "0.49787468", "0.4941635", "0.49326417", "0.49246904", "0.4922888", "0.49135435", "0.4902758", "0.4893153", "0.48848763", "0.48797563", "0.48775458", "0.48732615", "0.48496455", "0.4823839", "0.4818339", "0.4810629", "0.48057142", "0.48032248", "0.48029497", "0.48024648", "0.48004216", "0.47996503", "0.4795538" ]
0.7194981
0
CreateImage uses the override method CreateImageFn or the real implementation.
func (c *TestClient) CreateImage(project string, i *compute.Image) error { if c.CreateImageFn != nil { return c.CreateImageFn(project, i) } return c.client.CreateImage(project, i) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (p *OnPrem) CreateImage(ctx *Context) error {\n\treturn fmt.Errorf(\"Operation not supported\")\n}", "func (d *Docker) CreateImage(parentRef string) {}", "func (d *driverMock) CreateImage(ctx context.Context, id string) (core.Image, error) {\n\tif d.CreateImageErr != nil {\n\t\treturn core.Image{}, d.CreateImageErr\n\t}\n\td.CreateImageID = id\n\treturn core.Image{Id: &id}, nil\n}", "func CreateImage(width int, height int, background color.RGBA) *image.RGBA {\n\trect := image.Rect(0, 0, width, height)\n\timg := image.NewRGBA(rect)\n\tdraw.Draw(img, img.Bounds(), &image.Uniform{C: background}, image.ZP, draw.Src)\n\n\treturn img\n}", "func (img *CreateImageInput) CreateImage() (CreateImageResponse, error) {\n\n\tif status := support.DoesCloudSupports(strings.ToLower(img.Cloud.Name)); status != true {\n\t\treturn CreateImageResponse{}, fmt.Errorf(common.DefaultCloudResponse + \"CreateImage\")\n\t}\n\n\tswitch strings.ToLower(img.Cloud.Name) {\n\tcase \"aws\":\n\n\t\tcreds, crderr := common.GetCredentials(&common.GetCredentialsInput{Profile: img.Cloud.Profile, Cloud: img.Cloud.Name})\n\t\tif crderr != nil {\n\t\t\treturn CreateImageResponse{}, crderr\n\t\t}\n\t\t// I will establish session so that we can carry out the process in cloud\n\t\tsession_input := awssess.CreateSessionInput{Region: img.Cloud.Region, KeyId: creds.KeyId, AcessKey: creds.SecretAccess}\n\t\tsess := session_input.CreateAwsSession()\n\n\t\t//authorizing to request further\n\t\tauthinpt := auth.EstablishConnectionInput{Region: img.Cloud.Region, Resource: \"ec2\", Session: sess}\n\n\t\tresponse_image := make([]image.ImageResponse, 0)\n\n\t\tfor _, id := range img.InstanceIds {\n\t\t\timgcreate := new(image.ImageCreateInput)\n\t\t\timgcreate.InstanceId = id\n\t\t\timgcreate.GetRaw = img.Cloud.GetRaw\n\t\t\tresponse, imgerr := imgcreate.CreateImage(authinpt)\n\t\t\tif imgerr != nil {\n\t\t\t\treturn CreateImageResponse{}, imgerr\n\t\t\t}\n\t\t\tresponse_image = append(response_image, response)\n\t\t}\n\t\treturn CreateImageResponse{AwsResponse: response_image}, nil\n\n\tcase \"azure\":\n\t\treturn CreateImageResponse{}, fmt.Errorf(common.DefaultAzResponse)\n\tcase \"gcp\":\n\t\treturn CreateImageResponse{}, fmt.Errorf(common.DefaultGcpResponse)\n\tcase \"openstack\":\n\t\treturn CreateImageResponse{}, fmt.Errorf(common.DefaultOpResponse)\n\tdefault:\n\t\tlog.Info(\"\")\n\t\tlog.Error(common.DefaultCloudResponse + \"CreateImage\")\n\t\tlog.Info(\"\")\n\t\treturn CreateImageResponse{}, fmt.Errorf(common.DefaultCloudResponse + \"CreateImage\")\n\t}\n}", "func (i *LibpodAPI) CreateImage(call ioprojectatomicpodman.VarlinkCall) error {\n\treturn call.ReplyMethodNotImplemented(\"CreateImage\")\n}", "func (g GCPClient) CreateImage(name, storageURL, family string, nested, uefi, replace bool) error {\n\tif replace {\n\t\tif err := g.DeleteImage(name); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tlog.Infof(\"Creating image: %s\", name)\n\timgObj := &compute.Image{\n\t\tRawDisk: &compute.ImageRawDisk{\n\t\t\tSource: storageURL,\n\t\t},\n\t\tName: name,\n\t}\n\n\tif family != \"\" {\n\t\timgObj.Family = family\n\t}\n\n\tif nested {\n\t\timgObj.Licenses = []string{vmxImageLicence}\n\t}\n\n\tif uefi {\n\t\timgObj.GuestOsFeatures = []*compute.GuestOsFeature{\n\t\t\t{Type: uefiCompatibleFeature},\n\t\t}\n\t}\n\n\top, err := g.compute.Images.Insert(g.projectName, imgObj).Do()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err := g.pollOperationStatus(op.Name); err != nil {\n\t\treturn err\n\t}\n\tlog.Infof(\"Image %s created\", name)\n\treturn nil\n}", "func (im *imageContrller) CreateImage(data []color.RGBA, height, width int) *image.RGBA {\n\timg := new(image.RGBA)\n\n\tif height > 0 && width > 0 {\n\t\t// check data size\n\t\tif (height * width) == len(data) {\n\n\t\t\t// create image\n\t\t\tcanvas := image.NewRGBA(image.Rect(0, 0, width, height))\n\t\t\tfor i := 0; i < width; i++ {\n\t\t\t\tfor j := 0; j < height; j++ {\n\t\t\t\t\tindex := width*i + j\n\n\t\t\t\t\t// raw data\n\t\t\t\t\trawData := color.RGBA{\n\t\t\t\t\t\tR: data[index].R,\n\t\t\t\t\t\tG: data[index].G,\n\t\t\t\t\t\tB: data[index].B,\n\t\t\t\t\t\tA: 255,\n\t\t\t\t\t}\n\n\t\t\t\t\t// draw the raw data on canvas\n\t\t\t\t\tcanvas.Set(i, j, rawData)\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// update image\n\t\t\timg = canvas\n\t\t}\n\t}\n\treturn img\n}", "func CreateImage(source string, b []byte) *Image {\n\tcam := System.GetCamera(source)\n\tif cam.Dewarp {\n\t\tb = dewarpFisheye(b)\n\t}\n\n\t// image's stable ID is its hash\n\tpotato := sha256.New()\n\tpotato.Write(b)\n\thandle := hex.EncodeToString(potato.Sum(nil)[:32])\n\n\t// verify that the file doesn't somehow already exist\n\tdiskPath := Repository.dataPath(source, fmt.Sprintf(\"%s.%s\", handle, \"jpg\"))\n\tfi, err := os.Stat(diskPath)\n\tif err != nil {\n\t\tif !os.IsNotExist(err) {\n\t\t\tpanic(err)\n\t\t}\n\t}\n\tif fi != nil {\n\t\tlog.Error(\"Image.CreateImage\", \"file hash collision?! '%s'\", handle)\n\t} else {\n\t\t// write the actual file contents under its computed name\n\t\tif err := ioutil.WriteFile(diskPath, b, 0660); err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t}\n\n\tstat, err := os.Stat(diskPath)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn &Image{\n\t\tHandle: handle,\n\t\tSource: source,\n\t\tTimestamp: stat.ModTime(),\n\t\tHasVideo: false,\n\t}\n}", "func (o GetReposRepoTagOutput) ImageCreate() pulumi.IntOutput {\n\treturn o.ApplyT(func(v GetReposRepoTag) int { return v.ImageCreate }).(pulumi.IntOutput)\n}", "func (c Client) CreateImage(backedUpAt time.Time, anon []byte) (models.Image, error) {\n\tvar image models.Image\n\trequest := routes.CreateImageRequest{BackedUpAt: backedUpAt, Anon: string(anon)}\n\n\tvar payload bytes.Buffer\n\terr := jsonapi.MarshalOnePayloadWithoutIncluded(&payload, &request)\n\tif err != nil {\n\t\treturn image, err\n\t}\n\n\tresp, err := c.post(\"/images\", &payload)\n\tif err != nil {\n\t\treturn image, err\n\t}\n\n\tif resp.StatusCode != http.StatusCreated {\n\t\treturn image, parseError(resp.Body)\n\t}\n\n\terr = jsonapi.UnmarshalPayload(resp.Body, &image)\n\treturn image, err\n}", "func (v *IBM) CreateImage(ctx *lepton.Context, imagePath string) error {\n\t// also worth gzipping\n\n\ticow := imagePath + \".qcow2\"\n\n\targs := []string{\n\t\t\"convert\", \"-f\", \"raw\", \"-O\", \"qcow2\", imagePath, icow,\n\t}\n\n\tcmd := exec.Command(\"qemu-img\", args...)\n\tout, err := cmd.Output()\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\tfmt.Println(out)\n\t}\n\n\tstore := &Objects{\n\t\ttoken: v.iam,\n\t}\n\n\tv.Storage = store\n\terr = v.Storage.CopyToBucket(ctx.Config(), icow)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\timgName := ctx.Config().CloudConfig.ImageName\n\n\tv.createImage(ctx, icow, imgName)\n\treturn nil\n}", "func CreateImage(cli bce.Client, args *CreateImageArgs) (*CreateImageResult, error) {\n\t// Build the request\n\treq := &bce.BceRequest{}\n\treq.SetUri(getImageUri())\n\treq.SetMethod(http.POST)\n\n\tif args.ClientToken != \"\" {\n\t\treq.SetParam(\"clientToken\", args.ClientToken)\n\t}\n\n\tjsonBytes, err := json.Marshal(args)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbody, err := bce.NewBodyFromBytes(jsonBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq.SetBody(body)\n\n\t// Send request and get response\n\tresp := &bce.BceResponse{}\n\tif err := cli.SendRequest(req, resp); err != nil {\n\t\treturn nil, err\n\t}\n\tif resp.IsFail() {\n\t\treturn nil, resp.ServiceError()\n\t}\n\n\tjsonBody := &CreateImageResult{}\n\tif err := resp.ParseJsonBody(jsonBody); err != nil {\n\t\treturn nil, err\n\t}\n\treturn jsonBody, nil\n}", "func NewImage(repoName string, keychain authn.Keychain, ops ...ImageOption) (*Image, error) {\n\timageOpts := &options{}\n\tfor _, op := range ops {\n\t\tif err := op(imageOpts); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tplatform := defaultPlatform()\n\tif (imageOpts.platform != imgutil.Platform{}) {\n\t\tplatform = imageOpts.platform\n\t}\n\n\timage, err := emptyImage(platform)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tri := &Image{\n\t\tkeychain: keychain,\n\t\trepoName: repoName,\n\t\timage: image,\n\t}\n\n\tif imageOpts.prevImageRepoName != \"\" {\n\t\tif err := processPreviousImageOption(ri, imageOpts.prevImageRepoName, platform); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tif imageOpts.baseImageRepoName != \"\" {\n\t\tif err := processBaseImageOption(ri, imageOpts.baseImageRepoName, platform); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\timgOS, err := ri.OS()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif imgOS == \"windows\" {\n\t\tif err := prepareNewWindowsImage(ri); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn ri, nil\n}", "func CreateImageResource(xRefTable *XRefTable, r io.Reader, gray, sepia bool) (*types.IndirectRef, int, int, error) {\n\tsd, w, h, err := CreateImageStreamDict(xRefTable, r, gray, sepia)\n\tif err != nil {\n\t\treturn nil, 0, 0, err\n\t}\n\tindRef, err := xRefTable.IndRefForNewObject(*sd)\n\treturn indRef, w, h, err\n}", "func (s *Section) CreateImage() (*Image, error) {\n\titem := &Image{}\n\terr := s.AppendItem(item)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn item, nil\n}", "func CtrCreateImage(image images.Image) (images.Image, error) {\n\tif err := verifyCtr(); err != nil {\n\t\treturn images.Image{}, fmt.Errorf(\"CtrCreateImage: exception while verifying ctrd client: %s\", err.Error())\n\t}\n\treturn CtrdClient.ImageService().Create(ctrdCtx, image)\n}", "func CreateImage(client *gophercloud.ServiceClient, id string, opts CreateImageOptsBuilder) (r CreateImageResult) {\n\tb, err := opts.ToServerCreateImageMap()\n\tif err != nil {\n\t\tr.Err = err\n\t\treturn\n\t}\n\tresp, err := client.Post(actionURL(client, id), b, nil, &gophercloud.RequestOpts{\n\t\tOkCodes: []int{202},\n\t})\n\t_, r.Header, r.Err = gophercloud.ParseResponse(resp, err)\n\treturn\n}", "func NewImage(repoName string, dockerClient client.CommonAPIClient, ops ...ImageOption) (*Image, error) {\n\timageOpts := &options{}\n\tfor _, op := range ops {\n\t\tif err := op(imageOpts); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tplatform, err := defaultPlatform(dockerClient)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif (imageOpts.platform != imgutil.Platform{}) {\n\t\tif err := validatePlatformOption(platform, imageOpts.platform); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tplatform = imageOpts.platform\n\t}\n\n\tinspect := defaultInspect(platform)\n\n\timage := &Image{\n\t\tdocker: dockerClient,\n\t\trepoName: repoName,\n\t\tinspect: inspect,\n\t\tlayerPaths: make([]string, len(inspect.RootFS.Layers)),\n\t\tdownloadBaseOnce: &sync.Once{},\n\t}\n\n\tif imageOpts.prevImageRepoName != \"\" {\n\t\tif err := processPreviousImageOption(image, imageOpts.prevImageRepoName, platform, dockerClient); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tif imageOpts.baseImageRepoName != \"\" {\n\t\tif err := processBaseImageOption(image, imageOpts.baseImageRepoName, platform, dockerClient); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tif image.inspect.Os == \"windows\" {\n\t\tif err := prepareNewWindowsImage(image); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn image, nil\n}", "func (c *Client) CreateImage(request *CreateImageRequest) (response *CreateImageResponse, err error) {\n return c.CreateImageWithContext(context.Background(), request)\n}", "func (c *ImageController) Create(ctx *app.CreateImageContext) error {\n\t// ImageController_Create: start_implement\n\n\t// Put your logic here\n\n\t// ImageController_Create: end_implement\n\treturn nil\n}", "func (h *Handler) CreateImage(w http.ResponseWriter, r *http.Request) {\n\tvar newImage model.BlockImage\n\tbody, ok := handleReadBody(w, r, \"create image\")\n\tif !ok {\n\t\treturn\n\t}\n\n\tif err := json.Unmarshal(body, &newImage); err != nil {\n\t\tlogger.Errorf(\"failed to unmarshal create image request body '%s': %+v\", string(body), err)\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tif newImage.Name == \"\" || newImage.PoolName == \"\" || newImage.Size == 0 {\n\t\tlogger.Errorf(\"image missing required fields: %+v\", newImage)\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tcreatedImage, err := ceph.CreateImage(h.context, h.config.clusterInfo.Name, newImage.Name,\n\t\tnewImage.PoolName, newImage.Size)\n\tif err != nil {\n\t\tlogger.Errorf(\"failed to create image %+v: %+v\", newImage, err)\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tw.Write([]byte(fmt.Sprintf(\"succeeded created image %s\", createdImage.Name)))\n}", "func CreatedImage(id, ref string, created time.Time) imageapi.Image {\n\timage := ImageWithLayers(id, ref, nil, Layer1, Layer2, Layer3, Layer4, Layer5)\n\timage.CreationTimestamp = metav1.NewTime(created)\n\treturn image\n}", "func (d ImagefsDriver) Create(r *volume.CreateRequest) error {\n\tfmt.Printf(\"-> Create %+v\\n\", r)\n\tsource, ok := r.Options[\"source\"]\n\tif !ok {\n\t\treturn fmt.Errorf(\"no source volume specified\")\n\t}\n\n\t// pull the image\n\t/*readCloser, err := d.cli.ImagePull(context.Background(), source, types.ImagePullOptions{\n\t\t// HACK assume the registry ignores the auth header\n\t\tRegistryAuth: \"null\",\n\t})\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unexpected error: %s\", err)\n\t}\n\tscanner := bufio.NewScanner(readCloser)\n\tfor scanner.Scan() {\n\t}*/\n\n\tcontainerConfig := &container.Config{\n\t\tImage: source,\n\t\tEntrypoint: []string{\"/runtime/loop\"},\n\t\tLabels: map[string]string{\n\t\t\t\"com.docker.imagefs.version\": version,\n\t\t},\n\t\tNetworkDisabled: true,\n\t}\n\n\tif target, ok := r.Options[\"target\"]; ok {\n\t\tcontainerConfig.Labels[\"com.docker.imagefs.target\"] = target\n\t}\n\t// TODO handle error\n\thostConfig := &container.HostConfig{\n\t\tBinds: []string{\"/tmp/runtime:/runtime\"},\n\t\t//AutoRemove: true,\n\t}\n\n\tvar platform *specs.Platform\n\tif platformStr, ok := r.Options[\"platform\"]; ok {\n\t\tif versions.GreaterThanOrEqualTo(d.cli.ClientVersion(), \"1.41\") {\n\t\t\tp, err := platforms.Parse(platformStr)\n\t\t\tif err != nil {\n\t\t\t\treturn errors.Wrap(err, \"error parsing specified platform\")\n\t\t\t}\n\t\t\tplatform = &p\n\t\t}\n\t}\n\n\tnetworkConfig := &network.NetworkingConfig{}\n\tcont, err := d.cli.ContainerCreate(\n\t\tcontext.Background(),\n\t\tcontainerConfig,\n\t\thostConfig,\n\t\tnetworkConfig,\n\t\tplatform,\n\t\t// TODO(rabrams) namespace\n\t\tr.Name,\n\t)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unexpected error: %s\", err)\n\t}\n\tfmt.Printf(\"Temp container ID: %s\", cont.ID)\n\td.cli.ContainerStart(\n\t\tcontext.Background(),\n\t\tcont.ID,\n\t\ttypes.ContainerStartOptions{},\n\t)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unexpected error: %s\", err)\n\t}\n\treturn nil\n}", "func (c *Client) CreateImage(args *CreateImageArgs) (string, error) {\n\tresponse := CreateImageResponse{}\n\terr := c.Invoke(\"CreateImage\", args, &response)\n\tif err == nil {\n\t\treturn response.ImageId, nil\n\t}\n\treturn \"\", err\n}", "func createImageResource(ctx context.Context, d *schema.ResourceData, meta interface{}) diag.Diagnostics {\n\t// Warning or errors can be collected in a slice type\n\tvar diags diag.Diagnostics\n\n\tclient := (meta.(Client)).Client\n\tname := rdEntryStr(d, \"name\")\n\tid := rdEntryStr(d, \"id\")\n\terrMsgPrefix := fmt.Sprintf(\"[ERROR] Image %s (id: %s) Create Failed.\",\n\t\tname, id)\n\tif client == nil {\n\t\treturn diag.Errorf(\"%s nil Client\", errMsgPrefix)\n\t}\n\tcfg := &swagger_models.ImageConfig{\n\t\tName: &name,\n\t}\n\terr := updateImageCfgFromResourceData(cfg, d)\n\tif err != nil {\n\t\treturn diag.Errorf(\"%s %s\", errMsgPrefix, err.Error())\n\t}\n\tlog.Printf(\"[INFO] Creating Image: %s\", name)\n\tclient.XRequestIdPrefix = \"TF-image-create\"\n\trspData := &swagger_models.ZsrvResponse{}\n\t_, err = client.SendReq(\"POST\", imageUrlExtension, cfg, rspData)\n\tif err != nil {\n\t\treturn diag.Errorf(\"%s Err: %s\", errMsgPrefix, err.Error())\n\t}\n\tlog.Printf(\"Image %s (ID: %s) Successfully created\\n\",\n\t\trspData.ObjectName, rspData.ObjectID)\n\td.SetId(rspData.ObjectID)\n\treturn diags\n}", "func (i *ImagesModel) CreateImage(\n\tmetaConstructor *images.SoftwareImageMetaConstructor,\n\timageReader io.Reader) (string, error) {\n\tif metaConstructor == nil {\n\t\treturn \"\", controller.ErrModelMissingInputMetadata\n\t}\n\tif imageReader == nil {\n\t\treturn \"\", controller.ErrModelMissingInputArtifact\n\t}\n\tartifactID, err := i.handleArtifact(metaConstructor, imageReader)\n\t// try to remove artifact file from file storage on error\n\tif err != nil {\n\t\tif cleanupErr := i.fileStorage.Delete(artifactID); cleanupErr != nil {\n\t\t\treturn \"\", errors.Wrap(err, cleanupErr.Error())\n\t\t}\n\t}\n\treturn artifactID, err\n}", "func (g Goba) CreateImage(typ DatabaseType) (*Image, error) {\n\tfor _, handler := range g.handlers {\n\t\tif handler.Type() == typ {\n\t\t\treturn handler.CreateImage()\n\t\t}\n\t}\n\treturn nil, ErrNoSuchHandler\n}", "func newImage(uid int64, imgname string, tag int, des string) CRImage {\n\treturn CRImage{\n\t\tUserId: uid,\n\t\tImageName: imgname,\n\t\tTag: tag,\n\t\tStar: 0,\n\t\tFork: 0,\n\t\tComm: 0,\n\t\tStatus: 0,\n\t\tDescrip: des,\n\t\tDate: time.Now().Format(\"2006-01-02\"),\n\t}\n}", "func (v *ImageClient) Create(c Image) (Image, error) {\n\t//Construct composite key consisting of name\n\tkey := ImageKey{\n\t\tImageName: c.ImageName,\n\t}\n\n\t//Check if this Image already exists\n\t_, err := v.Get(c.ImageName)\n\tif err == nil {\n\t\treturn Image{}, pkgerrors.New(\"Image already exists\")\n\t}\n\n\terr = v.util.DBCreate(v.storeName, key, v.tagMeta, c)\n\tif err != nil {\n\t\treturn Image{}, pkgerrors.Wrap(err, \"Creating DB Entry\")\n\t}\n\n\terr = v.CreateFile(c)\n\tif err != nil {\n\t\treturn Image{}, pkgerrors.Wrap(err, \"Creating File in FS\")\n\t}\n\n\treturn c, nil\n}", "func (p *AWS) CreateImage(ctx *lepton.Context, imagePath string) error {\n\timageName := ctx.Config().CloudConfig.ImageName\n\n\ti, _ := p.findImageByName(imageName)\n\tif i != nil {\n\t\treturn fmt.Errorf(\"failed creating image: image with name %s already exists\", imageName)\n\t}\n\n\tc := ctx.Config()\n\n\tkey := c.CloudConfig.ImageName\n\n\tbar := progressbar.New(100)\n\n\tgo func() {\n\t\tfor {\n\t\t\ttime.Sleep(2 * time.Second)\n\n\t\t\tbar.Add64(1)\n\t\t\tbar.RenderBlank()\n\n\t\t\tif bar.State().CurrentPercent == 99 {\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}()\n\n\tctx.Logger().Info(\"Creating snapshot\")\n\tsnapshotID, err := p.createSnapshot(imagePath)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tbar.Set(100)\n\tbar.Finish()\n\tfmt.Println()\n\n\t// tag the volume\n\ttags, _ := buildAwsTags(c.CloudConfig.Tags, key)\n\n\tctx.Logger().Info(\"Tagging snapshot\")\n\t_, err = p.ec2.CreateTags(&ec2.CreateTagsInput{\n\t\tResources: aws.StringSlice([]string{snapshotID}),\n\t\tTags: tags,\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tt := time.Now().UnixNano()\n\ts := strconv.FormatInt(t, 10)\n\n\tamiName := key + s\n\n\t// register ami\n\tenaSupport := GetEnaSupportForFlavor(c.CloudConfig.Flavor)\n\n\trinput := &ec2.RegisterImageInput{\n\t\tName: aws.String(amiName),\n\t\tArchitecture: aws.String(\"x86_64\"),\n\t\tBlockDeviceMappings: []*ec2.BlockDeviceMapping{\n\t\t\t{\n\t\t\t\tDeviceName: aws.String(\"/dev/sda1\"),\n\t\t\t\tEbs: &ec2.EbsBlockDevice{\n\t\t\t\t\tDeleteOnTermination: aws.Bool(false),\n\t\t\t\t\tSnapshotId: aws.String(snapshotID),\n\t\t\t\t\tVolumeType: aws.String(\"gp2\"),\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\tDescription: aws.String(fmt.Sprintf(\"nanos image %s\", key)),\n\t\tRootDeviceName: aws.String(\"/dev/sda1\"),\n\t\tVirtualizationType: aws.String(\"hvm\"),\n\t\tEnaSupport: aws.Bool(enaSupport),\n\t}\n\n\tctx.Logger().Info(\"Registering image\")\n\tresreg, err := p.ec2.RegisterImage(rinput)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Add name tag to the created ami\n\tctx.Logger().Info(\"Tagging image\")\n\t_, err = p.ec2.CreateTags(&ec2.CreateTagsInput{\n\t\tResources: []*string{resreg.ImageId},\n\t\tTags: tags,\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (client *AWSClient) CreateImage(ctx context.Context, instanceID, name, now string) (string, error) {\n\tresult, err := client.svcEC2.CreateImageWithContext(ctx, &ec2.CreateImageInput{\n\t\tInstanceId: aws.String(instanceID),\n\t\tDescription: aws.String(\"create by go-create-image-backup\"),\n\t\tName: aws.String(fmt.Sprintf(\"%s-%s\", name, now)),\n\t\tNoReboot: aws.Bool(true),\n\t})\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\timageID := *result.ImageId\n\n\tif err := client.svcEC2.WaitUntilImageAvailableWithContext(\n\t\tctx,\n\t\t&ec2.DescribeImagesInput{\n\t\t\tImageIds: []*string{aws.String(imageID)},\n\t\t},\n\t\t[]request.WaiterOption{request.WithWaiterMaxAttempts(120)}...,\n\t); err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn imageID, nil\n}", "func CreateImage(name, size, base, dstDir string) (string, error) {\n\terr := os.MkdirAll(dstDir, os.ModePerm)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tparams := []string{\"qemu-img\", \"create\", \"-f\", \"qcow2\", \"-o\", \"cluster_size=2M\"}\n\tif base != \"\" {\n\t\tb, _ := filepath.Abs(filepath.Join(C.ImageDir, base))\n\t\to := fmt.Sprintf(\"backing_file=%s,backing_fmt=qcow2\", b)\n\t\tparams = append(params, \"-o\")\n\t\tparams = append(params, o)\n\t}\n\tp, _ := filepath.Abs(filepath.Join(dstDir, name+\".qcow2\"))\n\tparams = append(params, p)\n\tparams = append(params, size)\n\n\tlog.Println(\"Creating image: \", params)\n\terr = Execs([][]string{params})\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\terr = checkImageSize(p)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn p, nil\n}", "func (i ImageAnnotatorHandlerFuncs) CreateImageAnnotations(data interface{}, name string, count int) map[string]string {\n\tif i.ImageLabelCreationFunc != nil {\n\t\treturn i.ImageAnnotationCreationFunc(data, name, count)\n\t}\n\treturn make(map[string]string)\n}", "func NewImage(ctx *pulumi.Context,\n\tname string, args *ImageArgs, opts ...pulumi.ResourceOption) (*Image, error) {\n\tif args == nil {\n\t\targs = &ImageArgs{}\n\t}\n\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Image\n\terr := ctx.RegisterResource(\"alicloud:ecs/image:Image\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (img *Image) CreateImage(ctx context.Context) (output *dynamodb.PutItemOutput, err error) {\n\tdb := ctx.Value(\"db\").(*dynamodb.DynamoDB)\n\n\tid, _ := uuid.NewV4()\n\timg.ID = id.String()\n\timg.CreatedAt = time.Now()\n\n\titem, _ := dynamodbattribute.MarshalMap(img)\n\n\toutput, err = db.PutItem(&dynamodb.PutItemInput{\n\t\tTableName: aws.String(\"images\"),\n\t\tItem: item,\n\t})\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn output, err\n}", "func (c *PostImageClient) Create() *PostImageCreate {\n\tmutation := newPostImageMutation(c.config, OpCreate)\n\treturn &PostImageCreate{config: c.config, hooks: c.Hooks(), mutation: mutation}\n}", "func createImage(t *testing.T) (*Image, uint) {\n // Datastore ID 1 means default for image\n\tid, err := CreateImage(imageTpl, 1)\n\tif err != nil {\n\t\tt.Error(err)\n\t}\n\n\t// Get Image by ID\n\timage := NewImage(id)\n\n\terr = image.Info()\n\tif err != nil {\n\t\tt.Error(err)\n\t}\n\n\treturn image, id\n}", "func CreateImageObject(ctx context.Context, project string, name string) (*Image, error) {\n\tclient, err := daisyCompute.NewClient(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar apiImage *api.Image\n\tapiImage, err = client.GetImage(project, name)\n\treturn &Image{apiImage, client, project}, err\n}", "func (c *TestClient) CreateMachineImage(project string, mi *compute.MachineImage) error {\n\tif c.CreateMachineImageFn != nil {\n\t\treturn c.CreateMachineImageFn(project, mi)\n\t}\n\treturn c.client.CreateMachineImage(project, mi)\n}", "func (m *MockBundleImager) CreateImage(arg0 sheaf.Bundle) (v1.Image, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"CreateImage\", arg0)\n\tret0, _ := ret[0].(v1.Image)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func newImage(ctx context.Context, sys *types.SystemContext, s storageReference) (types.ImageCloser, error) {\n\tsrc, err := newImageSource(s)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\timg, err := image.FromSource(ctx, sys, src)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsize, err := src.getSize()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &storageImageCloser{ImageCloser: img, size: size}, nil\n}", "func createImage(w, h int) image.Image {\n\t// create a RGBA image from the sensor\n\tpixels := image.NewRGBA(image.Rect(0, 0, w, h))\n\tn := 0\n\tfor _, i := range grid {\n\t\tcolor := colors[getColorIndex(i)]\n\t\tpixels.Pix[n] = getR(color)\n\t\tpixels.Pix[n+1] = getG(color)\n\t\tpixels.Pix[n+2] = getB(color)\n\t\tpixels.Pix[n+3] = 0xFF // we don't need to use this\n\t\tn = n + 4\n\t}\n\tdest := resize.Resize(360, 0, pixels, resize.Lanczos3)\n\treturn dest\n}", "func (ref openshiftReference) NewImage(certPath string, tlsVerify bool) (types.Image, error) {\n\treturn nil, errors.New(\"Full Image support not implemented for atomic: image names\")\n}", "func NewImage() *Image {\n\treturn &Image{}\n}", "func (t Tasker) ImageCreation(url, assetID string, w, h int) error {\n\tb, err := t.Download(url)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\timg, _, _ := image.Decode(bytes.NewReader(b))\n\n\tanalyzer := smartcrop.NewAnalyzer(nfnt.NewDefaultResizer())\n\ttopCrop, _ := analyzer.FindBestCrop(img, w, h)\n\n\t// The crop will have the requested aspect ratio, but you need to copy/scale it yourself\n\tt.logger.Infof(\"Top crop: %+v\\n\", topCrop)\n\n\ttype SubImager interface {\n\t\tSubImage(r image.Rectangle) image.Image\n\t}\n\tcroppedimg := img.(SubImager).SubImage(topCrop)\n\n\twr, err := t.bucket.NewWriter(context.Background(), assetID+\"/thumbnail.jpg\", nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer wr.Close()\n\n\tif err = jpeg.Encode(wr, croppedimg, &jpeg.Options{Quality: 85}); err != nil {\n\t\treturn err\n\t}\n\n\tthumbnail := opendamclient.Asset{\n\t\tAssetId: uuid.New().String(),\n\t\tKind: \"image\",\n\t\tVersion: opendamclient.Version{},\n\t\tFile: opendamclient.AssetFile{\n\t\t\tName: \"thumbnail.jpg\",\n\t\t\tSource: assetID + \"/thumbnail.jpg\",\n\t\t\tContentType: \"image/jpeg\",\n\t\t\tWidth: float32(w),\n\t\t\tHeight: float32(h),\n\t\t\tSize: 0,\n\t\t},\n\t}\n\tjob, resp, err := t.api.PutAsset(context.Background(), assetID, opendamclient.AssetUpdate{\n\t\tFormats: []opendamclient.Asset{thumbnail},\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\tt.logger.WithFields(logrus.Fields{\n\t\t\"job\": job,\n\t\t\"resp\": resp,\n\t}).Debug(\"asset put response\")\n\treturn nil\n}", "func NewImage(name, kernel, initrd, rootfs string) *Image {\n\treturn &Image{\n\t\tName: name,\n\t\tkernel: kernel,\n\t\tinitrd: initrd,\n\t\trootfs: rootfs,\n\t}\n}", "func NewImage(idd uintptr) *Image {\n\treturn &Image{\n\t\tWindowBase: WindowBase{idd: idd},\n\t\tOnClicked: nil,\n\t}\n}", "func createImage(w http.ResponseWriter, r *http.Request) {\r\n\t//\tvars := mux.Vars(r)\r\n\t//\tid, _ := strconv.ParseInt(vars[\"id\"], 10, 64)\r\n\tvar ni newimage\r\n\tif err := json.NewDecoder(r.Body).Decode(&ni); err != nil {\r\n\t\tlogger.Warnf(\"error decoding image: %s\", err)\r\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\r\n\t\treturn\r\n\t}\r\n\tbi := baseImage{ni.BaseImage}\r\n\tcr := newImage(ni.UserId, ni.ImageName, ni.Tag, ni.Descrip)\r\n\tif err := cr.Add(); err != nil {\r\n\t\tlogger.Warnf(\"error creating image: %s\", err)\r\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\r\n\t\treturn\r\n\t}\r\n\tw.Header().Set(\"content-type\", \"application/json\")\r\n\tw.WriteHeader(http.StatusCreated)\r\n\tif err := json.NewEncoder(w).Encode(bi); err != nil {\r\n\t\tlogger.Error(err)\r\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\r\n\t}\r\n}", "func (c *containerdCAS) CreateImage(reference, blobHash string) error {\n\tindex, err := getIndexManifest(c, blobHash)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"CreateImage: Exception while fetching IndexManifest. %s\", err.Error())\n\t}\n\tmanifest, err := getManifestFromIndex(c, index)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"CreateImage: Exception while fetching Manifest. %s\", err.Error())\n\t}\n\timage := images.Image{\n\t\tName: reference,\n\t\tLabels: nil,\n\t\tTarget: spec.Descriptor{\n\t\t\tMediaType: images.MediaTypeDockerSchema2ManifestList,\n\t\t\tDigest: digest.Digest(blobHash),\n\t\t\tSize: manifest.Config.Size,\n\t\t},\n\t\tCreatedAt: time.Time{},\n\t\tUpdatedAt: time.Time{},\n\t}\n\t_, err = ctrdClient.ImageService().Create(ctrdCtx, image)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"CreateImage: Exception while creating reference: %s. %s\", reference, err.Error())\n\t}\n\treturn nil\n}", "func (i *ImageService) CreateLayer(container *container.Container, initFunc layer.MountInit) (layer.RWLayer, error) {\n\treturn nil, errdefs.NotImplemented(errdefs.NotImplemented(errors.New(\"not implemented\")))\n}", "func BuildImage(ctx context.Context, ow *rpc.OutputWriter, cli *client.Client, opts *docker.BuildImageOpts) Fixer {\n\treturn func() (string, error) {\n\t\tcreated, err := docker.EnsureImage(ctx, ow, cli, opts)\n\t\tif err != nil {\n\t\t\treturn \"failed to create custom image.\", err\n\t\t}\n\t\tif created {\n\t\t\treturn \"custom image already existed.\", nil\n\t\t}\n\t\treturn \"custom image created successfully.\", nil\n\t}\n}", "func NewImage(img image.Image, xmin, ymin, xmax, ymax float64) *Image {\n\tif src, ok := img.(*image.Uniform); ok {\n\t\timg = uniform{\n\t\t\tsrc,\n\t\t\timage.Rect(0, 0, int(xmax-xmin+0.5), int(ymax-ymin+0.5)),\n\t\t}\n\t}\n\tbounds := img.Bounds()\n\tcols := bounds.Dx()\n\trows := bounds.Dy()\n\tdx := math.Abs(xmax-xmin) / float64(cols)\n\tdy := math.Abs(ymax-ymin) / float64(rows)\n\treturn &Image{\n\t\timg: img,\n\t\tcols: cols,\n\t\trows: rows,\n\t\txmin: xmin,\n\t\txmax: xmax,\n\t\tdx: dx,\n\t\tymin: ymin,\n\t\tymax: ymax,\n\t\tdy: dy,\n\t}\n}", "func newImage(page Page, ir *ImageResource, centerX, centerY float64) *image {\n\treturn &image{\n\t\tgraphicsObject: graphicsObject{\n\t\t\tpage: page,\n\t\t},\n\t\tname: ir.name,\n\t\tcenterX: centerX,\n\t\tcenterY: centerY,\n\t\twidth: float64(ir.width),\n\t\theight: float64(ir.height),\n\t}\n}", "func CreateImageFlavor(flavorLabel string, outputFlavorFilePath string, inputImageFilePath string, outputEncImageFilePath string,\n\tkeyID string, integrityRequired bool) (string, error) {\n\tlog.Trace(\"pkg/imageflavor/create_image_flavors.go:CreateImageFlavor() Entering\")\n\tdefer log.Trace(\"pkg/imageflavor/create_image_flavors.go:CreateImageFlavor() Leaving\")\n\n\tvar err error\n\tvar wrappedKey []byte\n\tvar keyUrlString string\n\tencRequired := true\n\timageFilePath := inputImageFilePath\n\n\t//Determine if encryption is required\n\toutputEncImageFilePath = strings.TrimSpace(outputEncImageFilePath)\n\tif len(outputEncImageFilePath) <= 0 {\n\t\tencRequired = false\n\t}\n\n\t// set logger fields\n\tlog = log.WithFields(logrus.Fields{\n\t\t\"flavorLabel\": flavorLabel,\n\t\t\"encryptionRequired\": encRequired,\n\t\t\"integrityrequired\": integrityRequired,\n\t\t\"inputImageFilePath\": inputImageFilePath,\n\t\t\"outputFlavorFilePath\": outputFlavorFilePath,\n\t\t\"outputEncImageFilePath\": keyID,\n\t})\n\n\t//Error if image specified doesn't exist\n\t_, err = os.Stat(inputImageFilePath)\n\tif os.IsNotExist(err) {\n\t\treturn \"\", errors.Wrap(err, \"I/O error reading image file: \"+err.Error())\n\t}\n\n\t//Encrypt the image with the key\n\tif encRequired {\n\t\t// fetch the key to encrypt the image\n\t\twrappedKey, keyUrlString, err = util.FetchKey(keyID, \"\")\n\t\tif err != nil {\n\t\t\treturn \"\", errors.Wrap(err, \"Fetch key failed: \"+err.Error())\n\t\t}\n\t\t// encrypt the image with key retrieved from KBS\n\t\terr = util.Encrypt(inputImageFilePath, consts.EnvelopePrivatekeyLocation, outputEncImageFilePath, wrappedKey)\n\t\tif err != nil {\n\t\t\treturn \"\", errors.Wrap(err, \"Image encryption failed: \"+err.Error())\n\t\t}\n\t\timageFilePath = outputEncImageFilePath\n\t}\n\n\t//Check the encrypted image output file\n\timageFile, err := ioutil.ReadFile(imageFilePath)\n\tif err != nil {\n\t\treturn \"\", errors.Wrap(err, \"I/O Error creating encrypted image file: \"+err.Error())\n\t}\n\n\t//Take the digest of the encrypted image\n\tdigest := sha512.Sum384([]byte(imageFile))\n\n\t//Create image flavor\n\timageFlavor, err := flavor.GetImageFlavor(flavorLabel, encRequired, keyUrlString, base64.StdEncoding.EncodeToString(digest[:]))\n\tif err != nil {\n\t\treturn \"\", errors.Wrap(err, \"Error creating image flavor: \"+err.Error())\n\t}\n\n\t//Marshall the image flavor to a JSON string\n\timageFlavorJSON, err := json.Marshal(imageFlavor)\n\tif err != nil {\n\t\treturn \"\", errors.Wrap(err, \"Error while marshalling image flavor: \"+err.Error())\n\t}\n\n\tsignedFlavor, err := flavorUtil.GetSignedFlavor(string(imageFlavorJSON), consts.FlavorSigningKeyPath)\n\tif err != nil {\n\t\treturn \"\", errors.Wrap(err, \"Error signing flavor for image: \"+err.Error())\n\t}\n\n\tlog.Info(\"pkg/imageflavor/create_image_flavors.go:CreateImageFlavor() Successfully created image flavor\")\n\tlog.Debugf(\"pkg/imageflavor/create_image_flavors.go:CreateImageFlavor() Successfully created image flavor %s\", signedFlavor)\n\n\t//If no output flavor file path was specified, return the marshalled image flavor\n\tif len(strings.TrimSpace(outputFlavorFilePath)) <= 0 {\n\t\treturn signedFlavor, nil\n\t}\n\n\t//Otherwise, write it to the specified file\n\terr = ioutil.WriteFile(outputFlavorFilePath, []byte(signedFlavor), 0600)\n\tif err != nil {\n\t\treturn \"\", errors.Wrapf(err, \"I/O Error writing image flavor to output file %s\", outputFlavorFilePath)\n\t}\n\n\tlog.Info(\"pkg/imageflavor/create_image_flavors.go:CreateImageFlavor() Successfully wrote image flavor to file\")\n\treturn \"\", nil\n}", "func (m *Manager) CreateImage(reader io.Reader) (string, *schema.ImageManifest, error) {\n\thr := hashutil.NewSha512(reader)\n\tf, err := tempfile.New(hr)\n\tif err != nil {\n\t\treturn \"\", nil, err\n\t}\n\tdefer f.Close()\n\n\thash := fmt.Sprintf(\"sha512-%s\", hr.Sha512())\n\n\t// double check we don't already have it\n\tm.imagesLock.RLock()\n\tmanifest, exists := m.images[hash]\n\tm.imagesLock.RUnlock()\n\tif exists {\n\t\treturn hash, manifest, nil\n\t}\n\n\tdest := filepath.Join(m.Options.Directory, hash)\n\tif err := os.Mkdir(dest, os.FileMode(0755)); err != nil {\n\t\treturn \"\", nil, err\n\t}\n\n\tsuccessful := false\n\tdefer func() {\n\t\tif !successful {\n\t\t\tos.RemoveAll(dest)\n\t\t}\n\t}()\n\n\tfi, err := os.Stat(dest)\n\tif err != nil {\n\t\treturn \"\", nil, err\n\t}\n\n\t// untar the file\n\ttarfile := tarhelper.NewUntar(f, dest)\n\ttarfile.PreserveOwners = true\n\ttarfile.PreservePermissions = true\n\ttarfile.Compression = tarhelper.DETECT\n\ttarfile.AbsoluteRoot = dest\n\tif err := tarfile.Extract(); err != nil {\n\t\treturn \"\", nil, fmt.Errorf(\"failed to extract image filesystem: %v\", err)\n\t}\n\n\t// load the manifest and return it\n\tmanifest, err = m.loadFile(fi)\n\tif err != nil {\n\t\treturn \"\", nil, err\n\t}\n\tsuccessful = true\n\treturn hash, manifest, nil\n}", "func UserImageCreateService(userid string, label string, name string) (*domain.Image, error) {\n\timage, err := domain.UserCreateItem(userid, label, name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn image, nil\n}", "func MakeImage(svc ec2iface.EC2API, description, instanceID, name *string) (*ec2.CreateImageOutput, error) {\n // snippet-start:[ec2.go.create_image.call]\n opts := &ec2.CreateImageInput{\n Description: description,\n InstanceId: instanceID,\n Name: name,\n BlockDeviceMappings: []*ec2.BlockDeviceMapping{\n {\n DeviceName: aws.String(\"/dev/sda1\"),\n NoDevice: aws.String(\"\"),\n },\n {\n DeviceName: aws.String(\"/dev/sdb\"),\n NoDevice: aws.String(\"\"),\n },\n {\n DeviceName: aws.String(\"/dev/sdc\"),\n NoDevice: aws.String(\"\"),\n },\n },\n }\n resp, err := svc.CreateImage(opts)\n // snippet-end:[ec2.go.create_image.call]\n if err != nil {\n return nil, err\n }\n\n return resp, nil\n}", "func Create(ctx context.Context, dbConn *db.DB, rbmq *rabbitmq.RabbitMQ, cm *CreateImage) (*Image, error) {\n\tparams := []interface{}{cm.Title, cm.URL, cm.Slug, cm.Publisher}\n\tquery := \"INSERT INTO images(title, url, slug, publisher) VALUES($1,$2,$3,$4) RETURNING *\"\n\trow, err := dbConn.PSQLQueryRawx(ctx, query, params...)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, fmt.Sprintf(\"db.images.insert(%s)\", db.Query(query)))\n\t}\n\tvar img Image\n\tif err = row.StructScan(&img); err != nil {\n\t\treturn nil, errors.Wrap(err, fmt.Sprintf(\"db.images.insert(%s)StructScan\", db.Query(query)))\n\t}\n\tqn := \"image_created\"\n\timgJSON, err := json.Marshal(&img)\n\tif err != nil {\n\t\tlog.Warnf(\"RabbitMQ: failed to marshal obj %v for image creation: %v\", &img, err)\n\t}\n\trbmq.DeclareQueue(qn)\n\tif err != nil {\n\t\tlog.Warnf(\"RabbitMQ: cannot declare queue for image creation: %v\", err)\n\t}\n\terr = rbmq.Publish(&qn, imgJSON)\n\tif err != nil {\n\t\tlog.Warnf(\"RabbitMQ: failed to publish a message for image creation: %v\", err)\n\t}\n\treturn &img, nil\n}", "func AutoCreate() (image.Image, string) {\n\tnum := Calc.Rand(10000000, 99999999)\n\timg, err := ManualCreate(4, Calc.Int2String(num))\n\tif err != nil {\n\t\tlog.Print(err)\n\t}\n\treturn img, Calc.Int2String(num)\n}", "func CreatePng(filename string, f ComplexFunc, n int) (err error) {\n\t//create the file that will hold the image\n\tfile, err := os.Create(filename)\n\t//check for errors.\n\tif err != nil {\n\t\treturn\n\t}\n\t//when evertyhing else in this method is finished - close file\n\tdefer file.Close()\n\t//make the return variable by encoding the file using the function f and size n\n\terr = png.Encode(file, Julia(f, n))\n\n\treturn\n}", "func (client *Client) CreateVirtualImage(req *Request) (*Response, error) {\n\treturn client.Execute(&Request{\n\t\tMethod: \"POST\",\n\t\tPath: VirtualImagesPath,\n\t\tQueryParams: req.QueryParams,\n\t\tBody: req.Body,\n\t\tResult: &CreateVirtualImageResult{},\n\t})\n}", "func NewImage(imageFiles []ImageFile) (Image, error) {\n\treturn newImage(imageFiles, false)\n}", "func NewImage(img model.Image, cfg *config.Giffy) Image {\n\tif cfg.Meta.IsProdlike() && len(cfg.CloudFrontDNS) > 0 {\n\t\treturn Image{\n\t\t\tImage: img,\n\t\t\tS3ReadURL: fmt.Sprintf(\"https://%s/%s\", cfg.CloudFrontDNS, img.S3Key),\n\t\t}\n\t}\n\treturn Image{\n\t\tImage: img,\n\t\tS3ReadURL: fmt.Sprintf(\"https://%s.s3.%s.amazonaws.com/%s\", img.S3Bucket, cfg.Aws.RegionOrDefault(), img.S3Key),\n\t}\n}", "func CreateCImage(config *ImageConfig) *CImage {\n\tr := new(CImage)\n\tr.Paletted = image.NewPaletted(image.Rect(0, 0, config.Width, config.Height), randomPalette())\n\tr.config = config\n\tif nil == r.config.fontManager {\n\t\tfm := CreateFontManager()\n\t\tfor _, fontFile := range config.FontFiles {\n\t\t\tfm.AddFont(fontFile)\n\t\t}\n\t\tr.config.fontManager = fm\n\t}\n\n\treturn r\n}", "func NewImage(d *Docker, name string) *Image {\n\treturn &Image{\n\t\tdocker: d,\n\t\tname: name,\n\t}\n}", "func CreateProductImage(c buffalo.Context) error {\n\timages := &models.Images{}\n\n\ttx, ok := c.Value(\"tx\").(*pop.Connection)\n\tif !ok {\n\t\treturn errors.New(\"Transaction not found\")\n\t}\n\tif err := c.Bind(images); err != nil {\n\t\treturn errors.WithStack(err)\n\t}\n\n\tif err := tx.Create(images); err != nil {\n\t\treturn errors.WithStack(err)\n\t}\n\treturn c.Render(201, r.JSON(images))\n}", "func Image(image string) CreateOption {\n\treturn func(c *CreateOptions) {\n\t\tc.image = image\n\t}\n}", "func (mr *MockBundleImagerMockRecorder) CreateImage(arg0 interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"CreateImage\", reflect.TypeOf((*MockBundleImager)(nil).CreateImage), arg0)\n}", "func (ref ostreeReference) NewImage(ctx context.Context, sys *types.SystemContext) (types.ImageCloser, error) {\n\treturn image.FromReference(ctx, sys, ref)\n}", "func (i ImageAnnotatorHandlerFuncs) CreateImageLabels(data interface{}, name string, count int) map[string]string {\n\tif i.ImageLabelCreationFunc != nil {\n\t\treturn i.ImageLabelCreationFunc(data, name, count)\n\t}\n\treturn make(map[string]string)\n}", "func newImage(vres *artworksviews.ImageView) *Image {\n\tres := &Image{}\n\tif vres.ImageID != nil {\n\t\tres.ImageID = *vres.ImageID\n\t}\n\tif vres.ExpiresIn != nil {\n\t\tres.ExpiresIn = *vres.ExpiresIn\n\t}\n\treturn res\n}", "func (d *Driver) Create() error {\n\tif err := d.verifyRootPermissions(); err != nil {\n\t\treturn err\n\t}\n\n\tif err := copyFile(d.ImageSourcePath, d.getDiskPath()); err != nil {\n\t\treturn err\n\t}\n\n\tif _, err := d.resizeDiskImageIfNeeded(d.DiskCapacity); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (self *GameObjectCreator) ImageI(args ...interface{}) *Image{\n return &Image{self.Object.Call(\"image\", args)}\n}", "func CreateImage(d []Drawer, g GlobalConfig) {\n\t// Setup()\n\n\t// 先进行图片加载\n\t// TODO: 需要观察图片是否被正确加载\n\tprepareImage(d)\n\n\tfmt.Println(\"create image\", len(d), g)\n\n\tc := Canvas{canvas.New(g.Width, g.Height)}\n\n\tctx := canvas.NewContext(c)\n\n\t// to draw a red background\n\tctx.SetFillColor(color.RGBA{0xff, 0x00, 0x00, 0xff})\n\tctx.DrawPath(0, 0, canvas.Rectangle(g.Width, g.Height))\n\n\t// 对*canvas.Draw函数传入绘图数据\n\tc.Draw(d)\n\n\t// SavePNG的第二个参数是canvas导出时放大的倍数\n\t// 尽量导出2x或者3x的尺寸,但坐标是1x的,需要更多测试\n\tc.WriteFile(g.FileName, rasterizer.PNGWriter(1.0))\n\n\tfmt.Println(\"draw_data_finish\", len(d))\n}", "func CloudInitFatCreateImage(outputFile *os.File, size int64, inputFiles []CIFFile) error {\n\n\tif err := outputFile.Truncate(size); err != nil {\n\t\toutputFile.Close()\n\t\tos.Remove(outputFile.Name())\n\t\treturn err\n\t}\n\n\t// BlockDevice backed by a file\n\tdevice, err := fs.NewFileDisk(outputFile)\n\tif err != nil {\n\t\toutputFile.Close()\n\t\tos.Remove(outputFile.Name())\n\t\treturn err\n\t}\n\n\t// Format the block device so it contains a valid FAT filesystem\n\tformatConfig := &fat.SuperFloppyConfig{\n\t\tFATType: fat.FAT12,\n\t\tLabel: \"cidata\",\n\t\tOEMName: \"cidata\",\n\t}\n\n\tif fat.FormatSuperFloppy(device, formatConfig); err != nil {\n\t\toutputFile.Close()\n\t\tos.Remove(outputFile.Name())\n\t\treturn err\n\t}\n\n\tfilesys, err := fat.New(device)\n\tif err != nil {\n\t\toutputFile.Close()\n\t\tos.Remove(outputFile.Name())\n\t\treturn err\n\t}\n\n\trootDir, err := filesys.RootDir()\n\tif err != nil {\n\t\toutputFile.Close()\n\t\tos.Remove(outputFile.Name())\n\t\treturn err\n\t}\n\n\tfor _, file := range inputFiles {\n\t\terr = cloudInitFatAddFile(rootDir, file)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\toutputFile.Close()\n\treturn nil\n}", "func CreateRandomImage(namespace, name string) (*imageapiv1.Image, error) {\n\tconst layersCount = 2\n\n\tlayersDescs := make([]distribution.Descriptor, layersCount)\n\tfor i := range layersDescs {\n\t\t_, desc, err := MakeRandomLayer()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tlayersDescs[i] = desc\n\t}\n\n\tmanifest, err := MakeSchema1Manifest(\"unused-name\", \"unused-tag\", layersDescs)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t_, manifestSchema1, err := manifest.Payload()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn NewImageForManifest(\n\t\tfmt.Sprintf(\"%s/%s\", namespace, name),\n\t\tstring(manifestSchema1),\n\t\t\"\",\n\t\tfalse,\n\t)\n}", "func CreatePicture(camera *Camera, width int, height int) *Picture {\r\n\tp := &Picture{\r\n\t\tpixelWidth: width,\r\n\t\tpixelHeight: height,\r\n\t\tpixelWidthDistance: camera.GetWidth() / float64(width),\r\n\t\tpixelHeightDistance: camera.GetHeight() / float64(height),\r\n\t\tcamera: camera}\r\n\r\n\treturn p\r\n}", "func NewImage(width, height int, volatile bool) *Image {\n\ti := newImageWithoutInit(width, height, volatile)\n\ti.ReplacePixels(nil, 0, 0, width, height)\n\treturn i\n}", "func CreateBlankImage(dest string, size resource.Quantity, preallocate bool) error {\n\tklog.V(1).Infof(\"creating raw image with size %s, preallocation %v\", size.String(), preallocate)\n\treturn qemuIterface.CreateBlankImage(dest, size, preallocate)\n}", "func newImageWithoutInit(width, height int, volatile bool) *Image {\n\ti := &Image{\n\t\timage: graphics.NewImage(width, height),\n\t\tvolatile: volatile,\n\t}\n\ttheImages.add(i)\n\treturn i\n}", "func (p *PodmanTestIntegration) BuildImage(dockerfile, imageName string, layers string) {\n\t// TODO\n}", "func (s *REST) Create(ctx kapi.Context, obj runtime.Object) (runtime.Object, error) {\n\tif err := rest.BeforeCreate(Strategy, ctx, obj); err != nil {\n\t\treturn nil, err\n\t}\n\n\tmapping := obj.(*api.ImageStreamMapping)\n\n\tstream, err := s.findStreamForMapping(ctx, mapping)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\timage := mapping.Image\n\ttag := mapping.Tag\n\tif len(tag) == 0 {\n\t\ttag = api.DefaultImageTag\n\t}\n\n\tif err := s.imageRegistry.CreateImage(ctx, &image); err != nil && !errors.IsAlreadyExists(err) {\n\t\treturn nil, err\n\t}\n\n\tnext := api.TagEvent{\n\t\tCreated: util.Now(),\n\t\tDockerImageReference: image.DockerImageReference,\n\t\tImage: image.Name,\n\t}\n\n\tif !api.AddTagEventToImageStream(stream, tag, next) {\n\t\t// nothing actually changed\n\t\treturn &kapi.Status{Status: kapi.StatusSuccess}, nil\n\t}\n\n\tapi.UpdateTrackingTags(stream, tag, next)\n\n\tif _, err := s.imageStreamRegistry.UpdateImageStreamStatus(ctx, stream); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &kapi.Status{Status: kapi.StatusSuccess}, nil\n}", "func CreateEmptyImage(t *testing.T, client *gophercloud.ServiceClient) (*images.Image, error) {\n\tvar image *images.Image\n\n\tname := tools.RandomString(\"ACPTTEST\", 16)\n\tt.Logf(\"Attempting to create image: %s\", name)\n\n\tprotected := false\n\tvisibility := images.ImageVisibilityPrivate\n\tcreateOpts := &images.CreateOpts{\n\t\tName: name,\n\t\tContainerFormat: \"bare\",\n\t\tDiskFormat: \"qcow2\",\n\t\tMinDisk: 0,\n\t\tMinRAM: 0,\n\t\tProtected: &protected,\n\t\tVisibility: &visibility,\n\t\tProperties: map[string]string{\n\t\t\t\"architecture\": \"x86_64\",\n\t\t},\n\t\tTags: []string{\"foo\", \"bar\", \"baz\"},\n\t}\n\n\timage, err := images.Create(client, createOpts).Extract()\n\tif err != nil {\n\t\treturn image, err\n\t}\n\n\tt.Logf(\"Created image %s: %#v\", name, image)\n\treturn image, nil\n}", "func CreateSpeakerImage(speaker, city, year string) (imageFile string) {\n\tui := &input.UI{}\n\tsrcPath, err := ui.Ask(\"Path to speaker image. Must be a PNG or JPG file.\", &input.Options{\n\t\tRequired: true,\n\t\tLoop: true,\n\t\tHideOrder: true,\n\t\tValidateFunc: func(s string) error {\n\t\t\tif (s != \"\") && (helpers.ValidateField(s, \"filepath\") != true) {\n\t\t\t\treturn fmt.Errorf(\"please enter a proper path\")\n\t\t\t}\n\n\t\t\tif _, err := os.Stat(s); err == nil {\n\t\t\t\treturn fmt.Errorf(\"File not found.\")\n\t\t\t}\n\n\t\t\treturn nil\n\t\t},\n\t})\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\t// create the destPath better here\n\n\tre := regexp.MustCompile(\"\\\\.[^.]+$\")\n\text := strings.ToLower(re.FindString(srcPath))\n\tswitch ext {\n\tcase \"jpg\":\n\t\ts := []string{strings.TrimSpace(speaker), \".jpg\"}\n\t\tdestPath := filepath.Join(helpers.EventContentPath(city, year), \"speakers\", strings.Join(s, \"\"))\n\t\thelpers.ResizeImage(srcPath, destPath, \"jpg\", 600, 600)\n\t\treturn strings.Join(s, \"\")\n\tcase \"jpeg\":\n\t\ts := []string{strings.TrimSpace(speaker), \".jpg\"}\n\t\tdestPath := filepath.Join(helpers.EventContentPath(city, year), \"speakers\", strings.Join(s, \"\"))\n\t\thelpers.ResizeImage(srcPath, destPath, \"jpg\", 600, 600)\n\t\treturn strings.Join(s, \"\")\n\tcase \"png\":\n\t\ts := []string{strings.TrimSpace(speaker), \".png\"}\n\t\tdestPath := filepath.Join(helpers.EventContentPath(city, year), \"speakers\", strings.Join(s, \"\"))\n\t\thelpers.ResizeImage(srcPath, destPath, \"png\", 600, 600)\n\t\treturn strings.Join(s, \"\")\n\t}\n\treturn \"\"\n}", "func newImageResource(name string, width, height int, bitperComponent int, data []byte) *ImageResource {\n\treturn &ImageResource{\n\t\tname: name,\n\t\twidth: width,\n\t\theight: height,\n\t\tcolorSpace: colorSpaceDeviceGray,\n\t\tbitsPerComponent: bitperComponent,\n\t\tdata: data,\n\t}\n}", "func (c MockDockerClient) ImageExists(ctx context.Context, imageName string) bool {\n\tif c.ImageExistsFn != nil {\n\t\tfmt.Println(\"[MockDockerClient] In \", utils.CurrentFunctionName())\n\t\tfmt.Println(\"[MockDockerClient] - ctx: \", ctx)\n\t\tfmt.Println(\"[MockDockerClient] - imageName: \", imageName)\n\t\treturn c.ImageExistsFn(ctx, imageName)\n\t}\n\tpanic(fmt.Sprintf(\"No function defined for: %s\", utils.CurrentFunctionName()))\n}", "func NewImage(userID uint, thumb string, lowRes string, highRes string, thumbUUID uuid.UUID, LowResUUID uuid.UUID, HighResUUID uuid.UUID) *Image {\n\treturn &Image{\n\t\tUserID: userID,\n\t\tThumbURL: thumb,\n\t\tLowResURL: lowRes,\n\t\tHighResURL: highRes,\n\t\tThumbUUID: thumbUUID,\n\t\tLowResUUID: LowResUUID,\n\t\tHighResUUID: HighResUUID,\n\t}\n}", "func (im *imageContrller) CreateSolidImage(data color.RGBA, height, width int) *image.RGBA {\n\timg := new(image.RGBA)\n\n\tif height > 0 && width > 0 {\n\n\t\t// create image\n\t\tcanvas := image.NewRGBA(image.Rect(0, 0, width, height))\n\t\tfor i := 0; i < width; i++ {\n\t\t\tfor j := 0; j < height; j++ {\n\n\t\t\t\t// draw the raw data on canvas\n\t\t\t\tcanvas.Set(i, j, data)\n\t\t\t}\n\t\t}\n\n\t\t// update image\n\t\timg = canvas\n\n\t}\n\n\treturn img\n}", "func create() cli.ActionFunc {\n\treturn func(ctx *cli.Context) error {\n\t\tcreds := gobaclient.Credentials{\n\t\t\tUsername: ctx.Parent().GlobalString(\"username\"),\n\t\t\tPassword: ctx.Parent().GlobalString(\"password\"),\n\t\t}\n\n\t\turi, err := url.Parse(ctx.Parent().GlobalString(\"uri\"))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\ttyp := goba.DatabaseType(ctx.Parent().GlobalString(\"database-type\"))\n\n\t\timage, err := gobaclient.CreateImage(*uri, creds, typ)\n\t\tif err == nil {\n\t\t\tprintImageNames(*image)\n\t\t}\n\t\treturn err\n\t}\n}", "func (s *DataStore) CreateEngineImage(img *longhorn.EngineImage) (*longhorn.EngineImage, error) {\n\tret, err := s.lhClient.LonghornV1beta2().EngineImages(s.namespace).Create(context.TODO(), img, metav1.CreateOptions{})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif SkipListerCheck {\n\t\treturn ret, nil\n\t}\n\n\tobj, err := verifyCreation(ret.Name, \"engine image\", func(name string) (runtime.Object, error) {\n\t\treturn s.getEngineImageRO(name)\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tret, ok := obj.(*longhorn.EngineImage)\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"BUG: datastore: verifyCreation returned wrong type for engine image\")\n\t}\n\n\treturn ret.DeepCopy(), nil\n}", "func (self *GameObjectCreator) Image(x int, y int, key interface{}) *Image{\n return &Image{self.Object.Call(\"image\", x, y, key)}\n}", "func Create(p *CreateParam) error {\n\tvar a C.struct_ploop_create_param\n\n\t// default image file name\n\tif p.file == \"\" {\n\t\tp.file = \"root.hdd\"\n\t}\n\n\ta.size = convertSize(p.size)\n\ta.mode = C.int(p.mode)\n\ta.image = C.CString(p.file)\n\tdefer cfree(a.image)\n\ta.fstype = C.CString(\"ext4\")\n\tdefer cfree(a.fstype)\n\n\tret := C.ploop_create_image(&a)\n\treturn mkerr(ret)\n}", "func (p *PdfiumImplementation) FPDFBitmap_Create(request *requests.FPDFBitmap_Create) (*responses.FPDFBitmap_Create, error) {\n\tp.Lock()\n\tdefer p.Unlock()\n\n\tbitmap := C.FPDFBitmap_Create(C.int(request.Width), C.int(request.Height), C.int(request.Alpha))\n\tbitmapHandle := p.registerBitmap(bitmap)\n\n\treturn &responses.FPDFBitmap_Create{\n\t\tBitmap: bitmapHandle.nativeRef,\n\t}, nil\n}", "func NewImage(name string, project string) *Image {\n\treturn &Image{\n\t\t&compute.Image{Name: name, Labels: make(map[string]string)},\n\t\tproject,\n\t}\n}", "func create_init_img(srcImg image.Image, img_dbl bool, sigma float64) image.Image {\n\tvar sig_diff float64\n\tgreyImg := GrayImage(srcImg)\n\tif img_dbl {\n\t\tgreyImg = ResizeImageDouble(greyImg)\n\t\tsig_diff = (sigma*sigma - SIFT_INIT_SIGMA*SIFT_INIT_SIGMA*4)\n\t} else {\n\t\tsig_diff = (sigma*sigma - SIFT_INIT_SIGMA*SIFT_INIT_SIGMA)\n\t}\n\tgreyImg = resize.GaussianSmooth(greyImg, sig_diff, 3)\n\treturn greyImg\n}", "func CreateImageHTTP(rootDir, namespace string) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tfile, _, err := r.FormFile(\"image\")\n\t\tif err != nil {\n\t\t\tfmt.Println(err)\n\t\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\t\tdefer file.Close()\n\n\t\ts := newS3FileUploader(\"cm-personal-site-v2-bucket\", len(defaultBreakpoints))\n\t\t//writer := writer{}\n\t\tir := newImageResizeService(defaultBreakpoints, namespace, s)\n\n\t\tpath, err := ir.saveImageAllSizesUUID(file, rootDir)\n\t\tif err != nil {\n\t\t\tfmt.Println(err)\n\t\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\n\t\terr = s.getUploadErr()\n\t\tif err != nil {\n\t\t\tfmt.Println(err)\n\t\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\n\t\tfmt.Println(\"No errors uploading to \", path)\n\n\t\tjsonResp := struct {\n\t\t\tPath string `json:\"path\"`\n\t\t}{\n\t\t\tPath: path,\n\t\t}\n\n\t\tfmt.Println(jsonResp)\n\n\t\t//w.Header().Add(\"Content-Type\", \"application/json\")\n\t\tjson.NewEncoder(w).Encode(jsonResp)\n\t}\n}", "func NewImage(vres *artworksviews.Image) *Image {\n\treturn newImage(vres.Projected)\n}", "func (c *FakeImageSignatures) Create(ctx context.Context, imageSignature *v1.ImageSignature, opts metav1.CreateOptions) (result *v1.ImageSignature, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewRootCreateAction(imagesignaturesResource, imageSignature), &v1.ImageSignature{})\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\treturn obj.(*v1.ImageSignature), err\n}", "func (img *Image) createThumbnail(opt parseOptions) *image.NRGBA {\n\treturn imaging.Thumbnail(img.Image, opt.width, opt.height, opt.filter)\n}" ]
[ "0.7250976", "0.7120547", "0.68815017", "0.6879296", "0.6770811", "0.67653894", "0.65842426", "0.6559869", "0.65506613", "0.64892596", "0.64121974", "0.63556874", "0.6354497", "0.6282587", "0.62618643", "0.6247416", "0.6237735", "0.62349695", "0.61861354", "0.6158404", "0.61310697", "0.61279535", "0.61104834", "0.6108112", "0.60958636", "0.60709625", "0.6068229", "0.60645604", "0.6059382", "0.60396886", "0.6033249", "0.60331315", "0.6029301", "0.60168046", "0.60109645", "0.59960157", "0.5942701", "0.5930302", "0.59272885", "0.59179157", "0.5911607", "0.5910031", "0.5889946", "0.5889295", "0.58707756", "0.58654", "0.58521146", "0.5849136", "0.58398384", "0.58318824", "0.581281", "0.58029795", "0.5788548", "0.5763067", "0.575872", "0.5757556", "0.5742054", "0.57357574", "0.5715014", "0.5707491", "0.57041633", "0.5701021", "0.56916136", "0.5688619", "0.5643336", "0.5632043", "0.56316006", "0.5627223", "0.5626928", "0.56169826", "0.561109", "0.56016177", "0.5597713", "0.5583232", "0.55671746", "0.554943", "0.5545804", "0.5528333", "0.55224663", "0.55043495", "0.5485958", "0.5451265", "0.5442049", "0.54383254", "0.54354507", "0.54162693", "0.54129493", "0.53993446", "0.5366842", "0.53647465", "0.53592277", "0.5354986", "0.532346", "0.5311434", "0.5297278", "0.528743", "0.5283973", "0.52724695", "0.5269045", "0.5254662" ]
0.6677079
6
CreateInstance uses the override method CreateInstanceFn or the real implementation.
func (c *TestClient) CreateInstance(project, zone string, i *compute.Instance) error { if c.CreateInstanceFn != nil { return c.CreateInstanceFn(project, zone, i) } return c.client.CreateInstance(project, zone, i) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewCreateInstance(name string, plan string, team string) *CreateInstance {\n\tthis := CreateInstance{}\n\tthis.Name = name\n\tthis.Plan = plan\n\tthis.Team = team\n\treturn &this\n}", "func New(fn func() error) func() error {\n\treturn NewFactory()(fn)\n}", "func New(ctx resource.Context, cfg echo.Config) (i echo.Instance, err error) {\n\terr = resource.UnsupportedEnvironment(ctx.Environment())\n\n\tctx.Environment().Case(environment.Native, func() {\n\t\ti, err = native.New(ctx, cfg)\n\t})\n\n\tctx.Environment().Case(environment.Kube, func() {\n\t\ti, err = kube.New(ctx, cfg)\n\t})\n\treturn\n}", "func (f *FakeInstance) Create(_ context.Context, _ *govultr.InstanceCreateReq) (*govultr.Instance, *http.Response, error) {\n\tpanic(\"implement me\")\n}", "func (vp *scalewayProvider) CreateInstance(log *logging.Logger, options providers.CreateInstanceOptions, dnsProvider providers.DnsProvider) (providers.ClusterInstance, error) {\n\t// Create server\n\tid, err := vp.createServer(options)\n\tif err != nil {\n\t\treturn providers.ClusterInstance{}, maskAny(err)\n\t}\n\n\t// Wait for the server to be active\n\tserver, err := vp.waitUntilServerActive(id, false)\n\tif err != nil {\n\t\treturn providers.ClusterInstance{}, maskAny(err)\n\t}\n\n\tif options.RoleLoadBalancer {\n\t\tpublicIpv4 := server.PublicAddress.IP\n\t\tpublicIpv6 := \"\"\n\t\tif err := providers.RegisterInstance(vp.Logger, dnsProvider, options, server.Name, options.RoleLoadBalancer, publicIpv4, publicIpv6); err != nil {\n\t\t\treturn providers.ClusterInstance{}, maskAny(err)\n\t\t}\n\t}\n\n\tvp.Logger.Infof(\"Server '%s' is ready\", server.Name)\n\n\treturn vp.clusterInstance(server, false), nil\n}", "func CreateInstance(clsid *GUID, iid *GUID) (unk *IUnknown, err error) {\n\tif iid == nil {\n\t\tiid = IID_IUnknown\n\t}\n\thr, _, _ := procCoCreateInstance.Call(\n\t\tuintptr(unsafe.Pointer(clsid)),\n\t\t0,\n\t\tCLSCTX_SERVER,\n\t\tuintptr(unsafe.Pointer(iid)),\n\t\tuintptr(unsafe.Pointer(&unk)))\n\tif hr != 0 {\n\t\terr = NewError(hr)\n\t}\n\treturn\n}", "func NewCreateInstanceWithDefaults() *CreateInstance {\n\tthis := CreateInstance{}\n\treturn &this\n}", "func newInstance(moduleName, name string, priv interface{}) (*BaseInstance, error) {\n\tfactory, found := instanceFactories[moduleName]\n\tif !found {\n\t\treturn nil, fmt.Errorf(\"Module '%s' doesn't exist.\\n\", moduleName)\n\t}\n\n\trp, ok := ringParams[moduleName]\n\tif !ok {\n\t\trp = defaultRingParam\n\t}\n\n\tbi := &BaseInstance{name: name}\n\n\tringName := fmt.Sprintf(\"input-%s\", name)\n\tbi.input = dpdk.RingCreate(ringName, rp.Count, rp.SocketId, dpdk.RING_F_SC_DEQ)\n\tif bi.input == nil {\n\t\treturn nil, fmt.Errorf(\"Input ring creation faild for %s.\\n\", name)\n\t}\n\n\tif rp.SecondaryInput {\n\t\tringName := fmt.Sprintf(\"input2-%s\", name)\n\t\tbi.input2 = dpdk.RingCreate(ringName, rp.Count, rp.SocketId, dpdk.RING_F_SC_DEQ)\n\t\tif bi.input2 == nil {\n\t\t\treturn nil, fmt.Errorf(\"Second input ring creation failed for %s\", name)\n\t\t}\n\t}\n\n\tbi.rules = newRules()\n\n\tinstance, err := factory(bi, priv)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Creating module '%s' with name '%s' failed: %v\\n\", moduleName, name, err)\n\t}\n\tbi.instance = instance\n\n\treturn bi, nil\n}", "func (px *Paxos) CreateInstance(seq int) {\n\n\t//create instance and update max if necessary\n\n\tpx.instances[seq] = &Instance{-1, -1, nil, false}\n\tif seq > px.maxSeq {\n\t\tpx.maxSeq = seq\n\t}\n}", "func (gf *GOFactory) Create(typeID string) (IGameObject, error) {\n\tv, ok := gf.GoCreator[typeID]\n\n\t// not found\n\tif !ok {\n\t\treturn nil, errors.New(\"factory object not found \" + typeID)\n\t}\n\n\tbc := v.(ICreator)\n\n\tgologger.SLogger.Println(\"Factory Created Obj Of Type\", typeID)\n\n\t// call its create function\n\treturn bc.CreateObj(), nil\n}", "func (d *driverMock) CreateInstance(ctx context.Context, publicKey string) (string, error) {\n\tif d.CreateInstanceErr != nil {\n\t\treturn \"\", d.CreateInstanceErr\n\t}\n\n\td.CreateInstanceID = \"ocid1...\"\n\n\treturn d.CreateInstanceID, nil\n}", "func newInstance0(frame *rtda.Frame) {\n\tvars := frame.LocalVars()\n\tconstructorObj := vars.GetRef(0)\n\targArrObj := vars.GetRef(1)\n\n\tgoConstructor := getExtra(constructorObj)\n\tgoClass := goConstructor.Class()\n\tobj := goClass.NewObj()\n\tstack := frame.OperandStack()\n\tstack.PushRef(obj)\n\n\t// call <init>\n\targs := actualConstructorArgs(obj, argArrObj, goConstructor)\n\tframe.Thread().InvokeMethodWithShim(goConstructor, args)\n}", "func NewInstance(ctx *pulumi.Context,\n\tname string, args *InstanceArgs, opts ...pulumi.ResourceOption) (*Instance, error) {\n\tif args == nil || args.Type == nil {\n\t\treturn nil, errors.New(\"missing required argument 'Type'\")\n\t}\n\tif args == nil {\n\t\targs = &InstanceArgs{}\n\t}\n\tvar resource Instance\n\terr := ctx.RegisterResource(\"gcp:datafusion/instance:Instance\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func newInstance(moduleName, name string, priv interface{}) (*BaseInstance, error) {\n\tm, found := modules[moduleName]\n\tif !found {\n\t\treturn nil, fmt.Errorf(\"No such module: %s\", moduleName)\n\t}\n\n\tif _, exists := m.instance[name]; exists {\n\t\treturn nil, fmt.Errorf(\"%s already exists in %s\", name, moduleName)\n\t}\n\n\tbi := &BaseInstance{name: name, module: m, subinstance: false}\n\n\tringName := fmt.Sprintf(\"input-%s\", name)\n\tbi.input = dpdk.RingCreate(ringName, m.ringParam.Count, m.ringParam.SocketId, dpdk.RING_F_SC_DEQ)\n\tif bi.input == nil {\n\t\treturn nil, fmt.Errorf(\"Input ring creation faild for %s.\\n\", name)\n\t}\n\n\tif m.ringParam.SecondaryInput {\n\t\tringName := fmt.Sprintf(\"input2-%s\", name)\n\t\tbi.input2 = dpdk.RingCreate(ringName, m.ringParam.Count, m.ringParam.SocketId, dpdk.RING_F_SC_DEQ)\n\t\tif bi.input2 == nil {\n\t\t\treturn nil, fmt.Errorf(\"Second input ring creation failed for %s\", name)\n\t\t}\n\t}\n\n\tbi.rules = newRules()\n\n\tif m.moduleType == TypeInterface || m.moduleType == TypeRIF {\n\t\tbi.counter = NewCounter()\n\t}\n\n\tinstance, err := m.factory(bi, priv)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Creating module '%s' with name '%s' failed: %v\\n\", moduleName, name, err)\n\t}\n\tbi.instance = instance\n\n\t// Set rule observer, if the module complies to RulesNotify.\n\tif rn, ok := instance.(RulesNotify); ok {\n\t\tbi.rules.setRulesNotify(rn)\n\t}\n\n\tm.instance[name] = bi\n\n\treturn bi, nil\n}", "func (h *Adapter) CreateInstance(kubeconfig []byte, contextName string, ch *chan interface{}) error {\n\terr := h.validateKubeconfig(kubeconfig)\n\tif err != nil {\n\t\treturn ErrCreateInstance(err)\n\t}\n\n\terr = h.createKubeClient(kubeconfig)\n\tif err != nil {\n\t\treturn ErrCreateInstance(err)\n\t}\n\n\terr = h.createKubeconfig(kubeconfig)\n\tif err != nil {\n\t\treturn ErrCreateInstance(err)\n\t}\n\n\terr = h.createMesheryKubeclient(kubeconfig)\n\tif err != nil {\n\t\treturn ErrCreateInstance(err)\n\t}\n\n\th.ClientcmdConfig.CurrentContext = contextName\n\th.Channel = ch\n\n\treturn nil\n}", "func (h *Adapter) CreateInstance(kubeconfig []byte, contextName string, ch *chan interface{}) error {\n\terr := h.validateKubeconfig(kubeconfig)\n\tif err != nil {\n\t\treturn ErrCreateInstance(err)\n\t}\n\n\terr = h.createKubeconfig(kubeconfig)\n\tif err != nil {\n\t\treturn ErrCreateInstance(err)\n\t}\n\n\th.MesheryKubeclient, err = mesherykube.New(kubeconfig)\n\tif err != nil {\n\t\treturn ErrClientSet(err)\n\t}\n\n\th.DynamicKubeClient = h.MesheryKubeclient.DynamicKubeClient\n\th.RestConfig = h.MesheryKubeclient.RestConfig\n\n\th.KubeClient, err = kubernetes.NewForConfig(&h.RestConfig)\n\tif err != nil {\n\t\treturn ErrClientSet(err)\n\t}\n\n\th.ClientcmdConfig.CurrentContext = contextName\n\th.Channel = ch\n\n\treturn nil\n}", "func (c *Client) CloudCreateInstance(projectID, name, pubkeyID, flavorID, imageID, region string) (instance *types.CloudInstance, err error) {\n\tinstanceReq := types.CloudInstance{\n\t\tName: name,\n\t\tSSHKeyID: pubkeyID,\n\t\tFlavorID: flavorID,\n\t\tImageID: imageID,\n\t\tRegion: region,\n\t}\n\terr = c.Post(queryEscape(\"/cloud/project/%s/instance\", projectID), instanceReq, &instance)\n\treturn instance, err\n}", "func (p *OnPrem) CreateInstance(ctx *Context) error {\n\tc := ctx.config\n\n\thypervisor := HypervisorInstance()\n\tif hypervisor == nil {\n\t\tfmt.Println(\"No hypervisor found on $PATH\")\n\t\tfmt.Println(\"Please install OPS using curl https://ops.city/get.sh -sSfL | sh\")\n\t\tos.Exit(1)\n\t}\n\n\tinstancename := c.CloudConfig.ImageName\n\n\tfmt.Printf(\"booting %s ...\\n\", instancename)\n\n\topshome := GetOpsHome()\n\timgpath := path.Join(opshome, \"images\", instancename)\n\n\tc.RunConfig.BaseName = instancename\n\tc.RunConfig.Imagename = imgpath\n\tc.RunConfig.OnPrem = true\n\n\thypervisor.Start(&c.RunConfig)\n\n\treturn nil\n}", "func (c *Client) CreateInstance(displayName, availabilityDomain, compartmentID, nodeShape, nodeImageName, nodeSubnetID, sshUser, authorizedKeys string, nodeOCPUs, nodeMemoryInGBs int) (string, error) {\n\n\treq := identity.ListAvailabilityDomainsRequest{}\n\treq.CompartmentId = &compartmentID\n\tads, err := c.identityClient.ListAvailabilityDomains(context.Background(), req)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Just in case shortened or lower-case availability domain name was used\n\tlog.Debugf(\"Resolving availability domain from %s\", availabilityDomain)\n\tfor _, ad := range ads.Items {\n\t\tif strings.Contains(*ad.Name, strings.ToUpper(availabilityDomain)) {\n\t\t\tlog.Debugf(\"Availability domain %s\", *ad.Name)\n\t\t\tavailabilityDomain = *ad.Name\n\t\t}\n\t}\n\n\timageID, err := c.getImageID(compartmentID, nodeImageName)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Create the launch compute instance request\n\trequest := core.LaunchInstanceRequest{\n\t\tLaunchInstanceDetails: core.LaunchInstanceDetails{\n\t\t\tAvailabilityDomain: &availabilityDomain,\n\t\t\tCompartmentId: &compartmentID,\n\t\t\tShape: &nodeShape,\n\t\t\tCreateVnicDetails: &core.CreateVnicDetails{\n\t\t\t\tSubnetId: &nodeSubnetID,\n\t\t\t},\n\t\t\tDisplayName: &displayName,\n\t\t\tMetadata: map[string]string{\n\t\t\t\t\"ssh_authorized_keys\": authorizedKeys,\n\t\t\t\t\"user_data\": base64.StdEncoding.EncodeToString(createCloudInitScript(sshUser)),\n\t\t\t},\n\t\t\tSourceDetails: core.InstanceSourceViaImageDetails{\n\t\t\t\tImageId: imageID,\n\t\t\t},\n\t\t},\n\t}\n\n\tif nodeOCPUs > 0 {\n\t\toCPUs := float32(nodeOCPUs)\n\t\tmemoryInGBs := float32(nodeMemoryInGBs)\n\n\t\tLaunchInstanceShapeConfigDetails := core.LaunchInstanceShapeConfigDetails{\n\t\t\tOcpus: &oCPUs,\n\t\t\tMemoryInGBs: &memoryInGBs,\n\t\t}\n\t\trequest.ShapeConfig = &LaunchInstanceShapeConfigDetails\n\t}\n\n\tlog.Debugf(\"Launching instance with cloud-init: %s\", string(createCloudInitScript(sshUser)))\n\n\tcreateResp, err := c.computeClient.LaunchInstance(context.Background(), request)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// wait until lifecycle status is Running\n\tpollUntilRunning := func(r common.OCIOperationResponse) bool {\n\t\tif converted, ok := r.Response.(core.GetInstanceResponse); ok {\n\t\t\treturn converted.LifecycleState != core.InstanceLifecycleStateRunning\n\t\t}\n\t\treturn true\n\t}\n\n\t// create get instance request with a retry policy which takes a function\n\t// to determine shouldRetry or not\n\tpollingGetRequest := core.GetInstanceRequest{\n\t\tInstanceId: createResp.Instance.Id,\n\t\tRequestMetadata: helpers.GetRequestMetadataWithCustomizedRetryPolicy(pollUntilRunning),\n\t}\n\n\tinstance, pollError := c.computeClient.GetInstance(context.Background(), pollingGetRequest)\n\tif pollError != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn *instance.Id, nil\n}", "func doCreate(constructor func() base.IGameObject2D, isActive *bool) base.IGameObject2D {\r\n\tobj := constructor()\r\n\tobj.Obj().SetIGameObject2D(obj)\r\n\tapp.registerChannel <- resourceAccessRequest{\r\n\t\tpayload: obj,\r\n\t\tisActive: isActive,\r\n\t}\r\n\treturn obj\r\n}", "func newInstance(rt reflect.Type) reflect.Value {\n\tt := rt\n\tif t.Kind() == reflect.Ptr {\n\t\tt = t.Elem()\n\t}\n\treturn reflect.New(t)\n}", "func (a *Client) CreateInstance(params *CreateInstanceParams) (*CreateInstanceOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewCreateInstanceParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"createInstance\",\n\t\tMethod: \"POST\",\n\t\tPathPattern: \"/instances\",\n\t\tProducesMediaTypes: []string{\"\"},\n\t\tConsumesMediaTypes: []string{\"\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &CreateInstanceReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*CreateInstanceOK), nil\n\n}", "func (i *InstanceServiceHandler) Create(ctx context.Context, instanceReq *InstanceCreateReq) (*Instance, error) {\n\turi := fmt.Sprintf(\"%s\", instancePath)\n\n\treq, err := i.client.NewRequest(ctx, http.MethodPost, uri, instanceReq)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tinstance := new(instanceBase)\n\tif err = i.client.DoWithContext(ctx, req, instance); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn instance.Instance, nil\n}", "func newProtonInstance(proton core.Protoner) reflect.Value {\n\tbaseValue := reflect.ValueOf(proton)\n\n\t// try to create new value of proton\n\tmethod := baseValue.MethodByName(\"New\")\n\tif method.IsValid() {\n\t\treturns := method.Call(emptyParameters)\n\t\tif len(returns) <= 0 {\n\t\t\tpanic(fmt.Sprintf(\"Method New must has at least 1 returns. now %d\", len(returns)))\n\t\t}\n\t\treturn returns[0]\n\t} else {\n\t\t// return reflect.New(reflect.TypeOf(proton).Elem())\n\t\treturn newInstance(reflect.TypeOf(proton))\n\t}\n}", "func newInstance(x *runtime.Runtime, p *build.Instance, v *adt.Vertex) *Instance {\n\t// TODO: associate root source with structLit.\n\tinst := &Instance{\n\t\troot: v,\n\t\tinst: p,\n\t}\n\tif p != nil {\n\t\tinst.ImportPath = p.ImportPath\n\t\tinst.Dir = p.Dir\n\t\tinst.PkgName = p.PkgName\n\t\tinst.DisplayName = p.ImportPath\n\t\tif p.Err != nil {\n\t\t\tinst.setListOrError(p.Err)\n\t\t}\n\t}\n\n\tx.AddInst(p.ImportPath, v, p)\n\tx.SetBuildData(p, inst)\n\tinst.index = x\n\treturn inst\n}", "func NewCfnInstance(scope awscdk.Construct, id *string, props *CfnInstanceProps) CfnInstance {\n\t_init_.Initialize()\n\n\tj := jsiiProxy_CfnInstance{}\n\n\t_jsii_.Create(\n\t\t\"monocdk.aws_opsworks.CfnInstance\",\n\t\t[]interface{}{scope, id, props},\n\t\t&j,\n\t)\n\n\treturn &j\n}", "func (s *API) CreateInstance(req *CreateInstanceRequest, opts ...scw.RequestOption) (*Instance, error) {\n\tvar err error\n\n\tdefaultProjectID, exist := s.client.GetDefaultProjectID()\n\tif exist && req.OrganizationID == nil && req.ProjectID == nil {\n\t\treq.ProjectID = &defaultProjectID\n\t}\n\n\tdefaultOrganizationID, exist := s.client.GetDefaultOrganizationID()\n\tif exist && req.OrganizationID == nil && req.ProjectID == nil {\n\t\treq.OrganizationID = &defaultOrganizationID\n\t}\n\n\tif req.Region == \"\" {\n\t\tdefaultRegion, _ := s.client.GetDefaultRegion()\n\t\treq.Region = defaultRegion\n\t}\n\n\tif req.Name == \"\" {\n\t\treq.Name = namegenerator.GetRandomName(\"ins\")\n\t}\n\n\tif fmt.Sprint(req.Region) == \"\" {\n\t\treturn nil, errors.New(\"field Region cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"POST\",\n\t\tPath: \"/rdb/v1/regions/\" + fmt.Sprint(req.Region) + \"/instances\",\n\t\tHeaders: http.Header{},\n\t}\n\n\terr = scwReq.SetBody(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar resp Instance\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func Create(constructor func() base.IGameObject2D) base.IGameObject2D {\r\n\treturn doCreate(constructor, infra.BoolPtr_True)\r\n}", "func NewCfnInstance_Override(c CfnInstance, scope awscdk.Construct, id *string, props *CfnInstanceProps) {\n\t_init_.Initialize()\n\n\t_jsii_.Create(\n\t\t\"monocdk.aws_opsworks.CfnInstance\",\n\t\t[]interface{}{scope, id, props},\n\t\tc,\n\t)\n}", "func CreateInstance(client daisyCompute.Client, project, zone string, i *api.Instance) (*Instance, error) {\n\tif err := client.CreateInstance(project, zone, i); err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Instance{Instance: i, client: client, Project: project, Zone: zone}, nil\n}", "func (h *handler) CreateInstance(kubeconfig []byte, contextName string, ch *chan interface{}) error {\n\n\tvar err error\n\th.channel = ch\n\th.kubeConfigPath = \"/Users/abishekk/.kube/config\"\n\t// h.kubeConfigPath, err = h.config.GetKey(\"kube-config-path\")\n\t// if err != nil {\n\t// \treturn ErrClientConfig(err)\n\t// }\n\n\tconfig, err := h.clientConfig(kubeconfig, contextName)\n\tif err != nil {\n\t\treturn ErrClientConfig(err)\n\t}\n\n\t// creates the clientset\n\tclientset, err := kubernetes.NewForConfig(config)\n\tif err != nil {\n\t\treturn ErrClientSet(err)\n\t}\n\n\th.kubeClient = clientset\n\n\treturn nil\n}", "func (client BaseClient) CreateFeatureInstance(ctx context.Context, body *FeatureInstanceInputs) (result FeatureInstance, err error) {\n\tif err := validation.Validate([]validation.Validation{\n\t\t{TargetValue: body,\n\t\t\tConstraints: []validation.Constraint{{Target: \"body\", Name: validation.Null, Rule: false,\n\t\t\t\tChain: []validation.Constraint{{Target: \"body.FeatureName\", Name: validation.Null, Rule: true,\n\t\t\t\t\tChain: []validation.Constraint{{Target: \"body.FeatureName\", Name: validation.Pattern, Rule: `^[a-z0-9-]+$`, Chain: nil}}},\n\t\t\t\t\t{Target: \"body.FeatureVersion\", Name: validation.Null, Rule: true,\n\t\t\t\t\t\tChain: []validation.Constraint{{Target: \"body.FeatureVersion\", Name: validation.Pattern, Rule: `^v?((\\d+)\\.(\\d+)\\.(\\d+))(?:-([\\dA-Za-z\\-]+(?:\\.[\\dA-Za-z\\-]+)*))?(?:\\+([\\dA-Za-z\\-]+(?:\\.[\\dA-Za-z\\-]+)*))?$`, Chain: nil}}},\n\t\t\t\t\t{Target: \"body.InstanceName\", Name: validation.Null, Rule: true,\n\t\t\t\t\t\tChain: []validation.Constraint{{Target: \"body.InstanceName\", Name: validation.Pattern, Rule: `^[a-z0-9-]+$`, Chain: nil}}},\n\t\t\t\t}}}}}); err != nil {\n\t\treturn result, validation.NewError(\"beacon.BaseClient\", \"CreateFeatureInstance\", err.Error())\n\t}\n\n\treq, err := client.CreateFeatureInstancePreparer(ctx, body)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"CreateFeatureInstance\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresp, err := client.CreateFeatureInstanceSender(req)\n\tif err != nil {\n\t\tresult.Response = autorest.Response{Response: resp}\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"CreateFeatureInstance\", resp, \"Failure sending request\")\n\t\treturn\n\t}\n\n\tresult, err = client.CreateFeatureInstanceResponder(resp)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"CreateFeatureInstance\", resp, \"Failure responding to request\")\n\t}\n\n\treturn\n}", "func (c *Client) CreateInstance(args *CreateInstanceArgs) (*CreateInstanceResult, error) {\n\tif len(args.AdminPass) > 0 {\n\t\tcryptedPass, err := Aes128EncryptUseSecreteKey(c.Config.Credentials.SecretAccessKey, args.AdminPass)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\targs.AdminPass = cryptedPass\n\t}\n\n\tif args.RootDiskSizeInGb <= 0 {\n\t\targs.RootDiskSizeInGb = 20\n\t}\n\n\tif args.PurchaseCount < 1 {\n\t\targs.PurchaseCount = 1\n\t}\n\n\tjsonBytes, jsonErr := json.Marshal(args)\n\tif jsonErr != nil {\n\t\treturn nil, jsonErr\n\t}\n\tbody, err := bce.NewBodyFromBytes(jsonBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn CreateInstance(c, args, body)\n}", "func New(rnd *random.Random) *Fn {\n\treturn &Fn{\n\t\trnd: rnd,\n\t}\n}", "func NewCreateInstanceRequest(server string, params *CreateInstanceParams, body CreateInstanceJSONRequestBody) (*http.Request, error) {\n\tvar bodyReader io.Reader\n\tbuf, err := json.Marshal(body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbodyReader = bytes.NewReader(buf)\n\treturn NewCreateInstanceRequestWithBody(server, params, \"application/json\", bodyReader)\n}", "func (client *Client) CreateInstance(request *CreateInstanceRequest) (_result *CreateInstanceResponse, _err error) {\n\truntime := &util.RuntimeOptions{}\n\t_result = &CreateInstanceResponse{}\n\t_body, _err := client.CreateInstanceWithOptions(request, runtime)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\t_result = _body\n\treturn _result, _err\n}", "func (f *Factory) Create(componentName string) (interface{}, error) {\n\tif info, exists := f.registry[componentName]; exists {\n\t\treturn info.Constructor()\n\t}\n\treturn nil, fmt.Errorf(\"Factory error: component '%s' does not exist\", componentName)\n}", "func NewInstance(ctx *pulumi.Context,\n\tname string, args *InstanceArgs, opts ...pulumi.ResourceOption) (*Instance, error) {\n\tif args == nil || args.AvailabilityZone == nil {\n\t\treturn nil, errors.New(\"missing required argument 'AvailabilityZone'\")\n\t}\n\tif args == nil || args.BlueprintId == nil {\n\t\treturn nil, errors.New(\"missing required argument 'BlueprintId'\")\n\t}\n\tif args == nil || args.BundleId == nil {\n\t\treturn nil, errors.New(\"missing required argument 'BundleId'\")\n\t}\n\tif args == nil {\n\t\targs = &InstanceArgs{}\n\t}\n\tvar resource Instance\n\terr := ctx.RegisterResource(\"aws:lightsail/instance:Instance\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func NewInstance(value interface{}) interface{} {\n\tvar result interface{}\n\trv := reflect.ValueOf(value)\n\tif rv.Kind() == reflect.Ptr {\n\t\tresult = reflect.New(rv.Elem().Type()).Interface()\n\t} else {\n\t\tresult = reflect.New(reflect.TypeOf(value)).Elem().Interface()\n\t}\n\treturn result\n}", "func New(publisherFactories []factory.Contract) *instance {\n\treturn &instance{\n\t\tpublisherFactories: publisherFactories,\n\t}\n}", "func (t *MyTriggerFactory) New(config *trigger.Config) trigger.Trigger {\n\treturn &BootstrapTrigger{metadata: t.metadata, config: config}\n}", "func (s *stepCreateInstance) Run(ctx context.Context, state multistep.StateBag) multistep.StepAction {\n\tclient := state.Get(\"client\").(*civogo.Client)\n\tui := state.Get(\"ui\").(packer.Ui)\n\tc := state.Get(\"config\").(*Config)\n\tsshKeyID := state.Get(\"ssh_key_id\").(string)\n\n\t// Create the instance based on configuration\n\tui.Say(\"Creating instance...\")\n\n\ttemplate, err := client.FindTemplate(c.Template)\n\tif err != nil {\n\t\tui.Error(err.Error())\n\t}\n\n\tnetwork, _ := client.GetDefaultNetwork()\n\n\tInstanceConfig := &civogo.InstanceConfig{\n\t\tHostname: c.InstanceName,\n\t\tPublicIPRequired: c.PublicNetworking,\n\t\tRegion: c.Region,\n\t\tNetworkID: network.ID,\n\t\tInitialUser: c.Comm.SSHUsername,\n\t\tSize: c.Size,\n\t\tTemplateID: template.ID,\n\t\tSSHKeyID: sshKeyID,\n\t}\n\n\tlog.Printf(\"[DEBUG] Instance create paramaters: %+v\", InstanceConfig)\n\n\tinstance, err := client.CreateInstance(InstanceConfig)\n\tif err != nil {\n\t\terr := fmt.Errorf(\"Error creating instance: %s\", err)\n\t\tstate.Put(\"error\", err)\n\t\tui.Error(err.Error())\n\t\treturn multistep.ActionHalt\n\t}\n\n\t// We use this in cleanup\n\ts.instanceID = instance.ID\n\n\t// Store the instance id for later\n\tstate.Put(\"instance_id\", instance.ID)\n\n\treturn multistep.ActionContinue\n}", "func New(behavior Func) structmap.Behavior {\n\treturn behavior\n}", "func NewInstance(ctx *pulumi.Context,\n\tname string, args *InstanceArgs, opts ...pulumi.ResourceOption) (*Instance, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.Cluster == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'Cluster'\")\n\t}\n\tif args.InstanceId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'InstanceId'\")\n\t}\n\tif args.InstanceType == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'InstanceType'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Instance\n\terr := ctx.RegisterResource(\"gcp:alloydb/instance:Instance\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (i *Instance) SpawnInstance() (instance *Instance, err error) {\n\tvar res uintptr\n\tvar newUnknown *ole.IUnknown\n\n\tres, _, _ = syscall.SyscallN(\n\t\ti.vTable.SpawnInstance, // IWbemClassObject::SpawnInstance(\n\t\tuintptr(unsafe.Pointer(i.object)), // IWbemClassObject ptr\n\t\tuintptr(0), // [in] long lFlags,\n\t\tuintptr(unsafe.Pointer(&newUnknown))) // [out] IWbemClassObject **ppNewInstance)\n\tif res != 0 {\n\t\treturn nil, ole.NewError(res)\n\t}\n\n\treturn newInstance(newUnknown, i.service), nil\n}", "func NewInstance(ctx *pulumi.Context,\n\tname string, args *InstanceArgs, opts ...pulumi.ResourceOption) (*Instance, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.Description == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'Description'\")\n\t}\n\tif args.InstanceSeries == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'InstanceSeries'\")\n\t}\n\tif args.Specification == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'Specification'\")\n\t}\n\tif args.VswitchId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'VswitchId'\")\n\t}\n\tif args.ZoneId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'ZoneId'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Instance\n\terr := ctx.RegisterResource(\"alicloud:drds/instance:Instance\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func CreateTestInstance(t *testing.T, uuid dvid.UUID, typename, name string, config dvid.Config) {\n\tconfig.Set(\"typename\", typename)\n\tconfig.Set(\"dataname\", name)\n\tjsonData, err := config.MarshalJSON()\n\tif err != nil {\n\t\tt.Fatalf(\"Unable to make JSON for instance creation: %v\\n\", config)\n\t}\n\tapiStr := fmt.Sprintf(\"%srepo/%s/instance\", WebAPIPath, uuid)\n\tTestHTTP(t, \"POST\", apiStr, bytes.NewBuffer(jsonData))\n}", "func (p *ProxMox) CreateInstance(ctx *lepton.Context) error {\n\n\tvar err error\n\n\tconfig := ctx.Config()\n\n\tnextid := p.getNextID()\n\n\tp.instanceName = config.RunConfig.InstanceName\n\n\tp.isoStorageName = config.TargetConfig[\"IsoStorageName\"]\n\n\tp.imageName = config.CloudConfig.ImageName\n\n\tp.arch = \"x86_64\"\n\tif config.TargetConfig[\"Arch\"] != \"\" {\n\t\tp.arch = config.TargetConfig[\"Arch\"]\n\t}\n\n\tp.machine = \"q35\"\n\tif config.TargetConfig[\"Machine\"] != \"\" {\n\t\tp.machine = config.TargetConfig[\"Machine\"]\n\t}\n\n\tp.sockets = \"1\"\n\tif config.TargetConfig[\"Sockets\"] != \"\" {\n\t\tsocketsInt, err := strconv.Atoi(config.TargetConfig[\"Sockets\"])\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif socketsInt < 1 {\n\t\t\treturn errors.New(\"Bad configuration option; Sockets can only be set postitive starting from \\\"1\\\"\")\n\t\t}\n\t\tp.sockets = config.TargetConfig[\"Sockets\"]\n\t}\n\n\tp.cores = \"1\"\n\tif config.TargetConfig[\"Cores\"] != \"\" {\n\t\tcoresInt, err := strconv.Atoi(config.TargetConfig[\"Cores\"])\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif coresInt < 1 {\n\t\t\treturn errors.New(\"Bad configuration option; Cores can only be set postitive starting from \\\"1\\\"\")\n\t\t}\n\t\tp.cores = config.TargetConfig[\"Cores\"]\n\t}\n\n\tp.numa = \"0\"\n\tif config.TargetConfig[\"Numa\"] != \"\" {\n\t\tif config.TargetConfig[\"Numa\"] != \"0\" && config.TargetConfig[\"Numa\"] != \"1\" {\n\t\t\treturn errors.New(\"Bad configuration option; Numa can only be set to \\\"0\\\" or \\\"1\\\"\")\n\t\t}\n\t\tp.numa = config.TargetConfig[\"Numa\"]\n\t}\n\n\t// Memory\n\n\tp.memory = \"512\"\n\tif config.TargetConfig[\"Memory\"] != \"\" {\n\t\tmemoryInt, err := lepton.RAMInBytes(config.TargetConfig[\"Memory\"])\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tmemoryInt = memoryInt / 1024 / 1024\n\t\tp.memory = strconv.FormatInt(memoryInt, 10)\n\t}\n\n\t// Main storage\n\n\tp.storageName = \"local-lvm\"\n\tif config.TargetConfig[\"StorageName\"] != \"\" {\n\t\tp.storageName = config.TargetConfig[\"StorageName\"]\n\t}\n\n\t// Iso storage\n\n\tp.isoStorageName = \"local\"\n\tif config.TargetConfig[\"IsoStorageName\"] != \"\" {\n\t\tp.isoStorageName = config.TargetConfig[\"IsoStorageName\"]\n\t}\n\n\t// Bridge prefix\n\n\tp.bridgePrefix = \"vmbr\"\n\tif config.TargetConfig[\"BridgePrefix\"] != \"\" {\n\t\tp.bridgePrefix = config.TargetConfig[\"BridgePrefix\"]\n\t}\n\n\t// Onboot\n\n\tp.onboot = \"0\"\n\tif config.TargetConfig[\"Onboot\"] != \"\" {\n\t\tif config.TargetConfig[\"Onboot\"] != \"0\" && config.TargetConfig[\"Onboot\"] != \"1\" {\n\t\t\treturn errors.New(\"Bad configuration option; Onboot can only be set to \\\"0\\\" or \\\"1\\\"\")\n\t\t}\n\t\tp.onboot = config.TargetConfig[\"Onboot\"]\n\t}\n\n\t// Protection\n\n\tp.protection = \"0\"\n\tif config.TargetConfig[\"Protection\"] != \"\" {\n\t\tif config.TargetConfig[\"Protection\"] != \"0\" && config.TargetConfig[\"Protection\"] != \"1\" {\n\t\t\treturn errors.New(\"Bad configuration option; Protection can only be set to \\\"0\\\" or \\\"1\\\"\")\n\t\t}\n\t\tp.protection = config.TargetConfig[\"Protection\"]\n\t}\n\n\t// These two preventive checks here, because Proxmox will not return\n\t// an error if the storage is missing and a misconfigured instance will be created.\n\n\terr = p.CheckStorage(p.storageName, \"images\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = p.CheckStorage(p.isoStorageName, \"iso\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdata := url.Values{}\n\tdata.Set(\"vmid\", nextid)\n\tdata.Set(\"name\", p.instanceName)\n\tdata.Set(\"name\", p.imageName)\n\tdata.Set(\"machine\", p.machine)\n\tdata.Set(\"sockets\", p.sockets)\n\tdata.Set(\"cores\", p.cores)\n\tdata.Set(\"numa\", p.numa)\n\tdata.Set(\"memory\", p.memory)\n\tdata.Set(\"onboot\", p.onboot)\n\tdata.Set(\"protection\", p.protection)\n\tdata.Set(\"serial0\", \"socket\")\n\n\t// Configuring network interfaces\n\n\tnics := config.RunConfig.Nics\n\tfor i := 0; i < len(nics); i++ {\n\t\tis := strconv.Itoa(i)\n\t\tbrName := nics[i].BridgeName\n\t\tif brName == \"\" {\n\t\t\tbrName = p.bridgePrefix + is\n\t\t}\n\n\t\terr = p.CheckBridge(brName)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif nics[i].IPAddress != \"\" {\n\t\t\tcidr := \"24\"\n\n\t\t\tif nics[i].NetMask != \"\" {\n\t\t\t\tcidrInt := lepton.CCidr(nics[i].NetMask)\n\t\t\t\tcidr = strconv.FormatInt(int64(cidrInt), 10)\n\t\t\t}\n\n\t\t\tif nics[i].Gateway != \"\" {\n\t\t\t\tdata.Set(\"ipconfig\"+is, \"ip=\"+nics[i].IPAddress+\"/\"+cidr+\",\"+\"gw=\"+nics[i].Gateway)\n\t\t\t} else {\n\t\t\t\tdata.Set(\"ipconfig\"+is, \"ip=\"+nics[i].IPAddress+\"/\"+cidr)\n\t\t\t}\n\t\t} else {\n\t\t\tdata.Set(\"ipconfig\"+is, \"dhcp\")\n\t\t}\n\n\t\tdata.Set(\"net\"+is, \"model=virtio,bridge=\"+brName)\n\t}\n\tif len(nics) == 0 {\n\t\t// single dhcp nic\n\t\tdata.Set(\"net0\", \"model=virtio,bridge=vmbr0\")\n\t}\n\n\treq, err := http.NewRequest(\"POST\", p.apiURL+\"/api2/json/nodes/\"+p.nodeNAME+\"/qemu\", bytes.NewBufferString(data.Encode()))\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\ttr := &http.Transport{\n\t\tTLSClientConfig: &tls.Config{InsecureSkipVerify: true},\n\t}\n\tclient := &http.Client{Transport: tr}\n\n\treq.Header.Add(\"Authorization\", \"PVEAPIToken=\"+p.tokenID+\"=\"+p.secret)\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\tbody, err := io.ReadAll(resp.Body)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\tdebug := false\n\tif debug {\n\t\tfmt.Println(string(body))\n\t}\n\n\terr = p.CheckResultType(body, \"createinstance\", \"file=\"+p.isoStorageName+\":iso/\"+p.imageName+\".iso\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = p.addVirtioDisk(ctx, nextid)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = p.movDisk(ctx, nextid)\n\n\treturn err\n}", "func New(kind string, result metadata.Contract) *instance {\n\treturn &instance{\n\t\tkind: kind,\n\t\tresult: result,\n\t}\n}", "func (f *Function) CreateInstanceSnapshot() {\n\tlogger := log.WithFields(log.Fields{\"fID\": f.fID})\n\n\tlogger.Debug(\"Creating instance snapshot\")\n\n\tctx, cancel := context.WithTimeout(context.Background(), time.Second*5)\n\tdefer cancel()\n\n\terr := orch.PauseVM(ctx, f.vmID)\n\tif err != nil {\n\t\tlog.Panic(err)\n\t}\n\n\terr = orch.CreateSnapshot(ctx, f.vmID)\n\tif err != nil {\n\t\tlog.Panic(err)\n\t}\n\n\t_, err = orch.ResumeVM(ctx, f.vmID)\n\tif err != nil {\n\t\tlog.Panic(err)\n\t}\n}", "func NewInstance(ctx *pulumi.Context,\n\tname string, args *InstanceArgs, opts ...pulumi.ResourceOption) (*Instance, error) {\n\tif args == nil {\n\t\targs = &InstanceArgs{}\n\t}\n\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Instance\n\terr := ctx.RegisterResource(\"gcp:looker/instance:Instance\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func NewInstance(ctx *pulumi.Context,\n\tname string, args *InstanceArgs, opts ...pulumi.ResourceOption) (*Instance, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.PaymentType == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'PaymentType'\")\n\t}\n\tif args.VersionCode == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'VersionCode'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Instance\n\terr := ctx.RegisterResource(\"alicloud:threatdetection/instance:Instance\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func Create(conf *config.Config, logger *logrus.Logger) (Runner, error) {\n\tvar srv Runner\n\n\tswitch conf.Server.Type {\n\tcase \"grpc\":\n\t\tsrv = grpc.NewRunner(conf,logger)\n\tdefault:\n\t\treturn nil, errors.New(\"InvalidServerTypeError(#{conf.Server.Type})\")\n\t}\n\treturn srv, nil\n}", "func New() Go { return Go{} }", "func expectedNewInstance(jobID, datasetID string) *dataset.NewInstance {\n\tnewInstance := &dataset.NewInstance{\n\t\tLinks: &dataset.Links{\n\t\t\tDataset: dataset.Link{\n\t\t\t\tURL: \"http://localhost:22000/datasets/\" + datasetID,\n\t\t\t\tID: datasetID,\n\t\t\t},\n\t\t\tJob: dataset.Link{\n\t\t\t\tURL: \"http://import-api/jobs/\" + jobID,\n\t\t\t\tID: jobID,\n\t\t\t},\n\t\t},\n\t\tDimensions: []dataset.CodeList{},\n\t\tImportTasks: &dataset.InstanceImportTasks{\n\t\t\tImportObservations: &dataset.ImportObservationsTask{\n\t\t\t\tState: dataset.StateCreated.String(),\n\t\t\t},\n\t\t\tBuildHierarchyTasks: []*dataset.BuildHierarchyTask{},\n\t\t\tBuildSearchIndexTasks: []*dataset.BuildSearchIndexTask{},\n\t\t},\n\t\tType: \"cantabular_blob\",\n\t}\n\tif datasetID == \"dataset1\" {\n\t\tnewInstance.Dimensions = []dataset.CodeList{{ID: \"codelist11\"}, {ID: \"codelist12\"}}\n\t\tnewInstance.LowestGeography = \"lowest_geo\"\n\t} else if datasetID == \"dataset2\" {\n\t\tnewInstance.Dimensions = []dataset.CodeList{{ID: \"codelist21\"}, {ID: \"codelist22\"}, {ID: \"codelist23\"}}\n\t}\n\treturn newInstance\n}", "func (a *ACMEInstance) CreateInstance(challenge, dir, domains, domainIPs, dnsManager string, revoke bool, serverK server.KeyInstance, clientK client.KeyInstanceCert) {\n\n\tif ok := supportedChallenges[challenge]; !ok {\n\t\tfmt.Fprintln(os.Stderr, \"unsupported challengeType: \", challenge)\n\t\tos.Exit(2)\n\t}\n\ta.challengeType = challenge\n\ta.dirURL = dir\n\ta.domainList = strings.Split(domains, \";\")\n\n\tipList := strings.Split(domainIPs, \";\")\n\tipDom := make(map[string]string)\n\tfor n, domain := range a.domainList {\n\t\tipDom[domain] = ipList[n]\n\t}\n\ta.ipForDomain = ipDom\n\n\tif dnsManager == \"azuredns\" {\n\t\ta.dnsManager = dnsmanager.AZUREDNSCredentials{Login: \"dummyuser\",\n\t\t\tPassword: \"1234\",\n\t\t\tSubscriptionID: \"9fa587f1-4961-48a6-b6f6-ec69c6d724f1\",\n\t\t\tResourceGroups: \"fileTransfer\",\n\t\t\tAuthorization: \"Bearer eyJ0eXAiOiJKV1QiLCJhbGciOiJSUzI1NiIsIng1dCI6Imh1Tjk1SXZQZmVocTM0R3pCRFoxR1hHaXJuTSIsImtpZCI6Imh1Tjk1SXZQZmVocTM0R3pCRFoxR1hHaXJuTSJ9.eyJhdWQiOiJodHRwczovL21hbmFnZW1lbnQuY29yZS53aW5kb3dzLm5ldC8iLCJpc3MiOiJodHRwczovL3N0cy53aW5kb3dzLm5ldC9mMDhhODUxZS1hYzVjLTQ5NGItODk0MS00N2U2YTI2NTc4MWQvIiwiaWF0IjoxNTk3MDY5Njc5LCJuYmYiOjE1OTcwNjk2NzksImV4cCI6MTU5NzA3MzU3OSwiYWNyIjoiMSIsImFpbyI6IkFTUUEyLzhRQUFBQWJPNHQrWW5pYVp2NEUzRG4rYUdGOGxVc25RZzhJQ1VHNS9yZitLMEFrcWM9IiwiYW1yIjpbInB3ZCJdLCJhcHBpZCI6IjdmNTlhNzczLTJlYWYtNDI5Yy1hMDU5LTUwZmM1YmIyOGI0NCIsImFwcGlkYWNyIjoiMiIsImZhbWlseV9uYW1lIjoiTWVpZXIiLCJnaXZlbl9uYW1lIjoiRmlsaXAiLCJncm91cHMiOlsiNzA5YzZmM2YtNTEwYy00NTgwLTlkMGYtYzQ1OWJiMTcyMDE3Il0sImlwYWRkciI6IjUxLjE1NC41My4xNjUiLCJuYW1lIjoiRmlsaXAgTWVpZXIiLCJvaWQiOiJmMGNhMzk3Ni1lNjBlLTQzZDItYTM0ZS0wMDZhZGRlNDVhYmIiLCJwdWlkIjoiMTAwMzIwMDA2QjI2MUZDNyIsInJoIjoiMC5BVHdBSG9XSzhGeXNTMG1KUVVmbW9tVjRIWE9uV1gtdkxweENvRmxRX0Z1eWkwUThBTTQuIiwic2NwIjoidXNlcl9pbXBlcnNvbmF0aW9uIiwic3ViIjoiMzlKdHpVVUFCOFNYNWRWRVBOYWJsZzFiUkp5MkhGdW43TEZDRktxamRlVSIsInRpZCI6ImYwOGE4NTFlLWFjNWMtNDk0Yi04OTQxLTQ3ZTZhMjY1NzgxZCIsInVuaXF1ZV9uYW1lIjoiZmlsaXAubWVpZXJAOGRheXNhd2Vlay5jYyIsInVwbiI6ImZpbGlwLm1laWVyQDhkYXlzYXdlZWsuY2MiLCJ1dGkiOiJ1aXBUU3VVTmhrYXdqckpyNzRSSkFBIiwidmVyIjoiMS4wIiwid2lkcyI6WyI2MmU5MDM5NC02OWY1LTQyMzctOTE5MC0wMTIxNzcxNDVlMTAiXSwieG1zX3RjZHQiOjE1NjgxODgyNzZ9.nzU-Yj1uwpqPqcUJVE7iKunPtTCwFbQ4_pE-EiAiVGWpzi6A2e1t9YAW6s2BqnHVltNDO2xBJRZgyjSuetCekuX_nxvwZIU4hDppl5lrt6O85-PtQrYR34DOa05O2fg7a53lhP_b5uSy3XexZpqwNvpbC0dqAictuv59kN6rlZQyUoP_J70jVx-WhXwGQNpgn9uDs11SDgxioKIgrDh0rA1q0kJxJ-4pLbO6l2B2KfL0lrkSJinKFPslwhRhKTHFhqVbdSkiFV7gyK-Kc13iGzXUiB2aHu3M6B-Yy5fmRoF4SMFqFJelLvlctDPUiLK8b9_qQpX60aYIEnbur4amtg\",\n\t\t}\n\t}\n\n\ta.revokeCert = revoke\n\n\ta.challengeStatusMap = make(map[string]ChallengeStatus)\n\ta.challangeCreated = make(map[string]bool)\n\n\ta.serverKey = serverK //used to communicate with ACME server\n\ta.clientKey = clientK\n\n}", "func TypeCreateInstance(type_ Type) *TypeInstance {\n\tc_type := (C.GType)(type_)\n\n\tretC := C.g_type_create_instance(c_type)\n\tretGo := TypeInstanceNewFromC(unsafe.Pointer(retC))\n\n\treturn retGo\n}", "func CreateInstance(computeService *compute.Service) (*compute.Operation, error) {\n\tstartupMetadata := fmt.Sprintf(\"#! /bin/bash\\n \\n echo FILESHARE_MOUNT_PRIMARY=\\\"%s\\\" >> /root/.bashrc \\n echo FILESTORE_IP_PRIMARY=\\\"%s\\\" >> /root/.bashrc \\n echo FILESHARE_NAME_PRIMARY=\\\"%s\\\" >> /root/.bashrc \\n echo FILESHARE_MOUNT_SECONDARY=\\\"%s\\\" >> /root/.bashrc \\n echo FILESHARE_NAME_SECONDARY=\\\"%s\\\" >> /root/.bashrc \\n echo FILESTORE_IP_SECONDARY=\\\"%s\\\" >> /root/.bashrc \\n echo GCP_BUCKET_NAME=\\\"%s\\\" >> /root/.bashrc \\n source /root/.bashrc \\n wget -O /root/google-filestore-backup.sh https://raw.githubusercontent.com/sreesanpd/google-filestore-backup-cloud-function/master/resources/google-filestore-backup.sh \\n chmod +x /root/google-filestore-backup.sh \\n /bin/bash /root/google-filestore-backup.sh \\necho \\\"I am STARTING some work at $(date)\\\" | sudo tee -a $HOME/work.txt\", FileshareMountPrimary, FilestoreIpPrimary, FileshareNamePrimary, FileshareMountSecondary, FileshareNameSecondary, FilestoreIpSecondary, GcpBucketName)\n\tshutdownMetadata := fmt.Sprintf(\"#! /bin/bash\\n\\necho \\\"I am FINISHING some work on %s at $(date)\\\" | sudo tee -a $(HOME)/work.txt\", InstanceName)\n\n\tinstance := &compute.Instance{\n\t\tName: InstanceName,\n\t\tMachineType: fmt.Sprintf(\"zones/%s/machineTypes/f1-micro\", Zone),\n\t\tNetworkInterfaces: []*compute.NetworkInterface{\n\t\t\t{\n\t\t\t\tName: VPCName,\n\t\t\t\tSubnetwork: fmt.Sprintf(\"projects/%s/regions/%s/subnetworks/%s\", ProjectID, Region, SubnetName),\n\t\t\t\tAccessConfigs: []*compute.AccessConfig{\n\t\t\t\t\t{\n\t\t\t\t\t\tName: \"External NAT\",\n\t\t\t\t\t\tType: \"ONE_TO_ONE_NAT\",\n\t\t\t\t\t\tNetworkTier: \"PREMIUM\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\tScheduling: &compute.Scheduling{\n\t\t\tPreemptible: true,\n\t\t},\n\t\tDisks: []*compute.AttachedDisk{\n\t\t\t{\n\t\t\t\tBoot: true, // The first disk must be a boot disk.\n\t\t\t\tAutoDelete: true, //Optional\n\t\t\t\tMode: \"READ_WRITE\", //Mode should be READ_WRITE or READ_ONLY\n\t\t\t\tInterface: \"SCSI\", //SCSI or NVME - NVME only for SSDs\n\t\t\t\tInitializeParams: &compute.AttachedDiskInitializeParams{\n\t\t\t\t\tDiskName: \"worker-instance-boot-disk\",\n\t\t\t\t\tSourceImage: \"projects/debian-cloud/global/images/family/debian-9\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\tMetadata: &compute.Metadata{\n\t\t\tItems: []*compute.MetadataItems{\n\t\t\t\t{\n\t\t\t\t\tKey: \"startup-script\",\n\t\t\t\t\tValue: &startupMetadata,\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tKey: \"shutdown-script\",\n\t\t\t\t\tValue: &shutdownMetadata,\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\tServiceAccounts: []*compute.ServiceAccount{\n\t\t\t{\n\t\t\t\tEmail: \"filestorebackups123-gcs-sa@learn-gcp-173102.iam.gserviceaccount.com\",\n\t\t\t\tScopes: []string{\"https://www.googleapis.com/auth/cloud-platform\"},\n\t\t\t\t\n\t\t\t},\n\t\t\t},\n\n\t\t\n\t}\n\treturn computeService.Instances.Insert(ProjectID, Zone, instance).Do()\n}", "func (a *acceptor) newInstance() {\n\ta.instanceID++\n\ta.initForNewPaxosInstance()\n}", "func NewInstance(name string, opts ...Option) (Instance, error) {\n\tif name == \"\" {\n\t\treturn nil, ErrNoName\n\t}\n\ti := &instance{\n\t\tname: name,\n\t\tbindAddr: \"127.0.0.1\",\n\t\tbindPort: 8100,\n\t\tshutdownCh: make(chan struct{}),\n\t\tpeers: make(map[string]*Peer),\n\t\tcaches: make(map[string]*cache),\n\t\tlogger: log.New(os.Stdout, \"huton\", log.LstdFlags),\n\t\traftApplicationTimeout: 10 * time.Second,\n\t\traftTransportTimeout: 10 * time.Second,\n\t\traftRetainSnapshotCount: 2,\n\t\tserfEventChannel: make(chan serf.Event, 256),\n\t}\n\tfor _, opt := range opts {\n\t\topt(i)\n\t}\n\ti.logger.Println(\"Initializing RPC server...\")\n\tif err := i.setupRPC(); err != nil {\n\t\ti.Shutdown()\n\t\treturn i, err\n\t}\n\ti.logger.Println(\"Initializing Raft cluster...\")\n\tif err := i.setupRaft(); err != nil {\n\t\ti.Shutdown()\n\t\treturn i, err\n\t}\n\tip := net.ParseIP(i.bindAddr)\n\traftAddr := &net.TCPAddr{\n\t\tIP: ip,\n\t\tPort: i.bindPort + 1,\n\t}\n\trpcAddr := &net.TCPAddr{\n\t\tIP: ip,\n\t\tPort: i.bindPort + 2,\n\t}\n\n\ti.logger.Println(\"Initializing Serf cluster...\")\n\tif err := i.setupSerf(raftAddr, rpcAddr); err != nil {\n\t\ti.Shutdown()\n\t\treturn i, err\n\t}\n\tgo i.handleEvents()\n\treturn i, nil\n}", "func (t *LambdaFactory) New(config *trigger.Config) (trigger.Trigger, error) {\n\n\tif singleton == nil {\n\t\tsingleton = &LambdaTrigger{}\n\t\treturn singleton, nil\n\t}\n\n\tlog.RootLogger().Warn(\"Only one lambda trigger instance can be instantiated\")\n\n\treturn nil, nil\n}", "func (o *FakeObject) New(args ...interface{}) Object { return o.Invoke(args) }", "func (c *Client) Create(cfg Function) (err error) {\n\tc.progressListener.SetTotal(4)\n\tdefer c.progressListener.Done()\n\n\t// Initialize, writing out a template implementation and a config file.\n\t// TODO: the Function's Initialize parameters are slightly different than\n\t// the Initializer interface, and can thus cause confusion (one passes an\n\t// optional name the other passes root path). This could easily cause\n\t// confusion and thus we may want to rename Initalizer to the more specific\n\t// task it performs: ContextTemplateWriter or similar.\n\tc.progressListener.Increment(\"Initializing new Function project\")\n\terr = c.Initialize(cfg)\n\tif err != nil {\n\t\treturn\n\t}\n\n\t// Load the now-initialized Function.\n\tf, err := NewFunction(cfg.Root)\n\tif err != nil {\n\t\treturn\n\t}\n\n\t// Build the now-initialized Function\n\tc.progressListener.Increment(\"Building container image\")\n\tif err = c.Build(f.Root); err != nil {\n\t\treturn\n\t}\n\n\t// Deploy the initialized Function, returning its publicly\n\t// addressible name for possible registration.\n\tc.progressListener.Increment(\"Deploying Function to cluster\")\n\tif err = c.Deploy(f.Root); err != nil {\n\t\treturn\n\t}\n\n\t// Create an external route to the Function\n\tc.progressListener.Increment(\"Creating route to Function\")\n\tif err = c.Route(f.Root); err != nil {\n\t\treturn\n\t}\n\n\tc.progressListener.Complete(\"Create complete\")\n\n\t// TODO: use the knative client during deployment such that the actual final\n\t// route can be returned from the deployment step, passed to the DNS Router\n\t// for routing actual traffic, and returned here.\n\tif c.verbose {\n\t\tfmt.Printf(\"https://%v/\\n\", f.Name)\n\t}\n\treturn\n}", "func (c *TestClient) CreateTargetInstance(project, zone string, ti *compute.TargetInstance) error {\n\tif c.CreateTargetInstanceFn != nil {\n\t\treturn c.CreateTargetInstanceFn(project, zone, ti)\n\t}\n\treturn c.client.CreateTargetInstance(project, zone, ti)\n}", "func (env *Environment) New(testFn TestFunc) *Fixture {\n\treturn &Fixture{\n\t\tname: runtime.FuncForPC(reflect.ValueOf(testFn).Pointer()).Name(),\n\t\ttestFn: testFn,\n\t\tenv: env,\n\t}\n}", "func NewCreateFnDefault(code int) *CreateFnDefault {\n\treturn &CreateFnDefault{\n\t\t_statusCode: code,\n\t}\n}", "func (h *HealthCheck) createInstance() error {\n\tif h.frameworkError != nil {\n\t\treturn h.frameworkError\n\t}\n\tglog.V(4).Info(\"Creating a ServiceInstance\")\n\tinstance := &v1beta1.ServiceInstance{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tName: h.instanceName,\n\t\t\tNamespace: h.namespace.Name,\n\t\t},\n\t\tSpec: v1beta1.ServiceInstanceSpec{\n\t\t\tPlanReference: v1beta1.PlanReference{\n\t\t\t\tClusterServiceClassExternalName: h.serviceclassName,\n\t\t\t\tClusterServicePlanExternalName: \"default\",\n\t\t\t},\n\t\t},\n\t}\n\toperationStartTime := time.Now()\n\tvar err error\n\tinstance, err = h.serviceCatalogClientSet.ServicecatalogV1beta1().ServiceInstances(h.namespace.Name).Create(instance)\n\tif err != nil {\n\t\treturn h.setError(\"error creating instance: %v\", err.Error())\n\t}\n\n\tif instance == nil {\n\t\treturn h.setError(\"error creating instance - instance is null\")\n\t}\n\n\tglog.V(4).Info(\"Waiting for ServiceInstance to be ready\")\n\terr = util.WaitForInstanceCondition(h.serviceCatalogClientSet.ServicecatalogV1beta1(),\n\t\th.namespace.Name,\n\t\th.instanceName,\n\t\tv1beta1.ServiceInstanceCondition{\n\t\t\tType: v1beta1.ServiceInstanceConditionReady,\n\t\t\tStatus: v1beta1.ConditionTrue,\n\t\t},\n\t)\n\tif err != nil {\n\t\treturn h.setError(\"instance not ready: %v\", err.Error())\n\t}\n\tReportOperationCompleted(\"create_instance\", operationStartTime)\n\n\tglog.V(4).Info(\"Verifing references are resolved\")\n\tsc, err := h.serviceCatalogClientSet.ServicecatalogV1beta1().ServiceInstances(h.namespace.Name).Get(h.instanceName, metav1.GetOptions{})\n\tif err != nil {\n\t\treturn h.setError(\"error getting instance: %v\", err.Error())\n\t}\n\n\tif sc.Spec.ClusterServiceClassRef == nil {\n\t\treturn h.setError(\"ClusterServiceClassRef should not be null\")\n\t}\n\tif sc.Spec.ClusterServicePlanRef == nil {\n\t\treturn h.setError(\"ClusterServicePlanRef should not be null\")\n\t}\n\n\tif strings.Compare(sc.Spec.ClusterServiceClassRef.Name, h.serviceclassID) != 0 {\n\t\treturn h.setError(\"ClusterServiceClassRef.Name error: %v != %v\", sc.Spec.ClusterServiceClassRef.Name, h.serviceclassID)\n\t}\n\tif strings.Compare(sc.Spec.ClusterServicePlanRef.Name, h.serviceplanID) != 0 {\n\t\treturn h.setError(\"sc.Spec.ClusterServicePlanRef.Name error: %v != %v\", sc.Spec.ClusterServicePlanRef.Name, h.serviceplanID)\n\t}\n\treturn nil\n}", "func NewInstance(projectID string, metaData map[string]string, secGroup []map[string]interface{}) openstack.InstanceClient {\n\treturn &Instance{\n\t\tprojectID: projectID,\n\t\tmetaData: metaData,\n\t\tsecGroup: secGroup,\n\t\tcreated: time.Now(),\n\t}\n}", "func TestNewInstance(t *testing.T) {\n\tif _, err := NewInstance(nil); err == nil {\n\t\tt.Error(\"NewInstance: expected error with nil database handle\")\n\t}\n\n\tRunWithDB(func(db *sql.DB) {\n\t\tif _, err := NewInstance(db); err != nil {\n\t\t\tt.Fatal(\"NewInstance: got error:\\n\", err)\n\t\t}\n\t})\n}", "func (a *Client) CreateInstanceTransformationByObjectName2(params *CreateInstanceTransformationByObjectName2Params) (*CreateInstanceTransformationByObjectName2OK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewCreateInstanceTransformationByObjectName2Params()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"createInstanceTransformationByObjectName2\",\n\t\tMethod: \"POST\",\n\t\tPathPattern: \"/instances/{id}/transformations/{objectName}\",\n\t\tProducesMediaTypes: []string{\"\"},\n\t\tConsumesMediaTypes: []string{\"\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &CreateInstanceTransformationByObjectName2Reader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*CreateInstanceTransformationByObjectName2OK), nil\n\n}", "func New(opts ...InstanceOpt) (*Instance, error) {\n\tinstance := &Instance{}\n\n\tfor _, opt := range opts {\n\t\tif err := opt(instance); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tpool, err := dockertest.NewPool(\"\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresource, err := pool.Run(\"localstack/localstack\", \"\", []string{instance.serviceString()})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\twithDefaults(instance)\n\tinstance.resolver = instance.makeResolver()\n\tinstance.pool = pool\n\tinstance.resource = resource\n\n\treturn instance, nil\n}", "func (p FnBabysitErrorPolicy) New() ErrorPolicyFn {\n\treturn ErrorPolicyFn(p)\n}", "func NewSDKActor(execute func(string) string) SDKActor {\n sdkActor := SDKActor{}\n sdkActor.connector = newConnector()\n sdkActor.execute = execute\n return sdkActor\n}", "func (f Function) New(ctx string) error {\n\treturn f.unsafeWrap(errors.New(ctx), ctx, \"\")\n}", "func New(cfg Config) Backend {\n\treturn &instance{\n\t\tcli: cfg.Client,\n\t\tcfg: cfg.RestConfig,\n\t\tinitImage: cfg.InitImage,\n\t\tnamespace: cfg.Namespace,\n\t\trandomPorts: map[int]int{},\n\t\ttimeOut: int(cfg.TimeOut.Seconds()),\n\t}\n}", "func Open(name string, syscallService *bridge.SyscallService, db db.Database) (InstanceCreator, error) {\n\treturn defaultRegistry.Open(name, syscallService, db)\n}", "func New(flags *pflag.FlagSet) *Instance {\n\texecutionInstance := &execution.Instance{\n\t\tLogDir: logDir,\n\t\tGeneratedDir: generatedDir,\n\t\tTemplateDir: templateDir,\n\t\tDoRunParallel: getBoolFlagValue(flags, flag.DoRunParallel),\n\t\tDryRunEnabled: getBoolFlagValue(flags, flag.DryRun),\n\t\tReRun: getBoolFlagValue(flags, flag.ReRun),\n\t\tIgnoreIfPrefix: ignoreIfPrefix,\n\t\tTimeoutInterval: time.Minute * 15, //change later\n\t\tState: execution.State{\n\t\t\tStateFilePath: stateFilePath,\n\t\t\tStateFileDefaultname: stateFileDefaultName,\n\t\t},\n\t}\n\texecutionInstance.Init()\n\n\treturn &Instance{\n\t\tConfig: make(map[string]interface{}),\n\t\tFlags: flags,\n\t\tInstance: executionInstance,\n\t\tStartTime: time.Now(),\n\t}\n}", "func NewFactory() processor.Factory {\n\treturn processor.NewFactory(\n\t\tmetadata.Type,\n\t\tcreateDefaultConfig,\n\t\tprocessor.WithMetrics(createMetricsProcessor, metadata.MetricsStability))\n}", "func NewFactory() processor.Factory {\n\treturn processor.NewFactory(\n\t\tmetadata.Type,\n\t\tcreateDefaultConfig,\n\t\tprocessor.WithMetrics(createMetricsProcessor, metadata.MetricsStability))\n}", "func (asf authorizationStrategyFactory) Create(c *Credentials) Strategy {\n\treturn newExternalTokenStrategy(asf.create(c))\n}", "func NewFactory() extension.Factory {\n\treturn extension.NewFactory(\n\t\tmetadata.Type,\n\t\tcreateDefaultConfig,\n\t\tcreateExtension,\n\t\tmetadata.ExtensionStability,\n\t)\n}", "func NewFactory() extension.Factory {\n\treturn extension.NewFactory(\n\t\tmetadata.Type,\n\t\tcreateDefaultConfig,\n\t\tcreateExtension,\n\t\tmetadata.ExtensionStability,\n\t)\n}", "func NewCreateFnNotFound() *CreateFnNotFound {\n\treturn &CreateFnNotFound{}\n}", "func NewInstance(ctx *pulumi.Context,\n\tname string, args *InstanceArgs, opts ...pulumi.ResourceOption) (*Instance, error) {\n\tif args == nil {\n\t\targs = &InstanceArgs{}\n\t}\n\n\tif args.AdminPass != nil {\n\t\targs.AdminPass = pulumi.ToSecret(args.AdminPass).(pulumi.StringPtrInput)\n\t}\n\tsecrets := pulumi.AdditionalSecretOutputs([]string{\n\t\t\"adminPass\",\n\t})\n\topts = append(opts, secrets)\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Instance\n\terr := ctx.RegisterResource(\"openstack:compute/instance:Instance\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func NewInstance(ctx *pulumi.Context,\n\tname string, args *InstanceArgs, opts ...pulumi.ResourceOption) (*Instance, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.InstanceId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'InstanceId'\")\n\t}\n\tif args.MachineType == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'MachineType'\")\n\t}\n\treplaceOnChanges := pulumi.ReplaceOnChanges([]string{\n\t\t\"instanceId\",\n\t\t\"location\",\n\t\t\"project\",\n\t})\n\topts = append(opts, replaceOnChanges)\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Instance\n\terr := ctx.RegisterResource(\"google-native:notebooks/v1:Instance\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func NewInstanceExemplar(srcPtr interface{}) factory.Exemplar {\n\n\ttyp := reflect.TypeOf(srcPtr)\n\n\tif typ.Kind() != reflect.Ptr {\n\t\tpanic(\"Should be a pointer\")\n\t}\n\n\ttyp = typ.Elem()\n\n\treturn &nie{\n\t\ttyp: typ,\n\t\th: fmt.Sprintf(\"%x\", sha256.Sum256([]byte(typ.String()))),\n\t}\n}", "func CreateProtocolInstance(action ActionProps, filter FilteringProps) *ProtocolInstance {\r\n\treturn &ProtocolInstance{\r\n\t\tAction: action,\r\n\t\tFilter: filter,\r\n\t}\r\n}", "func (fd FeatureDefinition) Create() Feature {\n\tfeature := C.OGR_F_Create(fd.cval)\n\treturn Feature{feature}\n}", "func (f *fakeLB) CreateForwardingRule(_ context.Context, _ string, _ *govultr.ForwardingRule) (*govultr.ForwardingRule, *http.Response, error) {\n\tpanic(\"implement me\")\n}", "func (_e *MockTransport_Expecter) CreateTransportInstance(transportUrl interface{}, _a1 interface{}, _options ...interface{}) *MockTransport_CreateTransportInstance_Call {\n\treturn &MockTransport_CreateTransportInstance_Call{Call: _e.mock.On(\"CreateTransportInstance\",\n\t\tappend([]interface{}{transportUrl, _a1}, _options...)...)}\n}", "func New(instance InstanceIdentifier, state *state.State, name string, conf config.Device, volatileGet VolatileGetter, volatileSet VolatileSetter) (Device, error) {\n\tdevFunc := devTypes[conf[\"type\"]]\n\n\t// Check if top-level type is recognised, if it is known type it will return a function.\n\tif devFunc == nil {\n\t\treturn nil, ErrUnsupportedDevType\n\t}\n\n\t// Run the device create function and check it succeeds.\n\tdev := devFunc(conf)\n\tif dev == nil {\n\t\treturn nil, ErrUnsupportedDevType\n\t}\n\n\t// Init the device and run validation of supplied config.\n\tdev.init(instance, state, name, conf, volatileGet, volatileSet)\n\terr := dev.validateConfig()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn dev, nil\n}", "func (f *Factory) Create(file string, opts ...FactoryOpt) JGameObjectHandle {\n\tdefer debug.Trace().UnTrace()\n\n\t// read in game object data\n\n\tdata, err := support.OpenFile(file)\n\tif err != nil {\n\t\tsupport.LogFatal(\"Failed to open Config file: \" + file)\n\t}\n\n\tholder, err := support.ReadData(data)\n\tif err != nil {\n\t\tsupport.LogFatal(\"Failed to read in Config file: \" + file)\n\t\treturn nil\n\t}\n\n\tm := v.(map[string]interface{})\n\ttypename, err := m[\"Type\"]\n\tif err != nil {\n\t\tlog.Panic(err)\n\t}\n\tobj := factoryFunc(typename.(string))\n\tSerializeInPlace(obj, holder)\n\n\t// apply options to game object\n\tfor _, opt := range opts {\n\t\topt(obj)\n\t}\n\n\t// check for transform, check for dispatcher, name etc\n\n\tif obj.Name() == \"\" {\n\t\tobj.SetName(fmt.Sprint(\"obj\", len(f.ObjList)))\n\t}\n\n\t// add game object to factory\n\n\tf.ObjList = append(f.ObjList, obj)\n\th := JGameObjectHandle{id: len(f.ObjList)}\n\tf.NameMap[obj.Name()] = h\n\treturn h\n}", "func MakeInstantiable(m *MockWasmer) {\n\tm.CreateFn = HashOnlyCreateFn\n\tm.InstantiateFn = NoOpInstantiateFn\n\tm.AnalyzeCodeFn = WithoutIBCAnalyzeFn\n}", "func LookupInstance(ctx *pulumi.Context, args *LookupInstanceArgs, opts ...pulumi.InvokeOption) (*LookupInstanceResult, error) {\n\topts = internal.PkgInvokeDefaultOpts(opts)\n\tvar rv LookupInstanceResult\n\terr := ctx.Invoke(\"google-native:compute/v1:getInstance\", args, &rv, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &rv, nil\n}", "func NewFactory() processor.Factory {\n\treturn processor.NewFactory(\n\t\tmetadata.Type,\n\t\tcreateDefaultConfig,\n\t\tprocessor.WithTraces(createTracesProcessor, metadata.TracesStability))\n}", "func (c *TestClient) CreateInstanceBeta(project, zone string, i *computeBeta.Instance) error {\n\tif c.CreateInstanceBetaFn != nil {\n\t\treturn c.CreateInstanceBetaFn(project, zone, i)\n\t}\n\treturn c.client.CreateInstanceBeta(project, zone, i)\n}", "func (c *Client) CreateInstance(name, zone, machineType, ipxeURL string) error {\n\treq := &packngo.DeviceCreateRequest{\n\t\tHostname: name,\n\t\tPlan: machineType,\n\t\tProjectID: c.projectID,\n\t\tFacility: []string{zone},\n\t\tIPXEScriptURL: ipxeURL,\n\t\tOS: \"custom_ipxe\",\n\t\tDescription: \"eden test vm\",\n\t\tBillingCycle: \"hourly\",\n\t}\n\t_, _, err := c.client.Devices.Create(req)\n\treturn err\n}", "func New(config *Config) functions.Runner {\n\treturn &impl{*config}\n}", "func Factory() interface{} {\n\tinstance := new(components.Greeter)\n\tfmt.Println(\"Type\", reflect.TypeOf(instance))\n\treflect.New()\n\treturn instance\n}", "func (g GCPClient) CreateInstance(name, image, zone, machineType string, disks Disks, data *string, nested, vtpm, replace bool) error {\n\tif replace {\n\t\tif err := g.DeleteInstance(name, zone, true); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tlog.Infof(\"Creating instance %s from image %s (type: %s in %s)\", name, image, machineType, zone)\n\n\tenabled := new(string)\n\t*enabled = \"1\"\n\n\tk, err := ssh.NewPublicKey(g.privKey.Public())\n\tif err != nil {\n\t\treturn err\n\t}\n\tsshKey := new(string)\n\t*sshKey = fmt.Sprintf(\"moby:%s moby\", string(ssh.MarshalAuthorizedKey(k)))\n\n\t// check provided image to be compatible with provided options\n\top, err := g.compute.Images.Get(g.projectName, image).Do()\n\tif err != nil {\n\t\treturn err\n\t}\n\tuefiCompatible := false\n\tfor _, feature := range op.GuestOsFeatures {\n\t\tif feature != nil && feature.Type == uefiCompatibleFeature {\n\t\t\tuefiCompatible = true\n\t\t\tbreak\n\t\t}\n\t}\n\tif vtpm && !uefiCompatible {\n\t\treturn fmt.Errorf(\"cannot use vTPM without UEFI_COMPATIBLE image\")\n\t}\n\t// we should check for nested\n\tvmxLicense := false\n\tfor _, license := range op.Licenses {\n\t\t// we omit hostname and version when define license\n\t\tif strings.HasSuffix(license, vmxImageLicence) {\n\t\t\tvmxLicense = true\n\t\t\tbreak\n\t\t}\n\t}\n\tif nested && !vmxLicense {\n\t\treturn fmt.Errorf(\"cannot use nested virtualization without enable-vmx image\")\n\t}\n\n\tinstanceDisks := []*compute.AttachedDisk{\n\t\t{\n\t\t\tAutoDelete: true,\n\t\t\tBoot: true,\n\t\t\tInitializeParams: &compute.AttachedDiskInitializeParams{\n\t\t\t\tSourceImage: fmt.Sprintf(\"global/images/%s\", image),\n\t\t\t},\n\t\t},\n\t}\n\n\tfor i, disk := range disks {\n\t\tvar diskName string\n\t\tif disk.Path != \"\" {\n\t\t\tdiskName = disk.Path\n\t\t} else {\n\t\t\tdiskName = fmt.Sprintf(\"%s-disk-%d\", name, i)\n\t\t}\n\t\tvar diskSizeGb int64\n\t\tif disk.Size == 0 {\n\t\t\tdiskSizeGb = int64(1)\n\t\t} else {\n\t\t\tdiskSizeGb = int64(convertMBtoGB(disk.Size))\n\t\t}\n\t\tdiskObj := &compute.Disk{Name: diskName, SizeGb: diskSizeGb}\n\t\tif vtpm {\n\t\t\tdiskObj.GuestOsFeatures = []*compute.GuestOsFeature{\n\t\t\t\t{Type: uefiCompatibleFeature},\n\t\t\t}\n\t\t}\n\t\tdiskOp, err := g.compute.Disks.Insert(g.projectName, zone, diskObj).Do()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif err := g.pollZoneOperationStatus(diskOp.Name, zone); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tinstanceDisks = append(instanceDisks, &compute.AttachedDisk{\n\t\t\tAutoDelete: true,\n\t\t\tBoot: false,\n\t\t\tSource: fmt.Sprintf(\"zones/%s/disks/%s\", zone, diskName),\n\t\t})\n\t}\n\n\tinstanceObj := &compute.Instance{\n\t\tMachineType: fmt.Sprintf(\"zones/%s/machineTypes/%s\", zone, machineType),\n\t\tName: name,\n\t\tDisks: instanceDisks,\n\t\tNetworkInterfaces: []*compute.NetworkInterface{\n\t\t\t{\n\t\t\t\tNetwork: \"global/networks/default\",\n\t\t\t\tAccessConfigs: []*compute.AccessConfig{\n\t\t\t\t\t{\n\t\t\t\t\t\tType: \"ONE_TO_ONE_NAT\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\tMetadata: &compute.Metadata{\n\t\t\tItems: []*compute.MetadataItems{\n\t\t\t\t{\n\t\t\t\t\tKey: \"serial-port-enable\",\n\t\t\t\t\tValue: enabled,\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tKey: \"ssh-keys\",\n\t\t\t\t\tValue: sshKey,\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tKey: \"user-data\",\n\t\t\t\t\tValue: data,\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\n\tif nested {\n\t\tinstanceObj.MinCpuPlatform = \"Intel Haswell\"\n\t}\n\tif vtpm {\n\t\tinstanceObj.ShieldedInstanceConfig = &compute.ShieldedInstanceConfig{EnableVtpm: true}\n\t}\n\n\t// Don't wait for operation to complete!\n\t// A headstart is needed as by the time we've polled for this event to be\n\t// completed, the instance may have already terminated\n\t_, err = g.compute.Instances.Insert(g.projectName, zone, instanceObj).Do()\n\tif err != nil {\n\t\treturn err\n\t}\n\tlog.Infof(\"Instance created\")\n\treturn nil\n}" ]
[ "0.6088032", "0.58898586", "0.58512163", "0.58139443", "0.5727347", "0.5646979", "0.56438726", "0.56116194", "0.5563739", "0.5519419", "0.5513493", "0.55097413", "0.54723245", "0.54625016", "0.54587626", "0.54305315", "0.54281205", "0.54214644", "0.53558147", "0.533639", "0.5320407", "0.53186584", "0.5311436", "0.5304801", "0.52552724", "0.52458864", "0.5236449", "0.5231713", "0.5223588", "0.5222819", "0.5215566", "0.5205292", "0.51820046", "0.51457447", "0.51357347", "0.5135422", "0.512017", "0.5112746", "0.5109049", "0.5103823", "0.50827813", "0.50787127", "0.50778884", "0.5068884", "0.50601006", "0.50466627", "0.50221956", "0.50092953", "0.50079274", "0.50055647", "0.5003814", "0.49901095", "0.49886492", "0.49867132", "0.498655", "0.4983844", "0.49813867", "0.49760452", "0.49730432", "0.49680212", "0.49674425", "0.49605426", "0.49574465", "0.49565935", "0.49512377", "0.494899", "0.49464834", "0.49456513", "0.4941114", "0.49402946", "0.49378714", "0.4936051", "0.49346614", "0.49304494", "0.49159327", "0.49112737", "0.49010935", "0.48888662", "0.48888662", "0.4887517", "0.48828205", "0.48828205", "0.48798907", "0.4878398", "0.48775065", "0.48747104", "0.48719734", "0.48676667", "0.4864388", "0.48638874", "0.48609877", "0.4855799", "0.48528045", "0.4846115", "0.48440433", "0.48403803", "0.48377368", "0.48353025", "0.48274902", "0.4823234" ]
0.54121184
18
CreateNetwork uses the override method CreateNetworkFn or the real implementation.
func (c *TestClient) CreateNetwork(project string, n *compute.Network) error { if c.CreateNetworkFn != nil { return c.CreateNetworkFn(project, n) } return c.client.CreateNetwork(project, n) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (p Pipeline) CreateNetwork() error {\n\treturn p.localRunner.NetworkCreator(p.Network)()\n}", "func CreateNetwork(ctx context.Context) (*tc.Network, string) {\n\trandomUuid, _ := uuid.NewRandom()\n\tnetworkName := randomUuid.String()\n\tnet, err := tc.GenericNetwork(ctx, tc.GenericNetworkRequest{\n\t\tNetworkRequest: tc.NetworkRequest{\n\t\t\tName: networkName,\n\t\t\tCheckDuplicate: true,\n\t\t},\n\t})\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn &net, networkName\n}", "func (s *Stack) CreateNetwork(req abstract.NetworkRequest) (*abstract.Network, fail.Error) {\n\t// disable subnetwork auto-creation\n\tne := compute.Network{\n\t\tName: s.GcpConfig.NetworkName,\n\t\tAutoCreateSubnetworks: false,\n\t\tForceSendFields: []string{\"AutoCreateSubnetworks\"},\n\t}\n\n\tcompuService := s.ComputeService\n\n\trecreateSafescaleNetwork := true\n\trecnet, err := compuService.Networks.Get(s.GcpConfig.ProjectID, ne.Name).Do()\n\tif recnet != nil && err == nil {\n\t\trecreateSafescaleNetwork = false\n\t} else if err != nil {\n\t\tif gerr, ok := err.(*googleapi.Error); ok {\n\t\t\tif gerr.Code != 404 {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t} else {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tif recreateSafescaleNetwork {\n\t\topp, err := compuService.Networks.Insert(s.GcpConfig.ProjectID, &ne).Context(context.Background()).Do()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\toco := OpContext{\n\t\t\tOperation: opp,\n\t\t\tProjectID: s.GcpConfig.ProjectID,\n\t\t\tService: compuService,\n\t\t\tDesiredState: \"DONE\",\n\t\t}\n\n\t\terr = waitUntilOperationIsSuccessfulOrTimeout(oco, temporal.GetMinDelay(), 2*temporal.GetContextTimeout())\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tnecreated, err := compuService.Networks.Get(s.GcpConfig.ProjectID, ne.Name).Do()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tnet := abstract.NewNetwork()\n\tnet.ID = strconv.FormatUint(necreated.Id, 10)\n\tnet.Name = necreated.Name\n\n\t// Create subnetwork\n\n\ttheRegion := s.GcpConfig.Region\n\n\tsubnetReq := compute.Subnetwork{\n\t\tIpCidrRange: req.CIDR,\n\t\tName: req.Name,\n\t\tNetwork: fmt.Sprintf(\"projects/%s/global/networks/%s\", s.GcpConfig.ProjectID, s.GcpConfig.NetworkName),\n\t\tRegion: theRegion,\n\t}\n\n\topp, err := compuService.Subnetworks.Insert(\n\t\ts.GcpConfig.ProjectID, theRegion, &subnetReq,\n\t).Context(context.Background()).Do()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\toco := OpContext{\n\t\tOperation: opp,\n\t\tProjectID: s.GcpConfig.ProjectID,\n\t\tService: compuService,\n\t\tDesiredState: \"DONE\",\n\t}\n\n\terr = waitUntilOperationIsSuccessfulOrTimeout(oco, temporal.GetMinDelay(), 2*temporal.GetContextTimeout())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tgcpSubNet, err := compuService.Subnetworks.Get(s.GcpConfig.ProjectID, theRegion, req.Name).Do()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// FIXME: Add properties and GatewayID\n\tsubnet := abstract.NewNetwork()\n\tsubnet.ID = strconv.FormatUint(gcpSubNet.Id, 10)\n\tsubnet.Name = gcpSubNet.Name\n\tsubnet.CIDR = gcpSubNet.IpCidrRange\n\tsubnet.IPVersion = ipversion.IPv4\n\n\tbuildNewRule := true\n\tfirewallRuleName := fmt.Sprintf(\"%s-%s-all-in\", s.GcpConfig.NetworkName, gcpSubNet.Name)\n\n\tfws, err := compuService.Firewalls.Get(s.GcpConfig.ProjectID, firewallRuleName).Do()\n\tif fws != nil && err == nil {\n\t\tbuildNewRule = false\n\t} else if err != nil {\n\t\tif gerr, ok := err.(*googleapi.Error); ok {\n\t\t\tif gerr.Code != 404 {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t} else {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tif buildNewRule {\n\t\tfiw := compute.Firewall{\n\t\t\tAllowed: []*compute.FirewallAllowed{\n\t\t\t\t{\n\t\t\t\t\tIPProtocol: \"all\",\n\t\t\t\t},\n\t\t\t},\n\t\t\tDirection: \"INGRESS\",\n\t\t\tDisabled: false,\n\t\t\tName: firewallRuleName,\n\t\t\tNetwork: fmt.Sprintf(\n\t\t\t\t\"https://www.googleapis.com/compute/v1/projects/%s/global/networks/%s\", s.GcpConfig.ProjectID,\n\t\t\t\ts.GcpConfig.NetworkName,\n\t\t\t),\n\t\t\tPriority: 999,\n\t\t\tSourceRanges: []string{\"0.0.0.0/0\"},\n\t\t}\n\n\t\topp, err = compuService.Firewalls.Insert(s.GcpConfig.ProjectID, &fiw).Do()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\toco = OpContext{\n\t\t\tOperation: opp,\n\t\t\tProjectID: s.GcpConfig.ProjectID,\n\t\t\tService: compuService,\n\t\t\tDesiredState: \"DONE\",\n\t\t}\n\n\t\terr = waitUntilOperationIsSuccessfulOrTimeout(oco, temporal.GetMinDelay(), temporal.GetHostTimeout())\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tbuildNewNATRule := true\n\tnatRuleName := fmt.Sprintf(\"%s-%s-nat-allowed\", s.GcpConfig.NetworkName, gcpSubNet.Name)\n\n\trfs, err := compuService.Routes.Get(s.GcpConfig.ProjectID, natRuleName).Do()\n\tif rfs != nil && err == nil {\n\t\tbuildNewNATRule = false\n\t} else if err != nil {\n\t\tif gerr, ok := err.(*googleapi.Error); ok {\n\t\t\tif gerr.Code != 404 {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t} else {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tif buildNewNATRule {\n\t\troute := &compute.Route{\n\t\t\tDestRange: \"0.0.0.0/0\",\n\t\t\tName: natRuleName,\n\t\t\tNetwork: fmt.Sprintf(\n\t\t\t\t\"https://www.googleapis.com/compute/v1/projects/%s/global/networks/%s\", s.GcpConfig.ProjectID,\n\t\t\t\ts.GcpConfig.NetworkName,\n\t\t\t),\n\t\t\tNextHopInstance: fmt.Sprintf(\n\t\t\t\t\"projects/%s/zones/%s/instances/gw-%s\", s.GcpConfig.ProjectID, s.GcpConfig.Zone, req.Name,\n\t\t\t),\n\t\t\tPriority: 800,\n\t\t\tTags: []string{fmt.Sprintf(\"no-ip-%s\", gcpSubNet.Name)},\n\t\t}\n\t\topp, err := compuService.Routes.Insert(s.GcpConfig.ProjectID, route).Do()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\toco = OpContext{\n\t\t\tOperation: opp,\n\t\t\tProjectID: s.GcpConfig.ProjectID,\n\t\t\tService: compuService,\n\t\t\tDesiredState: \"DONE\",\n\t\t}\n\n\t\terr = waitUntilOperationIsSuccessfulOrTimeout(oco, temporal.GetMinDelay(), 2*temporal.GetContextTimeout())\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t}\n\n\t// FIXME: Validation before return...\n\treturn subnet, nil\n}", "func (net *NetworkCreateInput) CreateNetwork() (CreateNetworkResponse, error) {\n\n\tif status := support.DoesCloudSupports(strings.ToLower(net.Cloud.Name)); status != true {\n\t\treturn CreateNetworkResponse{}, fmt.Errorf(common.DefaultCloudResponse + \"CreateNetwork\")\n\t}\n\n\tswitch strings.ToLower(net.Cloud.Name) {\n\tcase \"aws\":\n\n\t\t// Gets the establish session so that it can carry out the process in cloud\n\t\tsess := (net.Cloud.Client).(*session.Session)\n\n\t\t//authorizing to request further\n\t\tauthinpt := auth.EstablishConnectionInput{Region: net.Cloud.Region, Resource: \"ec2\", Session: sess}\n\n\t\t// Fetching all the networks across cloud aws\n\t\tnetworkin := new(awsnetwork.NetworkCreateInput)\n\t\tnetworkin.Name = net.Name\n\t\tnetworkin.VpcCidr = net.VpcCidr\n\t\tnetworkin.SubCidrs = net.SubCidr\n\t\tnetworkin.Type = net.Type\n\t\tnetworkin.Ports = net.Ports\n\t\tnetworkin.GetRaw = net.Cloud.GetRaw\n\t\tresponse, netErr := networkin.CreateNetwork(authinpt)\n\t\tif netErr != nil {\n\t\t\treturn CreateNetworkResponse{}, netErr\n\t\t}\n\t\treturn CreateNetworkResponse{AwsResponse: response}, nil\n\n\tcase \"azure\":\n\t\treturn CreateNetworkResponse{}, fmt.Errorf(common.DefaultAzResponse)\n\tcase \"gcp\":\n\t\treturn CreateNetworkResponse{}, fmt.Errorf(common.DefaultGcpResponse)\n\tcase \"openstack\":\n\t\treturn CreateNetworkResponse{}, fmt.Errorf(common.DefaultOpResponse)\n\tdefault:\n\t\treturn CreateNetworkResponse{}, fmt.Errorf(common.DefaultCloudResponse + \"CreateNetwork\")\n\t}\n}", "func (l *Libvirt) NetworkCreate(Net Network) (err error) {\n\tvar buf []byte\n\n\targs := NetworkCreateArgs {\n\t\tNet: Net,\n\t}\n\n\tbuf, err = encode(&args)\n\tif err != nil {\n\t\treturn\n\t}\n\n\n\t_, err = l.requestStream(39, constants.Program, buf, nil, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func CreateNetwork(ctx context.Context, ow *rpc.OutputWriter, cli *client.Client, networkID string, netcfg network.IPAMConfig) Fixer {\n\treturn func() (string, error) {\n\t\t_, err := docker.EnsureBridgeNetwork(ctx, ow, cli, networkID, false, netcfg)\n\t\tif err != nil {\n\t\t\treturn \"could not create network.\", err\n\t\t}\n\t\treturn \"network created.\", nil\n\t}\n}", "func NewNetwork(ctx *pulumi.Context,\n\tname string, args *NetworkArgs, opts ...pulumi.ResourceOpt) (*Network, error) {\n\tinputs := make(map[string]interface{})\n\tif args == nil {\n\t\tinputs[\"attachable\"] = nil\n\t\tinputs[\"checkDuplicate\"] = nil\n\t\tinputs[\"driver\"] = nil\n\t\tinputs[\"ingress\"] = nil\n\t\tinputs[\"internal\"] = nil\n\t\tinputs[\"ipamConfigs\"] = nil\n\t\tinputs[\"ipamDriver\"] = nil\n\t\tinputs[\"ipv6\"] = nil\n\t\tinputs[\"labels\"] = nil\n\t\tinputs[\"name\"] = nil\n\t\tinputs[\"options\"] = nil\n\t} else {\n\t\tinputs[\"attachable\"] = args.Attachable\n\t\tinputs[\"checkDuplicate\"] = args.CheckDuplicate\n\t\tinputs[\"driver\"] = args.Driver\n\t\tinputs[\"ingress\"] = args.Ingress\n\t\tinputs[\"internal\"] = args.Internal\n\t\tinputs[\"ipamConfigs\"] = args.IpamConfigs\n\t\tinputs[\"ipamDriver\"] = args.IpamDriver\n\t\tinputs[\"ipv6\"] = args.Ipv6\n\t\tinputs[\"labels\"] = args.Labels\n\t\tinputs[\"name\"] = args.Name\n\t\tinputs[\"options\"] = args.Options\n\t}\n\tinputs[\"scope\"] = nil\n\ts, err := ctx.RegisterResource(\"docker:index/network:Network\", name, true, inputs, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Network{s: s}, nil\n}", "func (d *Driver) CreateNetwork(r *sdk.CreateNetworkRequest) error {\n\tvar netCidr *net.IPNet\n\tvar netGw string\n\tvar err error\n\tlog.Debugf(\"Network Create Called: [ %+v ]\", r)\n\tfor _, v4 := range r.IPv4Data {\n\t\tnetGw = v4.Gateway\n\t\t_, netCidr, err = net.ParseCIDR(v4.Pool)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Parse and validate the config. It should not be conflict with existing networks' config\n\tconfig, err := parseNetworkOptions(r.NetworkID, r.Options)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Generate a name for what will be the sandbox side pipe interface\n\tcontainerIfName, err := d.getContainerIfName(r)\n\tif err != nil {\n\t\treturn err\n\t}\n\tlog.Debugf(\"containerIfName:%v\", containerIfName)\n\tconfig.ContainerIfName = containerIfName\n\n\tn := &network{\n\t\tid: r.NetworkID,\n\t\tconfig: config,\n\t\tendpoints: endpointTable{},\n\t\tcidr: netCidr,\n\t\tgateway: netGw,\n\t}\n\n\tbName, err := getBridgeName(r)\n\tconfig.BridgeName = bName\n\tlog.Debugf(\"bridgeName:%v\", bName)\n\n\t// Initialize handle when needed\n\td.Lock()\n\tif d.nlh == nil {\n\t\td.nlh = NlHandle()\n\t}\n\td.Unlock()\n\n\t// Create or retrieve the bridge L3 interface\n\tbridgeIface, err := newInterface(d.nlh, bName)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tn.bridge = bridgeIface\n\tsetupDevice(bridgeIface)\n\tsetupDeviceUp(config, bridgeIface)\n\n\td.addNetwork(n)\n\treturn nil\n}", "func newNetwork(networkType NetworkModel) network {\n\tswitch networkType {\n\tcase NoopNetworkModel:\n\t\treturn &noopNetwork{}\n\tcase DefaultNetworkModel:\n\t\treturn &defNetwork{}\n\tdefault:\n\t\treturn &noopNetwork{}\n\t}\n}", "func networkCreateExample() string {\n\treturn `$ pouch network create -n pouchnet -d bridge --gateway 192.168.1.1 --subnet 192.168.1.0/24\npouchnet: e1d541722d68dc5d133cca9e7bd8fd9338603e1763096c8e853522b60d11f7b9`\n}", "func CreateNetwork(cli *docker.Client, config *NetworkConfig) (*docker.Network, error) {\r\n\t// Setting network\r\n\tvar netoworkOptions = docker.CreateNetworkOptions{\r\n\t\tCheckDuplicate: true,\r\n\t\tDriver: \"bridge\",\r\n\t\tName: config.NetoworkName,\r\n\t\tIPAM: docker.IPAMOptions{\r\n\t\t\tConfig: []docker.IPAMConfig{docker.IPAMConfig{\r\n\t\t\t\tSubnet: config.SubnetMast,\r\n\t\t\t\tIPRange: config.IPRange,\r\n\t\t\t\tGateway: config.Gateway,\r\n\t\t\t},\r\n\t\t\t},\r\n\t\t},\r\n\t}\r\n\t// Create network\r\n\treturn cli.CreateNetwork(netoworkOptions)\r\n}", "func (s SessionLease) CreateNetwork() error {\n\tnetworks, err := s.cli.NetworkList(context.Background(), types.NetworkListOptions{})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor _, network := range networks {\n\t\tif network.Name == \"local\" {\n\t\t\t// Network already exists\n\t\t\ts.Log.WithFields(logrus.Fields{\n\t\t\t\t\"networkID\": network.ID,\n\t\t\t\t\"networkName\": network.Name,\n\t\t\t\t\"networkDriver\": network.Driver,\n\t\t\t}).Info(\"Network already exists.\")\n\t\t\treturn nil\n\t\t}\n\t}\n\n\tresponse, err := s.cli.NetworkCreate(context.Background(), \"local\", types.NetworkCreate{\n\t\tCheckDuplicate: true,\n\t\tDriver: \"bridge\",\n\t\tIPAM: &network.IPAM{\n\t\t\tDriver: \"default\",\n\t\t},\n\t\tInternal: false,\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ts.Log.WithField(\"networkID\", response.ID).Debug(\"Network created.\")\n\treturn nil\n}", "func (s *Service) CreateNetwork() error {\n\t// Create docker client.\n\tvar (\n\t\tcli *client.Client\n\n\t\terr error\n\t\terrChanPeer = make(chan error, 1)\n\t\terrChanOrderer = make(chan error, 1)\n\n\t\twg sync.WaitGroup\n\t\twgPeerDone = make(chan bool) // todo change this to ctx.Done()\n\t\twgOrdererDone = make(chan bool) // todo change this to ctx.Done()\n\t)\n\n\tif cli, err = NewClient(); err != nil {\n\t\treturn errors.Wrap(err, \"NewClient failed with error\")\n\t}\n\ts.MyClient = cli\n\n\tctx := context.TODO()\n\trespNet, err := cli.NetworkCreate(ctx, \"giou_net\", types.NetworkCreate{})\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"NetworkCreate failed with error\")\n\t}\n\tlog.Println(\"Network has been created wth ID: \", respNet.ID)\n\n\t// In need of absolute path to bind/mount host:container paths.\n\tprojectPath, err := filepath.Abs(\"./\")\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"Failed to get project's path with error\")\n\t}\n\n\twg.Add(1)\n\n\tgo func(orgs []config.Organization) {\n\t\tdefer wg.Done()\n\n\t\tfor _, org := range orgs {\n\t\t\tfor i := range org.Peers {\n\n\t\t\t\tgo s.RunPeer(org.Name, org.Peers, projectPath, i, errChanPeer, wgPeerDone)\n\n\t\t\t\tselect {\n\t\t\t\tcase <-wgPeerDone:\n\t\t\t\t\tlog.Println(\"carry on...\")\n\n\t\t\t\t\tbreak\n\t\t\t\tcase err := <-errChanPeer:\n\t\t\t\t\tclose(errChanPeer)\n\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t//return err\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}(s.Cfg.Orgs[1:])\n\n\twg.Add(1)\n\n\tgo func(org config.Organization) {\n\t\tdefer wg.Done()\n\n\t\tfor i := range org.Orderers {\n\n\t\t\tgo s.RunOrderer(org.Orderers, projectPath, i, errChanOrderer, wgOrdererDone)\n\n\t\t\tselect {\n\t\t\tcase <-wgOrdererDone:\n\t\t\t\t// carry on\n\t\t\t\tlog.Println(\"carry on orderer...\")\n\n\t\t\t\tbreak\n\t\t\tcase err := <-errChanOrderer:\n\t\t\t\tclose(errChanOrderer)\n\t\t\t\tlog.Fatal(\"Error Orderer: \", err)\n\t\t\t\t//return err\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}(s.Cfg.Orgs[0])\n\n\t// waits until containers are up and running\n\twg.Wait()\n\n\treturn s.List()\n}", "func (orch *Docker) CreateNetwork(ctx context.Context, netID string) error {\n\t// Check if it already exists\n\t_, err := orch.cli.NetworkInspect(ctx, netID, types.NetworkInspectOptions{})\n\tif err != nil {\n\t\t// Create if doesn't exist\n\t\t_, err = orch.cli.NetworkCreate(ctx, netID, types.NetworkCreate{})\n\t}\n\n\treturn err\n}", "func NewNetwork(ctx *pulumi.Context,\n\tname string, args *NetworkArgs, opts ...pulumi.ResourceOption) (*Network, error) {\n\tif args == nil {\n\t\targs = &NetworkArgs{}\n\t}\n\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Network\n\terr := ctx.RegisterResource(\"openstack:networking/network:Network\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (d *driver) CreateNetwork(id string, option map[string]interface{}, nInfo driverapi.NetworkInfo, ipV4Data, ipV6Data []driverapi.IPAMData) error {\n\tif len(ipV4Data) == 0 || ipV4Data[0].Pool.String() == \"0.0.0.0/0\" {\n\t\treturn types.InvalidParameterErrorf(\"ipv4 pool is empty\")\n\t}\n\t// Sanity checks\n\td.Lock()\n\tif _, ok := d.networks[id]; ok {\n\t\td.Unlock()\n\t\treturn types.ForbiddenErrorf(\"network %s exists\", id)\n\t}\n\td.Unlock()\n\n\t// Parse and validate the config. It should not be conflict with existing networks' config\n\tconfig, err := parseNetworkOptions(id, option)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err = config.processIPAM(id, ipV4Data, ipV6Data); err != nil {\n\t\treturn err\n\t}\n\n\t// start the critical section, from this point onward we are dealing with the list of networks\n\t// so to be consistent we cannot allow that the list changes\n\td.configNetwork.Lock()\n\tdefer d.configNetwork.Unlock()\n\n\t// check network conflicts\n\tif err = d.checkConflict(config); err != nil {\n\t\tnerr, ok := err.(defaultBridgeNetworkConflict)\n\t\tif !ok {\n\t\t\treturn err\n\t\t}\n\t\t// Got a conflict with a stale default network, clean that up and continue\n\t\tlog.G(context.TODO()).Warn(nerr)\n\t\tif err := d.deleteNetwork(nerr.ID); err != nil {\n\t\t\tlog.G(context.TODO()).WithError(err).Debug(\"Error while cleaning up network on conflict\")\n\t\t}\n\t}\n\n\t// there is no conflict, now create the network\n\tif err = d.createNetwork(config); err != nil {\n\t\treturn err\n\t}\n\n\treturn d.storeUpdate(config)\n}", "func (s stack) CreateNetwork(req abstract.NetworkRequest) (res *abstract.Network, ferr fail.Error) {\n\tnullAN := abstract.NewNetwork()\n\tif s.IsNull() {\n\t\treturn nullAN, fail.InvalidInstanceError()\n\t}\n\n\tdefer debug.NewTracer(nil, tracing.ShouldTrace(\"stack.aws\") || tracing.ShouldTrace(\"stacks.network\"), \"(%v)\", req).WithStopwatch().Entering().Exiting()\n\n\t// Check if network already there\n\tvar xerr fail.Error\n\tif _, xerr = s.rpcDescribeVpcByName(aws.String(req.Name)); xerr != nil {\n\t\tswitch xerr.(type) {\n\t\tcase *fail.ErrNotFound:\n\t\t\tdebug.IgnoreError(xerr)\n\t\t\t// continue\n\t\tdefault:\n\t\t\treturn nullAN, xerr\n\t\t}\n\t} else {\n\t\treturn nullAN, fail.DuplicateError(\"a Network/VPC named '%s' already exists\")\n\t}\n\n\t// if not, create the network/VPC\n\ttheVpc, xerr := s.rpcCreateVpc(aws.String(req.Name), aws.String(req.CIDR))\n\tif xerr != nil {\n\t\treturn nullAN, fail.Wrap(xerr, \"failed to create VPC\")\n\t}\n\n\t// wait until available status\n\tif IsOperation(theVpc, \"State\", reflect.TypeOf(\"\")) {\n\t\tretryErr := retry.WhileUnsuccessful(\n\t\t\tfunc() error {\n\t\t\t\tvpcTmp, innerXErr := s.rpcDescribeVpcByID(theVpc.VpcId)\n\t\t\t\tif innerXErr != nil {\n\t\t\t\t\treturn innerXErr\n\t\t\t\t}\n\t\t\t\tif aws.StringValue(vpcTmp.State) != \"available\" {\n\t\t\t\t\treturn fail.NewError(\"not ready\")\n\t\t\t\t}\n\t\t\t\treturn nil\n\t\t\t},\n\t\t\ttemporal.GetMinDelay(),\n\t\t\ttemporal.GetDefaultDelay(),\n\t\t)\n\t\tif retryErr != nil {\n\t\t\tswitch retryErr.(type) {\n\t\t\tcase *retry.ErrStopRetry:\n\t\t\t\treturn nullAN, fail.Wrap(fail.Cause(retryErr), \"stopping retries\")\n\t\t\tcase *fail.ErrTimeout:\n\t\t\t\treturn nullAN, fail.Wrap(fail.Cause(retryErr), \"timeout\")\n\t\t\tdefault:\n\t\t\t\treturn nullAN, retryErr\n\t\t\t}\n\t\t}\n\t}\n\n\tdefer func() {\n\t\tif ferr != nil && !req.KeepOnFailure {\n\t\t\tif theVpc != nil {\n\t\t\t\tderr := s.DeleteNetwork(aws.StringValue(theVpc.VpcId))\n\t\t\t\tif derr != nil {\n\t\t\t\t\t_ = ferr.AddConsequence(derr)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\n\tgw, xerr := s.rpcCreateInternetGateway()\n\tif xerr != nil {\n\t\treturn nullAN, fail.Wrap(xerr, \"failed to create internet gateway\")\n\t}\n\n\tif xerr = s.rpcAttachInternetGateway(theVpc.VpcId, gw.InternetGatewayId); xerr != nil {\n\t\treturn nullAN, fail.Wrap(xerr, \"failed to attach internet gateway to Network\")\n\t}\n\n\tdefer func() {\n\t\tif ferr != nil && !req.KeepOnFailure {\n\t\t\tif derr := s.rpcDetachInternetGateway(theVpc.VpcId, gw.InternetGatewayId); derr != nil {\n\t\t\t\t_ = ferr.AddConsequence(normalizeError(derr))\n\t\t\t}\n\t\t}\n\t}()\n\n\ttables, xerr := s.rpcDescribeRouteTables(aws.String(\"vpc-id\"), []*string{theVpc.VpcId})\n\tif xerr != nil {\n\t\treturn nullAN, xerr\n\t}\n\tif len(tables) < 1 {\n\t\treturn nullAN, fail.InconsistentError(\"no Route Tables\")\n\t}\n\n\tif xerr = s.rpcCreateRoute(gw.InternetGatewayId, tables[0].RouteTableId, aws.String(\"0.0.0.0/0\")); xerr != nil {\n\t\treturn nullAN, fail.Wrap(xerr, \"failed to create route\")\n\t}\n\n\tdefer func() {\n\t\tif ferr != nil && !req.KeepOnFailure {\n\t\t\tif derr := s.rpcDeleteRoute(tables[0].RouteTableId, aws.String(\"0.0.0.0/0\")); derr != nil {\n\t\t\t\t_ = ferr.AddConsequence(normalizeError(derr))\n\t\t\t}\n\t\t}\n\t}()\n\n\tanet := abstract.NewNetwork()\n\tanet.ID = aws.StringValue(theVpc.VpcId)\n\tanet.Name = req.Name\n\tanet.CIDR = req.CIDR\n\tanet.DNSServers = req.DNSServers\n\n\t// Make sure we log warnings\n\t_ = anet.OK()\n\n\treturn anet, nil\n}", "func (d *Driver) createNetwork() error {\n\tconn, err := getConnection()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"getting libvirt connection\")\n\t}\n\tdefer conn.Close()\n\n\t// network: default\n\t// It is assumed that the libvirt/kvm installation has already created this network\n\n\t// network: private\n\n\t// Only create the private network if it does not already exist\n\tif _, err := conn.LookupNetworkByName(d.PrivateNetwork); err != nil {\n\t\t// create the XML for the private network from our networkTmpl\n\t\ttmpl := template.Must(template.New(\"network\").Parse(networkTmpl))\n\t\tvar networkXML bytes.Buffer\n\t\tif err := tmpl.Execute(&networkXML, d); err != nil {\n\t\t\treturn errors.Wrap(err, \"executing network template\")\n\t\t}\n\n\t\t// define the network using our template\n\t\tnetwork, err := conn.NetworkDefineXML(networkXML.String())\n\t\tif err != nil {\n\t\t\treturn errors.Wrapf(err, \"defining network from xml: %s\", networkXML.String())\n\t\t}\n\n\t\t// and finally create it\n\t\tif err := network.Create(); err != nil {\n\t\t\treturn errors.Wrapf(err, \"creating network %s\", d.PrivateNetwork)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (d *Driver) CreateNetwork(r *pluginNet.CreateNetworkRequest) error {\n\tdefer osl.InitOSContext()()\n\n\tid := r.NetworkID\n\topts := r.Options\n\tipV4Data := r.IPv4Data\n\tipV6Data := r.IPv6Data\n\tlogrus.Infof(\"CreateNetwork macvlan with networkID=%s,opts=%s\", id, opts)\n\n\tif id == \"\" {\n\t\treturn fmt.Errorf(\"invalid network id\")\n\t}\n\n\t// reject a null v4 network\n\tif len(ipV4Data) == 0 || ipV4Data[0].Pool == \"0.0.0.0/0\" {\n\t\treturn fmt.Errorf(\"ipv4 pool is empty\")\n\t}\n\n\t// parse and validate the config and bind to networkConfiguration\n\tconfig, err := parseNetworkOptions(id, opts)\n\tif err != nil {\n\t\tstr := fmt.Sprintf(\"CreateNetwork opts is invalid %s\", opts)\n\t\tlogrus.Errorf(str)\n\t\treturn fmt.Errorf(str)\n\t}\n\n\tconfig.ID = id\n\terr = config.processIPAM(id, ipV4Data, ipV6Data)\n\tif err != nil {\n\t\tstr := fmt.Sprintf(\"CreateNetwork ipV4Data is invalid %s\", ipV4Data)\n\t\tlogrus.Errorf(str)\n\t\treturn fmt.Errorf(str)\n\t}\n\t// verify the macvlan mode from -o macvlan_mode option\n\tswitch config.MacvlanMode {\n\tcase \"\", modeBridge:\n\t\t// default to macvlan bridge mode if -o macvlan_mode is empty\n\t\tconfig.MacvlanMode = modeBridge\n\tcase modePrivate:\n\t\tconfig.MacvlanMode = modePrivate\n\tcase modePassthru:\n\t\tconfig.MacvlanMode = modePassthru\n\tcase modeVepa:\n\t\tconfig.MacvlanMode = modeVepa\n\tdefault:\n\t\tstr := fmt.Sprintf(\"requested macvlan mode '%s' is not valid, 'bridge' mode is the macvlan driver default\", config.MacvlanMode)\n\t\tlogrus.Errorf(str)\n\t\treturn fmt.Errorf(str)\n\t}\n\t// loopback is not a valid parent link\n\tif config.Parent == \"lo\" {\n\t\tstr := fmt.Sprintf(\"loopback interface is not a valid %s parent link\", macvlanType)\n\t\tlogrus.Errorf(str)\n\t\treturn fmt.Errorf(str)\n\t}\n\t// if parent interface not specified, create a dummy type link to use named dummy+net_id\n\tif config.Parent == \"\" {\n\t\tconfig.Parent = getDummyName(stringid.TruncateID(config.ID))\n\t\t// empty parent and --internal are handled the same. Set here to update k/v\n\t\tconfig.Internal = true\n\t}\n\n\terr = d.createNetwork(config)\n\tif err != nil {\n\t\tstr := fmt.Sprintf(\"CreateNetwork is failed %v\", err)\n\t\tlogrus.Errorf(str)\n\t\treturn fmt.Errorf(str)\n\t}\n\n\treturn nil\n}", "func (account *SCloudaccount) createNetwork(ctx context.Context, wireId, networkType string, net CANetConf) error {\n\tnetwork := &SNetwork{}\n\tnetwork.Name = net.Name\n\tif hint, err := NetworkManager.NewIfnameHint(net.Name); err != nil {\n\t\tlog.Errorf(\"can't NewIfnameHint form hint %s\", net.Name)\n\t} else {\n\t\tnetwork.IfnameHint = hint\n\t}\n\tnetwork.GuestIpStart = net.IpStart\n\tnetwork.GuestIpEnd = net.IpEnd\n\tnetwork.GuestIpMask = net.IpMask\n\tnetwork.GuestGateway = net.Gateway\n\tnetwork.VlanId = int(net.VlanID)\n\tnetwork.WireId = wireId\n\tnetwork.ServerType = networkType\n\tnetwork.IsPublic = true\n\tnetwork.Status = api.NETWORK_STATUS_AVAILABLE\n\tnetwork.PublicScope = string(rbacscope.ScopeDomain)\n\tnetwork.ProjectId = account.ProjectId\n\tnetwork.DomainId = account.DomainId\n\tnetwork.Description = net.Description\n\n\tnetwork.SetModelManager(NetworkManager, network)\n\t// TODO: Prevent IP conflict\n\tlog.Infof(\"create network %s succussfully\", network.Id)\n\terr := NetworkManager.TableSpec().Insert(ctx, network)\n\treturn err\n}", "func (tester *ServiceTester) CreateNetwork(t *testing.T, name string, cidr string) *abstract.Network {\n\tnetwork, err := tester.Service.CreateNetwork(context.Background(), abstract.NetworkRequest{\n\t\tName: name,\n\t\tCIDR: cidr,\n\t})\n\trequire.NoError(t, err)\n\n\treturn network\n}", "func (self *AltaActor) createNetwork() error {\n\t// Loop thru each endpoint\n\tfor _, endpoint := range self.Model.Spec.Endpoints {\n\t\tnetwork, err := netCtrler.FindNetwork(endpoint.NetworkName)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Network %s not found while creating Alta: %+v\", self.Model.Spec)\n\t\t\treturn errors.New(\"network not found\")\n\t\t}\n\n\t\t// Send the network info to the node\n\t\tvar resp altaspec.ReqSuccess\n\t\turl := \"/network/create\"\n\t\terr = nodeCtrler.NodePostReq(self.Model.CurrNode, url, network.NetSpec, &resp)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error sending network info to node %s. Err: %v\",\n\t\t\t\tself.Model.CurrNode, err)\n\t\t}\n\t}\n\n\t// Move forward in fsm\n\tself.AltaEvent(\"createVol\")\n\n\treturn nil\n}", "func NewNetwork(dht *DHT, self *Node) (*Network, error) {\n\ts := &Network{\n\t\tdht: dht,\n\t\tself: self,\n\t\tdone: make(chan struct{}),\n\t}\n\t// init the rate limiter\n\ts.limiter = ratelimit.New(defaultConnRate)\n\n\taddr := fmt.Sprintf(\"%s:%d\", self.IP, self.Port)\n\t// new the network socket\n\tsocket, err := utp.NewSocket(\"udp\", addr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ts.socket = socket\n\n\treturn s, nil\n}", "func NewNetwork(dockerClient *client.Client, cfg NetworkConfig) (out *Network, err error) {\n\tscopes.Framework.Infof(\"Creating Docker network %s\", cfg.Name)\n\tresp, err := dockerClient.NetworkCreate(context.Background(), cfg.Name, types.NetworkCreate{\n\t\tCheckDuplicate: true,\n\t\tLabels: cfg.Labels,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tscopes.Framework.Infof(\"Docker network %s created (ID=%s)\", cfg.Name, resp.ID)\n\n\tn := &Network{\n\t\tNetworkConfig: cfg,\n\t\tdockerClient: dockerClient,\n\t\tid: resp.ID,\n\t}\n\tdefer func() {\n\t\tif err != nil {\n\t\t\t_ = n.Close()\n\t\t}\n\t}()\n\n\t// Retrieve the subnet for the network.\n\tiresp, err := dockerClient.NetworkInspect(context.Background(), resp.ID, types.NetworkInspectOptions{})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif _, n.Subnet, err = net.ParseCIDR(iresp.IPAM.Config[0].Subnet); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn n, nil\n}", "func (session *dummySession) CreateNetworkRequest(method string, url string, payload string, header map[string]string) networkModel.NetworkRequest {\n\tassert.Fail(session.t, \"Unexpected call to CreateNetworkRequest\")\n\treturn nil\n}", "func (c *Client) CreateNetwork(name string) (string, error) {\n\tendpoint := fmt.Sprintf(\"%snetworks/create\", baseAddr)\n\n\tmin := struct {\n\t\tName string `json:\"Name\"`\n\t\tDriver string `json:\"Driver\"`\n\t\tAttachable bool `json:\"Attachable\"`\n\t}{\n\t\tName: name,\n\t\tDriver: \"bridge\",\n\t\tAttachable: true,\n\t}\n\n\tb, err := json.Marshal(&min)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tr, err := c.http.Post(endpoint, \"application/json\", bytes.NewReader(b))\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif err = statusCode(r.StatusCode, http.StatusCreated); err != nil {\n\t\treturn \"\", err\n\t}\n\n\tres := struct {\n\t\tID string `json:\"Id\"`\n\t\tWarnings []interface{} `json:\"Warnings\"`\n\t}{}\n\n\treturn res.ID, json.NewDecoder(r.Body).Decode(&res)\n}", "func NewNetwork(opts ...options.Option) Network {\n\toptions := options.NewOptions(opts...)\n\n\t// get router\n\n\t// get proxy\n\n\treturn &network{\n\t\tOptions: options,\n\t\t// fill the blanks\n\t\t// router: r,\n\t\t// proxy: p,\n\t}\n}", "func (b *BridgeNetworkDriver) Create(name string, subnet string) (*Network, error) {\n\t// 取到网段字符串中的网关ip地址和网络的ip段\n\tip, IPRange, _ := net.ParseCIDR(subnet)\n\tIPRange.IP = ip\n\n\tn := &Network{\n\t\tName: name,\n\t\tIPRange: IPRange,\n\t\tDriver: b.Name(),\n\t}\n\n\terr := b.initBridge(n)\n\treturn n, err\n}", "func (c *clusterNetwork) Create(cn *sdnapi.ClusterNetwork) (result *sdnapi.ClusterNetwork, err error) {\n\tresult = &sdnapi.ClusterNetwork{}\n\terr = c.r.Post().Resource(\"clusterNetworks\").Body(cn).Do().Into(result)\n\treturn\n}", "func NetworkResourceCreate(w http.ResponseWriter, r *http.Request) {\n\tvar req openapi.NetworkResourceCreateRequest\n\tif err := json.NewDecoder(r.Body).Decode(&req); err != nil {\n\t\thandleError(w, http.StatusBadRequest, err.Error())\n\t\treturn\n\t}\n\n\tnetwork, err := ctr.CreateNetworkResource(&req)\n\tif err != nil {\n\t\thandleError(w, http.StatusInternalServerError, err.Error())\n\t\treturn\n\t}\n\n\thandleResult(w, http.StatusCreated, network)\n\treturn\n}", "func NewNetwork() *Network {\n\treturn &Network{\n\t\tport: 1234,\n\t}\n}", "func NewNetwork(svc iaas.Service) (resources.Network, fail.Error) {\n\tif svc == nil {\n\t\treturn NullValue(), fail.InvalidParameterCannotBeNilError(\"svc\")\n\t}\n\n\tcoreInstance, xerr := NewCore(svc, networkKind, networksFolderName, &abstract.Network{})\n\txerr = debug.InjectPlannedFail(xerr)\n\tif xerr != nil {\n\t\treturn NullValue(), xerr\n\t}\n\n\tinstance := &Network{\n\t\tMetadataCore: coreInstance,\n\t}\n\treturn instance, nil\n}", "func New() *NetworkCreateInput {\n\tnet := &NetworkCreateInput{}\n\treturn net\n}", "func newNetworkDef() libvirtxml.Network {\n\tconst defNetworkXML = `\n\t\t<network>\n\t\t <name>default</name>\n\t\t <forward mode='nat'>\n\t\t <nat>\n\t\t <port start='1024' end='65535'/>\n\t\t </nat>\n\t\t </forward>\n\t\t</network>`\n\tif d, err := newDefNetworkFromXML(defNetworkXML); err != nil {\n\t\tpanic(fmt.Sprintf(\"Unexpected error while parsing default network definition: %s\", err))\n\t} else {\n\t\treturn d\n\t}\n}", "func (h networkHandler) createNetworkHandler(w http.ResponseWriter, r *http.Request) {\n\tvar p netintents.Network\n\tvars := mux.Vars(r)\n\tclusterProvider := vars[\"provider-name\"]\n\tcluster := vars[\"cluster-name\"]\n\n\terr := json.NewDecoder(r.Body).Decode(&p)\n\n\tswitch {\n\tcase err == io.EOF:\n\t\thttp.Error(w, \"Empty body\", http.StatusBadRequest)\n\t\treturn\n\tcase err != nil:\n\t\thttp.Error(w, err.Error(), http.StatusUnprocessableEntity)\n\t\treturn\n\t}\n\n\terr, httpError := validation.ValidateJsonSchemaData(vnJSONFile, p)\nif err != nil {\n\thttp.Error(w, err.Error(), httpError)\n\treturn\n}\n\n\t// Name is required.\n\tif p.Metadata.Name == \"\" {\n\t\thttp.Error(w, \"Missing name in POST request\", http.StatusBadRequest)\n\t\treturn\n\t}\n\n\terr = validateNetworkInputs(p)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tret, err := h.client.CreateNetwork(p, clusterProvider, cluster, false)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(http.StatusCreated)\n\terr = json.NewEncoder(w).Encode(ret)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n}", "func NewNetwork() *Network {\n\tnet := &Network{\n\t\tvms: make(map[int]*intcode.VM),\n\t\tmsgs: make(map[int]chan int),\n\t\tpackets: make(chan packet, 128),\n\t}\n\tnet.nat = newNAT(net)\n\treturn net\n}", "func (n *NetworkServiceHandler) Create(ctx context.Context, regionID, description, cidrBlock string) (*Network, error) {\n\n\turi := \"/v1/network/create\"\n\n\tvalues := url.Values{\n\t\t\"DCID\": {regionID},\n\t}\n\n\t// Optional\n\tif cidrBlock != \"\" {\n\t\t_, ipNet, err := net.ParseCIDR(cidrBlock)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif v4Subnet := ipNet.IP.To4(); v4Subnet != nil {\n\t\t\tvalues.Add(\"v4_subnet\", v4Subnet.String())\n\t\t}\n\t\tmask, _ := ipNet.Mask.Size()\n\t\tvalues.Add(\"v4_subnet_mask\", strconv.Itoa(mask))\n\t}\n\n\tif description != \"\" {\n\t\tvalues.Add(\"description\", description)\n\t}\n\n\treq, err := n.client.NewRequest(ctx, http.MethodPost, uri, values)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tnetwork := new(Network)\n\terr = n.client.DoWithContext(ctx, req, network)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn network, nil\n}", "func NewNetwork(spec *types.NetworkSpec) (Network, error) {\n\tn := &network{\n\t\tname: spec.Name,\n\t\ttyp: spec.Type,\n\t\tuseNAT: spec.UseNAT,\n\t}\n\tif len(spec.Address) > 0 {\n\t\taddr, err := netlink.ParseAddr(spec.Address)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tn.addr = addr\n\t}\n\n\treturn n, nil\n}", "func (d *Driver) createNetwork(config *configuration) error {\n\tif !parentExists(config.Parent) {\n\t\t// if the --internal flag is set, create a dummy link\n\t\tif config.Internal {\n\t\t\terr := createDummyLink(config.Parent, getDummyName(stringid.TruncateID(config.ID)))\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tconfig.CreatedSlaveLink = true\n\t\t\t// notify the user in logs they have limited comunicatins\n\t\t\tif config.Parent == getDummyName(stringid.TruncateID(config.ID)) {\n\t\t\t\tlogrus.Infof(\"Empty -o parent= and --internal flags limit communications to other containers inside of network: %s\",\n\t\t\t\t\tconfig.Parent)\n\t\t\t}\n\t\t} else {\n\t\t\t// if the subinterface parent_iface.vlan_id checks do not pass, return err.\n\t\t\t// a valid example is 'eth0.10' for a parent iface 'eth0' with a vlan id '10'\n\t\t\terr := createVlanLink(config.Parent)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\t// if driver created the networks slave link, record it for future deletion\n\t\t\tconfig.CreatedSlaveLink = true\n\t\t}\n\t}\n\tn := &network{\n\t\tid: config.ID,\n\t\tdriver: d,\n\t\tendpoints: endpointTable{},\n\t\tconfig: config,\n\t}\n\t// add the *network\n\td.addNetwork(n)\n\n\treturn nil\n}", "func (d *Driver) createNetworks() error {\n\tif err := d.createNetwork(\"default\", defaultNetworkTmpl); err != nil {\n\t\treturn errors.Wrap(err, \"creating default network\")\n\t}\n\tif err := d.createNetwork(d.NetworkName, privateNetworkTmpl); err != nil {\n\t\treturn errors.Wrap(err, \"creating private network\")\n\t}\n\n\treturn nil\n}", "func NewNetwork(x int, n float64, a float64, act int, b float64, h int, r int, o int) Network{\n\tnet := Network{Lrate: n, Mom: a, Bias: b, Activation: act, Inputs: x, Outputs: o, Hidden: h}\n\tfor i := 0; i < r; i++ {\n\t\tif i == 0 {\n\t\t\tl := L.NewHidden(x, n, a, act, b, h)\n\t\t\tnet.HiddenLayers = append(net.HiddenLayers, l)\t\t\t\n\t\t} else {\n\t\t\tl := L.NewHidden(h, n, a, act, b, h)\t\n\t\t\tnet.HiddenLayers = append(net.HiddenLayers, l)\t\t\t\n\t\t}\n\t}\n\tnet.OutputLayer = L.NewOutput(h, n, a, act, b, o)\n\n\treturn net\n}", "func CreateNetworkSecurityGroupRule() {}", "func (instance *Network) Create(ctx context.Context, req abstract.NetworkRequest) (xerr fail.Error) {\n\tdefer fail.OnPanic(&xerr)\n\n\tif instance == nil || instance.IsNull() {\n\t\treturn fail.InvalidInstanceError()\n\t}\n\tif ctx == nil {\n\t\treturn fail.InvalidParameterCannotBeNilError(\"ctx\")\n\t}\n\n\ttask, xerr := concurrency.TaskFromContext(ctx)\n\txerr = debug.InjectPlannedFail(xerr)\n\tif xerr != nil {\n\t\treturn xerr\n\t}\n\n\tif task.Aborted() {\n\t\treturn fail.AbortedError(nil, \"aborted\")\n\t}\n\n\ttracer := debug.NewTracer(task, true, \"('%s', '%s')\", req.Name, req.CIDR).WithStopwatch().Entering()\n\tdefer tracer.Exiting()\n\n\tinstance.lock.Lock()\n\tdefer instance.lock.Unlock()\n\n\t// Check if subnet already exists and is managed by SafeScale\n\tsvc := instance.GetService()\n\tif existing, xerr := LoadNetwork(svc, req.Name); xerr == nil {\n\t\texisting.Released()\n\t\treturn fail.DuplicateError(\"Network '%s' already exists\", req.Name)\n\t}\n\n\tif task.Aborted() {\n\t\treturn fail.AbortedError(nil, \"aborted\")\n\t}\n\n\t// Verify if the subnet already exist and in this case is not managed by SafeScale\n\t_, xerr = svc.InspectNetworkByName(req.Name)\n\txerr = debug.InjectPlannedFail(xerr)\n\tif xerr != nil {\n\t\tswitch xerr.(type) {\n\t\tcase *fail.ErrNotFound:\n\t\t\t// continue\n\t\tdefault:\n\t\t\treturn xerr\n\t\t}\n\t} else {\n\t\treturn fail.DuplicateError(\"Network '%s' already exists (not managed by SafeScale)\", req.Name)\n\t}\n\n\t// Verify the CIDR is not routable\n\tif req.CIDR != \"\" {\n\t\troutable, xerr := netretry.IsCIDRRoutable(req.CIDR)\n\t\txerr = debug.InjectPlannedFail(xerr)\n\t\tif xerr != nil {\n\t\t\treturn fail.Wrap(xerr, \"failed to determine if CIDR is not routable\")\n\t\t}\n\n\t\tif routable {\n\t\t\treturn fail.InvalidRequestError(\"cannot create such a Networking, CIDR must not be routable; please choose an appropriate CIDR (RFC1918)\")\n\t\t}\n\t}\n\n\tif task.Aborted() {\n\t\treturn fail.AbortedError(nil, \"aborted\")\n\t}\n\n\t// Create the Network\n\tlogrus.Debugf(\"Creating Network '%s' with CIDR '%s'...\", req.Name, req.CIDR)\n\tan, xerr := svc.CreateNetwork(req)\n\txerr = debug.InjectPlannedFail(xerr)\n\tif xerr != nil {\n\t\treturn xerr\n\t}\n\n\tdefer func() {\n\t\tif xerr != nil && !req.KeepOnFailure {\n\t\t\tderr := svc.DeleteNetwork(an.ID)\n\t\t\tderr = debug.InjectPlannedFail(derr)\n\t\t\tif derr != nil {\n\t\t\t\t_ = xerr.AddConsequence(fail.Wrap(derr, \"cleaning up on failure, failed to delete Network\"))\n\t\t\t}\n\t\t}\n\t}()\n\n\tif task.Aborted() {\n\t\treturn fail.AbortedError(nil, \"aborted\")\n\t}\n\n\t// Write subnet object metadata\n\t// logrus.Debugf(\"Saving subnet metadata '%s' ...\", subnet.GetName)\n\treturn instance.carry(an)\n}", "func NewNetwork(in, out, all []*NNode, netId int) *Network {\n\tn := Network{\n\t\tId: netId,\n\t\tinputs: in,\n\t\tOutputs: out,\n\t\tallNodes: all,\n\t\tnumLinks: -1,\n\t\tallNodesMIMO: all,\n\t}\n\treturn &n\n}", "func TestNetworkCreateOnNonExistentTenant(t *testing.T) {\n\t// create netagent\n\tag, _, _ := createNetAgent(t)\n\tAssert(t, ag != nil, \"Failed to create agent %#v\", ag)\n\tdefer ag.Stop()\n\n\t// create network\n\tnt := netproto.Network{\n\t\tTypeMeta: api.TypeMeta{Kind: \"Network\"},\n\t\tObjectMeta: api.ObjectMeta{\n\t\t\tTenant: \"nonExistentNetwork\",\n\t\t\tNamespace: \"nonExistentNetwork\",\n\t\t\tName: \"default\",\n\t\t},\n\t}\n\n\t// create network\n\terr := ag.CreateNetwork(&nt)\n\tAssert(t, err != nil, \"Network create was expected to fail.\")\n}", "func NewNetwork(cfg SourceGeneratorConfig) Sourcer {\n\treturn &Network{\n\t\tschema: []idk.Field{\n\t\t\tidk.IDField{NameVal: \"id\"}, // 0\n\t\t\tidk.IDField{NameVal: \"qos_tier\"}, // 1\n\t\t\tidk.IntField{NameVal: \"source_equip_id\"}, // 2\n\t\t\tidk.IntField{NameVal: \"dest_equip_id\"}, // 3\n\t\t\tidk.IntField{NameVal: \"data_size\"}, // 4\n\t\t\tidk.StringField{NameVal: \"data_type\"}, // 5\n\t\t\tidk.StringField{NameVal: \"customer\"}, // 6\n\t\t\tidk.IntField{NameVal: \"timestamp\"}, // 7\n\t\t},\n\t}\n}", "func dockerNetCreate(subnet net.IPNet, gw net.IP, bridge string, subnetID string) error {\n\tcmd := exec.Command(\"docker\", \"network\", \"create\", \"-d=ciao\", \"--ipam-driver=ciao\",\n\t\t\"--subnet=\"+subnet.String(), \"--gateway=\"+gw.String(),\n\t\t\"--opt\", \"bridge=\"+bridge, subnetID)\n\n\t_, err := cmd.CombinedOutput()\n\treturn err\n}", "func (c *client) CreateNetwork(ctx context.Context, b *pipeline.Build) error {\n\tlogrus.Tracef(\"creating network for pipeline %s\", b.ID)\n\n\t// nolint: lll // ignore long line length due to link\n\t// create the network for the pod\n\t//\n\t// This is done due to the nature of how networking works inside the\n\t// pod. Each container inside the pod shares the same network IP and\n\t// port space. This allows them to communicate with each other via\n\t// localhost. However, to keep the runtime behavior consistent,\n\t// Vela adds DNS entries for each container that requires it.\n\t//\n\t// More info:\n\t// * https://kubernetes.io/docs/concepts/workloads/pods/pod/\n\t// * https://kubernetes.io/docs/concepts/services-networking/add-entries-to-pod-etc-hosts-with-host-aliases/\n\t//\n\t// https://pkg.go.dev/k8s.io/api/core/v1?tab=doc#HostAlias\n\tnetwork := v1.HostAlias{\n\t\tIP: \"127.0.0.1\",\n\t\tHostnames: []string{},\n\t}\n\n\t// iterate through all services in the pipeline\n\tfor _, service := range b.Services {\n\t\t// create the host entry for the pod container aliases\n\t\thost := fmt.Sprintf(\"%s.local\", service.Name)\n\n\t\t// add the host entry to the pod container aliases\n\t\tnetwork.Hostnames = append(network.Hostnames, host)\n\t}\n\n\t// iterate through all steps in the pipeline\n\tfor _, step := range b.Steps {\n\t\t// skip all steps not running in detached mode\n\t\tif !step.Detach {\n\t\t\tcontinue\n\t\t}\n\n\t\t// create the host entry for the pod container aliases\n\t\thost := fmt.Sprintf(\"%s.local\", step.Name)\n\n\t\t// add the host entry to the pod container aliases\n\t\tnetwork.Hostnames = append(network.Hostnames, host)\n\t}\n\n\t// iterate through all stages in the pipeline\n\tfor _, stage := range b.Stages {\n\t\t// iterate through all steps in each stage\n\t\tfor _, step := range stage.Steps {\n\t\t\t// skip all steps not running in detached mode\n\t\t\tif !step.Detach {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// create the host entry for the pod container aliases\n\t\t\thost := fmt.Sprintf(\"%s.local\", step.Name)\n\n\t\t\t// add the host entry to the pod container aliases\n\t\t\tnetwork.Hostnames = append(network.Hostnames, host)\n\t\t}\n\t}\n\n\t// add the network definition to the pod spec\n\t//\n\t// https://pkg.go.dev/k8s.io/api/core/v1?tab=doc#PodSpec\n\tc.Pod.Spec.HostAliases = append(c.Pod.Spec.HostAliases, network)\n\n\treturn nil\n}", "func createHnsNetwork(backend string, networkAdapter string) (string, error) {\n\tvar network hcsshim.HNSNetwork\n\tif backend == \"vxlan\" {\n\t\t// Ignoring the return because both true and false without an error represent that the firewall rule was created or already exists\n\t\tif _, err := wapi.FirewallRuleAdd(\"OverlayTraffic4789UDP\", \"Overlay network traffic UDP\", \"\", \"4789\", wapi.NET_FW_IP_PROTOCOL_UDP, wapi.NET_FW_PROFILE2_ALL); err != nil {\n\t\t\treturn \"\", fmt.Errorf(\"error creating firewall rules: %v\", err)\n\t\t}\n\t\tlogrus.Infof(\"Creating VXLAN network using the vxlanAdapter: %s\", networkAdapter)\n\t\tnetwork = hcsshim.HNSNetwork{\n\t\t\tType: \"Overlay\",\n\t\t\tName: CalicoHnsNetworkName,\n\t\t\tNetworkAdapterName: networkAdapter,\n\t\t\tSubnets: []hcsshim.Subnet{\n\t\t\t\t{\n\t\t\t\t\tAddressPrefix: \"192.168.255.0/30\",\n\t\t\t\t\tGatewayAddress: \"192.168.255.1\",\n\t\t\t\t\tPolicies: []json.RawMessage{\n\t\t\t\t\t\t[]byte(\"{ \\\"Type\\\": \\\"VSID\\\", \\\"VSID\\\": 9999 }\"),\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t}\n\t} else {\n\t\tnetwork = hcsshim.HNSNetwork{\n\t\t\tType: \"L2Bridge\",\n\t\t\tName: CalicoHnsNetworkName,\n\t\t\tNetworkAdapterName: networkAdapter,\n\t\t\tSubnets: []hcsshim.Subnet{\n\t\t\t\t{\n\t\t\t\t\tAddressPrefix: \"192.168.255.0/30\",\n\t\t\t\t\tGatewayAddress: \"192.168.255.1\",\n\t\t\t\t},\n\t\t\t},\n\t\t}\n\t}\n\n\tif _, err := network.Create(); err != nil {\n\t\treturn \"\", fmt.Errorf(\"error creating the %s network: %v\", CalicoHnsNetworkName, err)\n\t}\n\n\t// Check if network exists. If it does not after 5 minutes, fail\n\tfor start := time.Now(); time.Since(start) < 5*time.Minute; {\n\t\tnetwork, err := hcsshim.GetHNSNetworkByName(CalicoHnsNetworkName)\n\t\tif err == nil {\n\t\t\treturn network.ManagementIP, nil\n\t\t}\n\t}\n\n\treturn \"\", fmt.Errorf(\"failed to create %s network\", CalicoHnsNetworkName)\n}", "func (nb *NetBuilder) FindOrCreateNetwork(nw *Network) error {\n\t// Check that the HNS version is supported.\n\terr := nb.checkHNSVersion()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tnw.Name = nb.generateHNSNetworkName(nw)\n\thnsNetwork, err := hcsshim.GetHNSNetworkByName(nw.Name)\n\tif err == nil {\n\t\tlog.Infof(\"Found existing HNS network %s.\", nw.Name)\n\t\treturn nil\n\t}\n\n\t// If existing network flag is enabled, many of the parameters of netConfig become optional.\n\t// This can potentially lead to failure in network creation.\n\t// Therefore, return error at this point.\n\tif nw.UseExisting {\n\t\tlog.Errorf(\"Failed to find existing network: %s.\", nw.Name)\n\t\treturn fmt.Errorf(\"failed to find existing network %s\", nw.Name)\n\t}\n\n\t// Find the ENI link.\n\terr = nw.ENI.AttachToLink()\n\tif err != nil {\n\t\tlog.Errorf(\"Failed to find ENI link: %v.\", err)\n\t\treturn err\n\t}\n\n\t// Initialize the HNS network.\n\thnsNetwork = &hcsshim.HNSNetwork{\n\t\tName: nw.Name,\n\t\tType: hnsTransparentNetworkType,\n\t\tNetworkAdapterName: nw.ENI.GetLinkName(),\n\n\t\tSubnets: []hcsshim.Subnet{\n\t\t\t{\n\t\t\t\tAddressPrefix: vpc.GetSubnetPrefix(&nw.IPAddresses[0]).String(),\n\t\t\t},\n\t\t},\n\t}\n\n\t// Gateway IP addresses are optional, therefore, if they are available then add the first one.\n\tif len(nw.GatewayIPAddresses) != 0 {\n\t\thnsNetwork.Subnets[0].GatewayAddress = nw.GatewayIPAddresses[0].String()\n\t}\n\n\t// Create the HNS network.\n\tlog.Infof(\"Creating HNS network: %+v\", hnsNetwork)\n\thnsResponse, err := hnsNetwork.Create()\n\tif err != nil {\n\t\tlog.Errorf(\"Failed to create HNS network: %v.\", err)\n\t\treturn err\n\t}\n\n\tlog.Infof(\"Received HNS network response: %+v.\", hnsResponse)\n\n\t// For the new network, disable the vNIC in the host compartment.\n\tmgmtIface := fmt.Sprintf(vNICNameFormat, nw.ENI.GetLinkName())\n\terr = nb.disableInterface(mgmtIface)\n\tif err != nil {\n\t\t// This is a fatal error as the management vNIC must be disabled.\n\t\t_ = nb.DeleteNetwork(nw)\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (n *NetworkCreateCommand) runNetworkCreate(args []string) error {\n\tname := n.name\n\tif len(args) != 0 {\n\t\tname = args[0]\n\t}\n\tif name == \"\" {\n\t\treturn fmt.Errorf(\"network name cannot be empty\")\n\t}\n\n\tnetworkRequest, err := n.buildNetworkCreateRequest(name)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tctx := context.Background()\n\tapiClient := n.cli.Client()\n\tresp, err := apiClient.NetworkCreate(ctx, networkRequest)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif resp.Warning != \"\" {\n\t\tfmt.Printf(\"WARNING: %s \\n\", resp.Warning)\n\t}\n\tfmt.Printf(\"%s: %s\\n\", name, resp.ID)\n\n\treturn nil\n}", "func checkCreateNetwork(t *testing.T, expError bool, tenant, network, encap, subnet, gw string, tag int) {\n\tnet := client.Network{\n\t\tTenantName: tenant,\n\t\tNetworkName: network,\n\t\tEncap: encap,\n\t\tSubnet: subnet,\n\t\tGateway: gw,\n\t\tPktTag: tag,\n\t}\n\terr := contivClient.NetworkPost(&net)\n\tif err != nil && !expError {\n\t\tt.Fatalf(\"Error creating network {%+v}. Err: %v\", net, err)\n\t} else if err == nil && expError {\n\t\tt.Fatalf(\"Create network {%+v} succeded while expecing error\", net)\n\t} else if err == nil {\n\t\t// verify network is created\n\t\t_, err := contivClient.NetworkGet(tenant, network)\n\t\tif err != nil {\n\t\t\tt.Fatalf(\"Error getting network %s/%s. Err: %v\", tenant, network, err)\n\t\t}\n\t}\n}", "func NewNetwork(inputSize, outputSize, hiddenLayersCount int) Network {\n\tbias := rand.NormFloat64()\n\tweights := make([]float64, inputSize)\n\thiddenLayers := make([]Layer, hiddenLayersCount)\n\n\tfor i := 0; i < hiddenLayersCount; i++ {\n\t\thiddenLayers[i] = newLayer(inputSize)\n\t}\n\n\tfor i := 0; i < inputSize; i++ {\n\t\tweights[i] = rand.NormFloat64()\n\t}\n\n\top := Neuron{\n\t\tbias: bias,\n\t\tweights: weights,\n\t}\n\n\toutputs := make([]Neuron, outputSize)\n\tfor i := 0; i < outputSize; i++ {\n\t\toutputs[i] = Neuron{\n\t\t\tbias: rand.NormFloat64(),\n\t\t\tweights: weights,\n\t\t}\n\t}\n\n\tnetwork := Network{\n\t\tinputSize: inputSize,\n\t\toutputSize: outputSize,\n\t\top2: op,\n\t\top: outputs,\n\t\tlayers: hiddenLayers,\n\t}\n\n\treturn network\n}", "func newNetwork(cfg *config.Network, c *ec2.EC2) (*network, error) {\n\tlog.Debug(\"Initializing AWS Network\")\n\tn := &network{\n\t\tResources: resource.NewResources(),\n\t\tNetwork: cfg,\n\t\tec2: c,\n\t}\n\n\tvpc, err := newVpc(c, n)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tn.vpc = vpc\n\tn.Append(vpc)\n\n\trouteTables, err := newRouteTables(c, n)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tn.routeTables = routeTables\n\tn.Append(routeTables)\n\n\tinternetGateway, err := newInternetGateway(c, n, \"public\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tn.internetGateway = internetGateway\n\tn.Append(internetGateway)\n\n\t// Load the vpc since it is needed for the caches.\n\terr = n.vpc.Load()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tn.subnetCache, err = newSubnetCache(n)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tn.secgroupCache, err = newSecurityGroupCache(n)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn n, nil\n}", "func (d Docker) CreateNetworkIfNotPresent(ctx context.Context, name string) (string, bool, error) {\n\n\t// (0) create new docker client\n\tdocker, err := client.NewClientWithOpts(client.FromEnv, client.WithAPIVersionNegotiation())\n\tif err != nil {\n\t\tlog.Errorln(\"Failed to create docker client\")\n\t\treturn \"\", false, err\n\t}\n\tdefer docker.Close()\n\n\t// (1) configure list filters\n\targs := filters.NewArgs()\n\targs.Add(\"name\", name)\n\n\t// (2) get filtered list of networks\n\tnetworkList, err := docker.NetworkList(ctx, types.NetworkListOptions{\n\t\tFilters: args,\n\t})\n\tif err != nil {\n\t\tlog.Errorln(\"Failed to list docker networks\")\n\t\treturn \"\", false, err\n\t}\n\n\t// (2.1) If possible, return an existing network\n\tif len(networkList) > 1 {\n\t\tlog.Warnf(\"Found %d networks instead of only one. Choosing the first one: '%s'.\", len(networkList), networkList[0].ID)\n\t}\n\n\tif len(networkList) > 0 {\n\t\tlog.Infof(\"Network with name '%s' already exists with ID '%s'\", name, networkList[0].ID)\n\t\treturn networkList[0].ID, true, nil\n\t}\n\n\t// (3) Create a new network\n\tnetwork, err := docker.NetworkCreate(ctx, name, types.NetworkCreate{\n\t\tLabels: k3d.DefaultObjectLabels,\n\t})\n\tif err != nil {\n\t\tlog.Errorln(\"Failed to create network\")\n\t\treturn \"\", false, err\n\t}\n\n\tlog.Infof(\"Created network '%s'\", name)\n\treturn network.ID, false, nil\n}", "func CreateProspectNetwork(prospectID int64, ec2InstanceID string) (*Network, bool) {\n\tbase10Bits, ok := NextNetwork(ec2InstanceID)\n\tif !ok {\n\t\tlog.Printf(\"error: failed to get next network\")\n\t\treturn nil, false\n\t}\n\tsshPort, ok := NextPort(ec2InstanceID)\n\tif !ok {\n\t\tlog.Printf(\"error: failed to get next SSH port\")\n\t\treturn nil, false\n\t}\n\tipv4Network := calculateNetwork(base10Bits, 0)\n\trand.Seed(time.Now().Unix())\n\tnetwork := &Network{\n\t\tEC2InstanceID: ec2InstanceID,\n\t\tNetwork: ipv4Network,\n\t\tMask: defaultNetworkMask,\n\t\tProspectID: prospectID,\n\t\tExamContainerIP: calculateNetwork(base10Bits, 2),\n\t\tGatewayIP: calculateNetwork(base10Bits, 1),\n\t\tProctorContainerIP: calculateNetwork(\n\t\t\tbase10Bits,\n\t\t\trand.Int63n(24)+230),\n\t\tSSHPort: sshPort,\n\t}\n\terr := connection.QueryRow(`\n\tINSERT INTO prospect_network\n\t(ec2_instance_id, network, mask, prospect_id, exam_container_ip, proctor_container_ip, gateway_ip, ssh_port)\n\tVALUES\n\t($1, $2, $3, $4, $5, $6, $7, $8)\n\tRETURNING id, created, modified\n\t`,\n\t\tnetwork.EC2InstanceID,\n\t\tnetwork.Network,\n\t\tnetwork.Mask,\n\t\tnetwork.ProspectID,\n\t\tnetwork.ExamContainerIP,\n\t\tnetwork.ProctorContainerIP,\n\t\tnetwork.GatewayIP,\n\t\tnetwork.SSHPort,\n\t).Scan(&network.ID, &network.Created, &network.Modified)\n\tif err != nil {\n\t\tlog.Printf(\"error: failed to create network for %d on %s: %s\", prospectID, ec2InstanceID, err)\n\t\treturn nil, false\n\t}\n\treturn network, true\n}", "func CreateHnsNetwork(nwConfig cns.CreateHnsNetworkRequest) error {\n\treturn fmt.Errorf(\"CreateHnsNetwork shouldn't be called for linux platform\")\n}", "func newNetwork(config neuralNetConfig) *neuralNet {\n\treturn &neuralNet{config: config}\n}", "func (l *Libvirt) NetworkCreateXML(XML string) (rNet Network, err error) {\n\tvar buf []byte\n\n\targs := NetworkCreateXMLArgs {\n\t\tXML: XML,\n\t}\n\n\tbuf, err = encode(&args)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tvar r response\n\tr, err = l.requestStream(40, constants.Program, buf, nil, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\t// Return value unmarshaling\n\ttpd := typedParamDecoder{}\n\tct := map[string]xdr.TypeDecoder{\"libvirt.TypedParam\": tpd}\n\trdr := bytes.NewReader(r.Payload)\n\tdec := xdr.NewDecoderCustomTypes(rdr, 0, ct)\n\t// Net: Network\n\t_, err = dec.Decode(&rNet)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func NetworkPost(w http.ResponseWriter, r *http.Request) {\n\tvar newNet swl.NetworkCreate\n\n\tbody, err := ioutil.ReadAll(io.LimitReader(r.Body, 1048576))\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tif err := r.Body.Close(); err != nil {\n\t\tpanic(err)\n\t}\n\n\tif !swl.ValidateAndUnmarshalJSON(body, &newNet, w) {\n\t\treturn\n\t}\n\n\tresp, err := goreq.Request{\n\t\tMethod: \"POST\",\n\t\tUri: remoteURL + \"/networks\",\n\t\tBody: body,\n\t}.Do()\n\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tdefer resp.Body.Close()\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=UTF-8\")\n\tw.WriteHeader(resp.StatusCode)\n\tresbody, _ := ioutil.ReadAll(resp.Body)\n\tw.Write(resbody)\n}", "func makeNodeWithNetwork(t *testing.T, name string, networksAndIPs map[string]string) *v1.Node {\n\tt.Helper()\n\tnode := &v1.Node{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tName: name,\n\t\t\tAnnotations: map[string]string{},\n\t\t},\n\t}\n\tvar northInterfaces networkv1.NorthInterfacesAnnotation\n\tfor netName, ip := range networksAndIPs {\n\t\tnorthInterfaces = append(northInterfaces, networkv1.NorthInterface{\n\t\t\tNetwork: netName,\n\t\t\tIpAddress: ip,\n\t\t})\n\n\t}\n\tif len(northInterfaces) > 0 {\n\t\tannotation, err := networkv1.MarshalNorthInterfacesAnnotation(northInterfaces)\n\t\tif err != nil {\n\t\t\tt.Errorf(\"could not create node annotations\")\n\t\t}\n\t\tnode.ObjectMeta.Annotations[networkv1.NorthInterfacesAnnotationKey] = annotation\n\t}\n\treturn node\n}", "func NewNetwork(c *Config) *Network {\n\tn := c.Inputs\n\tr := c.ConnectionRatio\n\tw := c.SharedWeight\n\tif n <= 0 {\n\t\treturn nil\n\t}\n\t// Pre-allocate room for n neurons and set the shared weight to the configured value.\n\toutputNode := NewRandomNeuron()\n\tallNodes := make([]Neuron, 0, n+1)\n\tallNodes = append(allNodes, outputNode)\n\tnet := &Network{allNodes, make([]*Neuron, n), &outputNode, w, 100}\n\n\t// Initialize n input nodes that all are inputs to the one output node.\n\tfor i := 0; i < n; i++ {\n\t\tinputNode := NewRandomNeuron()\n\t\tallNodes = append(allNodes, inputNode)\n\t\tnet.InputNodes[i] = &inputNode\n\t\t// Make connections for all nodes where a random number between 0 and 1 are larger than r\n\t\tif rand.Float64() > r {\n\t\t\tif err := net.OutputNode.AddInput(net.InputNodes[i]); err != nil {\n\t\t\t\tpanic(err)\n\t\t\t}\n\t\t}\n\t}\n\n\tnet.AllNodes = allNodes\n\treturn net\n}", "func CreateNetns(name string) (ns.NetNS, error) {\n\truntime.LockOSThread()\n\tdefer runtime.UnlockOSThread()\n\t// Get current namespace.\n\tcurrentNs, err := ns.GetCurrentNS()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tnamespacePath := nsPath + name\n\terr = DeleteNetns(name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// Create a new network namespace.\n\t_, err = netns.NewNamed(name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tnetNamespace, err := ns.GetNS(namespacePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// Set back the original namespace.\n\tif err = currentNs.Set(); err != nil {\n\t\treturn nil, err\n\t}\n\treturn netNamespace, nil\n}", "func (client *Client) CreateNetworkPool(req *Request) (*Response, error) {\n\treturn client.Execute(&Request{\n\t\tMethod: \"POST\",\n\t\tPath: NetworkPoolsPath,\n\t\tQueryParams: req.QueryParams,\n\t\tBody: req.Body,\n\t\tResult: &CreateNetworkPoolResult{},\n\t})\n}", "func newNetworkLayer(ln net.Listener, publicRaftAddr string) *networkLayer {\n\taddr, _ := net.ResolveTCPAddr(\"tcp\", publicRaftAddr)\n\treturn &networkLayer{\n\t\tln: ln,\n\t\taddr: addr,\n\t}\n}", "func (osh *SystemHandler) NewNetworkNamespace() (ns netns.NsHandle, err error) {\n\treturn netns.New()\n}", "func (a *NetworkServerAPI) Create(ctx context.Context, req *pb.CreateNetworkServerRequest) (*pb.CreateNetworkServerResponse, error) {\n\tif err := a.validator.Validate(ctx,\n\t\tauth.ValidateNetworkServersAccess(auth.Create, 0),\n\t); err != nil {\n\t\treturn nil, grpc.Errorf(codes.Unauthenticated, \"authentication failed: %s\", err)\n\t}\n\n\tns := storage.NetworkServer{\n\t\tName: req.Name,\n\t\tServer: req.Server,\n\t\tCACert: req.CaCert,\n\t\tTLSCert: req.TlsCert,\n\t\tTLSKey: req.TlsKey,\n\t\tRoutingProfileCACert: req.RoutingProfileCACert,\n\t\tRoutingProfileTLSCert: req.RoutingProfileTLSCert,\n\t\tRoutingProfileTLSKey: req.RoutingProfileTLSKey,\n\t}\n\n\terr := storage.Transaction(config.C.PostgreSQL.DB, func(tx sqlx.Ext) error {\n\t\treturn storage.CreateNetworkServer(tx, &ns)\n\t})\n\tif err != nil {\n\t\treturn nil, errToRPCError(err)\n\t}\n\n\treturn &pb.CreateNetworkServerResponse{\n\t\tId: ns.ID,\n\t}, nil\n}", "func ensureNetwork(name string) error {\n\t// TODO: the network might already exist and not have ipv6 ... :|\n\t// discussion: https://github.com/kubernetes-sigs/kind/pull/1508#discussion_r414594198\n\texists, err := checkIfNetworkExists(name)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// network already exists, we're good\n\tif exists {\n\t\treturn nil\n\t}\n\n\t// Generate unique subnet per network based on the name\n\t// obtained from the ULA fc00::/8 range\n\t// Make N attempts with \"probing\" in case we happen to collide\n\tsubnet := generateULASubnetFromName(name, 0)\n\terr = createNetwork(name, subnet)\n\tif err == nil {\n\t\t// Success!\n\t\treturn nil\n\t}\n\n\t// On the first try check if ipv6 fails entirely on this machine\n\t// https://github.com/kubernetes-sigs/kind/issues/1544\n\t// Otherwise if it's not a pool overlap error, fail\n\t// If it is, make more attempts below\n\tif isIPv6UnavailableError(err) {\n\t\t// only one attempt, IPAM is automatic in ipv4 only\n\t\treturn createNetwork(name, \"\")\n\t} else if isPoolOverlapError(err) {\n\t\t// unknown error ...\n\t\treturn err\n\t}\n\n\t// keep trying for ipv6 subnets\n\tconst maxAttempts = 5\n\tfor attempt := int32(1); attempt < maxAttempts; attempt++ {\n\t\tsubnet := generateULASubnetFromName(name, attempt)\n\t\terr = createNetwork(name, subnet)\n\t\tif err == nil {\n\t\t\t// success!\n\t\t\treturn nil\n\t\t} else if !isPoolOverlapError(err) {\n\t\t\t// unknown error ...\n\t\t\treturn err\n\t\t}\n\t}\n\treturn errors.New(\"exhausted attempts trying to find a non-overlapping subnet\")\n}", "func CreateServerNetwork() (*ServerNetwork, error) {\n\tserverBroker, err := genericNetwork.NewNetwork(genericNetwork.MQTT)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tserverNet := ServerNetwork{\n\t\tIface: serverBroker,\n\t\tEvents: make(chan map[string]sd.SwitchStatus),\n\t}\n\treturn &serverNet, nil\n\n}", "func NewNetwork(prcId uint32, chanFromMutex chan string, chanToMutex chan string, mutex mutex.Mutex)*Network {\n\tnetwork := new(Network)\n\tnetwork.procId = prcId\n\tnetwork.in = make(chan string)\n\tnetwork.out = make(chan string)\n\tnetwork.entering = make(chan struct{\n\t\tuint32\n\t\tnet.Conn\n\t})\n\tnetwork.chanToMutex = chanToMutex\n\tnetwork.chanFromMutex = chanFromMutex\n\tnetwork.mutex = mutex\n\tnetwork.site = make(map[uint32]net.Conn)\n\treturn network\n\n}", "func newNetwork(peers ...stateMachine) *network {\n\treturn newNetworkWithConfig(nil, peers...)\n}", "func newNetwork(t *testing.T, cfg Config) *Network {\n\tt.Helper()\n\t// only one caller/test can create and use a network at a time\n\tt.Log(\"acquiring test network lock\")\n\tlock.Lock()\n\n\tbaseDir, err := ioutil.TempDir(t.TempDir(), cfg.ChainID)\n\trequire.NoError(t, err)\n\tt.Logf(\"created temporary directory: %s\", baseDir)\n\n\tnetwork := &Network{\n\t\tT: t,\n\t\tBaseDir: baseDir,\n\t\tValidators: make([]*Validator, cfg.NumValidators),\n\t\tConfig: cfg,\n\t}\n\n\tt.Log(\"preparing test network...\")\n\n\tmonikers := make([]string, cfg.NumValidators)\n\tnodeIDs := make([]string, cfg.NumValidators)\n\tvalPubKeys := make([]cryptotypes.PubKey, cfg.NumValidators)\n\n\tvar (\n\t\tgenAccounts []dclauthtypes.Account\n\t\tgenFiles []string\n\t)\n\n\tbuf := bufio.NewReader(os.Stdin)\n\n\t// generate private keys, node IDs, and initial transactions\n\tfor i := 0; i < cfg.NumValidators; i++ {\n\t\tappCfg := srvconfig.DefaultConfig()\n\t\tappCfg.Pruning = cfg.PruningStrategy\n\t\tappCfg.MinGasPrices = cfg.MinGasPrices\n\t\tappCfg.API.Enable = true\n\t\tappCfg.API.Swagger = false\n\t\tappCfg.Telemetry.Enabled = false\n\n\t\tctx := server.NewDefaultContext()\n\t\ttmCfg := ctx.Config\n\t\ttmCfg.Consensus.TimeoutCommit = cfg.TimeoutCommit\n\n\t\t// Only allow the first validator to expose an RPC, API and gRPC\n\t\t// server/client due to Tendermint in-process constraints.\n\t\tapiAddr := \"\"\n\t\ttmCfg.RPC.ListenAddress = \"\"\n\t\tappCfg.GRPC.Enable = false\n\t\tappCfg.GRPCWeb.Enable = false\n\t\tif i == 0 {\n\t\t\tapiListenAddr, _, err := server.FreeTCPAddr()\n\t\t\trequire.NoError(t, err)\n\t\t\tappCfg.API.Address = apiListenAddr\n\n\t\t\tapiURL, err := url.Parse(apiListenAddr)\n\t\t\trequire.NoError(t, err)\n\t\t\tapiAddr = \"http://\" + net.JoinHostPort(apiURL.Hostname(), apiURL.Port())\n\n\t\t\trpcAddr, _, err := server.FreeTCPAddr()\n\t\t\trequire.NoError(t, err)\n\t\t\ttmCfg.RPC.ListenAddress = rpcAddr\n\n\t\t\t_, grpcPort, err := server.FreeTCPAddr()\n\t\t\trequire.NoError(t, err)\n\t\t\tappCfg.GRPC.Address = fmt.Sprintf(\"0.0.0.0:%s\", grpcPort)\n\t\t\tappCfg.GRPC.Enable = true\n\n\t\t\t_, grpcWebPort, err := server.FreeTCPAddr()\n\t\t\trequire.NoError(t, err)\n\t\t\tappCfg.GRPCWeb.Address = fmt.Sprintf(\"0.0.0.0:%s\", grpcWebPort)\n\t\t\tappCfg.GRPCWeb.Enable = true\n\t\t}\n\n\t\tlogger := log.NewNopLogger()\n\t\tif cfg.EnableLogging {\n\t\t\tlogger = log.NewTMLogger(log.NewSyncWriter(os.Stdout))\n\t\t\tlogger, _ = tmflags.ParseLogLevel(\"info\", logger, tmcfg.DefaultLogLevel)\n\t\t}\n\n\t\tctx.Logger = logger\n\n\t\tnodeDirName := fmt.Sprintf(\"node%d\", i)\n\t\tnodeDir := filepath.Join(network.BaseDir, nodeDirName, \"simd\")\n\t\tclientDir := filepath.Join(network.BaseDir, nodeDirName, \"simcli\")\n\t\tgentxsDir := filepath.Join(network.BaseDir, \"gentxs\")\n\n\t\trequire.NoError(t, os.MkdirAll(filepath.Join(nodeDir, \"config\"), 0o755))\n\t\trequire.NoError(t, os.MkdirAll(clientDir, 0o755))\n\n\t\ttmCfg.SetRoot(nodeDir)\n\t\ttmCfg.Moniker = nodeDirName\n\t\tmonikers[i] = nodeDirName\n\n\t\tproxyAddr, _, err := server.FreeTCPAddr()\n\t\trequire.NoError(t, err)\n\t\ttmCfg.ProxyApp = proxyAddr\n\n\t\tp2pAddr, _, err := server.FreeTCPAddr()\n\t\trequire.NoError(t, err)\n\t\ttmCfg.P2P.ListenAddress = p2pAddr\n\t\ttmCfg.P2P.AddrBookStrict = false\n\t\ttmCfg.P2P.AllowDuplicateIP = true\n\n\t\tnodeID, pubKey, err := dclgenutil.InitializeNodeValidatorFiles(tmCfg)\n\t\trequire.NoError(t, err)\n\t\tnodeIDs[i] = nodeID\n\t\tvalPubKeys[i] = pubKey\n\n\t\tkb, err := keyring.New(sdk.KeyringServiceName(), keyring.BackendTest, clientDir, buf, cfg.KeyringOptions...)\n\t\trequire.NoError(t, err)\n\n\t\tkeyringAlgos, _ := kb.SupportedAlgorithms()\n\t\talgo, err := keyring.NewSigningAlgoFromString(cfg.SigningAlgo, keyringAlgos)\n\t\trequire.NoError(t, err)\n\n\t\taddr, secret, err := server.GenerateSaveCoinKey(kb, nodeDirName, true, algo)\n\t\trequire.NoError(t, err)\n\n\t\tinfo := map[string]string{\"secret\": secret}\n\t\tinfoBz, err := json.Marshal(info)\n\t\trequire.NoError(t, err)\n\n\t\t// save private key seed words\n\t\trequire.NoError(t, writeFile(fmt.Sprintf(\"%v.json\", \"key_seed\"), clientDir, infoBz))\n\n\t\tgenFiles = append(genFiles, tmCfg.GenesisFile())\n\n\t\tbaseAccount := authtypes.NewBaseAccount(addr, nil, 0, 0)\n\t\taccount := dclauthtypes.NewAccount(\n\t\t\tbaseAccount,\n\t\t\t[]dclauthtypes.AccountRole{\n\t\t\t\tdclauthtypes.Vendor,\n\t\t\t\tdclauthtypes.CertificationCenter,\n\t\t\t\tdclauthtypes.Trustee,\n\t\t\t\tdclauthtypes.NodeAdmin,\n\t\t\t},\n\t\t\tnil,\n\t\t\tnil,\n\t\t\ttestconstants.Vid,\n\t\t)\n\t\tgenAccounts = append(genAccounts, *account)\n\n\t\trequire.NoError(t, err)\n\n\t\tdescription := validatortypes.NewDescription(nodeDirName, \"\", \"\", \"\")\n\n\t\tcreateValMsg, err := validatortypes.NewMsgCreateValidator(\n\t\t\tsdk.ValAddress(addr),\n\t\t\tvalPubKeys[i],\n\t\t\t&description,\n\t\t)\n\t\trequire.NoError(t, err)\n\n\t\tp2pURL, err := url.Parse(p2pAddr)\n\t\trequire.NoError(t, err)\n\n\t\tmemo := fmt.Sprintf(\"%s@%s:%s\", nodeIDs[i], p2pURL.Hostname(), p2pURL.Port())\n\t\tfee := sdk.NewCoins(sdk.NewCoin(fmt.Sprintf(\"%stoken\", nodeDirName), sdk.NewInt(0)))\n\t\ttxBuilder := cfg.TxConfig.NewTxBuilder()\n\t\trequire.NoError(t, txBuilder.SetMsgs(createValMsg))\n\t\ttxBuilder.SetFeeAmount(fee) // Arbitrary fee\n\t\ttxBuilder.SetGasLimit(1000000) // Need at least 100386\n\t\ttxBuilder.SetMemo(memo)\n\n\t\ttxFactory := tx.Factory{}\n\t\ttxFactory = txFactory.\n\t\t\tWithChainID(cfg.ChainID).\n\t\t\tWithMemo(memo).\n\t\t\tWithKeybase(kb).\n\t\t\tWithTxConfig(cfg.TxConfig)\n\n\t\terr = tx.Sign(txFactory, nodeDirName, txBuilder, true)\n\t\trequire.NoError(t, err)\n\n\t\ttxBz, err := cfg.TxConfig.TxJSONEncoder()(txBuilder.GetTx())\n\t\trequire.NoError(t, err)\n\t\trequire.NoError(t, writeFile(fmt.Sprintf(\"%v.json\", nodeDirName), gentxsDir, txBz))\n\n\t\tsrvconfig.WriteConfigFile(filepath.Join(nodeDir, \"config/app.toml\"), appCfg)\n\n\t\tclientCtx := client.Context{}.\n\t\t\tWithKeyringDir(clientDir).\n\t\t\tWithKeyring(kb).\n\t\t\tWithHomeDir(tmCfg.RootDir).\n\t\t\tWithChainID(cfg.ChainID).\n\t\t\tWithInterfaceRegistry(cfg.InterfaceRegistry).\n\t\t\tWithCodec(cfg.Codec).\n\t\t\tWithLegacyAmino(cfg.LegacyAmino).\n\t\t\tWithTxConfig(cfg.TxConfig).\n\t\t\tWithAccountRetriever(cfg.AccountRetriever)\n\n\t\tnetwork.Validators[i] = &Validator{\n\t\t\tAppConfig: appCfg,\n\t\t\tClientCtx: clientCtx,\n\t\t\tCtx: ctx,\n\t\t\tDir: filepath.Join(network.BaseDir, nodeDirName),\n\t\t\tNodeID: nodeID,\n\t\t\tPubKey: pubKey,\n\t\t\tMoniker: nodeDirName,\n\t\t\tRPCAddress: tmCfg.RPC.ListenAddress,\n\t\t\tP2PAddress: tmCfg.P2P.ListenAddress,\n\t\t\tAPIAddress: apiAddr,\n\t\t\tAddress: addr,\n\t\t\tValAddress: sdk.ValAddress(addr),\n\t\t}\n\t}\n\n\trequire.NoError(t, initGenFiles(cfg, genAccounts, genFiles))\n\trequire.NoError(t, collectGenFiles(cfg, network.Validators, network.BaseDir))\n\n\tt.Log(\"starting test network...\")\n\tfor _, v := range network.Validators {\n\t\trequire.NoError(t, startInProcess(cfg, v))\n\t}\n\n\tt.Log(\"started test network\")\n\n\t// Ensure we cleanup incase any test was abruptly halted (e.g. SIGINT) as any\n\t// defer in a test would not be called.\n\tserver.TrapSignal(network.Cleanup)\n\n\treturn network\n}", "func (a *ASA) CreateNetworkService(n *NetworkService, duplicateAction int) error {\n\tvar err error\n\t_, err = a.Post(apiNetworkServicesEndpoint, n)\n\tif err != nil {\n\t\tasaErr := err.(ASAError)\n\n\t\tn.ObjectID = asaErr.Details\n\t\tif asaErr.Code == errorDuplicate {\n\t\t\tif a.debug {\n\t\t\t\tglog.Warningf(\"This is a duplicate\\n\")\n\t\t\t}\n\t\t\tif duplicateAction == DuplicateActionError {\n\t\t\t\treturn err\n\t\t\t}\n\t\t} else {\n\t\t\tif a.debug {\n\t\t\t\tglog.Errorf(\"Error: %s\\n\", err)\n\t\t\t}\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\tn.ObjectID = n.Name\n\t}\n\n\treturn nil\n}", "func (s *Service) CreateNetworkRule(req CreateNetworkRuleRequest) (TaskReference, error) {\n\tbody, err := s.createNetworkRuleResponseBody(req)\n\n\treturn body.Data, err\n}", "func CreateDefaultExtNetwork(networkType string) error {\n\treturn fmt.Errorf(\"CreateDefaultExtNetwork shouldn't be called for linux platform\")\n}", "func (service *ContrailService) CreateNetworkPolicy(\n\tctx context.Context,\n\trequest *models.CreateNetworkPolicyRequest) (*models.CreateNetworkPolicyResponse, error) {\n\tmodel := request.NetworkPolicy\n\tif model.UUID == \"\" {\n\t\tmodel.UUID = uuid.NewV4().String()\n\t}\n\tauth := common.GetAuthCTX(ctx)\n\tif auth == nil {\n\t\treturn nil, common.ErrorUnauthenticated\n\t}\n\n\tif model.FQName == nil {\n\t\tif model.DisplayName != \"\" {\n\t\t\tmodel.FQName = []string{auth.DomainID(), auth.ProjectID(), model.DisplayName}\n\t\t} else {\n\t\t\tmodel.FQName = []string{auth.DomainID(), auth.ProjectID(), model.UUID}\n\t\t}\n\t}\n\tmodel.Perms2 = &models.PermType2{}\n\tmodel.Perms2.Owner = auth.ProjectID()\n\tif err := common.DoInTransaction(\n\t\tservice.DB,\n\t\tfunc(tx *sql.Tx) error {\n\t\t\treturn db.CreateNetworkPolicy(ctx, tx, request)\n\t\t}); err != nil {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"err\": err,\n\t\t\t\"resource\": \"network_policy\",\n\t\t}).Debug(\"db create failed on create\")\n\t\treturn nil, common.ErrorInternal\n\t}\n\treturn &models.CreateNetworkPolicyResponse{\n\t\tNetworkPolicy: request.NetworkPolicy,\n\t}, nil\n}", "func InitializeNewNetwork(structure []int, learningRate, momentumFactor float64, epochCount int, activationFunction, activationFunctionDerivative func(float64) float64) ConnectedNetwork {\n\n\trand.Seed(time.Now().UnixNano())\n\n\t//Initialize the network object\n\tn := ConnectedNetwork{\n\t\tStructure: structure,\n\t\tWeights: [][][]float64{},\n\t\tActivations: [][]float64{},\n\t\tDeltas: [][]float64{},\n\t\tChanges: [][]float64{},\n\t\tLearningRate: learningRate,\n\t\tMomentumFactor: momentumFactor,\n\t\tEpochCount: epochCount,\n\t\tActivationFunction: activationFunction,\n\t\tActivationFunctionDerivative: activationFunctionDerivative,\n\t}\n\n\ttmp := []float64{}\n\tfor i := 0; i < structure[0]; i++ {\n\t\ttmp = append(tmp, 0)\n\t}\n\tn.Activations = append(n.Activations, tmp)\n\tn.Changes = append(n.Changes, tmp)\n\n\t//Initialize the weights of the network to random values\n\tfor i := 1; i < len(n.Structure); i++ {\n\t\ttemp := [][]float64{}\n\t\ttempa := []float64{}\n\t\tfor j := 0; j < n.Structure[i]; j++ {\n\t\t\ttempSec := []float64{}\n\t\t\ttempa = append(tempa, float64(0))\n\t\t\tfor k := 0; k < n.Structure[i-1]; k++ {\n\t\t\t\ttempSec = append(tempSec, rand.Float64())\n\t\t\t}\n\t\t\ttemp = append(temp, tempSec)\n\t\t}\n\t\tn.Weights = append(n.Weights, temp)\n\t\tn.Activations = append(n.Activations, tempa)\n\t\tn.Changes = append(n.Changes, CreateVector(len(tempa), 0.0))\n\t\tn.Deltas = append(n.Deltas, CreateVector(len(tempa), 0.0))\n\t}\n\treturn n\n}", "func MakeNetwork(numInputs, numOutputs int, hiddenLayers []int, fullyConnected, bias, multipleActivations bool, mainActivation string, activationList []string, connections [][][]int) Network {\n //first check for proper inputs -_-\n if hiddenLayers == nil {panic(\"No layers requested\") } //if no layers given\n if !fullyConnected && connections == nil {panic(\"Unknown connections for partially connected network\") } //if not fully connected but no connections provided\n if !fullyConnected && len(connections) != len(hiddenLayers)+1 {panic(\"Wrong number of partially connected connections provided\") } //if not fully connected but the wrong number of connection lists provided\n if multipleActivations && activationList == nil {panic(\"Not enough activations provided\") } //if not singularly activated but other options not provided\n if multipleActivations && len(activationList) != len(hiddenLayers)+1 {panic(\"Confused by the wrong number of activations provided\") } //If the wrong number of activations provided\n\n //Now we can actually start\n\n //add the output layer to the list\n layerslist := append(hiddenLayers, numOutputs)\n\n //make the output variable\n var output Network\n\n //load the easy stuff\n output.NumInputs = numInputs\n output.NumLayers = len(layerslist)\n output.NumOutputs = numOutputs\n output.FullyConnected = fullyConnected\n\n //make the layers\n for index,value := range layerslist { //for each layer to be created\n\n\t\t//connections handling\n\t\tvar cons [][]int\n\t\tif fullyConnected {\n\t\t\tcons = nil\n\t\t} else {\n\t\t\tcons = connections[index]\n\t\t}\n\n\t\t//activation handling\n\t\tvar act string\n\t\tif multipleActivations {\n\t\t\tact = activationList[index]\n\t\t} else {\n\t\t\tact = mainActivation\n\t\t}\n\n\t\t//last layer handling\n\t\tvar deaths_allowed bool\n\t\tif index == output.NumLayers - 1 {\n\t\t\tdeaths_allowed = false\n\t\t} else {\n\t\t\tdeaths_allowed = true\n\t\t}\n\n\t\t//last number of nodes (if its not the first layer, its layerlist's previous index, otherwise, its the input numbers)\n\t\tvar lastNodes int\n\t\tif index > 0 {\n\t\t\tlastNodes = layerslist[index - 1]\n\t\t} else {\n\t\t\tlastNodes = numInputs\n\t\t}\n\n\t\t//actually make the layers\n newLayer := l.MakeLayer(index, lastNodes, value, fullyConnected, bias, deaths_allowed, act, cons)\n\n //save the new layer\n output.Layers = append(output.Layers, newLayer)\n\t}\n\n output.LayerListing = layerslist //FOR DEBUG ONLY\n output.Bias = bias //for BP\n\n\treturn output\n\n}", "func (nls *NetworkListService) CreateNetworkList(opts CreateNetworkListOptions) (*NetworkListResponse, *ClientResponse, error) {\n\n\tapiURI := apiPaths[\"network_list\"]\n\n\tvar k *NetworkListResponse\n\tresp, err := nls.client.NewRequest(\"POST\", apiURI, opts, &k)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\n\treturn k, resp, err\n}", "func (client IdentityClient) CreateNetworkSource(ctx context.Context, request CreateNetworkSourceRequest) (response CreateNetworkSourceResponse, err error) {\n\tvar ociResponse common.OCIResponse\n\tpolicy := common.NoRetryPolicy()\n\tif client.RetryPolicy() != nil {\n\t\tpolicy = *client.RetryPolicy()\n\t}\n\tif request.RetryPolicy() != nil {\n\t\tpolicy = *request.RetryPolicy()\n\t}\n\n\tif !(request.OpcRetryToken != nil && *request.OpcRetryToken != \"\") {\n\t\trequest.OpcRetryToken = common.String(common.RetryToken())\n\t}\n\n\tociResponse, err = common.Retry(ctx, request, client.createNetworkSource, policy)\n\tif err != nil {\n\t\tif ociResponse != nil {\n\t\t\tif httpResponse := ociResponse.HTTPResponse(); httpResponse != nil {\n\t\t\t\topcRequestId := httpResponse.Header.Get(\"opc-request-id\")\n\t\t\t\tresponse = CreateNetworkSourceResponse{RawResponse: httpResponse, OpcRequestId: &opcRequestId}\n\t\t\t} else {\n\t\t\t\tresponse = CreateNetworkSourceResponse{}\n\t\t\t}\n\t\t}\n\t\treturn\n\t}\n\tif convertedResponse, ok := ociResponse.(CreateNetworkSourceResponse); ok {\n\t\tcommon.EcContext.UpdateEndOfWindow(time.Duration(240 * time.Second))\n\t\tresponse = convertedResponse\n\t} else {\n\t\terr = fmt.Errorf(\"failed to convert OCIResponse into CreateNetworkSourceResponse\")\n\t}\n\treturn\n}", "func (client *NetworkToNetworkInterconnectsClient) createCreateRequest(ctx context.Context, resourceGroupName string, networkFabricName string, networkToNetworkInterconnectName string, body NetworkToNetworkInterconnect, options *NetworkToNetworkInterconnectsClientBeginCreateOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.ManagedNetworkFabric/networkFabrics/{networkFabricName}/networkToNetworkInterconnects/{networkToNetworkInterconnectName}\"\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif networkFabricName == \"\" {\n\t\treturn nil, errors.New(\"parameter networkFabricName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{networkFabricName}\", url.PathEscape(networkFabricName))\n\tif networkToNetworkInterconnectName == \"\" {\n\t\treturn nil, errors.New(\"parameter networkToNetworkInterconnectName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{networkToNetworkInterconnectName}\", url.PathEscape(networkToNetworkInterconnectName))\n\treq, err := runtime.NewRequest(ctx, http.MethodPut, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2023-06-15\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, runtime.MarshalAsJSON(req, body)\n}", "func New(options configapi.NodeConfig, clusterDomain string, proxyConfig *kubeproxyconfig.KubeProxyConfiguration, enableProxy, enableDNS bool) (*NetworkConfig, error) {\n\tkubeConfig, err := configapi.GetKubeConfigOrInClusterConfig(options.MasterKubeConfig, options.MasterClientConnectionOverrides)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif err := configureKubeConfigForClientCertRotation(options, kubeConfig); err != nil {\n\t\tutilruntime.HandleError(fmt.Errorf(\"Unable to enable client certificate rotation for network components: %v\", err))\n\t}\n\tinternalKubeClient, err := kclientsetinternal.NewForConfig(kubeConfig)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\texternalKubeClient, err := kclientsetexternal.NewForConfig(kubeConfig)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tkubeClient, err := kclientset.NewForConfig(kubeConfig)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tnetworkClient, err := networkclient.NewForConfig(kubeConfig)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tinternalKubeInformers := kinternalinformers.NewSharedInformerFactory(internalKubeClient, proxyConfig.ConfigSyncPeriod.Duration)\n\n\tconfig := &NetworkConfig{\n\t\tKubeClientset: kubeClient,\n\t\tExternalKubeClientset: externalKubeClient,\n\t\tInternalKubeInformers: internalKubeInformers,\n\n\t\tProxyConfig: proxyConfig,\n\t\tEnableUnidling: options.EnableUnidling,\n\t}\n\n\tif network.IsOpenShiftNetworkPlugin(options.NetworkConfig.NetworkPluginName) {\n\t\tconfig.NetworkInformers = networkinformers.NewSharedInformerFactory(networkClient, network.DefaultInformerResyncPeriod)\n\n\t\tconfig.SDNNode, config.SDNProxy, err = NewSDNInterfaces(options, networkClient, kubeClient, internalKubeClient, internalKubeInformers, config.NetworkInformers, proxyConfig)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"SDN initialization failed: %v\", err)\n\t\t}\n\t}\n\n\tif enableDNS {\n\t\tdnsConfig, err := dns.NewServerDefaults()\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"DNS configuration was not possible: %v\", err)\n\t\t}\n\t\tif len(options.DNSBindAddress) > 0 {\n\t\t\tdnsConfig.DnsAddr = options.DNSBindAddress\n\t\t}\n\t\tdnsConfig.Domain = clusterDomain + \".\"\n\t\tdnsConfig.Local = \"openshift.default.svc.\" + dnsConfig.Domain\n\n\t\t// identify override nameservers\n\t\tvar nameservers []string\n\t\tfor _, s := range options.DNSNameservers {\n\t\t\tnameservers = append(nameservers, s)\n\t\t}\n\t\tif len(options.DNSRecursiveResolvConf) > 0 {\n\t\t\tc, err := miekgdns.ClientConfigFromFile(options.DNSRecursiveResolvConf)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, fmt.Errorf(\"could not start DNS, unable to read config file: %v\", err)\n\t\t\t}\n\t\t\tfor _, s := range c.Servers {\n\t\t\t\tnameservers = append(nameservers, net.JoinHostPort(s, c.Port))\n\t\t\t}\n\t\t}\n\n\t\tif len(nameservers) > 0 {\n\t\t\tdnsConfig.Nameservers = nameservers\n\t\t}\n\n\t\tservices, err := dns.NewCachedServiceAccessor(internalKubeInformers.Core().InternalVersion().Services())\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"could not start DNS: failed to add ClusterIP index: %v\", err)\n\t\t}\n\n\t\tendpoints, err := dns.NewCachedEndpointsAccessor(internalKubeInformers.Core().InternalVersion().Endpoints())\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"could not start DNS: failed to add HostnameIP index: %v\", err)\n\t\t}\n\n\t\t// TODO: use kubeletConfig.ResolverConfig as an argument to etcd in the event the\n\t\t// user sets it, instead of passing it to the kubelet.\n\t\tconfig.DNSServer = dns.NewServer(\n\t\t\tdnsConfig,\n\t\t\tservices,\n\t\t\tendpoints,\n\t\t\t\"node\",\n\t\t)\n\t}\n\n\treturn config, nil\n}", "func NewNetwork(ctx context.Context, num int) ([]*Peer, error) {\n\tvar span opentracing.Span\n\tspan, ctx = opentracing.StartSpanFromContext(ctx, \"NewNetwork\")\n\tpeers := make([]*Peer, 0, num)\n\n\tfor i := 0; i < num; i++ {\n\t\tparams, err := testutil.RandPeerNetParams()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tp, err := NewPeer(ctx, params)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"error creating peer: %s\", err.Error())\n\t\t\treturn nil, err\n\t\t}\n\n\t\tpeers = append(peers, p)\n\t}\n\tspan.Finish()\n\treturn peers, nil\n}", "func (r *ProtocolIncus) CreateNetworkACL(acl api.NetworkACLsPost) error {\n\tif !r.HasExtension(\"network_acl\") {\n\t\treturn fmt.Errorf(`The server is missing the required \"network_acl\" API extension`)\n\t}\n\n\t// Send the request.\n\t_, _, err := r.query(\"POST\", \"/network-acls\", acl, \"\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func NewNetworkV2(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) {\n\teo.ApplyDefaults(\"network\")\n\turl, err := client.EndpointLocator(eo)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &gophercloud.ServiceClient{ProviderClient: client, Endpoint: url}, nil\n}", "func NewServerNetwork(\n\tnet, laddr string,\n\thandler func(conn Conn, cmd Command),\n\taccept func(conn Conn) bool,\n\tclosed func(conn Conn, err error),\n) *Server {\n\tif handler == nil {\n\t\tpanic(\"handler is nil\")\n\t}\n\ts := newServer()\n\ts.net = net\n\ts.laddr = laddr\n\ts.handler = handler\n\ts.accept = accept\n\ts.closed = closed\n\treturn s\n}", "func NewNetworkBuilder(name string, iprange schema.IPRange, explorer *client.Client) *NetworkBuilder {\n\treturn &NetworkBuilder{\n\t\tNetwork: workloads.Network{\n\t\t\tName: name,\n\t\t\tIprange: iprange,\n\t\t\tNetworkResources: []workloads.NetworkNetResource{},\n\t\t},\n\t\texplorer: explorer,\n\t}\n}", "func TestNetworkCreateDelete(t *testing.T) {\n\t// create netagent\n\tag, _, _ := createNetAgent(t)\n\tAssert(t, ag != nil, \"Failed to create agent %#v\", ag)\n\tdefer ag.Stop()\n\n\t// network message\n\tnt := netproto.Network{\n\t\tTypeMeta: api.TypeMeta{Kind: \"Network\"},\n\t\tObjectMeta: api.ObjectMeta{\n\t\t\tTenant: \"default\",\n\t\t\tName: \"default\",\n\t\t\tNamespace: \"default\",\n\t\t},\n\t\tSpec: netproto.NetworkSpec{\n\t\t\tVlanID: 42,\n\t\t},\n\t}\n\n\t// make create network call\n\terr := ag.CreateNetwork(&nt)\n\tAssertOk(t, err, \"Error creating network\")\n\ttnt, err := ag.FindNetwork(nt.ObjectMeta)\n\tAssertOk(t, err, \"Network was not found in DB\")\n\tAssert(t, tnt.Spec.VlanID == 42, \"Network VLAN didn't match\", tnt)\n\n\t// verify duplicate network creations succeed\n\terr = ag.CreateNetwork(&nt)\n\tAssertOk(t, err, \"Error creating duplicate network\")\n\n\t// verify duplicate network name with different content does not succeed\n\tnnt := netproto.Network{\n\t\tTypeMeta: api.TypeMeta{Kind: \"Network\"},\n\t\tObjectMeta: api.ObjectMeta{\n\t\t\tTenant: \"default\",\n\t\t\tName: \"default\",\n\t\t\tNamespace: \"default\",\n\t\t},\n\t\tSpec: netproto.NetworkSpec{\n\t\t\tVlanID: 84,\n\t\t},\n\t}\n\terr = ag.CreateNetwork(&nnt)\n\tAssert(t, (err != nil), \"conflicting network creation succeeded\")\n\n\t// verify list api works\n\tnetList := ag.ListNetwork()\n\tAssert(t, len(netList) == 2, \"Incorrect number of networks\")\n\n\t// delete the network and verify its gone from db\n\terr = ag.DeleteNetwork(nt.Tenant, nt.Namespace, nt.Name)\n\tAssertOk(t, err, \"Error deleting network\")\n\t_, err = ag.FindNetwork(nt.ObjectMeta)\n\tAssert(t, err != nil, \"Network was still found in database after deleting\", ag)\n\n\t// verify you can not delete non-existing network\n\terr = ag.DeleteNetwork(nt.Tenant, nt.Namespace, nt.Name)\n\tAssert(t, err != nil, \"deleting non-existing network succeeded\", ag)\n}", "func (n *Network) ValidateCreate() error {\n\tvar allErrors field.ErrorList\n\n\tnetworklog.Info(\"validate create\", \"name\", n.Name)\n\n\t// shared validation rules with update\n\tallErrors = append(allErrors, n.Validate()...)\n\n\tif len(allErrors) == 0 {\n\t\treturn nil\n\t}\n\n\treturn apierrors.NewInvalid(schema.GroupKind{}, n.Name, allErrors)\n}", "func New(t *testing.T, configs ...Config) *Network {\n\tt.Helper()\n\tif len(configs) > 1 {\n\t\tpanic(\"at most one config should be provided\")\n\t}\n\tvar cfg Config\n\tif len(configs) == 0 {\n\t\tcfg = DefaultConfig()\n\t} else {\n\t\tcfg = configs[0]\n\t}\n\tnet := newNetwork(t, cfg)\n\tt.Cleanup(net.Cleanup)\n\n\treturn net\n}", "func NewNetworkClient(settings Settings) Network {\n\tn := Network{}\n\tn.connect(settings.Address)\n\treturn n\n}", "func CreateVNet(vNetName string, rgroupName string) {\n\ta := wow.New(os.Stdout, spin.Get(spin.Dots), \"Creating virtual network : \"+vNetName)\n\ta.Start()\n\ttime.Sleep(2 * time.Second)\n\ta.Text(\"This would take a few minutes...\").Spinner(spin.Get(spin.Dots))\n\tcmd := exec.Command(\"az\", \"network\", \"vnet\", \"create\", \"--name\",\n\t\tvNetName, \"--resource-group\", rgroupName)\n\tvar out bytes.Buffer\n\tvar stderr bytes.Buffer\n\tcmd.Stdout = &out\n\tcmd.Stderr = &stderr\n\terr := cmd.Run()\n\tif err != nil {\n\t\tfmt.Println(fmt.Sprint(err) + \": \" + stderr.String())\n\t\treturn\n\t}\n\ta.PersistWith(spin.Spinner{}, \"....\")\n\tfmt.Println(\"Result: \" + out.String())\n\temoji.Println(\":beer: Cheers!!!\")\n}", "func newClusterNetwork(c *Client) *clusterNetwork {\n\treturn &clusterNetwork{\n\t\tr: c,\n\t}\n}", "func Create(strinput2 string, tc net.Conn) {\n\tnetwork.SendDataMessage(&tc, 4, 5, 1, strinput2)\n}", "func NewNetworkProvider(myNetID string, port int, nodeKeyPair *key.Pair, suite kyber.Group, log *logger.Logger) (*NetImpl, error) {\n\tif err := peering.CheckMyNetID(myNetID, port); err != nil {\n\t\t// can't continue because NetID parameter is not correct\n\t\tlog.Panicf(\"checkMyNetworkID: '%v'. || Check the 'netid' parameter in config.json\", err)\n\t\treturn nil, err\n\t}\n\tn := NetImpl{\n\t\tmyNetID: myNetID,\n\t\tport: port,\n\t\tpeers: make(map[string]*peer),\n\t\tpeersMutex: &sync.RWMutex{},\n\t\tnodeKeyPair: nodeKeyPair,\n\t\tsuite: suite,\n\t\tlog: log,\n\t}\n\tn.events = events.NewEvent(n.eventHandler)\n\treturn &n, nil\n}", "func (client IdentityClient) createNetworkSource(ctx context.Context, request common.OCIRequest, binaryReqBody *common.OCIReadSeekCloser, extraHeaders map[string]string) (common.OCIResponse, error) {\n\n\thttpRequest, err := request.HTTPRequest(http.MethodPost, \"/networkSources\", binaryReqBody, extraHeaders)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar response CreateNetworkSourceResponse\n\tvar httpResponse *http.Response\n\thttpResponse, err = client.Call(ctx, &httpRequest)\n\tdefer common.CloseBodyIfValid(httpResponse)\n\tresponse.RawResponse = httpResponse\n\tif err != nil {\n\t\treturn response, err\n\t}\n\n\terr = common.UnmarshalResponse(httpResponse, &response)\n\treturn response, err\n}", "func (p *Provisioner) networkProvisionImpl(ctx context.Context, reservation *provision.Reservation) error {\n\tnr := pkg.NetResource{}\n\tif err := json.Unmarshal(reservation.Data, &nr); err != nil {\n\t\treturn fmt.Errorf(\"failed to unmarshal network from reservation: %w\", err)\n\t}\n\n\tif err := validateNR(nr); err != nil {\n\t\treturn fmt.Errorf(\"validation of the network resource failed: %w\", err)\n\t}\n\n\tnr.NetID = networkID(reservation.User, nr.Name)\n\n\tmgr := stubs.NewNetworkerStub(p.zbus)\n\tlog.Debug().Str(\"network\", fmt.Sprintf(\"%+v\", nr)).Msg(\"provision network\")\n\n\t_, err := mgr.CreateNR(nr)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"failed to create network resource for network %s\", nr.NetID)\n\t}\n\n\treturn nil\n}", "func (o RegionNetworkEndpointGroupOutput) Network() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *RegionNetworkEndpointGroup) pulumi.StringPtrOutput { return v.Network }).(pulumi.StringPtrOutput)\n}", "func TestAddNetwork(t *testing.T) {\n\t// initialize rsrcMgr since we use it for resource allocation\n\trsrcMgr.Init(nil)\n\n\t// Initialize the ctrler\n\tInit()\n\n\t// Create network\n\tnetwork, err := NewNetwork(\"default\")\n\tif err != nil {\n\t\tt.Errorf(\"Error creating network default. Err: %v\", err)\n\t\treturn\n\t}\n\n\tlog.Infof(\"Successfully Created network: %+v\", network)\n\n\t// Create new endpoint\n\tep, err := network.NewEndPoint(\"alta1234.0\")\n\tif err != nil {\n\t\tt.Errorf(\"Error creating network endpoint. Err: %v\", err)\n\t\treturn\n\t}\n\n\tlog.Infof(\"Successfully Created endpoint: %+v\", ep)\n}", "func NewNetworkDevice(t, id string) (NetDev, error) {\n\tvar netdev NetDev\n\n\tif t != \"user\" && t != \"tap\" {\n\t\treturn netdev, fmt.Errorf(\"Unsupported netdev type\")\n\t}\n\tif len(id) == 0 {\n\t\treturn netdev, fmt.Errorf(\"You must specify a netdev ID\")\n\t}\n\n\tnetdev.Type = t\n\tnetdev.ID = id\n\n\treturn netdev, nil\n}" ]
[ "0.7904054", "0.74405885", "0.7367915", "0.73364156", "0.72807366", "0.724993", "0.7249608", "0.720847", "0.7190325", "0.71097827", "0.7107186", "0.7097025", "0.6948547", "0.6921092", "0.69033426", "0.6875668", "0.68736416", "0.68717694", "0.6846474", "0.6840961", "0.6768673", "0.67512393", "0.66800326", "0.6669319", "0.66416186", "0.6640618", "0.66219616", "0.6610562", "0.6608851", "0.6599996", "0.65918267", "0.65865475", "0.65690416", "0.6562306", "0.65583503", "0.6544835", "0.6527006", "0.6474513", "0.64267176", "0.6417737", "0.63911915", "0.63851446", "0.63798004", "0.6359891", "0.6337329", "0.63152957", "0.6275991", "0.62714356", "0.6270296", "0.62686825", "0.626231", "0.6248733", "0.6211044", "0.61774826", "0.6173037", "0.6101074", "0.6091276", "0.60816634", "0.6080191", "0.6043613", "0.5995769", "0.5993364", "0.59848803", "0.5983413", "0.596986", "0.59364146", "0.5934349", "0.5917059", "0.5876769", "0.5849481", "0.584783", "0.5837532", "0.58344436", "0.58198136", "0.58177423", "0.5788399", "0.5766457", "0.5748942", "0.57390505", "0.573466", "0.5724964", "0.5723745", "0.5720955", "0.56988645", "0.56660223", "0.56643707", "0.566139", "0.56612974", "0.5659225", "0.5647562", "0.56273025", "0.5609194", "0.55983824", "0.5596327", "0.5595614", "0.5593982", "0.55764586", "0.5575265", "0.5574708", "0.5572434" ]
0.6956183
12
CreateSubnetwork uses the override method CreateSubnetworkFn or the real implementation.
func (c *TestClient) CreateSubnetwork(project, region string, n *compute.Subnetwork) error { if c.CreateSubnetworkFn != nil { return c.CreateSubnetworkFn(project, region, n) } return c.client.CreateSubnetwork(project, region, n) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *SubnetListener) Create(inctx context.Context, in *protocol.SubnetCreateRequest) (_ *protocol.Subnet, err error) {\n\tdefer fail.OnExitConvertToGRPCStatus(inctx, &err)\n\tdefer fail.OnExitLogError(inctx, &err)\n\tdefer fail.OnExitWrapError(inctx, &err, \"cannot create Subnet\")\n\n\tif s == nil {\n\t\treturn nil, fail.InvalidInstanceError()\n\t}\n\tif in == nil {\n\t\treturn nil, fail.InvalidParameterError(\"in\", \"cannot be nil\")\n\t}\n\tif inctx == nil {\n\t\treturn nil, fail.InvalidParameterError(\"inctx\", \"cannot be nil\")\n\t}\n\n\tnetworkRef, _ := srvutils.GetReference(in.GetNetwork())\n\tif networkRef == \"\" {\n\t\treturn nil, fail.InvalidParameterError(\"in.Network\", \"must contain an ID or a Name\")\n\t}\n\n\tjob, xerr := PrepareJob(inctx, in.GetNetwork().GetTenantId(), fmt.Sprintf(\"/subnet/%s/create\", networkRef))\n\tif xerr != nil {\n\t\treturn nil, xerr\n\t}\n\tdefer job.Close()\n\n\tctx := job.Context()\n\n\tvar sizing *abstract.HostSizingRequirements\n\tif in.GetGateway() != nil {\n\t\tif in.GetGateway().SizingAsString != \"\" {\n\t\t\tsizing, _, xerr = converters.HostSizingRequirementsFromStringToAbstract(in.GetGateway().GetSizingAsString())\n\t\t\tif xerr != nil {\n\t\t\t\treturn nil, xerr\n\t\t\t}\n\t\t} else if in.GetGateway().GetSizing() != nil {\n\t\t\tsizing = converters.HostSizingRequirementsFromProtocolToAbstract(in.GetGateway().GetSizing())\n\t\t}\n\t}\n\tif sizing == nil {\n\t\tsizing = &abstract.HostSizingRequirements{MinGPU: -1}\n\t}\n\tsizing.Image = in.GetGateway().GetImageId()\n\n\treq := abstract.SubnetRequest{\n\t\tName: in.GetName(),\n\t\tCIDR: in.GetCidr(),\n\t\tDomain: in.GetDomain(),\n\t\tHA: in.GetFailOver(),\n\t\tDefaultSSHPort: in.GetGateway().GetSshPort(),\n\t\tKeepOnFailure: in.GetKeepOnFailure(),\n\t}\n\n\thandler := handlers.NewSubnetHandler(job)\n\tsubnetInstance, xerr := handler.Create(networkRef, req, \"\", *sizing)\n\tif xerr != nil {\n\t\treturn nil, xerr\n\t}\n\n\tlogrus.WithContext(ctx).Infof(\"Subnet '%s' successfully created.\", req.Name)\n\treturn subnetInstance.ToProtocol(ctx)\n}", "func (s *Stack) CreateNetwork(req abstract.NetworkRequest) (*abstract.Network, fail.Error) {\n\t// disable subnetwork auto-creation\n\tne := compute.Network{\n\t\tName: s.GcpConfig.NetworkName,\n\t\tAutoCreateSubnetworks: false,\n\t\tForceSendFields: []string{\"AutoCreateSubnetworks\"},\n\t}\n\n\tcompuService := s.ComputeService\n\n\trecreateSafescaleNetwork := true\n\trecnet, err := compuService.Networks.Get(s.GcpConfig.ProjectID, ne.Name).Do()\n\tif recnet != nil && err == nil {\n\t\trecreateSafescaleNetwork = false\n\t} else if err != nil {\n\t\tif gerr, ok := err.(*googleapi.Error); ok {\n\t\t\tif gerr.Code != 404 {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t} else {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tif recreateSafescaleNetwork {\n\t\topp, err := compuService.Networks.Insert(s.GcpConfig.ProjectID, &ne).Context(context.Background()).Do()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\toco := OpContext{\n\t\t\tOperation: opp,\n\t\t\tProjectID: s.GcpConfig.ProjectID,\n\t\t\tService: compuService,\n\t\t\tDesiredState: \"DONE\",\n\t\t}\n\n\t\terr = waitUntilOperationIsSuccessfulOrTimeout(oco, temporal.GetMinDelay(), 2*temporal.GetContextTimeout())\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tnecreated, err := compuService.Networks.Get(s.GcpConfig.ProjectID, ne.Name).Do()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tnet := abstract.NewNetwork()\n\tnet.ID = strconv.FormatUint(necreated.Id, 10)\n\tnet.Name = necreated.Name\n\n\t// Create subnetwork\n\n\ttheRegion := s.GcpConfig.Region\n\n\tsubnetReq := compute.Subnetwork{\n\t\tIpCidrRange: req.CIDR,\n\t\tName: req.Name,\n\t\tNetwork: fmt.Sprintf(\"projects/%s/global/networks/%s\", s.GcpConfig.ProjectID, s.GcpConfig.NetworkName),\n\t\tRegion: theRegion,\n\t}\n\n\topp, err := compuService.Subnetworks.Insert(\n\t\ts.GcpConfig.ProjectID, theRegion, &subnetReq,\n\t).Context(context.Background()).Do()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\toco := OpContext{\n\t\tOperation: opp,\n\t\tProjectID: s.GcpConfig.ProjectID,\n\t\tService: compuService,\n\t\tDesiredState: \"DONE\",\n\t}\n\n\terr = waitUntilOperationIsSuccessfulOrTimeout(oco, temporal.GetMinDelay(), 2*temporal.GetContextTimeout())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tgcpSubNet, err := compuService.Subnetworks.Get(s.GcpConfig.ProjectID, theRegion, req.Name).Do()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// FIXME: Add properties and GatewayID\n\tsubnet := abstract.NewNetwork()\n\tsubnet.ID = strconv.FormatUint(gcpSubNet.Id, 10)\n\tsubnet.Name = gcpSubNet.Name\n\tsubnet.CIDR = gcpSubNet.IpCidrRange\n\tsubnet.IPVersion = ipversion.IPv4\n\n\tbuildNewRule := true\n\tfirewallRuleName := fmt.Sprintf(\"%s-%s-all-in\", s.GcpConfig.NetworkName, gcpSubNet.Name)\n\n\tfws, err := compuService.Firewalls.Get(s.GcpConfig.ProjectID, firewallRuleName).Do()\n\tif fws != nil && err == nil {\n\t\tbuildNewRule = false\n\t} else if err != nil {\n\t\tif gerr, ok := err.(*googleapi.Error); ok {\n\t\t\tif gerr.Code != 404 {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t} else {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tif buildNewRule {\n\t\tfiw := compute.Firewall{\n\t\t\tAllowed: []*compute.FirewallAllowed{\n\t\t\t\t{\n\t\t\t\t\tIPProtocol: \"all\",\n\t\t\t\t},\n\t\t\t},\n\t\t\tDirection: \"INGRESS\",\n\t\t\tDisabled: false,\n\t\t\tName: firewallRuleName,\n\t\t\tNetwork: fmt.Sprintf(\n\t\t\t\t\"https://www.googleapis.com/compute/v1/projects/%s/global/networks/%s\", s.GcpConfig.ProjectID,\n\t\t\t\ts.GcpConfig.NetworkName,\n\t\t\t),\n\t\t\tPriority: 999,\n\t\t\tSourceRanges: []string{\"0.0.0.0/0\"},\n\t\t}\n\n\t\topp, err = compuService.Firewalls.Insert(s.GcpConfig.ProjectID, &fiw).Do()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\toco = OpContext{\n\t\t\tOperation: opp,\n\t\t\tProjectID: s.GcpConfig.ProjectID,\n\t\t\tService: compuService,\n\t\t\tDesiredState: \"DONE\",\n\t\t}\n\n\t\terr = waitUntilOperationIsSuccessfulOrTimeout(oco, temporal.GetMinDelay(), temporal.GetHostTimeout())\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tbuildNewNATRule := true\n\tnatRuleName := fmt.Sprintf(\"%s-%s-nat-allowed\", s.GcpConfig.NetworkName, gcpSubNet.Name)\n\n\trfs, err := compuService.Routes.Get(s.GcpConfig.ProjectID, natRuleName).Do()\n\tif rfs != nil && err == nil {\n\t\tbuildNewNATRule = false\n\t} else if err != nil {\n\t\tif gerr, ok := err.(*googleapi.Error); ok {\n\t\t\tif gerr.Code != 404 {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t} else {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tif buildNewNATRule {\n\t\troute := &compute.Route{\n\t\t\tDestRange: \"0.0.0.0/0\",\n\t\t\tName: natRuleName,\n\t\t\tNetwork: fmt.Sprintf(\n\t\t\t\t\"https://www.googleapis.com/compute/v1/projects/%s/global/networks/%s\", s.GcpConfig.ProjectID,\n\t\t\t\ts.GcpConfig.NetworkName,\n\t\t\t),\n\t\t\tNextHopInstance: fmt.Sprintf(\n\t\t\t\t\"projects/%s/zones/%s/instances/gw-%s\", s.GcpConfig.ProjectID, s.GcpConfig.Zone, req.Name,\n\t\t\t),\n\t\t\tPriority: 800,\n\t\t\tTags: []string{fmt.Sprintf(\"no-ip-%s\", gcpSubNet.Name)},\n\t\t}\n\t\topp, err := compuService.Routes.Insert(s.GcpConfig.ProjectID, route).Do()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\toco = OpContext{\n\t\t\tOperation: opp,\n\t\t\tProjectID: s.GcpConfig.ProjectID,\n\t\t\tService: compuService,\n\t\t\tDesiredState: \"DONE\",\n\t\t}\n\n\t\terr = waitUntilOperationIsSuccessfulOrTimeout(oco, temporal.GetMinDelay(), 2*temporal.GetContextTimeout())\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t}\n\n\t// FIXME: Validation before return...\n\treturn subnet, nil\n}", "func (u *User) CreateSubnet(nodeID, data string, idempotencyKey ...string) (map[string]interface{}, error) {\n\tlog.info(\"========== CREATE SUBNET ==========\")\n\turl := buildURL(path[\"users\"], u.UserID, path[\"nodes\"], nodeID, path[\"subnets\"])\n\n\treturn u.do(\"POST\", url, data, idempotencyKey)\n}", "func (o RegionNetworkEndpointGroupOutput) Subnetwork() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *RegionNetworkEndpointGroup) pulumi.StringPtrOutput { return v.Subnetwork }).(pulumi.StringPtrOutput)\n}", "func CreateNetworkSecurityGroupRule() {}", "func (o NetworkInterfaceOutput) Subnetwork() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v NetworkInterface) *string { return v.Subnetwork }).(pulumi.StringPtrOutput)\n}", "func (d *Driver) createNetwork(config *configuration) error {\n\tif !parentExists(config.Parent) {\n\t\t// if the --internal flag is set, create a dummy link\n\t\tif config.Internal {\n\t\t\terr := createDummyLink(config.Parent, getDummyName(stringid.TruncateID(config.ID)))\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tconfig.CreatedSlaveLink = true\n\t\t\t// notify the user in logs they have limited comunicatins\n\t\t\tif config.Parent == getDummyName(stringid.TruncateID(config.ID)) {\n\t\t\t\tlogrus.Infof(\"Empty -o parent= and --internal flags limit communications to other containers inside of network: %s\",\n\t\t\t\t\tconfig.Parent)\n\t\t\t}\n\t\t} else {\n\t\t\t// if the subinterface parent_iface.vlan_id checks do not pass, return err.\n\t\t\t// a valid example is 'eth0.10' for a parent iface 'eth0' with a vlan id '10'\n\t\t\terr := createVlanLink(config.Parent)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\t// if driver created the networks slave link, record it for future deletion\n\t\t\tconfig.CreatedSlaveLink = true\n\t\t}\n\t}\n\tn := &network{\n\t\tid: config.ID,\n\t\tdriver: d,\n\t\tendpoints: endpointTable{},\n\t\tconfig: config,\n\t}\n\t// add the *network\n\td.addNetwork(n)\n\n\treturn nil\n}", "func AddSubnet(connectionName string, rsType string, vpcName string, reqInfo cres.SubnetInfo) (*cres.VPCInfo, error) {\n\tcblog.Info(\"call AddSubnet()\")\n\n\t// check empty and trim user inputs\n connectionName, err := EmptyCheckAndTrim(\"connectionName\", connectionName)\n if err != nil {\n\t\tcblog.Error(err)\n return nil, err\n }\n\n vpcName, err = EmptyCheckAndTrim(\"vpcName\", vpcName)\n if err != nil {\n\t\tcblog.Error(err)\n return nil, err\n }\n\n\tcldConn, err := ccm.GetCloudConnection(connectionName)\n\tif err != nil {\n\t\tcblog.Error(err)\n\t\treturn nil, err\n\t}\n\n\thandler, err := cldConn.CreateVPCHandler()\n\tif err != nil {\n\t\tcblog.Error(err)\n\t\treturn nil, err\n\t}\n\n\tvpcSPLock.Lock(connectionName, vpcName)\n\tdefer vpcSPLock.Unlock(connectionName, vpcName)\n\t// (1) check exist(NameID)\n\tbool_ret, err := iidRWLock.IsExistIID(iidm.SUBNETGROUP, connectionName, vpcName, reqInfo.IId)\n\tif err != nil {\n\t\tcblog.Error(err)\n\t\treturn nil, err\n\t}\n\n\tif bool_ret == true {\n\t\terr := fmt.Errorf(rsType + \"-\" + reqInfo.IId.NameId + \" already exists!\")\n\t\tcblog.Error(err)\n\t\treturn nil, err\n\t}\n\t// (2) create Resource\n\tiidVPCInfo, err := iidRWLock.GetIID(iidm.IIDSGROUP, connectionName, rsVPC, cres.IID{vpcName, \"\"})\n\tif err != nil {\n\t\tcblog.Error(err)\n\t\treturn nil, err\n\t}\n\n\tsubnetUUID, err := iidm.New(connectionName, rsType, reqInfo.IId.NameId)\n\tif err != nil {\n cblog.Error(err)\n return nil, err\n }\n\n\t// driverIID for driver\n\tsubnetReqNameId := reqInfo.IId.NameId\n\treqInfo.IId = cres.IID{subnetUUID, \"\"}\n\tinfo, err := handler.AddSubnet(getDriverIID(iidVPCInfo.IId), reqInfo)\n\tif err != nil {\n\t\tcblog.Error(err)\n\t\treturn nil, err\n\t}\n\n\t// (3) insert IID\n\t// for Subnet list\n\tfor _, subnetInfo := range info.SubnetInfoList {\t\t\n\t\tif subnetInfo.IId.NameId == reqInfo.IId.NameId { // NameId => SS-UUID\n\t\t\t// key-value structure: ~/{SUBNETGROUP}/{ConnectionName}/{VPC-NameId}/{Subnet-reqNameId} [subnet-driverNameId:subnet-driverSystemId] # VPC NameId => rsType\n\t\t\tsubnetSpiderIId := cres.IID{subnetReqNameId, subnetInfo.IId.NameId + \":\" + subnetInfo.IId.SystemId}\n\t\t\t_, err := iidRWLock.CreateIID(iidm.SUBNETGROUP, connectionName, vpcName, subnetSpiderIId)\n\t\t\tif err != nil {\n\t\t\t\tcblog.Error(err)\n\t\t\t\t// rollback\n\t\t\t\t// (1) for resource\n\t\t\t\tcblog.Info(\"<<ROLLBACK:TRY:VPC-SUBNET-CSP>> \" + subnetInfo.IId.SystemId)\n\t\t\t\t_, err2 := handler.RemoveSubnet(getDriverIID(iidVPCInfo.IId), subnetInfo.IId)\n\t\t\t\tif err2 != nil {\n\t\t\t\t\tcblog.Error(err2)\n\t\t\t\t\treturn nil, fmt.Errorf(err.Error() + \", \" + err2.Error())\n\t\t\t\t}\n\t\t\t\t// (2) for Subnet IID\n\t\t\t\tcblog.Info(\"<<ROLLBACK:TRY:VPC-SUBNET-IID>> \" + subnetInfo.IId.NameId)\n\t\t\t\t_, err3 := iidRWLock.DeleteIID(iidm.SUBNETGROUP, connectionName, vpcName, subnetSpiderIId) // vpcName => rsType\n\t\t\t\tif err3 != nil {\n\t\t\t\t\tcblog.Error(err3)\n\t\t\t\t\treturn nil, fmt.Errorf(err.Error() + \", \" + err3.Error())\n\t\t\t\t}\n\t\t\t\tcblog.Error(err)\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t}\n\t}\n\n\t// (3) set ResourceInfo(userIID)\n\tinfo.IId = getUserIID(iidVPCInfo.IId)\n\n\t// set NameId for SubnetInfo List\n\t// create new SubnetInfo List\n\tsubnetInfoList := []cres.SubnetInfo{}\n\tfor _, subnetInfo := range info.SubnetInfoList {\t\t\n\t\tsubnetIIDInfo, err := iidRWLock.GetIIDbySystemID(iidm.SUBNETGROUP, connectionName, vpcName, subnetInfo.IId) // vpcName => rsType\n\t\tif err != nil {\n\t\t\tcblog.Error(err)\n\t\t\treturn nil, err\n\t\t}\n\t\tif subnetIIDInfo.IId.NameId != \"\" { // insert only this user created.\n\t\t\tsubnetInfo.IId = getUserIID(subnetIIDInfo.IId)\n\t\t\tsubnetInfoList = append(subnetInfoList, subnetInfo)\n\t\t}\n\t}\n\tinfo.SubnetInfoList = subnetInfoList\n\n\treturn &info, nil\n}", "func (tester *ServiceTester) CreateSubnet(t *testing.T, networkID, name string, withGW bool, cidr string) (*abstract.Subnet, *abstract.HostFull) {\n\tctx := context.Background()\n\tsubnet, err := tester.Service.CreateSubnet(context.Background(), abstract.SubnetRequest{\n\t\tName: name,\n\t\tIPVersion: ipversion.IPv4,\n\t\tNetworkID: networkID,\n\t\tCIDR: cidr,\n\t})\n\trequire.NoError(t, err)\n\n\tvar gateway *abstract.HostFull\n\tif withGW {\n\t\ttpls, err := tester.Service.ListTemplatesBySizing(ctx, abstract.HostSizingRequirements{\n\t\t\tMinCores: 1,\n\t\t\tMinRAMSize: 1,\n\t\t\tMinDiskSize: 0,\n\t\t}, false)\n\t\trequire.Nil(t, err)\n\t\timg, err := tester.Service.SearchImage(ctx, \"Ubuntu 20.04\")\n\t\trequire.Nil(t, err)\n\t\tkeypair, err := tester.Service.CreateKeyPair(ctx, \"kp_\"+subnet.Name)\n\t\trequire.Nil(t, err)\n\n\t\tgwRequest := abstract.HostRequest{\n\t\t\tImageID: img.ID,\n\t\t\tSubnets: []*abstract.Subnet{subnet},\n\t\t\tKeyPair: keypair,\n\t\t\tTemplateID: tpls[0].ID,\n\t\t\tResourceName: \"gw-\" + name,\n\t\t\tIsGateway: true,\n\t\t}\n\n\t\tgateway, _, err = tester.Service.CreateHost(context.Background(), gwRequest, nil)\n\t\trequire.Nil(t, err)\n\t\tsubnet.GatewayIDs = []string{gateway.Core.ID}\n\t}\n\n\treturn subnet, gateway\n}", "func (o RouterInterfaceOutput) Subnetwork() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v RouterInterface) *string { return v.Subnetwork }).(pulumi.StringPtrOutput)\n}", "func MythicRPCTaskCreateSubtask(input MythicRPCTaskCreateSubtaskMessage) MythicRPCTaskCreateSubtaskMessageResponse {\n\tresponse := MythicRPCTaskCreateSubtaskMessageResponse{\n\t\tSuccess: false,\n\t}\n\ttaskingLocation := \"mythic_rpc\"\n\tcreateTaskInput := CreateTaskInput{\n\t\tParentTaskID: &input.TaskID,\n\t\tCommandName: input.CommandName,\n\t\tParams: input.Params,\n\t\tToken: input.Token,\n\t\tParameterGroupName: input.ParameterGroupName,\n\t\tSubtaskCallbackFunction: input.SubtaskCallbackFunction,\n\t\tTaskingLocation: &taskingLocation,\n\t}\n\ttask := databaseStructs.Task{}\n\toperatorOperation := databaseStructs.Operatoroperation{}\n\tif err := database.DB.Get(&task, `SELECT \n\tcallback.id \"callback.id\",\n\tcallback.display_id \"callback.display_id\",\n\tcallback.operation_id \"callback.operation_id\",\n\toperator.id \"operator.id\",\n\toperator.admin \"operator.admin\" \n\tFROM task\n\tJOIN callback ON task.callback_id = callback.id \n\tJOIN operator ON task.operator_id = operator.id\n\tWHERE task.id=$1`, input.TaskID); err != nil {\n\t\tresponse.Error = err.Error()\n\t\tlogging.LogError(err, \"Failed to fetch task/callback information when creating subtask\")\n\t\treturn response\n\t} else if err := database.DB.Get(&operatorOperation, `SELECT\n\tbase_disabled_commands_id\n\tFROM operatoroperation\n\tWHERE operator_id = $1 AND operation_id = $2\n\t`, task.Operator.ID, task.Callback.OperationID); err != nil {\n\t\tlogging.LogError(err, \"Failed to get operation information when creating subtask\")\n\t\tresponse.Error = err.Error()\n\t\treturn response\n\t} else {\n\t\tcreateTaskInput.IsOperatorAdmin = task.Operator.Admin\n\t\tcreateTaskInput.CallbackDisplayID = task.Callback.DisplayID\n\t\tcreateTaskInput.CurrentOperationID = task.Callback.OperationID\n\t\tif operatorOperation.BaseDisabledCommandsID.Valid {\n\t\t\tbaseDisabledCommandsID := int(operatorOperation.BaseDisabledCommandsID.Int64)\n\t\t\tcreateTaskInput.DisabledCommandID = &baseDisabledCommandsID\n\t\t}\n\t\tcreateTaskInput.OperatorID = task.Operator.ID\n\t\t// create a subtask of this task\n\t\tcreationResponse := CreateTask(createTaskInput)\n\t\tif creationResponse.Status == \"success\" {\n\t\t\tresponse.Success = true\n\t\t\tresponse.TaskID = creationResponse.TaskID\n\t\t} else {\n\t\t\tresponse.Error = creationResponse.Error\n\t\t}\n\t\treturn response\n\t}\n\n}", "func (b *BridgeNetworkDriver) Create(name string, subnet string) (*Network, error) {\n\t// 取到网段字符串中的网关ip地址和网络的ip段\n\tip, IPRange, _ := net.ParseCIDR(subnet)\n\tIPRange.IP = ip\n\n\tn := &Network{\n\t\tName: name,\n\t\tIPRange: IPRange,\n\t\tDriver: b.Name(),\n\t}\n\n\terr := b.initBridge(n)\n\treturn n, err\n}", "func dockerNetCreate(subnet net.IPNet, gw net.IP, bridge string, subnetID string) error {\n\tcmd := exec.Command(\"docker\", \"network\", \"create\", \"-d=ciao\", \"--ipam-driver=ciao\",\n\t\t\"--subnet=\"+subnet.String(), \"--gateway=\"+gw.String(),\n\t\t\"--opt\", \"bridge=\"+bridge, subnetID)\n\n\t_, err := cmd.CombinedOutput()\n\treturn err\n}", "func (o LookupRegionNetworkEndpointGroupResultOutput) Subnetwork() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupRegionNetworkEndpointGroupResult) string { return v.Subnetwork }).(pulumi.StringOutput)\n}", "func (p Pipeline) CreateNetwork() error {\n\treturn p.localRunner.NetworkCreator(p.Network)()\n}", "func (o RouterInterfaceResponseOutput) Subnetwork() pulumi.StringOutput {\n\treturn o.ApplyT(func(v RouterInterfaceResponse) string { return v.Subnetwork }).(pulumi.StringOutput)\n}", "func (o NetworkInterfaceResponseOutput) Subnetwork() pulumi.StringOutput {\n\treturn o.ApplyT(func(v NetworkInterfaceResponse) string { return v.Subnetwork }).(pulumi.StringOutput)\n}", "func resourceVolterraVirtualNetworkCreate(d *schema.ResourceData, meta interface{}) error {\n\tclient := meta.(*APIClient)\n\n\tcreateMeta := &ves_io_schema.ObjectCreateMetaType{}\n\tcreateSpec := &ves_io_schema_virtual_network.CreateSpecType{}\n\tcreateReq := &ves_io_schema_virtual_network.CreateRequest{\n\t\tMetadata: createMeta,\n\t\tSpec: createSpec,\n\t}\n\n\tif v, ok := d.GetOk(\"annotations\"); ok && !isIntfNil(v) {\n\n\t\tms := map[string]string{}\n\n\t\tfor k, v := range v.(map[string]interface{}) {\n\t\t\tval := v.(string)\n\t\t\tms[k] = val\n\t\t}\n\t\tcreateMeta.Annotations = ms\n\t}\n\n\tif v, ok := d.GetOk(\"description\"); ok && !isIntfNil(v) {\n\t\tcreateMeta.Description =\n\t\t\tv.(string)\n\t}\n\n\tif v, ok := d.GetOk(\"disable\"); ok && !isIntfNil(v) {\n\t\tcreateMeta.Disable =\n\t\t\tv.(bool)\n\t}\n\n\tif v, ok := d.GetOk(\"labels\"); ok && !isIntfNil(v) {\n\n\t\tms := map[string]string{}\n\n\t\tfor k, v := range v.(map[string]interface{}) {\n\t\t\tval := v.(string)\n\t\t\tms[k] = val\n\t\t}\n\t\tcreateMeta.Labels = ms\n\t}\n\n\tif v, ok := d.GetOk(\"name\"); ok && !isIntfNil(v) {\n\t\tcreateMeta.Name =\n\t\t\tv.(string)\n\t}\n\n\tif v, ok := d.GetOk(\"namespace\"); ok && !isIntfNil(v) {\n\t\tcreateMeta.Namespace =\n\t\t\tv.(string)\n\t}\n\n\t//network_choice\n\n\tnetworkChoiceTypeFound := false\n\n\tif v, ok := d.GetOk(\"global_network\"); ok && !networkChoiceTypeFound {\n\n\t\tnetworkChoiceTypeFound = true\n\n\t\tif v.(bool) {\n\t\t\tnetworkChoiceInt := &ves_io_schema_virtual_network.CreateSpecType_GlobalNetwork{}\n\t\t\tnetworkChoiceInt.GlobalNetwork = &ves_io_schema.Empty{}\n\t\t\tcreateSpec.NetworkChoice = networkChoiceInt\n\t\t}\n\n\t}\n\n\tif v, ok := d.GetOk(\"legacy_type\"); ok && !networkChoiceTypeFound {\n\n\t\tnetworkChoiceTypeFound = true\n\t\tnetworkChoiceInt := &ves_io_schema_virtual_network.CreateSpecType_LegacyType{}\n\n\t\tcreateSpec.NetworkChoice = networkChoiceInt\n\n\t\tnetworkChoiceInt.LegacyType = ves_io_schema.VirtualNetworkType(ves_io_schema.VirtualNetworkType_value[v.(string)])\n\n\t}\n\n\tif v, ok := d.GetOk(\"site_local_inside_network\"); ok && !networkChoiceTypeFound {\n\n\t\tnetworkChoiceTypeFound = true\n\n\t\tif v.(bool) {\n\t\t\tnetworkChoiceInt := &ves_io_schema_virtual_network.CreateSpecType_SiteLocalInsideNetwork{}\n\t\t\tnetworkChoiceInt.SiteLocalInsideNetwork = &ves_io_schema.Empty{}\n\t\t\tcreateSpec.NetworkChoice = networkChoiceInt\n\t\t}\n\n\t}\n\n\tif v, ok := d.GetOk(\"site_local_network\"); ok && !networkChoiceTypeFound {\n\n\t\tnetworkChoiceTypeFound = true\n\n\t\tif v.(bool) {\n\t\t\tnetworkChoiceInt := &ves_io_schema_virtual_network.CreateSpecType_SiteLocalNetwork{}\n\t\t\tnetworkChoiceInt.SiteLocalNetwork = &ves_io_schema.Empty{}\n\t\t\tcreateSpec.NetworkChoice = networkChoiceInt\n\t\t}\n\n\t}\n\n\tif v, ok := d.GetOk(\"srv6_network\"); ok && !networkChoiceTypeFound {\n\n\t\tnetworkChoiceTypeFound = true\n\t\tnetworkChoiceInt := &ves_io_schema_virtual_network.CreateSpecType_Srv6Network{}\n\t\tnetworkChoiceInt.Srv6Network = &ves_io_schema_virtual_network.PerSiteSrv6NetworkType{}\n\t\tcreateSpec.NetworkChoice = networkChoiceInt\n\n\t\tsl := v.(*schema.Set).List()\n\t\tfor _, set := range sl {\n\t\t\tcs := set.(map[string]interface{})\n\n\t\t\tif v, ok := cs[\"access_network_rtargets\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tsl := v.([]interface{})\n\t\t\t\taccessNetworkRtargets := make([]*ves_io_schema.RouteTarget, len(sl))\n\t\t\t\tnetworkChoiceInt.Srv6Network.AccessNetworkRtargets = accessNetworkRtargets\n\t\t\t\tfor i, set := range sl {\n\t\t\t\t\taccessNetworkRtargets[i] = &ves_io_schema.RouteTarget{}\n\t\t\t\t\taccessNetworkRtargetsMapStrToI := set.(map[string]interface{})\n\n\t\t\t\t\trtargetChoiceTypeFound := false\n\n\t\t\t\t\tif v, ok := accessNetworkRtargetsMapStrToI[\"asn2byte_rtarget\"]; ok && !isIntfNil(v) && !rtargetChoiceTypeFound {\n\n\t\t\t\t\t\trtargetChoiceTypeFound = true\n\t\t\t\t\t\trtargetChoiceInt := &ves_io_schema.RouteTarget_Asn2ByteRtarget{}\n\t\t\t\t\t\trtargetChoiceInt.Asn2ByteRtarget = &ves_io_schema.RouteTarget2ByteAsn{}\n\t\t\t\t\t\taccessNetworkRtargets[i].RtargetChoice = rtargetChoiceInt\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t\t\tif v, ok := cs[\"as_number\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Asn2ByteRtarget.AsNumber = uint32(v.(int))\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tif v, ok := cs[\"value\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Asn2ByteRtarget.Value = uint32(v.(int))\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := accessNetworkRtargetsMapStrToI[\"asn4byte_rtarget\"]; ok && !isIntfNil(v) && !rtargetChoiceTypeFound {\n\n\t\t\t\t\t\trtargetChoiceTypeFound = true\n\t\t\t\t\t\trtargetChoiceInt := &ves_io_schema.RouteTarget_Asn4ByteRtarget{}\n\t\t\t\t\t\trtargetChoiceInt.Asn4ByteRtarget = &ves_io_schema.RouteTarget4ByteAsn{}\n\t\t\t\t\t\taccessNetworkRtargets[i].RtargetChoice = rtargetChoiceInt\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t\t\tif v, ok := cs[\"as_number\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Asn4ByteRtarget.AsNumber = uint32(v.(int))\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tif v, ok := cs[\"value\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Asn4ByteRtarget.Value = uint32(v.(int))\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := accessNetworkRtargetsMapStrToI[\"ipv4_addr_rtarget\"]; ok && !isIntfNil(v) && !rtargetChoiceTypeFound {\n\n\t\t\t\t\t\trtargetChoiceTypeFound = true\n\t\t\t\t\t\trtargetChoiceInt := &ves_io_schema.RouteTarget_Ipv4AddrRtarget{}\n\t\t\t\t\t\trtargetChoiceInt.Ipv4AddrRtarget = &ves_io_schema.RouteTargetIPv4Addr{}\n\t\t\t\t\t\taccessNetworkRtargets[i].RtargetChoice = rtargetChoiceInt\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t\t\tif v, ok := cs[\"address\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Ipv4AddrRtarget.Address = v.(string)\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tif v, ok := cs[\"value\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Ipv4AddrRtarget.Value = uint32(v.(int))\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tdefaultVipChoiceTypeFound := false\n\n\t\t\tif v, ok := cs[\"anycast_vip\"]; ok && !isIntfNil(v) && !defaultVipChoiceTypeFound {\n\n\t\t\t\tdefaultVipChoiceTypeFound = true\n\t\t\t\tdefaultVipChoiceInt := &ves_io_schema_virtual_network.PerSiteSrv6NetworkType_AnycastVip{}\n\n\t\t\t\tnetworkChoiceInt.Srv6Network.DefaultVipChoice = defaultVipChoiceInt\n\n\t\t\t\tdefaultVipChoiceInt.AnycastVip = v.(string)\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"fleet_vip\"]; ok && !isIntfNil(v) && !defaultVipChoiceTypeFound {\n\n\t\t\t\tdefaultVipChoiceTypeFound = true\n\t\t\t\tdefaultVipChoiceInt := &ves_io_schema_virtual_network.PerSiteSrv6NetworkType_FleetVip{}\n\t\t\t\tdefaultVipChoiceInt.FleetVip = &ves_io_schema_virtual_network.AnyCastVIPFleetType{}\n\t\t\t\tnetworkChoiceInt.Srv6Network.DefaultVipChoice = defaultVipChoiceInt\n\n\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\tfor _, set := range sl {\n\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\tif v, ok := cs[\"vip_allocator\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tvipAllocatorInt := &ves_io_schema_views.ObjectRefType{}\n\t\t\t\t\t\tdefaultVipChoiceInt.FleetVip.VipAllocator = vipAllocatorInt\n\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tvaMapToStrVal := set.(map[string]interface{})\n\t\t\t\t\t\t\tif val, ok := vaMapToStrVal[\"name\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\t\t\tvipAllocatorInt.Name = val.(string)\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\tif val, ok := vaMapToStrVal[\"namespace\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\t\t\tvipAllocatorInt.Namespace = val.(string)\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tif val, ok := vaMapToStrVal[\"tenant\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\t\t\tvipAllocatorInt.Tenant = val.(string)\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"interface_ip_vip\"]; ok && !isIntfNil(v) && !defaultVipChoiceTypeFound {\n\n\t\t\t\tdefaultVipChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\tdefaultVipChoiceInt := &ves_io_schema_virtual_network.PerSiteSrv6NetworkType_InterfaceIpVip{}\n\t\t\t\t\tdefaultVipChoiceInt.InterfaceIpVip = &ves_io_schema.Empty{}\n\t\t\t\t\tnetworkChoiceInt.Srv6Network.DefaultVipChoice = defaultVipChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"enterprise_network_rtargets\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tsl := v.([]interface{})\n\t\t\t\tenterpriseNetworkRtargets := make([]*ves_io_schema.RouteTarget, len(sl))\n\t\t\t\tnetworkChoiceInt.Srv6Network.EnterpriseNetworkRtargets = enterpriseNetworkRtargets\n\t\t\t\tfor i, set := range sl {\n\t\t\t\t\tenterpriseNetworkRtargets[i] = &ves_io_schema.RouteTarget{}\n\t\t\t\t\tenterpriseNetworkRtargetsMapStrToI := set.(map[string]interface{})\n\n\t\t\t\t\trtargetChoiceTypeFound := false\n\n\t\t\t\t\tif v, ok := enterpriseNetworkRtargetsMapStrToI[\"asn2byte_rtarget\"]; ok && !isIntfNil(v) && !rtargetChoiceTypeFound {\n\n\t\t\t\t\t\trtargetChoiceTypeFound = true\n\t\t\t\t\t\trtargetChoiceInt := &ves_io_schema.RouteTarget_Asn2ByteRtarget{}\n\t\t\t\t\t\trtargetChoiceInt.Asn2ByteRtarget = &ves_io_schema.RouteTarget2ByteAsn{}\n\t\t\t\t\t\tenterpriseNetworkRtargets[i].RtargetChoice = rtargetChoiceInt\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t\t\tif v, ok := cs[\"as_number\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Asn2ByteRtarget.AsNumber = uint32(v.(int))\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tif v, ok := cs[\"value\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Asn2ByteRtarget.Value = uint32(v.(int))\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := enterpriseNetworkRtargetsMapStrToI[\"asn4byte_rtarget\"]; ok && !isIntfNil(v) && !rtargetChoiceTypeFound {\n\n\t\t\t\t\t\trtargetChoiceTypeFound = true\n\t\t\t\t\t\trtargetChoiceInt := &ves_io_schema.RouteTarget_Asn4ByteRtarget{}\n\t\t\t\t\t\trtargetChoiceInt.Asn4ByteRtarget = &ves_io_schema.RouteTarget4ByteAsn{}\n\t\t\t\t\t\tenterpriseNetworkRtargets[i].RtargetChoice = rtargetChoiceInt\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t\t\tif v, ok := cs[\"as_number\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Asn4ByteRtarget.AsNumber = uint32(v.(int))\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tif v, ok := cs[\"value\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Asn4ByteRtarget.Value = uint32(v.(int))\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := enterpriseNetworkRtargetsMapStrToI[\"ipv4_addr_rtarget\"]; ok && !isIntfNil(v) && !rtargetChoiceTypeFound {\n\n\t\t\t\t\t\trtargetChoiceTypeFound = true\n\t\t\t\t\t\trtargetChoiceInt := &ves_io_schema.RouteTarget_Ipv4AddrRtarget{}\n\t\t\t\t\t\trtargetChoiceInt.Ipv4AddrRtarget = &ves_io_schema.RouteTargetIPv4Addr{}\n\t\t\t\t\t\tenterpriseNetworkRtargets[i].RtargetChoice = rtargetChoiceInt\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t\t\tif v, ok := cs[\"address\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Ipv4AddrRtarget.Address = v.(string)\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tif v, ok := cs[\"value\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Ipv4AddrRtarget.Value = uint32(v.(int))\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"export_rtargets\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tsl := v.([]interface{})\n\t\t\t\texportRtargets := make([]*ves_io_schema.RouteTarget, len(sl))\n\t\t\t\tnetworkChoiceInt.Srv6Network.ExportRtargets = exportRtargets\n\t\t\t\tfor i, set := range sl {\n\t\t\t\t\texportRtargets[i] = &ves_io_schema.RouteTarget{}\n\t\t\t\t\texportRtargetsMapStrToI := set.(map[string]interface{})\n\n\t\t\t\t\trtargetChoiceTypeFound := false\n\n\t\t\t\t\tif v, ok := exportRtargetsMapStrToI[\"asn2byte_rtarget\"]; ok && !isIntfNil(v) && !rtargetChoiceTypeFound {\n\n\t\t\t\t\t\trtargetChoiceTypeFound = true\n\t\t\t\t\t\trtargetChoiceInt := &ves_io_schema.RouteTarget_Asn2ByteRtarget{}\n\t\t\t\t\t\trtargetChoiceInt.Asn2ByteRtarget = &ves_io_schema.RouteTarget2ByteAsn{}\n\t\t\t\t\t\texportRtargets[i].RtargetChoice = rtargetChoiceInt\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t\t\tif v, ok := cs[\"as_number\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Asn2ByteRtarget.AsNumber = uint32(v.(int))\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tif v, ok := cs[\"value\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Asn2ByteRtarget.Value = uint32(v.(int))\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := exportRtargetsMapStrToI[\"asn4byte_rtarget\"]; ok && !isIntfNil(v) && !rtargetChoiceTypeFound {\n\n\t\t\t\t\t\trtargetChoiceTypeFound = true\n\t\t\t\t\t\trtargetChoiceInt := &ves_io_schema.RouteTarget_Asn4ByteRtarget{}\n\t\t\t\t\t\trtargetChoiceInt.Asn4ByteRtarget = &ves_io_schema.RouteTarget4ByteAsn{}\n\t\t\t\t\t\texportRtargets[i].RtargetChoice = rtargetChoiceInt\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t\t\tif v, ok := cs[\"as_number\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Asn4ByteRtarget.AsNumber = uint32(v.(int))\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tif v, ok := cs[\"value\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Asn4ByteRtarget.Value = uint32(v.(int))\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := exportRtargetsMapStrToI[\"ipv4_addr_rtarget\"]; ok && !isIntfNil(v) && !rtargetChoiceTypeFound {\n\n\t\t\t\t\t\trtargetChoiceTypeFound = true\n\t\t\t\t\t\trtargetChoiceInt := &ves_io_schema.RouteTarget_Ipv4AddrRtarget{}\n\t\t\t\t\t\trtargetChoiceInt.Ipv4AddrRtarget = &ves_io_schema.RouteTargetIPv4Addr{}\n\t\t\t\t\t\texportRtargets[i].RtargetChoice = rtargetChoiceInt\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t\t\tif v, ok := cs[\"address\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Ipv4AddrRtarget.Address = v.(string)\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tif v, ok := cs[\"value\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Ipv4AddrRtarget.Value = uint32(v.(int))\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"fleets\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tsl := v.([]interface{})\n\t\t\t\tfleetsInt := make([]*ves_io_schema_views.ObjectRefType, len(sl))\n\t\t\t\tnetworkChoiceInt.Srv6Network.Fleets = fleetsInt\n\t\t\t\tfor i, ps := range sl {\n\n\t\t\t\t\tfMapToStrVal := ps.(map[string]interface{})\n\t\t\t\t\tfleetsInt[i] = &ves_io_schema_views.ObjectRefType{}\n\n\t\t\t\t\tif v, ok := fMapToStrVal[\"name\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\tfleetsInt[i].Name = v.(string)\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := fMapToStrVal[\"namespace\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\tfleetsInt[i].Namespace = v.(string)\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := fMapToStrVal[\"tenant\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\tfleetsInt[i].Tenant = v.(string)\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"internet_rtargets\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tsl := v.([]interface{})\n\t\t\t\tinternetRtargets := make([]*ves_io_schema.RouteTarget, len(sl))\n\t\t\t\tnetworkChoiceInt.Srv6Network.InternetRtargets = internetRtargets\n\t\t\t\tfor i, set := range sl {\n\t\t\t\t\tinternetRtargets[i] = &ves_io_schema.RouteTarget{}\n\t\t\t\t\tinternetRtargetsMapStrToI := set.(map[string]interface{})\n\n\t\t\t\t\trtargetChoiceTypeFound := false\n\n\t\t\t\t\tif v, ok := internetRtargetsMapStrToI[\"asn2byte_rtarget\"]; ok && !isIntfNil(v) && !rtargetChoiceTypeFound {\n\n\t\t\t\t\t\trtargetChoiceTypeFound = true\n\t\t\t\t\t\trtargetChoiceInt := &ves_io_schema.RouteTarget_Asn2ByteRtarget{}\n\t\t\t\t\t\trtargetChoiceInt.Asn2ByteRtarget = &ves_io_schema.RouteTarget2ByteAsn{}\n\t\t\t\t\t\tinternetRtargets[i].RtargetChoice = rtargetChoiceInt\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t\t\tif v, ok := cs[\"as_number\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Asn2ByteRtarget.AsNumber = uint32(v.(int))\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tif v, ok := cs[\"value\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Asn2ByteRtarget.Value = uint32(v.(int))\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := internetRtargetsMapStrToI[\"asn4byte_rtarget\"]; ok && !isIntfNil(v) && !rtargetChoiceTypeFound {\n\n\t\t\t\t\t\trtargetChoiceTypeFound = true\n\t\t\t\t\t\trtargetChoiceInt := &ves_io_schema.RouteTarget_Asn4ByteRtarget{}\n\t\t\t\t\t\trtargetChoiceInt.Asn4ByteRtarget = &ves_io_schema.RouteTarget4ByteAsn{}\n\t\t\t\t\t\tinternetRtargets[i].RtargetChoice = rtargetChoiceInt\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t\t\tif v, ok := cs[\"as_number\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Asn4ByteRtarget.AsNumber = uint32(v.(int))\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tif v, ok := cs[\"value\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Asn4ByteRtarget.Value = uint32(v.(int))\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := internetRtargetsMapStrToI[\"ipv4_addr_rtarget\"]; ok && !isIntfNil(v) && !rtargetChoiceTypeFound {\n\n\t\t\t\t\t\trtargetChoiceTypeFound = true\n\t\t\t\t\t\trtargetChoiceInt := &ves_io_schema.RouteTarget_Ipv4AddrRtarget{}\n\t\t\t\t\t\trtargetChoiceInt.Ipv4AddrRtarget = &ves_io_schema.RouteTargetIPv4Addr{}\n\t\t\t\t\t\tinternetRtargets[i].RtargetChoice = rtargetChoiceInt\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t\t\tif v, ok := cs[\"address\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Ipv4AddrRtarget.Address = v.(string)\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tif v, ok := cs[\"value\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\trtargetChoiceInt.Ipv4AddrRtarget.Value = uint32(v.(int))\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tnamespaceChoiceTypeFound := false\n\n\t\t\tif v, ok := cs[\"no_namespace_network\"]; ok && !isIntfNil(v) && !namespaceChoiceTypeFound {\n\n\t\t\t\tnamespaceChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\tnamespaceChoiceInt := &ves_io_schema_virtual_network.PerSiteSrv6NetworkType_NoNamespaceNetwork{}\n\t\t\t\t\tnamespaceChoiceInt.NoNamespaceNetwork = &ves_io_schema.Empty{}\n\t\t\t\t\tnetworkChoiceInt.Srv6Network.NamespaceChoice = namespaceChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"srv6_network_ns_params\"]; ok && !isIntfNil(v) && !namespaceChoiceTypeFound {\n\n\t\t\t\tnamespaceChoiceTypeFound = true\n\t\t\t\tnamespaceChoiceInt := &ves_io_schema_virtual_network.PerSiteSrv6NetworkType_Srv6NetworkNsParams{}\n\t\t\t\tnamespaceChoiceInt.Srv6NetworkNsParams = &ves_io_schema_virtual_network.Srv6NetworkNsParametersType{}\n\t\t\t\tnetworkChoiceInt.Srv6Network.NamespaceChoice = namespaceChoiceInt\n\n\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\tfor _, set := range sl {\n\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\tif v, ok := cs[\"namespace\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tnamespaceChoiceInt.Srv6NetworkNsParams.Namespace = v.(string)\n\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"remote_sid_stats_plen\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tnetworkChoiceInt.Srv6Network.RemoteSidStatsPlen = uint32(v.(int))\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"slice\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\tsliceInt := &ves_io_schema_views.ObjectRefType{}\n\t\t\t\tnetworkChoiceInt.Srv6Network.Slice = sliceInt\n\n\t\t\t\tfor _, set := range sl {\n\t\t\t\t\tsMapToStrVal := set.(map[string]interface{})\n\t\t\t\t\tif val, ok := sMapToStrVal[\"name\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\tsliceInt.Name = val.(string)\n\t\t\t\t\t}\n\t\t\t\t\tif val, ok := sMapToStrVal[\"namespace\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\tsliceInt.Namespace = val.(string)\n\t\t\t\t\t}\n\n\t\t\t\t\tif val, ok := sMapToStrVal[\"tenant\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\tsliceInt.Tenant = val.(string)\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tsnatPoolChoiceTypeFound := false\n\n\t\t\tif v, ok := cs[\"fleet_snat_pool\"]; ok && !isIntfNil(v) && !snatPoolChoiceTypeFound {\n\n\t\t\t\tsnatPoolChoiceTypeFound = true\n\t\t\t\tsnatPoolChoiceInt := &ves_io_schema_virtual_network.PerSiteSrv6NetworkType_FleetSnatPool{}\n\t\t\t\tsnatPoolChoiceInt.FleetSnatPool = &ves_io_schema_virtual_network.SNATPoolFleetType{}\n\t\t\t\tnetworkChoiceInt.Srv6Network.SnatPoolChoice = snatPoolChoiceInt\n\n\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\tfor _, set := range sl {\n\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\tif v, ok := cs[\"snat_pool_allocator\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tsnatPoolAllocatorInt := &ves_io_schema_views.ObjectRefType{}\n\t\t\t\t\t\tsnatPoolChoiceInt.FleetSnatPool.SnatPoolAllocator = snatPoolAllocatorInt\n\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tspaMapToStrVal := set.(map[string]interface{})\n\t\t\t\t\t\t\tif val, ok := spaMapToStrVal[\"name\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\t\t\tsnatPoolAllocatorInt.Name = val.(string)\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\tif val, ok := spaMapToStrVal[\"namespace\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\t\t\tsnatPoolAllocatorInt.Namespace = val.(string)\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tif val, ok := spaMapToStrVal[\"tenant\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\t\t\tsnatPoolAllocatorInt.Tenant = val.(string)\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"interface_ip_snat_pool\"]; ok && !isIntfNil(v) && !snatPoolChoiceTypeFound {\n\n\t\t\t\tsnatPoolChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\tsnatPoolChoiceInt := &ves_io_schema_virtual_network.PerSiteSrv6NetworkType_InterfaceIpSnatPool{}\n\t\t\t\t\tsnatPoolChoiceInt.InterfaceIpSnatPool = &ves_io_schema.Empty{}\n\t\t\t\t\tnetworkChoiceInt.Srv6Network.SnatPoolChoice = snatPoolChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"site_snat_pool\"]; ok && !isIntfNil(v) && !snatPoolChoiceTypeFound {\n\n\t\t\t\tsnatPoolChoiceTypeFound = true\n\t\t\t\tsnatPoolChoiceInt := &ves_io_schema_virtual_network.PerSiteSrv6NetworkType_SiteSnatPool{}\n\t\t\t\tsnatPoolChoiceInt.SiteSnatPool = &ves_io_schema_virtual_network.SNATPoolSiteType{}\n\t\t\t\tnetworkChoiceInt.Srv6Network.SnatPoolChoice = snatPoolChoiceInt\n\n\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\tfor _, set := range sl {\n\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\tif v, ok := cs[\"node_snat_pool\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tnodeSnatPool := make(map[string]*ves_io_schema_virtual_network.SNATPoolType)\n\t\t\t\t\t\tsnatPoolChoiceInt.SiteSnatPool.NodeSnatPool = nodeSnatPool\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tnodeSnatPoolMapStrToI := set.(map[string]interface{})\n\t\t\t\t\t\t\tkey, ok := nodeSnatPoolMapStrToI[\"name\"]\n\t\t\t\t\t\t\tif ok && !isIntfNil(key) {\n\t\t\t\t\t\t\t\tnodeSnatPool[key.(string)] = &ves_io_schema_virtual_network.SNATPoolType{}\n\t\t\t\t\t\t\t\tval, _ := nodeSnatPoolMapStrToI[\"value\"]\n\n\t\t\t\t\t\t\t\tnodeSnatPoolVals := val.(*schema.Set).List()\n\t\t\t\t\t\t\t\tfor _, intVal := range nodeSnatPoolVals {\n\n\t\t\t\t\t\t\t\t\tnodeSnatPoolStaticMap := intVal.(map[string]interface{})\n\n\t\t\t\t\t\t\t\t\tif w, ok := nodeSnatPoolStaticMap[\"ipv4_prefixes\"]; ok && !isIntfNil(w) {\n\t\t\t\t\t\t\t\t\t\tnodeSnatPool[key.(string)].Ipv4Prefixes = nil\n\t\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t\t\tif w, ok := nodeSnatPoolStaticMap[\"ipv6_prefixes\"]; ok && !isIntfNil(w) {\n\t\t\t\t\t\t\t\t\t\tnodeSnatPool[key.(string)].Ipv6Prefixes = nil\n\t\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t\t\t// break after one loop\n\t\t\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t}\n\n\t}\n\n\t//static_routes\n\tif v, ok := d.GetOk(\"static_routes\"); ok && !isIntfNil(v) {\n\n\t\tsl := v.([]interface{})\n\t\tstaticRoutes := make([]*ves_io_schema_virtual_network.StaticRouteViewType, len(sl))\n\t\tcreateSpec.StaticRoutes = staticRoutes\n\t\tfor i, set := range sl {\n\t\t\tstaticRoutes[i] = &ves_io_schema_virtual_network.StaticRouteViewType{}\n\t\t\tstaticRoutesMapStrToI := set.(map[string]interface{})\n\n\t\t\tif v, ok := staticRoutesMapStrToI[\"attrs\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tattrsList := []ves_io_schema.RouteAttrType{}\n\t\t\t\tfor _, j := range v.([]interface{}) {\n\t\t\t\t\tattrsList = append(attrsList, ves_io_schema.RouteAttrType(ves_io_schema.RouteAttrType_value[j.(string)]))\n\t\t\t\t}\n\t\t\t\tstaticRoutes[i].Attrs = attrsList\n\n\t\t\t}\n\n\t\t\tif w, ok := staticRoutesMapStrToI[\"ip_prefixes\"]; ok && !isIntfNil(w) {\n\t\t\t\tls := make([]string, len(w.([]interface{})))\n\t\t\t\tfor i, v := range w.([]interface{}) {\n\t\t\t\t\tls[i] = v.(string)\n\t\t\t\t}\n\t\t\t\tstaticRoutes[i].IpPrefixes = ls\n\t\t\t}\n\n\t\t\tnextHopChoiceTypeFound := false\n\n\t\t\tif v, ok := staticRoutesMapStrToI[\"default_gateway\"]; ok && !isIntfNil(v) && !nextHopChoiceTypeFound {\n\n\t\t\t\tnextHopChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\tnextHopChoiceInt := &ves_io_schema_virtual_network.StaticRouteViewType_DefaultGateway{}\n\t\t\t\t\tnextHopChoiceInt.DefaultGateway = &ves_io_schema.Empty{}\n\t\t\t\t\tstaticRoutes[i].NextHopChoice = nextHopChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := staticRoutesMapStrToI[\"interface\"]; ok && !isIntfNil(v) && !nextHopChoiceTypeFound {\n\n\t\t\t\tnextHopChoiceTypeFound = true\n\t\t\t\tnextHopChoiceInt := &ves_io_schema_virtual_network.StaticRouteViewType_Interface{}\n\t\t\t\tnextHopChoiceInt.Interface = &ves_io_schema_views.ObjectRefType{}\n\t\t\t\tstaticRoutes[i].NextHopChoice = nextHopChoiceInt\n\n\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\tfor _, set := range sl {\n\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\tif v, ok := cs[\"name\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tnextHopChoiceInt.Interface.Name = v.(string)\n\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := cs[\"namespace\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tnextHopChoiceInt.Interface.Namespace = v.(string)\n\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := cs[\"tenant\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tnextHopChoiceInt.Interface.Tenant = v.(string)\n\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := staticRoutesMapStrToI[\"ip_address\"]; ok && !isIntfNil(v) && !nextHopChoiceTypeFound {\n\n\t\t\t\tnextHopChoiceTypeFound = true\n\t\t\t\tnextHopChoiceInt := &ves_io_schema_virtual_network.StaticRouteViewType_IpAddress{}\n\n\t\t\t\tstaticRoutes[i].NextHopChoice = nextHopChoiceInt\n\n\t\t\t\tnextHopChoiceInt.IpAddress = v.(string)\n\n\t\t\t}\n\n\t\t}\n\n\t}\n\n\t//static_v6_routes\n\tif v, ok := d.GetOk(\"static_v6_routes\"); ok && !isIntfNil(v) {\n\n\t\tsl := v.([]interface{})\n\t\tstaticV6Routes := make([]*ves_io_schema_virtual_network.StaticV6RouteViewType, len(sl))\n\t\tcreateSpec.StaticV6Routes = staticV6Routes\n\t\tfor i, set := range sl {\n\t\t\tstaticV6Routes[i] = &ves_io_schema_virtual_network.StaticV6RouteViewType{}\n\t\t\tstaticV6RoutesMapStrToI := set.(map[string]interface{})\n\n\t\t\tif v, ok := staticV6RoutesMapStrToI[\"attrs\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tattrsList := []ves_io_schema.RouteAttrType{}\n\t\t\t\tfor _, j := range v.([]interface{}) {\n\t\t\t\t\tattrsList = append(attrsList, ves_io_schema.RouteAttrType(ves_io_schema.RouteAttrType_value[j.(string)]))\n\t\t\t\t}\n\t\t\t\tstaticV6Routes[i].Attrs = attrsList\n\n\t\t\t}\n\n\t\t\tif w, ok := staticV6RoutesMapStrToI[\"ip_prefixes\"]; ok && !isIntfNil(w) {\n\t\t\t\tls := make([]string, len(w.([]interface{})))\n\t\t\t\tfor i, v := range w.([]interface{}) {\n\t\t\t\t\tls[i] = v.(string)\n\t\t\t\t}\n\t\t\t\tstaticV6Routes[i].IpPrefixes = ls\n\t\t\t}\n\n\t\t\tnextHopChoiceTypeFound := false\n\n\t\t\tif v, ok := staticV6RoutesMapStrToI[\"default_gateway\"]; ok && !isIntfNil(v) && !nextHopChoiceTypeFound {\n\n\t\t\t\tnextHopChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\tnextHopChoiceInt := &ves_io_schema_virtual_network.StaticV6RouteViewType_DefaultGateway{}\n\t\t\t\t\tnextHopChoiceInt.DefaultGateway = &ves_io_schema.Empty{}\n\t\t\t\t\tstaticV6Routes[i].NextHopChoice = nextHopChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := staticV6RoutesMapStrToI[\"interface\"]; ok && !isIntfNil(v) && !nextHopChoiceTypeFound {\n\n\t\t\t\tnextHopChoiceTypeFound = true\n\t\t\t\tnextHopChoiceInt := &ves_io_schema_virtual_network.StaticV6RouteViewType_Interface{}\n\t\t\t\tnextHopChoiceInt.Interface = &ves_io_schema_views.ObjectRefType{}\n\t\t\t\tstaticV6Routes[i].NextHopChoice = nextHopChoiceInt\n\n\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\tfor _, set := range sl {\n\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\tif v, ok := cs[\"name\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tnextHopChoiceInt.Interface.Name = v.(string)\n\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := cs[\"namespace\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tnextHopChoiceInt.Interface.Namespace = v.(string)\n\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := cs[\"tenant\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tnextHopChoiceInt.Interface.Tenant = v.(string)\n\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := staticV6RoutesMapStrToI[\"ip_address\"]; ok && !isIntfNil(v) && !nextHopChoiceTypeFound {\n\n\t\t\t\tnextHopChoiceTypeFound = true\n\t\t\t\tnextHopChoiceInt := &ves_io_schema_virtual_network.StaticV6RouteViewType_IpAddress{}\n\n\t\t\t\tstaticV6Routes[i].NextHopChoice = nextHopChoiceInt\n\n\t\t\t\tnextHopChoiceInt.IpAddress = v.(string)\n\n\t\t\t}\n\n\t\t}\n\n\t}\n\n\tlog.Printf(\"[DEBUG] Creating Volterra VirtualNetwork object with struct: %+v\", createReq)\n\n\tcreateVirtualNetworkResp, err := client.CreateObject(context.Background(), ves_io_schema_virtual_network.ObjectType, createReq)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error creating VirtualNetwork: %s\", err)\n\t}\n\td.SetId(createVirtualNetworkResp.GetObjSystemMetadata().GetUid())\n\n\treturn resourceVolterraVirtualNetworkRead(d, meta)\n}", "func (c *TestClient) GetSubnetwork(project, region, name string) (*compute.Subnetwork, error) {\n\tif c.GetSubnetworkFn != nil {\n\t\treturn c.GetSubnetworkFn(project, region, name)\n\t}\n\treturn c.client.GetSubnetwork(project, region, name)\n}", "func (instance *Network) AdoptSubnet(ctx context.Context, subnet resources.Subnet) (xerr fail.Error) {\n\tdefer fail.OnPanic(&xerr)\n\n\tif instance == nil || instance.IsNull() {\n\t\treturn fail.InvalidInstanceError()\n\t}\n\tif ctx == nil {\n\t\treturn fail.InvalidParameterCannotBeNilError(\"ctx\")\n\t}\n\tif subnet == nil {\n\t\treturn fail.InvalidParameterCannotBeNilError(\"subnet\")\n\t}\n\n\ttask, xerr := concurrency.TaskFromContext(ctx)\n\txerr = debug.InjectPlannedFail(xerr)\n\tif xerr != nil {\n\t\treturn xerr\n\t}\n\n\tif task.Aborted() {\n\t\treturn fail.AbortedError(nil, \"aborted\")\n\t}\n\n\tinstance.lock.Lock()\n\tdefer instance.lock.Unlock()\n\n\tparentNetwork, xerr := subnet.InspectNetwork()\n\txerr = debug.InjectPlannedFail(xerr)\n\tif xerr != nil {\n\t\treturn xerr\n\t}\n\tdefer parentNetwork.Released()\n\n\tif parentNetwork.GetName() != instance.GetName() {\n\t\treturn fail.InvalidRequestError(\"cannot adopt Subnet '%s' because Network '%s' does not own it\", subnet.GetName(), instance.GetName())\n\t}\n\n\treturn instance.Alter(func(_ data.Clonable, props *serialize.JSONProperties) fail.Error {\n\t\treturn props.Alter(networkproperty.SubnetsV1, func(clonable data.Clonable) fail.Error {\n\t\t\tnsV1, ok := clonable.(*propertiesv1.NetworkSubnets)\n\t\t\tif !ok {\n\t\t\t\treturn fail.InconsistentError(\"'*propertiesv1.NetworkSubnets' expected, '%s' provided\", reflect.TypeOf(clonable).String())\n\t\t\t}\n\n\t\t\tname := subnet.GetName()\n\t\t\tid := subnet.GetID()\n\t\t\tnsV1.ByID[id] = name\n\t\t\tnsV1.ByName[name] = id\n\t\t\treturn nil\n\t\t})\n\t})\n}", "func (vcd *TestVCD) testCreateOrgVdcNetworkRouted(check *C, ipSubnet string, subInterface, distributed bool) {\n\tfmt.Printf(\"Running: %s\\n\", check.TestName())\n\tnetworkName := TestCreateOrgVdcNetworkRouted\n\n\tgateway := ipSubnet + \".1\"\n\tstartAddress := ipSubnet + \".2\"\n\tendAddress := ipSubnet + \".50\"\n\n\tif subInterface {\n\t\tnetworkName += \"-sub\"\n\t}\n\terr := RemoveOrgVdcNetworkIfExists(*vcd.vdc, networkName)\n\tif err != nil {\n\t\tcheck.Skip(fmt.Sprintf(\"Error deleting network : %s\", err))\n\t}\n\n\tedgeGWName := vcd.config.VCD.EdgeGateway\n\tif edgeGWName == \"\" {\n\t\tcheck.Skip(\"Edge Gateway not provided\")\n\t}\n\tedgeGateway, err := vcd.vdc.GetEdgeGatewayByName(edgeGWName, false)\n\tif err != nil {\n\t\tcheck.Skip(fmt.Sprintf(\"Edge Gateway %s not found\", edgeGWName))\n\t}\n\n\tnetworkDescription := \"Created by govcd tests\"\n\tvar networkConfig = types.OrgVDCNetwork{\n\t\tXmlns: types.XMLNamespaceVCloud,\n\t\tName: networkName,\n\t\tDescription: networkDescription,\n\t\tConfiguration: &types.NetworkConfiguration{\n\t\t\tFenceMode: types.FenceModeNAT,\n\t\t\tIPScopes: &types.IPScopes{\n\t\t\t\tIPScope: []*types.IPScope{&types.IPScope{\n\t\t\t\t\tIsInherited: false,\n\t\t\t\t\tGateway: gateway,\n\t\t\t\t\tNetmask: \"255.255.255.0\",\n\t\t\t\t\tIPRanges: &types.IPRanges{\n\t\t\t\t\t\tIPRange: []*types.IPRange{\n\t\t\t\t\t\t\t&types.IPRange{\n\t\t\t\t\t\t\t\tStartAddress: startAddress,\n\t\t\t\t\t\t\t\tEndAddress: endAddress,\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t\tBackwardCompatibilityMode: true,\n\t\t},\n\t\tEdgeGateway: &types.Reference{\n\t\t\tHREF: edgeGateway.EdgeGateway.HREF,\n\t\t\tID: edgeGateway.EdgeGateway.ID,\n\t\t\tName: edgeGateway.EdgeGateway.Name,\n\t\t\tType: edgeGateway.EdgeGateway.Type,\n\t\t},\n\t\tIsShared: false,\n\t}\n\tif subInterface && distributed {\n\t\tcheck.Skip(\"A network can't be at the same time distributed and subInterface\")\n\t}\n\tif subInterface {\n\t\tnetworkConfig.Configuration.SubInterface = &subInterface\n\t}\n\n\tif distributed {\n\t\tdistributedRoutingEnabled := edgeGateway.EdgeGateway.Configuration.DistributedRoutingEnabled\n\t\tif distributedRoutingEnabled != nil && *distributedRoutingEnabled {\n\t\t\tnetworkConfig.Configuration.DistributedInterface = &distributed\n\t\t} else {\n\t\t\tcheck.Skip(fmt.Sprintf(\"edge gateway %s doesn't have distributed routing enabled\", edgeGWName))\n\t\t}\n\t}\n\n\tLogNetwork(networkConfig)\n\terr = vcd.vdc.CreateOrgVDCNetworkWait(&networkConfig)\n\tif err != nil {\n\t\tfmt.Printf(\"error creating Network <%s>: %s\\n\", networkName, err)\n\t}\n\tcheck.Assert(err, IsNil)\n\tAddToCleanupList(networkName,\n\t\t\"network\",\n\t\tvcd.org.Org.Name+\"|\"+vcd.vdc.Vdc.Name,\n\t\t\"Test_CreateOrgVdcNetworkRouted\")\n\tnetwork, err := vcd.vdc.GetOrgVdcNetworkByName(networkName, true)\n\tcheck.Assert(err, IsNil)\n\tcheck.Assert(network, NotNil)\n\tcheck.Assert(network.OrgVDCNetwork.Name, Equals, networkName)\n\tcheck.Assert(network.OrgVDCNetwork.Description, Equals, networkDescription)\n\tif subInterface {\n\t\tcheck.Assert(network.OrgVDCNetwork.Configuration.SubInterface, NotNil)\n\t\tcheck.Assert(*network.OrgVDCNetwork.Configuration.SubInterface, Equals, true)\n\t}\n\n\t// Tests FindEdgeGatewayNameByNetwork\n\t// Note: is should work without refreshing either VDC or edge gateway\n\tconnectedGw, err := vcd.vdc.FindEdgeGatewayNameByNetwork(networkName)\n\tcheck.Assert(err, IsNil)\n\tcheck.Assert(connectedGw, Equals, edgeGWName)\n\n\ttask, err := network.Delete()\n\tcheck.Assert(err, IsNil)\n\terr = task.WaitTaskCompletion()\n\tcheck.Assert(err, IsNil)\n}", "func (c *TestClient) DeleteSubnetwork(project, region, name string) error {\n\tif c.DeleteSubnetworkFn != nil {\n\t\treturn c.DeleteSubnetworkFn(project, region, name)\n\t}\n\treturn c.client.DeleteSubnetwork(project, region, name)\n}", "func CreateRouterInterfaceOnSubnet(t *testing.T, client *gophercloud.ServiceClient, subnetID, routerID string) (*routers.InterfaceInfo, error) {\n\tt.Logf(\"Attempting to add subnet %s to router %s\", subnetID, routerID)\n\n\taiOpts := routers.AddInterfaceOpts{\n\t\tSubnetID: subnetID,\n\t}\n\n\tiface, err := routers.AddInterface(client, routerID, aiOpts).Extract()\n\tif err != nil {\n\t\treturn iface, err\n\t}\n\n\tif err := WaitForRouterInterfaceToAttach(client, iface.PortID); err != nil {\n\t\treturn iface, err\n\t}\n\n\tt.Logf(\"Successfully added subnet %s to router %s\", subnetID, routerID)\n\treturn iface, nil\n}", "func (instance *Network) Create(ctx context.Context, req abstract.NetworkRequest) (xerr fail.Error) {\n\tdefer fail.OnPanic(&xerr)\n\n\tif instance == nil || instance.IsNull() {\n\t\treturn fail.InvalidInstanceError()\n\t}\n\tif ctx == nil {\n\t\treturn fail.InvalidParameterCannotBeNilError(\"ctx\")\n\t}\n\n\ttask, xerr := concurrency.TaskFromContext(ctx)\n\txerr = debug.InjectPlannedFail(xerr)\n\tif xerr != nil {\n\t\treturn xerr\n\t}\n\n\tif task.Aborted() {\n\t\treturn fail.AbortedError(nil, \"aborted\")\n\t}\n\n\ttracer := debug.NewTracer(task, true, \"('%s', '%s')\", req.Name, req.CIDR).WithStopwatch().Entering()\n\tdefer tracer.Exiting()\n\n\tinstance.lock.Lock()\n\tdefer instance.lock.Unlock()\n\n\t// Check if subnet already exists and is managed by SafeScale\n\tsvc := instance.GetService()\n\tif existing, xerr := LoadNetwork(svc, req.Name); xerr == nil {\n\t\texisting.Released()\n\t\treturn fail.DuplicateError(\"Network '%s' already exists\", req.Name)\n\t}\n\n\tif task.Aborted() {\n\t\treturn fail.AbortedError(nil, \"aborted\")\n\t}\n\n\t// Verify if the subnet already exist and in this case is not managed by SafeScale\n\t_, xerr = svc.InspectNetworkByName(req.Name)\n\txerr = debug.InjectPlannedFail(xerr)\n\tif xerr != nil {\n\t\tswitch xerr.(type) {\n\t\tcase *fail.ErrNotFound:\n\t\t\t// continue\n\t\tdefault:\n\t\t\treturn xerr\n\t\t}\n\t} else {\n\t\treturn fail.DuplicateError(\"Network '%s' already exists (not managed by SafeScale)\", req.Name)\n\t}\n\n\t// Verify the CIDR is not routable\n\tif req.CIDR != \"\" {\n\t\troutable, xerr := netretry.IsCIDRRoutable(req.CIDR)\n\t\txerr = debug.InjectPlannedFail(xerr)\n\t\tif xerr != nil {\n\t\t\treturn fail.Wrap(xerr, \"failed to determine if CIDR is not routable\")\n\t\t}\n\n\t\tif routable {\n\t\t\treturn fail.InvalidRequestError(\"cannot create such a Networking, CIDR must not be routable; please choose an appropriate CIDR (RFC1918)\")\n\t\t}\n\t}\n\n\tif task.Aborted() {\n\t\treturn fail.AbortedError(nil, \"aborted\")\n\t}\n\n\t// Create the Network\n\tlogrus.Debugf(\"Creating Network '%s' with CIDR '%s'...\", req.Name, req.CIDR)\n\tan, xerr := svc.CreateNetwork(req)\n\txerr = debug.InjectPlannedFail(xerr)\n\tif xerr != nil {\n\t\treturn xerr\n\t}\n\n\tdefer func() {\n\t\tif xerr != nil && !req.KeepOnFailure {\n\t\t\tderr := svc.DeleteNetwork(an.ID)\n\t\t\tderr = debug.InjectPlannedFail(derr)\n\t\t\tif derr != nil {\n\t\t\t\t_ = xerr.AddConsequence(fail.Wrap(derr, \"cleaning up on failure, failed to delete Network\"))\n\t\t\t}\n\t\t}\n\t}()\n\n\tif task.Aborted() {\n\t\treturn fail.AbortedError(nil, \"aborted\")\n\t}\n\n\t// Write subnet object metadata\n\t// logrus.Debugf(\"Saving subnet metadata '%s' ...\", subnet.GetName)\n\treturn instance.carry(an)\n}", "func ProtoToSubnetwork(p *computepb.ComputeSubnetwork) *compute.Subnetwork {\n\tobj := &compute.Subnetwork{\n\t\tCreationTimestamp: dcl.StringOrNil(p.GetCreationTimestamp()),\n\t\tDescription: dcl.StringOrNil(p.GetDescription()),\n\t\tGatewayAddress: dcl.StringOrNil(p.GetGatewayAddress()),\n\t\tIPCidrRange: dcl.StringOrNil(p.GetIpCidrRange()),\n\t\tName: dcl.StringOrNil(p.GetName()),\n\t\tNetwork: dcl.StringOrNil(p.GetNetwork()),\n\t\tFingerprint: dcl.StringOrNil(p.GetFingerprint()),\n\t\tPurpose: ProtoToComputeSubnetworkPurposeEnum(p.GetPurpose()),\n\t\tRole: ProtoToComputeSubnetworkRoleEnum(p.GetRole()),\n\t\tPrivateIPGoogleAccess: dcl.Bool(p.GetPrivateIpGoogleAccess()),\n\t\tRegion: dcl.StringOrNil(p.GetRegion()),\n\t\tLogConfig: ProtoToComputeSubnetworkLogConfig(p.GetLogConfig()),\n\t\tProject: dcl.StringOrNil(p.GetProject()),\n\t\tSelfLink: dcl.StringOrNil(p.GetSelfLink()),\n\t\tEnableFlowLogs: dcl.Bool(p.GetEnableFlowLogs()),\n\t}\n\tfor _, r := range p.GetSecondaryIpRanges() {\n\t\tobj.SecondaryIPRanges = append(obj.SecondaryIPRanges, *ProtoToComputeSubnetworkSecondaryIPRanges(r))\n\t}\n\treturn obj\n}", "func (ggSession *GreengrassSession) CreateSub(source, target, subject string) error {\n\tsourceArn := ggSession.mapSubToArn(source)\n\ttargetArn := ggSession.mapSubToArn(target)\n\n\tnewUUID, err := uuid.NewV4()\n\tif err != nil {\n\t\treturn err\n\t}\n\tuuidString := newUUID.String()\n\n\t// Check if we need to create the initial version\n\tif ggSession.config.SubscriptionDefinition.ID == \"\" {\n\t\tnewSubscription, err := ggSession.greengrass.CreateSubscriptionDefinition(&greengrass.CreateSubscriptionDefinitionInput{\n\t\t\tInitialVersion: &greengrass.SubscriptionDefinitionVersion{\n\t\t\t\tSubscriptions: []*greengrass.Subscription{\n\t\t\t\t\t&greengrass.Subscription{\n\t\t\t\t\t\tSource: &sourceArn,\n\t\t\t\t\t\tTarget: &targetArn,\n\t\t\t\t\t\tSubject: &subject,\n\t\t\t\t\t\tId: &uuidString,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t})\n\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfmt.Printf(\"Created new subscription\\n\")\n\t\tggSession.config.SubscriptionDefinition.ID = *newSubscription.Id\n\t\tggSession.config.SubscriptionDefinition.VersionArn = *newSubscription.LatestVersionArn\n\n\t\tggSession.updateGroup()\n\n\t\treturn nil\n\t}\n\n\t// Add subscription to existing\n\tsubscription, _ := ggSession.greengrass.GetSubscriptionDefinition(&greengrass.GetSubscriptionDefinitionInput{\n\t\tSubscriptionDefinitionId: &ggSession.config.SubscriptionDefinition.ID,\n\t})\n\n\tsubscriptionVersion, _ := ggSession.greengrass.GetSubscriptionDefinitionVersion(&greengrass.GetSubscriptionDefinitionVersionInput{\n\t\tSubscriptionDefinitionId: subscription.Id,\n\t\tSubscriptionDefinitionVersionId: subscription.LatestVersion,\n\t})\n\tsubscriptions := subscriptionVersion.Definition.Subscriptions\n\n\tsubscriptions = append(subscriptions, &greengrass.Subscription{\n\t\tSource: &sourceArn,\n\t\tTarget: &targetArn,\n\t\tSubject: &subject,\n\t\tId: &uuidString,\n\t})\n\n\toutput, err := ggSession.greengrass.CreateSubscriptionDefinitionVersion(&greengrass.CreateSubscriptionDefinitionVersionInput{\n\t\tSubscriptionDefinitionId: subscription.Id,\n\t\tSubscriptions: subscriptions,\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tggSession.config.SubscriptionDefinition.VersionArn = *output.Arn\n\tfmt.Printf(\"Updated subscription\\n\")\n\n\tggSession.updateGroup()\n\n\treturn nil\n}", "func SubnetworkToProto(resource *compute.Subnetwork) *computepb.ComputeSubnetwork {\n\tp := &computepb.ComputeSubnetwork{}\n\tp.SetCreationTimestamp(dcl.ValueOrEmptyString(resource.CreationTimestamp))\n\tp.SetDescription(dcl.ValueOrEmptyString(resource.Description))\n\tp.SetGatewayAddress(dcl.ValueOrEmptyString(resource.GatewayAddress))\n\tp.SetIpCidrRange(dcl.ValueOrEmptyString(resource.IPCidrRange))\n\tp.SetName(dcl.ValueOrEmptyString(resource.Name))\n\tp.SetNetwork(dcl.ValueOrEmptyString(resource.Network))\n\tp.SetFingerprint(dcl.ValueOrEmptyString(resource.Fingerprint))\n\tp.SetPurpose(ComputeSubnetworkPurposeEnumToProto(resource.Purpose))\n\tp.SetRole(ComputeSubnetworkRoleEnumToProto(resource.Role))\n\tp.SetPrivateIpGoogleAccess(dcl.ValueOrEmptyBool(resource.PrivateIPGoogleAccess))\n\tp.SetRegion(dcl.ValueOrEmptyString(resource.Region))\n\tp.SetLogConfig(ComputeSubnetworkLogConfigToProto(resource.LogConfig))\n\tp.SetProject(dcl.ValueOrEmptyString(resource.Project))\n\tp.SetSelfLink(dcl.ValueOrEmptyString(resource.SelfLink))\n\tp.SetEnableFlowLogs(dcl.ValueOrEmptyBool(resource.EnableFlowLogs))\n\tsSecondaryIPRanges := make([]*computepb.ComputeSubnetworkSecondaryIPRanges, len(resource.SecondaryIPRanges))\n\tfor i, r := range resource.SecondaryIPRanges {\n\t\tsSecondaryIPRanges[i] = ComputeSubnetworkSecondaryIPRangesToProto(&r)\n\t}\n\tp.SetSecondaryIpRanges(sSecondaryIPRanges)\n\n\treturn p\n}", "func createHnsNetwork(backend string, networkAdapter string) (string, error) {\n\tvar network hcsshim.HNSNetwork\n\tif backend == \"vxlan\" {\n\t\t// Ignoring the return because both true and false without an error represent that the firewall rule was created or already exists\n\t\tif _, err := wapi.FirewallRuleAdd(\"OverlayTraffic4789UDP\", \"Overlay network traffic UDP\", \"\", \"4789\", wapi.NET_FW_IP_PROTOCOL_UDP, wapi.NET_FW_PROFILE2_ALL); err != nil {\n\t\t\treturn \"\", fmt.Errorf(\"error creating firewall rules: %v\", err)\n\t\t}\n\t\tlogrus.Infof(\"Creating VXLAN network using the vxlanAdapter: %s\", networkAdapter)\n\t\tnetwork = hcsshim.HNSNetwork{\n\t\t\tType: \"Overlay\",\n\t\t\tName: CalicoHnsNetworkName,\n\t\t\tNetworkAdapterName: networkAdapter,\n\t\t\tSubnets: []hcsshim.Subnet{\n\t\t\t\t{\n\t\t\t\t\tAddressPrefix: \"192.168.255.0/30\",\n\t\t\t\t\tGatewayAddress: \"192.168.255.1\",\n\t\t\t\t\tPolicies: []json.RawMessage{\n\t\t\t\t\t\t[]byte(\"{ \\\"Type\\\": \\\"VSID\\\", \\\"VSID\\\": 9999 }\"),\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t}\n\t} else {\n\t\tnetwork = hcsshim.HNSNetwork{\n\t\t\tType: \"L2Bridge\",\n\t\t\tName: CalicoHnsNetworkName,\n\t\t\tNetworkAdapterName: networkAdapter,\n\t\t\tSubnets: []hcsshim.Subnet{\n\t\t\t\t{\n\t\t\t\t\tAddressPrefix: \"192.168.255.0/30\",\n\t\t\t\t\tGatewayAddress: \"192.168.255.1\",\n\t\t\t\t},\n\t\t\t},\n\t\t}\n\t}\n\n\tif _, err := network.Create(); err != nil {\n\t\treturn \"\", fmt.Errorf(\"error creating the %s network: %v\", CalicoHnsNetworkName, err)\n\t}\n\n\t// Check if network exists. If it does not after 5 minutes, fail\n\tfor start := time.Now(); time.Since(start) < 5*time.Minute; {\n\t\tnetwork, err := hcsshim.GetHNSNetworkByName(CalicoHnsNetworkName)\n\t\tif err == nil {\n\t\t\treturn network.ManagementIP, nil\n\t\t}\n\t}\n\n\treturn \"\", fmt.Errorf(\"failed to create %s network\", CalicoHnsNetworkName)\n}", "func NewSubnet(ctx *pulumi.Context,\n\tname string, args *SubnetArgs, opts ...pulumi.ResourceOption) (*Subnet, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.VpcId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'VpcId'\")\n\t}\n\tvar resource Subnet\n\terr := ctx.RegisterResource(\"aws:ec2/subnet:Subnet\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (s *PublicTransactionPoolAPI) CreateSubAccount(ctx context.Context, maxSub hexutil.Uint64, addr *common.Address) (bool, error) {\n\terr := s.wallet.CreateSubAccount(uint64(maxSub), addr)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\treturn true, nil\n}", "func TestNewSubnet(t *testing.T) {\n\ttype args struct {\n\t\tanIp string\n\t\tnetmask string\n\t}\n\ttests := []struct {\n\t\tname string\n\t\targs args\n\t\twant *Subnet\n\t\twantErr bool\n\t}{\n\t\t{\n\t\t\tname: \"test subnet 192.168\",\n\t\t\targs: args{anIp: \"192.168.100.4\", netmask: \"255.255.255.0\"},\n\t\t\twant: &Subnet{start: 3232261120, current: 3232261120, Len: 255},\n\t\t\twantErr: false,\n\t\t},\n\t\t{\n\t\t\tname: \"test subnet 172.31 with /20 cidr\",\n\t\t\targs: args{anIp: \"172.31.17.69\", netmask: \"255.255.240.0\"},\n\t\t\twant: &Subnet{start: 2887716864, current: 2887716864, Len: 4095},\n\t\t\twantErr: false,\n\t\t},\n\t}\n\tfor _, tt := range tests {\n\t\tt.Run(tt.name, func(t *testing.T) {\n\t\t\tgot, err := NewSubnet(tt.args.anIp, tt.args.netmask)\n\t\t\tif (err != nil) != tt.wantErr {\n\t\t\t\tt.Errorf(\"NewSubnet() error = %v, wantErr %v\", err, tt.wantErr)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif !reflect.DeepEqual(got, tt.want) {\n\t\t\t\tt.Errorf(\"NewSubnet() = %v, want %v\", got, tt.want)\n\t\t\t}\n\t\t})\n\t}\n}", "func CreateNIC(stack *Stack, id string, subnetVlan uint16) (netlink.Link, error) {\n\tif subnetVlan > 4096 {\n\t\treturn nil, fmt.Errorf(\"subnet out of range\")\n\t}\n\n\tnic := &netlink.Macvtap{\n\t\tMacvlan: netlink.Macvlan{\n\t\t\tLinkAttrs: netlink.LinkAttrs{\n\t\t\t\tName: fmt.Sprintf(nicPattern, id),\n\t\t\t\tMTU: 1500,\n\t\t\t\tParentIndex: stack.Bridge.Index,\n\t\t\t},\n\t\t\tMode: netlink.MACVLAN_MODE_BRIDGE,\n\t\t},\n\t}\n\n\tif err := netlink.LinkAdd(nic); err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to add tap device: %s\", err)\n\t}\n\n\t// if err := netlink.BridgeVlanAdd(nic, subnetVlan, false, true, false, false); err != nil {\n\t// \treturn nil, fmt.Errorf(\"Failed to add VLAN to nic: %s\", err)\n\t// }\n\n\tif err := netlink.LinkSetUp(nic); err != nil {\n\t\treturn nic, fmt.Errorf(\"failed to set nic to up: %s\", err)\n\t}\n\n\tstack.Nics[id] = &VNic{id: id, vlan: subnetVlan, link: nic}\n\n\treturn nic, nil\n}", "func CreateSubnet(ctx context.Context, log logr.Logger, awsClient *awsclient.Client, vpcID string, cidr string, name string) (string, error) {\n\toutput, err := awsClient.EC2.CreateSubnet(&ec2.CreateSubnetInput{\n\t\tCidrBlock: awssdk.String(cidr),\n\t\tVpcId: awssdk.String(vpcID),\n\t\tTagSpecifications: []*ec2.TagSpecification{\n\t\t\t{\n\t\t\t\tResourceType: awssdk.String(\"subnet\"),\n\t\t\t\tTags: []*ec2.Tag{\n\t\t\t\t\t{\n\t\t\t\t\t\tKey: awssdk.String(\"Name\"),\n\t\t\t\t\t\tValue: awssdk.String(name),\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t})\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tsubnetID := output.Subnet.SubnetId\n\n\tif err := wait.PollUntil(5*time.Second, func() (bool, error) {\n\t\tlog.Info(\"Waiting until subnet is available...\", \"subnetID\", *subnetID)\n\n\t\toutput, err := awsClient.EC2.DescribeSubnets(&ec2.DescribeSubnetsInput{\n\t\t\tSubnetIds: []*string{subnetID},\n\t\t})\n\t\tif err != nil {\n\t\t\treturn false, err\n\t\t}\n\n\t\tsubnet := output.Subnets[0]\n\t\tif *subnet.State != \"available\" {\n\t\t\treturn false, nil\n\t\t}\n\n\t\treturn true, nil\n\t}, ctx.Done()); err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn *subnetID, nil\n}", "func (a *Azure) CreateVirtualNetworkSubnet(ctx context.Context, vnetName, subnetName string) (subnet *network.Subnet, err error) {\n\tsubnetsClient, err := a.getSubnetsClient()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfuture, err := subnetsClient.CreateOrUpdate(\n\t\tctx,\n\t\ta.groupName,\n\t\tvnetName,\n\t\tsubnetName,\n\t\tnetwork.Subnet{\n\t\t\tSubnetPropertiesFormat: &network.SubnetPropertiesFormat{\n\t\t\t\tAddressPrefix: to.StringPtr(\"10.0.0.0/16\"),\n\t\t\t},\n\t\t})\n\tif err != nil {\n\t\treturn subnet, fmt.Errorf(\"cannot create subnet: %v\", err)\n\t}\n\n\terr = future.WaitForCompletionRef(ctx, subnetsClient.Client)\n\tif err != nil {\n\t\treturn subnet, fmt.Errorf(\"cannot get the subnet create or update future response: %v\", err)\n\t}\n\n\tsc, err := future.Result(*subnetsClient)\n\n\treturn &sc, err\n}", "func DeleteVirtualNetworkSubnet() {}", "func DeleteVirtualNetworkSubnet() {}", "func createSubscriber(ctx context.Context, m manipulate.Manipulator, namespace string) manipulate.Subscriber {\n\tfilter := elemental.NewPushFilter()\n\tfilter.FilterIdentity(gaia.NetworkAccessPolicyIdentity.Name)\n\tfilter.FilterIdentity(gaia.ExternalNetworkIdentity.Name)\n\n\ts := maniphttp.NewSubscriber(\n\t\tm,\n\t\tmaniphttp.SubscriberOptionRecursive(true), // get events from children as well.\n\t\tmaniphttp.SubscriberOptionNamespace(namespace), //start at namespace\n\t)\n\ts.Start(ctx, filter)\n\n\treturn s\n}", "func (s *SubnetworkServer) applySubnetwork(ctx context.Context, c *compute.Client, request *computepb.ApplyComputeSubnetworkRequest) (*computepb.ComputeSubnetwork, error) {\n\tp := ProtoToSubnetwork(request.GetResource())\n\tres, err := c.ApplySubnetwork(ctx, p)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tr := SubnetworkToProto(res)\n\treturn r, nil\n}", "func (o LookupWorkstationClusterResultOutput) Subnetwork() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupWorkstationClusterResult) string { return v.Subnetwork }).(pulumi.StringOutput)\n}", "func networkCreateExample() string {\n\treturn `$ pouch network create -n pouchnet -d bridge --gateway 192.168.1.1 --subnet 192.168.1.0/24\npouchnet: e1d541722d68dc5d133cca9e7bd8fd9338603e1763096c8e853522b60d11f7b9`\n}", "func New(svc iaas.Service, terraform bool) (resources.Subnet, fail.Error) {\n\tif terraform {\n\t\treturn operations.NewTerraformSubnet(svc)\n\t}\n\treturn operations.NewSubnet(svc)\n}", "func CreateSubnetWithNetowrkSecurityGroup(ctx context.Context, vnetName, subnetName, addressPrefix, nsgName string) (subnet network.Subnet, err error) {\n\tnsg, err := GetNetworkSecurityGroup(ctx, nsgName)\n\tif err != nil {\n\t\treturn subnet, fmt.Errorf(\"cannot get nsg: %v\", err)\n\t}\n\n\tsubnetsClient := getSubnetsClient()\n\tfuture, err := subnetsClient.CreateOrUpdate(\n\t\tctx,\n\t\tconfig.GroupName(),\n\t\tvnetName,\n\t\tsubnetName,\n\t\tnetwork.Subnet{\n\t\t\tSubnetPropertiesFormat: &network.SubnetPropertiesFormat{\n\t\t\t\tAddressPrefix: to.StringPtr(addressPrefix),\n\t\t\t\tNetworkSecurityGroup: &nsg,\n\t\t\t},\n\t\t})\n\tif err != nil {\n\t\treturn subnet, fmt.Errorf(\"cannot create subnet: %v\", err)\n\t}\n\n\terr = future.WaitForCompletion(ctx, subnetsClient.Client)\n\tif err != nil {\n\t\treturn subnet, fmt.Errorf(\"cannot get the subnet create or update future response: %v\", err)\n\t}\n\n\treturn future.Result(subnetsClient)\n}", "func SetupExistingVNet(ctx context.Context, vnetCidr string, cpSubnetCidrs, nodeSubnetCidrs map[string]string, bastionSubnetName, bastionSubnetCidr string) func() {\n\tBy(\"creating Azure clients with the workload cluster's subscription\")\n\tsettings, err := auth.GetSettingsFromEnvironment()\n\tExpect(err).NotTo(HaveOccurred())\n\tsubscriptionID := settings.GetSubscriptionID()\n\tauthorizer, err := settings.GetAuthorizer()\n\tExpect(err).NotTo(HaveOccurred())\n\tgroupClient := resources.NewGroupsClient(subscriptionID)\n\tgroupClient.Authorizer = authorizer\n\tvnetClient := network.NewVirtualNetworksClient(subscriptionID)\n\tvnetClient.Authorizer = authorizer\n\tnsgClient := network.NewSecurityGroupsClient(subscriptionID)\n\tnsgClient.Authorizer = authorizer\n\troutetableClient := network.NewRouteTablesClient(subscriptionID)\n\troutetableClient.Authorizer = authorizer\n\n\tBy(\"creating a resource group\")\n\tgroupName := os.Getenv(AzureResourceGroup)\n\t_, err = groupClient.CreateOrUpdate(ctx, groupName, resources.Group{\n\t\tLocation: pointer.StringPtr(os.Getenv(AzureLocation)),\n\t\tTags: map[string]*string{\n\t\t\t\"jobName\": pointer.StringPtr(os.Getenv(JobName)),\n\t\t\t\"creationTimestamp\": pointer.StringPtr(os.Getenv(Timestamp)),\n\t\t},\n\t})\n\tExpect(err).To(BeNil())\n\n\tBy(\"creating a network security group\")\n\tnsgName := \"control-plane-nsg\"\n\tsecurityRules := []network.SecurityRule{\n\t\t{\n\t\t\tName: pointer.StringPtr(\"allow_ssh\"),\n\t\t\tSecurityRulePropertiesFormat: &network.SecurityRulePropertiesFormat{\n\t\t\t\tDescription: pointer.StringPtr(\"Allow SSH\"),\n\t\t\t\tPriority: pointer.Int32Ptr(2200),\n\t\t\t\tProtocol: network.SecurityRuleProtocolTCP,\n\t\t\t\tAccess: network.SecurityRuleAccessAllow,\n\t\t\t\tDirection: network.SecurityRuleDirectionInbound,\n\t\t\t\tSourceAddressPrefix: pointer.StringPtr(\"*\"),\n\t\t\t\tSourcePortRange: pointer.StringPtr(\"*\"),\n\t\t\t\tDestinationAddressPrefix: pointer.StringPtr(\"*\"),\n\t\t\t\tDestinationPortRange: pointer.StringPtr(\"22\"),\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: pointer.StringPtr(\"allow_apiserver\"),\n\t\t\tSecurityRulePropertiesFormat: &network.SecurityRulePropertiesFormat{\n\t\t\t\tDescription: pointer.StringPtr(\"Allow API Server\"),\n\t\t\t\tSourcePortRange: pointer.StringPtr(\"*\"),\n\t\t\t\tDestinationPortRange: pointer.StringPtr(\"6443\"),\n\t\t\t\tSourceAddressPrefix: pointer.StringPtr(\"*\"),\n\t\t\t\tDestinationAddressPrefix: pointer.StringPtr(\"*\"),\n\t\t\t\tProtocol: network.SecurityRuleProtocolTCP,\n\t\t\t\tAccess: network.SecurityRuleAccessAllow,\n\t\t\t\tDirection: network.SecurityRuleDirectionInbound,\n\t\t\t\tPriority: pointer.Int32Ptr(2201),\n\t\t\t},\n\t\t},\n\t}\n\tnsgFuture, err := nsgClient.CreateOrUpdate(ctx, groupName, nsgName, network.SecurityGroup{\n\t\tLocation: pointer.StringPtr(os.Getenv(AzureLocation)),\n\t\tSecurityGroupPropertiesFormat: &network.SecurityGroupPropertiesFormat{\n\t\t\tSecurityRules: &securityRules,\n\t\t},\n\t})\n\tExpect(err).To(BeNil())\n\terr = nsgFuture.WaitForCompletionRef(ctx, nsgClient.Client)\n\tExpect(err).To(BeNil())\n\n\tBy(\"creating a node security group\")\n\tnsgNodeName := \"node-nsg\"\n\tsecurityRulesNode := []network.SecurityRule{}\n\tnsgNodeFuture, err := nsgClient.CreateOrUpdate(ctx, groupName, nsgNodeName, network.SecurityGroup{\n\t\tLocation: pointer.StringPtr(os.Getenv(AzureLocation)),\n\t\tSecurityGroupPropertiesFormat: &network.SecurityGroupPropertiesFormat{\n\t\t\tSecurityRules: &securityRulesNode,\n\t\t},\n\t})\n\tExpect(err).To(BeNil())\n\terr = nsgNodeFuture.WaitForCompletionRef(ctx, nsgClient.Client)\n\tExpect(err).To(BeNil())\n\n\tBy(\"creating a node routetable\")\n\trouteTableName := \"node-routetable\"\n\trouteTable := network.RouteTable{\n\t\tLocation: pointer.StringPtr(os.Getenv(AzureLocation)),\n\t\tRouteTablePropertiesFormat: &network.RouteTablePropertiesFormat{},\n\t}\n\troutetableFuture, err := routetableClient.CreateOrUpdate(ctx, groupName, routeTableName, routeTable)\n\tExpect(err).To(BeNil())\n\terr = routetableFuture.WaitForCompletionRef(ctx, routetableClient.Client)\n\tExpect(err).To(BeNil())\n\n\tBy(\"creating a virtual network\")\n\tvar subnets []network.Subnet\n\tfor name, cidr := range cpSubnetCidrs {\n\t\tsubnets = append(subnets, network.Subnet{\n\t\t\tSubnetPropertiesFormat: &network.SubnetPropertiesFormat{\n\t\t\t\tAddressPrefix: pointer.StringPtr(cidr),\n\t\t\t\tNetworkSecurityGroup: &network.SecurityGroup{\n\t\t\t\t\tID: pointer.StringPtr(fmt.Sprintf(\"/subscriptions/%s/resourceGroups/%s/providers/Microsoft.Network/networkSecurityGroups/%s\", subscriptionID, groupName, nsgName)),\n\t\t\t\t},\n\t\t\t},\n\t\t\tName: pointer.StringPtr(name),\n\t\t})\n\t}\n\tfor name, cidr := range nodeSubnetCidrs {\n\t\tsubnets = append(subnets, network.Subnet{\n\t\t\tSubnetPropertiesFormat: &network.SubnetPropertiesFormat{\n\t\t\t\tAddressPrefix: pointer.StringPtr(cidr),\n\t\t\t\tNetworkSecurityGroup: &network.SecurityGroup{\n\t\t\t\t\tID: pointer.StringPtr(fmt.Sprintf(\"/subscriptions/%s/resourceGroups/%s/providers/Microsoft.Network/networkSecurityGroups/%s\", subscriptionID, groupName, nsgNodeName)),\n\t\t\t\t},\n\t\t\t\tRouteTable: &network.RouteTable{\n\t\t\t\t\tID: pointer.StringPtr(fmt.Sprintf(\"/subscriptions/%s/resourceGroups/%s/providers/Microsoft.Network/routeTables/%s\", subscriptionID, groupName, routeTableName)),\n\t\t\t\t},\n\t\t\t},\n\t\t\tName: pointer.StringPtr(name),\n\t\t})\n\t}\n\n\t// Create the AzureBastion subnet.\n\tsubnets = append(subnets, network.Subnet{\n\t\tSubnetPropertiesFormat: &network.SubnetPropertiesFormat{\n\t\t\tAddressPrefix: pointer.StringPtr(bastionSubnetCidr),\n\t\t},\n\t\tName: pointer.StringPtr(bastionSubnetName),\n\t})\n\n\tvnetFuture, err := vnetClient.CreateOrUpdate(ctx, groupName, os.Getenv(AzureCustomVNetName), network.VirtualNetwork{\n\t\tLocation: pointer.StringPtr(os.Getenv(AzureLocation)),\n\t\tVirtualNetworkPropertiesFormat: &network.VirtualNetworkPropertiesFormat{\n\t\t\tAddressSpace: &network.AddressSpace{\n\t\t\t\tAddressPrefixes: &[]string{vnetCidr},\n\t\t\t},\n\t\t\tSubnets: &subnets,\n\t\t},\n\t})\n\tif err != nil {\n\t\tfmt.Print(err.Error())\n\t}\n\tExpect(err).To(BeNil())\n\terr = vnetFuture.WaitForCompletionRef(ctx, vnetClient.Client)\n\tExpect(err).To(BeNil())\n\n\treturn func() {\n\t\tLogf(\"deleting an existing virtual network %q\", os.Getenv(AzureCustomVNetName))\n\t\tvFuture, err := vnetClient.Delete(ctx, groupName, os.Getenv(AzureCustomVNetName))\n\t\tExpect(err).NotTo(HaveOccurred())\n\t\tExpect(vFuture.WaitForCompletionRef(ctx, vnetClient.Client)).To(Succeed())\n\n\t\tLogf(\"deleting an existing route table %q\", routeTableName)\n\t\trtFuture, err := routetableClient.Delete(ctx, groupName, routeTableName)\n\t\tExpect(err).NotTo(HaveOccurred())\n\t\tExpect(rtFuture.WaitForCompletionRef(ctx, routetableClient.Client)).To(Succeed())\n\n\t\tLogf(\"deleting an existing network security group %q\", nsgNodeName)\n\t\tnsgFuture, err := nsgClient.Delete(ctx, groupName, nsgNodeName)\n\t\tExpect(err).NotTo(HaveOccurred())\n\t\tExpect(nsgFuture.WaitForCompletionRef(ctx, nsgClient.Client)).To(Succeed())\n\n\t\tLogf(\"deleting an existing network security group %q\", nsgName)\n\t\tnsgFuture, err = nsgClient.Delete(ctx, groupName, nsgName)\n\t\tExpect(err).NotTo(HaveOccurred())\n\t\tExpect(nsgFuture.WaitForCompletionRef(ctx, nsgClient.Client)).To(Succeed())\n\n\t\tLogf(\"verifying the existing resource group %q is empty\", groupName)\n\t\tresClient := resources.NewClient(subscriptionID)\n\t\tresClient.Authorizer = authorizer\n\t\tEventually(func() ([]resources.GenericResourceExpanded, error) {\n\t\t\tpage, err := resClient.ListByResourceGroup(ctx, groupName, \"\", \"provisioningState\", to.Int32Ptr(10))\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\n\t\t\t// for each resource do a GET directly for that resource to avoid hitting Azure list cache\n\t\t\tvar foundResources []resources.GenericResourceExpanded\n\t\t\tfor _, genericResource := range page.Values() {\n\t\t\t\tapiversion, err := getAPIVersion(*genericResource.ID)\n\t\t\t\tif err != nil {\n\t\t\t\t\tLogWarningf(\"failed to get API version for %q with %+v\", *genericResource.ID, err)\n\t\t\t\t}\n\n\t\t\t\t_, err = resClient.GetByID(ctx, *genericResource.ID, apiversion)\n\t\t\t\tif err != nil && azure.ResourceNotFound(err) {\n\t\t\t\t\t// the resources is returned in the list, but it's actually 404\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\t// unexpected error calling GET on the resource\n\t\t\t\tif err != nil {\n\t\t\t\t\tLogWarningf(\"failed GETing resource %q with %+v\", *genericResource.ID, err)\n\t\t\t\t\treturn nil, err\n\t\t\t\t}\n\n\t\t\t\t// if resource is still there, then append to foundResources\n\t\t\t\tfoundResources = append(foundResources, genericResource)\n\t\t\t}\n\t\t\treturn foundResources, nil\n\t\t\t// add some tolerance for Azure caching of resource group resource caching\n\t\t}, deleteOperationTimeout, retryableOperationTimeout).Should(BeEmpty(), \"Expect the manually created resource group is empty after removing the manually created resources.\")\n\n\t\tLogf(\"deleting the existing resource group %q\", groupName)\n\t\tgrpFuture, err := groupClient.Delete(ctx, groupName)\n\t\tExpect(err).NotTo(HaveOccurred())\n\t\tExpect(grpFuture.WaitForCompletionRef(ctx, nsgClient.Client)).To(Succeed())\n\t}\n}", "func (f *FakeInstance) CreateReverseIPv6(_ context.Context, _ string, _ *govultr.ReverseIP) error {\n\tpanic(\"implement me\")\n}", "func TestNetworkCreateOnNonExistentTenant(t *testing.T) {\n\t// create netagent\n\tag, _, _ := createNetAgent(t)\n\tAssert(t, ag != nil, \"Failed to create agent %#v\", ag)\n\tdefer ag.Stop()\n\n\t// create network\n\tnt := netproto.Network{\n\t\tTypeMeta: api.TypeMeta{Kind: \"Network\"},\n\t\tObjectMeta: api.ObjectMeta{\n\t\t\tTenant: \"nonExistentNetwork\",\n\t\t\tNamespace: \"nonExistentNetwork\",\n\t\t\tName: \"default\",\n\t\t},\n\t}\n\n\t// create network\n\terr := ag.CreateNetwork(&nt)\n\tAssert(t, err != nil, \"Network create was expected to fail.\")\n}", "func CreateNIC(ctx context.Context, vnetName, subnetName, nsgName, ipName, nicName string) (nic network.Interface, err error) {\n\tsubnet, err := GetVirtualNetworkSubnet(ctx, vnetName, subnetName)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to get subnet: %v\", err)\n\t}\n\n\tip, err := GetPublicIP(ctx, ipName)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to get ip address: %v\", err)\n\t}\n\n\tnicParams := network.Interface{\n\t\tName: to.StringPtr(nicName),\n\t\tLocation: to.StringPtr(config.Location()),\n\t\tInterfacePropertiesFormat: &network.InterfacePropertiesFormat{\n\t\t\tIPConfigurations: &[]network.InterfaceIPConfiguration{\n\t\t\t\t{\n\t\t\t\t\tName: to.StringPtr(\"ipConfig1\"),\n\t\t\t\t\tInterfaceIPConfigurationPropertiesFormat: &network.InterfaceIPConfigurationPropertiesFormat{\n\t\t\t\t\t\tSubnet: &subnet,\n\t\t\t\t\t\tPrivateIPAllocationMethod: network.Dynamic,\n\t\t\t\t\t\tPublicIPAddress: &ip,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\n\tif nsgName != \"\" {\n\t\tnsg, err := GetNetworkSecurityGroup(ctx, nsgName)\n\t\tif err != nil {\n\t\t\tlog.Fatalf(\"failed to get nsg: %v\", err)\n\t\t}\n\t\tnicParams.NetworkSecurityGroup = &nsg\n\t}\n\n\tnicClient := getNicClient()\n\tfuture, err := nicClient.CreateOrUpdate(ctx, config.GroupName(), nicName, nicParams)\n\tif err != nil {\n\t\treturn nic, fmt.Errorf(\"cannot create nic: %v\", err)\n\t}\n\n\terr = future.WaitForCompletionRef(ctx, nicClient.Client)\n\tif err != nil {\n\t\treturn nic, fmt.Errorf(\"cannot get nic create or update future response: %v\", err)\n\t}\n\n\treturn future.Result(nicClient)\n}", "func (s *Stream) CreateSubStream(headers http.Header, fin bool) (*Stream, error) {\n\treturn s.conn.CreateStream(headers, s, fin)\n}", "func CreateNIC(ctx context.Context, vnetName, subnetName, nsgName, ipName, nicName string) (nic network.Interface, err error) {\n\tsubnet, err := GetVirtualNetworkSubnet(ctx, vnetName, subnetName)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to get subnet: %v\", err)\n\t}\n\n\tip, err := GetPublicIP(ctx, ipName)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to get ip address: %v\", err)\n\t}\n\n\tnicParams := network.Interface{\n\t\tName: to.StringPtr(nicName),\n\t\tLocation: to.StringPtr(config.Location()),\n\t\tInterfacePropertiesFormat: &network.InterfacePropertiesFormat{\n\t\t\tIPConfigurations: &[]network.InterfaceIPConfiguration{\n\t\t\t\t{\n\t\t\t\t\tName: to.StringPtr(\"ipConfig1\"),\n\t\t\t\t\tInterfaceIPConfigurationPropertiesFormat: &network.InterfaceIPConfigurationPropertiesFormat{\n\t\t\t\t\t\tSubnet: &subnet,\n\t\t\t\t\t\tPrivateIPAllocationMethod: network.Dynamic,\n\t\t\t\t\t\tPublicIPAddress: &ip,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\n\tif nsgName != \"\" {\n\t\tnsg, err := GetNetworkSecurityGroup(ctx, nsgName)\n\t\tif err != nil {\n\t\t\tlog.Fatalf(\"failed to get nsg: %v\", err)\n\t\t}\n\t\tnicParams.NetworkSecurityGroup = &nsg\n\t}\n\n\tnicClient := getNicClient()\n\tfuture, err := nicClient.CreateOrUpdate(ctx, config.GroupName(), nicName, nicParams)\n\tif err != nil {\n\t\treturn nic, fmt.Errorf(\"cannot create nic: %v\", err)\n\t}\n\n\terr = future.WaitForCompletion(ctx, nicClient.Client)\n\tif err != nil {\n\t\treturn nic, fmt.Errorf(\"cannot get nic create or update future response: %v\", err)\n\t}\n\n\treturn future.Result(nicClient)\n}", "func (o ClusterOutput) Subnetwork() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Cluster) pulumi.StringOutput { return v.Subnetwork }).(pulumi.StringOutput)\n}", "func CreateVNet(vNetName string, rgroupName string) {\n\ta := wow.New(os.Stdout, spin.Get(spin.Dots), \"Creating virtual network : \"+vNetName)\n\ta.Start()\n\ttime.Sleep(2 * time.Second)\n\ta.Text(\"This would take a few minutes...\").Spinner(spin.Get(spin.Dots))\n\tcmd := exec.Command(\"az\", \"network\", \"vnet\", \"create\", \"--name\",\n\t\tvNetName, \"--resource-group\", rgroupName)\n\tvar out bytes.Buffer\n\tvar stderr bytes.Buffer\n\tcmd.Stdout = &out\n\tcmd.Stderr = &stderr\n\terr := cmd.Run()\n\tif err != nil {\n\t\tfmt.Println(fmt.Sprint(err) + \": \" + stderr.String())\n\t\treturn\n\t}\n\ta.PersistWith(spin.Spinner{}, \"....\")\n\tfmt.Println(\"Result: \" + out.String())\n\temoji.Println(\":beer: Cheers!!!\")\n}", "func CreateVirtualNetworkSubnet(ctx context.Context, vnetName, subnetName string) (subnet network.Subnet, err error) {\n\tsubnetsClient := getSubnetsClient()\n\n\tfuture, err := subnetsClient.CreateOrUpdate(\n\t\tctx,\n\t\tconfig.GroupName(),\n\t\tvnetName,\n\t\tsubnetName,\n\t\tnetwork.Subnet{\n\t\t\tSubnetPropertiesFormat: &network.SubnetPropertiesFormat{\n\t\t\t\tAddressPrefix: to.StringPtr(\"10.0.0.0/16\"),\n\t\t\t},\n\t\t})\n\tif err != nil {\n\t\treturn subnet, fmt.Errorf(\"cannot create subnet: %v\", err)\n\t}\n\n\terr = future.WaitForCompletionRef(ctx, subnetsClient.Client)\n\tif err != nil {\n\t\treturn subnet, fmt.Errorf(\"cannot get the subnet create or update future response: %v\", err)\n\t}\n\n\treturn future.Result(subnetsClient)\n}", "func CreateNetwork(ctx context.Context) (*tc.Network, string) {\n\trandomUuid, _ := uuid.NewRandom()\n\tnetworkName := randomUuid.String()\n\tnet, err := tc.GenericNetwork(ctx, tc.GenericNetworkRequest{\n\t\tNetworkRequest: tc.NetworkRequest{\n\t\t\tName: networkName,\n\t\t\tCheckDuplicate: true,\n\t\t},\n\t})\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn &net, networkName\n}", "func CreateVirtualNetworkSubnet(ctx context.Context, vnetName, subnetName string) (subnet network.Subnet, err error) {\n\tsubnetsClient := getSubnetsClient()\n\n\tfuture, err := subnetsClient.CreateOrUpdate(\n\t\tctx,\n\t\tconfig.GroupName(),\n\t\tvnetName,\n\t\tsubnetName,\n\t\tnetwork.Subnet{\n\t\t\tSubnetPropertiesFormat: &network.SubnetPropertiesFormat{\n\t\t\t\tAddressPrefix: to.StringPtr(\"10.0.0.0/16\"),\n\t\t\t},\n\t\t})\n\tif err != nil {\n\t\treturn subnet, fmt.Errorf(\"cannot create subnet: %v\", err)\n\t}\n\n\terr = future.WaitForCompletion(ctx, subnetsClient.Client)\n\tif err != nil {\n\t\treturn subnet, fmt.Errorf(\"cannot get the subnet create or update future response: %v\", err)\n\t}\n\n\treturn future.Result(subnetsClient)\n}", "func (s *Service) CreateNetwork() error {\n\t// Create docker client.\n\tvar (\n\t\tcli *client.Client\n\n\t\terr error\n\t\terrChanPeer = make(chan error, 1)\n\t\terrChanOrderer = make(chan error, 1)\n\n\t\twg sync.WaitGroup\n\t\twgPeerDone = make(chan bool) // todo change this to ctx.Done()\n\t\twgOrdererDone = make(chan bool) // todo change this to ctx.Done()\n\t)\n\n\tif cli, err = NewClient(); err != nil {\n\t\treturn errors.Wrap(err, \"NewClient failed with error\")\n\t}\n\ts.MyClient = cli\n\n\tctx := context.TODO()\n\trespNet, err := cli.NetworkCreate(ctx, \"giou_net\", types.NetworkCreate{})\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"NetworkCreate failed with error\")\n\t}\n\tlog.Println(\"Network has been created wth ID: \", respNet.ID)\n\n\t// In need of absolute path to bind/mount host:container paths.\n\tprojectPath, err := filepath.Abs(\"./\")\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"Failed to get project's path with error\")\n\t}\n\n\twg.Add(1)\n\n\tgo func(orgs []config.Organization) {\n\t\tdefer wg.Done()\n\n\t\tfor _, org := range orgs {\n\t\t\tfor i := range org.Peers {\n\n\t\t\t\tgo s.RunPeer(org.Name, org.Peers, projectPath, i, errChanPeer, wgPeerDone)\n\n\t\t\t\tselect {\n\t\t\t\tcase <-wgPeerDone:\n\t\t\t\t\tlog.Println(\"carry on...\")\n\n\t\t\t\t\tbreak\n\t\t\t\tcase err := <-errChanPeer:\n\t\t\t\t\tclose(errChanPeer)\n\t\t\t\t\tlog.Fatal(\"Error: \", err)\n\t\t\t\t\t//return err\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}(s.Cfg.Orgs[1:])\n\n\twg.Add(1)\n\n\tgo func(org config.Organization) {\n\t\tdefer wg.Done()\n\n\t\tfor i := range org.Orderers {\n\n\t\t\tgo s.RunOrderer(org.Orderers, projectPath, i, errChanOrderer, wgOrdererDone)\n\n\t\t\tselect {\n\t\t\tcase <-wgOrdererDone:\n\t\t\t\t// carry on\n\t\t\t\tlog.Println(\"carry on orderer...\")\n\n\t\t\t\tbreak\n\t\t\tcase err := <-errChanOrderer:\n\t\t\t\tclose(errChanOrderer)\n\t\t\t\tlog.Fatal(\"Error Orderer: \", err)\n\t\t\t\t//return err\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}(s.Cfg.Orgs[0])\n\n\t// waits until containers are up and running\n\twg.Wait()\n\n\treturn s.List()\n}", "func CreateVirtualNetworkAndSubnets(ctx context.Context, vnetName, subnet1Name, subnet2Name string) (vnet network.VirtualNetwork, err error) {\n\tvnetClient := getVnetClient()\n\tfuture, err := vnetClient.CreateOrUpdate(\n\t\tctx,\n\t\tconfig.GroupName(),\n\t\tvnetName,\n\t\tnetwork.VirtualNetwork{\n\t\t\tLocation: to.StringPtr(config.Location()),\n\t\t\tVirtualNetworkPropertiesFormat: &network.VirtualNetworkPropertiesFormat{\n\t\t\t\tAddressSpace: &network.AddressSpace{\n\t\t\t\t\tAddressPrefixes: &[]string{\"10.0.0.0/8\"},\n\t\t\t\t},\n\t\t\t\tSubnets: &[]network.Subnet{\n\t\t\t\t\t{\n\t\t\t\t\t\tName: to.StringPtr(subnet1Name),\n\t\t\t\t\t\tSubnetPropertiesFormat: &network.SubnetPropertiesFormat{\n\t\t\t\t\t\t\tAddressPrefix: to.StringPtr(\"10.0.0.0/16\"),\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t\t{\n\t\t\t\t\t\tName: to.StringPtr(subnet2Name),\n\t\t\t\t\t\tSubnetPropertiesFormat: &network.SubnetPropertiesFormat{\n\t\t\t\t\t\t\tAddressPrefix: to.StringPtr(\"10.1.0.0/16\"),\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t})\n\n\tif err != nil {\n\t\treturn vnet, fmt.Errorf(\"cannot create virtual network: %v\", err)\n\t}\n\n\terr = future.WaitForCompletion(ctx, vnetClient.Client)\n\tif err != nil {\n\t\treturn vnet, fmt.Errorf(\"cannot get the vnet create or update future response: %v\", err)\n\t}\n\n\treturn future.Result(vnetClient)\n}", "func (az *Cloud) CreateOrUpdateSubnet(service *v1.Service, subnet network.Subnet) error {\n\tctx, cancel := getContextWithCancel()\n\tdefer cancel()\n\n\tvar rg string\n\tif len(az.VnetResourceGroup) > 0 {\n\t\trg = az.VnetResourceGroup\n\t} else {\n\t\trg = az.ResourceGroup\n\t}\n\n\trerr := az.SubnetsClient.CreateOrUpdate(ctx, rg, az.VnetName, *subnet.Name, subnet)\n\tklog.V(10).Infof(\"SubnetClient.CreateOrUpdate(%s): end\", *subnet.Name)\n\tif rerr != nil {\n\t\tklog.Errorf(\"SubnetClient.CreateOrUpdate(%s) failed: %s\", *subnet.Name, rerr.Error().Error())\n\t\taz.Event(service, v1.EventTypeWarning, \"CreateOrUpdateSubnet\", rerr.Error().Error())\n\t\treturn rerr.Error()\n\t}\n\n\treturn nil\n}", "func CreateNetwork(cli *docker.Client, config *NetworkConfig) (*docker.Network, error) {\r\n\t// Setting network\r\n\tvar netoworkOptions = docker.CreateNetworkOptions{\r\n\t\tCheckDuplicate: true,\r\n\t\tDriver: \"bridge\",\r\n\t\tName: config.NetoworkName,\r\n\t\tIPAM: docker.IPAMOptions{\r\n\t\t\tConfig: []docker.IPAMConfig{docker.IPAMConfig{\r\n\t\t\t\tSubnet: config.SubnetMast,\r\n\t\t\t\tIPRange: config.IPRange,\r\n\t\t\t\tGateway: config.Gateway,\r\n\t\t\t},\r\n\t\t\t},\r\n\t\t},\r\n\t}\r\n\t// Create network\r\n\treturn cli.CreateNetwork(netoworkOptions)\r\n}", "func (c *MockSubnetsClient) CreateOrUpdate(ctx context.Context, resourceGroupName, virtualNetworkName, subnetName string, parameters network.Subnet) (*network.Subnet, error) {\n\t// Ignore resourceGroupName and virtualNetworkName for simplicity.\n\tif _, ok := c.Subnets[subnetName]; ok {\n\t\treturn nil, fmt.Errorf(\"update not supported\")\n\t}\n\tparameters.Name = &subnetName\n\tc.Subnets[subnetName] = parameters\n\treturn &parameters, nil\n}", "func (d *Driver) createNetwork() error {\n\tconn, err := getConnection()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"getting libvirt connection\")\n\t}\n\tdefer conn.Close()\n\n\t// network: default\n\t// It is assumed that the libvirt/kvm installation has already created this network\n\n\t// network: private\n\n\t// Only create the private network if it does not already exist\n\tif _, err := conn.LookupNetworkByName(d.PrivateNetwork); err != nil {\n\t\t// create the XML for the private network from our networkTmpl\n\t\ttmpl := template.Must(template.New(\"network\").Parse(networkTmpl))\n\t\tvar networkXML bytes.Buffer\n\t\tif err := tmpl.Execute(&networkXML, d); err != nil {\n\t\t\treturn errors.Wrap(err, \"executing network template\")\n\t\t}\n\n\t\t// define the network using our template\n\t\tnetwork, err := conn.NetworkDefineXML(networkXML.String())\n\t\tif err != nil {\n\t\t\treturn errors.Wrapf(err, \"defining network from xml: %s\", networkXML.String())\n\t\t}\n\n\t\t// and finally create it\n\t\tif err := network.Create(); err != nil {\n\t\t\treturn errors.Wrapf(err, \"creating network %s\", d.PrivateNetwork)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (de *DockerEngine) getSubNet() (string, error) {\n\tde.subNetMu.Lock()\n\tdefer de.subNetMu.Unlock()\n\n\taddrs, err := net.InterfaceAddrs()\n\n\tif err != nil {\n\t\treturn \"\", errors.Wrap(err, \"Error getting network addresses\")\n\t}\n\n\tvar nets []*net.IPNet\n\n\tfor _, addr := range addrs {\n\t\tdockerLog.Debugf(\"Inspecting interface %s\", addr.String())\n\n\t\t_, n, err := net.ParseCIDR(addr.String())\n\n\t\tif err != nil {\n\t\t\tdockerLog.Warningf(\"Error parsing address: %s\", addr.String())\n\n\t\t\tcontinue\n\t\t}\n\n\t\tnets = append(nets, n)\n\t}\n\n\tnetaddr := func() string {\n\t\ttpl := \"10.%d.%d.0/24\"\n\n\t\treturn fmt.Sprintf(tpl, de.subNetOct1, de.subNetOct2)\n\t}\n\n\t_, pnet, _ := net.ParseCIDR(netaddr())\n\n\tfor {\n\t\t// Find non-overlapping network\n\t\toverlap := false\n\n\t\tfor _, n := range nets {\n\t\t\tif lib.NetsOverlap(pnet, n) {\n\t\t\t\toverlap = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\n\t\tif overlap {\n\t\t\tde.subNetOct2 += 1\n\n\t\t\tif de.subNetOct2 > 255 {\n\t\t\t\tde.subNetOct1 += 1\n\t\t\t\tde.subNetOct2 = 0\n\t\t\t}\n\n\t\t\t_, pnet, _ = net.ParseCIDR(netaddr())\n\t\t} else {\n\t\t\tbreak\n\t\t}\n\t}\n\n\treturn netaddr(), nil\n}", "func (gb *gcpBalancer) newSubConn() {\n\tgb.mu.Lock()\n\tdefer gb.mu.Unlock()\n\n\t// there are chances the newly created subconns are still connecting,\n\t// we can wait on those new subconns.\n\tfor _, scState := range gb.scStates {\n\t\tif scState == connectivity.Connecting {\n\t\t\treturn\n\t\t}\n\t}\n\n\tsc, err := gb.cc.NewSubConn(\n\t\tgb.addrs,\n\t\tbalancer.NewSubConnOptions{HealthCheckEnabled: healthCheckEnabled},\n\t)\n\tif err != nil {\n\t\tgrpclog.Errorf(\"grpcgcp.gcpBalancer: failed to NewSubConn: %v\", err)\n\t\treturn\n\t}\n\tgb.scRefs[sc] = &subConnRef{\n\t\tsubConn: sc,\n\t}\n\tgb.scStates[sc] = connectivity.Idle\n\tsc.Connect()\n}", "func CreateSubCgroupPath(group, prefix string) (string, error) {\n\tbase := path.Join(basePath, group, prefix)\n\tEnsureDirExists(base)\n\treturn ioutil.TempDir(base, \"\")\n}", "func (ec2 *EC2) CreateNetworkInterface(opts CreateNetworkInterface) (resp *CreateNetworkInterfaceResp, err error) {\n\tparams := makeParams(\"CreateNetworkInterface\")\n\tparams[\"SubnetId\"] = opts.SubnetId\n\tfor i, ip := range opts.PrivateIPs {\n\t\tprefix := fmt.Sprintf(\"PrivateIpAddresses.%d.\", i+1)\n\t\tparams[prefix+\"PrivateIpAddress\"] = ip.Address\n\t\tparams[prefix+\"Primary\"] = strconv.FormatBool(ip.IsPrimary)\n\t}\n\tif opts.Description != \"\" {\n\t\tparams[\"Description\"] = opts.Description\n\t}\n\tif opts.SecondaryPrivateIPCount > 0 {\n\t\tcount := strconv.Itoa(opts.SecondaryPrivateIPCount)\n\t\tparams[\"SecondaryPrivateIpAddressCount\"] = count\n\t}\n\tfor i, groupId := range opts.SecurityGroupIds {\n\t\tparams[\"SecurityGroupId.\"+strconv.Itoa(i+1)] = groupId\n\t}\n\tresp = &CreateNetworkInterfaceResp{}\n\terr = ec2.query(params, resp)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn resp, nil\n}", "func (instance *Network) ToProtocol() (_ *protocol.Network, xerr fail.Error) {\n\tdefer fail.OnPanic(&xerr)\n\n\tif instance == nil || instance.IsNull() {\n\t\treturn nil, fail.InvalidInstanceError()\n\t}\n\n\tinstance.lock.RLock()\n\tdefer instance.lock.RUnlock()\n\n\tvar pn *protocol.Network\n\txerr = instance.Review(func(clonable data.Clonable, props *serialize.JSONProperties) fail.Error {\n\t\tan, ok := clonable.(*abstract.Network)\n\t\tif !ok {\n\t\t\treturn fail.InconsistentError(\"'*abstract.Networking' expected, '%s' provided\", reflect.TypeOf(clonable).String())\n\n\t\t}\n\n\t\tpn = &protocol.Network{\n\t\t\tId: an.ID,\n\t\t\tName: an.Name,\n\t\t\tCidr: an.CIDR,\n\t\t}\n\n\t\treturn props.Inspect(networkproperty.SubnetsV1, func(clonable data.Clonable) fail.Error {\n\t\t\tnsV1, ok := clonable.(*propertiesv1.NetworkSubnets)\n\t\t\tif !ok {\n\t\t\t\treturn fail.InconsistentError(\"'*propertiesv1.NetworkSubnets' expected, '%s' provided\", reflect.TypeOf(clonable).String())\n\t\t\t}\n\t\t\tfor k := range nsV1.ByName {\n\t\t\t\tpn.Subnets = append(pn.Subnets, k)\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\t})\n\txerr = debug.InjectPlannedFail(xerr)\n\tif xerr != nil {\n\t\treturn nil, xerr\n\t}\n\n\treturn pn, nil\n}", "func (a *Azure) DeleteSubnetwork(ctx *lepton.Context, subnetID string) error {\n\tlogger := ctx.Logger()\n\n\tsubnetsClient, err := a.getSubnetsClient()\n\tif err != nil {\n\t\treturn err\n\t}\n\tvnetClient, err := a.getVnetClient()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tsubnetName := getAzureResourceNameFromID(subnetID)\n\tvnName := getAzureVirtualNetworkFromID(subnetID)\n\n\tsubnet, err := subnetsClient.Get(context.TODO(), a.groupName, vnName, subnetName, \"\")\n\tif err != nil {\n\t\tctx.Logger().Error(err)\n\t\treturn fmt.Errorf(\"failed getting subnet\")\n\t}\n\n\tvirtualNetwork, err := vnetClient.Get(context.TODO(), a.groupName, vnName, \"\")\n\tif err != nil {\n\t\tctx.Logger().Error(err)\n\t\treturn errors.New(\"failed getting virtual network\")\n\t}\n\n\tif hasAzureOpsTags(virtualNetwork.Tags) && (subnet.IPConfigurations == nil || len(*subnet.IPConfigurations) == 0) {\n\t\tlogger.Infof(\"Deleting %s...\", *subnet.ID)\n\t\tsubnetDeleteTask, err := subnetsClient.Delete(context.TODO(), a.groupName, subnetName, subnetName)\n\t\tif err != nil {\n\t\t\tlogger.Error(err)\n\t\t\treturn errors.New(\"error deleting subnet\")\n\t\t}\n\n\t\terr = subnetDeleteTask.WaitForCompletionRef(context.TODO(), subnetsClient.Client)\n\t\tif err != nil {\n\t\t\tlogger.Error(err)\n\t\t\treturn errors.New(\"error waiting for subnet deletion\")\n\t\t}\n\n\t\tlogger.Infof(\"Deleting virtualNetworks/%s\", vnName)\n\t\tvnDeleteTask, err := vnetClient.Delete(context.TODO(), a.groupName, vnName)\n\t\tif err != nil {\n\t\t\tlogger.Error(err)\n\t\t\treturn errors.New(\"error deleting virtual network\")\n\t\t}\n\n\t\terr = vnDeleteTask.WaitForCompletionRef(context.TODO(), vnetClient.Client)\n\t\tif err != nil {\n\t\t\tlogger.Error(err)\n\t\t\treturn errors.New(\"error waiting for virtual network deletion\")\n\t\t}\n\t} else {\n\t\treturn errors.New(\"other devices are connected to the same subnet\")\n\t}\n\n\treturn nil\n}", "func (a *Azure) CreateNIC(ctx context.Context, location string, vnetName, subnetName, nsgName, ipName, nicName string, enableIPForwarding bool) (nic network.Interface, err error) {\n\tsubnet, err := a.GetVirtualNetworkSubnet(ctx, vnetName, subnetName)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to get subnet: %v\", err)\n\t}\n\n\tip, err := a.GetPublicIP(ctx, ipName)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to get ip address: %v\", err)\n\t}\n\n\tnicParams := network.Interface{\n\t\tName: to.StringPtr(nicName),\n\t\tLocation: to.StringPtr(location),\n\t\tInterfacePropertiesFormat: &network.InterfacePropertiesFormat{\n\t\t\tIPConfigurations: &[]network.InterfaceIPConfiguration{\n\t\t\t\t{\n\t\t\t\t\tName: to.StringPtr(\"ipConfig1\"),\n\t\t\t\t\tInterfaceIPConfigurationPropertiesFormat: &network.InterfaceIPConfigurationPropertiesFormat{\n\t\t\t\t\t\tSubnet: subnet,\n\t\t\t\t\t\tPrivateIPAllocationMethod: network.Dynamic,\n\t\t\t\t\t\tPublicIPAddress: &ip,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t\tEnableIPForwarding: &enableIPForwarding,\n\t\t},\n\t\tTags: getAzureDefaultTags(),\n\t}\n\n\tif nsgName != \"\" {\n\t\tnsg, err := a.GetNetworkSecurityGroup(ctx, nsgName)\n\t\tif err != nil {\n\t\t\tlog.Fatalf(\"failed to get nsg: %v\", err)\n\t\t}\n\t\tnicParams.NetworkSecurityGroup = nsg\n\t}\n\n\tnicClient := a.getNicClient()\n\n\tfuture, err := nicClient.CreateOrUpdate(ctx, a.groupName, nicName, nicParams)\n\tif err != nil {\n\t\treturn nic, fmt.Errorf(\"cannot create nic: %v\", err)\n\t}\n\n\terr = future.WaitForCompletionRef(ctx, nicClient.Client)\n\tif err != nil {\n\t\treturn nic, fmt.Errorf(\"cannot get nic create or update future response: %v\", err)\n\t}\n\n\tnic, err = future.Result(*nicClient)\n\treturn\n}", "func CreateSubnetWithNetworkSecurityGroup(ctx context.Context, vnetName, subnetName, addressPrefix, nsgName string) (subnet network.Subnet, err error) {\n\tnsg, err := GetNetworkSecurityGroup(ctx, nsgName)\n\tif err != nil {\n\t\treturn subnet, fmt.Errorf(\"cannot get nsg: %v\", err)\n\t}\n\n\tsubnetsClient := getSubnetsClient()\n\tfuture, err := subnetsClient.CreateOrUpdate(\n\t\tctx,\n\t\tconfig.GroupName(),\n\t\tvnetName,\n\t\tsubnetName,\n\t\tnetwork.Subnet{\n\t\t\tSubnetPropertiesFormat: &network.SubnetPropertiesFormat{\n\t\t\t\tAddressPrefix: to.StringPtr(addressPrefix),\n\t\t\t\tNetworkSecurityGroup: &nsg,\n\t\t\t},\n\t\t})\n\tif err != nil {\n\t\treturn subnet, fmt.Errorf(\"cannot create subnet: %v\", err)\n\t}\n\n\terr = future.WaitForCompletionRef(ctx, subnetsClient.Client)\n\tif err != nil {\n\t\treturn subnet, fmt.Errorf(\"cannot get the subnet create or update future response: %v\", err)\n\t}\n\n\treturn future.Result(subnetsClient)\n}", "func NetworkResourceCreate(w http.ResponseWriter, r *http.Request) {\n\tvar req openapi.NetworkResourceCreateRequest\n\tif err := json.NewDecoder(r.Body).Decode(&req); err != nil {\n\t\thandleError(w, http.StatusBadRequest, err.Error())\n\t\treturn\n\t}\n\n\tnetwork, err := ctr.CreateNetworkResource(&req)\n\tif err != nil {\n\t\thandleError(w, http.StatusInternalServerError, err.Error())\n\t\treturn\n\t}\n\n\thandleResult(w, http.StatusCreated, network)\n\treturn\n}", "func (self *AltaActor) createNetwork() error {\n\t// Loop thru each endpoint\n\tfor _, endpoint := range self.Model.Spec.Endpoints {\n\t\tnetwork, err := netCtrler.FindNetwork(endpoint.NetworkName)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Network %s not found while creating Alta: %+v\", self.Model.Spec)\n\t\t\treturn errors.New(\"network not found\")\n\t\t}\n\n\t\t// Send the network info to the node\n\t\tvar resp altaspec.ReqSuccess\n\t\turl := \"/network/create\"\n\t\terr = nodeCtrler.NodePostReq(self.Model.CurrNode, url, network.NetSpec, &resp)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error sending network info to node %s. Err: %v\",\n\t\t\t\tself.Model.CurrNode, err)\n\t\t}\n\t}\n\n\t// Move forward in fsm\n\tself.AltaEvent(\"createVol\")\n\n\treturn nil\n}", "func subnet(base *net.IPNet, newBits int, num int) (*net.IPNet, error) {\n\tip := base.IP\n\tmask := base.Mask\n\n\tbaseLength, addressLength := mask.Size()\n\tnewPrefixLen := baseLength + newBits\n\n\t// check if there is sufficient address space to extend the network prefix\n\tif newPrefixLen > addressLength {\n\t\treturn nil, fmt.Errorf(\"not enought space to extend prefix of %d by %d\", baseLength, newBits)\n\t}\n\n\t// calculate the maximum network number\n\tmaxNetNum := uint64(1<<uint64(newBits)) - 1\n\tif uint64(num) > maxNetNum {\n\t\treturn nil, fmt.Errorf(\"prefix extension of %d does not accommodate a subnet numbered %d\", newBits, num)\n\t}\n\n\treturn &net.IPNet{\n\t\tIP: insertNetworkNumIntoIP(ip, num, newPrefixLen),\n\t\tMask: net.CIDRMask(newPrefixLen, addressLength),\n\t}, nil\n}", "func (a *Azure) CreateSubnetWithNetworkSecurityGroup(ctx context.Context, vnetName, subnetName, addressPrefix, nsgName string) (subnet *network.Subnet, err error) {\n\tnsg, err := a.GetNetworkSecurityGroup(ctx, nsgName)\n\tif err != nil {\n\t\treturn subnet, fmt.Errorf(\"cannot get nsg: %v\", err)\n\t}\n\n\tsubnetsClient, err := a.getSubnetsClient()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfuture, err := subnetsClient.CreateOrUpdate(\n\t\tctx,\n\t\ta.groupName,\n\t\tvnetName,\n\t\tsubnetName,\n\t\tnetwork.Subnet{\n\t\t\tSubnetPropertiesFormat: &network.SubnetPropertiesFormat{\n\t\t\t\tAddressPrefix: to.StringPtr(addressPrefix),\n\t\t\t\tNetworkSecurityGroup: nsg,\n\t\t\t},\n\t\t})\n\tif err != nil {\n\t\treturn subnet, fmt.Errorf(\"cannot create subnet: %v\", err)\n\t}\n\n\terr = future.WaitForCompletionRef(ctx, subnetsClient.Client)\n\tif err != nil {\n\t\treturn subnet, fmt.Errorf(\"cannot get the subnet create or update future response: %v\", err)\n\t}\n\n\tsc, err := future.Result(*subnetsClient)\n\n\treturn &sc, err\n}", "func (g *graphviz) createSubgraph(project string, packages map[string][]string) {\n\t// If there's only a single package and that's the project root, do not\n\t// create a subgraph. Just create a node.\n\tif children, ok := packages[project]; ok && len(packages) == 1 {\n\t\tg.createNode(project, \"\", children)\n\t\treturn\n\t}\n\n\t// Sort and use the packages for consistent output.\n\tpkgs := []gvnode{}\n\n\tfor name, children := range packages {\n\t\tpkgs = append(pkgs, gvnode{project: name, children: children})\n\t}\n\n\tsort.Sort(byGvnode(pkgs))\n\n\tsubgraphPkgs := []string{}\n\trootChildren := []string{}\n\tfor _, p := range pkgs {\n\t\tif p.project == project {\n\t\t\t// Do not create a separate node for the root package.\n\t\t\trootChildren = append(rootChildren, p.children...)\n\t\t\tcontinue\n\t\t}\n\t\tg.createNode(p.project, \"\", p.children)\n\t\tsubgraphPkgs = append(subgraphPkgs, p.project)\n\t}\n\n\tsg := &gvsubgraph{\n\t\tproject:\tproject,\n\t\tpackages:\tsubgraphPkgs,\n\t\tindex:\t\tlen(g.clusters),\n\t\tchildren:\trootChildren,\n\t}\n\n\tg.h[project] = sg.hash()\n\tg.clusters[project] = sg\n}", "func (s stack) CreateNetwork(req abstract.NetworkRequest) (res *abstract.Network, ferr fail.Error) {\n\tnullAN := abstract.NewNetwork()\n\tif s.IsNull() {\n\t\treturn nullAN, fail.InvalidInstanceError()\n\t}\n\n\tdefer debug.NewTracer(nil, tracing.ShouldTrace(\"stack.aws\") || tracing.ShouldTrace(\"stacks.network\"), \"(%v)\", req).WithStopwatch().Entering().Exiting()\n\n\t// Check if network already there\n\tvar xerr fail.Error\n\tif _, xerr = s.rpcDescribeVpcByName(aws.String(req.Name)); xerr != nil {\n\t\tswitch xerr.(type) {\n\t\tcase *fail.ErrNotFound:\n\t\t\tdebug.IgnoreError(xerr)\n\t\t\t// continue\n\t\tdefault:\n\t\t\treturn nullAN, xerr\n\t\t}\n\t} else {\n\t\treturn nullAN, fail.DuplicateError(\"a Network/VPC named '%s' already exists\")\n\t}\n\n\t// if not, create the network/VPC\n\ttheVpc, xerr := s.rpcCreateVpc(aws.String(req.Name), aws.String(req.CIDR))\n\tif xerr != nil {\n\t\treturn nullAN, fail.Wrap(xerr, \"failed to create VPC\")\n\t}\n\n\t// wait until available status\n\tif IsOperation(theVpc, \"State\", reflect.TypeOf(\"\")) {\n\t\tretryErr := retry.WhileUnsuccessful(\n\t\t\tfunc() error {\n\t\t\t\tvpcTmp, innerXErr := s.rpcDescribeVpcByID(theVpc.VpcId)\n\t\t\t\tif innerXErr != nil {\n\t\t\t\t\treturn innerXErr\n\t\t\t\t}\n\t\t\t\tif aws.StringValue(vpcTmp.State) != \"available\" {\n\t\t\t\t\treturn fail.NewError(\"not ready\")\n\t\t\t\t}\n\t\t\t\treturn nil\n\t\t\t},\n\t\t\ttemporal.GetMinDelay(),\n\t\t\ttemporal.GetDefaultDelay(),\n\t\t)\n\t\tif retryErr != nil {\n\t\t\tswitch retryErr.(type) {\n\t\t\tcase *retry.ErrStopRetry:\n\t\t\t\treturn nullAN, fail.Wrap(fail.Cause(retryErr), \"stopping retries\")\n\t\t\tcase *fail.ErrTimeout:\n\t\t\t\treturn nullAN, fail.Wrap(fail.Cause(retryErr), \"timeout\")\n\t\t\tdefault:\n\t\t\t\treturn nullAN, retryErr\n\t\t\t}\n\t\t}\n\t}\n\n\tdefer func() {\n\t\tif ferr != nil && !req.KeepOnFailure {\n\t\t\tif theVpc != nil {\n\t\t\t\tderr := s.DeleteNetwork(aws.StringValue(theVpc.VpcId))\n\t\t\t\tif derr != nil {\n\t\t\t\t\t_ = ferr.AddConsequence(derr)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\n\tgw, xerr := s.rpcCreateInternetGateway()\n\tif xerr != nil {\n\t\treturn nullAN, fail.Wrap(xerr, \"failed to create internet gateway\")\n\t}\n\n\tif xerr = s.rpcAttachInternetGateway(theVpc.VpcId, gw.InternetGatewayId); xerr != nil {\n\t\treturn nullAN, fail.Wrap(xerr, \"failed to attach internet gateway to Network\")\n\t}\n\n\tdefer func() {\n\t\tif ferr != nil && !req.KeepOnFailure {\n\t\t\tif derr := s.rpcDetachInternetGateway(theVpc.VpcId, gw.InternetGatewayId); derr != nil {\n\t\t\t\t_ = ferr.AddConsequence(normalizeError(derr))\n\t\t\t}\n\t\t}\n\t}()\n\n\ttables, xerr := s.rpcDescribeRouteTables(aws.String(\"vpc-id\"), []*string{theVpc.VpcId})\n\tif xerr != nil {\n\t\treturn nullAN, xerr\n\t}\n\tif len(tables) < 1 {\n\t\treturn nullAN, fail.InconsistentError(\"no Route Tables\")\n\t}\n\n\tif xerr = s.rpcCreateRoute(gw.InternetGatewayId, tables[0].RouteTableId, aws.String(\"0.0.0.0/0\")); xerr != nil {\n\t\treturn nullAN, fail.Wrap(xerr, \"failed to create route\")\n\t}\n\n\tdefer func() {\n\t\tif ferr != nil && !req.KeepOnFailure {\n\t\t\tif derr := s.rpcDeleteRoute(tables[0].RouteTableId, aws.String(\"0.0.0.0/0\")); derr != nil {\n\t\t\t\t_ = ferr.AddConsequence(normalizeError(derr))\n\t\t\t}\n\t\t}\n\t}()\n\n\tanet := abstract.NewNetwork()\n\tanet.ID = aws.StringValue(theVpc.VpcId)\n\tanet.Name = req.Name\n\tanet.CIDR = req.CIDR\n\tanet.DNSServers = req.DNSServers\n\n\t// Make sure we log warnings\n\t_ = anet.OK()\n\n\treturn anet, nil\n}", "func (d *DestinationClient) CreateSubaccountDestination(newDestination Destination) error {\n\n\tvar errResponse ErrorMessage\n\n\tresponse, err := d.restyClient.R().\n\t\tSetBody(newDestination).\n\t\tSetError(&errResponse).\n\t\tPost(\"/subaccountDestinations\")\n\n\tif err != nil {\n\t\treturn err\n\t}\n\tif response.StatusCode() != 201 {\n\t\terrResponse.statusCode = response.StatusCode()\n\t\treturn errResponse\n\t}\n\treturn nil\n}", "func (d *Driver) CreateNetwork(r *pluginNet.CreateNetworkRequest) error {\n\tdefer osl.InitOSContext()()\n\n\tid := r.NetworkID\n\topts := r.Options\n\tipV4Data := r.IPv4Data\n\tipV6Data := r.IPv6Data\n\tlogrus.Infof(\"CreateNetwork macvlan with networkID=%s,opts=%s\", id, opts)\n\n\tif id == \"\" {\n\t\treturn fmt.Errorf(\"invalid network id\")\n\t}\n\n\t// reject a null v4 network\n\tif len(ipV4Data) == 0 || ipV4Data[0].Pool == \"0.0.0.0/0\" {\n\t\treturn fmt.Errorf(\"ipv4 pool is empty\")\n\t}\n\n\t// parse and validate the config and bind to networkConfiguration\n\tconfig, err := parseNetworkOptions(id, opts)\n\tif err != nil {\n\t\tstr := fmt.Sprintf(\"CreateNetwork opts is invalid %s\", opts)\n\t\tlogrus.Errorf(str)\n\t\treturn fmt.Errorf(str)\n\t}\n\n\tconfig.ID = id\n\terr = config.processIPAM(id, ipV4Data, ipV6Data)\n\tif err != nil {\n\t\tstr := fmt.Sprintf(\"CreateNetwork ipV4Data is invalid %s\", ipV4Data)\n\t\tlogrus.Errorf(str)\n\t\treturn fmt.Errorf(str)\n\t}\n\t// verify the macvlan mode from -o macvlan_mode option\n\tswitch config.MacvlanMode {\n\tcase \"\", modeBridge:\n\t\t// default to macvlan bridge mode if -o macvlan_mode is empty\n\t\tconfig.MacvlanMode = modeBridge\n\tcase modePrivate:\n\t\tconfig.MacvlanMode = modePrivate\n\tcase modePassthru:\n\t\tconfig.MacvlanMode = modePassthru\n\tcase modeVepa:\n\t\tconfig.MacvlanMode = modeVepa\n\tdefault:\n\t\tstr := fmt.Sprintf(\"requested macvlan mode '%s' is not valid, 'bridge' mode is the macvlan driver default\", config.MacvlanMode)\n\t\tlogrus.Errorf(str)\n\t\treturn fmt.Errorf(str)\n\t}\n\t// loopback is not a valid parent link\n\tif config.Parent == \"lo\" {\n\t\tstr := fmt.Sprintf(\"loopback interface is not a valid %s parent link\", macvlanType)\n\t\tlogrus.Errorf(str)\n\t\treturn fmt.Errorf(str)\n\t}\n\t// if parent interface not specified, create a dummy type link to use named dummy+net_id\n\tif config.Parent == \"\" {\n\t\tconfig.Parent = getDummyName(stringid.TruncateID(config.ID))\n\t\t// empty parent and --internal are handled the same. Set here to update k/v\n\t\tconfig.Internal = true\n\t}\n\n\terr = d.createNetwork(config)\n\tif err != nil {\n\t\tstr := fmt.Sprintf(\"CreateNetwork is failed %v\", err)\n\t\tlogrus.Errorf(str)\n\t\treturn fmt.Errorf(str)\n\t}\n\n\treturn nil\n}", "func (a *Azure) CreateVirtualNetworkAndSubnets(ctx context.Context, location string, vnetName, subnet1Name, subnet2Name string) (vnet *network.VirtualNetwork, err error) {\n\tvnetClient, err := a.getVnetClient()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfuture, err := vnetClient.CreateOrUpdate(\n\t\tctx,\n\t\ta.groupName,\n\t\tvnetName,\n\t\tnetwork.VirtualNetwork{\n\t\t\tLocation: to.StringPtr(location),\n\t\t\tVirtualNetworkPropertiesFormat: &network.VirtualNetworkPropertiesFormat{\n\t\t\t\tAddressSpace: &network.AddressSpace{\n\t\t\t\t\tAddressPrefixes: &[]string{\"10.0.0.0/8\"},\n\t\t\t\t},\n\t\t\t\tSubnets: &[]network.Subnet{\n\t\t\t\t\t{\n\t\t\t\t\t\tName: to.StringPtr(subnet1Name),\n\t\t\t\t\t\tSubnetPropertiesFormat: &network.SubnetPropertiesFormat{\n\t\t\t\t\t\t\tAddressPrefix: to.StringPtr(\"10.0.0.0/16\"),\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t\t{\n\t\t\t\t\t\tName: to.StringPtr(subnet2Name),\n\t\t\t\t\t\tSubnetPropertiesFormat: &network.SubnetPropertiesFormat{\n\t\t\t\t\t\t\tAddressPrefix: to.StringPtr(\"10.1.0.0/16\"),\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t})\n\n\tif err != nil {\n\t\treturn vnet, fmt.Errorf(\"cannot create virtual network: %v\", err)\n\t}\n\n\terr = future.WaitForCompletionRef(ctx, vnetClient.Client)\n\tif err != nil {\n\t\treturn vnet, fmt.Errorf(\"cannot get the vnet create or update future response: %v\", err)\n\t}\n\n\tvn, err := future.Result(*vnetClient)\n\n\treturn &vn, err\n}", "func (m *MockClient) CreateSubnet(arg0 *ec2.CreateSubnetInput) (*ec2.CreateSubnetOutput, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"CreateSubnet\", arg0)\n\tret0, _ := ret[0].(*ec2.CreateSubnetOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func CreateSubCgroupPath(group, prefix string) (string, error) {\n\tbase := path.Join(basePath, group, prefix)\n\tEnsureDirExists(base)\n\treturn os.MkdirTemp(base, \"\")\n}", "func RunNamespacesCreate(c *CmdConfig) error {\n\tlabel, _ := c.Doit.GetString(c.NS, \"label\")\n\tregion, _ := c.Doit.GetString(c.NS, \"region\")\n\tskipConnect, _ := c.Doit.GetBool(c.NS, \"no-connect\")\n\tif label == \"\" || region == \"\" {\n\t\treturn fmt.Errorf(\"the '--label' and '--region' flags are both required\")\n\t}\n\tvalidRegion := getValidRegion(region)\n\tif validRegion == \"\" {\n\t\tfmt.Fprintf(c.Out, \"Valid region values are %+v\\n\", getValidRegions())\n\t\treturn fmt.Errorf(\"'%s' is not a valid region value\", region)\n\t}\n\tss := c.Serverless()\n\tctx := context.TODO()\n\tuniq, err := isLabelUnique(ctx, ss, label)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif !uniq {\n\t\treturn fmt.Errorf(\"you are using label '%s' for another namespace; labels should be unique\", label)\n\t}\n\tif !skipConnect && ss.CheckServerlessStatus() == do.ErrServerlessNotInstalled {\n\t\tskipConnect = true\n\t\tfmt.Fprintln(c.Out, \"Warning: namespace will be created but not connected (serverless software is not installed)\")\n\t}\n\tcreds, err := ss.CreateNamespace(ctx, label, validRegion)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif skipConnect {\n\t\tfmt.Fprintf(c.Out, \"New namespace %s created, but not connected.\\n\", creds.Namespace)\n\t\treturn nil\n\t}\n\terr = ss.WriteCredentials(creds)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfmt.Fprintf(c.Out, \"Connected to functions namespace '%s' on API host '%s'\\n\", creds.Namespace, creds.APIHost)\n\treturn nil\n}", "func newVirtualNetworkClient(subID string, authorizer auth.Authorizer) (*client, error) {\n\tc, err := wssdcloudclient.GetVirtualNetworkClient(&subID, authorizer)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &client{c}, nil\n}", "func (plugin *Plugin) createPATNetworkNamespace(\n\tpatNetNSName string,\n\ttrunk *eni.Trunk,\n\tbranchName string,\n\tbranchMACAddress net.HardwareAddr,\n\tbranchVlanID int,\n\tbranchIPAddress *net.IPNet,\n\tbranchSubnet *vpc.Subnet,\n\tbridgeIPAddress *net.IPNet) (netns.NetNS, error) {\n\t// Create the PAT network namespace.\n\tlog.Infof(\"Creating PAT netns %s.\", patNetNSName)\n\tpatNetNS, err := netns.NewNetNS(patNetNSName)\n\tif err != nil {\n\t\tlog.Errorf(\"Failed to create PAT netns %s: %v.\", patNetNSName, err)\n\t\treturn nil, err\n\t}\n\n\t// Create the branch ENI.\n\tbranch, err := eni.NewBranch(trunk, branchName, branchMACAddress, branchVlanID)\n\tif err != nil {\n\t\tlog.Errorf(\"Failed to create branch interface %s in PAT netns %s: %v.\",\n\t\t\tbranchName, patNetNSName, err)\n\t\treturn nil, err\n\t}\n\n\t// Create a link for the branch ENI.\n\tlog.Infof(\"Creating branch link %s in PAT netns %s.\", branchName, patNetNSName)\n\tif err = branch.AttachToLink(true); err != nil {\n\t\tlog.Errorf(\"Failed to attach branch interface %s in %s: %v.\",\n\t\t\tbranchName, patNetNSName, err)\n\t\treturn nil, err\n\t}\n\n\t// Move branch ENI to the PAT network namespace.\n\tlog.Infof(\"Moving branch link %s to PAT netns %s.\", branchName, patNetNSName)\n\tif err = branch.SetNetNS(patNetNS); err != nil {\n\t\tlog.Errorf(\"Failed to move branch link %s to PAT netns %s: %v.\",\n\t\t\tbranchName, patNetNSName, err)\n\t\treturn nil, err\n\t}\n\n\t// Configure the PAT network namespace.\n\tlog.Infof(\"Setting up PAT netns %s.\", patNetNSName)\n\terr = patNetNS.Run(func() error {\n\t\treturn plugin.setupPATNetworkNamespace(patNetNSName,\n\t\t\tbridgeName, bridgeIPAddress, branch, branchIPAddress, branchSubnet)\n\t})\n\tif err != nil {\n\t\tlog.Errorf(\"Failed to setup PAT netns %s: %v.\", patNetNSName, err)\n\t\treturn nil, err\n\t}\n\treturn patNetNS, nil\n}", "func Subnets(prefix cty.Value, newbits ...cty.Value) (cty.Value, error) {\n\targs := make([]cty.Value, len(newbits)+1)\n\targs[0] = prefix\n\tcopy(args[1:], newbits)\n\treturn SubnetsFunc.Call(args)\n}", "func createSingleHostNetworking(ctx context.Context, svc iaas.Service, singleHostRequest abstract.HostRequest) (_ resources.Subnet, _ func() fail.Error, ferr fail.Error) {\n\t// Build network name\n\tcfg, xerr := svc.GetConfigurationOptions(ctx)\n\tif xerr != nil {\n\t\treturn nil, nil, xerr\n\t}\n\n\tbucketName := cfg.GetString(\"MetadataBucketName\")\n\tif bucketName == \"\" {\n\t\treturn nil, nil, fail.InconsistentError(\"missing service configuration option 'MetadataBucketName'\")\n\t}\n\n\t// Trim and TrimPrefix don't do the same thing\n\tnetworkName := fmt.Sprintf(\"sfnet-%s\", strings.TrimPrefix(bucketName, objectstorage.BucketNamePrefix+\"-\"))\n\n\t// Create network if needed\n\tnetworkInstance, xerr := LoadNetwork(ctx, svc, networkName)\n\tif xerr != nil {\n\t\tswitch xerr.(type) {\n\t\tcase *fail.ErrNotFound:\n\t\t\tnetworkInstance, xerr = NewNetwork(svc)\n\t\t\tif xerr != nil {\n\t\t\t\treturn nil, nil, xerr\n\t\t\t}\n\n\t\t\trequest := abstract.NetworkRequest{\n\t\t\t\tName: networkName,\n\t\t\t\tCIDR: abstract.SingleHostNetworkCIDR,\n\t\t\t\tKeepOnFailure: true,\n\t\t\t}\n\t\t\txerr = networkInstance.Create(ctx, &request, nil)\n\t\t\tif xerr != nil {\n\t\t\t\t// handle a particular case of *fail.ErrDuplicate...\n\t\t\t\tswitch cerr := xerr.(type) {\n\t\t\t\tcase *fail.ErrDuplicate:\n\t\t\t\t\tvalue, found := cerr.Annotation(\"managed\")\n\t\t\t\t\tif found && value != nil {\n\t\t\t\t\t\tmanaged, ok := value.(bool)\n\t\t\t\t\t\tif ok && !managed {\n\t\t\t\t\t\t\treturn nil, nil, xerr\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\tdefault:\n\t\t\t\t}\n\t\t\t\t// ... otherwise, try to get Network that is created by another goroutine\n\t\t\t\tswitch xerr.(type) {\n\t\t\t\tcase *fail.ErrDuplicate, *fail.ErrNotAvailable:\n\t\t\t\t\t// If these errors occurred, another goroutine is running to create the same Network, so wait for it\n\t\t\t\t\tnetworkInstance, xerr = LoadNetwork(ctx, svc, networkName)\n\t\t\t\t\tif xerr != nil {\n\t\t\t\t\t\treturn nil, nil, xerr\n\t\t\t\t\t}\n\t\t\t\tdefault:\n\t\t\t\t\treturn nil, nil, xerr\n\t\t\t\t}\n\t\t\t}\n\t\tdefault:\n\t\t\treturn nil, nil, xerr\n\t\t}\n\t}\n\n\tnid, err := networkInstance.GetID()\n\tif err != nil {\n\t\treturn nil, nil, fail.ConvertError(err)\n\t}\n\n\t// Check if Subnet exists\n\tvar (\n\t\tsubnetRequest abstract.SubnetRequest\n\t\tcidrIndex uint\n\t)\n\tsubnetInstance, xerr := LoadSubnet(ctx, svc, nid, singleHostRequest.ResourceName)\n\tif xerr != nil {\n\t\tswitch xerr.(type) {\n\t\tcase *fail.ErrNotFound:\n\t\t\tsubnetInstance, xerr = NewSubnet(svc)\n\t\t\tif xerr != nil {\n\t\t\t\treturn nil, nil, xerr\n\t\t\t}\n\n\t\t\tvar (\n\t\t\t\tsubnetCIDR string\n\t\t\t)\n\n\t\t\tsubnetCIDR, cidrIndex, xerr = ReserveCIDRForSingleHost(ctx, networkInstance)\n\t\t\txerr = debug.InjectPlannedFail(xerr)\n\t\t\tif xerr != nil {\n\t\t\t\treturn nil, nil, xerr\n\t\t\t}\n\n\t\t\tvar dnsServers []string\n\t\t\topts, xerr := svc.GetConfigurationOptions(ctx)\n\t\t\txerr = debug.InjectPlannedFail(xerr)\n\t\t\tif xerr != nil {\n\t\t\t\tswitch xerr.(type) {\n\t\t\t\tcase *fail.ErrNotFound:\n\t\t\t\tdefault:\n\t\t\t\t\treturn nil, nil, xerr\n\t\t\t\t}\n\t\t\t} else if servers := strings.TrimSpace(opts.GetString(\"DNSServers\")); servers != \"\" {\n\t\t\t\tdnsServers = strings.Split(servers, \",\")\n\t\t\t}\n\n\t\t\tsubnetRequest.Name = singleHostRequest.ResourceName\n\t\t\tsubnetRequest.NetworkID, err = networkInstance.GetID()\n\t\t\tif err != nil {\n\t\t\t\treturn nil, nil, fail.ConvertError(err)\n\t\t\t}\n\t\t\tsubnetRequest.IPVersion = ipversion.IPv4\n\t\t\tsubnetRequest.CIDR = subnetCIDR\n\t\t\tsubnetRequest.DNSServers = dnsServers\n\t\t\tsubnetRequest.HA = false\n\n\t\t\txerr = subnetInstance.CreateSubnetWithoutGateway(ctx, subnetRequest)\n\t\t\txerr = debug.InjectPlannedFail(xerr)\n\t\t\tif xerr != nil {\n\t\t\t\treturn nil, nil, xerr\n\t\t\t}\n\n\t\t\tdefer func() {\n\t\t\t\tferr = debug.InjectPlannedFail(ferr)\n\t\t\t\tif ferr != nil && !singleHostRequest.KeepOnFailure {\n\t\t\t\t\tderr := subnetInstance.Delete(cleanupContextFrom(ctx))\n\t\t\t\t\tif derr != nil {\n\t\t\t\t\t\t_ = ferr.AddConsequence(\n\t\t\t\t\t\t\tfail.Wrap(\n\t\t\t\t\t\t\t\tderr, \"cleaning up on failure, failed to delete Subnet '%s'\",\n\t\t\t\t\t\t\t\tsingleHostRequest.ResourceName,\n\t\t\t\t\t\t\t),\n\t\t\t\t\t\t)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}()\n\n\t\t\t// Sets the CIDR index in instance metadata\n\t\t\txerr = subnetInstance.Alter(ctx, func(clonable data.Clonable, _ *serialize.JSONProperties) fail.Error {\n\t\t\t\tas, ok := clonable.(*abstract.Subnet)\n\t\t\t\tif !ok {\n\t\t\t\t\treturn fail.InconsistentError(\n\t\t\t\t\t\t\"'*abstract.Subnet' expected, '%s' provided\", reflect.TypeOf(clonable).String(),\n\t\t\t\t\t)\n\t\t\t\t}\n\n\t\t\t\tas.SingleHostCIDRIndex = cidrIndex\n\t\t\t\treturn nil\n\t\t\t})\n\t\t\tif xerr != nil {\n\t\t\t\treturn nil, nil, xerr\n\t\t\t}\n\t\tdefault:\n\t\t\treturn nil, nil, xerr\n\t\t}\n\t} else {\n\t\treturn nil, nil, fail.DuplicateError(\"there is already a Subnet named '%s'\", singleHostRequest.ResourceName)\n\t}\n\n\tundoFunc := func() fail.Error {\n\t\tvar errs []error\n\t\tif !singleHostRequest.KeepOnFailure {\n\t\t\tderr := subnetInstance.Delete(cleanupContextFrom(ctx))\n\t\t\tif derr != nil {\n\t\t\t\terrs = append(\n\t\t\t\t\terrs, fail.Wrap(\n\t\t\t\t\t\tderr, \"cleaning up on failure, failed to delete Subnet '%s'\", singleHostRequest.ResourceName,\n\t\t\t\t\t),\n\t\t\t\t)\n\t\t\t}\n\t\t\tderr = FreeCIDRForSingleHost(cleanupContextFrom(ctx), networkInstance, cidrIndex)\n\t\t\tif derr != nil {\n\t\t\t\terrs = append(\n\t\t\t\t\terrs, fail.Wrap(\n\t\t\t\t\t\tderr, \"cleaning up on failure, failed to free CIDR slot in Network '%s'\",\n\t\t\t\t\t\tnetworkInstance.GetName(),\n\t\t\t\t\t),\n\t\t\t\t)\n\t\t\t}\n\t\t}\n\t\tif len(errs) > 0 {\n\t\t\treturn fail.NewErrorList(errs)\n\t\t}\n\t\treturn nil\n\t}\n\n\treturn subnetInstance, undoFunc, nil\n}", "func CreateFailover(request *msgs.CreateFailoverRequest, ns, pgouser string) msgs.CreateFailoverResponse {\n\tvar err error\n\tresp := msgs.CreateFailoverResponse{}\n\tresp.Status.Code = msgs.Ok\n\tresp.Status.Msg = \"\"\n\tresp.Results = make([]string, 0)\n\n\t_, err = validateClusterName(request.ClusterName, ns)\n\tif err != nil {\n\t\tresp.Status.Code = msgs.Error\n\t\tresp.Status.Msg = err.Error()\n\t\treturn resp\n\t}\n\n\tif request.Target != \"\" {\n\t\t_, err = isValidFailoverTarget(request.Target, request.ClusterName, ns)\n\t\tif err != nil {\n\t\t\tresp.Status.Code = msgs.Error\n\t\t\tresp.Status.Msg = err.Error()\n\t\t\treturn resp\n\t\t}\n\t}\n\n\tlog.Debugf(\"create failover called for %s\", request.ClusterName)\n\n\t// Create a pgtask\n\tspec := crv1.PgtaskSpec{}\n\tspec.Namespace = ns\n\tspec.Name = request.ClusterName + \"-\" + config.LABEL_FAILOVER\n\n\t// previous failovers will leave a pgtask so remove it first\n\tkubeapi.Deletepgtask(apiserver.RESTClient, spec.Name, ns)\n\n\tspec.TaskType = crv1.PgtaskFailover\n\tspec.Parameters = make(map[string]string)\n\tspec.Parameters[request.ClusterName] = request.ClusterName\n\n\tlabels := make(map[string]string)\n\tlabels[\"target\"] = request.Target\n\tlabels[config.LABEL_PG_CLUSTER] = request.ClusterName\n\tlabels[config.LABEL_PGOUSER] = pgouser\n\n\tnewInstance := &crv1.Pgtask{\n\t\tObjectMeta: meta_v1.ObjectMeta{\n\t\t\tName: spec.Name,\n\t\t\tLabels: labels,\n\t\t},\n\t\tSpec: spec,\n\t}\n\n\terr = kubeapi.Createpgtask(apiserver.RESTClient,\n\t\tnewInstance, ns)\n\tif err != nil {\n\t\tresp.Status.Code = msgs.Error\n\t\tresp.Status.Msg = err.Error()\n\t\treturn resp\n\t}\n\n\tresp.Results = append(resp.Results, \"created Pgtask (failover) for cluster \"+request.ClusterName)\n\n\treturn resp\n}", "func CreateProspectNetwork(prospectID int64, ec2InstanceID string) (*Network, bool) {\n\tbase10Bits, ok := NextNetwork(ec2InstanceID)\n\tif !ok {\n\t\tlog.Printf(\"error: failed to get next network\")\n\t\treturn nil, false\n\t}\n\tsshPort, ok := NextPort(ec2InstanceID)\n\tif !ok {\n\t\tlog.Printf(\"error: failed to get next SSH port\")\n\t\treturn nil, false\n\t}\n\tipv4Network := calculateNetwork(base10Bits, 0)\n\trand.Seed(time.Now().Unix())\n\tnetwork := &Network{\n\t\tEC2InstanceID: ec2InstanceID,\n\t\tNetwork: ipv4Network,\n\t\tMask: defaultNetworkMask,\n\t\tProspectID: prospectID,\n\t\tExamContainerIP: calculateNetwork(base10Bits, 2),\n\t\tGatewayIP: calculateNetwork(base10Bits, 1),\n\t\tProctorContainerIP: calculateNetwork(\n\t\t\tbase10Bits,\n\t\t\trand.Int63n(24)+230),\n\t\tSSHPort: sshPort,\n\t}\n\terr := connection.QueryRow(`\n\tINSERT INTO prospect_network\n\t(ec2_instance_id, network, mask, prospect_id, exam_container_ip, proctor_container_ip, gateway_ip, ssh_port)\n\tVALUES\n\t($1, $2, $3, $4, $5, $6, $7, $8)\n\tRETURNING id, created, modified\n\t`,\n\t\tnetwork.EC2InstanceID,\n\t\tnetwork.Network,\n\t\tnetwork.Mask,\n\t\tnetwork.ProspectID,\n\t\tnetwork.ExamContainerIP,\n\t\tnetwork.ProctorContainerIP,\n\t\tnetwork.GatewayIP,\n\t\tnetwork.SSHPort,\n\t).Scan(&network.ID, &network.Created, &network.Modified)\n\tif err != nil {\n\t\tlog.Printf(\"error: failed to create network for %d on %s: %s\", prospectID, ec2InstanceID, err)\n\t\treturn nil, false\n\t}\n\treturn network, true\n}", "func (scope *Scope) CreateSubScope() *Scope {\n\treturn CreateScope(scope)\n}", "func (d *DestinationClient) CreateSubaccountCertificate(cert Certificate) error {\n\n\tvar errResponse ErrorMessage\n\n\tresponse, err := d.restyClient.R().\n\t\tSetBody(cert).\n\t\tSetError(&errResponse).\n\t\tPost(\"/subaccountCertificates\")\n\n\tif err != nil {\n\t\treturn err\n\t}\n\tif response.StatusCode() != 201 {\n\t\terrResponse.statusCode = response.StatusCode()\n\t\treturn errResponse\n\t}\n\treturn nil\n}", "func (client *NetworkToNetworkInterconnectsClient) createCreateRequest(ctx context.Context, resourceGroupName string, networkFabricName string, networkToNetworkInterconnectName string, body NetworkToNetworkInterconnect, options *NetworkToNetworkInterconnectsClientBeginCreateOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.ManagedNetworkFabric/networkFabrics/{networkFabricName}/networkToNetworkInterconnects/{networkToNetworkInterconnectName}\"\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif networkFabricName == \"\" {\n\t\treturn nil, errors.New(\"parameter networkFabricName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{networkFabricName}\", url.PathEscape(networkFabricName))\n\tif networkToNetworkInterconnectName == \"\" {\n\t\treturn nil, errors.New(\"parameter networkToNetworkInterconnectName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{networkToNetworkInterconnectName}\", url.PathEscape(networkToNetworkInterconnectName))\n\treq, err := runtime.NewRequest(ctx, http.MethodPut, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2023-06-15\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, runtime.MarshalAsJSON(req, body)\n}", "func (s *Subscribe) CreateSub(contact, api string, admin, ignore bool) *Subscriber {\n\tfor i := range s.Subscribers {\n\t\tif contact == s.Subscribers[i].Contact && api == s.Subscribers[i].API {\n\t\t\ts.Subscribers[i].Admin = admin\n\t\t\ts.Subscribers[i].Ignored = ignore\n\t\t\t// Already exists, return it.\n\t\t\treturn s.Subscribers[i]\n\t\t}\n\t}\n\n\ts.Subscribers = append(s.Subscribers, &Subscriber{\n\t\tContact: contact,\n\t\tAPI: api,\n\t\tAdmin: admin,\n\t\tIgnored: ignore,\n\t\tEvents: &Events{\n\t\t\tMap: make(map[string]*Rules),\n\t\t},\n\t})\n\n\treturn s.Subscribers[len(s.Subscribers)-1]\n}", "func (c *client) CreateNetwork(ctx context.Context, b *pipeline.Build) error {\n\tlogrus.Tracef(\"creating network for pipeline %s\", b.ID)\n\n\t// nolint: lll // ignore long line length due to link\n\t// create the network for the pod\n\t//\n\t// This is done due to the nature of how networking works inside the\n\t// pod. Each container inside the pod shares the same network IP and\n\t// port space. This allows them to communicate with each other via\n\t// localhost. However, to keep the runtime behavior consistent,\n\t// Vela adds DNS entries for each container that requires it.\n\t//\n\t// More info:\n\t// * https://kubernetes.io/docs/concepts/workloads/pods/pod/\n\t// * https://kubernetes.io/docs/concepts/services-networking/add-entries-to-pod-etc-hosts-with-host-aliases/\n\t//\n\t// https://pkg.go.dev/k8s.io/api/core/v1?tab=doc#HostAlias\n\tnetwork := v1.HostAlias{\n\t\tIP: \"127.0.0.1\",\n\t\tHostnames: []string{},\n\t}\n\n\t// iterate through all services in the pipeline\n\tfor _, service := range b.Services {\n\t\t// create the host entry for the pod container aliases\n\t\thost := fmt.Sprintf(\"%s.local\", service.Name)\n\n\t\t// add the host entry to the pod container aliases\n\t\tnetwork.Hostnames = append(network.Hostnames, host)\n\t}\n\n\t// iterate through all steps in the pipeline\n\tfor _, step := range b.Steps {\n\t\t// skip all steps not running in detached mode\n\t\tif !step.Detach {\n\t\t\tcontinue\n\t\t}\n\n\t\t// create the host entry for the pod container aliases\n\t\thost := fmt.Sprintf(\"%s.local\", step.Name)\n\n\t\t// add the host entry to the pod container aliases\n\t\tnetwork.Hostnames = append(network.Hostnames, host)\n\t}\n\n\t// iterate through all stages in the pipeline\n\tfor _, stage := range b.Stages {\n\t\t// iterate through all steps in each stage\n\t\tfor _, step := range stage.Steps {\n\t\t\t// skip all steps not running in detached mode\n\t\t\tif !step.Detach {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// create the host entry for the pod container aliases\n\t\t\thost := fmt.Sprintf(\"%s.local\", step.Name)\n\n\t\t\t// add the host entry to the pod container aliases\n\t\t\tnetwork.Hostnames = append(network.Hostnames, host)\n\t\t}\n\t}\n\n\t// add the network definition to the pod spec\n\t//\n\t// https://pkg.go.dev/k8s.io/api/core/v1?tab=doc#PodSpec\n\tc.Pod.Spec.HostAliases = append(c.Pod.Spec.HostAliases, network)\n\n\treturn nil\n}", "func resourceVolterraNetworkInterfaceCreate(d *schema.ResourceData, meta interface{}) error {\n\tclient := meta.(*APIClient)\n\n\tcreateMeta := &ves_io_schema.ObjectCreateMetaType{}\n\tcreateSpec := &ves_io_schema_network_interface.CreateSpecType{}\n\tcreateReq := &ves_io_schema_network_interface.CreateRequest{\n\t\tMetadata: createMeta,\n\t\tSpec: createSpec,\n\t}\n\n\tif v, ok := d.GetOk(\"annotations\"); ok && !isIntfNil(v) {\n\n\t\tms := map[string]string{}\n\n\t\tfor k, v := range v.(map[string]interface{}) {\n\t\t\tval := v.(string)\n\t\t\tms[k] = val\n\t\t}\n\t\tcreateMeta.Annotations = ms\n\t}\n\n\tif v, ok := d.GetOk(\"description\"); ok && !isIntfNil(v) {\n\t\tcreateMeta.Description =\n\t\t\tv.(string)\n\t}\n\n\tif v, ok := d.GetOk(\"disable\"); ok && !isIntfNil(v) {\n\t\tcreateMeta.Disable =\n\t\t\tv.(bool)\n\t}\n\n\tif v, ok := d.GetOk(\"labels\"); ok && !isIntfNil(v) {\n\n\t\tms := map[string]string{}\n\n\t\tfor k, v := range v.(map[string]interface{}) {\n\t\t\tval := v.(string)\n\t\t\tms[k] = val\n\t\t}\n\t\tcreateMeta.Labels = ms\n\t}\n\n\tif v, ok := d.GetOk(\"name\"); ok && !isIntfNil(v) {\n\t\tcreateMeta.Name =\n\t\t\tv.(string)\n\t}\n\n\tif v, ok := d.GetOk(\"namespace\"); ok && !isIntfNil(v) {\n\t\tcreateMeta.Namespace =\n\t\t\tv.(string)\n\t}\n\n\t//interface_choice\n\n\tinterfaceChoiceTypeFound := false\n\n\tif v, ok := d.GetOk(\"dedicated_interface\"); ok && !interfaceChoiceTypeFound {\n\n\t\tinterfaceChoiceTypeFound = true\n\t\tinterfaceChoiceInt := &ves_io_schema_network_interface.CreateSpecType_DedicatedInterface{}\n\t\tinterfaceChoiceInt.DedicatedInterface = &ves_io_schema_network_interface.DedicatedInterfaceType{}\n\t\tcreateSpec.InterfaceChoice = interfaceChoiceInt\n\n\t\tsl := v.(*schema.Set).List()\n\t\tfor _, set := range sl {\n\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t// device\n\n\t\t\tif v, ok := cs[\"device\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tinterfaceChoiceInt.DedicatedInterface.Device = v.(string)\n\t\t\t}\n\n\t\t\t// monitoring_choice\n\n\t\t\tmonitoringChoiceTypeFound := false\n\n\t\t\tif v, ok := cs[\"monitor\"]; ok && !isIntfNil(v) && !monitoringChoiceTypeFound {\n\n\t\t\t\tmonitoringChoiceTypeFound = true\n\t\t\t\t_ = v\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"monitor_disabled\"]; ok && !isIntfNil(v) && !monitoringChoiceTypeFound {\n\n\t\t\t\tmonitoringChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\tmonitoringChoiceInt := &ves_io_schema_network_interface.DedicatedInterfaceType_MonitorDisabled{}\n\t\t\t\t\tmonitoringChoiceInt.MonitorDisabled = &ves_io_schema.Empty{}\n\t\t\t\t\tinterfaceChoiceInt.DedicatedInterface.MonitoringChoice = monitoringChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\t// mtu\n\n\t\t\tif v, ok := cs[\"mtu\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tinterfaceChoiceInt.DedicatedInterface.Mtu = uint32(v.(int))\n\t\t\t}\n\n\t\t\t// node_choice\n\n\t\t\tnodeChoiceTypeFound := false\n\n\t\t\tif v, ok := cs[\"cluster\"]; ok && !isIntfNil(v) && !nodeChoiceTypeFound {\n\n\t\t\t\tnodeChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\tnodeChoiceInt := &ves_io_schema_network_interface.DedicatedInterfaceType_Cluster{}\n\t\t\t\t\tnodeChoiceInt.Cluster = &ves_io_schema.Empty{}\n\t\t\t\t\tinterfaceChoiceInt.DedicatedInterface.NodeChoice = nodeChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"node\"]; ok && !isIntfNil(v) && !nodeChoiceTypeFound {\n\n\t\t\t\tnodeChoiceTypeFound = true\n\t\t\t\tnodeChoiceInt := &ves_io_schema_network_interface.DedicatedInterfaceType_Node{}\n\n\t\t\t\tinterfaceChoiceInt.DedicatedInterface.NodeChoice = nodeChoiceInt\n\n\t\t\t\tnodeChoiceInt.Node = v.(string)\n\n\t\t\t}\n\n\t\t\t// primary_choice\n\n\t\t\tprimaryChoiceTypeFound := false\n\n\t\t\tif v, ok := cs[\"is_primary\"]; ok && !isIntfNil(v) && !primaryChoiceTypeFound {\n\n\t\t\t\tprimaryChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\tprimaryChoiceInt := &ves_io_schema_network_interface.DedicatedInterfaceType_IsPrimary{}\n\t\t\t\t\tprimaryChoiceInt.IsPrimary = &ves_io_schema.Empty{}\n\t\t\t\t\tinterfaceChoiceInt.DedicatedInterface.PrimaryChoice = primaryChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"not_primary\"]; ok && !isIntfNil(v) && !primaryChoiceTypeFound {\n\n\t\t\t\tprimaryChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\tprimaryChoiceInt := &ves_io_schema_network_interface.DedicatedInterfaceType_NotPrimary{}\n\t\t\t\t\tprimaryChoiceInt.NotPrimary = &ves_io_schema.Empty{}\n\t\t\t\t\tinterfaceChoiceInt.DedicatedInterface.PrimaryChoice = primaryChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\t// priority\n\n\t\t\tif v, ok := cs[\"priority\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tinterfaceChoiceInt.DedicatedInterface.Priority = uint32(v.(int))\n\t\t\t}\n\n\t\t}\n\n\t}\n\n\tif v, ok := d.GetOk(\"dedicated_management_interface\"); ok && !interfaceChoiceTypeFound {\n\n\t\tinterfaceChoiceTypeFound = true\n\t\tinterfaceChoiceInt := &ves_io_schema_network_interface.CreateSpecType_DedicatedManagementInterface{}\n\t\tinterfaceChoiceInt.DedicatedManagementInterface = &ves_io_schema_network_interface.DedicatedManagementInterfaceType{}\n\t\tcreateSpec.InterfaceChoice = interfaceChoiceInt\n\n\t\tsl := v.(*schema.Set).List()\n\t\tfor _, set := range sl {\n\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t// device\n\n\t\t\tif v, ok := cs[\"device\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tinterfaceChoiceInt.DedicatedManagementInterface.Device = v.(string)\n\t\t\t}\n\n\t\t\t// mtu\n\n\t\t\tif v, ok := cs[\"mtu\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tinterfaceChoiceInt.DedicatedManagementInterface.Mtu = uint32(v.(int))\n\t\t\t}\n\n\t\t\t// node_choice\n\n\t\t\tnodeChoiceTypeFound := false\n\n\t\t\tif v, ok := cs[\"cluster\"]; ok && !isIntfNil(v) && !nodeChoiceTypeFound {\n\n\t\t\t\tnodeChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\tnodeChoiceInt := &ves_io_schema_network_interface.DedicatedManagementInterfaceType_Cluster{}\n\t\t\t\t\tnodeChoiceInt.Cluster = &ves_io_schema.Empty{}\n\t\t\t\t\tinterfaceChoiceInt.DedicatedManagementInterface.NodeChoice = nodeChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"node\"]; ok && !isIntfNil(v) && !nodeChoiceTypeFound {\n\n\t\t\t\tnodeChoiceTypeFound = true\n\t\t\t\tnodeChoiceInt := &ves_io_schema_network_interface.DedicatedManagementInterfaceType_Node{}\n\n\t\t\t\tinterfaceChoiceInt.DedicatedManagementInterface.NodeChoice = nodeChoiceInt\n\n\t\t\t\tnodeChoiceInt.Node = v.(string)\n\n\t\t\t}\n\n\t\t}\n\n\t}\n\n\tif v, ok := d.GetOk(\"ethernet_interface\"); ok && !interfaceChoiceTypeFound {\n\n\t\tinterfaceChoiceTypeFound = true\n\t\tinterfaceChoiceInt := &ves_io_schema_network_interface.CreateSpecType_EthernetInterface{}\n\t\tinterfaceChoiceInt.EthernetInterface = &ves_io_schema_network_interface.EthernetInterfaceType{}\n\t\tcreateSpec.InterfaceChoice = interfaceChoiceInt\n\n\t\tsl := v.(*schema.Set).List()\n\t\tfor _, set := range sl {\n\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t// address_choice\n\n\t\t\taddressChoiceTypeFound := false\n\n\t\t\tif v, ok := cs[\"dhcp_client\"]; ok && !isIntfNil(v) && !addressChoiceTypeFound {\n\n\t\t\t\taddressChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\taddressChoiceInt := &ves_io_schema_network_interface.EthernetInterfaceType_DhcpClient{}\n\t\t\t\t\taddressChoiceInt.DhcpClient = &ves_io_schema.Empty{}\n\t\t\t\t\tinterfaceChoiceInt.EthernetInterface.AddressChoice = addressChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"dhcp_server\"]; ok && !isIntfNil(v) && !addressChoiceTypeFound {\n\n\t\t\t\taddressChoiceTypeFound = true\n\t\t\t\taddressChoiceInt := &ves_io_schema_network_interface.EthernetInterfaceType_DhcpServer{}\n\t\t\t\taddressChoiceInt.DhcpServer = &ves_io_schema_network_interface.DHCPServerParametersType{}\n\t\t\t\tinterfaceChoiceInt.EthernetInterface.AddressChoice = addressChoiceInt\n\n\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\tfor _, set := range sl {\n\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t// dhcp_networks\n\n\t\t\t\t\tif v, ok := cs[\"dhcp_networks\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tsl := v.([]interface{})\n\t\t\t\t\t\tdhcpNetworks := make([]*ves_io_schema_network_interface.DHCPNetworkType, len(sl))\n\t\t\t\t\t\taddressChoiceInt.DhcpServer.DhcpNetworks = dhcpNetworks\n\t\t\t\t\t\tfor i, set := range sl {\n\t\t\t\t\t\t\tdhcpNetworks[i] = &ves_io_schema_network_interface.DHCPNetworkType{}\n\t\t\t\t\t\t\tdhcpNetworksMapStrToI := set.(map[string]interface{})\n\n\t\t\t\t\t\t\t// dns_choice\n\n\t\t\t\t\t\t\tdnsChoiceTypeFound := false\n\n\t\t\t\t\t\t\tif v, ok := dhcpNetworksMapStrToI[\"dns_address\"]; ok && !isIntfNil(v) && !dnsChoiceTypeFound {\n\n\t\t\t\t\t\t\t\tdnsChoiceTypeFound = true\n\t\t\t\t\t\t\t\tdnsChoiceInt := &ves_io_schema_network_interface.DHCPNetworkType_DnsAddress{}\n\n\t\t\t\t\t\t\t\tdhcpNetworks[i].DnsChoice = dnsChoiceInt\n\n\t\t\t\t\t\t\t\tdnsChoiceInt.DnsAddress = v.(string)\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tif v, ok := dhcpNetworksMapStrToI[\"same_as_dgw\"]; ok && !isIntfNil(v) && !dnsChoiceTypeFound {\n\n\t\t\t\t\t\t\t\tdnsChoiceTypeFound = true\n\n\t\t\t\t\t\t\t\tif v.(bool) {\n\t\t\t\t\t\t\t\t\tdnsChoiceInt := &ves_io_schema_network_interface.DHCPNetworkType_SameAsDgw{}\n\t\t\t\t\t\t\t\t\tdnsChoiceInt.SameAsDgw = &ves_io_schema.Empty{}\n\t\t\t\t\t\t\t\t\tdhcpNetworks[i].DnsChoice = dnsChoiceInt\n\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t// gateway_choice\n\n\t\t\t\t\t\t\tgatewayChoiceTypeFound := false\n\n\t\t\t\t\t\t\tif v, ok := dhcpNetworksMapStrToI[\"dgw_address\"]; ok && !isIntfNil(v) && !gatewayChoiceTypeFound {\n\n\t\t\t\t\t\t\t\tgatewayChoiceTypeFound = true\n\t\t\t\t\t\t\t\tgatewayChoiceInt := &ves_io_schema_network_interface.DHCPNetworkType_DgwAddress{}\n\n\t\t\t\t\t\t\t\tdhcpNetworks[i].GatewayChoice = gatewayChoiceInt\n\n\t\t\t\t\t\t\t\tgatewayChoiceInt.DgwAddress = v.(string)\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tif v, ok := dhcpNetworksMapStrToI[\"first_address\"]; ok && !isIntfNil(v) && !gatewayChoiceTypeFound {\n\n\t\t\t\t\t\t\t\tgatewayChoiceTypeFound = true\n\n\t\t\t\t\t\t\t\tif v.(bool) {\n\t\t\t\t\t\t\t\t\tgatewayChoiceInt := &ves_io_schema_network_interface.DHCPNetworkType_FirstAddress{}\n\t\t\t\t\t\t\t\t\tgatewayChoiceInt.FirstAddress = &ves_io_schema.Empty{}\n\t\t\t\t\t\t\t\t\tdhcpNetworks[i].GatewayChoice = gatewayChoiceInt\n\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tif v, ok := dhcpNetworksMapStrToI[\"last_address\"]; ok && !isIntfNil(v) && !gatewayChoiceTypeFound {\n\n\t\t\t\t\t\t\t\tgatewayChoiceTypeFound = true\n\n\t\t\t\t\t\t\t\tif v.(bool) {\n\t\t\t\t\t\t\t\t\tgatewayChoiceInt := &ves_io_schema_network_interface.DHCPNetworkType_LastAddress{}\n\t\t\t\t\t\t\t\t\tgatewayChoiceInt.LastAddress = &ves_io_schema.Empty{}\n\t\t\t\t\t\t\t\t\tdhcpNetworks[i].GatewayChoice = gatewayChoiceInt\n\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t// network_prefix_choice\n\n\t\t\t\t\t\t\tnetworkPrefixChoiceTypeFound := false\n\n\t\t\t\t\t\t\tif v, ok := dhcpNetworksMapStrToI[\"network_prefix\"]; ok && !isIntfNil(v) && !networkPrefixChoiceTypeFound {\n\n\t\t\t\t\t\t\t\tnetworkPrefixChoiceTypeFound = true\n\t\t\t\t\t\t\t\tnetworkPrefixChoiceInt := &ves_io_schema_network_interface.DHCPNetworkType_NetworkPrefix{}\n\n\t\t\t\t\t\t\t\tdhcpNetworks[i].NetworkPrefixChoice = networkPrefixChoiceInt\n\n\t\t\t\t\t\t\t\tnetworkPrefixChoiceInt.NetworkPrefix = v.(string)\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tif v, ok := dhcpNetworksMapStrToI[\"network_prefix_allocator\"]; ok && !isIntfNil(v) && !networkPrefixChoiceTypeFound {\n\n\t\t\t\t\t\t\t\tnetworkPrefixChoiceTypeFound = true\n\t\t\t\t\t\t\t\tnetworkPrefixChoiceInt := &ves_io_schema_network_interface.DHCPNetworkType_NetworkPrefixAllocator{}\n\t\t\t\t\t\t\t\tnetworkPrefixChoiceInt.NetworkPrefixAllocator = &ves_io_schema_views.ObjectRefType{}\n\t\t\t\t\t\t\t\tdhcpNetworks[i].NetworkPrefixChoice = networkPrefixChoiceInt\n\n\t\t\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t\t\t\t\t// name\n\n\t\t\t\t\t\t\t\t\tif v, ok := cs[\"name\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\t\t\tnetworkPrefixChoiceInt.NetworkPrefixAllocator.Name = v.(string)\n\t\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t\t\t// namespace\n\n\t\t\t\t\t\t\t\t\tif v, ok := cs[\"namespace\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\t\t\tnetworkPrefixChoiceInt.NetworkPrefixAllocator.Namespace = v.(string)\n\t\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t\t\t// tenant\n\n\t\t\t\t\t\t\t\t\tif v, ok := cs[\"tenant\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\t\t\tnetworkPrefixChoiceInt.NetworkPrefixAllocator.Tenant = v.(string)\n\t\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t// pool_settings\n\n\t\t\t\t\t\t\tif v, ok := dhcpNetworksMapStrToI[\"pool_settings\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\tdhcpNetworks[i].PoolSettings = ves_io_schema_network_interface.DHCPPoolSettingType(ves_io_schema_network_interface.DHCPPoolSettingType_value[v.(string)])\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t// pools\n\n\t\t\t\t\t\t\tif v, ok := dhcpNetworksMapStrToI[\"pools\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\tsl := v.([]interface{})\n\t\t\t\t\t\t\t\tpools := make([]*ves_io_schema_network_interface.DHCPPoolType, len(sl))\n\t\t\t\t\t\t\t\tdhcpNetworks[i].Pools = pools\n\t\t\t\t\t\t\t\tfor i, set := range sl {\n\t\t\t\t\t\t\t\t\tpools[i] = &ves_io_schema_network_interface.DHCPPoolType{}\n\t\t\t\t\t\t\t\t\tpoolsMapStrToI := set.(map[string]interface{})\n\n\t\t\t\t\t\t\t\t\t// end_ip\n\n\t\t\t\t\t\t\t\t\tif w, ok := poolsMapStrToI[\"end_ip\"]; ok && !isIntfNil(w) {\n\t\t\t\t\t\t\t\t\t\tpools[i].EndIp = w.(string)\n\t\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t\t\t// exclude\n\n\t\t\t\t\t\t\t\t\tif w, ok := poolsMapStrToI[\"exclude\"]; ok && !isIntfNil(w) {\n\t\t\t\t\t\t\t\t\t\tpools[i].Exclude = w.(bool)\n\t\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t\t\t// start_ip\n\n\t\t\t\t\t\t\t\t\tif w, ok := poolsMapStrToI[\"start_ip\"]; ok && !isIntfNil(w) {\n\t\t\t\t\t\t\t\t\t\tpools[i].StartIp = w.(string)\n\t\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t\t// dhcp_option82_tag\n\n\t\t\t\t\tif v, ok := cs[\"dhcp_option82_tag\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\taddressChoiceInt.DhcpServer.DhcpOption82Tag = v.(string)\n\t\t\t\t\t}\n\n\t\t\t\t\t// fixed_ip_map\n\n\t\t\t\t\tif v, ok := cs[\"fixed_ip_map\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tms := map[string]string{}\n\t\t\t\t\t\tfor k, v := range v.(map[string]interface{}) {\n\t\t\t\t\t\t\tms[k] = v.(string)\n\t\t\t\t\t\t}\n\t\t\t\t\t\taddressChoiceInt.DhcpServer.FixedIpMap = ms\n\t\t\t\t\t}\n\n\t\t\t\t\t// interfaces_addressing_choice\n\n\t\t\t\t\tinterfacesAddressingChoiceTypeFound := false\n\n\t\t\t\t\tif v, ok := cs[\"automatic_from_end\"]; ok && !isIntfNil(v) && !interfacesAddressingChoiceTypeFound {\n\n\t\t\t\t\t\tinterfacesAddressingChoiceTypeFound = true\n\n\t\t\t\t\t\tif v.(bool) {\n\t\t\t\t\t\t\tinterfacesAddressingChoiceInt := &ves_io_schema_network_interface.DHCPServerParametersType_AutomaticFromEnd{}\n\t\t\t\t\t\t\tinterfacesAddressingChoiceInt.AutomaticFromEnd = &ves_io_schema.Empty{}\n\t\t\t\t\t\t\taddressChoiceInt.DhcpServer.InterfacesAddressingChoice = interfacesAddressingChoiceInt\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := cs[\"automatic_from_start\"]; ok && !isIntfNil(v) && !interfacesAddressingChoiceTypeFound {\n\n\t\t\t\t\t\tinterfacesAddressingChoiceTypeFound = true\n\n\t\t\t\t\t\tif v.(bool) {\n\t\t\t\t\t\t\tinterfacesAddressingChoiceInt := &ves_io_schema_network_interface.DHCPServerParametersType_AutomaticFromStart{}\n\t\t\t\t\t\t\tinterfacesAddressingChoiceInt.AutomaticFromStart = &ves_io_schema.Empty{}\n\t\t\t\t\t\t\taddressChoiceInt.DhcpServer.InterfacesAddressingChoice = interfacesAddressingChoiceInt\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := cs[\"interface_ip_map\"]; ok && !isIntfNil(v) && !interfacesAddressingChoiceTypeFound {\n\n\t\t\t\t\t\tinterfacesAddressingChoiceTypeFound = true\n\t\t\t\t\t\tinterfacesAddressingChoiceInt := &ves_io_schema_network_interface.DHCPServerParametersType_InterfaceIpMap{}\n\t\t\t\t\t\tinterfacesAddressingChoiceInt.InterfaceIpMap = &ves_io_schema_network_interface.DHCPInterfaceIPType{}\n\t\t\t\t\t\taddressChoiceInt.DhcpServer.InterfacesAddressingChoice = interfacesAddressingChoiceInt\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t\t\t// interface_ip_map\n\n\t\t\t\t\t\t\tif v, ok := cs[\"interface_ip_map\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\tms := map[string]string{}\n\t\t\t\t\t\t\t\tfor k, v := range v.(map[string]interface{}) {\n\t\t\t\t\t\t\t\t\tms[k] = v.(string)\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\tinterfacesAddressingChoiceInt.InterfaceIpMap.InterfaceIpMap = ms\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"static_ip\"]; ok && !isIntfNil(v) && !addressChoiceTypeFound {\n\n\t\t\t\taddressChoiceTypeFound = true\n\t\t\t\taddressChoiceInt := &ves_io_schema_network_interface.EthernetInterfaceType_StaticIp{}\n\t\t\t\taddressChoiceInt.StaticIp = &ves_io_schema_network_interface.StaticIPParametersType{}\n\t\t\t\tinterfaceChoiceInt.EthernetInterface.AddressChoice = addressChoiceInt\n\n\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\tfor _, set := range sl {\n\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t// network_prefix_choice\n\n\t\t\t\t\tnetworkPrefixChoiceTypeFound := false\n\n\t\t\t\t\tif v, ok := cs[\"cluster_static_ip\"]; ok && !isIntfNil(v) && !networkPrefixChoiceTypeFound {\n\n\t\t\t\t\t\tnetworkPrefixChoiceTypeFound = true\n\t\t\t\t\t\tnetworkPrefixChoiceInt := &ves_io_schema_network_interface.StaticIPParametersType_ClusterStaticIp{}\n\t\t\t\t\t\tnetworkPrefixChoiceInt.ClusterStaticIp = &ves_io_schema_network_interface.StaticIpParametersClusterType{}\n\t\t\t\t\t\taddressChoiceInt.StaticIp.NetworkPrefixChoice = networkPrefixChoiceInt\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t\t\t// interface_ip_map\n\n\t\t\t\t\t\t\tif v, ok := cs[\"interface_ip_map\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\t\t\tinterfaceIpMap := make(map[string]*ves_io_schema_network_interface.StaticIpParametersNodeType)\n\t\t\t\t\t\t\t\tnetworkPrefixChoiceInt.ClusterStaticIp.InterfaceIpMap = interfaceIpMap\n\t\t\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\t\t\tinterfaceIpMapMapStrToI := set.(map[string]interface{})\n\t\t\t\t\t\t\t\t\tkey, ok := interfaceIpMapMapStrToI[\"name\"]\n\t\t\t\t\t\t\t\t\tif ok && !isIntfNil(key) {\n\t\t\t\t\t\t\t\t\t\tinterfaceIpMap[key.(string)] = &ves_io_schema_network_interface.StaticIpParametersNodeType{}\n\t\t\t\t\t\t\t\t\t\tval, _ := interfaceIpMapMapStrToI[\"value\"]\n\n\t\t\t\t\t\t\t\t\t\tinterfaceIpMapVals := val.(*schema.Set).List()\n\t\t\t\t\t\t\t\t\t\tfor _, intVal := range interfaceIpMapVals {\n\n\t\t\t\t\t\t\t\t\t\t\tinterfaceIpMapStaticMap := intVal.(map[string]interface{})\n\n\t\t\t\t\t\t\t\t\t\t\tif w, ok := interfaceIpMapStaticMap[\"default_gw\"]; ok && !isIntfNil(w) {\n\t\t\t\t\t\t\t\t\t\t\t\tinterfaceIpMap[key.(string)].DefaultGw = w.(string)\n\t\t\t\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t\t\t\t\tif w, ok := interfaceIpMapStaticMap[\"dns_server\"]; ok && !isIntfNil(w) {\n\t\t\t\t\t\t\t\t\t\t\t\tinterfaceIpMap[key.(string)].DnsServer = w.(string)\n\t\t\t\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t\t\t\t\tif w, ok := interfaceIpMapStaticMap[\"ip_address\"]; ok && !isIntfNil(w) {\n\t\t\t\t\t\t\t\t\t\t\t\tinterfaceIpMap[key.(string)].IpAddress = w.(string)\n\t\t\t\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t\t\t\t\t// break after one loop\n\t\t\t\t\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := cs[\"fleet_static_ip\"]; ok && !isIntfNil(v) && !networkPrefixChoiceTypeFound {\n\n\t\t\t\t\t\tnetworkPrefixChoiceTypeFound = true\n\t\t\t\t\t\tnetworkPrefixChoiceInt := &ves_io_schema_network_interface.StaticIPParametersType_FleetStaticIp{}\n\t\t\t\t\t\tnetworkPrefixChoiceInt.FleetStaticIp = &ves_io_schema_network_interface.StaticIpParametersFleetType{}\n\t\t\t\t\t\taddressChoiceInt.StaticIp.NetworkPrefixChoice = networkPrefixChoiceInt\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t\t\t// default_gw\n\n\t\t\t\t\t\t\tif v, ok := cs[\"default_gw\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\tnetworkPrefixChoiceInt.FleetStaticIp.DefaultGw = v.(string)\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t// dns_server\n\n\t\t\t\t\t\t\tif v, ok := cs[\"dns_server\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\tnetworkPrefixChoiceInt.FleetStaticIp.DnsServer = v.(string)\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t// network_prefix_allocator\n\n\t\t\t\t\t\t\tif v, ok := cs[\"network_prefix_allocator\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\tnetworkPrefixAllocatorInt := &ves_io_schema_views.ObjectRefType{}\n\t\t\t\t\t\t\t\tnetworkPrefixChoiceInt.FleetStaticIp.NetworkPrefixAllocator = networkPrefixAllocatorInt\n\n\t\t\t\t\t\t\t\tnpaMapToStrVal := v.(map[string]interface{})\n\t\t\t\t\t\t\t\tif val, ok := npaMapToStrVal[\"name\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\t\t\t\tnetworkPrefixAllocatorInt.Name = val.(string)\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\tif val, ok := npaMapToStrVal[\"namespace\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\t\t\t\tnetworkPrefixAllocatorInt.Namespace = val.(string)\n\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t\tif val, ok := npaMapToStrVal[\"tenant\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\t\t\t\tnetworkPrefixAllocatorInt.Tenant = val.(string)\n\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := cs[\"node_static_ip\"]; ok && !isIntfNil(v) && !networkPrefixChoiceTypeFound {\n\n\t\t\t\t\t\tnetworkPrefixChoiceTypeFound = true\n\t\t\t\t\t\tnetworkPrefixChoiceInt := &ves_io_schema_network_interface.StaticIPParametersType_NodeStaticIp{}\n\t\t\t\t\t\tnetworkPrefixChoiceInt.NodeStaticIp = &ves_io_schema_network_interface.StaticIpParametersNodeType{}\n\t\t\t\t\t\taddressChoiceInt.StaticIp.NetworkPrefixChoice = networkPrefixChoiceInt\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t\t\t// default_gw\n\n\t\t\t\t\t\t\tif v, ok := cs[\"default_gw\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\tnetworkPrefixChoiceInt.NodeStaticIp.DefaultGw = v.(string)\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t// dns_server\n\n\t\t\t\t\t\t\tif v, ok := cs[\"dns_server\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\tnetworkPrefixChoiceInt.NodeStaticIp.DnsServer = v.(string)\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t// ip_address\n\n\t\t\t\t\t\t\tif v, ok := cs[\"ip_address\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\tnetworkPrefixChoiceInt.NodeStaticIp.IpAddress = v.(string)\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\t// device\n\n\t\t\tif v, ok := cs[\"device\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tinterfaceChoiceInt.EthernetInterface.Device = v.(string)\n\t\t\t}\n\n\t\t\t// monitoring_choice\n\n\t\t\tmonitoringChoiceTypeFound := false\n\n\t\t\tif v, ok := cs[\"monitor\"]; ok && !isIntfNil(v) && !monitoringChoiceTypeFound {\n\n\t\t\t\tmonitoringChoiceTypeFound = true\n\t\t\t\t_ = v\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"monitor_disabled\"]; ok && !isIntfNil(v) && !monitoringChoiceTypeFound {\n\n\t\t\t\tmonitoringChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\tmonitoringChoiceInt := &ves_io_schema_network_interface.EthernetInterfaceType_MonitorDisabled{}\n\t\t\t\t\tmonitoringChoiceInt.MonitorDisabled = &ves_io_schema.Empty{}\n\t\t\t\t\tinterfaceChoiceInt.EthernetInterface.MonitoringChoice = monitoringChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\t// mtu\n\n\t\t\tif v, ok := cs[\"mtu\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tinterfaceChoiceInt.EthernetInterface.Mtu = uint32(v.(int))\n\t\t\t}\n\n\t\t\t// network_choice\n\n\t\t\tnetworkChoiceTypeFound := false\n\n\t\t\tif v, ok := cs[\"inside_network\"]; ok && !isIntfNil(v) && !networkChoiceTypeFound {\n\n\t\t\t\tnetworkChoiceTypeFound = true\n\t\t\t\tnetworkChoiceInt := &ves_io_schema_network_interface.EthernetInterfaceType_InsideNetwork{}\n\t\t\t\tnetworkChoiceInt.InsideNetwork = &ves_io_schema_views.ObjectRefType{}\n\t\t\t\tinterfaceChoiceInt.EthernetInterface.NetworkChoice = networkChoiceInt\n\n\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\tfor _, set := range sl {\n\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t// name\n\n\t\t\t\t\tif v, ok := cs[\"name\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tnetworkChoiceInt.InsideNetwork.Name = v.(string)\n\t\t\t\t\t}\n\n\t\t\t\t\t// namespace\n\n\t\t\t\t\tif v, ok := cs[\"namespace\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tnetworkChoiceInt.InsideNetwork.Namespace = v.(string)\n\t\t\t\t\t}\n\n\t\t\t\t\t// tenant\n\n\t\t\t\t\tif v, ok := cs[\"tenant\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tnetworkChoiceInt.InsideNetwork.Tenant = v.(string)\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"site_local_inside_network\"]; ok && !isIntfNil(v) && !networkChoiceTypeFound {\n\n\t\t\t\tnetworkChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\tnetworkChoiceInt := &ves_io_schema_network_interface.EthernetInterfaceType_SiteLocalInsideNetwork{}\n\t\t\t\t\tnetworkChoiceInt.SiteLocalInsideNetwork = &ves_io_schema.Empty{}\n\t\t\t\t\tinterfaceChoiceInt.EthernetInterface.NetworkChoice = networkChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"site_local_network\"]; ok && !isIntfNil(v) && !networkChoiceTypeFound {\n\n\t\t\t\tnetworkChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\tnetworkChoiceInt := &ves_io_schema_network_interface.EthernetInterfaceType_SiteLocalNetwork{}\n\t\t\t\t\tnetworkChoiceInt.SiteLocalNetwork = &ves_io_schema.Empty{}\n\t\t\t\t\tinterfaceChoiceInt.EthernetInterface.NetworkChoice = networkChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"storage_network\"]; ok && !isIntfNil(v) && !networkChoiceTypeFound {\n\n\t\t\t\tnetworkChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\tnetworkChoiceInt := &ves_io_schema_network_interface.EthernetInterfaceType_StorageNetwork{}\n\t\t\t\t\tnetworkChoiceInt.StorageNetwork = &ves_io_schema.Empty{}\n\t\t\t\t\tinterfaceChoiceInt.EthernetInterface.NetworkChoice = networkChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\t// node_choice\n\n\t\t\tnodeChoiceTypeFound := false\n\n\t\t\tif v, ok := cs[\"cluster\"]; ok && !isIntfNil(v) && !nodeChoiceTypeFound {\n\n\t\t\t\tnodeChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\tnodeChoiceInt := &ves_io_schema_network_interface.EthernetInterfaceType_Cluster{}\n\t\t\t\t\tnodeChoiceInt.Cluster = &ves_io_schema.Empty{}\n\t\t\t\t\tinterfaceChoiceInt.EthernetInterface.NodeChoice = nodeChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"node\"]; ok && !isIntfNil(v) && !nodeChoiceTypeFound {\n\n\t\t\t\tnodeChoiceTypeFound = true\n\t\t\t\tnodeChoiceInt := &ves_io_schema_network_interface.EthernetInterfaceType_Node{}\n\n\t\t\t\tinterfaceChoiceInt.EthernetInterface.NodeChoice = nodeChoiceInt\n\n\t\t\t\tnodeChoiceInt.Node = v.(string)\n\n\t\t\t}\n\n\t\t\t// primary_choice\n\n\t\t\tprimaryChoiceTypeFound := false\n\n\t\t\tif v, ok := cs[\"is_primary\"]; ok && !isIntfNil(v) && !primaryChoiceTypeFound {\n\n\t\t\t\tprimaryChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\tprimaryChoiceInt := &ves_io_schema_network_interface.EthernetInterfaceType_IsPrimary{}\n\t\t\t\t\tprimaryChoiceInt.IsPrimary = &ves_io_schema.Empty{}\n\t\t\t\t\tinterfaceChoiceInt.EthernetInterface.PrimaryChoice = primaryChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"not_primary\"]; ok && !isIntfNil(v) && !primaryChoiceTypeFound {\n\n\t\t\t\tprimaryChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\tprimaryChoiceInt := &ves_io_schema_network_interface.EthernetInterfaceType_NotPrimary{}\n\t\t\t\t\tprimaryChoiceInt.NotPrimary = &ves_io_schema.Empty{}\n\t\t\t\t\tinterfaceChoiceInt.EthernetInterface.PrimaryChoice = primaryChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\t// priority\n\n\t\t\tif v, ok := cs[\"priority\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tinterfaceChoiceInt.EthernetInterface.Priority = uint32(v.(int))\n\t\t\t}\n\n\t\t\t// vlan_choice\n\n\t\t\tvlanChoiceTypeFound := false\n\n\t\t\tif v, ok := cs[\"untagged\"]; ok && !isIntfNil(v) && !vlanChoiceTypeFound {\n\n\t\t\t\tvlanChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\tvlanChoiceInt := &ves_io_schema_network_interface.EthernetInterfaceType_Untagged{}\n\t\t\t\t\tvlanChoiceInt.Untagged = &ves_io_schema.Empty{}\n\t\t\t\t\tinterfaceChoiceInt.EthernetInterface.VlanChoice = vlanChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"vlan_id\"]; ok && !isIntfNil(v) && !vlanChoiceTypeFound {\n\n\t\t\t\tvlanChoiceTypeFound = true\n\t\t\t\tvlanChoiceInt := &ves_io_schema_network_interface.EthernetInterfaceType_VlanId{}\n\n\t\t\t\tinterfaceChoiceInt.EthernetInterface.VlanChoice = vlanChoiceInt\n\n\t\t\t\tvlanChoiceInt.VlanId =\n\t\t\t\t\tuint32(v.(int))\n\n\t\t\t}\n\n\t\t}\n\n\t}\n\n\tif v, ok := d.GetOk(\"legacy_interface\"); ok && !interfaceChoiceTypeFound {\n\n\t\tinterfaceChoiceTypeFound = true\n\t\tinterfaceChoiceInt := &ves_io_schema_network_interface.CreateSpecType_LegacyInterface{}\n\t\tinterfaceChoiceInt.LegacyInterface = &ves_io_schema_network_interface.LegacyInterfaceType{}\n\t\tcreateSpec.InterfaceChoice = interfaceChoiceInt\n\n\t\tsl := v.(*schema.Set).List()\n\t\tfor _, set := range sl {\n\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t// DHCP_server\n\n\t\t\tif v, ok := cs[\"dhcp_server\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tinterfaceChoiceInt.LegacyInterface.DHCPServer = ves_io_schema_network_interface.NetworkInterfaceDHCPServer(ves_io_schema_network_interface.NetworkInterfaceDHCPServer_value[v.(string)])\n\n\t\t\t}\n\n\t\t\t// DNS_server\n\n\t\t\tif v, ok := cs[\"dns_server\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\tdnsServer := &ves_io_schema_network_interface.NetworkInterfaceDNS{}\n\t\t\t\tinterfaceChoiceInt.LegacyInterface.DNSServer = dnsServer\n\t\t\t\tfor _, set := range sl {\n\t\t\t\t\tdnsServerMapStrToI := set.(map[string]interface{})\n\n\t\t\t\t\t// dns_mode\n\n\t\t\t\t\tif v, ok := dnsServerMapStrToI[\"dns_mode\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tdnsServer.DnsMode = ves_io_schema_network_interface.NetworkInterfaceDNSMode(ves_io_schema_network_interface.NetworkInterfaceDNSMode_value[v.(string)])\n\n\t\t\t\t\t}\n\n\t\t\t\t\t// dns_server\n\n\t\t\t\t\tif v, ok := dnsServerMapStrToI[\"dns_server\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tsl := v.([]interface{})\n\t\t\t\t\t\tdnsServerIpv4s := make([]*ves_io_schema.Ipv4AddressType, len(sl))\n\t\t\t\t\t\tdnsServer.DnsServer = dnsServerIpv4s\n\t\t\t\t\t\tfor i, set := range sl {\n\t\t\t\t\t\t\tdnsServerIpv4s[i] = &ves_io_schema.Ipv4AddressType{}\n\t\t\t\t\t\t\tdnsServerMapStrToI := set.(map[string]interface{})\n\n\t\t\t\t\t\t\t// addr\n\n\t\t\t\t\t\t\tif w, ok := dnsServerMapStrToI[\"addr\"]; ok && !isIntfNil(w) {\n\t\t\t\t\t\t\t\tdnsServerIpv4s[i].Addr = w.(string)\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\t// address_allocator\n\n\t\t\tif v, ok := cs[\"address_allocator\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tsl := v.([]interface{})\n\t\t\t\taddressAllocatorInt := make([]*ves_io_schema.ObjectRefType, len(sl))\n\t\t\t\tinterfaceChoiceInt.LegacyInterface.AddressAllocator = addressAllocatorInt\n\t\t\t\tfor i, ps := range sl {\n\n\t\t\t\t\taaMapToStrVal := ps.(map[string]interface{})\n\t\t\t\t\taddressAllocatorInt[i] = &ves_io_schema.ObjectRefType{}\n\n\t\t\t\t\taddressAllocatorInt[i].Kind = \"address_allocator\"\n\n\t\t\t\t\tif v, ok := aaMapToStrVal[\"name\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\taddressAllocatorInt[i].Name = v.(string)\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := aaMapToStrVal[\"namespace\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\taddressAllocatorInt[i].Namespace = v.(string)\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := aaMapToStrVal[\"tenant\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\taddressAllocatorInt[i].Tenant = v.(string)\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := aaMapToStrVal[\"uid\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\taddressAllocatorInt[i].Uid = v.(string)\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\t// default_gateway\n\n\t\t\tif v, ok := cs[\"default_gateway\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\tdefaultGateway := &ves_io_schema_network_interface.NetworkInterfaceDFGW{}\n\t\t\t\tinterfaceChoiceInt.LegacyInterface.DefaultGateway = defaultGateway\n\t\t\t\tfor _, set := range sl {\n\t\t\t\t\tdefaultGatewayMapStrToI := set.(map[string]interface{})\n\n\t\t\t\t\t// default_gateway_address\n\n\t\t\t\t\tif v, ok := defaultGatewayMapStrToI[\"default_gateway_address\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tdefaultGatewayAddress := &ves_io_schema.Ipv4AddressType{}\n\t\t\t\t\t\tdefaultGateway.DefaultGatewayAddress = defaultGatewayAddress\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tdefaultGatewayAddressMapStrToI := set.(map[string]interface{})\n\n\t\t\t\t\t\t\t// addr\n\n\t\t\t\t\t\t\tif w, ok := defaultGatewayAddressMapStrToI[\"addr\"]; ok && !isIntfNil(w) {\n\t\t\t\t\t\t\t\tdefaultGatewayAddress.Addr = w.(string)\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t\t// default_gateway_mode\n\n\t\t\t\t\tif v, ok := defaultGatewayMapStrToI[\"default_gateway_mode\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tdefaultGateway.DefaultGatewayMode = ves_io_schema_network_interface.NetworkInterfaceGatewayMode(ves_io_schema_network_interface.NetworkInterfaceGatewayMode_value[v.(string)])\n\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\t// device_name\n\n\t\t\tif v, ok := cs[\"device_name\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tinterfaceChoiceInt.LegacyInterface.DeviceName = v.(string)\n\t\t\t}\n\n\t\t\t// dhcp_address\n\n\t\t\tif v, ok := cs[\"dhcp_address\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tinterfaceChoiceInt.LegacyInterface.DhcpAddress = ves_io_schema_network_interface.NetworkInterfaceDHCP(ves_io_schema_network_interface.NetworkInterfaceDHCP_value[v.(string)])\n\n\t\t\t}\n\n\t\t\t// monitoring_choice\n\n\t\t\tmonitoringChoiceTypeFound := false\n\n\t\t\tif v, ok := cs[\"monitor\"]; ok && !isIntfNil(v) && !monitoringChoiceTypeFound {\n\n\t\t\t\tmonitoringChoiceTypeFound = true\n\t\t\t\t_ = v\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"monitor_disabled\"]; ok && !isIntfNil(v) && !monitoringChoiceTypeFound {\n\n\t\t\t\tmonitoringChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\tmonitoringChoiceInt := &ves_io_schema_network_interface.LegacyInterfaceType_MonitorDisabled{}\n\t\t\t\t\tmonitoringChoiceInt.MonitorDisabled = &ves_io_schema.Empty{}\n\t\t\t\t\tinterfaceChoiceInt.LegacyInterface.MonitoringChoice = monitoringChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\t// mtu\n\n\t\t\tif v, ok := cs[\"mtu\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tinterfaceChoiceInt.LegacyInterface.Mtu = uint32(v.(int))\n\t\t\t}\n\n\t\t\t// priority\n\n\t\t\tif v, ok := cs[\"priority\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tinterfaceChoiceInt.LegacyInterface.Priority = uint32(v.(int))\n\t\t\t}\n\n\t\t\t// static_addresses\n\n\t\t\tif v, ok := cs[\"static_addresses\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tsl := v.([]interface{})\n\t\t\t\tstaticAddresses := make([]*ves_io_schema.Ipv4SubnetType, len(sl))\n\t\t\t\tinterfaceChoiceInt.LegacyInterface.StaticAddresses = staticAddresses\n\t\t\t\tfor i, set := range sl {\n\t\t\t\t\tstaticAddresses[i] = &ves_io_schema.Ipv4SubnetType{}\n\t\t\t\t\tstaticAddressesMapStrToI := set.(map[string]interface{})\n\n\t\t\t\t\t// plen\n\n\t\t\t\t\tif w, ok := staticAddressesMapStrToI[\"plen\"]; ok && !isIntfNil(w) {\n\t\t\t\t\t\tstaticAddresses[i].Plen = uint32(w.(int))\n\t\t\t\t\t}\n\n\t\t\t\t\t// prefix\n\n\t\t\t\t\tif w, ok := staticAddressesMapStrToI[\"prefix\"]; ok && !isIntfNil(w) {\n\t\t\t\t\t\tstaticAddresses[i].Prefix = w.(string)\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\t// tunnel\n\n\t\t\tif v, ok := cs[\"tunnel\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\ttunnel := &ves_io_schema_network_interface.NetworkInterfaceTunnel{}\n\t\t\t\tinterfaceChoiceInt.LegacyInterface.Tunnel = tunnel\n\t\t\t\tfor _, set := range sl {\n\t\t\t\t\ttunnelMapStrToI := set.(map[string]interface{})\n\n\t\t\t\t\t// tunnel\n\n\t\t\t\t\tif v, ok := tunnelMapStrToI[\"tunnel\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tsl := v.([]interface{})\n\t\t\t\t\t\ttunnelInt := make([]*ves_io_schema.ObjectRefType, len(sl))\n\t\t\t\t\t\ttunnel.Tunnel = tunnelInt\n\t\t\t\t\t\tfor i, ps := range sl {\n\n\t\t\t\t\t\t\ttMapToStrVal := ps.(map[string]interface{})\n\t\t\t\t\t\t\ttunnelInt[i] = &ves_io_schema.ObjectRefType{}\n\n\t\t\t\t\t\t\ttunnelInt[i].Kind = \"tunnel\"\n\n\t\t\t\t\t\t\tif v, ok := tMapToStrVal[\"name\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\t\t\ttunnelInt[i].Name = v.(string)\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tif v, ok := tMapToStrVal[\"namespace\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\t\t\ttunnelInt[i].Namespace = v.(string)\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tif v, ok := tMapToStrVal[\"tenant\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\t\t\ttunnelInt[i].Tenant = v.(string)\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\tif v, ok := tMapToStrVal[\"uid\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\t\t\ttunnelInt[i].Uid = v.(string)\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\t// type\n\n\t\t\tif v, ok := cs[\"type\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tinterfaceChoiceInt.LegacyInterface.Type = ves_io_schema_network_interface.NetworkInterfaceType(ves_io_schema_network_interface.NetworkInterfaceType_value[v.(string)])\n\n\t\t\t}\n\n\t\t\t// virtual_network\n\n\t\t\tif v, ok := cs[\"virtual_network\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tsl := v.([]interface{})\n\t\t\t\tvirtualNetworkInt := make([]*ves_io_schema.ObjectRefType, len(sl))\n\t\t\t\tinterfaceChoiceInt.LegacyInterface.VirtualNetwork = virtualNetworkInt\n\t\t\t\tfor i, ps := range sl {\n\n\t\t\t\t\tvnMapToStrVal := ps.(map[string]interface{})\n\t\t\t\t\tvirtualNetworkInt[i] = &ves_io_schema.ObjectRefType{}\n\n\t\t\t\t\tvirtualNetworkInt[i].Kind = \"virtual_network\"\n\n\t\t\t\t\tif v, ok := vnMapToStrVal[\"name\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\tvirtualNetworkInt[i].Name = v.(string)\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := vnMapToStrVal[\"namespace\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\tvirtualNetworkInt[i].Namespace = v.(string)\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := vnMapToStrVal[\"tenant\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\tvirtualNetworkInt[i].Tenant = v.(string)\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := vnMapToStrVal[\"uid\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\tvirtualNetworkInt[i].Uid = v.(string)\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\t// vlan_tag\n\n\t\t\tif v, ok := cs[\"vlan_tag\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tinterfaceChoiceInt.LegacyInterface.VlanTag = uint32(v.(int))\n\t\t\t}\n\n\t\t\t// vlan_tagging\n\n\t\t\tif v, ok := cs[\"vlan_tagging\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tinterfaceChoiceInt.LegacyInterface.VlanTagging = ves_io_schema_network_interface.NetworkInterfaceVLANTagging(ves_io_schema_network_interface.NetworkInterfaceVLANTagging_value[v.(string)])\n\n\t\t\t}\n\n\t\t}\n\n\t}\n\n\tif v, ok := d.GetOk(\"tunnel_interface\"); ok && !interfaceChoiceTypeFound {\n\n\t\tinterfaceChoiceTypeFound = true\n\t\tinterfaceChoiceInt := &ves_io_schema_network_interface.CreateSpecType_TunnelInterface{}\n\t\tinterfaceChoiceInt.TunnelInterface = &ves_io_schema_network_interface.TunnelInterfaceType{}\n\t\tcreateSpec.InterfaceChoice = interfaceChoiceInt\n\n\t\tsl := v.(*schema.Set).List()\n\t\tfor _, set := range sl {\n\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t// mtu\n\n\t\t\tif v, ok := cs[\"mtu\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tinterfaceChoiceInt.TunnelInterface.Mtu = uint32(v.(int))\n\t\t\t}\n\n\t\t\t// network_choice\n\n\t\t\tnetworkChoiceTypeFound := false\n\n\t\t\tif v, ok := cs[\"inside_network\"]; ok && !isIntfNil(v) && !networkChoiceTypeFound {\n\n\t\t\t\tnetworkChoiceTypeFound = true\n\t\t\t\tnetworkChoiceInt := &ves_io_schema_network_interface.TunnelInterfaceType_InsideNetwork{}\n\t\t\t\tnetworkChoiceInt.InsideNetwork = &ves_io_schema_views.ObjectRefType{}\n\t\t\t\tinterfaceChoiceInt.TunnelInterface.NetworkChoice = networkChoiceInt\n\n\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\tfor _, set := range sl {\n\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t// name\n\n\t\t\t\t\tif v, ok := cs[\"name\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tnetworkChoiceInt.InsideNetwork.Name = v.(string)\n\t\t\t\t\t}\n\n\t\t\t\t\t// namespace\n\n\t\t\t\t\tif v, ok := cs[\"namespace\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tnetworkChoiceInt.InsideNetwork.Namespace = v.(string)\n\t\t\t\t\t}\n\n\t\t\t\t\t// tenant\n\n\t\t\t\t\tif v, ok := cs[\"tenant\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\tnetworkChoiceInt.InsideNetwork.Tenant = v.(string)\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"site_local_inside_network\"]; ok && !isIntfNil(v) && !networkChoiceTypeFound {\n\n\t\t\t\tnetworkChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\tnetworkChoiceInt := &ves_io_schema_network_interface.TunnelInterfaceType_SiteLocalInsideNetwork{}\n\t\t\t\t\tnetworkChoiceInt.SiteLocalInsideNetwork = &ves_io_schema.Empty{}\n\t\t\t\t\tinterfaceChoiceInt.TunnelInterface.NetworkChoice = networkChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"site_local_network\"]; ok && !isIntfNil(v) && !networkChoiceTypeFound {\n\n\t\t\t\tnetworkChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\tnetworkChoiceInt := &ves_io_schema_network_interface.TunnelInterfaceType_SiteLocalNetwork{}\n\t\t\t\t\tnetworkChoiceInt.SiteLocalNetwork = &ves_io_schema.Empty{}\n\t\t\t\t\tinterfaceChoiceInt.TunnelInterface.NetworkChoice = networkChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\t// node_choice\n\n\t\t\tnodeChoiceTypeFound := false\n\n\t\t\tif v, ok := cs[\"cluster\"]; ok && !isIntfNil(v) && !nodeChoiceTypeFound {\n\n\t\t\t\tnodeChoiceTypeFound = true\n\n\t\t\t\tif v.(bool) {\n\t\t\t\t\tnodeChoiceInt := &ves_io_schema_network_interface.TunnelInterfaceType_Cluster{}\n\t\t\t\t\tnodeChoiceInt.Cluster = &ves_io_schema.Empty{}\n\t\t\t\t\tinterfaceChoiceInt.TunnelInterface.NodeChoice = nodeChoiceInt\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\tif v, ok := cs[\"node\"]; ok && !isIntfNil(v) && !nodeChoiceTypeFound {\n\n\t\t\t\tnodeChoiceTypeFound = true\n\t\t\t\tnodeChoiceInt := &ves_io_schema_network_interface.TunnelInterfaceType_Node{}\n\n\t\t\t\tinterfaceChoiceInt.TunnelInterface.NodeChoice = nodeChoiceInt\n\n\t\t\t\tnodeChoiceInt.Node = v.(string)\n\n\t\t\t}\n\n\t\t\t// priority\n\n\t\t\tif v, ok := cs[\"priority\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tinterfaceChoiceInt.TunnelInterface.Priority = uint32(v.(int))\n\t\t\t}\n\n\t\t\t// static_ip\n\n\t\t\tif v, ok := cs[\"static_ip\"]; ok && !isIntfNil(v) {\n\n\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\tstaticIp := &ves_io_schema_network_interface.StaticIPParametersType{}\n\t\t\t\tinterfaceChoiceInt.TunnelInterface.StaticIp = staticIp\n\t\t\t\tfor _, set := range sl {\n\t\t\t\t\tstaticIpMapStrToI := set.(map[string]interface{})\n\n\t\t\t\t\t// network_prefix_choice\n\n\t\t\t\t\tnetworkPrefixChoiceTypeFound := false\n\n\t\t\t\t\tif v, ok := staticIpMapStrToI[\"cluster_static_ip\"]; ok && !isIntfNil(v) && !networkPrefixChoiceTypeFound {\n\n\t\t\t\t\t\tnetworkPrefixChoiceTypeFound = true\n\t\t\t\t\t\tnetworkPrefixChoiceInt := &ves_io_schema_network_interface.StaticIPParametersType_ClusterStaticIp{}\n\t\t\t\t\t\tnetworkPrefixChoiceInt.ClusterStaticIp = &ves_io_schema_network_interface.StaticIpParametersClusterType{}\n\t\t\t\t\t\tstaticIp.NetworkPrefixChoice = networkPrefixChoiceInt\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t\t\t// interface_ip_map\n\n\t\t\t\t\t\t\tif v, ok := cs[\"interface_ip_map\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\t\t\tinterfaceIpMap := make(map[string]*ves_io_schema_network_interface.StaticIpParametersNodeType)\n\t\t\t\t\t\t\t\tnetworkPrefixChoiceInt.ClusterStaticIp.InterfaceIpMap = interfaceIpMap\n\t\t\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\t\t\tinterfaceIpMapMapStrToI := set.(map[string]interface{})\n\t\t\t\t\t\t\t\t\tkey, ok := interfaceIpMapMapStrToI[\"name\"]\n\t\t\t\t\t\t\t\t\tif ok && !isIntfNil(key) {\n\t\t\t\t\t\t\t\t\t\tinterfaceIpMap[key.(string)] = &ves_io_schema_network_interface.StaticIpParametersNodeType{}\n\t\t\t\t\t\t\t\t\t\tval, _ := interfaceIpMapMapStrToI[\"value\"]\n\n\t\t\t\t\t\t\t\t\t\tinterfaceIpMapVals := val.(*schema.Set).List()\n\t\t\t\t\t\t\t\t\t\tfor _, intVal := range interfaceIpMapVals {\n\n\t\t\t\t\t\t\t\t\t\t\tinterfaceIpMapStaticMap := intVal.(map[string]interface{})\n\n\t\t\t\t\t\t\t\t\t\t\tif w, ok := interfaceIpMapStaticMap[\"default_gw\"]; ok && !isIntfNil(w) {\n\t\t\t\t\t\t\t\t\t\t\t\tinterfaceIpMap[key.(string)].DefaultGw = w.(string)\n\t\t\t\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t\t\t\t\tif w, ok := interfaceIpMapStaticMap[\"dns_server\"]; ok && !isIntfNil(w) {\n\t\t\t\t\t\t\t\t\t\t\t\tinterfaceIpMap[key.(string)].DnsServer = w.(string)\n\t\t\t\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t\t\t\t\tif w, ok := interfaceIpMapStaticMap[\"ip_address\"]; ok && !isIntfNil(w) {\n\t\t\t\t\t\t\t\t\t\t\t\tinterfaceIpMap[key.(string)].IpAddress = w.(string)\n\t\t\t\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t\t\t\t\t// break after one loop\n\t\t\t\t\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := staticIpMapStrToI[\"fleet_static_ip\"]; ok && !isIntfNil(v) && !networkPrefixChoiceTypeFound {\n\n\t\t\t\t\t\tnetworkPrefixChoiceTypeFound = true\n\t\t\t\t\t\tnetworkPrefixChoiceInt := &ves_io_schema_network_interface.StaticIPParametersType_FleetStaticIp{}\n\t\t\t\t\t\tnetworkPrefixChoiceInt.FleetStaticIp = &ves_io_schema_network_interface.StaticIpParametersFleetType{}\n\t\t\t\t\t\tstaticIp.NetworkPrefixChoice = networkPrefixChoiceInt\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t\t\t// default_gw\n\n\t\t\t\t\t\t\tif v, ok := cs[\"default_gw\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\tnetworkPrefixChoiceInt.FleetStaticIp.DefaultGw = v.(string)\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t// dns_server\n\n\t\t\t\t\t\t\tif v, ok := cs[\"dns_server\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\tnetworkPrefixChoiceInt.FleetStaticIp.DnsServer = v.(string)\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t// network_prefix_allocator\n\n\t\t\t\t\t\t\tif v, ok := cs[\"network_prefix_allocator\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\tnetworkPrefixAllocatorInt := &ves_io_schema_views.ObjectRefType{}\n\t\t\t\t\t\t\t\tnetworkPrefixChoiceInt.FleetStaticIp.NetworkPrefixAllocator = networkPrefixAllocatorInt\n\n\t\t\t\t\t\t\t\tnpaMapToStrVal := v.(map[string]interface{})\n\t\t\t\t\t\t\t\tif val, ok := npaMapToStrVal[\"name\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\t\t\t\tnetworkPrefixAllocatorInt.Name = val.(string)\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\tif val, ok := npaMapToStrVal[\"namespace\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\t\t\t\tnetworkPrefixAllocatorInt.Namespace = val.(string)\n\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t\tif val, ok := npaMapToStrVal[\"tenant\"]; ok && !isIntfNil(v) {\n\t\t\t\t\t\t\t\t\tnetworkPrefixAllocatorInt.Tenant = val.(string)\n\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t\tif v, ok := staticIpMapStrToI[\"node_static_ip\"]; ok && !isIntfNil(v) && !networkPrefixChoiceTypeFound {\n\n\t\t\t\t\t\tnetworkPrefixChoiceTypeFound = true\n\t\t\t\t\t\tnetworkPrefixChoiceInt := &ves_io_schema_network_interface.StaticIPParametersType_NodeStaticIp{}\n\t\t\t\t\t\tnetworkPrefixChoiceInt.NodeStaticIp = &ves_io_schema_network_interface.StaticIpParametersNodeType{}\n\t\t\t\t\t\tstaticIp.NetworkPrefixChoice = networkPrefixChoiceInt\n\n\t\t\t\t\t\tsl := v.(*schema.Set).List()\n\t\t\t\t\t\tfor _, set := range sl {\n\t\t\t\t\t\t\tcs := set.(map[string]interface{})\n\n\t\t\t\t\t\t\t// default_gw\n\n\t\t\t\t\t\t\tif v, ok := cs[\"default_gw\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\tnetworkPrefixChoiceInt.NodeStaticIp.DefaultGw = v.(string)\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t// dns_server\n\n\t\t\t\t\t\t\tif v, ok := cs[\"dns_server\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\tnetworkPrefixChoiceInt.NodeStaticIp.DnsServer = v.(string)\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t// ip_address\n\n\t\t\t\t\t\t\tif v, ok := cs[\"ip_address\"]; ok && !isIntfNil(v) {\n\n\t\t\t\t\t\t\t\tnetworkPrefixChoiceInt.NodeStaticIp.IpAddress = v.(string)\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t\t// tunnel\n\n\t\t\tif v, ok := cs[\"tunnel\"]; ok && !isIntfNil(v) {\n\n\t\t\t\ttunnelInt := &ves_io_schema_views.ObjectRefType{}\n\t\t\t\tinterfaceChoiceInt.TunnelInterface.Tunnel = tunnelInt\n\n\t\t\t\ttMapToStrVal := v.(map[string]interface{})\n\t\t\t\tif val, ok := tMapToStrVal[\"name\"]; ok && !isIntfNil(v) {\n\t\t\t\t\ttunnelInt.Name = val.(string)\n\t\t\t\t}\n\t\t\t\tif val, ok := tMapToStrVal[\"namespace\"]; ok && !isIntfNil(v) {\n\t\t\t\t\ttunnelInt.Namespace = val.(string)\n\t\t\t\t}\n\n\t\t\t\tif val, ok := tMapToStrVal[\"tenant\"]; ok && !isIntfNil(v) {\n\t\t\t\t\ttunnelInt.Tenant = val.(string)\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t}\n\n\t}\n\n\tlog.Printf(\"[DEBUG] Creating Volterra NetworkInterface object with struct: %+v\", createReq)\n\n\tcreateNetworkInterfaceResp, err := client.CreateObject(context.Background(), ves_io_schema_network_interface.ObjectType, createReq)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error creating NetworkInterface: %s\", err)\n\t}\n\td.SetId(createNetworkInterfaceResp.GetObjSystemMetadata().GetUid())\n\n\treturn resourceVolterraNetworkInterfaceRead(d, meta)\n}", "func (net *NetworkCreateInput) CreateNetwork() (CreateNetworkResponse, error) {\n\n\tif status := support.DoesCloudSupports(strings.ToLower(net.Cloud.Name)); status != true {\n\t\treturn CreateNetworkResponse{}, fmt.Errorf(common.DefaultCloudResponse + \"CreateNetwork\")\n\t}\n\n\tswitch strings.ToLower(net.Cloud.Name) {\n\tcase \"aws\":\n\n\t\t// Gets the establish session so that it can carry out the process in cloud\n\t\tsess := (net.Cloud.Client).(*session.Session)\n\n\t\t//authorizing to request further\n\t\tauthinpt := auth.EstablishConnectionInput{Region: net.Cloud.Region, Resource: \"ec2\", Session: sess}\n\n\t\t// Fetching all the networks across cloud aws\n\t\tnetworkin := new(awsnetwork.NetworkCreateInput)\n\t\tnetworkin.Name = net.Name\n\t\tnetworkin.VpcCidr = net.VpcCidr\n\t\tnetworkin.SubCidrs = net.SubCidr\n\t\tnetworkin.Type = net.Type\n\t\tnetworkin.Ports = net.Ports\n\t\tnetworkin.GetRaw = net.Cloud.GetRaw\n\t\tresponse, netErr := networkin.CreateNetwork(authinpt)\n\t\tif netErr != nil {\n\t\t\treturn CreateNetworkResponse{}, netErr\n\t\t}\n\t\treturn CreateNetworkResponse{AwsResponse: response}, nil\n\n\tcase \"azure\":\n\t\treturn CreateNetworkResponse{}, fmt.Errorf(common.DefaultAzResponse)\n\tcase \"gcp\":\n\t\treturn CreateNetworkResponse{}, fmt.Errorf(common.DefaultGcpResponse)\n\tcase \"openstack\":\n\t\treturn CreateNetworkResponse{}, fmt.Errorf(common.DefaultOpResponse)\n\tdefault:\n\t\treturn CreateNetworkResponse{}, fmt.Errorf(common.DefaultCloudResponse + \"CreateNetwork\")\n\t}\n}", "func (account *SCloudaccount) createNetwork(ctx context.Context, wireId, networkType string, net CANetConf) error {\n\tnetwork := &SNetwork{}\n\tnetwork.Name = net.Name\n\tif hint, err := NetworkManager.NewIfnameHint(net.Name); err != nil {\n\t\tlog.Errorf(\"can't NewIfnameHint form hint %s\", net.Name)\n\t} else {\n\t\tnetwork.IfnameHint = hint\n\t}\n\tnetwork.GuestIpStart = net.IpStart\n\tnetwork.GuestIpEnd = net.IpEnd\n\tnetwork.GuestIpMask = net.IpMask\n\tnetwork.GuestGateway = net.Gateway\n\tnetwork.VlanId = int(net.VlanID)\n\tnetwork.WireId = wireId\n\tnetwork.ServerType = networkType\n\tnetwork.IsPublic = true\n\tnetwork.Status = api.NETWORK_STATUS_AVAILABLE\n\tnetwork.PublicScope = string(rbacscope.ScopeDomain)\n\tnetwork.ProjectId = account.ProjectId\n\tnetwork.DomainId = account.DomainId\n\tnetwork.Description = net.Description\n\n\tnetwork.SetModelManager(NetworkManager, network)\n\t// TODO: Prevent IP conflict\n\tlog.Infof(\"create network %s succussfully\", network.Id)\n\terr := NetworkManager.TableSpec().Insert(ctx, network)\n\treturn err\n}", "func TestPrivateSharedSubnet(t *testing.T) {\n\tnewIntegrationTest(\"private-shared-subnet.example.com\", \"private-shared-subnet\").\n\t\twithAddons(\n\t\t\tawsEBSCSIAddon,\n\t\t\tdnsControllerAddon,\n\t\t\tawsCCMAddon,\n\t\t).\n\t\twithPrivate().\n\t\trunTestTerraformAWS(t)\n}", "func (v *CnciVnic) create() error {\n\tvar err error\n\n\tif v.GlobalID == \"\" {\n\t\treturn netError(v, \"create cannot create an unnamed cnci vnic\")\n\t}\n\n\tif v.LinkName == \"\" {\n\t\tif v.LinkName, err = genIface(v, true); err != nil {\n\t\t\treturn netError(v, \"create geniface %v %v\", v.GlobalID, err)\n\t\t}\n\n\t\tif _, err := netlink.LinkByAlias(v.GlobalID); err == nil {\n\t\t\treturn netError(v, \"create interface exists %v %v\", v.GlobalID, err)\n\t\t}\n\t}\n\n\tv.Link.Name = v.LinkName\n\tif v.Link.ParentIndex == 0 {\n\t\treturn netError(v, \"create parent index not set %v %v\", v.GlobalID, v.Link)\n\t}\n\n\tif err := netlink.LinkAdd(v.Link); err != nil {\n\t\treturn netError(v, \"create netlink.LinkAdd %v %v\", v.GlobalID, err)\n\t}\n\n\tlink, err := netlink.LinkByName(v.LinkName)\n\tif err != nil {\n\t\treturn netError(v, \"create netlink.LinkAdd %v %v\", v.GlobalID, err)\n\t}\n\n\tvl, ok := link.(*netlink.Macvtap)\n\tif !ok {\n\t\treturn netError(v, \"create incorrect interface type %v %v\", v.GlobalID, link.Type())\n\t}\n\n\tv.Link = vl\n\n\tif err := v.setAlias(v.GlobalID); err != nil {\n\t\terr1 := v.destroy()\n\t\treturn netError(v, \"create set alias [%v] [%v] [%v]\", v.GlobalID, err, err1)\n\t}\n\n\tif v.MACAddr != nil {\n\t\tif err := v.setHardwareAddr(*v.MACAddr); err != nil {\n\t\t\terr1 := v.destroy()\n\t\t\treturn netError(v, \"create set hardware addr [%v] [%v] [%v] [%v]\",\n\t\t\t\tv.MACAddr.String(), v.GlobalID, err, err1)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (orch *Docker) CreateNetwork(ctx context.Context, netID string) error {\n\t// Check if it already exists\n\t_, err := orch.cli.NetworkInspect(ctx, netID, types.NetworkInspectOptions{})\n\tif err != nil {\n\t\t// Create if doesn't exist\n\t\t_, err = orch.cli.NetworkCreate(ctx, netID, types.NetworkCreate{})\n\t}\n\n\treturn err\n}", "func (n *NetworkCreateCommand) runNetworkCreate(args []string) error {\n\tname := n.name\n\tif len(args) != 0 {\n\t\tname = args[0]\n\t}\n\tif name == \"\" {\n\t\treturn fmt.Errorf(\"network name cannot be empty\")\n\t}\n\n\tnetworkRequest, err := n.buildNetworkCreateRequest(name)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tctx := context.Background()\n\tapiClient := n.cli.Client()\n\tresp, err := apiClient.NetworkCreate(ctx, networkRequest)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif resp.Warning != \"\" {\n\t\tfmt.Printf(\"WARNING: %s \\n\", resp.Warning)\n\t}\n\tfmt.Printf(\"%s: %s\\n\", name, resp.ID)\n\n\treturn nil\n}", "func CreateSubSpace(t *testing.T, awaitilities wait.Awaitilities, opts ...SpaceOption) *toolchainv1alpha1.Space {\n\tspace := NewSpace(t, awaitilities, opts...)\n\n\terr := awaitilities.Host().CreateWithCleanup(t, space)\n\trequire.NoError(t, err)\n\n\treturn space\n}", "func (r *ReactorGraph) CreateCompute2(a Cell, b Cell, f func(int, int) int) ComputeCell {\n\tn := &Node{a: a.(*Node), b: b.(*Node), f2: f, value: f(a.Value(), b.Value())}\n\ta.(*Node).addDependency(n)\n\tb.(*Node).addDependency(n)\n\treturn n\n}", "func rewriteNetworkInContainerCreate(body []byte, space_id string) (b []byte) {\n\tvar sep = []byte(\"\\\"NetworkMode\\\":\\\"\")\n\ti := bytes.Index(body, sep)\n\ti += 15 //position of net name\n\tj := bytes.Index(body[i:], []byte(\"\\\"\")) // j == len of name\n\tj += i //position of double-quote after net name\n\n\tglog.Infof(\"i=%d j=%d\", i, j)\n\t//var nameBytes []byte\n\t//nameBytes = make([]byte, j-i+1)\n\tglog.Infof(\"**%s**\", string(body[i:j]))\n\n\t//copy(nameBytes, body[i:j])\n\t//buf := bytes.NewBuffer(nameBytes)\n\t//nameString := buf.String()\n\tnameString := string(body[i:j])\n\t//nameString = strings.TrimSpace(nameString)\n\tglog.Infof(\"nameString=**%s**\", nameString)\n\n\tfullnameString := uniqueNetName(nameString, space_id)\n\n\tnewBodyStr := string(body[:i]) + fullnameString + string(body[j:])\n\n\t//b = make([]byte, len(newBody))\n\tb = []byte(newBodyStr)\n\n\tglog.Infof(\"rewriteNetworkInContainerCreate: New Body=**%s**\", newBodyStr)\n\n\treturn\n}" ]
[ "0.6361011", "0.61224127", "0.59131217", "0.58526635", "0.58484566", "0.5747736", "0.5738869", "0.5733577", "0.57101136", "0.5664078", "0.56429756", "0.5593993", "0.558692", "0.5578184", "0.55734694", "0.5531528", "0.55255383", "0.548908", "0.54657733", "0.5454603", "0.54461867", "0.54232746", "0.54214185", "0.5385212", "0.53820664", "0.5374957", "0.53716177", "0.5348333", "0.5347199", "0.53226227", "0.53130937", "0.53086513", "0.52967006", "0.52874947", "0.52783084", "0.52783084", "0.52447456", "0.5242747", "0.5237132", "0.5234592", "0.5231083", "0.52271", "0.5217402", "0.5197552", "0.5190987", "0.51641965", "0.51637316", "0.5148585", "0.51456285", "0.5143565", "0.5132968", "0.5128487", "0.51035005", "0.5101985", "0.5099988", "0.50920767", "0.50919604", "0.50906223", "0.50892735", "0.5076822", "0.5061149", "0.5045971", "0.50430626", "0.5040268", "0.5036042", "0.502292", "0.5018618", "0.5015191", "0.5011428", "0.50043875", "0.50003237", "0.49992967", "0.49966314", "0.4995943", "0.49868596", "0.49755296", "0.49672788", "0.4965385", "0.49652112", "0.49512723", "0.49432158", "0.49308908", "0.4930835", "0.4920906", "0.49123865", "0.4905549", "0.49033204", "0.48746425", "0.48619518", "0.48608562", "0.48536846", "0.48531184", "0.4846627", "0.48434472", "0.48413426", "0.48397887", "0.48390928", "0.48229706", "0.4822644", "0.48109412" ]
0.72097254
0
CreateTargetInstance uses the override method CreateTargetInstanceFn or the real implementation.
func (c *TestClient) CreateTargetInstance(project, zone string, ti *compute.TargetInstance) error { if c.CreateTargetInstanceFn != nil { return c.CreateTargetInstanceFn(project, zone, ti) } return c.client.CreateTargetInstance(project, zone, ti) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (m *Client) CreateTarget(arg0 context.Context, arg1 zendesk.Target) (zendesk.Target, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"CreateTarget\", arg0, arg1)\n\tret0, _ := ret[0].(zendesk.Target)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func NewTarget(namespace, pod, container string) *Target {\n\treturn &Target{\n\t\tNamespace: namespace,\n\t\tPod: pod,\n\t\tContainer: container,\n\t}\n}", "func (p *ProxMox) CreateInstance(ctx *lepton.Context) error {\n\n\tvar err error\n\n\tconfig := ctx.Config()\n\n\tnextid := p.getNextID()\n\n\tp.instanceName = config.RunConfig.InstanceName\n\n\tp.isoStorageName = config.TargetConfig[\"IsoStorageName\"]\n\n\tp.imageName = config.CloudConfig.ImageName\n\n\tp.arch = \"x86_64\"\n\tif config.TargetConfig[\"Arch\"] != \"\" {\n\t\tp.arch = config.TargetConfig[\"Arch\"]\n\t}\n\n\tp.machine = \"q35\"\n\tif config.TargetConfig[\"Machine\"] != \"\" {\n\t\tp.machine = config.TargetConfig[\"Machine\"]\n\t}\n\n\tp.sockets = \"1\"\n\tif config.TargetConfig[\"Sockets\"] != \"\" {\n\t\tsocketsInt, err := strconv.Atoi(config.TargetConfig[\"Sockets\"])\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif socketsInt < 1 {\n\t\t\treturn errors.New(\"Bad configuration option; Sockets can only be set postitive starting from \\\"1\\\"\")\n\t\t}\n\t\tp.sockets = config.TargetConfig[\"Sockets\"]\n\t}\n\n\tp.cores = \"1\"\n\tif config.TargetConfig[\"Cores\"] != \"\" {\n\t\tcoresInt, err := strconv.Atoi(config.TargetConfig[\"Cores\"])\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif coresInt < 1 {\n\t\t\treturn errors.New(\"Bad configuration option; Cores can only be set postitive starting from \\\"1\\\"\")\n\t\t}\n\t\tp.cores = config.TargetConfig[\"Cores\"]\n\t}\n\n\tp.numa = \"0\"\n\tif config.TargetConfig[\"Numa\"] != \"\" {\n\t\tif config.TargetConfig[\"Numa\"] != \"0\" && config.TargetConfig[\"Numa\"] != \"1\" {\n\t\t\treturn errors.New(\"Bad configuration option; Numa can only be set to \\\"0\\\" or \\\"1\\\"\")\n\t\t}\n\t\tp.numa = config.TargetConfig[\"Numa\"]\n\t}\n\n\t// Memory\n\n\tp.memory = \"512\"\n\tif config.TargetConfig[\"Memory\"] != \"\" {\n\t\tmemoryInt, err := lepton.RAMInBytes(config.TargetConfig[\"Memory\"])\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tmemoryInt = memoryInt / 1024 / 1024\n\t\tp.memory = strconv.FormatInt(memoryInt, 10)\n\t}\n\n\t// Main storage\n\n\tp.storageName = \"local-lvm\"\n\tif config.TargetConfig[\"StorageName\"] != \"\" {\n\t\tp.storageName = config.TargetConfig[\"StorageName\"]\n\t}\n\n\t// Iso storage\n\n\tp.isoStorageName = \"local\"\n\tif config.TargetConfig[\"IsoStorageName\"] != \"\" {\n\t\tp.isoStorageName = config.TargetConfig[\"IsoStorageName\"]\n\t}\n\n\t// Bridge prefix\n\n\tp.bridgePrefix = \"vmbr\"\n\tif config.TargetConfig[\"BridgePrefix\"] != \"\" {\n\t\tp.bridgePrefix = config.TargetConfig[\"BridgePrefix\"]\n\t}\n\n\t// Onboot\n\n\tp.onboot = \"0\"\n\tif config.TargetConfig[\"Onboot\"] != \"\" {\n\t\tif config.TargetConfig[\"Onboot\"] != \"0\" && config.TargetConfig[\"Onboot\"] != \"1\" {\n\t\t\treturn errors.New(\"Bad configuration option; Onboot can only be set to \\\"0\\\" or \\\"1\\\"\")\n\t\t}\n\t\tp.onboot = config.TargetConfig[\"Onboot\"]\n\t}\n\n\t// Protection\n\n\tp.protection = \"0\"\n\tif config.TargetConfig[\"Protection\"] != \"\" {\n\t\tif config.TargetConfig[\"Protection\"] != \"0\" && config.TargetConfig[\"Protection\"] != \"1\" {\n\t\t\treturn errors.New(\"Bad configuration option; Protection can only be set to \\\"0\\\" or \\\"1\\\"\")\n\t\t}\n\t\tp.protection = config.TargetConfig[\"Protection\"]\n\t}\n\n\t// These two preventive checks here, because Proxmox will not return\n\t// an error if the storage is missing and a misconfigured instance will be created.\n\n\terr = p.CheckStorage(p.storageName, \"images\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = p.CheckStorage(p.isoStorageName, \"iso\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdata := url.Values{}\n\tdata.Set(\"vmid\", nextid)\n\tdata.Set(\"name\", p.instanceName)\n\tdata.Set(\"name\", p.imageName)\n\tdata.Set(\"machine\", p.machine)\n\tdata.Set(\"sockets\", p.sockets)\n\tdata.Set(\"cores\", p.cores)\n\tdata.Set(\"numa\", p.numa)\n\tdata.Set(\"memory\", p.memory)\n\tdata.Set(\"onboot\", p.onboot)\n\tdata.Set(\"protection\", p.protection)\n\tdata.Set(\"serial0\", \"socket\")\n\n\t// Configuring network interfaces\n\n\tnics := config.RunConfig.Nics\n\tfor i := 0; i < len(nics); i++ {\n\t\tis := strconv.Itoa(i)\n\t\tbrName := nics[i].BridgeName\n\t\tif brName == \"\" {\n\t\t\tbrName = p.bridgePrefix + is\n\t\t}\n\n\t\terr = p.CheckBridge(brName)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif nics[i].IPAddress != \"\" {\n\t\t\tcidr := \"24\"\n\n\t\t\tif nics[i].NetMask != \"\" {\n\t\t\t\tcidrInt := lepton.CCidr(nics[i].NetMask)\n\t\t\t\tcidr = strconv.FormatInt(int64(cidrInt), 10)\n\t\t\t}\n\n\t\t\tif nics[i].Gateway != \"\" {\n\t\t\t\tdata.Set(\"ipconfig\"+is, \"ip=\"+nics[i].IPAddress+\"/\"+cidr+\",\"+\"gw=\"+nics[i].Gateway)\n\t\t\t} else {\n\t\t\t\tdata.Set(\"ipconfig\"+is, \"ip=\"+nics[i].IPAddress+\"/\"+cidr)\n\t\t\t}\n\t\t} else {\n\t\t\tdata.Set(\"ipconfig\"+is, \"dhcp\")\n\t\t}\n\n\t\tdata.Set(\"net\"+is, \"model=virtio,bridge=\"+brName)\n\t}\n\tif len(nics) == 0 {\n\t\t// single dhcp nic\n\t\tdata.Set(\"net0\", \"model=virtio,bridge=vmbr0\")\n\t}\n\n\treq, err := http.NewRequest(\"POST\", p.apiURL+\"/api2/json/nodes/\"+p.nodeNAME+\"/qemu\", bytes.NewBufferString(data.Encode()))\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\ttr := &http.Transport{\n\t\tTLSClientConfig: &tls.Config{InsecureSkipVerify: true},\n\t}\n\tclient := &http.Client{Transport: tr}\n\n\treq.Header.Add(\"Authorization\", \"PVEAPIToken=\"+p.tokenID+\"=\"+p.secret)\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\tbody, err := io.ReadAll(resp.Body)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\tdebug := false\n\tif debug {\n\t\tfmt.Println(string(body))\n\t}\n\n\terr = p.CheckResultType(body, \"createinstance\", \"file=\"+p.isoStorageName+\":iso/\"+p.imageName+\".iso\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = p.addVirtioDisk(ctx, nextid)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = p.movDisk(ctx, nextid)\n\n\treturn err\n}", "func (a *AdminApiService) CreateTarget(ctx _context.Context, target Target) (*_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodPost\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/admin/target\"\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{\"application/json\"}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &target\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 400 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 401 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 403 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 409 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 500 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t}\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func (_e *MockTransport_Expecter) CreateTransportInstance(transportUrl interface{}, _a1 interface{}, _options ...interface{}) *MockTransport_CreateTransportInstance_Call {\n\treturn &MockTransport_CreateTransportInstance_Call{Call: _e.mock.On(\"CreateTransportInstance\",\n\t\tappend([]interface{}{transportUrl, _a1}, _options...)...)}\n}", "func NewTarget(node RegistNode) error {\n\tlock.Lock()\n\tdefer lock.Unlock()\n\tif _, ok := registryMap[node.Domain]; !ok {\n\t\tif registryMap == nil {\n\t\t\tregistryMap = map[string]RegistNode{}\n\t\t}\n\n\t\tregistryMap[node.Domain] = node\n\t\treturn nil\n\t}\n\treturn ErrServiceExisted\n}", "func CreateTaskTarget(prov provider.Provider, userVariables config.Config, serviceParams config.ServiceParams,\n\tfetcher state.Fetcher, localPath, remotePath string, convertNewline, overwrite, useStorage bool,\n\targs []string, uploadFiles []string, downloadFiles []string) (controller.Thing, error) {\n\treturn &taskState{\n\t\tstatus: NotRunning,\n\t\tprovider: prov,\n\t\tuserVariables: userVariables,\n\t\tlocalPath: localPath,\n\t\tremotePath: remotePath,\n\t\targs: args,\n\t\toverwrite: overwrite,\n\t\tconvertNewline: convertNewline,\n\t\tuseStorage: useStorage,\n\t\tclient: nil,\n\t\tfetcher: fetcher,\n\t\tserviceParameters: serviceParams,\n\t\tuploadFiles: uploadFiles,\n\t\tdownloadFiles: downloadFiles,\n\t}, nil\n}", "func NewTarget(typ string, g *graph.Graph, n *graph.Node, capture *types.Capture, uuids flow.UUIDs, bpf *flow.BPF, fta *flow.TableAllocator) (Target, error) {\n\tswitch typ {\n\tcase \"netflowv5\":\n\t\treturn NewNetFlowV5Target(g, n, capture, uuids)\n\tcase \"erspanv1\":\n\t\treturn NewERSpanTarget(g, n, capture)\n\tcase \"\", \"local\":\n\t\treturn NewLocalTarget(g, n, capture, uuids, fta)\n\t}\n\n\treturn nil, ErrTargetTypeUnknown\n}", "func (mr *ClientMockRecorder) CreateTarget(arg0, arg1 interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"CreateTarget\", reflect.TypeOf((*Client)(nil).CreateTarget), arg0, arg1)\n}", "func NewTarget() Target {\n\treturn Target{Alias: \"$tag_host $tag_name\", DsType: \"influxdb\"}\n}", "func (p *OnPrem) CreateInstance(ctx *Context) error {\n\tc := ctx.config\n\n\thypervisor := HypervisorInstance()\n\tif hypervisor == nil {\n\t\tfmt.Println(\"No hypervisor found on $PATH\")\n\t\tfmt.Println(\"Please install OPS using curl https://ops.city/get.sh -sSfL | sh\")\n\t\tos.Exit(1)\n\t}\n\n\tinstancename := c.CloudConfig.ImageName\n\n\tfmt.Printf(\"booting %s ...\\n\", instancename)\n\n\topshome := GetOpsHome()\n\timgpath := path.Join(opshome, \"images\", instancename)\n\n\tc.RunConfig.BaseName = instancename\n\tc.RunConfig.Imagename = imgpath\n\tc.RunConfig.OnPrem = true\n\n\thypervisor.Start(&c.RunConfig)\n\n\treturn nil\n}", "func (f *FakeInstance) Create(_ context.Context, _ *govultr.InstanceCreateReq) (*govultr.Instance, *http.Response, error) {\n\tpanic(\"implement me\")\n}", "func NewTarget(labels, discoveredLabels labels.Labels, params url.Values) *Target {\n\treturn &Target{\n\t\tlabels: labels,\n\t\tdiscoveredLabels: discoveredLabels,\n\t\tparams: params,\n\t\thealth: HealthUnknown,\n\t}\n}", "func NewTarget(url string) (t *Target) {\n t = &Target{Url:url, method:defaultMethod, header:http.Header{}}\n return t\n}", "func createTarget(s *scope, args []pyObject) *core.BuildTarget {\n\tisTruthy := func(i int) bool {\n\t\treturn args[i] != nil && args[i] != None && args[i].IsTruthy()\n\t}\n\tname := string(args[nameBuildRuleArgIdx].(pyString))\n\ttestCmd := args[testCMDBuildRuleArgIdx]\n\ttest := isTruthy(testBuildRuleArgIdx)\n\t// A bunch of error checking first\n\ts.NAssert(name == \"all\", \"'all' is a reserved build target name.\")\n\ts.NAssert(name == \"\", \"Target name is empty\")\n\ts.NAssert(strings.ContainsRune(name, '/'), \"/ is a reserved character in build target names\")\n\ts.NAssert(strings.ContainsRune(name, ':'), \": is a reserved character in build target names\")\n\n\tif tag := args[tagBuildRuleArgIdx]; tag != nil {\n\t\tif tagStr := string(tag.(pyString)); tagStr != \"\" {\n\t\t\tname = tagName(name, tagStr)\n\t\t}\n\t}\n\tlabel, err := core.TryNewBuildLabel(s.pkg.Name, name)\n\ts.Assert(err == nil, \"Invalid build target name %s\", name)\n\tlabel.Subrepo = s.pkg.SubrepoName\n\n\ttarget := core.NewBuildTarget(label)\n\ttarget.Subrepo = s.pkg.Subrepo\n\ttarget.IsBinary = isTruthy(binaryBuildRuleArgIdx)\n\ttarget.IsSubrepo = isTruthy(subrepoArgIdx)\n\ttarget.NeedsTransitiveDependencies = isTruthy(needsTransitiveDepsBuildRuleArgIdx)\n\ttarget.OutputIsComplete = isTruthy(outputIsCompleteBuildRuleArgIdx)\n\ttarget.Sandbox = isTruthy(sandboxBuildRuleArgIdx)\n\ttarget.TestOnly = test || isTruthy(testOnlyBuildRuleArgIdx)\n\ttarget.ShowProgress.Set(isTruthy(progressBuildRuleArgIdx))\n\ttarget.IsRemoteFile = isTruthy(urlsBuildRuleArgIdx)\n\ttarget.IsTextFile = args[cmdBuildRuleArgIdx] == textFileCommand\n\ttarget.Local = isTruthy(localBuildRuleArgIdx)\n\ttarget.ExitOnError = isTruthy(exitOnErrorArgIdx)\n\tfor _, o := range asStringList(s, args[outDirsBuildRuleArgIdx], \"output_dirs\") {\n\t\ttarget.AddOutputDirectory(o)\n\t}\n\n\tvar size *core.Size\n\tif args[sizeBuildRuleArgIdx] != None {\n\t\tname := string(args[sizeBuildRuleArgIdx].(pyString))\n\t\tsize = mustSize(s, name)\n\t\ttarget.AddLabel(name)\n\t}\n\tif args[passEnvBuildRuleArgIdx] != None {\n\t\tl := asStringList(s, mustList(args[passEnvBuildRuleArgIdx]), \"pass_env\")\n\t\ttarget.PassEnv = &l\n\t}\n\n\ttarget.BuildTimeout = sizeAndTimeout(s, size, args[buildTimeoutBuildRuleArgIdx], s.state.Config.Build.Timeout)\n\ttarget.Stamp = isTruthy(stampBuildRuleArgIdx)\n\ttarget.IsFilegroup = args[cmdBuildRuleArgIdx] == filegroupCommand\n\tif desc := args[buildingDescriptionBuildRuleArgIdx]; desc != nil && desc != None {\n\t\ttarget.BuildingDescription = string(desc.(pyString))\n\t}\n\tif target.IsBinary {\n\t\ttarget.AddLabel(\"bin\")\n\t}\n\tif target.IsRemoteFile {\n\t\ttarget.AddLabel(\"remote\")\n\t}\n\ttarget.Command, target.Commands = decodeCommands(s, args[cmdBuildRuleArgIdx])\n\tif test {\n\t\ttarget.Test = new(core.TestFields)\n\n\t\tif flaky := args[flakyBuildRuleArgIdx]; flaky != nil {\n\t\t\tif flaky == True {\n\t\t\t\ttarget.Test.Flakiness = defaultFlakiness\n\t\t\t\ttarget.AddLabel(\"flaky\") // Automatically label flaky tests\n\t\t\t} else if flaky == False {\n\t\t\t\ttarget.Test.Flakiness = 1\n\t\t\t} else if i, ok := flaky.(pyInt); ok {\n\t\t\t\tif int(i) <= 1 {\n\t\t\t\t\ttarget.Test.Flakiness = 1\n\t\t\t\t} else {\n\t\t\t\t\ttarget.Test.Flakiness = uint8(i)\n\t\t\t\t\ttarget.AddLabel(\"flaky\")\n\t\t\t\t}\n\t\t\t}\n\t\t} else {\n\t\t\ttarget.Test.Flakiness = 1\n\t\t}\n\t\tif testCmd != nil && testCmd != None {\n\t\t\ttarget.Test.Command, target.Test.Commands = decodeCommands(s, args[testCMDBuildRuleArgIdx])\n\t\t}\n\t\ttarget.Test.Timeout = sizeAndTimeout(s, size, args[testTimeoutBuildRuleArgIdx], s.state.Config.Test.Timeout)\n\t\ttarget.Test.Sandbox = isTruthy(testSandboxBuildRuleArgIdx)\n\t\ttarget.Test.NoOutput = isTruthy(noTestOutputBuildRuleArgIdx)\n\t}\n\n\tif err := validateSandbox(s.state, target); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tif s.state.Config.Build.Config == \"dbg\" {\n\t\ttarget.Debug = new(core.DebugFields)\n\t\ttarget.Debug.Command, _ = decodeCommands(s, args[debugCMDBuildRuleArgIdx])\n\t}\n\treturn target\n}", "func (s *Session) createTarget(header *ws.Header, frame io.Reader) (Target, error) {\n\tvar opBytes [indexBytesSize]byte\n\tif _, err := io.ReadFull(frame, opBytes[:]); err != nil {\n\t\treturn nil, err\n\t}\n\n\tws.Cipher(opBytes[:], header.Mask, 0)\n\theader.Mask = shiftCipher(header.Mask, indexBytesSize)\n\theader.Length -= indexBytesSize\n\tindex := int(binary.BigEndian.Uint16(opBytes[:]))\n\n\tif index == controlIndex {\n\t\treturn NewRPCTarget(s.readCopyBuffer, s), nil\n\t}\n\n\tcnx := s.GetConnection(index)\n\tif cnx == nil {\n\t\treturn nil, UnknownConnection\n\t}\n\n\treturn &ConnectionTarget{cnx}, nil\n}", "func (client BaseClient) CreateFeatureInstanceResponder(resp *http.Response) (result FeatureInstance, err error) {\n\terr = autorest.Respond(\n\t\tresp,\n\t\tclient.ByInspecting(),\n\t\tazure.WithErrorUnlessStatusCode(http.StatusOK),\n\t\tautorest.ByUnmarshallingJSON(&result),\n\t\tautorest.ByClosing())\n\tresult.Response = autorest.Response{Response: resp}\n\treturn\n}", "func NewCfnInstance_Override(c CfnInstance, scope awscdk.Construct, id *string, props *CfnInstanceProps) {\n\t_init_.Initialize()\n\n\t_jsii_.Create(\n\t\t\"monocdk.aws_opsworks.CfnInstance\",\n\t\t[]interface{}{scope, id, props},\n\t\tc,\n\t)\n}", "func (h *Adapter) CreateInstance(kubeconfig []byte, contextName string, ch *chan interface{}) error {\n\terr := h.validateKubeconfig(kubeconfig)\n\tif err != nil {\n\t\treturn ErrCreateInstance(err)\n\t}\n\n\terr = h.createKubeClient(kubeconfig)\n\tif err != nil {\n\t\treturn ErrCreateInstance(err)\n\t}\n\n\terr = h.createKubeconfig(kubeconfig)\n\tif err != nil {\n\t\treturn ErrCreateInstance(err)\n\t}\n\n\terr = h.createMesheryKubeclient(kubeconfig)\n\tif err != nil {\n\t\treturn ErrCreateInstance(err)\n\t}\n\n\th.ClientcmdConfig.CurrentContext = contextName\n\th.Channel = ch\n\n\treturn nil\n}", "func (awsI *Ec2Instance) CreateInstance(volumeSize int, zookeeperSize int, idleTimeout int, bastionVolSnapshotId string) error {\n\terr := awsI.runTerraformInit()\n\n\tif (bastionVolSnapshotId != \"\") {\n\t\tbastionVolData, err := Asset(\"etc/extras/bastion_volume.tf\")\n\t\tif err != nil {\n\t\t\tawsI.Ctx.ConsoleLog(0, \"Error while initializing terraform for the additional bastion volume: %s\\n\", err)\n\t\t\treturn nil\n\t\t}\n\t\tbastionVolumePath := path.Join(awsI.DeployDir, \"etc\", \"terraform\", \"instance\", \"bastion_volume.tf\")\n\t\tbastionVolContents := string(bastionVolData[:])\n\t\tsdutils.WriteFile(bastionVolumePath, fmt.Sprintf(bastionVolContents, bastionVolSnapshotId, awsI.PrivateKey))\n\t}\n\n\terr = awsI.runTerraformApply(volumeSize, zookeeperSize, \"0.0.0.0/32\", idleTimeout, \"Creating the instance VMs...\")\n\tif err != nil {\n\t\tawsI.Ctx.ConsoleLog(1, \"Failed to create the instance.\\n\")\n\t\treturn err\n\t}\n\tawsI.Ctx.ConsoleLog(1, \"Successfully created the instance.\\n\")\n\treturn nil\n}", "func NewTarget(dependencyMap map[string]*Dependencies, targetArg string, extendFlag bool) (target Target, err error) {\n\n\ttarget = Target{\n\t\tinitial: []string{},\n\t\tdependencies: []string{},\n\t}\n\n\tinitialTarget := cfg.ContainersForReference(targetArg)\n\tfor _, c := range initialTarget {\n\t\tif includes(allowed, c) {\n\t\t\ttarget.initial = append(target.initial, c)\n\t\t}\n\t}\n\n\tif extendFlag {\n\t\tvar (\n\t\t\tdependenciesSet = make(map[string]struct{})\n\t\t\tcascadingSeeds = []string{}\n\t\t)\n\t\t// start from the explicitly targeted target\n\t\tfor _, name := range target.initial {\n\t\t\tdependenciesSet[name] = struct{}{}\n\t\t\tcascadingSeeds = append(cascadingSeeds, name)\n\t\t}\n\n\t\t// Cascade until the dependency map has been fully traversed\n\t\t// according to the cascading flags.\n\t\tfor len(cascadingSeeds) > 0 {\n\t\t\tnextCascadingSeeds := []string{}\n\t\t\tfor _, seed := range cascadingSeeds {\n\t\t\t\tif dependencies, ok := dependencyMap[seed]; ok {\n\t\t\t\t\t// Queue direct dependencies if we haven't already considered them\n\t\t\t\t\tfor _, name := range dependencies.All {\n\t\t\t\t\t\tif _, alreadyIncluded := dependenciesSet[name]; !alreadyIncluded {\n\t\t\t\t\t\t\tdependenciesSet[name] = struct{}{}\n\t\t\t\t\t\t\tnextCascadingSeeds = append(nextCascadingSeeds, name)\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t\tcascadingSeeds = nextCascadingSeeds\n\t\t}\n\n\t\tfor name := range dependenciesSet {\n\t\t\tif !includes(target.initial, name) {\n\t\t\t\ttarget.dependencies = append(target.dependencies, name)\n\t\t\t}\n\t\t}\n\n\t\tsort.Strings(target.dependencies)\n\t}\n\n\treturn\n}", "func (i *InstanceServiceHandler) Create(ctx context.Context, instanceReq *InstanceCreateReq) (*Instance, error) {\n\turi := fmt.Sprintf(\"%s\", instancePath)\n\n\treq, err := i.client.NewRequest(ctx, http.MethodPost, uri, instanceReq)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tinstance := new(instanceBase)\n\tif err = i.client.DoWithContext(ctx, req, instance); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn instance.Instance, nil\n}", "func NewTarget(path string, kvstore store.Store, log *zap.Logger) *Target {\n\t// make sure we have a trailing slash for trimming future updates\n\tif !strings.HasSuffix(path, \"/\") {\n\t\tpath = path + \"/\"\n\t}\n\n\teventTypePathWatcher := NewWatcher(path+\"eventtypes\", kvstore, log)\n\tfunctionPathWatcher := NewWatcher(path+\"functions\", kvstore, log)\n\tsubscriptionPathWatcher := NewWatcher(path+\"subscriptions\", kvstore, log)\n\tcorsPathWatcher := NewWatcher(path+\"cors\", kvstore, log)\n\n\t// serves lookups for event types\n\teventTypeCache := newEventTypeCache(log)\n\t// serves lookups for function info\n\tfunctionCache := newFunctionCache(log)\n\t// serves lookups for which functions are subscribed to an event\n\tsubscriptionCache := newSubscriptionCache(log)\n\t// serves lookups for cors configuration\n\tcorsCache := newCORSCache(log)\n\n\t// start reacting to changes\n\tshutdown := make(chan struct{})\n\teventTypePathWatcher.React(eventTypeCache, shutdown)\n\tfunctionPathWatcher.React(functionCache, shutdown)\n\tsubscriptionPathWatcher.React(subscriptionCache, shutdown)\n\tcorsPathWatcher.React(corsCache, shutdown)\n\n\treturn &Target{\n\t\tlog: log,\n\t\tshutdown: shutdown,\n\t\teventTypeCache: eventTypeCache,\n\t\tfunctionCache: functionCache,\n\t\tsubscriptionCache: subscriptionCache,\n\t\tcorsCache: corsCache,\n\t}\n}", "func (c *FakeAWSSNSTargets) Create(ctx context.Context, aWSSNSTarget *v1alpha1.AWSSNSTarget, opts v1.CreateOptions) (result *v1alpha1.AWSSNSTarget, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewCreateAction(awssnstargetsResource, c.ns, aWSSNSTarget), &v1alpha1.AWSSNSTarget{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\treturn obj.(*v1alpha1.AWSSNSTarget), err\n}", "func (px *Paxos) CreateInstance(seq int) {\n\n\t//create instance and update max if necessary\n\n\tpx.instances[seq] = &Instance{-1, -1, nil, false}\n\tif seq > px.maxSeq {\n\t\tpx.maxSeq = seq\n\t}\n}", "func (vp *scalewayProvider) CreateInstance(log *logging.Logger, options providers.CreateInstanceOptions, dnsProvider providers.DnsProvider) (providers.ClusterInstance, error) {\n\t// Create server\n\tid, err := vp.createServer(options)\n\tif err != nil {\n\t\treturn providers.ClusterInstance{}, maskAny(err)\n\t}\n\n\t// Wait for the server to be active\n\tserver, err := vp.waitUntilServerActive(id, false)\n\tif err != nil {\n\t\treturn providers.ClusterInstance{}, maskAny(err)\n\t}\n\n\tif options.RoleLoadBalancer {\n\t\tpublicIpv4 := server.PublicAddress.IP\n\t\tpublicIpv6 := \"\"\n\t\tif err := providers.RegisterInstance(vp.Logger, dnsProvider, options, server.Name, options.RoleLoadBalancer, publicIpv4, publicIpv6); err != nil {\n\t\t\treturn providers.ClusterInstance{}, maskAny(err)\n\t\t}\n\t}\n\n\tvp.Logger.Infof(\"Server '%s' is ready\", server.Name)\n\n\treturn vp.clusterInstance(server, false), nil\n}", "func (c *TestClient) GetTargetInstance(project, zone, name string) (*compute.TargetInstance, error) {\n\tif c.GetTargetInstanceFn != nil {\n\t\treturn c.GetTargetInstanceFn(project, zone, name)\n\t}\n\treturn c.client.GetTargetInstance(project, zone, name)\n}", "func (s *TargetCRUD) Create(arg ...crud.Arg) (crud.Arg, error) {\n\tevent := eventFromArg(arg[0])\n\ttarget := targetFromStuct(event)\n\tprint.CreatePrintln(\"creating target\", *target.Target.Target,\n\t\t\"on upstream\", *target.Upstream.ID)\n\treturn target, nil\n}", "func (r *ReconcileCanary) CreateServiceForTargetRef(instance *kharonv1alpha1.Canary) (*corev1.Service, error) {\n\t// We have to check if there is a Service called as the TargetRef.Name, otherwise create it\n\ttargetService := &corev1.Service{}\n\terr := r.client.Get(context.TODO(), types.NamespacedName{Name: instance.Spec.TargetRef.Name, Namespace: instance.Namespace}, targetService)\n\tif err != nil && errors.IsNotFound(err) {\n\t\tportName := instance.Spec.TargetRefContainerPort.StrVal\n\t\tif len(portName) <= 0 {\n\t\t\tportName = fmt.Sprintf(\"%d-%s\", instance.Spec.TargetRefContainerPort.IntVal, strings.ToLower(string(instance.Spec.TargetRefContainerProtocol)))\n\t\t}\n\t\t// The Service we need should be named as the Deployment because exposes the Deployment logic (as a canary)\n\t\ttargetServiceDef := &TargetServiceDef{\n\t\t\tserviceName: instance.Spec.TargetRef.Name,\n\t\t\tnamespace: instance.Namespace,\n\t\t\tselector: instance.Spec.TargetRefSelector,\n\t\t\tportName: portName,\n\t\t\tprotocol: instance.Spec.TargetRefContainerProtocol,\n\t\t\tport: instance.Spec.TargetRefContainerPort.IntVal,\n\t\t\ttargetPort: instance.Spec.TargetRefContainerPort,\n\t\t}\n\t\ttargetService = newServiceFromTargetServiceDef(targetServiceDef)\n\t\t// Set Canary instance as the owner and controller\n\t\tif err := controllerutil.SetControllerReference(instance, targetService, r.scheme); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tlog.Info(\"Creating the canary service\", \"CanaryService.Namespace\", targetService.Namespace, \"CanaryService.Name\", targetService.Name)\n\t\terr = r.client.Create(context.TODO(), targetService)\n\t\tif err != nil && !errors.IsAlreadyExists(err) {\n\t\t\treturn nil, err\n\t\t}\n\t} else if err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn targetService, nil\n}", "func (c *TestClient) DeleteTargetInstance(project, zone, name string) error {\n\tif c.DeleteTargetInstanceFn != nil {\n\t\treturn c.DeleteTargetInstanceFn(project, zone, name)\n\t}\n\treturn c.client.DeleteTargetInstance(project, zone, name)\n}", "func (a *Client) CreateInstance(params *CreateInstanceParams) (*CreateInstanceOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewCreateInstanceParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"createInstance\",\n\t\tMethod: \"POST\",\n\t\tPathPattern: \"/instances\",\n\t\tProducesMediaTypes: []string{\"\"},\n\t\tConsumesMediaTypes: []string{\"\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &CreateInstanceReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*CreateInstanceOK), nil\n\n}", "func (ts *TargetSyncer) NewTarget(session sarama.ConsumerGroupSession, claim sarama.ConsumerGroupClaim) (RunnableTarget, error) {\n\tdiscoveredLabels := model.LabelSet{\n\t\t\"__meta_kafka_topic\": model.LabelValue(claim.Topic()),\n\t\t\"__meta_kafka_partition\": model.LabelValue(fmt.Sprintf(\"%d\", claim.Partition())),\n\t\t\"__meta_kafka_member_id\": model.LabelValue(session.MemberID()),\n\t\t\"__meta_kafka_group_id\": model.LabelValue(ts.cfg.KafkaConfig.GroupID),\n\t}\n\tdetails := newDetails(session, claim)\n\tlabelMap := make(map[string]string)\n\tfor k, v := range discoveredLabels.Clone().Merge(ts.cfg.KafkaConfig.Labels) {\n\t\tlabelMap[string(k)] = string(v)\n\t}\n\tlabelOut := format(labels.FromMap(labelMap), ts.cfg.RelabelConfigs)\n\tif len(labelOut) == 0 {\n\t\tlevel.Warn(ts.logger).Log(\"msg\", \"dropping target\", \"reason\", \"no labels\", \"details\", details, \"discovered_labels\", discoveredLabels.String())\n\t\treturn &runnableDroppedTarget{\n\t\t\tTarget: target.NewDroppedTarget(\"dropping target, no labels\", discoveredLabels),\n\t\t\trunFn: func() {\n\t\t\t\tfor range claim.Messages() {\n\t\t\t\t}\n\t\t\t},\n\t\t}, nil\n\t}\n\tt := NewKafkaTarget(\n\t\tts.logger,\n\t\tsession,\n\t\tclaim,\n\t\tdiscoveredLabels,\n\t\tlabelOut,\n\t\tts.cfg.RelabelConfigs,\n\t\tts.client,\n\t\tts.cfg.KafkaConfig.UseIncomingTimestamp,\n\t\tts.messageParser,\n\t)\n\n\treturn t, nil\n}", "func (client BaseClient) CreateFeatureInstanceSender(req *http.Request) (*http.Response, error) {\n\treturn autorest.SendWithSender(client, req,\n\t\tautorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n}", "func (c *TestClient) CreateInstance(project, zone string, i *compute.Instance) error {\n\tif c.CreateInstanceFn != nil {\n\t\treturn c.CreateInstanceFn(project, zone, i)\n\t}\n\treturn c.client.CreateInstance(project, zone, i)\n}", "func (h *Adapter) CreateInstance(kubeconfig []byte, contextName string, ch *chan interface{}) error {\n\terr := h.validateKubeconfig(kubeconfig)\n\tif err != nil {\n\t\treturn ErrCreateInstance(err)\n\t}\n\n\terr = h.createKubeconfig(kubeconfig)\n\tif err != nil {\n\t\treturn ErrCreateInstance(err)\n\t}\n\n\th.MesheryKubeclient, err = mesherykube.New(kubeconfig)\n\tif err != nil {\n\t\treturn ErrClientSet(err)\n\t}\n\n\th.DynamicKubeClient = h.MesheryKubeclient.DynamicKubeClient\n\th.RestConfig = h.MesheryKubeclient.RestConfig\n\n\th.KubeClient, err = kubernetes.NewForConfig(&h.RestConfig)\n\tif err != nil {\n\t\treturn ErrClientSet(err)\n\t}\n\n\th.ClientcmdConfig.CurrentContext = contextName\n\th.Channel = ch\n\n\treturn nil\n}", "func (c *MockFileStorageClient) CreateMountTarget(ctx context.Context, details filestorage.CreateMountTargetDetails) (*filestorage.MountTarget, error) {\n\treturn nil, nil\n}", "func (d *driverMock) CreateInstance(ctx context.Context, publicKey string) (string, error) {\n\tif d.CreateInstanceErr != nil {\n\t\treturn \"\", d.CreateInstanceErr\n\t}\n\n\td.CreateInstanceID = \"ocid1...\"\n\n\treturn d.CreateInstanceID, nil\n}", "func CreateInstance(computeService *compute.Service) (*compute.Operation, error) {\n\tstartupMetadata := fmt.Sprintf(\"#! /bin/bash\\n \\n echo FILESHARE_MOUNT_PRIMARY=\\\"%s\\\" >> /root/.bashrc \\n echo FILESTORE_IP_PRIMARY=\\\"%s\\\" >> /root/.bashrc \\n echo FILESHARE_NAME_PRIMARY=\\\"%s\\\" >> /root/.bashrc \\n echo FILESHARE_MOUNT_SECONDARY=\\\"%s\\\" >> /root/.bashrc \\n echo FILESHARE_NAME_SECONDARY=\\\"%s\\\" >> /root/.bashrc \\n echo FILESTORE_IP_SECONDARY=\\\"%s\\\" >> /root/.bashrc \\n echo GCP_BUCKET_NAME=\\\"%s\\\" >> /root/.bashrc \\n source /root/.bashrc \\n wget -O /root/google-filestore-backup.sh https://raw.githubusercontent.com/sreesanpd/google-filestore-backup-cloud-function/master/resources/google-filestore-backup.sh \\n chmod +x /root/google-filestore-backup.sh \\n /bin/bash /root/google-filestore-backup.sh \\necho \\\"I am STARTING some work at $(date)\\\" | sudo tee -a $HOME/work.txt\", FileshareMountPrimary, FilestoreIpPrimary, FileshareNamePrimary, FileshareMountSecondary, FileshareNameSecondary, FilestoreIpSecondary, GcpBucketName)\n\tshutdownMetadata := fmt.Sprintf(\"#! /bin/bash\\n\\necho \\\"I am FINISHING some work on %s at $(date)\\\" | sudo tee -a $(HOME)/work.txt\", InstanceName)\n\n\tinstance := &compute.Instance{\n\t\tName: InstanceName,\n\t\tMachineType: fmt.Sprintf(\"zones/%s/machineTypes/f1-micro\", Zone),\n\t\tNetworkInterfaces: []*compute.NetworkInterface{\n\t\t\t{\n\t\t\t\tName: VPCName,\n\t\t\t\tSubnetwork: fmt.Sprintf(\"projects/%s/regions/%s/subnetworks/%s\", ProjectID, Region, SubnetName),\n\t\t\t\tAccessConfigs: []*compute.AccessConfig{\n\t\t\t\t\t{\n\t\t\t\t\t\tName: \"External NAT\",\n\t\t\t\t\t\tType: \"ONE_TO_ONE_NAT\",\n\t\t\t\t\t\tNetworkTier: \"PREMIUM\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\tScheduling: &compute.Scheduling{\n\t\t\tPreemptible: true,\n\t\t},\n\t\tDisks: []*compute.AttachedDisk{\n\t\t\t{\n\t\t\t\tBoot: true, // The first disk must be a boot disk.\n\t\t\t\tAutoDelete: true, //Optional\n\t\t\t\tMode: \"READ_WRITE\", //Mode should be READ_WRITE or READ_ONLY\n\t\t\t\tInterface: \"SCSI\", //SCSI or NVME - NVME only for SSDs\n\t\t\t\tInitializeParams: &compute.AttachedDiskInitializeParams{\n\t\t\t\t\tDiskName: \"worker-instance-boot-disk\",\n\t\t\t\t\tSourceImage: \"projects/debian-cloud/global/images/family/debian-9\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\tMetadata: &compute.Metadata{\n\t\t\tItems: []*compute.MetadataItems{\n\t\t\t\t{\n\t\t\t\t\tKey: \"startup-script\",\n\t\t\t\t\tValue: &startupMetadata,\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tKey: \"shutdown-script\",\n\t\t\t\t\tValue: &shutdownMetadata,\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\tServiceAccounts: []*compute.ServiceAccount{\n\t\t\t{\n\t\t\t\tEmail: \"filestorebackups123-gcs-sa@learn-gcp-173102.iam.gserviceaccount.com\",\n\t\t\t\tScopes: []string{\"https://www.googleapis.com/auth/cloud-platform\"},\n\t\t\t\t\n\t\t\t},\n\t\t\t},\n\n\t\t\n\t}\n\treturn computeService.Instances.Insert(ProjectID, Zone, instance).Do()\n}", "func CreateTestInstance(t *testing.T, uuid dvid.UUID, typename, name string, config dvid.Config) {\n\tconfig.Set(\"typename\", typename)\n\tconfig.Set(\"dataname\", name)\n\tjsonData, err := config.MarshalJSON()\n\tif err != nil {\n\t\tt.Fatalf(\"Unable to make JSON for instance creation: %v\\n\", config)\n\t}\n\tapiStr := fmt.Sprintf(\"%srepo/%s/instance\", WebAPIPath, uuid)\n\tTestHTTP(t, \"POST\", apiStr, bytes.NewBuffer(jsonData))\n}", "func (c *FakeCloudwatchEventTargets) Create(cloudwatchEventTarget *v1alpha1.CloudwatchEventTarget) (result *v1alpha1.CloudwatchEventTarget, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewCreateAction(cloudwatcheventtargetsResource, c.ns, cloudwatchEventTarget), &v1alpha1.CloudwatchEventTarget{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\treturn obj.(*v1alpha1.CloudwatchEventTarget), err\n}", "func (a *Client) CreateInstanceTransformationByObjectName2(params *CreateInstanceTransformationByObjectName2Params) (*CreateInstanceTransformationByObjectName2OK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewCreateInstanceTransformationByObjectName2Params()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"createInstanceTransformationByObjectName2\",\n\t\tMethod: \"POST\",\n\t\tPathPattern: \"/instances/{id}/transformations/{objectName}\",\n\t\tProducesMediaTypes: []string{\"\"},\n\t\tConsumesMediaTypes: []string{\"\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &CreateInstanceTransformationByObjectName2Reader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*CreateInstanceTransformationByObjectName2OK), nil\n\n}", "func (c *Client) CreateInstance(displayName, availabilityDomain, compartmentID, nodeShape, nodeImageName, nodeSubnetID, sshUser, authorizedKeys string, nodeOCPUs, nodeMemoryInGBs int) (string, error) {\n\n\treq := identity.ListAvailabilityDomainsRequest{}\n\treq.CompartmentId = &compartmentID\n\tads, err := c.identityClient.ListAvailabilityDomains(context.Background(), req)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Just in case shortened or lower-case availability domain name was used\n\tlog.Debugf(\"Resolving availability domain from %s\", availabilityDomain)\n\tfor _, ad := range ads.Items {\n\t\tif strings.Contains(*ad.Name, strings.ToUpper(availabilityDomain)) {\n\t\t\tlog.Debugf(\"Availability domain %s\", *ad.Name)\n\t\t\tavailabilityDomain = *ad.Name\n\t\t}\n\t}\n\n\timageID, err := c.getImageID(compartmentID, nodeImageName)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Create the launch compute instance request\n\trequest := core.LaunchInstanceRequest{\n\t\tLaunchInstanceDetails: core.LaunchInstanceDetails{\n\t\t\tAvailabilityDomain: &availabilityDomain,\n\t\t\tCompartmentId: &compartmentID,\n\t\t\tShape: &nodeShape,\n\t\t\tCreateVnicDetails: &core.CreateVnicDetails{\n\t\t\t\tSubnetId: &nodeSubnetID,\n\t\t\t},\n\t\t\tDisplayName: &displayName,\n\t\t\tMetadata: map[string]string{\n\t\t\t\t\"ssh_authorized_keys\": authorizedKeys,\n\t\t\t\t\"user_data\": base64.StdEncoding.EncodeToString(createCloudInitScript(sshUser)),\n\t\t\t},\n\t\t\tSourceDetails: core.InstanceSourceViaImageDetails{\n\t\t\t\tImageId: imageID,\n\t\t\t},\n\t\t},\n\t}\n\n\tif nodeOCPUs > 0 {\n\t\toCPUs := float32(nodeOCPUs)\n\t\tmemoryInGBs := float32(nodeMemoryInGBs)\n\n\t\tLaunchInstanceShapeConfigDetails := core.LaunchInstanceShapeConfigDetails{\n\t\t\tOcpus: &oCPUs,\n\t\t\tMemoryInGBs: &memoryInGBs,\n\t\t}\n\t\trequest.ShapeConfig = &LaunchInstanceShapeConfigDetails\n\t}\n\n\tlog.Debugf(\"Launching instance with cloud-init: %s\", string(createCloudInitScript(sshUser)))\n\n\tcreateResp, err := c.computeClient.LaunchInstance(context.Background(), request)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// wait until lifecycle status is Running\n\tpollUntilRunning := func(r common.OCIOperationResponse) bool {\n\t\tif converted, ok := r.Response.(core.GetInstanceResponse); ok {\n\t\t\treturn converted.LifecycleState != core.InstanceLifecycleStateRunning\n\t\t}\n\t\treturn true\n\t}\n\n\t// create get instance request with a retry policy which takes a function\n\t// to determine shouldRetry or not\n\tpollingGetRequest := core.GetInstanceRequest{\n\t\tInstanceId: createResp.Instance.Id,\n\t\tRequestMetadata: helpers.GetRequestMetadataWithCustomizedRetryPolicy(pollUntilRunning),\n\t}\n\n\tinstance, pollError := c.computeClient.GetInstance(context.Background(), pollingGetRequest)\n\tif pollError != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn *instance.Id, nil\n}", "func (client BaseClient) CreateFeatureInstancePreparer(ctx context.Context, body *FeatureInstanceInputs) (*http.Request, error) {\n\tpreparer := autorest.CreatePreparer(\n\t\tautorest.AsContentType(\"application/json; charset=utf-8\"),\n\t\tautorest.AsPost(),\n\t\tautorest.WithBaseURL(client.BaseURI),\n\t\tautorest.WithPath(\"/api/features/instances\"))\n\tif body != nil {\n\t\tpreparer = autorest.DecoratePreparer(preparer,\n\t\t\tautorest.WithJSON(body))\n\t}\n\treturn preparer.Prepare((&http.Request{}).WithContext(ctx))\n}", "func (a *BackendOptionsApiService) CreateLogTarget(ctx _context.Context, parentName string, parentType string, logTarget LogTarget, localVarOptionals *CreateLogTargetOpts) (LogTarget, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodPost\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue LogTarget\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/services/haproxy/configuration/log_targets\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tlocalVarQueryParams.Add(\"parent_name\", parameterToString(parentName, \"\"))\n\tlocalVarQueryParams.Add(\"parent_type\", parameterToString(parentType, \"\"))\n\tif localVarOptionals != nil && localVarOptionals.TransactionId.IsSet() {\n\t\tlocalVarQueryParams.Add(\"transaction_id\", parameterToString(localVarOptionals.TransactionId.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Version.IsSet() {\n\t\tlocalVarQueryParams.Add(\"version\", parameterToString(localVarOptionals.Version.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.ForceReload.IsSet() {\n\t\tlocalVarQueryParams.Add(\"force_reload\", parameterToString(localVarOptionals.ForceReload.Value(), \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{\"application/json\"}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &logTarget\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 201 {\n\t\t\tvar v LogTarget\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 202 {\n\t\t\tvar v LogTarget\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 400 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 409 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tvar v ModelError\n\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\tif err != nil {\n\t\t\tnewErr.error = err.Error()\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\tnewErr.model = v\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func (client BaseClient) CreateFeatureInstance(ctx context.Context, body *FeatureInstanceInputs) (result FeatureInstance, err error) {\n\tif err := validation.Validate([]validation.Validation{\n\t\t{TargetValue: body,\n\t\t\tConstraints: []validation.Constraint{{Target: \"body\", Name: validation.Null, Rule: false,\n\t\t\t\tChain: []validation.Constraint{{Target: \"body.FeatureName\", Name: validation.Null, Rule: true,\n\t\t\t\t\tChain: []validation.Constraint{{Target: \"body.FeatureName\", Name: validation.Pattern, Rule: `^[a-z0-9-]+$`, Chain: nil}}},\n\t\t\t\t\t{Target: \"body.FeatureVersion\", Name: validation.Null, Rule: true,\n\t\t\t\t\t\tChain: []validation.Constraint{{Target: \"body.FeatureVersion\", Name: validation.Pattern, Rule: `^v?((\\d+)\\.(\\d+)\\.(\\d+))(?:-([\\dA-Za-z\\-]+(?:\\.[\\dA-Za-z\\-]+)*))?(?:\\+([\\dA-Za-z\\-]+(?:\\.[\\dA-Za-z\\-]+)*))?$`, Chain: nil}}},\n\t\t\t\t\t{Target: \"body.InstanceName\", Name: validation.Null, Rule: true,\n\t\t\t\t\t\tChain: []validation.Constraint{{Target: \"body.InstanceName\", Name: validation.Pattern, Rule: `^[a-z0-9-]+$`, Chain: nil}}},\n\t\t\t\t}}}}}); err != nil {\n\t\treturn result, validation.NewError(\"beacon.BaseClient\", \"CreateFeatureInstance\", err.Error())\n\t}\n\n\treq, err := client.CreateFeatureInstancePreparer(ctx, body)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"CreateFeatureInstance\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresp, err := client.CreateFeatureInstanceSender(req)\n\tif err != nil {\n\t\tresult.Response = autorest.Response{Response: resp}\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"CreateFeatureInstance\", resp, \"Failure sending request\")\n\t\treturn\n\t}\n\n\tresult, err = client.CreateFeatureInstanceResponder(resp)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"CreateFeatureInstance\", resp, \"Failure responding to request\")\n\t}\n\n\treturn\n}", "func NewTarget(ctx context.Context, envAcc pkgadapter.EnvConfigAccessor, ceClient cloudevents.Client) pkgadapter.Adapter {\n\tenv := envAcc.(*envAccessor)\n\tlogger := logging.FromContext(ctx)\n\tmetrics.MustRegisterEventProcessingStatsView()\n\n\tclient, err := mongo.Connect(ctx, options.Client().ApplyURI(env.ServerURL))\n\tif err != nil {\n\t\tlogger.Panicw(\"error connecting to mongodb\", zap.Error(err))\n\t\treturn nil\n\t}\n\n\treplier, err := targetce.New(env.Component, logger.Named(\"replier\"),\n\t\ttargetce.ReplierWithStatefulHeaders(env.BridgeIdentifier),\n\t\ttargetce.ReplierWithStaticResponseType(v1alpha1.EventTypeMongoDBStaticResponse),\n\t\ttargetce.ReplierWithPayloadPolicy(targetce.PayloadPolicy(env.CloudEventPayloadPolicy)))\n\tif err != nil {\n\t\tlogger.Panicf(\"Error creating CloudEvents replier: %v\", err)\n\t}\n\n\tmt := &pkgadapter.MetricTag{\n\t\tResourceGroup: targets.MongoDBTargetResource.String(),\n\t\tNamespace: envAcc.GetNamespace(),\n\t\tName: envAcc.GetName(),\n\t}\n\n\treturn &adapter{\n\t\tmclient: client,\n\t\tdefaultDatabase: env.DefaultDatabase,\n\t\tdefaultCollection: env.DefaultCollection,\n\n\t\treplier: replier,\n\t\tceClient: ceClient,\n\t\tlogger: logger,\n\t\tsr: metrics.MustNewEventProcessingStatsReporter(mt),\n\t}\n}", "func (c *Client) CreateInstance(name, zone, machineType, ipxeURL string) error {\n\treq := &packngo.DeviceCreateRequest{\n\t\tHostname: name,\n\t\tPlan: machineType,\n\t\tProjectID: c.projectID,\n\t\tFacility: []string{zone},\n\t\tIPXEScriptURL: ipxeURL,\n\t\tOS: \"custom_ipxe\",\n\t\tDescription: \"eden test vm\",\n\t\tBillingCycle: \"hourly\",\n\t}\n\t_, _, err := c.client.Devices.Create(req)\n\treturn err\n}", "func (h *handler) CreateInstance(kubeconfig []byte, contextName string, ch *chan interface{}) error {\n\n\tvar err error\n\th.channel = ch\n\th.kubeConfigPath = \"/Users/abishekk/.kube/config\"\n\t// h.kubeConfigPath, err = h.config.GetKey(\"kube-config-path\")\n\t// if err != nil {\n\t// \treturn ErrClientConfig(err)\n\t// }\n\n\tconfig, err := h.clientConfig(kubeconfig, contextName)\n\tif err != nil {\n\t\treturn ErrClientConfig(err)\n\t}\n\n\t// creates the clientset\n\tclientset, err := kubernetes.NewForConfig(config)\n\tif err != nil {\n\t\treturn ErrClientSet(err)\n\t}\n\n\th.kubeClient = clientset\n\n\treturn nil\n}", "func (s *stepCreateInstance) Run(ctx context.Context, state multistep.StateBag) multistep.StepAction {\n\tclient := state.Get(\"client\").(*civogo.Client)\n\tui := state.Get(\"ui\").(packer.Ui)\n\tc := state.Get(\"config\").(*Config)\n\tsshKeyID := state.Get(\"ssh_key_id\").(string)\n\n\t// Create the instance based on configuration\n\tui.Say(\"Creating instance...\")\n\n\ttemplate, err := client.FindTemplate(c.Template)\n\tif err != nil {\n\t\tui.Error(err.Error())\n\t}\n\n\tnetwork, _ := client.GetDefaultNetwork()\n\n\tInstanceConfig := &civogo.InstanceConfig{\n\t\tHostname: c.InstanceName,\n\t\tPublicIPRequired: c.PublicNetworking,\n\t\tRegion: c.Region,\n\t\tNetworkID: network.ID,\n\t\tInitialUser: c.Comm.SSHUsername,\n\t\tSize: c.Size,\n\t\tTemplateID: template.ID,\n\t\tSSHKeyID: sshKeyID,\n\t}\n\n\tlog.Printf(\"[DEBUG] Instance create paramaters: %+v\", InstanceConfig)\n\n\tinstance, err := client.CreateInstance(InstanceConfig)\n\tif err != nil {\n\t\terr := fmt.Errorf(\"Error creating instance: %s\", err)\n\t\tstate.Put(\"error\", err)\n\t\tui.Error(err.Error())\n\t\treturn multistep.ActionHalt\n\t}\n\n\t// We use this in cleanup\n\ts.instanceID = instance.ID\n\n\t// Store the instance id for later\n\tstate.Put(\"instance_id\", instance.ID)\n\n\treturn multistep.ActionContinue\n}", "func New(ctx resource.Context, cfg echo.Config) (i echo.Instance, err error) {\n\terr = resource.UnsupportedEnvironment(ctx.Environment())\n\n\tctx.Environment().Case(environment.Native, func() {\n\t\ti, err = native.New(ctx, cfg)\n\t})\n\n\tctx.Environment().Case(environment.Kube, func() {\n\t\ti, err = kube.New(ctx, cfg)\n\t})\n\treturn\n}", "func (g GCPClient) CreateInstance(name, image, zone, machineType string, disks Disks, data *string, nested, vtpm, replace bool) error {\n\tif replace {\n\t\tif err := g.DeleteInstance(name, zone, true); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tlog.Infof(\"Creating instance %s from image %s (type: %s in %s)\", name, image, machineType, zone)\n\n\tenabled := new(string)\n\t*enabled = \"1\"\n\n\tk, err := ssh.NewPublicKey(g.privKey.Public())\n\tif err != nil {\n\t\treturn err\n\t}\n\tsshKey := new(string)\n\t*sshKey = fmt.Sprintf(\"moby:%s moby\", string(ssh.MarshalAuthorizedKey(k)))\n\n\t// check provided image to be compatible with provided options\n\top, err := g.compute.Images.Get(g.projectName, image).Do()\n\tif err != nil {\n\t\treturn err\n\t}\n\tuefiCompatible := false\n\tfor _, feature := range op.GuestOsFeatures {\n\t\tif feature != nil && feature.Type == uefiCompatibleFeature {\n\t\t\tuefiCompatible = true\n\t\t\tbreak\n\t\t}\n\t}\n\tif vtpm && !uefiCompatible {\n\t\treturn fmt.Errorf(\"cannot use vTPM without UEFI_COMPATIBLE image\")\n\t}\n\t// we should check for nested\n\tvmxLicense := false\n\tfor _, license := range op.Licenses {\n\t\t// we omit hostname and version when define license\n\t\tif strings.HasSuffix(license, vmxImageLicence) {\n\t\t\tvmxLicense = true\n\t\t\tbreak\n\t\t}\n\t}\n\tif nested && !vmxLicense {\n\t\treturn fmt.Errorf(\"cannot use nested virtualization without enable-vmx image\")\n\t}\n\n\tinstanceDisks := []*compute.AttachedDisk{\n\t\t{\n\t\t\tAutoDelete: true,\n\t\t\tBoot: true,\n\t\t\tInitializeParams: &compute.AttachedDiskInitializeParams{\n\t\t\t\tSourceImage: fmt.Sprintf(\"global/images/%s\", image),\n\t\t\t},\n\t\t},\n\t}\n\n\tfor i, disk := range disks {\n\t\tvar diskName string\n\t\tif disk.Path != \"\" {\n\t\t\tdiskName = disk.Path\n\t\t} else {\n\t\t\tdiskName = fmt.Sprintf(\"%s-disk-%d\", name, i)\n\t\t}\n\t\tvar diskSizeGb int64\n\t\tif disk.Size == 0 {\n\t\t\tdiskSizeGb = int64(1)\n\t\t} else {\n\t\t\tdiskSizeGb = int64(convertMBtoGB(disk.Size))\n\t\t}\n\t\tdiskObj := &compute.Disk{Name: diskName, SizeGb: diskSizeGb}\n\t\tif vtpm {\n\t\t\tdiskObj.GuestOsFeatures = []*compute.GuestOsFeature{\n\t\t\t\t{Type: uefiCompatibleFeature},\n\t\t\t}\n\t\t}\n\t\tdiskOp, err := g.compute.Disks.Insert(g.projectName, zone, diskObj).Do()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif err := g.pollZoneOperationStatus(diskOp.Name, zone); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tinstanceDisks = append(instanceDisks, &compute.AttachedDisk{\n\t\t\tAutoDelete: true,\n\t\t\tBoot: false,\n\t\t\tSource: fmt.Sprintf(\"zones/%s/disks/%s\", zone, diskName),\n\t\t})\n\t}\n\n\tinstanceObj := &compute.Instance{\n\t\tMachineType: fmt.Sprintf(\"zones/%s/machineTypes/%s\", zone, machineType),\n\t\tName: name,\n\t\tDisks: instanceDisks,\n\t\tNetworkInterfaces: []*compute.NetworkInterface{\n\t\t\t{\n\t\t\t\tNetwork: \"global/networks/default\",\n\t\t\t\tAccessConfigs: []*compute.AccessConfig{\n\t\t\t\t\t{\n\t\t\t\t\t\tType: \"ONE_TO_ONE_NAT\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\tMetadata: &compute.Metadata{\n\t\t\tItems: []*compute.MetadataItems{\n\t\t\t\t{\n\t\t\t\t\tKey: \"serial-port-enable\",\n\t\t\t\t\tValue: enabled,\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tKey: \"ssh-keys\",\n\t\t\t\t\tValue: sshKey,\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tKey: \"user-data\",\n\t\t\t\t\tValue: data,\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\n\tif nested {\n\t\tinstanceObj.MinCpuPlatform = \"Intel Haswell\"\n\t}\n\tif vtpm {\n\t\tinstanceObj.ShieldedInstanceConfig = &compute.ShieldedInstanceConfig{EnableVtpm: true}\n\t}\n\n\t// Don't wait for operation to complete!\n\t// A headstart is needed as by the time we've polled for this event to be\n\t// completed, the instance may have already terminated\n\t_, err = g.compute.Instances.Insert(g.projectName, zone, instanceObj).Do()\n\tif err != nil {\n\t\treturn err\n\t}\n\tlog.Infof(\"Instance created\")\n\treturn nil\n}", "func CustomTarget(httpClient util.HTTPClient, baseURL string, tlsOptions TLSOptions, retryInterval time.Duration) Target {\n\treturn &customTarget{\n\t\ttargetType: TargetCustom,\n\t\tbaseURL: baseURL,\n\t\thttpClient: httpClient,\n\t\ttlsOptions: tlsOptions,\n\t\tretryInterval: retryInterval,\n\t}\n}", "func New(opts ...InstanceOpt) (*Instance, error) {\n\tinstance := &Instance{}\n\n\tfor _, opt := range opts {\n\t\tif err := opt(instance); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tpool, err := dockertest.NewPool(\"\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresource, err := pool.Run(\"localstack/localstack\", \"\", []string{instance.serviceString()})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\twithDefaults(instance)\n\tinstance.resolver = instance.makeResolver()\n\tinstance.pool = pool\n\tinstance.resource = resource\n\n\treturn instance, nil\n}", "func NewCfnInstance(scope awscdk.Construct, id *string, props *CfnInstanceProps) CfnInstance {\n\t_init_.Initialize()\n\n\tj := jsiiProxy_CfnInstance{}\n\n\t_jsii_.Create(\n\t\t\"monocdk.aws_opsworks.CfnInstance\",\n\t\t[]interface{}{scope, id, props},\n\t\t&j,\n\t)\n\n\treturn &j\n}", "func newTargetFailure(target Target, err error) TargetStatus {\n\terrs := []error{fmt.Errorf(\"%s: %s: %s\", errorNewTargetFailure, target.Type, err.Error())}\n\treturn TargetStatus{\n\t\tName: target.Name,\n\t\tErrors: errs,\n\t\tSteps: nil,\n\t}\n}", "func (f *Function) CreateInstanceSnapshot() {\n\tlogger := log.WithFields(log.Fields{\"fID\": f.fID})\n\n\tlogger.Debug(\"Creating instance snapshot\")\n\n\tctx, cancel := context.WithTimeout(context.Background(), time.Second*5)\n\tdefer cancel()\n\n\terr := orch.PauseVM(ctx, f.vmID)\n\tif err != nil {\n\t\tlog.Panic(err)\n\t}\n\n\terr = orch.CreateSnapshot(ctx, f.vmID)\n\tif err != nil {\n\t\tlog.Panic(err)\n\t}\n\n\t_, err = orch.ResumeVM(ctx, f.vmID)\n\tif err != nil {\n\t\tlog.Panic(err)\n\t}\n}", "func CreateInstance(clsid *GUID, iid *GUID) (unk *IUnknown, err error) {\n\tif iid == nil {\n\t\tiid = IID_IUnknown\n\t}\n\thr, _, _ := procCoCreateInstance.Call(\n\t\tuintptr(unsafe.Pointer(clsid)),\n\t\t0,\n\t\tCLSCTX_SERVER,\n\t\tuintptr(unsafe.Pointer(iid)),\n\t\tuintptr(unsafe.Pointer(&unk)))\n\tif hr != 0 {\n\t\terr = NewError(hr)\n\t}\n\treturn\n}", "func (obj *Device) CreateRenderTarget(\n\twidth uint,\n\theight uint,\n\tformat FORMAT,\n\tmultiSample MULTISAMPLE_TYPE,\n\tmultisampleQuality uint32,\n\tlockable bool,\n\tsharedHandle uintptr,\n) (*Surface, Error) {\n\tvar surface *Surface\n\tret, _, _ := syscall.Syscall9(\n\t\tobj.vtbl.CreateRenderTarget,\n\t\t9,\n\t\tuintptr(unsafe.Pointer(obj)),\n\t\tuintptr(width),\n\t\tuintptr(height),\n\t\tuintptr(format),\n\t\tuintptr(multiSample),\n\t\tuintptr(multisampleQuality),\n\t\tuintptrBool(lockable),\n\t\tuintptr(unsafe.Pointer(&surface)),\n\t\tsharedHandle,\n\t)\n\treturn surface, toErr(ret)\n}", "func NewInstance(ctx *pulumi.Context,\n\tname string, args *InstanceArgs, opts ...pulumi.ResourceOption) (*Instance, error) {\n\tif args == nil || args.Type == nil {\n\t\treturn nil, errors.New(\"missing required argument 'Type'\")\n\t}\n\tif args == nil {\n\t\targs = &InstanceArgs{}\n\t}\n\tvar resource Instance\n\terr := ctx.RegisterResource(\"gcp:datafusion/instance:Instance\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (t *Target) CreateApplyManifest(spec *TestSpec, base string) error {\n\tmanifestPath := t.GetManifestPath(spec, base)\n\tgetTemplate := func(tmpl string) (*bytes.Buffer, error) {\n\t\tmetadata := map[string]interface{}{\n\t\t\t\"spec\": spec,\n\t\t\t\"target\": t,\n\t\t\t\"targetName\": t.GetServiceName(spec),\n\t\t}\n\t\tt, err := template.New(\"\").Parse(tmpl)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tcontent := new(bytes.Buffer)\n\t\terr = t.Execute(content, metadata)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn content, nil\n\t}\n\n\tswitch t.Kind {\n\tcase service:\n\t\t// As default services are listen on port 80.\n\t\tt.PortNumber = 80\n\t\tservice := `{\n\t\t\"apiVersion\": \"v1\",\n\t\t\"kind\": \"Service\",\n\t\t\"metadata\": {\n\t\t\t\"name\": \"{{ .targetName }}\",\n\t\t\t\"labels\": {\n\t\t\t\t\"test\": \"policygen\"\n\t\t\t}\n\t\t},\n\t\t\"spec\": {\n\t\t\t\"ports\": [\n\t\t\t\t{ \"port\": {{ .target.PortNumber }} }\n\t\t\t],\n\t\t\t\"selector\": {\n\t\t\t\t\"id\": \"{{ .spec.DestPod }}\"\n\t\t\t}\n\t\t}}`\n\t\tdata, err := getTemplate(service)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"cannot render template: %s\", err)\n\t\t}\n\t\terr = spec.Kub.RenderTemplateToFile(t.GetManifestName(spec), data.String(), os.ModePerm)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\tcase nodePort:\n\t\tt.SetPortNumber()\n\t\tnodePort := `\n\t\t{\n\t\t \"apiVersion\": \"v1\",\n\t\t \"kind\": \"Service\",\n\t\t \"metadata\": {\n\t\t\t\"name\": \"{{ .targetName }}\",\n\t\t\t\"labels\": {\n\t\t\t\t\"test\": \"policygen\"\n\t\t\t}\n\t\t },\n\t\t \"spec\": {\n\t\t\t\"type\": \"NodePort\",\n\t\t\t\"ports\": [\n\t\t\t {\n\t\t\t\t\"targetPort\": 80,\n\t\t\t\t\"port\": {{ .target.PortNumber }},\n\t\t\t\t\"protocol\": \"TCP\"\n\t\t\t }\n\t\t\t],\n\t\t\t\"selector\": {\n\t\t\t \"id\": \"{{ .spec.DestPod }}\"\n\t\t\t}\n\t\t }\n\t\t}`\n\n\t\tdata, err := getTemplate(nodePort)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"cannot render template: %s\", err)\n\t\t}\n\n\t\terr = spec.Kub.RenderTemplateToFile(t.GetManifestName(spec), data.String(), os.ModePerm)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\tcase direct:\n\t\tt.PortNumber = 80\n\t\treturn nil\n\t}\n\tres := spec.Kub.ApplyDefault(manifestPath)\n\tif !res.WasSuccessful() {\n\t\treturn fmt.Errorf(\"%s\", res.CombineOutput())\n\t}\n\treturn nil\n}", "func NewCreateInstance(name string, plan string, team string) *CreateInstance {\n\tthis := CreateInstance{}\n\tthis.Name = name\n\tthis.Plan = plan\n\tthis.Team = team\n\treturn &this\n}", "func (h *HealthCheck) createInstance() error {\n\tif h.frameworkError != nil {\n\t\treturn h.frameworkError\n\t}\n\tglog.V(4).Info(\"Creating a ServiceInstance\")\n\tinstance := &v1beta1.ServiceInstance{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tName: h.instanceName,\n\t\t\tNamespace: h.namespace.Name,\n\t\t},\n\t\tSpec: v1beta1.ServiceInstanceSpec{\n\t\t\tPlanReference: v1beta1.PlanReference{\n\t\t\t\tClusterServiceClassExternalName: h.serviceclassName,\n\t\t\t\tClusterServicePlanExternalName: \"default\",\n\t\t\t},\n\t\t},\n\t}\n\toperationStartTime := time.Now()\n\tvar err error\n\tinstance, err = h.serviceCatalogClientSet.ServicecatalogV1beta1().ServiceInstances(h.namespace.Name).Create(instance)\n\tif err != nil {\n\t\treturn h.setError(\"error creating instance: %v\", err.Error())\n\t}\n\n\tif instance == nil {\n\t\treturn h.setError(\"error creating instance - instance is null\")\n\t}\n\n\tglog.V(4).Info(\"Waiting for ServiceInstance to be ready\")\n\terr = util.WaitForInstanceCondition(h.serviceCatalogClientSet.ServicecatalogV1beta1(),\n\t\th.namespace.Name,\n\t\th.instanceName,\n\t\tv1beta1.ServiceInstanceCondition{\n\t\t\tType: v1beta1.ServiceInstanceConditionReady,\n\t\t\tStatus: v1beta1.ConditionTrue,\n\t\t},\n\t)\n\tif err != nil {\n\t\treturn h.setError(\"instance not ready: %v\", err.Error())\n\t}\n\tReportOperationCompleted(\"create_instance\", operationStartTime)\n\n\tglog.V(4).Info(\"Verifing references are resolved\")\n\tsc, err := h.serviceCatalogClientSet.ServicecatalogV1beta1().ServiceInstances(h.namespace.Name).Get(h.instanceName, metav1.GetOptions{})\n\tif err != nil {\n\t\treturn h.setError(\"error getting instance: %v\", err.Error())\n\t}\n\n\tif sc.Spec.ClusterServiceClassRef == nil {\n\t\treturn h.setError(\"ClusterServiceClassRef should not be null\")\n\t}\n\tif sc.Spec.ClusterServicePlanRef == nil {\n\t\treturn h.setError(\"ClusterServicePlanRef should not be null\")\n\t}\n\n\tif strings.Compare(sc.Spec.ClusterServiceClassRef.Name, h.serviceclassID) != 0 {\n\t\treturn h.setError(\"ClusterServiceClassRef.Name error: %v != %v\", sc.Spec.ClusterServiceClassRef.Name, h.serviceclassID)\n\t}\n\tif strings.Compare(sc.Spec.ClusterServicePlanRef.Name, h.serviceplanID) != 0 {\n\t\treturn h.setError(\"sc.Spec.ClusterServicePlanRef.Name error: %v != %v\", sc.Spec.ClusterServicePlanRef.Name, h.serviceplanID)\n\t}\n\treturn nil\n}", "func (o *CloudTargetCreateParams) WithTimeout(timeout time.Duration) *CloudTargetCreateParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func (client *Client) CreateInstance(request *CreateInstanceRequest) (_result *CreateInstanceResponse, _err error) {\n\truntime := &util.RuntimeOptions{}\n\t_result = &CreateInstanceResponse{}\n\t_body, _err := client.CreateInstanceWithOptions(request, runtime)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\t_result = _body\n\treturn _result, _err\n}", "func expectedNewInstance(jobID, datasetID string) *dataset.NewInstance {\n\tnewInstance := &dataset.NewInstance{\n\t\tLinks: &dataset.Links{\n\t\t\tDataset: dataset.Link{\n\t\t\t\tURL: \"http://localhost:22000/datasets/\" + datasetID,\n\t\t\t\tID: datasetID,\n\t\t\t},\n\t\t\tJob: dataset.Link{\n\t\t\t\tURL: \"http://import-api/jobs/\" + jobID,\n\t\t\t\tID: jobID,\n\t\t\t},\n\t\t},\n\t\tDimensions: []dataset.CodeList{},\n\t\tImportTasks: &dataset.InstanceImportTasks{\n\t\t\tImportObservations: &dataset.ImportObservationsTask{\n\t\t\t\tState: dataset.StateCreated.String(),\n\t\t\t},\n\t\t\tBuildHierarchyTasks: []*dataset.BuildHierarchyTask{},\n\t\t\tBuildSearchIndexTasks: []*dataset.BuildSearchIndexTask{},\n\t\t},\n\t\tType: \"cantabular_blob\",\n\t}\n\tif datasetID == \"dataset1\" {\n\t\tnewInstance.Dimensions = []dataset.CodeList{{ID: \"codelist11\"}, {ID: \"codelist12\"}}\n\t\tnewInstance.LowestGeography = \"lowest_geo\"\n\t} else if datasetID == \"dataset2\" {\n\t\tnewInstance.Dimensions = []dataset.CodeList{{ID: \"codelist21\"}, {ID: \"codelist22\"}, {ID: \"codelist23\"}}\n\t}\n\treturn newInstance\n}", "func (p *EC2Provisioner) CreateInstance(opts EC2CreateInstanceOptions) (*cfg.Remote, error) {\n\t// Set requested region\n\tp.WithRegion(opts.Region)\n\n\t// set highlighter\n\tvar highlight = out.NewColorer(out.CY)\n\n\t// Generate authentication\n\tvar keyName = fmt.Sprintf(\"%s_%s_inertia_key_%d\", opts.Name, p.user, time.Now().UnixNano())\n\tout.Fprintf(p.out, highlight.Sf(\":key: Generating key pair '%s'...\\n\", keyName))\n\tkeyResp, err := p.client.CreateKeyPair(&ec2.CreateKeyPairInput{\n\t\tKeyName: aws.String(keyName),\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\thomeDir, err := os.UserHomeDir()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Save key\n\tvar keyPath = filepath.Join(homeDir, \".ssh\", *keyResp.KeyName)\n\tout.Fprintf(p.out, highlight.Sf(\":inbox_tray: Saving key to '%s'...\\n\", keyPath))\n\tif err = local.SaveKey(*keyResp.KeyMaterial, keyPath); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Create security group for network configuration\n\tvar secGroup = fmt.Sprintf(\"%s-%d\", opts.Name, time.Now().UnixNano())\n\tout.Fprintf(p.out, highlight.Sf(\":circus_tent: Creating security group '%s'...\\n\", secGroup))\n\tgroup, err := p.client.CreateSecurityGroup(&ec2.CreateSecurityGroupInput{\n\t\tGroupName: aws.String(secGroup),\n\t\tDescription: aws.String(\n\t\t\tfmt.Sprintf(\"Rules for project %s on %s\", opts.ProjectName, opts.Name),\n\t\t),\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Set rules for ports\n\tout.Fprintf(p.out, highlight.Sf(\":electric_plug: Exposing ports '%s'...\\n\", secGroup))\n\tif err = p.exposePorts(*group.GroupId, opts.DaemonPort, opts.Ports); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Start up instance\n\tout.Fprintf(p.out, highlight.Sf(\":boat: Requesting instance '%s'...\\n\", secGroup))\n\trunResp, err := p.client.RunInstances(&ec2.RunInstancesInput{\n\t\tImageId: aws.String(opts.ImageID),\n\t\tInstanceType: aws.String(opts.InstanceType),\n\t\tMinCount: aws.Int64(1),\n\t\tMaxCount: aws.Int64(1),\n\n\t\t// Security options\n\t\tKeyName: keyResp.KeyName,\n\t\tSecurityGroupIds: []*string{group.GroupId},\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Check response validity\n\tif runResp.Instances == nil || len(runResp.Instances) == 0 {\n\t\treturn nil, errors.New(\"Unable to start instances: \" + runResp.String())\n\t}\n\tout.Fprintf(p.out, highlight.Sf(\"A %s instance has been provisioned\", opts.InstanceType))\n\n\t// Loop until intance is running\n\tvar instance ec2.Instance\n\tfor {\n\t\t// Wait briefly between checks\n\t\ttime.Sleep(3 * time.Second)\n\n\t\t// Request instance status\n\t\tout.Fprintf(p.out, \"Checking status of the requested instance...\\n\")\n\t\tresult, err := p.client.DescribeInstances(&ec2.DescribeInstancesInput{\n\t\t\tInstanceIds: []*string{runResp.Instances[0].InstanceId},\n\t\t})\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\t// Check if reservations are present\n\t\tif result.Reservations == nil || len(result.Reservations) == 0 ||\n\t\t\tlen(result.Reservations[0].Instances) == 0 {\n\t\t\t// A reservation corresponds to a command to start instances\n\t\t\t// If nothing is here... we gotta keep waiting\n\t\t\tfmt.Fprintln(p.out, \"No reservations found yet.\")\n\t\t\tcontinue\n\t\t}\n\n\t\t// Get status\n\t\ts := result.Reservations[0].Instances[0].State\n\t\tif s == nil {\n\t\t\tfmt.Println(p.out, \"Status unknown.\")\n\t\t\tcontinue\n\t\t}\n\n\t\t// Code 16 means instance has started, and we can continue!\n\t\tif s.Code != nil && *s.Code == codeEC2InstanceStarted {\n\t\t\tfmt.Fprintln(p.out, \"Instance is running!\")\n\t\t\tinstance = *result.Reservations[0].Instances[0]\n\t\t\tbreak\n\t\t}\n\n\t\t// Otherwise, keep polling\n\t\tif s.Name != nil {\n\t\t\tfmt.Fprintln(p.out, \"Instance status: \"+*s.Name)\n\t\t} else {\n\t\t\tfmt.Fprintln(p.out, \"Instance status: \"+s.String())\n\t\t}\n\t\tcontinue\n\t}\n\n\t// Check instance validity\n\tif instance.PublicDnsName == nil {\n\t\treturn nil, errors.New(\"Unable to find public IP address for instance: \" + instance.String())\n\t}\n\n\t// Set tags\n\tout.Fprintf(p.out, \"Setting tags on instance...\\n\")\n\tif _, err = p.client.CreateTags(&ec2.CreateTagsInput{\n\t\tResources: []*string{instance.InstanceId},\n\t\tTags: []*ec2.Tag{\n\t\t\t{\n\t\t\t\tKey: aws.String(\"Name\"),\n\t\t\t\tValue: aws.String(opts.Name),\n\t\t\t},\n\t\t\t{\n\t\t\t\tKey: aws.String(\"Purpose\"),\n\t\t\t\tValue: aws.String(\"Inertia Continuous Deployment\"),\n\t\t\t},\n\t\t},\n\t}); err != nil {\n\t\tfmt.Fprintln(p.out, \"Failed to set tags: \"+err.Error())\n\t}\n\n\t// Poll for SSH port to open\n\tfmt.Fprintln(p.out, \"Waiting for ports to open...\")\n\tfor {\n\t\ttime.Sleep(3 * time.Second)\n\t\tfmt.Fprintln(p.out, \"Checking ports...\")\n\t\tif conn, err := net.Dial(\"tcp\", *instance.PublicDnsName+\":22\"); err == nil {\n\t\t\tfmt.Fprintln(p.out, \"Connection established!\")\n\t\t\tconn.Close()\n\t\t\tbreak\n\t\t}\n\t}\n\n\t// Generate webhook secret\n\tout.Fprintf(p.out, \"Generating a webhook secret...\\n\")\n\twebhookSecret, err := common.GenerateRandomString()\n\tif err != nil {\n\t\tfmt.Fprintln(p.out, err.Error())\n\t\tfmt.Fprintln(p.out, \"Using default secret 'inertia'\")\n\t\twebhookSecret = \"interia\"\n\t} else {\n\t\tfmt.Fprintf(p.out, \"Generated webhook secret: '%s'\\n\", webhookSecret)\n\t}\n\n\t// Return remote configuration\n\treturn &cfg.Remote{\n\t\tName: opts.Name,\n\t\tIP: *instance.PublicDnsName,\n\t\tSSH: &cfg.SSH{\n\t\t\tUser: p.user,\n\t\t\tIdentityFile: keyPath,\n\t\t\tSSHPort: \"22\",\n\t\t},\n\t\tDaemon: &cfg.Daemon{\n\t\t\tPort: strconv.FormatInt(opts.DaemonPort, 10),\n\t\t\tWebHookSecret: webhookSecret,\n\t\t},\n\t\tProfiles: make(map[string]string),\n\t}, nil\n}", "func doCreate(constructor func() base.IGameObject2D, isActive *bool) base.IGameObject2D {\r\n\tobj := constructor()\r\n\tobj.Obj().SetIGameObject2D(obj)\r\n\tapp.registerChannel <- resourceAccessRequest{\r\n\t\tpayload: obj,\r\n\t\tisActive: isActive,\r\n\t}\r\n\treturn obj\r\n}", "func (c *TestClient) CreateInstanceBeta(project, zone string, i *computeBeta.Instance) error {\n\tif c.CreateInstanceBetaFn != nil {\n\t\treturn c.CreateInstanceBetaFn(project, zone, i)\n\t}\n\treturn c.client.CreateInstanceBeta(project, zone, i)\n}", "func NewInstance(ctx *pulumi.Context,\n\tname string, args *InstanceArgs, opts ...pulumi.ResourceOption) (*Instance, error) {\n\tif args == nil || args.AvailabilityZone == nil {\n\t\treturn nil, errors.New(\"missing required argument 'AvailabilityZone'\")\n\t}\n\tif args == nil || args.BlueprintId == nil {\n\t\treturn nil, errors.New(\"missing required argument 'BlueprintId'\")\n\t}\n\tif args == nil || args.BundleId == nil {\n\t\treturn nil, errors.New(\"missing required argument 'BundleId'\")\n\t}\n\tif args == nil {\n\t\targs = &InstanceArgs{}\n\t}\n\tvar resource Instance\n\terr := ctx.RegisterResource(\"aws:lightsail/instance:Instance\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func NewInstance(projectID string, metaData map[string]string, secGroup []map[string]interface{}) openstack.InstanceClient {\n\treturn &Instance{\n\t\tprojectID: projectID,\n\t\tmetaData: metaData,\n\t\tsecGroup: secGroup,\n\t\tcreated: time.Now(),\n\t}\n}", "func (t *Target) Create(n usm.Number) usm.String {\n\treturn fmt.Sprintf(`make([]byte, %v)`, n)\n}", "func (client *Client) CreateTargetsWithOptions(request *CreateTargetsRequest, runtime *util.RuntimeOptions) (_result *CreateTargetsResponse, _err error) {\n\t_err = util.ValidateModel(request)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\t_result = &CreateTargetsResponse{}\n\t_body, _err := client.DoRequest(tea.String(\"createTargets\"), tea.String(\"HTTP\"), tea.String(\"POST\"), tea.String(\"/openapi/createTargets\"), nil, tea.ToMap(request), runtime)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\t_err = tea.Convert(_body, &_result)\n\treturn _result, _err\n}", "func CreateTaskReflection(handlerSymbol interface{}) TaskReflection {\n\thandlerType := reflect.TypeOf(handlerSymbol)\n\n\treturn TaskReflection{\n\t\tHandler: reflect.ValueOf(handlerSymbol),\n\t\tEventType: handlerType.In(1),\n\t}\n}", "func newTarget(path string, xt, yt int) *target {\n\treturn &target{path: path, xTiles: xt, yTiles: yt}\n}", "func (c *Client) CreateInstance(args *CreateInstanceArgs) (*CreateInstanceResult, error) {\n\tif len(args.AdminPass) > 0 {\n\t\tcryptedPass, err := Aes128EncryptUseSecreteKey(c.Config.Credentials.SecretAccessKey, args.AdminPass)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\targs.AdminPass = cryptedPass\n\t}\n\n\tif args.RootDiskSizeInGb <= 0 {\n\t\targs.RootDiskSizeInGb = 20\n\t}\n\n\tif args.PurchaseCount < 1 {\n\t\targs.PurchaseCount = 1\n\t}\n\n\tjsonBytes, jsonErr := json.Marshal(args)\n\tif jsonErr != nil {\n\t\treturn nil, jsonErr\n\t}\n\tbody, err := bce.NewBodyFromBytes(jsonBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn CreateInstance(c, args, body)\n}", "func (c *Client) CreateMountTarget(ctx context.Context, params *CreateMountTargetInput, optFns ...func(*Options)) (*CreateMountTargetOutput, error) {\n\tif params == nil {\n\t\tparams = &CreateMountTargetInput{}\n\t}\n\n\tresult, metadata, err := c.invokeOperation(ctx, \"CreateMountTarget\", params, optFns, c.addOperationCreateMountTargetMiddlewares)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tout := result.(*CreateMountTargetOutput)\n\tout.ResultMetadata = metadata\n\treturn out, nil\n}", "func (c *Client) CreateInstanceByLabel(args *CreateSpecialInstanceArgs) (*CreateInstanceResult, error) {\n\tif len(args.AdminPass) > 0 {\n\t\tcryptedPass, err := Aes128EncryptUseSecreteKey(c.Config.Credentials.SecretAccessKey, args.AdminPass)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\targs.AdminPass = cryptedPass\n\t}\n\n\tif args.RootDiskSizeInGb <= 0 {\n\t\targs.RootDiskSizeInGb = 20\n\t}\n\n\tif args.PurchaseCount < 1 {\n\t\targs.PurchaseCount = 1\n\t}\n\n\tjsonBytes, jsonErr := json.Marshal(args)\n\tif jsonErr != nil {\n\t\treturn nil, jsonErr\n\t}\n\tbody, err := bce.NewBodyFromBytes(jsonBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn CreateInstanceByLabel(c, args, body)\n}", "func CreateMountTarget(e *efs.EFS, i string, s string) (*efs.MountTargetDescription, error) {\n\tvar security []*string\n\n\t// Determine if we need to assign a security group to this mount point, otherwise defer\n\t// to the default group.\n\tif *cliSecurity != \"\" {\n\t\tsecurity = []*string{\n\t\t\tcliSecurity,\n\t\t}\n\t}\n\n\tparams := &efs.CreateMountTargetInput{\n\t\tFileSystemId: aws.String(i),\n\t\tSubnetId: aws.String(s),\n\t\tSecurityGroups: security,\n\t}\n\tresp, err := e.CreateMountTarget(params)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Wait for the mount point to become available.\n\tfor {\n\t\tmnt, err := DescribeMountTarget(e, i)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif len(mnt.MountTargets) > 0 {\n\t\t\tif *mnt.MountTargets[0].LifeCycleState == efsAvail {\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\ttime.Sleep(10 * time.Second)\n\t}\n\n\treturn resp, nil\n}", "func NewTargetService(fqdn string) TargetService {\n\treturn TargetService{fQDN: fqdn}\n}", "func (c *Client) CreateMountTarget(ctx context.Context, params *CreateMountTargetInput, optFns ...func(*Options)) (*CreateMountTargetOutput, error) {\n\tif params == nil {\n\t\tparams = &CreateMountTargetInput{}\n\t}\n\n\tresult, metadata, err := c.invokeOperation(ctx, \"CreateMountTarget\", params, optFns, addOperationCreateMountTargetMiddlewares)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tout := result.(*CreateMountTargetOutput)\n\tout.ResultMetadata = metadata\n\treturn out, nil\n}", "func (c *Client) CreateMountTarget(ctx context.Context, params *CreateMountTargetInput, optFns ...func(*Options)) (*CreateMountTargetOutput, error) {\n\tif params == nil {\n\t\tparams = &CreateMountTargetInput{}\n\t}\n\n\tresult, metadata, err := c.invokeOperation(ctx, \"CreateMountTarget\", params, optFns, addOperationCreateMountTargetMiddlewares)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tout := result.(*CreateMountTargetOutput)\n\tout.ResultMetadata = metadata\n\treturn out, nil\n}", "func (f *fakeLB) CreateForwardingRule(_ context.Context, _ string, _ *govultr.ForwardingRule) (*govultr.ForwardingRule, *http.Response, error) {\n\tpanic(\"implement me\")\n}", "func (s *API) CreateInstance(req *CreateInstanceRequest, opts ...scw.RequestOption) (*Instance, error) {\n\tvar err error\n\n\tdefaultProjectID, exist := s.client.GetDefaultProjectID()\n\tif exist && req.OrganizationID == nil && req.ProjectID == nil {\n\t\treq.ProjectID = &defaultProjectID\n\t}\n\n\tdefaultOrganizationID, exist := s.client.GetDefaultOrganizationID()\n\tif exist && req.OrganizationID == nil && req.ProjectID == nil {\n\t\treq.OrganizationID = &defaultOrganizationID\n\t}\n\n\tif req.Region == \"\" {\n\t\tdefaultRegion, _ := s.client.GetDefaultRegion()\n\t\treq.Region = defaultRegion\n\t}\n\n\tif req.Name == \"\" {\n\t\treq.Name = namegenerator.GetRandomName(\"ins\")\n\t}\n\n\tif fmt.Sprint(req.Region) == \"\" {\n\t\treturn nil, errors.New(\"field Region cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"POST\",\n\t\tPath: \"/rdb/v1/regions/\" + fmt.Sprint(req.Region) + \"/instances\",\n\t\tHeaders: http.Header{},\n\t}\n\n\terr = scwReq.SetBody(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar resp Instance\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func (o *CloudTargetCreateParams) WithHTTPClient(client *http.Client) *CloudTargetCreateParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func NewInstance(ctx *pulumi.Context,\n\tname string, args *InstanceArgs, opts ...pulumi.ResourceOption) (*Instance, error) {\n\tif args == nil {\n\t\targs = &InstanceArgs{}\n\t}\n\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Instance\n\terr := ctx.RegisterResource(\"gcp:looker/instance:Instance\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func NewSDKActor(execute func(string) string) SDKActor {\n sdkActor := SDKActor{}\n sdkActor.connector = newConnector()\n sdkActor.execute = execute\n return sdkActor\n}", "func NewEventTarget(ctx *pulumi.Context,\n\tname string, args *EventTargetArgs, opts ...pulumi.ResourceOption) (*EventTarget, error) {\n\tif args == nil || args.Arn == nil {\n\t\treturn nil, errors.New(\"missing required argument 'Arn'\")\n\t}\n\tif args == nil || args.Rule == nil {\n\t\treturn nil, errors.New(\"missing required argument 'Rule'\")\n\t}\n\tif args == nil {\n\t\targs = &EventTargetArgs{}\n\t}\n\tvar resource EventTarget\n\terr := ctx.RegisterResource(\"aws:cloudwatch/eventTarget:EventTarget\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func NewInstance(ctx *pulumi.Context,\n\tname string, args *InstanceArgs, opts ...pulumi.ResourceOption) (*Instance, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.Description == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'Description'\")\n\t}\n\tif args.InstanceSeries == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'InstanceSeries'\")\n\t}\n\tif args.Specification == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'Specification'\")\n\t}\n\tif args.VswitchId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'VswitchId'\")\n\t}\n\tif args.ZoneId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'ZoneId'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Instance\n\terr := ctx.RegisterResource(\"alicloud:drds/instance:Instance\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func NewInstance(ctx *pulumi.Context,\n\tname string, args *InstanceArgs, opts ...pulumi.ResourceOption) (*Instance, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.PaymentType == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'PaymentType'\")\n\t}\n\tif args.VersionCode == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'VersionCode'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Instance\n\terr := ctx.RegisterResource(\"alicloud:threatdetection/instance:Instance\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func TargetBuilder() *Target {\n\treturn &Target{\n\t\terr: nil,\n\t\tinfService: nil,\n\t\texp: nil,\n\t\tk8sclient: nil,\n\t\tretries: 18,\n\t\tinterval: 10,\n\t}\n}", "func NewInstance(ctx *pulumi.Context,\n\tname string, args *InstanceArgs, opts ...pulumi.ResourceOption) (*Instance, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.InstanceId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'InstanceId'\")\n\t}\n\tif args.MachineType == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'MachineType'\")\n\t}\n\treplaceOnChanges := pulumi.ReplaceOnChanges([]string{\n\t\t\"instanceId\",\n\t\t\"location\",\n\t\t\"project\",\n\t})\n\topts = append(opts, replaceOnChanges)\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Instance\n\terr := ctx.RegisterResource(\"google-native:notebooks/v1:Instance\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (a *AssembliesApiService) CreateInstance(ctx _context.Context, did string, wid string, eid string, btAssemblyInstanceDefinitionParams BtAssemblyInstanceDefinitionParams) ([]BtOccurrence74, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodPost\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue []BtOccurrence74\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/api/assemblies/d/{did}/w/{wid}/e/{eid}/instances\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"did\"+\"}\", _neturl.QueryEscape(parameterToString(did, \"\")) , -1)\n\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"wid\"+\"}\", _neturl.QueryEscape(parameterToString(wid, \"\")) , -1)\n\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"eid\"+\"}\", _neturl.QueryEscape(parameterToString(eid, \"\")) , -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{\"application/json;charset=UTF-8; qs=0.09\"}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/vnd.onshape.v1+json;charset=UTF-8;qs=0.1\", \"application/json;charset=UTF-8; qs=0.09\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &btAssemblyInstanceDefinitionParams\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 200 {\n\t\t\tvar v []BtOccurrence74\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func (cce *CCEClient) Create(ctx context.Context, cluster *clusterv1.Cluster, machine *clusterv1.Machine) error {\n\tglog.V(4).Infof(\"Create machine: %+v\", machine.Name)\n\tinstance, err := cce.instanceIfExists(cluster, machine)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif instance != nil {\n\t\tglog.Infof(\"Skipped creating a VM that already exists, instanceID %s\", instance.InstanceID)\n\t}\n\n\tmachineCfg, err := machineProviderFromProviderConfig(machine.Spec.ProviderSpec)\n\tif err != nil {\n\t\tglog.Errorf(\"parse machine config err: %s\", err.Error())\n\t\treturn err\n\t}\n\tglog.V(4).Infof(\"machine config: %+v\", machineCfg)\n\n\tbccArgs := &bcc.CreateInstanceArgs{\n\t\tName: machine.Name,\n\t\tImageID: machineCfg.ImageID, // ubuntu-16.04-amd64\n\t\tBilling: billing.Billing{\n\t\t\tPaymentTiming: \"Postpaid\",\n\t\t},\n\t\tCPUCount: machineCfg.CPUCount,\n\t\tMemoryCapacityInGB: machineCfg.MemoryCapacityInGB,\n\t\tAdminPass: machineCfg.AdminPass,\n\t\tPurchaseCount: 1,\n\t\tInstanceType: \"N3\", // Normal 3\n\t\tNetworkCapacityInMbps: 1, //EIP bandwidth\n\t}\n\n\t// TODO support different regions\n\tinstanceIDs, err := cce.computeService.Bcc().CreateInstances(bccArgs, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif len(instanceIDs) != 1 {\n\t\treturn fmt.Errorf(\"CreateVMError\")\n\t}\n\n\tglog.Infof(\"Created a new VM, instanceID %s\", instanceIDs[0])\n\tif machine.ObjectMeta.Annotations == nil {\n\t\tmachine.ObjectMeta.Annotations = map[string]string{}\n\t}\n\tif cluster.ObjectMeta.Annotations == nil {\n\t\tcluster.ObjectMeta.Annotations = map[string]string{}\n\t}\n\tmachine.ObjectMeta.Annotations[TagInstanceID] = instanceIDs[0]\n\tmachine.ObjectMeta.Annotations[TagInstanceStatus] = \"Created\"\n\tmachine.ObjectMeta.Annotations[TagInstanceAdminPass] = machineCfg.AdminPass\n\tmachine.ObjectMeta.Annotations[TagKubeletVersion] = machine.Spec.Versions.Kubelet\n\n\ttoken, err := cce.getKubeadmToken()\n\tif err != nil {\n\t\tglog.Errorf(\"getKubeadmToken err: %+v\", err)\n\t\treturn err\n\t}\n\n\tif machineCfg.Role == \"master\" {\n\t\tcluster.ObjectMeta.Annotations[TagMasterInstanceID] = instanceIDs[0]\n\t\tcluster.ObjectMeta.Annotations[TagClusterToken] = token\n\t\tmachine.ObjectMeta.Annotations[TagInstanceRole] = \"master\"\n\t} else {\n\t\tmachine.ObjectMeta.Annotations[TagInstanceRole] = \"node\"\n\t}\n\n\tglog.V(4).Infof(\"new machine: %+v, annotation %+v\", machine.Name, machine.Annotations)\n\tcce.client.Update(context.Background(), cluster)\n\tcce.client.Update(context.Background(), machine)\n\n\t// TODO rewrite\n\tgo cce.postCreate(ctx, cluster, machine)\n\treturn nil\n}", "func Create(constructor func() base.IGameObject2D) base.IGameObject2D {\r\n\treturn doCreate(constructor, infra.BoolPtr_True)\r\n}", "func CreateInstance(client daisyCompute.Client, project, zone string, i *api.Instance) (*Instance, error) {\n\tif err := client.CreateInstance(project, zone, i); err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Instance{Instance: i, client: client, Project: project, Zone: zone}, nil\n}", "func (c *Client) CreateSystemInstance(ctx context.Context, params *CreateSystemInstanceInput, optFns ...func(*Options)) (*CreateSystemInstanceOutput, error) {\n\tif params == nil {\n\t\tparams = &CreateSystemInstanceInput{}\n\t}\n\n\tresult, metadata, err := c.invokeOperation(ctx, \"CreateSystemInstance\", params, optFns, addOperationCreateSystemInstanceMiddlewares)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tout := result.(*CreateSystemInstanceOutput)\n\tout.ResultMetadata = metadata\n\treturn out, nil\n}", "func HandleInstanceCreate(w rest.ResponseWriter, r *rest.Request) {\n\t// get ima\n\tima := Ima{}\n\terr := r.DecodeJsonPayload(&ima)\n\tif err != nil {\n\t\trest.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\tif ima.Ima == \"\" {\n\t\trest.Error(w, \"ima required\", 400)\n\t\treturn\n\t}\n\tif ima.Mem == 0 {\n\t\trest.Error(w, \"memory required\", 400)\n\t\treturn\n\t}\n\tif ima.Cpu == 0 {\n\t\trest.Error(w, \"cpu required\", 400)\n\t\treturn\n\t}\n\n\t// start the instance\n\tos := getImaOs(ima.Ima)\n\tswitch os {\n\tcase \"freebsd\":\n\t\t// clone ima to instance\n\t\tinstanceid := allocateInstanceId()\n\t\tcloneIma(ima.Ima, instanceid)\n\n\t\t// create network interface and bring up\n\t\ttap := allocateTap()\n\t\tif tap == \"\" {\n\t\t\treturn\n\t\t}\n\t\tsaveTap(tap, instanceid)\n\t\tbridge := findBridge()\n\t\taddTapToBridge(tap, bridge)\n\t\tbridgeUp(bridge)\n\n\t\tnmdm := \"/dev/nmdm-\" + instanceid + \"-A\"\n\t\tsaveCpu(ima.Cpu, instanceid)\n\t\tsaveMem(ima.Mem, instanceid)\n\t\tgo startFreeBSDVM(nmdm, ima.Cpu, ima.Mem, tap, instanceid)\n\t\tw.WriteJson(&instanceid)\n\tcase \"linux\":\n\t\t// clone ima to instance\n\t\tinstanceid := allocateInstanceId()\n\t\tcloneIma(ima.Ima, instanceid)\n\n\t\t// create network interface and bring up\n\t\ttap := allocateTap()\n\t\tif tap == \"\" {\n\t\t\treturn\n\t\t}\n\t\tsaveTap(tap, instanceid)\n\t\tbridge := findBridge()\n\t\taddTapToBridge(tap, bridge)\n\t\tbridgeUp(bridge)\n\n\t\t//nmdm := \"/dev/nmdm-\" + instanceid + \"-A\"\n\t\tsaveCpu(ima.Cpu, instanceid)\n\t\tsaveMem(ima.Mem, instanceid)\n\t\tbhyveDestroy(instanceid)\n\t\tnmdm := \"/dev/nmdm-\" + instanceid + \"-A\"\n\t\tgo startLinuxVM(nmdm, ima.Cpu, ima.Mem, tap, instanceid)\n\t\tw.WriteJson(&instanceid)\n\tdefault:\n\t\trest.Error(w, \"unknown OS\", 400)\n\t}\n}", "func (s *DataStore) CreateBackupTarget(backupTarget *longhorn.BackupTarget) (*longhorn.BackupTarget, error) {\n\tret, err := s.lhClient.LonghornV1beta2().BackupTargets(s.namespace).Create(context.TODO(), backupTarget, metav1.CreateOptions{})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif SkipListerCheck {\n\t\treturn ret, nil\n\t}\n\n\tobj, err := verifyCreation(ret.Name, \"backup target\", func(name string) (runtime.Object, error) {\n\t\treturn s.GetBackupTargetRO(name)\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tret, ok := obj.(*longhorn.BackupTarget)\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"BUG: datastore: verifyCreation returned wrong type for BackupTarget\")\n\t}\n\treturn ret.DeepCopy(), nil\n}", "func (t *TargetToExtent) Create(server *Server) (*http.Response, error) {\n\tendpoint := \"/api/v1.0/services/iscsi/targettoextent/\"\n\tvar targetToExtent TargetToExtent\n\tvar e interface{}\n\tresp, err := server.getSlingConnection().Post(endpoint).BodyJSON(t).Receive(&targetToExtent, &e)\n\tif err != nil {\n\t\tglog.Warningln(err)\n\t\treturn resp, err\n\t}\n\n\tif resp.StatusCode != 201 {\n\t\tbody, _ := json.Marshal(e)\n\t\treturn resp, fmt.Errorf(\"Error creating TargetToExtent for %+v - message: %s, status: %d\", *t, string(body), resp.StatusCode)\n\t}\n\n\tt.CopyFrom(&targetToExtent)\n\n\treturn resp, nil\n}", "func (a *Agent) spawnInstance(ctx context.Context, c instance.Config) {\n\tinst, err := a.instanceFactory(a.cfg.Global, c, a.cfg.WALDir, a.logger)\n\tif err != nil {\n\t\tlevel.Error(a.logger).Log(\"msg\", \"failed to create instance\", \"err\", err)\n\t\treturn\n\t}\n\n\tfor {\n\t\terr = inst.Run(ctx)\n\t\tif err != nil && err != context.Canceled {\n\t\t\tinstanceAbnormalExits.WithLabelValues(c.Name).Inc()\n\t\t\tlevel.Error(a.logger).Log(\"msg\", \"instance stopped abnormally, restarting after backoff period\", \"err\", err, \"backoff\", a.cfg.InstanceRestartBackoff, \"instance\", c.Name)\n\t\t\ttime.Sleep(a.cfg.InstanceRestartBackoff)\n\t\t} else {\n\t\t\tlevel.Info(a.logger).Log(\"msg\", \"stopped instance\", \"instance\", c.Name)\n\t\t\tbreak\n\t\t}\n\t}\n}" ]
[ "0.6227713", "0.58265316", "0.5731458", "0.57192355", "0.566241", "0.56205326", "0.5602566", "0.55804175", "0.55354637", "0.54952204", "0.5468607", "0.5446037", "0.5443763", "0.53265125", "0.53243345", "0.5311738", "0.52985287", "0.52229476", "0.52093345", "0.5201521", "0.5199547", "0.51970685", "0.519303", "0.5192264", "0.51906615", "0.51882493", "0.5153226", "0.5124199", "0.5124096", "0.5122056", "0.51190805", "0.5116488", "0.51074857", "0.5101954", "0.5101877", "0.50772333", "0.5074724", "0.5059843", "0.5047617", "0.5035017", "0.5015301", "0.5009899", "0.5009494", "0.50031143", "0.5001115", "0.4997837", "0.4992233", "0.49801418", "0.4946369", "0.49415344", "0.49367616", "0.49011284", "0.49003488", "0.4899473", "0.48851672", "0.48779777", "0.48719746", "0.48647672", "0.48615715", "0.48460117", "0.48318326", "0.48237687", "0.48025224", "0.47885427", "0.47878027", "0.47796565", "0.47768158", "0.47735068", "0.47562137", "0.47559562", "0.47457618", "0.47404352", "0.47391543", "0.4734715", "0.473023", "0.47285095", "0.47271168", "0.4724966", "0.472412", "0.47222972", "0.47222972", "0.47221163", "0.4711888", "0.4707391", "0.46853557", "0.46833932", "0.46803856", "0.46779865", "0.46746188", "0.4656299", "0.46471885", "0.46461946", "0.46458006", "0.46434432", "0.46430963", "0.46426645", "0.46335825", "0.4629893", "0.4623473", "0.4622169" ]
0.7435546
0
StartInstance uses the override method StartInstanceFn or the real implementation.
func (c *TestClient) StartInstance(project, zone, name string) error { if c.StartInstanceFn != nil { return c.StartInstanceFn(project, zone, name) } return c.client.StartInstance(project, zone, name) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (p *OnPrem) StartInstance(ctx *Context, instancename string) error {\n\treturn fmt.Errorf(\"Operation not supported\")\n}", "func (p *OnPrem) StartInstance(ctx *Context, instancename string) error {\n\treturn fmt.Errorf(\"Operation not supported\")\n}", "func (p *ProxMox) StartInstance(ctx *lepton.Context, instanceID string) error {\n\n\treq, err := http.NewRequest(\"POST\", p.apiURL+\"/api2/json/nodes/\"+p.nodeNAME+\"/qemu/\"+instanceID+\"/status/start\", nil)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\ttr := &http.Transport{\n\t\tTLSClientConfig: &tls.Config{InsecureSkipVerify: true},\n\t}\n\tclient := &http.Client{Transport: tr}\n\n\treq.Header.Add(\"Authorization\", \"PVEAPIToken=\"+p.tokenID+\"=\"+p.secret)\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\t_, err = io.ReadAll(resp.Body)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (p *ProcessDefinition) StartInstance(by QueryProcessDefinitionBy, req ReqStartInstance) (processDefinition *ResStartedProcessDefinition, err error) {\n\tprocessDefinition = &ResStartedProcessDefinition{}\n\tres, err := p.client.doPostJson(\"/process-definition/\"+by.String()+\"/start\", map[string]string{}, &req)\n\tif err != nil {\n\t\treturn\n\t}\n\n\terr = p.client.readJsonResponse(res, processDefinition)\n\treturn\n}", "func (i *InstanceServiceHandler) Start(ctx context.Context, instanceID string) error {\n\turi := fmt.Sprintf(\"%s/%s/start\", instancePath, instanceID)\n\n\treq, err := i.client.NewRequest(ctx, http.MethodPost, uri, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn i.client.DoWithContext(ctx, req, nil)\n}", "func (f *FakeInstance) Start(_ context.Context, _ string) error {\n\tpanic(\"implement me\")\n}", "func (c *Client) StartInstance(id string) error {\n\n\tactionRequest := core.InstanceActionRequest{}\n\tactionRequest.Action = core.InstanceActionActionStart\n\tactionRequest.InstanceId = &id\n\n\tstartResp, err := c.computeClient.InstanceAction(context.Background(), actionRequest)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// wait until lifecycle status is Running\n\tpollUntilRunning := func(r common.OCIOperationResponse) bool {\n\t\tif converted, ok := r.Response.(core.GetInstanceResponse); ok {\n\t\t\treturn converted.LifecycleState != core.InstanceLifecycleStateRunning\n\t\t}\n\t\treturn true\n\t}\n\n\tpollingGetRequest := core.GetInstanceRequest{\n\t\tInstanceId: startResp.Instance.Id,\n\t\tRequestMetadata: helpers.GetRequestMetadataWithCustomizedRetryPolicy(pollUntilRunning),\n\t}\n\n\t_, err = c.computeClient.GetInstance(context.Background(), pollingGetRequest)\n\n\treturn err\n}", "func StartInstance(computeService *compute.Service) (*compute.Operation, error) {\n\treturn computeService.Instances.Start(ProjectID, Zone, InstanceName).Do()\n}", "func startInstance(ec2Service *ec2.EC2, instance *ec2.Instance) error {\n\tinstanceState, err := getInstanceState(*instance)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif instanceState == \"shutting-down\" || instanceState == \"terminated\" || instanceState == \"stopping\" || instanceState == \"stopped\" {\n\t\tfmt.Errorf(\"instance stopped or terminated\")\n\t}\n\n\tinstanceId := *instance.InstanceId\n\tinput := &ec2.StartInstancesInput{\n\t\tInstanceIds: []*string {\n\t\t\taws.String(instanceId),\n\t\t},\n\t}\n\n\t_, err = ec2Service.StartInstances(input)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfmt.Println(\"starting...\")\n\treturn nil\n}", "func (c *Client) StartInstance(instanceId string) error {\n\treturn StartInstance(c, instanceId)\n}", "func (cracker *Firecracker) Start() error {\n\treturn cracker.action(\"InstanceStart\", \"\")\n}", "func (ins *EC2RemoteClient) startInstance() error {\n\tlog.Printf(\"Starting EC2 Instance %s\", ins.InstanceID)\n\t_, err := ins.ec2Client.StartInstances(&ec2.StartInstancesInput{InstanceIds: aws.StringSlice([]string{ins.InstanceID})})\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Error starting instance : %s\", err)\n\t}\n\tlog.Printf(\"Waiting for Instance %s to become ready (may take a few minutes)\", ins.InstanceID)\n\terr = ins.ec2Client.WaitUntilInstanceStatusOk(&ec2.DescribeInstanceStatusInput{InstanceIds: aws.StringSlice([]string{ins.InstanceID})})\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Error waiting for instance to become available : %s\", err)\n\t}\n\treturn err\n}", "func (c *Client) Start() error {\n\t// memguard is used to securely keep sensitive information in memory.\n\t// This call makes sure all data will be destroy when the program exits.\n\tdefer memguard.Purge()\n\n\tif c.Conf.Id == \"\" {\n\t\tu, err := uuid.NewV4()\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"could not auto generate app instance id: %v\", err)\n\t\t}\n\t\tc.Conf.Id = u.String()[:8]\n\t}\n\n\tr, err := c.Running()\n\tif err != nil {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"id\": c.Conf.Id,\n\t\t}).WithError(err).Error(\"error while checking for another instance using the same id\")\n\n\t\treturn err\n\t}\n\n\tif r {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"id\": c.Conf.Id,\n\t\t}).Error(\"can't start. Another instance is already using the same id\")\n\n\t\treturn fmt.Errorf(\"can't start. Another instance is already using the same id %s\", c.Conf.Id)\n\t}\n\n\tlog.Infof(\"instance identifier is %s\", c.Conf.Id)\n\n\tif c.Conf.Detach {\n\t\tvar err error\n\n\t\tic, err := NewDetachedInstance(c.Conf.Id)\n\t\tif err != nil {\n\t\t\tlog.WithError(err).Errorf(\"error while creating directory to store mole instance related files\")\n\t\t\treturn err\n\t\t}\n\n\t\terr = startDaemonProcess(ic)\n\t\tif err != nil {\n\t\t\tlog.WithFields(log.Fields{\n\t\t\t\t\"id\": c.Conf.Id,\n\t\t\t}).WithError(err).Error(\"error starting ssh tunnel\")\n\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\tgo c.handleSignals()\n\t}\n\n\tif c.Conf.Verbose {\n\t\tlog.SetLevel(log.DebugLevel)\n\t}\n\n\td, err := fsutils.CreateInstanceDir(c.Conf.Id)\n\tif err != nil {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"id\": c.Conf.Id,\n\t\t}).WithError(err).Error(\"error creating directory for mole instance\")\n\n\t\treturn err\n\t}\n\n\tif c.Conf.Rpc {\n\t\taddr, err := rpc.Start(c.Conf.RpcAddress)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\trd := filepath.Join(d.Dir, \"rpc\")\n\n\t\terr = ioutil.WriteFile(rd, []byte(addr.String()), 0644)\n\t\tif err != nil {\n\t\t\tlog.WithFields(log.Fields{\n\t\t\t\t\"id\": c.Conf.Id,\n\t\t\t}).WithError(err).Error(\"error creating file with rpc address\")\n\n\t\t\treturn err\n\t\t}\n\n\t\tc.Conf.RpcAddress = addr.String()\n\n\t\tlog.Infof(\"rpc server address saved on %s\", rd)\n\t}\n\n\tt, err := createTunnel(c.Conf)\n\tif err != nil {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"id\": c.Conf.Id,\n\t\t}).WithError(err).Error(\"error creating tunnel\")\n\n\t\treturn err\n\t}\n\n\tc.Tunnel = t\n\n\tif err = c.Tunnel.Start(); err != nil {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"tunnel\": c.Tunnel.String(),\n\t\t}).WithError(err).Error(\"error while starting tunnel\")\n\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (m *Machine) Start(ctx context.Context) error {\n\tm.logger.Debug(\"Called Machine.Start()\")\n\talreadyStarted := true\n\tm.startOnce.Do(func() {\n\t\tm.logger.Debug(\"Marking Machine as Started\")\n\t\talreadyStarted = false\n\t})\n\tif alreadyStarted {\n\t\treturn ErrAlreadyStarted\n\t}\n\n\tvar err error\n\tdefer func() {\n\t\tif err != nil {\n\t\t\tif cleanupErr := m.doCleanup(); cleanupErr != nil {\n\t\t\t\tm.Logger().Errorf(\n\t\t\t\t\t\"failed to cleanup VM after previous start failure: %v\", cleanupErr)\n\t\t\t}\n\t\t}\n\t}()\n\n\terr = m.Handlers.Run(ctx, m)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = m.startInstance(ctx)\n\treturn err\n}", "func (broker *Broker) Start() error {\n\tif broker.inbox != nil { // Start has been called at least twice\n\t\tlog.Error(\"Could not start instance\")\n\t\treturn StartInstanceError(errors.New(\"instance already started\"))\n\t}\n\tbroker.inbox = broker.network.RegisterGossipProtocol(ProtoName)\n\tgo broker.dispatcher()\n\n\treturn nil\n}", "func (vm *VirtualMachine) Start(args *DomainXML, reply *bool) error {\n\t// Passing the false parameter to ensure the start vm task is not added to waitgroup if there is pending signal termination on rpc\n\t_, err := proc.AddTask(false)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"rpc/server:Start() Could not add task for vm start\")\n\t}\n\tdefer proc.TaskDone()\n\tlog.Trace(\"rpc/server:Start() Entering\")\n\tdefer log.Trace(\"rpc/server:Start() Leaving\")\n\n\tif err = validation.ValidateXMLString(args.XML); err != nil {\n\t\tsecLog.Errorf(\"rpc:server() Start: %s, Invalid domain XML format\", message.InvalidInputBadParam)\n\t\treturn nil\n\t}\n\n\t// pass in vm.Watcher to get the instance to the File System Watcher\n\t*reply = wlavm.Start(args.XML, vm.Watcher)\n\treturn nil\n}", "func startInstance(cs *compute.Service, w http.ResponseWriter) {\n\toperation, err := StartInstance(cs)\n\tif err != nil {\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\tlog.Fatal(err)\n\t}\n\tw.WriteHeader(http.StatusOK)\n\tdata, _ := operation.MarshalJSON()\n\tw.Write(data)\n}", "func (d *driver) LaunchInstance(\n\tvimInstance *catalogue.VIMInstance,\n\thostname, image, flavour, keyPair string,\n\tnetworks []*catalogue.VNFDConnectionPoint,\n\tsecurityGroups []string,\n\tuserData string) (*catalogue.Server, error) {\n\n\ttag := util.FuncName()\n\n\td.WithFields(log.Fields{\n\t\t\"tag\": tag,\n\t}).Debug(\"received request\")\n\n\treturn d.LaunchInstanceAndWaitWithIPs(vimInstance, hostname, image, flavour, keyPair, networks, securityGroups, userData, nil, nil)\n}", "func (m *Master) AddAndStart(req AddInstanceReq, out *StartRsp) error {\n\tprocSign := req.Instance.ProcSign\n\tm.mux.Lock()\n\t_, exists := m.instances[procSign]\n\tm.mux.Unlock()\n\n\tif exists {\n\t\tif req.Override {\n\t\t\t_, err := m.StopInstance(procSign, syscall.SIGINT)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t} else {\n\t\t\treturn fmt.Errorf(\"process: %s has exists\", procSign)\n\t\t}\n\t}\n\t// else: add instance and start\n\tnInst := exportInstanceFromReq(req.Instance)\n\tm.instances[procSign] = nInst\n\t// start\n\treturn m.Start(procSign, out)\n}", "func Start(s Startable) error {\n\treturn s.Start()\n}", "func (instance *Host) Start(ctx context.Context) (ferr fail.Error) {\n\tdefer fail.OnPanic(&ferr)\n\n\tif valid.IsNil(instance) {\n\t\treturn fail.InvalidInstanceError()\n\t}\n\tif ctx == nil {\n\t\treturn fail.InvalidParameterCannotBeNilError(\"ctx\")\n\t}\n\n\thostName := instance.GetName()\n\thostID, err := instance.GetID()\n\tif err != nil {\n\t\treturn fail.ConvertError(err)\n\t}\n\n\tsvc := instance.Service()\n\ttimings, xerr := svc.Timings()\n\tif xerr != nil {\n\t\treturn xerr\n\t}\n\n\txerr = svc.StartHost(ctx, hostID)\n\txerr = debug.InjectPlannedFail(xerr)\n\tif xerr != nil {\n\t\treturn xerr\n\t}\n\n\txerr = retry.WhileUnsuccessful(\n\t\tfunc() error {\n\t\t\tselect {\n\t\t\tcase <-ctx.Done():\n\t\t\t\treturn retry.StopRetryError(ctx.Err())\n\t\t\tdefault:\n\t\t\t}\n\n\t\t\ths, err := instance.ForceGetState(ctx)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tif hs != hoststate.Started {\n\t\t\t\treturn fail.NewError(\"%s not started yet: %s\", hostName, hs.String())\n\t\t\t}\n\n\t\t\treturn nil\n\t\t},\n\t\ttimings.NormalDelay(),\n\t\ttimings.ExecutionTimeout(),\n\t)\n\txerr = debug.InjectPlannedFail(xerr)\n\tif xerr != nil {\n\t\tswitch xerr.(type) {\n\t\tcase *fail.ErrAborted:\n\t\t\tif cerr := fail.ConvertError(fail.Cause(xerr)); cerr != nil {\n\t\t\t\treturn cerr\n\t\t\t}\n\t\t\treturn xerr\n\t\tcase *retry.ErrTimeout:\n\t\t\treturn fail.Wrap(xerr, \"timeout waiting Host '%s' to be started\", hostName)\n\t\tdefault:\n\t\t\treturn xerr\n\t\t}\n\t}\n\n\t// Now unsafeReload\n\txerr = instance.unsafeReload(ctx)\n\txerr = debug.InjectPlannedFail(xerr)\n\tif xerr != nil {\n\t\treturn xerr\n\t}\n\n\treturn nil\n}", "func (p *OnPrem) CreateInstance(ctx *Context) error {\n\tc := ctx.config\n\n\thypervisor := HypervisorInstance()\n\tif hypervisor == nil {\n\t\tfmt.Println(\"No hypervisor found on $PATH\")\n\t\tfmt.Println(\"Please install OPS using curl https://ops.city/get.sh -sSfL | sh\")\n\t\tos.Exit(1)\n\t}\n\n\tinstancename := c.CloudConfig.ImageName\n\n\tfmt.Printf(\"booting %s ...\\n\", instancename)\n\n\topshome := GetOpsHome()\n\timgpath := path.Join(opshome, \"images\", instancename)\n\n\tc.RunConfig.BaseName = instancename\n\tc.RunConfig.Imagename = imgpath\n\tc.RunConfig.OnPrem = true\n\n\thypervisor.Start(&c.RunConfig)\n\n\treturn nil\n}", "func (a *Agent) spawnInstance(ctx context.Context, c instance.Config) {\n\tinst, err := a.instanceFactory(a.cfg.Global, c, a.cfg.WALDir, a.logger)\n\tif err != nil {\n\t\tlevel.Error(a.logger).Log(\"msg\", \"failed to create instance\", \"err\", err)\n\t\treturn\n\t}\n\n\tfor {\n\t\terr = inst.Run(ctx)\n\t\tif err != nil && err != context.Canceled {\n\t\t\tinstanceAbnormalExits.WithLabelValues(c.Name).Inc()\n\t\t\tlevel.Error(a.logger).Log(\"msg\", \"instance stopped abnormally, restarting after backoff period\", \"err\", err, \"backoff\", a.cfg.InstanceRestartBackoff, \"instance\", c.Name)\n\t\t\ttime.Sleep(a.cfg.InstanceRestartBackoff)\n\t\t} else {\n\t\t\tlevel.Info(a.logger).Log(\"msg\", \"stopped instance\", \"instance\", c.Name)\n\t\t\tbreak\n\t\t}\n\t}\n}", "func (p *PVT24) Start() error { panic(\"not implemented\") }", "func (i *Instance) Start() {\n\tvar err error\n\tvar router = mux.NewRouter()\n\n\t// Establish database connection\n\n\trepo, err := repository.NewMySQLRepository(i.Config.DB)\n\tif err != nil {\n\t\tlogrus.WithError(err).Fatal(\"Could not create mysql repository\")\n\t}\n\n\ti.Service, err = service.NewProvider(i.Config.Service, repo)\n\tif err != nil {\n\t\tlogrus.WithError(err).Fatal(\"Could not create service provider\")\n\t}\n\n\t// Initialize API\n\ti.API, err = api.New(i.Config.API, i.Service, router)\n\tif err != nil {\n\t\tlogrus.WithError(err).Fatal(\"Could not create API instance\")\n\t}\n\n\t// Startup the HTTP Server in a way that we can gracefully shut it down again\n\ti.httpServer = &http.Server{\n\t\tAddr: i.Config.ListenAddress,\n\t\tHandler: router,\n\t}\n\n\terr = i.httpServer.ListenAndServe()\n\tif err != http.ErrServerClosed {\n\t\tlogrus.WithError(err).Error(\"HTTP Server stopped unexpected\")\n\t\ti.Shutdown()\n\t} else {\n\t\tlogrus.WithError(err).Info(\"HTTP Server stopped\")\n\t}\n}", "func (instance *Instance) Start() error {\n\t// Init Twitter Streaming API client\n\tclient, err := createTwitterClient(instance.config)\n\tif err != nil {\n\t\treturn err\n\t}\n\tinstance.client = client\n\n\treturn nil\n}", "func InitInstance() {\n\tvar err error\n\tvar token string\n\tif Inst().ConfigMap != \"\" {\n\t\tlogrus.Infof(\"Using Config Map: %s \", Inst().ConfigMap)\n\t\ttoken, err = Inst().S.GetTokenFromConfigMap(Inst().ConfigMap)\n\t\texpect(err).NotTo(haveOccurred())\n\t\tlogrus.Infof(\"Token used for initializing: %s \", token)\n\t} else {\n\t\ttoken = \"\"\n\t}\n\n\terr = Inst().S.Init(scheduler.InitOptions{\n\t\tSpecDir: Inst().SpecDir,\n\t\tVolDriverName: Inst().V.String(),\n\t\tNodeDriverName: Inst().N.String(),\n\t\tSecretConfigMapName: Inst().ConfigMap,\n\t\tCustomAppConfig: Inst().CustomAppConfig,\n\t\tStorageProvisioner: Inst().Provisioner,\n\t\tSecretType: Inst().SecretType,\n\t\tVaultAddress: Inst().VaultAddress,\n\t\tVaultToken: Inst().VaultToken,\n\t\tPureVolumes: Inst().PureVolumes,\n\t\tHelmValuesConfigMapName: Inst().HelmValuesConfigMap,\n\t})\n\texpect(err).NotTo(haveOccurred())\n\n\terr = Inst().N.Init(node.InitOptions{\n\t\tSpecDir: Inst().SpecDir,\n\t})\n\texpect(err).NotTo(haveOccurred())\n\n\terr = Inst().V.Init(Inst().S.String(), Inst().N.String(), token, Inst().Provisioner, Inst().CsiGenericDriverConfigMap)\n\texpect(err).NotTo(haveOccurred())\n\n\tif Inst().Backup != nil {\n\t\terr = Inst().Backup.Init(Inst().S.String(), Inst().N.String(), Inst().V.String(), token)\n\t\texpect(err).NotTo(haveOccurred())\n\t}\n\tif testRailHostname != \"\" && testRailUsername != \"\" && testRailPassword != \"\" {\n\t\terr = testrailuttils.Init(testRailHostname, testRailUsername, testRailPassword)\n\t\tif err == nil {\n\t\t\tif testrailuttils.MilestoneName == \"\" || testrailuttils.RunName == \"\" || testrailuttils.JobRunID == \"\" {\n\t\t\t\tprocessError(fmt.Errorf(\"not all details provided to update testrail\"))\n\t\t\t}\n\t\t\ttestrailuttils.CreateMilestone()\n\t\t}\n\t} else {\n\t\tlogrus.Debugf(\"Not all information to connect to testrail is provided, skipping updates to testrail\")\n\t}\n\n\tif jiraUserName != \"\" && jiraToken != \"\" {\n\t\tlogrus.Info(\"Initializing JIRA connection\")\n\t\tjirautils.Init(jiraUserName, jiraToken)\n\n\t} else {\n\t\tlogrus.Debugf(\"Not all information to connect to JIRA is provided.\")\n\t}\n}", "func (px *Paxos) Start(seq int, v interface{}) {\n\tpx.mu.Lock()\n\tdefer px.mu.Unlock()\n\n\t// Your code here.\n\tlog.Println(\"Paxos.Start me:\", px.me, \"Seq:\", seq, \"px.min:\", px.min, \"px.max:\", px.max, \"Value:\", v)\n\tif seq < px.min {\n\t\treturn\n\t}\n\n\tinstance, ok := px.instances[seq]\n\tif !ok {\n\t\tinstance := new(Instance)\n\t\tinstance.seq = seq\n\t\tinstance.status = Pending\n\t\tinstance.pNum = -1\n\t\tinstance.aNum = -1\n\t\tinstance.aValue = nil\n\t\tpx.instances[seq] = instance\n\t} else {\n\t\tif instance.status == Decided {\n\t\t\tlog.Println(\"Paxos.Start me:\", px.me, \"Seq:\", seq, \"px.min:\", px.min, \"px.max:\", px.max, \"Value:\", v, \"SEQ HAS BEEN DECIDED\")\n\t\t\treturn\n\t\t}\n\t}\n\n\tgo px.propose(seq, v)\n}", "func InitInstance(instance Instance) error {\n\tif C.isProcAddrSet() == 0 {\n\t\treturn errors.New(\"vulkan: GetInstanceProcAddr is not set\")\n\t}\n\tret := C.vkInitInstance((C.VkInstance)(instance))\n\tif ret < 0 {\n\t\treturn errors.New(\"vkInitInstance failed\")\n\t}\n\treturn nil\n}", "func Start(numInstances int) error {\n\t// Ideally we should let the socket choose the port, but then\n\t// some things like the logger will not be set correctly.\n\tvar peers []gubernator.PeerInfo\n\tport := 1111\n\tfor i := 0; i < numInstances; i++ {\n\t\tpeers = append(peers, gubernator.PeerInfo{\n\t\t\tHTTPAddress: fmt.Sprintf(\"localhost:%d\", port),\n\t\t\tGRPCAddress: fmt.Sprintf(\"localhost:%d\", port+1),\n\t\t})\n\t\tport += 2\n\t}\n\treturn StartWith(peers)\n}", "func HandleInstanceCreate(w rest.ResponseWriter, r *rest.Request) {\n\t// get ima\n\tima := Ima{}\n\terr := r.DecodeJsonPayload(&ima)\n\tif err != nil {\n\t\trest.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\tif ima.Ima == \"\" {\n\t\trest.Error(w, \"ima required\", 400)\n\t\treturn\n\t}\n\tif ima.Mem == 0 {\n\t\trest.Error(w, \"memory required\", 400)\n\t\treturn\n\t}\n\tif ima.Cpu == 0 {\n\t\trest.Error(w, \"cpu required\", 400)\n\t\treturn\n\t}\n\n\t// start the instance\n\tos := getImaOs(ima.Ima)\n\tswitch os {\n\tcase \"freebsd\":\n\t\t// clone ima to instance\n\t\tinstanceid := allocateInstanceId()\n\t\tcloneIma(ima.Ima, instanceid)\n\n\t\t// create network interface and bring up\n\t\ttap := allocateTap()\n\t\tif tap == \"\" {\n\t\t\treturn\n\t\t}\n\t\tsaveTap(tap, instanceid)\n\t\tbridge := findBridge()\n\t\taddTapToBridge(tap, bridge)\n\t\tbridgeUp(bridge)\n\n\t\tnmdm := \"/dev/nmdm-\" + instanceid + \"-A\"\n\t\tsaveCpu(ima.Cpu, instanceid)\n\t\tsaveMem(ima.Mem, instanceid)\n\t\tgo startFreeBSDVM(nmdm, ima.Cpu, ima.Mem, tap, instanceid)\n\t\tw.WriteJson(&instanceid)\n\tcase \"linux\":\n\t\t// clone ima to instance\n\t\tinstanceid := allocateInstanceId()\n\t\tcloneIma(ima.Ima, instanceid)\n\n\t\t// create network interface and bring up\n\t\ttap := allocateTap()\n\t\tif tap == \"\" {\n\t\t\treturn\n\t\t}\n\t\tsaveTap(tap, instanceid)\n\t\tbridge := findBridge()\n\t\taddTapToBridge(tap, bridge)\n\t\tbridgeUp(bridge)\n\n\t\t//nmdm := \"/dev/nmdm-\" + instanceid + \"-A\"\n\t\tsaveCpu(ima.Cpu, instanceid)\n\t\tsaveMem(ima.Mem, instanceid)\n\t\tbhyveDestroy(instanceid)\n\t\tnmdm := \"/dev/nmdm-\" + instanceid + \"-A\"\n\t\tgo startLinuxVM(nmdm, ima.Cpu, ima.Mem, tap, instanceid)\n\t\tw.WriteJson(&instanceid)\n\tdefault:\n\t\trest.Error(w, \"unknown OS\", 400)\n\t}\n}", "func (p *GenericPlugin) Start(Args ...interface{}) error {\n\n\tp.mu.Lock()\n\tdefer p.mu.Unlock()\n\n\tswitch p.state {\n\tcase stateNotLoaded:\n\t\tif err := p.Load(); err != nil {\n\t\t\treturn err\n\t\t}\n\tcase stateLoaded:\n\tcase stateActive:\n\t\tp.agent.Logger().Printf(\"Cannot Start() module [ %s ], already running\", p.Name())\n\t\treturn nil\n\t}\n\n\tif err := p.ReadStatus(); err != nil {\n\t\treturn err\n\t}\n\n\tp.state = stateActive\n\tp.started = time.Now()\n\n\tswitch {\n\tcase p.init != nil:\n\t\t{\n\t\t\tArgs = append(p.args, Args...)\n\t\t\tif err := p.init(Args...); err != nil {\n\t\t\t\tp.stop()\n\t\t\t\tp.state = stateLoaded\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\tdefault:\n\t\treturn fmt.Errorf(\"plugin error: No Init() function loaded for module [ %s ]\", p.Name())\n\t}\n\n\tp.agent.Logger().Printf(\"Started module [ %s ]\", p.Name())\n\n\treturn nil\n}", "func SpawnInstance(user *twitter.User, twitterClient *twitter.Client, appConfig *AppConfig, tweetRefID *int64) error {\n\tvar count int\n\tif err := CountActiveSession(&count); err != nil {\n\t\treturn err\n\t}\n\n\tif count > appConfig.Scw.Limit {\n\t\treturn errors.New(\"Server limit reached\")\n\t}\n\n\ts, _ := scw.NewScalewayAPI(appConfig.Scw.AccessKey, appConfig.Scw.SecretKey, \"SCW-Twt\", appConfig.Scw.Region)\n\tpassword := NewPassword(6)\n\tserver := scw.ScalewayServerDefinition{\n\t\tName: user.ScreenName,\n\t\tImage: &appConfig.Scw.Image,\n\t\tOrganization: appConfig.Scw.AccessKey,\n\t\tCommercialType: appConfig.Scw.InstanceType,\n\t\tTags: []string{password},\n\t}\n\n\tid, err := s.PostServer(server)\n\n\tsession := &Session{\n\t\tUserID: user.ID,\n\t\tScreenName: user.ScreenName,\n\t\tServerID: id,\n\t\tState: \"starting\",\n\t}\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err := CreateSession(session); err != nil {\n\t\treturn err\n\t}\n\n\tlog.WithFields(logrus.Fields{\n\t\t\"twitter ID\": user.ID,\n\t\t\"server ID\": id,\n\t}).Info(\"Server created\")\n\tTweet(\n\t\ttwitterClient,\n\t\tfmt.Sprintf(\"@%s %s\", user.ScreenName, appConfig.Messages.StartingInProgress),\n\t\ttweetRefID,\n\t)\n\n\tif err := scw.StartServer(s, id, true); err != nil {\n\t\treturn err\n\n\t}\n\tlog.WithFields(logrus.Fields{\n\t\t\"twitter ID\": user.ID,\n\t\t\"server ID\": id,\n\t}).Info(\"Server started\")\n\n\tserver_info, _ := s.GetServer(id)\n\n\tnow := time.Now()\n\tsession.SessionEndAt = now.Add(time.Duration(appConfig.Scw.SessionTime * 60000000000))\n\tsession.State = \"running\"\n\n\tif err := UpdateSession(session); err != nil {\n\t\treturn err\n\t}\n\n\tTweet(\n\t\ttwitterClient,\n\t\tfmt.Sprintf(\"@%s %s\", user.ScreenName, appConfig.Messages.ServerReady),\n\t\ttweetRefID,\n\t)\n\n\tDM(\n\t\ttwitterClient,\n\t\tuser.ID,\n\t\tuser.ScreenName,\n\t\tfmt.Sprintf(\"%s %s ubuntu@%s\", appConfig.Messages.DmServerReady, password, server_info.PublicAddress.IP),\n\t)\n\treturn nil\n}", "func RunWithInstance(fn func(*Instance)) {\n\tRunWithDB(func(db *sql.DB) {\n\t\tif instance, err := NewInstance(db); err != nil {\n\t\t\tpanic(err)\n\t\t} else {\n\t\t\tfn(instance)\n\t\t}\n\t})\n}", "func (p *Plugin) Start(ctx context.Context, in *plugin.StartRequest) (*plugin.StartResponse, error) {\n\terr := p.Service.Start(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &plugin.StartResponse{}, nil\n}", "func (sn *SimNode) Start(snapshots map[string][]byte) error {\n\tnewService := func(name string) func(ctx *node.ServiceContext) (node.Service, error) {\n\t\treturn func(nodeCtx *node.ServiceContext) (node.Service, error) {\n\t\t\tctx := &ServiceContext{\n\t\t\t\tRPCDialer: sn.adapter,\n\t\t\t\tNodeContext: nodeCtx,\n\t\t\t\tConfig: sn.config,\n\t\t\t}\n\t\t\tif snapshots != nil {\n\t\t\t\tctx.Snapshot = snapshots[name]\n\t\t\t}\n\t\t\tserviceFunc := sn.adapter.services[name]\n\t\t\tservice, err := serviceFunc(ctx)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tsn.running[name] = service\n\t\t\treturn service, nil\n\t\t}\n\t}\n\n\t// ensure we only register the services once in the case of the node\n\t// being stopped and then started again\n\tvar regErr error\n\tsn.registerOnce.Do(func() {\n\t\tfor _, name := range sn.config.Services {\n\t\t\tif err := sn.node.Register(newService(name)); err != nil {\n\t\t\t\tregErr = err\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t})\n\tif regErr != nil {\n\t\treturn regErr\n\t}\n\n\tif err := sn.node.Start(); err != nil {\n\t\treturn err\n\t}\n\n\t// create an in-process RPC client\n\thandler, err := sn.node.RPCHandler()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tsn.lock.Lock()\n\tsn.client = rpc.DialInProc(handler)\n\tsn.lock.Unlock()\n\n\treturn nil\n}", "func Start(ctx context.Context, fun TaskFunc) Task {\n\treturn (&taskGroup{ctx: ctx}).Task(fun)\n}", "func (d *Driver) Start() error {\n\tcs := d.client()\n\t_, err := cs.AsyncRequest(&egoscale.StartVirtualMachine{\n\t\tID: d.ID,\n\t}, d.async)\n\n\treturn err\n}", "func (instance *ApplicationRuntime) Start() error{\n\tif nil != instance {\n\t\t// start application server\n\t\treturn instance.server.Start()\n\t}\n\treturn application_types.PanicSystemError\n}", "func (p *Init) Start(ctx context.Context) error {\n\tp.mu.Lock()\n\tdefer p.mu.Unlock()\n\n\treturn p.initState.Start(ctx)\n}", "func (r *ProjectsLocationsInstancesService) Start(name string, startinstancerequest *StartInstanceRequest) *ProjectsLocationsInstancesStartCall {\n\tc := &ProjectsLocationsInstancesStartCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.name = name\n\tc.startinstancerequest = startinstancerequest\n\treturn c\n}", "func (oh *OperatingHours) StartInstances(startInstances []*string) error {\n\n\tlog.WithField(\"len\", len(startInstances)).Info(\"startInstances\")\n\n\tif len(startInstances) > 0 {\n\t\t_, err := oh.EC2API.StartInstances(&ec2.StartInstancesInput{InstanceIds: startInstances})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfor _, instanceID := range startInstances {\n\t\t\tlog.WithField(\"instanceID:\", aws.StringValue(instanceID)).Info(\"start\")\n\t\t}\n\t}\n\n\treturn nil\n}", "func (i *Instance) SpawnInstance() (instance *Instance, err error) {\n\tvar res uintptr\n\tvar newUnknown *ole.IUnknown\n\n\tres, _, _ = syscall.SyscallN(\n\t\ti.vTable.SpawnInstance, // IWbemClassObject::SpawnInstance(\n\t\tuintptr(unsafe.Pointer(i.object)), // IWbemClassObject ptr\n\t\tuintptr(0), // [in] long lFlags,\n\t\tuintptr(unsafe.Pointer(&newUnknown))) // [out] IWbemClassObject **ppNewInstance)\n\tif res != 0 {\n\t\treturn nil, ole.NewError(res)\n\t}\n\n\treturn newInstance(newUnknown, i.service), nil\n}", "func (f *FakeInstance) MassStart(_ context.Context, _ []string) error {\n\tpanic(\"implement me\")\n}", "func (px *Paxos) Start(seq int, v interface{}) {\n\t// Your code here.\n\n\tmin := px.Min()\n\tpx.mu.Lock()\n\tif px.instances[seq] == nil && seq >= min {\n\t\tpx.CreateInstance(seq)\n\t}\n\tpx.mu.Unlock()\n\n\tstatus, _ := px.Status(seq)\n\tif status == Decided || status == Forgotten {\n\t\treturn\n\t}\n\tgo px.Proposer(seq, v)\n\t//execute goroutine to start agreement process so this can return immediately\n\t//goroutine needs to propose value from this peer, also run acceptor and learner?\n\t//actual paxos agreement for single instance should be run by this\n\t//goroutine also needs to transmit done value for each peer in the messages and update min and delete from map\n\t//set na to 0\n}", "func (t *SelfTester) Start() {}", "func (m *Master) Start(procSign string, out *StartRsp) error {\n\t// TODO\n\tfflags := NewFFlags(m.rootPath)\n\tcmd, err := m.StartInstance(procSign)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// when cmd is not nil, the process is starting for the first time\n\tif cmd != nil {\n\t\t// wait 3 secs to see if process still exists, if so\n\t\trunOK := waitForRunning(procSign, cmd)\n\t\tif runOK {\n\t\t\tfflags.SetForRunning(procSign, cmd.Process.Pid)\n\t\t\t*out = StartRsp{\n\t\t\t\tProcSign: procSign,\n\t\t\t\tPid: cmd.Process.Pid,\n\t\t\t\tHasRunning: false,\n\t\t\t}\n\t\t\treturn nil\n\t\t}\n\t\treturn fmt.Errorf(\"process: %s exits too quickly\", procSign)\n\t}\n\n\tf := NewFFlags(m.rootPath)\n\tpid := f.ReadPid(procSign)\n\t*out = StartRsp{\n\t\tProcSign: procSign,\n\t\tPid: pid,\n\t\tHasRunning: true,\n\t}\n\treturn nil\n}", "func (s stack) StartHost(ctx context.Context, hostParam stacks.HostParameter) fail.Error {\n\tif valid.IsNil(s) {\n\t\treturn fail.InvalidInstanceError()\n\t}\n\treturn fail.NotImplementedError(\"implement me\")\n}", "func (c *Client) StartInstances(instanceIds []string) ([]InstanceStateChange, error) {\n\tresponse := InstancesStateChangeResponse{}\n\targs := &StartInstancesArgs{InstanceIds: instanceIds}\n\terr := c.Invoke(\"StartInstances\", args, &response)\n\tif err == nil {\n\t\treturn response.InstancesSet.Item, nil\n\t}\n\treturn nil, err\n}", "func (service *Service) Start(ctx context.Context, initFn func() error) {\n\thandleErr(service.openSQLDBConnections(ctx))\n\thandleErr(service.openRedisConnections(ctx))\n\thandleErr(service.openExternalConnections(ctx))\n\thandleErr(service.initGRPC(ctx))\n\thandleErr(initFn())\n\thandleErr(service.run(ctx))\n}", "func (v *VMInstances) StartVMInstances(project string, zone string, name string) {\n\t_, err := v.computeService.Instances.Start(project, zone, name).Context(v.Ctx).Do()\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tfmt.Printf(\"Started instance %s\", name)\n}", "func (_m *IProvider) Start(_a0 map[string]*model.WorkerConfig, _a1 []*message.Message) {\n\t_m.Called(_a0, _a1)\n}", "func (f *Function) AddInstance() *metrics.Metric {\n\tf.Lock()\n\tdefer f.Unlock()\n\n\tlogger := log.WithFields(log.Fields{\"fID\": f.fID})\n\n\tlogger.Debug(\"Adding instance\")\n\n\tvar metr *metrics.Metric = nil\n\n\tctx, cancel := context.WithTimeout(context.Background(), time.Minute*5)\n\tdefer cancel()\n\n\tif f.isSnapshotReady {\n\t\tmetr = f.LoadInstance()\n\t} else {\n\t\tresp, _, err := orch.StartVM(ctx, f.getVMID(), f.imageName)\n\t\tif err != nil {\n\t\t\tlog.Panic(err)\n\t\t}\n\t\tf.guestIP = resp.GuestIP\n\t\tf.vmID = f.getVMID()\n\t\tf.lastInstanceID++\n\t}\n\n\ttStart := time.Now()\n\tfuncClient, err := f.getFuncClient()\n\tif metr != nil {\n\t\tmetr.MetricMap[metrics.ConnectFuncClient] = metrics.ToUS(time.Since(tStart))\n\t}\n\tif err != nil {\n\t\tlogger.Panic(\"Failed to acquire func client\")\n\t}\n\tf.funcClient = &funcClient\n\n\tf.stats.IncStarted(f.fID)\n\n\treturn metr\n}", "func (s *Supervisor) Start() {\n\ts.StartWithContext(context.Background())\n}", "func (f *Function) CreateInstanceSnapshot() {\n\tlogger := log.WithFields(log.Fields{\"fID\": f.fID})\n\n\tlogger.Debug(\"Creating instance snapshot\")\n\n\tctx, cancel := context.WithTimeout(context.Background(), time.Second*5)\n\tdefer cancel()\n\n\terr := orch.PauseVM(ctx, f.vmID)\n\tif err != nil {\n\t\tlog.Panic(err)\n\t}\n\n\terr = orch.CreateSnapshot(ctx, f.vmID)\n\tif err != nil {\n\t\tlog.Panic(err)\n\t}\n\n\t_, err = orch.ResumeVM(ctx, f.vmID)\n\tif err != nil {\n\t\tlog.Panic(err)\n\t}\n}", "func (d *Dameon) Start(netinterface string) error {\n\n\td.pm = NewPM()\n\tgo d.monitorProcs()\n\n\tvar err error\n\tvar absPath string\n\n\t// Ensure single instance\n\tif absPath, err = filepath.Abs(utils.Options.LockFile); err != nil {\n\t\treturn err\n\t}\n\tif d.lockfile, err = lockfile.New(absPath); err != nil {\n\t\treturn err\n\t}\n\n\tif err = d.lockfile.TryLock(); err != nil {\n\t\treturn err\n\t}\n\n\t// Start rpc\n\tgop := NewGOP(d)\n\trpc.Register(gop)\n\trpc.HandleHTTP()\n\n\tif d.listener, err = net.Listen(\"tcp\", netinterface); err != nil {\n\t\treturn err\n\t}\n\n\treturn http.Serve(d.listener, nil)\n}", "func (s LifecyclerRPC) BeforeInstance(state *isclib.Instance) error {\n\tvar resp struct{}\n\terr := s.client.Call(\"Plugin.BeforeInstance\", state, &resp)\n\treturn err\n}", "func (w *GovernanceWorker) startMicroserviceInstForAgreement(msdef *persistence.MicroserviceDefinition, agreementId string, dependencyPath []persistence.ServiceInstancePathElement, protocol string) error {\n\tglog.V(3).Infof(logString(fmt.Sprintf(\"start service instance %v for agreement %v\", msdef.SpecRef, agreementId)))\n\n\tvar msi *persistence.MicroserviceInstance\n\tneeds_new_ms := false\n\n\t// Always start a new instance if the sharing mode is multiple.\n\tif msdef.Sharable == exchange.MS_SHARING_MODE_MULTIPLE {\n\t\tneeds_new_ms = true\n\t\t// For other sharing modes, start a new instance only if there is no existing one.\n\t\t// The \"exclusive\" sharing mode is handled by maxAgreements=1 in the node side policy file. This ensures that agbots and nodes will\n\t\t// only support one agreement at any time.\n\t} else if ms_insts, err := persistence.FindMicroserviceInstances(w.db, []persistence.MIFilter{persistence.AllInstancesMIFilter(msdef.SpecRef, msdef.Version), persistence.UnarchivedMIFilter()}); err != nil {\n\t\teventlog.LogDatabaseEvent(w.db, persistence.SEVERITY_ERROR,\n\t\t\tfmt.Sprintf(\"Error retrieving all the service instances from db for %v version %v key %v. %v\", msdef.SpecRef, msdef.Version, msdef.Id, err),\n\t\t\tpersistence.EC_DATABASE_ERROR)\n\t\treturn fmt.Errorf(logString(fmt.Sprintf(\"Error retrieving all the service instances from db for %v version %v key %v. %v\", msdef.SpecRef, msdef.Version, msdef.Id, err)))\n\t} else if ms_insts == nil || len(ms_insts) == 0 {\n\t\tneeds_new_ms = true\n\t} else {\n\t\tmsi = &ms_insts[0]\n\t}\n\n\tif needs_new_ms {\n\t\tvar inst_err error\n\t\tif msi, inst_err = w.StartMicroservice(msdef.Id, agreementId, dependencyPath); inst_err != nil {\n\n\t\t\teventlog.LogServiceEvent2(w.db, persistence.SEVERITY_ERROR,\n\t\t\t\tfmt.Sprintf(\"Service starting failed for %v version %v, error: %v\", msdef.SpecRef, msdef.Version, inst_err),\n\t\t\t\tpersistence.EC_ERROR_START_SERVICE,\n\t\t\t\t\"\", msdef.SpecRef, msdef.Org, msdef.Version, msdef.Arch, []string{agreementId})\n\n\t\t\t// Try to downgrade the service/microservice to a lower version.\n\t\t\tglog.V(3).Infof(logString(fmt.Sprintf(\"Ending the agreement: %v because service %v failed to start\", agreementId, msdef.SpecRef)))\n\t\t\tag_reason_code := w.producerPH[protocol].GetTerminationCode(producer.TERM_REASON_MS_DOWNGRADE_REQUIRED)\n\t\t\tag_reason_text := w.producerPH[protocol].GetTerminationReason(ag_reason_code)\n\t\t\tif agreementId != \"\" {\n\t\t\t\tw.cancelAgreement(agreementId, protocol, ag_reason_code, ag_reason_text)\n\t\t\t}\n\n\t\t\tglog.V(3).Infof(logString(fmt.Sprintf(\"Downgrading service %v because version %v key %v failed to start. Error: %v\", msdef.SpecRef, msdef.Version, msdef.Id, inst_err)))\n\t\t\teventlog.LogServiceEvent2(w.db, persistence.SEVERITY_INFO,\n\t\t\t\tfmt.Sprintf(\"Start downgrading service %v version %v because service for aagreement failed to start.\", msdef.SpecRef, msdef.Version),\n\t\t\t\tpersistence.EC_START_DOWNGRADE_SERVICE,\n\t\t\t\t\"\", msdef.SpecRef, msdef.Org, msdef.Version, msdef.Arch, []string{agreementId})\n\n\t\t\tif err := w.RollbackMicroservice(msdef); err != nil {\n\t\t\t\teventlog.LogServiceEvent2(w.db, persistence.SEVERITY_INFO,\n\t\t\t\t\tfmt.Sprintf(\"Error downgrading service %v version %v. %v\", msdef.SpecRef, msdef.Version, err),\n\t\t\t\t\tpersistence.EC_ERROR_DOWNGRADE_SERVICE,\n\t\t\t\t\t\"\", msdef.SpecRef, msdef.Org, msdef.Version, msdef.Arch, []string{agreementId})\n\t\t\t\tglog.Errorf(logString(fmt.Sprintf(\"Error downgrading service %v version %v key %v. %v\", msdef.SpecRef, msdef.Version, msdef.Id, err)))\n\t\t\t}\n\n\t\t\treturn fmt.Errorf(logString(fmt.Sprintf(\"Failed to start service instance for %v version %v key %v. %v\", msdef.SpecRef, msdef.Version, msdef.Id, inst_err)))\n\t\t}\n\t} else if _, err := persistence.UpdateMSInstanceAddDependencyPath(w.db, msi.GetKey(), &dependencyPath); err != nil {\n\t\treturn fmt.Errorf(logString(fmt.Sprintf(\"error adding dependency path %v to the service %v: %v\", dependencyPath, msi.GetKey(), err)))\n\t}\n\n\t// Add the agreement id into the instance so that the workload containers know which instance to associate with.\n\tif agreementId != \"\" {\n\t\tif _, err := persistence.UpdateMSInstanceAssociatedAgreements(w.db, msi.GetKey(), true, agreementId); err != nil {\n\t\t\treturn fmt.Errorf(logString(fmt.Sprintf(\"error adding agreement id %v to the service %v: %v\", agreementId, msi.GetKey(), err)))\n\t\t}\n\t} else {\n\t\tif _, err := persistence.UpdateMSInstanceAgreementLess(w.db, msi.GetKey()); err != nil {\n\t\t\treturn fmt.Errorf(logString(fmt.Sprintf(\"error setting agreement-less on the service %v: %v\", msi.GetKey(), err)))\n\t\t}\n\t}\n\n\treturn nil\n}", "func (b *baseSysInit) Start(svcName string) error {\n\tb.StartCmd.AppendArgs(sysInitCmdArgs(b.sysInitType, svcName, \"start\")...)\n\t_, err := b.StartCmd.RunCombined()\n\treturn err\n}", "func (vm *VirtualMachine) Start(client SkytapClient) (*VirtualMachine, error) {\n\tlog.WithFields(log.Fields{\"vmId\": vm.Id}).Info(\"Starting VM\")\n\n\treturn vm.ChangeRunstate(client, RunStateStart, RunStateStart)\n}", "func (s *svc) Start() error {\n\ts.Lock()\n\tdefer s.Unlock()\n\n\ts.status = &router.Status{\n\t\tCode: router.Running,\n\t\tError: nil,\n\t}\n\n\treturn nil\n}", "func (e StaticFeeEstimator) Start() error {\n\treturn nil\n}", "func (p *FuncPool) AddInstance(fID, imageName string) (string, error) {\n\tf := p.getFunction(fID, imageName)\n\n\tlogger := log.WithFields(log.Fields{\"fID\": f.fID})\n\n\tf.OnceAddInstance.Do(\n\t\tfunc() {\n\t\t\tlogger.Debug(\"Function is inactive, starting the instance...\")\n\t\t\tf.AddInstance()\n\t\t})\n\n\treturn \"Instance started\", nil\n}", "func (mg *Groups) Start(instances int, force bool) error {\n\n\tif mg.group != nil && len(mg.group.ID) > 0 {\n\t\tif appClient := application.New(mg.client); appClient != nil {\n\n\t\t\tcallbackFunc := func(appID string) error {\n\n\t\t\t\tif err := appClient.Get(appID).Start(instances, force); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\treturn mg.traverseGroupsWithAppID(mg.group, callbackFunc)\n\t\t}\n\t\treturn fmt.Errorf(\"unnable to connect\")\n\t}\n\treturn errors.New(\"group cannot be null nor empty\")\n}", "func (c *Client) CreateInstance(displayName, availabilityDomain, compartmentID, nodeShape, nodeImageName, nodeSubnetID, sshUser, authorizedKeys string, nodeOCPUs, nodeMemoryInGBs int) (string, error) {\n\n\treq := identity.ListAvailabilityDomainsRequest{}\n\treq.CompartmentId = &compartmentID\n\tads, err := c.identityClient.ListAvailabilityDomains(context.Background(), req)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Just in case shortened or lower-case availability domain name was used\n\tlog.Debugf(\"Resolving availability domain from %s\", availabilityDomain)\n\tfor _, ad := range ads.Items {\n\t\tif strings.Contains(*ad.Name, strings.ToUpper(availabilityDomain)) {\n\t\t\tlog.Debugf(\"Availability domain %s\", *ad.Name)\n\t\t\tavailabilityDomain = *ad.Name\n\t\t}\n\t}\n\n\timageID, err := c.getImageID(compartmentID, nodeImageName)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Create the launch compute instance request\n\trequest := core.LaunchInstanceRequest{\n\t\tLaunchInstanceDetails: core.LaunchInstanceDetails{\n\t\t\tAvailabilityDomain: &availabilityDomain,\n\t\t\tCompartmentId: &compartmentID,\n\t\t\tShape: &nodeShape,\n\t\t\tCreateVnicDetails: &core.CreateVnicDetails{\n\t\t\t\tSubnetId: &nodeSubnetID,\n\t\t\t},\n\t\t\tDisplayName: &displayName,\n\t\t\tMetadata: map[string]string{\n\t\t\t\t\"ssh_authorized_keys\": authorizedKeys,\n\t\t\t\t\"user_data\": base64.StdEncoding.EncodeToString(createCloudInitScript(sshUser)),\n\t\t\t},\n\t\t\tSourceDetails: core.InstanceSourceViaImageDetails{\n\t\t\t\tImageId: imageID,\n\t\t\t},\n\t\t},\n\t}\n\n\tif nodeOCPUs > 0 {\n\t\toCPUs := float32(nodeOCPUs)\n\t\tmemoryInGBs := float32(nodeMemoryInGBs)\n\n\t\tLaunchInstanceShapeConfigDetails := core.LaunchInstanceShapeConfigDetails{\n\t\t\tOcpus: &oCPUs,\n\t\t\tMemoryInGBs: &memoryInGBs,\n\t\t}\n\t\trequest.ShapeConfig = &LaunchInstanceShapeConfigDetails\n\t}\n\n\tlog.Debugf(\"Launching instance with cloud-init: %s\", string(createCloudInitScript(sshUser)))\n\n\tcreateResp, err := c.computeClient.LaunchInstance(context.Background(), request)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// wait until lifecycle status is Running\n\tpollUntilRunning := func(r common.OCIOperationResponse) bool {\n\t\tif converted, ok := r.Response.(core.GetInstanceResponse); ok {\n\t\t\treturn converted.LifecycleState != core.InstanceLifecycleStateRunning\n\t\t}\n\t\treturn true\n\t}\n\n\t// create get instance request with a retry policy which takes a function\n\t// to determine shouldRetry or not\n\tpollingGetRequest := core.GetInstanceRequest{\n\t\tInstanceId: createResp.Instance.Id,\n\t\tRequestMetadata: helpers.GetRequestMetadataWithCustomizedRetryPolicy(pollUntilRunning),\n\t}\n\n\tinstance, pollError := c.computeClient.GetInstance(context.Background(), pollingGetRequest)\n\tif pollError != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn *instance.Id, nil\n}", "func (s *LifecyclerRPCServer) BeforeInstance(state *isclib.Instance, resp *struct{}) (err error) {\n\treturn s.Plugin.BeforeInstance(state)\n}", "func Start(args ...string) {\n runInstances(\"Start\", func(i int, id string) error {\n return runDaemon(\"run\", settingsToParams(i, true)...)\n })\n\n if cfg.UseNginx {\n UpdateNginxConf()\n }\n}", "func (c *Client) StartDBInstance(ctx context.Context, params *StartDBInstanceInput, optFns ...func(*Options)) (*StartDBInstanceOutput, error) {\n\tif params == nil {\n\t\tparams = &StartDBInstanceInput{}\n\t}\n\n\tresult, metadata, err := c.invokeOperation(ctx, \"StartDBInstance\", params, optFns, c.addOperationStartDBInstanceMiddlewares)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tout := result.(*StartDBInstanceOutput)\n\tout.ResultMetadata = metadata\n\treturn out, nil\n}", "func (m *MockedManager) Start(context.Context) error {\n\treturn nil\n}", "func (c *InstancesStartCall) Do(call *v1.InstancesStartCall, opts ...googleapi.CallOption) (*v1.Operation, error) {\n\treturn call.Do(opts...)\n}", "func (mgr *Manager) Start(register string) error {\n\terr := mgr.Register(register)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif mgr.tagsMap == nil || mgr.tagsMap[\"pindex\"] {\n\t\tmldd := mgr.options[\"managerLoadDataDir\"]\n\t\tif mldd == \"sync\" || mldd == \"async\" || mldd == \"\" {\n\t\t\terr := mgr.LoadDataDir()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tif mgr.tagsMap == nil || mgr.tagsMap[\"planner\"] {\n\t\tgo mgr.PlannerLoop()\n\t\tgo mgr.PlannerKick(\"start\")\n\t}\n\n\tif mgr.tagsMap == nil ||\n\t\t(mgr.tagsMap[\"pindex\"] && mgr.tagsMap[\"janitor\"]) {\n\t\tgo mgr.JanitorLoop()\n\t\tgo mgr.JanitorKick(\"start\")\n\t}\n\n\treturn mgr.StartCfg()\n}", "func (s *SafeTestingTBOnStart) Start(t testing.TB) error {\n\ts.SetTestingTB(t)\n\treturn nil\n}", "func (s *API) CreateInstance(req *CreateInstanceRequest, opts ...scw.RequestOption) (*Instance, error) {\n\tvar err error\n\n\tdefaultProjectID, exist := s.client.GetDefaultProjectID()\n\tif exist && req.OrganizationID == nil && req.ProjectID == nil {\n\t\treq.ProjectID = &defaultProjectID\n\t}\n\n\tdefaultOrganizationID, exist := s.client.GetDefaultOrganizationID()\n\tif exist && req.OrganizationID == nil && req.ProjectID == nil {\n\t\treq.OrganizationID = &defaultOrganizationID\n\t}\n\n\tif req.Region == \"\" {\n\t\tdefaultRegion, _ := s.client.GetDefaultRegion()\n\t\treq.Region = defaultRegion\n\t}\n\n\tif req.Name == \"\" {\n\t\treq.Name = namegenerator.GetRandomName(\"ins\")\n\t}\n\n\tif fmt.Sprint(req.Region) == \"\" {\n\t\treturn nil, errors.New(\"field Region cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"POST\",\n\t\tPath: \"/rdb/v1/regions/\" + fmt.Sprint(req.Region) + \"/instances\",\n\t\tHeaders: http.Header{},\n\t}\n\n\terr = scwReq.SetBody(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar resp Instance\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func CreateInstance(computeService *compute.Service) (*compute.Operation, error) {\n\tstartupMetadata := fmt.Sprintf(\"#! /bin/bash\\n \\n echo FILESHARE_MOUNT_PRIMARY=\\\"%s\\\" >> /root/.bashrc \\n echo FILESTORE_IP_PRIMARY=\\\"%s\\\" >> /root/.bashrc \\n echo FILESHARE_NAME_PRIMARY=\\\"%s\\\" >> /root/.bashrc \\n echo FILESHARE_MOUNT_SECONDARY=\\\"%s\\\" >> /root/.bashrc \\n echo FILESHARE_NAME_SECONDARY=\\\"%s\\\" >> /root/.bashrc \\n echo FILESTORE_IP_SECONDARY=\\\"%s\\\" >> /root/.bashrc \\n echo GCP_BUCKET_NAME=\\\"%s\\\" >> /root/.bashrc \\n source /root/.bashrc \\n wget -O /root/google-filestore-backup.sh https://raw.githubusercontent.com/sreesanpd/google-filestore-backup-cloud-function/master/resources/google-filestore-backup.sh \\n chmod +x /root/google-filestore-backup.sh \\n /bin/bash /root/google-filestore-backup.sh \\necho \\\"I am STARTING some work at $(date)\\\" | sudo tee -a $HOME/work.txt\", FileshareMountPrimary, FilestoreIpPrimary, FileshareNamePrimary, FileshareMountSecondary, FileshareNameSecondary, FilestoreIpSecondary, GcpBucketName)\n\tshutdownMetadata := fmt.Sprintf(\"#! /bin/bash\\n\\necho \\\"I am FINISHING some work on %s at $(date)\\\" | sudo tee -a $(HOME)/work.txt\", InstanceName)\n\n\tinstance := &compute.Instance{\n\t\tName: InstanceName,\n\t\tMachineType: fmt.Sprintf(\"zones/%s/machineTypes/f1-micro\", Zone),\n\t\tNetworkInterfaces: []*compute.NetworkInterface{\n\t\t\t{\n\t\t\t\tName: VPCName,\n\t\t\t\tSubnetwork: fmt.Sprintf(\"projects/%s/regions/%s/subnetworks/%s\", ProjectID, Region, SubnetName),\n\t\t\t\tAccessConfigs: []*compute.AccessConfig{\n\t\t\t\t\t{\n\t\t\t\t\t\tName: \"External NAT\",\n\t\t\t\t\t\tType: \"ONE_TO_ONE_NAT\",\n\t\t\t\t\t\tNetworkTier: \"PREMIUM\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\tScheduling: &compute.Scheduling{\n\t\t\tPreemptible: true,\n\t\t},\n\t\tDisks: []*compute.AttachedDisk{\n\t\t\t{\n\t\t\t\tBoot: true, // The first disk must be a boot disk.\n\t\t\t\tAutoDelete: true, //Optional\n\t\t\t\tMode: \"READ_WRITE\", //Mode should be READ_WRITE or READ_ONLY\n\t\t\t\tInterface: \"SCSI\", //SCSI or NVME - NVME only for SSDs\n\t\t\t\tInitializeParams: &compute.AttachedDiskInitializeParams{\n\t\t\t\t\tDiskName: \"worker-instance-boot-disk\",\n\t\t\t\t\tSourceImage: \"projects/debian-cloud/global/images/family/debian-9\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\tMetadata: &compute.Metadata{\n\t\t\tItems: []*compute.MetadataItems{\n\t\t\t\t{\n\t\t\t\t\tKey: \"startup-script\",\n\t\t\t\t\tValue: &startupMetadata,\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tKey: \"shutdown-script\",\n\t\t\t\t\tValue: &shutdownMetadata,\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\tServiceAccounts: []*compute.ServiceAccount{\n\t\t\t{\n\t\t\t\tEmail: \"filestorebackups123-gcs-sa@learn-gcp-173102.iam.gserviceaccount.com\",\n\t\t\t\tScopes: []string{\"https://www.googleapis.com/auth/cloud-platform\"},\n\t\t\t\t\n\t\t\t},\n\t\t\t},\n\n\t\t\n\t}\n\treturn computeService.Instances.Insert(ProjectID, Zone, instance).Do()\n}", "func (d *Driver) Start() error {\n\tclient, err := d.getClient()\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn client.StartVirtualMachine(d.vmName())\n}", "func (b *NoopLifecycle) Start(t testing.TB) error {\n\treturn nil\n}", "func (ds *ServerImpl) Start() {\n\t// Gossip the version info so that other nodes don't plan incompatible flows\n\t// for us.\n\tif g, ok := ds.ServerConfig.Gossip.Optional(MultiTenancyIssueNo); ok {\n\t\tif nodeID, ok := ds.ServerConfig.NodeID.OptionalNodeID(); ok {\n\t\t\tif err := g.AddInfoProto(\n\t\t\t\tgossip.MakeDistSQLNodeVersionKey(nodeID),\n\t\t\t\t&execinfrapb.DistSQLVersionGossipInfo{\n\t\t\t\t\tVersion: execinfra.Version,\n\t\t\t\t\tMinAcceptedVersion: execinfra.MinAcceptedVersion,\n\t\t\t\t},\n\t\t\t\t0, // ttl - no expiration\n\t\t\t); err != nil {\n\t\t\t\tpanic(err)\n\t\t\t}\n\t\t}\n\t}\n\n\tif err := ds.setDraining(false); err != nil {\n\t\tpanic(err)\n\t}\n\n\tds.flowScheduler.Start()\n}", "func (s *stepCreateInstance) Run(ctx context.Context, state multistep.StateBag) multistep.StepAction {\n\tclient := state.Get(\"client\").(*civogo.Client)\n\tui := state.Get(\"ui\").(packer.Ui)\n\tc := state.Get(\"config\").(*Config)\n\tsshKeyID := state.Get(\"ssh_key_id\").(string)\n\n\t// Create the instance based on configuration\n\tui.Say(\"Creating instance...\")\n\n\ttemplate, err := client.FindTemplate(c.Template)\n\tif err != nil {\n\t\tui.Error(err.Error())\n\t}\n\n\tnetwork, _ := client.GetDefaultNetwork()\n\n\tInstanceConfig := &civogo.InstanceConfig{\n\t\tHostname: c.InstanceName,\n\t\tPublicIPRequired: c.PublicNetworking,\n\t\tRegion: c.Region,\n\t\tNetworkID: network.ID,\n\t\tInitialUser: c.Comm.SSHUsername,\n\t\tSize: c.Size,\n\t\tTemplateID: template.ID,\n\t\tSSHKeyID: sshKeyID,\n\t}\n\n\tlog.Printf(\"[DEBUG] Instance create paramaters: %+v\", InstanceConfig)\n\n\tinstance, err := client.CreateInstance(InstanceConfig)\n\tif err != nil {\n\t\terr := fmt.Errorf(\"Error creating instance: %s\", err)\n\t\tstate.Put(\"error\", err)\n\t\tui.Error(err.Error())\n\t\treturn multistep.ActionHalt\n\t}\n\n\t// We use this in cleanup\n\ts.instanceID = instance.ID\n\n\t// Store the instance id for later\n\tstate.Put(\"instance_id\", instance.ID)\n\n\treturn multistep.ActionContinue\n}", "func (e *exec) start(ctx context.Context) {\n\t// Lock the mutex to prevent race conditions with Stop\n\te.execMutex.Lock()\n\tdefer e.execMutex.Unlock()\n\n\t// Do the startup sequence once until the shutdown sequence resets\n\te.startOnce.Do(func() {\n\t\tdefer func() {\n\t\t\t// reset stopOnce so the shutdown sequence can happen again\n\t\t\te.stopOnce = sync.Once{}\n\t\t}()\n\t\te.startFn(ctx)\n\t})\n}", "func (px *Paxos) Start(seq int, v interface{}) {\n\t//DPrintf(\"Start(%d, %v)\\n\", seq, v)\n\tgo px.propose(seq, v)\n}", "func (p *spaDevProxy) Start(ctx context.Context) error {\n\tif _, err := os.Stat(p.options.Dir); err != nil {\n\t\treturn err\n\t}\n\tpath, args := prepareRunner(p.options.RunnerType, p.options.ScriptName, p.options.Args...)\n\tp.cmd = newCommand(ctx, path, args...)\n\tp.cmd.SysProcAttr = &syscall.SysProcAttr{Setpgid: true}\n\tp.cmd.Env = append(p.options.Env, fmt.Sprintf(\"PATH=%s\", os.Getenv(\"PATH\")))\n\tp.cmd.Dir = p.options.Dir\n\n\tstdout, err := p.cmd.StdoutPipe()\n\tif err != nil {\n\t\treturn err\n\t}\n\tstderr, err := p.cmd.StderrPipe()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdone := p.forwardOutput(stdout, stderr)\n\n\terr = p.cmd.Start()\n\n\t<-done\n\n\treturn err\n}", "func NewInstance(ctx *pulumi.Context,\n\tname string, args *InstanceArgs, opts ...pulumi.ResourceOption) (*Instance, error) {\n\tif args == nil || args.Type == nil {\n\t\treturn nil, errors.New(\"missing required argument 'Type'\")\n\t}\n\tif args == nil {\n\t\targs = &InstanceArgs{}\n\t}\n\tvar resource Instance\n\terr := ctx.RegisterResource(\"gcp:datafusion/instance:Instance\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (sm *SM) Start() error {\n\treturn nil\n}", "func (runner *MockRunner) Start(runnable runner.Runnable) {\n\trunner.Called(runnable)\n}", "func (s *Service) Start(sc service.Context) error {\n\ts.sc = sc\n\n\ts.routines.Add(1)\n\tgo s.executionManagerLoop()\n\n\treturn nil\n}", "func (s *Server) Start() error {\n\t// Check if the server is running\n\tif s.IsRunning() {\n\t\treturn errors.New(\"Attempted to start a server that is already running at address: \" + s.instance.Addr)\n\t}\n\n\t// Set routes and middleware\n\ts.SetRoutes()\n\ts.GetInstance().Handler = middleware.NewMiddleware().Then(s.GetInstance().Handler)\n\n\tm := \"Listening for requests...\"\n\tlog.Info(m)\n\n\tgo s.GetInstance().ListenAndServe()\n\n\ts.running = true\n\n\treturn nil\n}", "func (g GCPClient) CreateInstance(name, image, zone, machineType string, disks Disks, data *string, nested, vtpm, replace bool) error {\n\tif replace {\n\t\tif err := g.DeleteInstance(name, zone, true); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tlog.Infof(\"Creating instance %s from image %s (type: %s in %s)\", name, image, machineType, zone)\n\n\tenabled := new(string)\n\t*enabled = \"1\"\n\n\tk, err := ssh.NewPublicKey(g.privKey.Public())\n\tif err != nil {\n\t\treturn err\n\t}\n\tsshKey := new(string)\n\t*sshKey = fmt.Sprintf(\"moby:%s moby\", string(ssh.MarshalAuthorizedKey(k)))\n\n\t// check provided image to be compatible with provided options\n\top, err := g.compute.Images.Get(g.projectName, image).Do()\n\tif err != nil {\n\t\treturn err\n\t}\n\tuefiCompatible := false\n\tfor _, feature := range op.GuestOsFeatures {\n\t\tif feature != nil && feature.Type == uefiCompatibleFeature {\n\t\t\tuefiCompatible = true\n\t\t\tbreak\n\t\t}\n\t}\n\tif vtpm && !uefiCompatible {\n\t\treturn fmt.Errorf(\"cannot use vTPM without UEFI_COMPATIBLE image\")\n\t}\n\t// we should check for nested\n\tvmxLicense := false\n\tfor _, license := range op.Licenses {\n\t\t// we omit hostname and version when define license\n\t\tif strings.HasSuffix(license, vmxImageLicence) {\n\t\t\tvmxLicense = true\n\t\t\tbreak\n\t\t}\n\t}\n\tif nested && !vmxLicense {\n\t\treturn fmt.Errorf(\"cannot use nested virtualization without enable-vmx image\")\n\t}\n\n\tinstanceDisks := []*compute.AttachedDisk{\n\t\t{\n\t\t\tAutoDelete: true,\n\t\t\tBoot: true,\n\t\t\tInitializeParams: &compute.AttachedDiskInitializeParams{\n\t\t\t\tSourceImage: fmt.Sprintf(\"global/images/%s\", image),\n\t\t\t},\n\t\t},\n\t}\n\n\tfor i, disk := range disks {\n\t\tvar diskName string\n\t\tif disk.Path != \"\" {\n\t\t\tdiskName = disk.Path\n\t\t} else {\n\t\t\tdiskName = fmt.Sprintf(\"%s-disk-%d\", name, i)\n\t\t}\n\t\tvar diskSizeGb int64\n\t\tif disk.Size == 0 {\n\t\t\tdiskSizeGb = int64(1)\n\t\t} else {\n\t\t\tdiskSizeGb = int64(convertMBtoGB(disk.Size))\n\t\t}\n\t\tdiskObj := &compute.Disk{Name: diskName, SizeGb: diskSizeGb}\n\t\tif vtpm {\n\t\t\tdiskObj.GuestOsFeatures = []*compute.GuestOsFeature{\n\t\t\t\t{Type: uefiCompatibleFeature},\n\t\t\t}\n\t\t}\n\t\tdiskOp, err := g.compute.Disks.Insert(g.projectName, zone, diskObj).Do()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif err := g.pollZoneOperationStatus(diskOp.Name, zone); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tinstanceDisks = append(instanceDisks, &compute.AttachedDisk{\n\t\t\tAutoDelete: true,\n\t\t\tBoot: false,\n\t\t\tSource: fmt.Sprintf(\"zones/%s/disks/%s\", zone, diskName),\n\t\t})\n\t}\n\n\tinstanceObj := &compute.Instance{\n\t\tMachineType: fmt.Sprintf(\"zones/%s/machineTypes/%s\", zone, machineType),\n\t\tName: name,\n\t\tDisks: instanceDisks,\n\t\tNetworkInterfaces: []*compute.NetworkInterface{\n\t\t\t{\n\t\t\t\tNetwork: \"global/networks/default\",\n\t\t\t\tAccessConfigs: []*compute.AccessConfig{\n\t\t\t\t\t{\n\t\t\t\t\t\tType: \"ONE_TO_ONE_NAT\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\tMetadata: &compute.Metadata{\n\t\t\tItems: []*compute.MetadataItems{\n\t\t\t\t{\n\t\t\t\t\tKey: \"serial-port-enable\",\n\t\t\t\t\tValue: enabled,\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tKey: \"ssh-keys\",\n\t\t\t\t\tValue: sshKey,\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tKey: \"user-data\",\n\t\t\t\t\tValue: data,\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\n\tif nested {\n\t\tinstanceObj.MinCpuPlatform = \"Intel Haswell\"\n\t}\n\tif vtpm {\n\t\tinstanceObj.ShieldedInstanceConfig = &compute.ShieldedInstanceConfig{EnableVtpm: true}\n\t}\n\n\t// Don't wait for operation to complete!\n\t// A headstart is needed as by the time we've polled for this event to be\n\t// completed, the instance may have already terminated\n\t_, err = g.compute.Instances.Insert(g.projectName, zone, instanceObj).Do()\n\tif err != nil {\n\t\treturn err\n\t}\n\tlog.Infof(\"Instance created\")\n\treturn nil\n}", "func NewInstance(name string, opts ...Option) (Instance, error) {\n\tif name == \"\" {\n\t\treturn nil, ErrNoName\n\t}\n\ti := &instance{\n\t\tname: name,\n\t\tbindAddr: \"127.0.0.1\",\n\t\tbindPort: 8100,\n\t\tshutdownCh: make(chan struct{}),\n\t\tpeers: make(map[string]*Peer),\n\t\tcaches: make(map[string]*cache),\n\t\tlogger: log.New(os.Stdout, \"huton\", log.LstdFlags),\n\t\traftApplicationTimeout: 10 * time.Second,\n\t\traftTransportTimeout: 10 * time.Second,\n\t\traftRetainSnapshotCount: 2,\n\t\tserfEventChannel: make(chan serf.Event, 256),\n\t}\n\tfor _, opt := range opts {\n\t\topt(i)\n\t}\n\ti.logger.Println(\"Initializing RPC server...\")\n\tif err := i.setupRPC(); err != nil {\n\t\ti.Shutdown()\n\t\treturn i, err\n\t}\n\ti.logger.Println(\"Initializing Raft cluster...\")\n\tif err := i.setupRaft(); err != nil {\n\t\ti.Shutdown()\n\t\treturn i, err\n\t}\n\tip := net.ParseIP(i.bindAddr)\n\traftAddr := &net.TCPAddr{\n\t\tIP: ip,\n\t\tPort: i.bindPort + 1,\n\t}\n\trpcAddr := &net.TCPAddr{\n\t\tIP: ip,\n\t\tPort: i.bindPort + 2,\n\t}\n\n\ti.logger.Println(\"Initializing Serf cluster...\")\n\tif err := i.setupSerf(raftAddr, rpcAddr); err != nil {\n\t\ti.Shutdown()\n\t\treturn i, err\n\t}\n\tgo i.handleEvents()\n\treturn i, nil\n}", "func Start(\n\tctx context.Context,\n\tconfig Config,\n\tdbConfig fw.DBConfig,\n\tdbConnector fw.DBConnector,\n\tdbMigrationTool fw.DBMigrationTool,\n\tsecurityPolicy fw.SecurityPolicy,\n\teventDispatcher fw.Dispatcher,\n) {\n\tdb, err := dbConnector.Connect(dbConfig)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\terr = dbMigrationTool.MigrateUp(db, config.MigrationRoot)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tgRpcService, err := dep.InitGRpcService(\n\t\tconfig.ServiceName,\n\t\tconfig.LogLevel,\n\t\tprovider.ServiceEmailAddress(config.ServiceEmailAddress),\n\t\tdb,\n\t\tsecurityPolicy,\n\t\tprovider.SendGridAPIKey(config.SendGridAPIKey),\n\t\tprovider.TemplateRootDir(config.TemplateRootDir),\n\t\tprovider.CacheSize(config.CacheSize),\n\t\teventDispatcher,\n\t)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tgo func() {\n\t\t<-ctx.Done()\n\t\tgRpcService.Stop()\n\t}()\n\n\tgRpcService.Start(config.GRpcAPIPort)\n}", "func (c *MockController) Start() error {\n\tc.StartFuncCalled++\n\n\treturn c.StartFunc()\n}", "func (this *ActionHandler) Start() {\n\tgo wait.Until(this.getNextSucceededVMTEvent, 0, this.config.StopEverything)\n\tgo wait.Until(this.getNextFailedVMTEvent, 0, this.config.StopEverything)\n\n\tthis.actionSupervisor.Start()\n}", "func (wc *workflowClient) StartWorkflow(\n\tctx context.Context,\n\toptions StartWorkflowOptions,\n\tworkflowFunc interface{},\n\targs ...interface{},\n) (*WorkflowExecution, error) {\n\tworkflowID := options.ID\n\tif len(workflowID) == 0 {\n\t\tworkflowID = uuid.NewRandom().String()\n\t}\n\n\tif options.TaskList == \"\" {\n\t\treturn nil, errors.New(\"missing TaskList\")\n\t}\n\n\texecutionTimeout := common.Int32Ceil(options.ExecutionStartToCloseTimeout.Seconds())\n\tif executionTimeout <= 0 {\n\t\treturn nil, errors.New(\"missing or invalid ExecutionStartToCloseTimeout\")\n\t}\n\n\tdecisionTaskTimeout := common.Int32Ceil(options.DecisionTaskStartToCloseTimeout.Seconds())\n\tif decisionTaskTimeout < 0 {\n\t\treturn nil, errors.New(\"negative DecisionTaskStartToCloseTimeout provided\")\n\t}\n\tif decisionTaskTimeout == 0 {\n\t\tdecisionTaskTimeout = defaultDecisionTaskTimeoutInSecs\n\t}\n\n\t// Validate type and its arguments.\n\tworkflowType, input, err := getValidatedWorkflowFunction(workflowFunc, args, wc.dataConverter, wc.registry)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tmemo, err := getWorkflowMemo(options.Memo, wc.dataConverter)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tsearchAttr, err := serializeSearchAttributes(options.SearchAttributes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdelayStartSeconds := common.Int32Ceil(options.DelayStart.Seconds())\n\tif delayStartSeconds < 0 {\n\t\treturn nil, errors.New(\"Invalid DelayStart option\")\n\t}\n\n\tjitterStartSeconds := common.Int32Ceil(options.JitterStart.Seconds())\n\tif jitterStartSeconds < 0 {\n\t\treturn nil, errors.New(\"Invalid JitterStart option\")\n\t}\n\n\t// create a workflow start span and attach it to the context object.\n\t// N.B. we need to finish this immediately as jaeger does not give us a way\n\t// to recreate a span given a span context - which means we will run into\n\t// issues during replay. we work around this by creating and ending the\n\t// workflow start span and passing in that context to the workflow. So\n\t// everything beginning with the StartWorkflowExecutionRequest will be\n\t// parented by the created start workflow span.\n\tctx, span := createOpenTracingWorkflowSpan(ctx, wc.tracer, time.Now(), fmt.Sprintf(\"StartWorkflow-%s\", workflowType.Name), workflowID)\n\tspan.Finish()\n\n\t// get workflow headers from the context\n\theader := wc.getWorkflowHeader(ctx)\n\n\t// run propagators to extract information about tracing and other stuff, store in headers field\n\tstartRequest := &s.StartWorkflowExecutionRequest{\n\t\tDomain: common.StringPtr(wc.domain),\n\t\tRequestId: common.StringPtr(uuid.New()),\n\t\tWorkflowId: common.StringPtr(workflowID),\n\t\tWorkflowType: workflowTypePtr(*workflowType),\n\t\tTaskList: common.TaskListPtr(s.TaskList{Name: common.StringPtr(options.TaskList)}),\n\t\tInput: input,\n\t\tExecutionStartToCloseTimeoutSeconds: common.Int32Ptr(executionTimeout),\n\t\tTaskStartToCloseTimeoutSeconds: common.Int32Ptr(decisionTaskTimeout),\n\t\tIdentity: common.StringPtr(wc.identity),\n\t\tWorkflowIdReusePolicy: options.WorkflowIDReusePolicy.toThriftPtr(),\n\t\tRetryPolicy: convertRetryPolicy(options.RetryPolicy),\n\t\tCronSchedule: common.StringPtr(options.CronSchedule),\n\t\tMemo: memo,\n\t\tSearchAttributes: searchAttr,\n\t\tHeader: header,\n\t\tDelayStartSeconds: common.Int32Ptr(delayStartSeconds),\n\t\tJitterStartSeconds: common.Int32Ptr(jitterStartSeconds),\n\t}\n\n\tvar response *s.StartWorkflowExecutionResponse\n\n\t// Start creating workflow request.\n\terr = backoff.Retry(ctx,\n\t\tfunc() error {\n\t\t\ttchCtx, cancel, opt := newChannelContext(ctx, wc.featureFlags)\n\t\t\tdefer cancel()\n\n\t\t\tvar err1 error\n\t\t\tresponse, err1 = wc.workflowService.StartWorkflowExecution(tchCtx, startRequest, opt...)\n\t\t\treturn err1\n\t\t}, createDynamicServiceRetryPolicy(ctx), isServiceTransientError)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif wc.metricsScope != nil {\n\t\tscope := wc.metricsScope.GetTaggedScope(tagTaskList, options.TaskList, tagWorkflowType, workflowType.Name)\n\t\tscope.Counter(metrics.WorkflowStartCounter).Inc(1)\n\t}\n\n\texecutionInfo := &WorkflowExecution{\n\t\tID: workflowID,\n\t\tRunID: response.GetRunId()}\n\treturn executionInfo, nil\n}", "func (s *AutoscalerSuite) TestInstanceLaunching(c *check.C) {\n\tclusterName := \"bob\"\n\tinstance := &gaws.Instance{\n\t\tID: \"instance-1\",\n\t}\n\tqueue := newMockQueue(\"queue-1\")\n\tec := newMockEC2(&ec2.Instance{\n\t\tInstanceId: aws.String(\"instance-1\"),\n\t})\n\ta, err := New(Config{\n\t\tClusterName: clusterName,\n\t\tNewLocalInstance: func() (*gaws.Instance, error) {\n\t\t\treturn instance, nil\n\t\t},\n\t\tQueue: queue,\n\t\tCloud: ec,\n\t})\n\tc.Assert(err, check.IsNil)\n\tc.Assert(a, check.NotNil)\n\n\tctx, cancel := context.WithCancel(context.TODO())\n\tdefer cancel()\n\top := newMockOperator(ops.Site{\n\t\tAccountID: \"1\",\n\t\tDomain: \"example.com\",\n\t\tClusterState: storage.ClusterState{},\n\t})\n\tgo a.ProcessEvents(ctx, queue.url, op)\n\n\t// send launched event\n\tinstanceID := \"instance-1\"\n\tmsg := &message{\n\t\treceipt: \"message-1\",\n\t\tbody: mustMarshalHook(HookEvent{\n\t\t\tInstanceID: instanceID,\n\t\t\tType: InstanceLaunching,\n\t\t}),\n\t}\n\tselect {\n\tcase <-time.After(time.Second):\n\t\tc.Fatalf(\"timeout\")\n\tcase queue.messagesC <- msg:\n\t}\n\n\t// expect the turn off request to arrive\n\tselect {\n\tcase input := <-ec.modifyC:\n\t\tc.Assert(aws.StringValue(input.InstanceId), check.DeepEquals, instanceID)\n\t\tc.Assert(input.SourceDestCheck.Value, check.DeepEquals, aws.Bool(false))\n\tcase <-time.After(time.Second):\n\t\tc.Fatalf(\"timeout\")\n\t}\n\n\t// expect the message to be scheduled for deletion\n\tselect {\n\tcase m := <-queue.deletedC:\n\t\tc.Assert(aws.StringValue(m.ReceiptHandle), check.DeepEquals, msg.receipt)\n\tcase <-time.After(time.Second):\n\t\tc.Fatalf(\"timeout\")\n\t}\n}", "func (p *process) Start() error {\n\tif p.ID() == InitProcessID {\n\t\tvar (\n\t\t\terrC = make(chan error, 1)\n\t\t\targs = append(p.container.runtimeArgs, \"start\", p.container.id)\n\t\t\tcmd = exec.Command(p.container.runtime, args...)\n\t\t)\n\t\tgo func() {\n\t\t\tout, err := cmd.CombinedOutput()\n\t\t\tif err != nil {\n\t\t\t\terrC <- fmt.Errorf(\"%s: %q\", err.Error(), out)\n\t\t\t}\n\t\t\terrC <- nil\n\t\t}()\n\t\tselect {\n\t\tcase err := <-errC:\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\tcase <-p.cmdDoneCh:\n\t\t\tif !p.cmdSuccess {\n\t\t\t\tif cmd.Process != nil {\n\t\t\t\t\tcmd.Process.Kill()\n\t\t\t\t}\n\t\t\t\tcmd.Wait()\n\t\t\t\treturn ErrShimExited\n\t\t\t}\n\t\t\terr := <-errC\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func (i *InstanceServiceHandler) MassStart(ctx context.Context, instanceList []string) error {\n\turi := fmt.Sprintf(\"%s/start\", instancePath)\n\n\treqBody := RequestBody{\"instance_ids\": instanceList}\n\treq, err := i.client.NewRequest(ctx, http.MethodPost, uri, reqBody)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn i.client.DoWithContext(ctx, req, nil)\n}", "func (m *mockAPI) Start() {\n\tif m.isServing() {\n\t\treturn\n\t}\n\n\tm.Server.Start()\n\tm.Lock()\n\tdefer m.Unlock()\n\tm.serving = true\n}", "func (m *Manager) Start(ctx context.Context, uuid string) error {\n\tm.mu.Lock()\n\tdefer m.mu.Unlock()\n\n\t// Check the manager is running.\n\tif m.ctx == nil {\n\t\treturn fmt.Errorf(\"manager not running\")\n\t}\n\n\trw, ok := m.workflows[uuid]\n\tif !ok {\n\t\treturn fmt.Errorf(\"cannot find workflow %v in the workflow list\", uuid)\n\t}\n\n\tif rw.wi.State != workflowpb.WorkflowState_NotStarted {\n\t\treturn fmt.Errorf(\"workflow with uuid %v is in state %v\", uuid, rw.wi.State)\n\t}\n\n\t// Change its state in the topo server. Note we do that first,\n\t// so if the running part fails, we will retry next time.\n\trw.wi.State = workflowpb.WorkflowState_Running\n\trw.wi.StartTime = time.Now().Unix()\n\tif err := m.ts.SaveWorkflow(ctx, rw.wi); err != nil {\n\t\treturn err\n\t}\n\n\trw.rootNode.State = workflowpb.WorkflowState_Running\n\trw.rootNode.BroadcastChanges(false /* updateChildren */)\n\n\tm.runWorkflow(rw)\n\treturn nil\n}", "func InitializeInstance(instance IService) {\n\tInstance = instance\n}", "func (s *ServerFx) Start() error {\n\terr := s.app.Start(context.Background())\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif s.startupSynchronizationMode.blockingStart {\n\t\t// If s.so.interruptCh is nil this will wait forever.\n\t\tinterruptSignal := <-s.startupSynchronizationMode.interruptCh\n\t\ts.logger.Info(\"Received interrupt signal, stopping the server.\", tag.Value(interruptSignal))\n\t\treturn s.Stop()\n\t}\n\n\treturn nil\n}", "func NewInstance(opts *Options) (Instance, error) {\n\ti := &instance{\n\t\topts: opts,\n\t}\n\tif err := i.startChild(); err != nil {\n\t\treturn nil, err\n\t}\n\treturn i, nil\n}" ]
[ "0.7707315", "0.7707315", "0.67058474", "0.6648973", "0.6615673", "0.6606206", "0.66045177", "0.6551344", "0.6458486", "0.6319141", "0.62782216", "0.61157715", "0.60638297", "0.5930062", "0.5837253", "0.5821408", "0.58164716", "0.5769161", "0.5676736", "0.56507826", "0.5643034", "0.5584573", "0.55774844", "0.5492427", "0.54570913", "0.54554087", "0.5447109", "0.54445016", "0.5438144", "0.54078215", "0.5398018", "0.53613764", "0.5351213", "0.53475314", "0.53367233", "0.5335751", "0.53285265", "0.5325632", "0.531986", "0.53021926", "0.53015095", "0.53004", "0.5279434", "0.526382", "0.52587354", "0.5251064", "0.52339727", "0.52233315", "0.5215909", "0.52157396", "0.52101284", "0.52023774", "0.51750696", "0.5173915", "0.5163141", "0.5145306", "0.51404935", "0.5138206", "0.51302433", "0.512065", "0.51132596", "0.51005334", "0.5087043", "0.50836307", "0.5081399", "0.50709486", "0.50682706", "0.5066453", "0.50655365", "0.5065101", "0.50625587", "0.5047114", "0.504263", "0.5042089", "0.5041069", "0.5040695", "0.5033321", "0.50317144", "0.5016755", "0.50165075", "0.5004218", "0.5001589", "0.4997671", "0.4996238", "0.49941126", "0.49892578", "0.49864435", "0.4980555", "0.49771452", "0.49718705", "0.49677557", "0.4961693", "0.49613565", "0.49607274", "0.49606836", "0.49551946", "0.49511722", "0.49503955", "0.49471393", "0.4945058" ]
0.7028185
2
StopInstance uses the override method StopInstanceFn or the real implementation.
func (c *TestClient) StopInstance(project, zone, name string) error { if c.StopInstanceFn != nil { return c.StopInstanceFn(project, zone, name) } return c.client.StopInstance(project, zone, name) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (p *OnPrem) StopInstance(ctx *Context, instancename string) error {\n\treturn fmt.Errorf(\"Operation not supported\")\n}", "func (p *OnPrem) StopInstance(ctx *Context, instancename string) error {\n\treturn fmt.Errorf(\"Operation not supported\")\n}", "func (c *Client) StopInstance(instanceId string, forceStop bool) error {\n\targs := &StopInstanceArgs{\n\t\tForceStop: forceStop,\n\t}\n\n\tjsonBytes, jsonErr := json.Marshal(args)\n\tif jsonErr != nil {\n\t\treturn jsonErr\n\t}\n\tbody, err := bce.NewBodyFromBytes(jsonBytes)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn StopInstance(c, instanceId, body)\n}", "func StopInstance(computeService *compute.Service) (*compute.Operation, error) {\n\treturn computeService.Instances.Stop(ProjectID, Zone, InstanceName).Do()\n}", "func (c *Client) StopInstance(id string) error {\n\n\tactionRequest := core.InstanceActionRequest{}\n\tactionRequest.Action = core.InstanceActionActionStop\n\tactionRequest.InstanceId = &id\n\n\tstopResp, err := c.computeClient.InstanceAction(context.Background(), actionRequest)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// wait until lifecycle status is Stopped\n\tpollUntilStopped := func(r common.OCIOperationResponse) bool {\n\t\tif converted, ok := r.Response.(core.GetInstanceResponse); ok {\n\t\t\treturn converted.LifecycleState != core.InstanceLifecycleStateStopped\n\t\t}\n\t\treturn true\n\t}\n\n\tpollingGetRequest := core.GetInstanceRequest{\n\t\tInstanceId: stopResp.Instance.Id,\n\t\tRequestMetadata: helpers.GetRequestMetadataWithCustomizedRetryPolicy(pollUntilStopped),\n\t}\n\n\t_, err = c.computeClient.GetInstance(context.Background(), pollingGetRequest)\n\n\treturn err\n}", "func (p *ProxMox) StopInstance(ctx *lepton.Context, instanceID string) error {\n\n\treq, err := http.NewRequest(\"POST\", p.apiURL+\"/api2/json/nodes/\"+p.nodeNAME+\"/qemu/\"+instanceID+\"/status/stop\", nil)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\ttr := &http.Transport{\n\t\tTLSClientConfig: &tls.Config{InsecureSkipVerify: true},\n\t}\n\tclient := &http.Client{Transport: tr}\n\n\treq.Header.Add(\"Authorization\", \"PVEAPIToken=\"+p.tokenID+\"=\"+p.secret)\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\t_, err = io.ReadAll(resp.Body)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func Stop() {\n\tinstance.stop()\n}", "func (inst *Instance) Stop(signal os.Signal) error {\n\tautoRestartHandle := inst.autoRestartHandle\n\n\t// send stop signal\n\tautoRestartHandle.mask()\n\treturn inst.stop(signal)\n}", "func stopInstance(cs *compute.Service, w http.ResponseWriter) {\n\toperation, err := StopInstance(cs)\n\tif err != nil {\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\tlog.Fatal(err)\n\t}\n\tw.WriteHeader(http.StatusOK)\n\tdata, _ := operation.MarshalJSON()\n\tw.Write(data)\n}", "func (client *Client) StopInstancesWithCallback(request *StopInstancesRequest, callback func(response *StopInstancesResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *StopInstancesResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.StopInstances(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (this *AVTransport) Stop(instanceId uint32) error {\n\ttype Response struct {\n\t\tXMLName xml.Name\n\t\tErrorResponse\n\t}\n\targs := []Arg{\n\t\t{\"InstanceID\", instanceId},\n\t}\n\tresponse := this.Svc.Call(\"Stop\", args)\n\tdoc := Response{}\n\txml.Unmarshal([]byte(response), &doc)\n\treturn doc.Error()\n}", "func TerminateInstance() {\n\tlog.Println(\"Starting to run terminate instance process\")\n\t// Get instance id and region from metadata\n\tinstanceId, instanceRegion := getInstanceID()\n\tlog.Printf(\"Working on %v in %v region\", instanceId, instanceRegion)\n\n\t// Init aws session\n\tawsSession,_ := initSession(instanceRegion)\n\tlog.Println(\"Initialized aws session\")\n\n\t// Init Aws auto scaling session\n\tinitAutoScalingAwsSession(awsSession)\n\tlog.Println(\"Initialized auto scaling session\")\n\n\t// Get auto scaling group name\n\tinstanceAutoScaleGroupName := getAutoScalingName(instanceId)\n\tlog.Printf(\"Instance %v auto scaling group name is: %v\", instanceId, instanceAutoScaleGroupName)\n\n\t// Set instance scale in policy to false\n\tsuccess := setScaleInProtectionToInstance(instanceAutoScaleGroupName, instanceId)\n\n\t// Terminate ec2 instance after setting scale in policy to false\n\tif success{\n\t\tterminateInstance(instanceId)\n\t}\n}", "func Stop(args ...string) {\n switch {\n case cfg.Kill:\n Kill(args...)\n default:\n runInstances(\"Stopping\", func(i int, id string) error {\n defer os.Remove(pidFileName(i))\n return run(\"stop\", id)\n })\n }\n}", "func (instance *Host) Stop(ctx context.Context) (ferr fail.Error) {\n\tdefer fail.OnPanic(&ferr)\n\n\tif valid.IsNil(instance) {\n\t\treturn fail.InvalidInstanceError()\n\t}\n\tif ctx == nil {\n\t\treturn fail.InvalidParameterCannotBeNilError(\"ctx\")\n\t}\n\n\thostName := instance.GetName()\n\thostID, err := instance.GetID()\n\tif err != nil {\n\t\treturn fail.ConvertError(err)\n\t}\n\n\tsvc := instance.Service()\n\n\ttimings, xerr := instance.Service().Timings()\n\tif xerr != nil {\n\t\treturn xerr\n\t}\n\n\txerr = instance.Sync(ctx)\n\tif xerr != nil {\n\t\tlogrus.WithContext(ctx).Debugf(\"failure trying to sync: %v\", xerr)\n\t}\n\n\txerr = svc.StopHost(ctx, hostID, false)\n\txerr = debug.InjectPlannedFail(xerr)\n\tif xerr != nil {\n\t\treturn xerr\n\t}\n\n\txerr = retry.WhileUnsuccessful(\n\t\tfunc() error {\n\t\t\tselect {\n\t\t\tcase <-ctx.Done():\n\t\t\t\treturn retry.StopRetryError(ctx.Err())\n\t\t\tdefault:\n\t\t\t}\n\n\t\t\ths, err := instance.ForceGetState(ctx)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tif hs != hoststate.Stopped {\n\t\t\t\treturn fail.NewError(\"%s not stopped yet: %s\", hostName, hs.String())\n\t\t\t}\n\n\t\t\treturn nil\n\t\t},\n\t\ttimings.NormalDelay(),\n\t\ttimings.ExecutionTimeout(),\n\t)\n\txerr = debug.InjectPlannedFail(xerr)\n\tif xerr != nil {\n\t\tswitch xerr.(type) {\n\t\tcase *fail.ErrAborted:\n\t\t\tif cerr := fail.ConvertError(fail.Cause(xerr)); cerr != nil {\n\t\t\t\treturn cerr\n\t\t\t}\n\t\t\treturn xerr\n\t\tcase *retry.ErrTimeout:\n\t\t\treturn fail.Wrap(xerr, \"timeout waiting Host '%s' to be stopped\", hostName)\n\t\tdefault:\n\t\t\treturn xerr\n\t\t}\n\t}\n\n\t// Now unsafeReload\n\txerr = instance.unsafeReload(ctx)\n\txerr = debug.InjectPlannedFail(xerr)\n\tif xerr != nil {\n\t\treturn xerr\n\t}\n\n\treturn nil\n}", "func (a API) Stop(ctx context.Context) error {\n\treturn a.srv.Shutdown(ctx)\n}", "func (c *TestClient) InstanceStopped(project, zone, name string) (bool, error) {\n\tif c.InstanceStoppedFn != nil {\n\t\treturn c.InstanceStoppedFn(project, zone, name)\n\t}\n\treturn c.client.InstanceStopped(project, zone, name)\n}", "func (f *RemoteRuntime) Stop() {\n\tf.server.Stop()\n}", "func (a *App) Stop() error {\n\tif a.opts.registrar != nil && a.instance != nil {\n\t\tctx, cancel := context.WithTimeout(a.opts.ctx, a.opts.registrarTimeout)\n\t\tdefer cancel()\n\t\tif err := a.opts.registrar.Deregister(ctx, a.instance); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tif a.cancel != nil {\n\t\ta.cancel()\n\t}\n\treturn nil\n}", "func (m *DomainMonitor) Stop() {\n\tm.mutex.Lock()\n\tif m.instance != nil {\n\t\tm.instance.Close() // TODO: Decide whether blocking here is acceptable\n\t\tm.instance = nil\n\t}\n\tm.mutex.Unlock()\n}", "func (il *InstanceLogger) Stop() {\n\t// Closes the client and flushes the buffer to Stackdriver\n\tif il.client != nil {\n\t\til.client.Close()\n\t\til.client = nil\n\t} else if il.cancelFunc != nil {\n\t\til.cancelFunc()\n\t\til.cancelFunc = nil\n\t}\n}", "func (m *Master) Stop(procSign string, out *StopRsp) error {\n\trtnCode, err := m.StopInstance(procSign, syscall.SIGINT)\n\tif err != nil {\n\t\treturn err\n\t}\n\t*out = StopRsp{\n\t\tReturnCode: rtnCode,\n\t\tProcSign: procSign,\n\t}\n\treturn nil\n}", "func (d *driverMock) TerminateInstance(ctx context.Context, id string) error {\n\tif d.TerminateInstanceErr != nil {\n\t\treturn d.TerminateInstanceErr\n\t}\n\n\td.TerminateInstanceID = id\n\n\treturn nil\n}", "func (vm *VirtualMachine) Stop(args *DomainXML, reply *bool) error {\n\t// Passing the true parameter to ensure the stop vm task is added to waitgroup as this action needs to be completed\n\t// even if there is pending signal termination on rpc\n\t_, err := proc.AddTask(true)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"rpc/server:Stop() Could not add task for vm stop\")\n\t}\n\tdefer proc.TaskDone()\n\n\tlog.Trace(\"rpc/server:Stop() Entering\")\n\tdefer log.Trace(\"rpc/server:Stop() Leaving\")\n\n\twlaMtx.Lock()\n\tdefer wlaMtx.Unlock()\n\n\tif err = validation.ValidateXMLString(args.XML); err != nil {\n\t\tsecLog.Errorf(\"rpc:server() Stop: %s, Invalid domain XML format\", message.InvalidInputBadParam)\n\t\treturn nil\n\t}\n\n\t// pass in vm.Watcher to get the instance to the File System Watcher\n\t*reply = wlavm.Stop(args.XML, vm.Watcher)\n\treturn nil\n}", "func StopSingleInstance() {\n\tif singleInstance != nil {\n\t\tsingleInstanceLock.Lock()\n\t\tdefer singleInstanceLock.Unlock()\n\n\t\tif singleInstance != nil {\n\t\t\tstopSingleInstance(singleInstance.path)\n\t\t}\n\t}\n}", "func (f *ControlPlane) Stop() error {\n\tif err := f.APIServer.Stop(); err != nil {\n\t\treturn err\n\t}\n\treturn f.Etcd.Stop()\n}", "func (e StaticFeeEstimator) Stop() error {\n\treturn nil\n}", "func (c *MockAzureCloud) DetachInstance(i *cloudinstances.CloudInstance) error {\n\treturn errors.New(\"DetachInstance not implemented on azureCloud\")\n}", "func (vm *VirtualMachine) Stop(client SkytapClient) (*VirtualMachine, error) {\n\tlog.WithFields(log.Fields{\"vmId\": vm.Id}).Info(\"Stopping VM\")\n\n /*\n\t Need to check current machine state as transitioning from suspended to stopped is not valid.\n\t*/\n\tcheckVm, err := GetVirtualMachine(client, vm.Id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif checkVm.Runstate == RunStatePause {\n\t\treturn nil, fmt.Errorf(\"Unable to stop a suspended VM.\")\n\t}\n\n\t/*\n There are cases where the call will succeed but the VM cannot be transitioned\n\t to stopped. Generally this is a case where the VM was started and immediately\n\t stopped. In this case the VMware tools didn't have an opportunity to full load.\n\t The VMware tools are required to send a graceful shutdown to the VM.\n\t*/\n\tnewVm, err := vm.ChangeRunstate(client, RunStateStop, RunStateStop, RunStateStart)\n\tif err != nil {\n\t\treturn newVm, err\n\t}\n if newVm.Error != false {\n\t return nil, fmt.Errorf(\"Error stopping VM %s, error: %+v\", vm.Id, newVm.Error)\n }\n\treturn newVm, err\n\n}", "func (app *App) Stop() {}", "func (m *manager) Stop(hive.HookContext) error {\n\tif m.workerpool != nil {\n\t\tif err := m.workerpool.Close(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tm.mutex.Lock()\n\tdefer m.mutex.Unlock()\n\n\treturn nil\n}", "func (client *Client) StopInstances(request *StopInstancesRequest) (response *StopInstancesResponse, err error) {\n\tresponse = CreateStopInstancesResponse()\n\terr = client.DoAction(request, response)\n\treturn\n}", "func Stop() {\n\t// /bin/dbus-send --system --dest=org.ganesha.nfsd --type=method_call /org/ganesha/nfsd/admin org.ganesha.nfsd.admin.shutdown\n}", "func (m *Manager) Stop(ctx context.Context, uuid string) error {\n\t// Find the workflow, mark it as stopped.\n\tm.mu.Lock()\n\trw, ok := m.workflows[uuid]\n\tif !ok {\n\t\tm.mu.Unlock()\n\t\treturn fmt.Errorf(\"no running workflow with uuid %v\", uuid)\n\t}\n\trw.stopped = true\n\tm.mu.Unlock()\n\n\t// Cancel the running guy, and waits for it.\n\trw.cancel()\n\tselect {\n\tcase <-rw.done:\n\t\tbreak\n\tcase <-ctx.Done():\n\t\treturn ctx.Err()\n\t}\n\treturn nil\n}", "func (a *actorsRuntime) Stop() {\n\tif a.placement != nil {\n\t\ta.placement.Stop()\n\t}\n}", "func (ms *MarvinServer) Stop() {\n\n}", "func (f *Function) RemoveInstance(isSync bool) (string, error) {\n\tf.Lock()\n\tdefer f.Unlock()\n\n\tlogger := log.WithFields(log.Fields{\"fID\": f.fID, \"isSync\": isSync})\n\n\tlogger.Debug(\"Removing instance\")\n\n\tvar (\n\t\tr string\n\t\terr error\n\t)\n\n\tf.OnceAddInstance = new(sync.Once)\n\n\tif orch.GetSnapshotsEnabled() {\n\t\tf.OffloadInstance()\n\t\tr = \"Successfully offloaded instance \" + f.vmID\n\t} else {\n\t\tif isSync {\n\t\t\terr = orch.StopSingleVM(context.Background(), f.vmID)\n\t\t} else {\n\t\t\tf.RemoveInstanceAsync()\n\t\t\tr = \"Successfully removed (async) instance \" + f.vmID\n\t\t}\n\t}\n\n\treturn r, err\n}", "func (f *FakeTunnel) Stop() error {\n\tf.active = false\n\treturn nil\n}", "func (o *AudioStreamPlayer) Stop() {\n\t//log.Println(\"Calling AudioStreamPlayer.Stop()\")\n\n\t// Build out the method's arguments\n\tptrArguments := make([]gdnative.Pointer, 0, 0)\n\n\t// Get the method bind\n\tmethodBind := gdnative.NewMethodBind(\"AudioStreamPlayer\", \"stop\")\n\n\t// Call the parent method.\n\t// void\n\tretPtr := gdnative.NewEmptyVoid()\n\tgdnative.MethodBindPtrCall(methodBind, o.GetBaseObject(), ptrArguments, retPtr)\n\n}", "func (t *target) Stop(err error) {\n\tif !daemon.stopping.Load() {\n\t\t// vs metasync\n\t\tt.regstate.mu.Lock()\n\t\tdaemon.stopping.Store(true)\n\t\tt.regstate.mu.Unlock()\n\t}\n\tif err == nil {\n\t\tnlog.Infoln(\"Stopping \" + t.String())\n\t} else {\n\t\tnlog.Warningf(\"Stopping %s: %v\", t, err)\n\t}\n\txreg.AbortAll(err)\n\tt.htrun.stop(t.netServ.pub.s != nil && !isErrNoUnregister(err) /*rm from Smap*/)\n}", "func (m *ManagerImpl) Stop() error {\n\tif !m.lifecycle.BeginShutdown() {\n\t\treturn ErrNotRunning\n\t}\n\tm.statusTracker.NotifySSEShutdownExpected()\n\tm.withRefreshTokenLock(func() {\n\t\tif m.nextRefresh != nil {\n\t\t\tm.nextRefresh.Stop()\n\t\t}\n\t})\n\tm.StopWorkers()\n\tm.sseClient.StopStreaming()\n\tm.lifecycle.AwaitShutdownComplete()\n\treturn nil\n}", "func (c *Controller) Stop(ctx hive.HookContext) error {\n\tdoneChan := make(chan struct{})\n\tgo func() {\n\t\tc.workerpool.Close()\n\t\tclose(doneChan)\n\t}()\n\n\tselect {\n\tcase <-ctx.Done():\n\t\treturn ctx.Err()\n\tcase <-doneChan:\n\t}\n\n\treturn nil\n}", "func (c *Controller) Stop(ctx hive.HookContext) error {\n\tdoneChan := make(chan struct{})\n\tgo func() {\n\t\tc.workerpool.Close()\n\t\tclose(doneChan)\n\t}()\n\n\tselect {\n\tcase <-ctx.Done():\n\t\treturn ctx.Err()\n\tcase <-doneChan:\n\t}\n\n\treturn nil\n}", "func (c *Client) TerminateInstance(id string) error {\n\t_, err := c.computeClient.TerminateInstance(context.Background(), core.TerminateInstanceRequest{InstanceId: &id})\n\treturn err\n}", "func (a API) Stop(cmd *None) (e error) {\n\tRPCHandlers[\"stop\"].Call <-API{a.Ch, cmd, nil}\n\treturn\n}", "func (f *FakeOutput) Stop() error { return nil }", "func (e *Engine) Stop() {\n\tif atomic.CompareAndSwapInt32(&e.stopping, 0, 1) {\n\t\te.wg.Wait()\n\t\te.running = 0\n\t\te.stopping = 0\n\t}\n}", "func (l *Lifecycle) Stop(ctx context.Context) error {\n\tl.mu.Lock()\n\tl.stopRecords = make(HookRecords, 0, l.numStarted)\n\tl.mu.Unlock()\n\n\t// Run backward from last successful OnStart.\n\tvar errs []error\n\tfor ; l.numStarted > 0; l.numStarted-- {\n\t\thook := l.hooks[l.numStarted-1]\n\t\tif hook.OnStop == nil {\n\t\t\tcontinue\n\t\t}\n\n\t\tl.mu.Lock()\n\t\tl.runningHook = hook\n\t\tl.mu.Unlock()\n\n\t\truntime, err := l.runStopHook(ctx, hook)\n\t\tif err != nil {\n\t\t\t// For best-effort cleanup, keep going after errors.\n\t\t\terrs = append(errs, err)\n\t\t}\n\n\t\tl.mu.Lock()\n\t\tl.stopRecords = append(l.stopRecords, HookRecord{\n\t\t\tCallerFrame: hook.callerFrame,\n\t\t\tFunc: hook.OnStop,\n\t\t\tRuntime: runtime,\n\t\t})\n\t\tl.mu.Unlock()\n\t}\n\n\treturn multierr.Combine(errs...)\n}", "func (e *exec) stop(ctx context.Context) {\n\t// Lock the mutex to prevent race conditions with Start\n\te.execMutex.Lock()\n\tdefer e.execMutex.Unlock()\n\n\t// Do the shutdown sequence once until the startup sequence resets\n\te.stopOnce.Do(func() {\n\t\tdefer func() {\n\t\t\t// reset startOnce so the startup sequence can happen again\n\t\t\te.startOnce = sync.Once{}\n\t\t}()\n\t\te.stopFn(ctx)\n\t})\n}", "func (p *NacosRegisterPlugin) Stop() error {\r\n\t_, ip, port, _ := util.ParseRpcxAddress(p.ServiceAddress)\r\n\r\n\tfor _, name := range p.Services {\r\n\t\tinst := vo.DeregisterInstanceParam{\r\n\t\t\tIp: ip,\r\n\t\t\tEphemeral: true,\r\n\t\t\tPort: uint64(port),\r\n\t\t\tServiceName: name,\r\n\t\t\tCluster: p.Cluster,\r\n\t\t\tGroupName: p.Group,\r\n\t\t}\r\n\t\t_, err := p.namingClient.DeregisterInstance(inst)\r\n\t\tif err != nil {\r\n\t\t\tlog.Errorf(\"faield to deregister %s: %v\", name, err)\r\n\t\t}\r\n\t}\r\n\r\n\treturn nil\r\n}", "func (r *ReflectorCtl) Stop(ctx context.Context) {\n\tstart := time.Now()\n\tdefer func() {\n\t\tstats.Observe(\"reflector-ctl-latency\", time.Now().Sub(start), stats.Tag{\n\t\t\tName: \"op\",\n\t\t\tValue: \"stop\",\n\t\t})\n\t}()\n\tevents.Log(\"Stopping reflector\")\n\tr.assertNotClosed()\n\tr.initLifecycle(ctx)\n\tresult := make(chan error)\n\tselect {\n\tcase r.messages <- reflectorCtlMsg{desired: false, result: result}:\n\tcase <-ctx.Done():\n\t\treturn\n\t}\n\tselect {\n\tcase err := <-result:\n\t\tif err != nil {\n\t\t\tpanic(\"could not stop reflector: \" + err.Error())\n\t\t}\n\tcase <-ctx.Done():\n\t\treturn\n\t}\n}", "func (hSvr *HTTPServer) Stop(ctx context.Context) error {\n\treturn hSvr.svr.Shutdown(ctx)\n}", "func (o *os) NativeVideoStop() {\n\to.ensureSingleton()\n\t//log.Println(\"Calling _OS.NativeVideoStop()\")\n\n\t// Build out the method's arguments\n\tptrArguments := make([]gdnative.Pointer, 0, 0)\n\n\t// Get the method bind\n\tmethodBind := gdnative.NewMethodBind(\"_OS\", \"native_video_stop\")\n\n\t// Call the parent method.\n\t// void\n\tretPtr := gdnative.NewEmptyVoid()\n\tgdnative.MethodBindPtrCall(methodBind, o.GetBaseObject(), ptrArguments, retPtr)\n\n}", "func (f *feeService) stop() {\n\tif err := f.srv.Shutdown(context.Background()); err != nil {\n\t\tfmt.Printf(\"error: cannot stop fee api: %v\", err)\n\t}\n\n\tf.wg.Wait()\n}", "func (p *plugin) Destroy(instance instance.ID, context instance.Context) error {\n\tlog.Info(fmt.Sprintf(\"Currently running %s on instance: %v\", context, instance))\n\t// Spawn a goroutine to delete in the background\n\tgo func() {\n\t\t// TODO: Checks need adding to examine the instance that are quick enough not to trip timeout\n\t\tvar err error\n\t\tif ignoreVMs == true {\n\t\t\terr = ignoreVM(p, string(instance))\n\t\t} else {\n\t\t\terr = deleteVM(p, string(instance))\n\t\t}\n\t\tif err != nil {\n\t\t\tlog.Error(\"Destroying Instance failed\", \"err\", err)\n\t\t}\n\t}()\n\n\t// TODO: Ideally the goroutine should return the error otherwise this function can never fail for InfraKit\n\treturn nil\n}", "func (api *API) Stop() error {\n\n\t// context: wait for 3 seconds\n\tctx, cancel := context.WithTimeout(context.Background(), 3*time.Second)\n\tdefer cancel()\n\n\terr := api.srv.Shutdown(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (inst *Instance) ForceStop() error {\n\tautoRestartHandle := inst.autoRestartHandle\n\n\tautoRestartHandle.mask()\n\treturn inst.kill()\n}", "func (e *Engine) Stop() {\n\te.running = false\n}", "func (d *Driver) Stop() error {\n\tcs := d.client()\n\t_, err := cs.AsyncRequest(&egoscale.StopVirtualMachine{\n\t\tID: d.ID,\n\t}, d.async)\n\n\treturn err\n}", "func (el *Launcher) Stop() error {\n\tlogrus.Debugf(\"engine launcher %v: prepare to stop engine %v\",\n\t\tel.LauncherName, el.currentEngine.EngineName)\n\n\tif _, err := el.currentEngine.Stop(); err != nil {\n\t\treturn err\n\t}\n\tel.lock.Lock()\n\tif !el.isStopping {\n\t\tclose(el.doneCh)\n\t}\n\tel.isStopping = true\n\tel.lock.Unlock()\n\n\tel.updateCh <- el\n\n\tlogrus.Debugf(\"engine launcher %v: succeed to stop engine %v at %v\",\n\t\tel.LauncherName, el.currentEngine.EngineName, el.currentEngine.GetListen())\n\treturn nil\n\n}", "func (s *Plugin) Stop() error {\n\treturn nil\n}", "func (b *Batch) Stop() {\n\tb.cancelFunc()\n}", "func (oh *OperatingHours) StopInstances(stopInstances []*string) error {\n\n\tlog.WithField(\"len\", len(stopInstances)).Info(\"stopInstances\")\n\n\tif len(stopInstances) > 0 {\n\t\t_, err := oh.EC2API.StopInstances(&ec2.StopInstancesInput{InstanceIds: stopInstances})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfor _, instanceID := range stopInstances {\n\t\t\tlog.WithField(\"instanceID:\", aws.StringValue(instanceID)).Info(\"stop\")\n\t\t}\n\t}\n\n\treturn nil\n}", "func (ctx *actorContext) Stop(pid *PID) {\n\tpid.ref().Stop(pid)\n}", "func (m *Mock) StopPlugin(name string, cancelFlag task.CancelFlag) (err error) {\n\treturn nil\n}", "func (rcsw *RemoteClusterServiceWatcher) Stop(cleanupState bool) {\n\trcsw.probeEventsSink.send(&ClusterNotRegistered{\n\t\tclusterName: rcsw.clusterName,\n\t})\n\tclose(rcsw.stopper)\n\tif cleanupState {\n\t\trcsw.eventsQueue.Add(&ClusterUnregistered{})\n\t}\n\trcsw.eventsQueue.ShutDown()\n}", "func (p *FuncPool) RemoveInstance(fID, imageName string, isSync bool) (string, error) {\n\tf := p.getFunction(fID, imageName)\n\n\treturn f.RemoveInstance(isSync)\n}", "func (f *Function) OffloadInstance() {\n\tlogger := log.WithFields(log.Fields{\"fID\": f.fID})\n\n\tlogger.Debug(\"Offloading instance\")\n\n\tctx, cancel := context.WithTimeout(context.Background(), time.Second*15)\n\tdefer cancel()\n\n\terr := orch.Offload(ctx, f.vmID)\n\tif err != nil {\n\t\tlog.Panic(err)\n\t}\n\tf.conn.Close()\n}", "func (hsp HistoryServicePrecacher) Stop() { hsp.pc.Stop() }", "func (a *appsec) stop() {\n\ta.unregisterWAF()\n\ta.limiter.Stop()\n}", "func Stop() {\n\tstopRunning <- true\n\n}", "func (d *Driver) Stop() error {\n\tclient, err := d.getClient()\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn client.ShutdownVirtualMachine(d.vmName(), true)\n}", "func (rtspService *RTSPService) Stop(msg *wssapi.Msg) (err error) {\n\treturn\n}", "func (s *Service) Stop() error {\n\t<-s.ReadyForCleanup\n\tif s.Proc == nil {\n\t\treturn fmt.Errorf(\"cannot stop %s (not started)\", s.InstanceName)\n\t}\n\tif err := s.Proc.Kill(); err != nil {\n\t\treturn fmt.Errorf(\"cannot kill %s: %s\", s.InstanceName, err)\n\t}\n\treturn nil\n}", "func (j *JobRunner) Stop() error {\n\terr := j.StaticTG.Stop()\n\tif err != nil {\n\t\treturn errors.AddContext(err, \"can't stop thread group\")\n\t}\n\treturn nil\n}", "func (s *Streamer) Stop() error {\n\ts.mu.Lock()\n\tif s.state != stateRunning {\n\t\ts.mu.Unlock()\n\t\treturn ErrNotRunning\n\t}\n\ts.state = stateStopping\n\ts.mu.Unlock()\n\n\ts.fsNotify.Close() // trigger stop chain: fsNotify -> (sendChangeEvents,logNotifyErrors) -> eventsRouter\n\ts.threads.Wait()\n\n\ts.mu.Lock()\n\ts.state = stateStopped\n\ts.mu.Unlock()\n\n\treturn nil\n}", "func (b *NoopLifecycle) Stop(t testing.TB) error {\n\treturn nil\n}", "func Stop() {\n\ts.Stop()\n}", "func (p *Plugin) Stop(ctx context.Context, in *plugin.StopRequest) (*plugin.StopResponse, error) {\n\terr := p.Service.Stop(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &plugin.StopResponse{}, nil\n}", "func (taskService TaskService) StopTaskInstanceExecutions(w http.ResponseWriter, r *http.Request) {\n\tvar payload struct {\n\t\tInstanceIDs []gocql.UUID `json:\"instanceIDs\"`\n\t}\n\n\tif err := validator.ExtractStructFromRequest(r, &payload); err != nil {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorCantDecodeInputData, \"TaskService.StopTaskInstanceExecutions: error while unmarshaling request body. Err=%v\", err)\n\t\tcommon.SendBadRequest(w, r, errorcode.ErrorCantDecodeInputData)\n\t\treturn\n\t}\n\n\tctx := r.Context()\n\ttaskInstances, err := taskService.taskInstancePersistence.GetByIDs(ctx, payload.InstanceIDs...)\n\tif err != nil {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorCantGetTaskInstances, \"TaskService.StopTaskInstanceExecutions: can't get a Task Instances by TaskInstanceIDs %v. Err=%v\", payload.InstanceIDs, err)\n\t\tcommon.SendInternalServerError(w, r, errorcode.ErrorCantGetTaskInstances)\n\t\treturn\n\t}\n\n\tif len(taskInstances) == 0 {\n\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorCantGetTaskInstances, \"TaskService.StopTaskInstanceExecutions: error can't get Task Instances by id. Err=%v\", err)\n\t\tcommon.SendBadRequest(w, r, errorcode.ErrorCantGetTaskInstances)\n\t\treturn\n\t}\n\n\tfor i := range taskInstances {\n\t\t// isStopped describes if future execution was stopped at least on one device\n\t\tisStopped := false\n\t\tfor deviceID := range taskInstances[i].Statuses {\n\t\t\tif taskInstances[i].Statuses[deviceID] == statuses.TaskInstancePending {\n\t\t\t\ttaskInstances[i].Statuses[deviceID] = statuses.TaskInstanceStopped\n\t\t\t\tisStopped = true\n\t\t\t}\n\t\t}\n\n\t\tif isStopped {\n\t\t\ttaskInstances[i].OverallStatus = statuses.TaskInstanceStopped\n\t\t}\n\t}\n\n\tfor _, instance := range taskInstances {\n\t\tif err = taskService.taskInstancePersistence.Insert(ctx, instance); err != nil {\n\t\t\tlogger.Log.ErrfCtx(r.Context(), errorcode.ErrorCantUpdateTaskInstances, \"TaskService.StopTaskInstanceExecutions: error while updating Task Instance (ID: %v). Err=%v\", instance.ID, err)\n\t\t\tcommon.SendInternalServerError(w, r, errorcode.ErrorCantUpdateTaskInstances)\n\t\t\treturn\n\t\t}\n\t}\n\n\tlogger.Log.InfofCtx(r.Context(), \"TaskService.StopTaskInstanceExecutions: Tasks' executions are successfully stopped\")\n\tcommon.RenderJSON(w, struct {\n\t\tStatus string\n\t}{Status: \"Complete\"})\n}", "func (f *framework) stop() {\n\tclose(f.epochChan)\n}", "func (c *InstancesStopCall) Do(call *v1.InstancesStopCall, opts ...googleapi.CallOption) (*v1.Operation, error) {\n\treturn call.Do(opts...)\n}", "func (inst *Instancer) Stop() {\n\tinst.cancel()\n\tinst.cache.Stop()\n\tinst.wg.Wait()\n}", "func (pier *Pier) Stop(isAux bool) error {\n\tif pier.config.Mode.Type != repo.UnionMode {\n\t\tif err := pier.monitor.Stop(); err != nil {\n\t\t\treturn fmt.Errorf(\"monitor stop: %w\", err)\n\t\t}\n\n\t\tif !isAux {\n\t\t\t// stop appchain plugin first and kill plugin process\n\t\t\tpier.grpcPlugin.Kill()\n\t\t}\n\n\t\tif err := pier.exec.Stop(); err != nil {\n\t\t\treturn fmt.Errorf(\"executor stop: %w\", err)\n\t\t}\n\t}\n\n\tif err := pier.lite.Stop(); err != nil {\n\t\treturn fmt.Errorf(\"lite stop: %w\", err)\n\t}\n\n\tif err := pier.exchanger.Stop(); err != nil {\n\t\treturn fmt.Errorf(\"exchanger stop: %w\", err)\n\t}\n\treturn nil\n}", "func (p *Proxy) Stop() error {\n\tp.svr.Stop()\n\treturn nil\n}", "func (b *Bootstrapper) Stop() error {\n\treturn nil\n}", "func (b *Bootstrapper) Stop() error {\n\treturn nil\n}", "func (g *RESTFrontend) Stop() {\n\tif g.IsRunning {\n\t\tg.Server.Shutdown(context.TODO())\n\t\tg.Listener.Close()\n\t}\n}", "func (_m *TimeTicker) Stop() {\n\t_m.Called()\n}", "func (t *TimerSnapshot) Stop() {}", "func (d *Driver) Stop() error {\n\tif err := d.verifyRootPermissions(); err != nil {\n\t\treturn err\n\t}\n\n\ts, err := d.GetState()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif s != state.Stopped {\n\t\terr := d.sendSignal(syscall.SIGTERM)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"hyperkit sigterm failed\")\n\t\t}\n\t\t// wait 120s for graceful shutdown\n\t\tfor i := 0; i < 60; i++ {\n\t\t\ttime.Sleep(2 * time.Second)\n\t\t\ts, _ := d.GetState()\n\t\t\tlog.Debugf(\"VM state: %s\", s)\n\t\t\tif s == state.Stopped {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t\treturn errors.New(\"VM Failed to gracefully shutdown, try the kill command\")\n\t}\n\treturn nil\n}", "func (p *PrivNegAPI) Stop() {\n\tif err := p.server.Shutdown(nil); err != nil {\n\t\tpanic(err)\n\t}\n}", "func (p *OnPrem) DeleteInstance(ctx *Context, instancename string) error {\n\treturn fmt.Errorf(\"Operation not supported\")\n}", "func (s *StepTeardownInstance) Run(state multistep.StateBag) multistep.StepAction {\n\tconfig := state.Get(\"config\").(*Config)\n\tdriver := state.Get(\"driver\").(Driver)\n\tui := state.Get(\"ui\").(packer.Ui)\n\n\tname := config.InstanceName\n\tif name == \"\" {\n\t\treturn multistep.ActionHalt\n\t}\n\n\tui.Say(\"Deleting instance...\")\n\tinstanceLog, _ := driver.GetSerialPortOutput(config.Zone, name)\n\tstate.Put(\"instance_log\", instanceLog)\n\terrCh, err := driver.DeleteInstance(config.Zone, name)\n\tif err == nil {\n\t\tselect {\n\t\tcase err = <-errCh:\n\t\tcase <-time.After(config.stateTimeout):\n\t\t\terr = errors.New(\"time out while waiting for instance to delete\")\n\t\t}\n\t}\n\n\tif err != nil {\n\t\tui.Error(fmt.Sprintf(\n\t\t\t\"Error deleting instance. Please delete it manually.\\n\\n\"+\n\t\t\t\t\"Name: %s\\n\"+\n\t\t\t\t\"Error: %s\", name, err))\n\t\treturn multistep.ActionHalt\n\t}\n\tui.Message(\"Instance has been deleted!\")\n\tstate.Put(\"instance_name\", \"\")\n\n\treturn multistep.ActionContinue\n}", "func stop() error {\n\tif spammerInstance == nil {\n\t\treturn ErrSpammerDisabled\n\t}\n\n\tspammerLock.Lock()\n\tdefer spammerLock.Unlock()\n\n\tstopWithoutLocking()\n\n\tisRunning = false\n\n\treturn nil\n}", "func Stop(label string, kind Kind) error {\n\tif kind == Daemon {\n\t\terr := isRoot()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t_, err := run(\"stop\", label)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (p *OnPrem) DeleteInstance(ctx *Context, instancename string) error {\n\n\tpid, err := strconv.Atoi(instancename)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\n\t// yolo\n\terr = sysKill(pid)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\n\topshome := GetOpsHome()\n\tipath := path.Join(opshome, \"instances\", instancename)\n\terr = os.Remove(ipath)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (client BaseClient) DisableFeatureInstance(ctx context.Context, featureName string, featureVersion string, instanceName string) (result FeatureInstance, err error) {\n\treq, err := client.DisableFeatureInstancePreparer(ctx, featureName, featureVersion, instanceName)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"DisableFeatureInstance\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresp, err := client.DisableFeatureInstanceSender(req)\n\tif err != nil {\n\t\tresult.Response = autorest.Response{Response: resp}\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"DisableFeatureInstance\", resp, \"Failure sending request\")\n\t\treturn\n\t}\n\n\tresult, err = client.DisableFeatureInstanceResponder(resp)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"DisableFeatureInstance\", resp, \"Failure responding to request\")\n\t}\n\n\treturn\n}", "func (r *ReceiveFuncState) Stop() {\n\tmu.Lock()\n\tdefer mu.Unlock()\n\tr.running = false\n\tselfCheckLocked()\n}", "func (s *svc) Stop() error {\n\ts.Lock()\n\tdefer s.Unlock()\n\n\tselect {\n\tcase <-s.exit:\n\t\treturn nil\n\tdefault:\n\t\tclose(s.exit)\n\t}\n\n\treturn nil\n}", "func (p *noop) Stop() {}" ]
[ "0.7653172", "0.7653172", "0.71888673", "0.7103518", "0.70696676", "0.69321007", "0.6709341", "0.63744724", "0.6367101", "0.6186668", "0.6054368", "0.6039993", "0.5997645", "0.59660184", "0.59182656", "0.587392", "0.5872844", "0.58528763", "0.58509845", "0.5787912", "0.5786989", "0.5723807", "0.57212716", "0.5685353", "0.5651963", "0.5642627", "0.5640793", "0.56306565", "0.5614821", "0.5585409", "0.55751437", "0.5573862", "0.5552428", "0.5547945", "0.5541383", "0.55335", "0.55258626", "0.55252165", "0.5522807", "0.55150425", "0.5503755", "0.5503755", "0.55025727", "0.5501079", "0.54880786", "0.54870707", "0.54832655", "0.5482132", "0.547663", "0.5473273", "0.5472292", "0.5469105", "0.5459465", "0.54512334", "0.5450873", "0.5445899", "0.54357535", "0.5421834", "0.54182816", "0.54112494", "0.5407677", "0.5398955", "0.5393831", "0.5393696", "0.5392514", "0.5388566", "0.53851604", "0.5384063", "0.5383478", "0.538237", "0.5366404", "0.5363244", "0.53614354", "0.5361361", "0.5360229", "0.53594524", "0.53519565", "0.5349633", "0.53478044", "0.53469723", "0.5345417", "0.53422534", "0.53408134", "0.5338268", "0.53341866", "0.53341866", "0.53314984", "0.5327147", "0.5326275", "0.53240436", "0.53211725", "0.53205794", "0.53200173", "0.53186256", "0.5316912", "0.5316171", "0.5310807", "0.5306776", "0.5305012", "0.5290329" ]
0.72906727
2
DeleteDisk uses the override method DeleteDiskFn or the real implementation.
func (c *TestClient) DeleteDisk(project, zone, name string) error { if c.DeleteDiskFn != nil { return c.DeleteDiskFn(project, zone, name) } return c.client.DeleteDisk(project, zone, name) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *Module) DiskDelete(name string) error {\n\tpath, err := s.findDisk(name)\n\tif os.IsNotExist(err) {\n\t\treturn nil\n\t} else if err != nil {\n\t\treturn err\n\t}\n\n\tif err := os.Remove(path); err != nil && !os.IsNotExist(err) {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func DeleteDisk(id int) {\n\tdb.Session().Delete(&model.Disk{}, \"id = ?\", id)\n}", "func DiskDelete(w rest.ResponseWriter, r *rest.Request) {\n\treq := DiskDeleteRequest{}\n\terr := r.DecodeJsonPayload(&req)\n\tif err != nil {\n\t\tlogit.Error.Println(err.Error())\n\t\trest.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\tlogit.Info.Println(\"DiskDelete called \" + req.Path)\n\n\tvar cmd *exec.Cmd\n\tcmd = exec.Command(\"deletevolume\", req.Path)\n\tvar out bytes.Buffer\n\tvar stderr bytes.Buffer\n\tcmd.Stdout = &out\n\tcmd.Stderr = &stderr\n\terr = cmd.Run()\n\tif err != nil {\n\t\tlogit.Error.Println(err.Error())\n\t\trest.Error(w, err.Error(), 400)\n\t\treturn\n\t}\n\n\tvar response DiskDeleteResponse\n\tresponse.Output = out.String()\n\tresponse.Status = \"OK\"\n\tw.WriteJson(&response)\n}", "func (c *MockDisksClient) Delete(ctx context.Context, resourceGroupName, diskName string) error {\n\t// Ignore resourceGroupName for simplicity.\n\tif _, ok := c.Disks[diskName]; !ok {\n\t\treturn fmt.Errorf(\"%s does not exist\", diskName)\n\t}\n\tdelete(c.Disks, diskName)\n\treturn nil\n}", "func (m *DisksClientMock) Delete(ctx context.Context, resourceGroupName string, diskName string) *retry.Error {\n\targs := m.Called(resourceGroupName, diskName)\n\tif args.Error(1) != nil {\n\t\treturn &retry.Error{RawError: args.Error(1)}\n\t}\n\treturn nil\n}", "func (d *Disk) Delete(name string) error {\n\treturn d.Runner.Delete(name)\n}", "func DeleteDisk(path string) {\n\tvar directory string = \"\"\n\tif strings.Contains(path, \"\\\"\") {\n\t\tdirectory, _ = SetDirectory(path)\n\t}\n\tif directory != \"\" {\n\t\tpath = directory\n\t}\n\treader := bufio.NewReader(os.Stdin)\n\tif _, err := os.Stat(path); err == nil {\n\t\tfor {\n\t\t\tfmt.Println(\"SEGURO QUE DESEA ELIMINAR EL DISCO \", path, \"? (y/n)\")\n\t\t\ttext, _ := reader.ReadString('\\n')\n\t\t\ttext = strings.Replace(text, \"\\n\", \"\", -1)\n\n\t\t\tif strings.Compare(\"Y\", text) == 0 || strings.Compare(\"y\", text) == 0 {\n\t\t\t\tdeleteFile(path)\n\t\t\t\tfmt.Println(\"SE ELIMINO EL DISCO CORRECTAMENTE\")\n\t\t\t\tPause()\n\t\t\t\tbreak\n\t\t\t} else if strings.Compare(\"N\", text) == 0 || strings.Compare(\"n\", text) == 0 {\n\t\t\t\treturn\n\t\t\t}\n\n\t\t}\n\t} else if os.IsNotExist(err) {\n\t\tPause()\n\t\tfmt.Println(err)\n\t}\n\n}", "func GuestDeleteDisks(endpoint string, guestid string, body GuestDeleteDiskBody) (int, []byte) {\n\tdeleteReq, _ := json.Marshal(body)\n\n\tbuffer := getEndpointwithGuests(endpoint)\n\tbuffer.WriteString(\"/\")\n\tbuffer.WriteString(guestid)\n\tbuffer.WriteString(\"/disks\")\n\n\tstatus, data := hq.Delete(buffer.String(), deleteReq)\n\n\treturn status, data\n}", "func (m *Manager) DeleteDiskAndMedia(diskName string) *Data {\n\n\turi := fmt.Sprintf(\"https://management.core.windows.net/%s/services/disks/%s?comp=media\", m.SubscrId, diskName)\n\n\tdata := &Data{\n\t\tVerb: \"DELETE\",\n\t\tUri: uri,\n\t}\n\n\treturn data\n}", "func fDiskEliminar(comando string) {\n\t//verifico que el comando exista\n\tif strings.Compare(comando, \"\") != 0 {\n\t\ttipoEliminacion := \"\"\n\t\tnombreEliminacion := \"\"\n\t\tpathEliminacion := \"\"\n\t\tpathOk := 0\n\t\ts := strings.Split(comando, \" -\")\n\t\tif len(s) > 3 {\n\t\t\tfor i := 1; i < len(s); i++ {\n\t\t\t\ts1 := strings.Split(s[i], \"->\")\n\t\t\t\tif len(s1) > 1 {\n\t\t\t\t\tswitch strings.ToLower(strings.TrimSpace(s1[0])) {\n\t\t\t\t\tcase \"delete\":\n\t\t\t\t\t\ttipoEliminacion = atributoDelete(strings.ToLower(strings.TrimSpace(strings.ReplaceAll(s1[1], \"\\\"\", \"\"))))\n\t\t\t\t\t\t//tipoEliminacion = \"full\"\n\t\t\t\t\tcase \"name\":\n\t\t\t\t\t\tnombreEliminacion = strings.ToLower(strings.TrimSpace(strings.ReplaceAll(s1[1], \"\\\"\", \"\")))\n\t\t\t\t\tcase \"path\":\n\t\t\t\t\t\tpathOk, pathEliminacion = verificarPath(strings.ToLower(strings.TrimSpace(strings.ReplaceAll(s1[1], \"\\\"\", \"\"))))\n\t\t\t\t\tdefault:\n\t\t\t\t\t\tfmt.Println(\"RESULTADO: El atributo \" + s1[0] + \" no se reconoce para el comando DELETE\")\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t\tif pathOk == 1 {\n\t\t\t\t//si existe el path\n\t\t\t\tif strings.Compare(tipoEliminacion, \"error\") != 0 {\n\t\t\t\t\t//si esta bien el tipo de eliminacion\n\t\t\t\t\tif strings.Compare(nombreEliminacion, \"\") != 0 {\n\t\t\t\t\t\t//si esta bien el nombre de la particion\n\t\t\t\t\t\teliminarParticion(pathEliminacion, nombreEliminacion, tipoEliminacion)\n\t\t\t\t\t} else {\n\t\t\t\t\t\tfmt.Println(\"RESULTADO: Debe ingresar el nombre de la particion a eliminar\")\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\tfmt.Println(\"RESULTADO: Error en el tipo de eliminacion de la particion\")\n\t\t\t\t}\n\t\t\t} else if pathOk == 2 {\n\t\t\t\tfmt.Println(\"RESULTADO: El archivo indicado no representa un disco\")\n\t\t\t} else {\n\t\t\t\tfmt.Println(\"RESULTADO: No existe el archivo especificado\")\n\t\t\t}\n\t\t} else {\n\t\t\tfmt.Println(\"RESULTADO: Faltan atributos obligatorios para el comando FDISK DELETE\")\n\t\t}\n\t}\n}", "func (base *Base) Delete(ctx context.Context, path string) error {\n\tctx, done := dcontext.WithTrace(ctx)\n\tdefer done(\"%s.Delete(%q)\", base.Name(), path)\n\n\tif !storagedriver.PathRegexp.MatchString(path) {\n\t\treturn storagedriver.InvalidPathError{Path: path, DriverName: base.StorageDriver.Name()}\n\t}\n\n\tstart := time.Now()\n\terr := base.setDriverName(base.StorageDriver.Delete(ctx, path))\n\tstorageAction.WithValues(base.Name(), \"Delete\").UpdateSince(start)\n\treturn err\n}", "func (d *Driver) Delete(ctx context.Context, f *functions.Function) error {\n\tspan, ctx := trace.Trace(ctx, \"\")\n\tdefer span.Finish()\n\tif err := d.deleteContainers(ctx, f, true); err != nil {\n\t\t// no error wrapping, delete already does that\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *clustermgrClient) DeleteMigratingDisk(ctx context.Context, taskType proto.TaskType, diskID proto.DiskID) (err error) {\n\treturn c.client.DeleteKV(ctx, genMigratingDiskID(taskType, diskID))\n}", "func NewCmdDiskDelete() *cobra.Command {\n\tvar yes *bool\n\tvar udiskIDs *[]string\n\treq := base.BizClient.NewDeleteUDiskRequest()\n\tcmd := &cobra.Command{\n\t\tUse: \"delete\",\n\t\tShort: \"Delete udisk instances\",\n\t\tLong: \"Delete udisk instances\",\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\tif !*yes {\n\t\t\t\tsure, err := ux.Prompt(fmt.Sprintf(\"Are you sure to delete udisk(s)?\"))\n\t\t\t\tif err != nil {\n\t\t\t\t\tbase.Cxt.PrintErr(err)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tif !sure {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t\tfor _, id := range *udiskIDs {\n\t\t\t\tid := base.PickResourceID(id)\n\t\t\t\treq.UDiskId = &id\n\t\t\t\t_, err := base.BizClient.DeleteUDisk(req)\n\t\t\t\tif err != nil {\n\t\t\t\t\tbase.HandleError(err)\n\t\t\t\t\tcontinue\n\t\t\t\t} else {\n\t\t\t\t\tbase.Cxt.Printf(\"udisk[%s] deleted\\n\", *req.UDiskId)\n\t\t\t\t}\n\t\t\t}\n\t\t},\n\t}\n\tflags := cmd.Flags()\n\tflags.SortFlags = false\n\tudiskIDs = flags.StringSlice(\"udisk-id\", nil, \"Required. The Resource ID of udisks to delete\")\n\treq.ProjectId = flags.String(\"project-id\", base.ConfigInstance.ProjectID, \"Optional. Assign project-id\")\n\treq.Region = flags.String(\"region\", base.ConfigInstance.Region, \"Optional. Assign region\")\n\treq.Zone = flags.String(\"zone\", base.ConfigInstance.Zone, \"Optional. Assign availability zone\")\n\tyes = flags.BoolP(\"yes\", \"y\", false, \"Optional. Do not prompt for confirmation.\")\n\n\tflags.SetFlagValuesFunc(\"udisk-id\", func() []string {\n\t\treturn getDiskList([]string{status.DISK_AVAILABLE, status.DISK_FAILED}, *req.ProjectId, *req.Region, *req.Zone)\n\t})\n\n\tcmd.MarkFlagRequired(\"udisk-id\")\n\n\treturn cmd\n}", "func (httpfs *FS) Delete(ctx context.Context, path string) (err error) {\n\treturn qfs.ErrReadOnly\n}", "func DeleteVM(c config.Cpi, extInput bosh.MethodArguments) error {\n\tvar cid string\n\tif reflect.TypeOf(extInput[0]) != reflect.TypeOf(cid) {\n\t\treturn errors.New(\"Received unexpected type for vm cid\")\n\t}\n\n\tcid = extInput[0].(string)\n\tnode, err := rackhdapi.GetNodeByVMCID(c, cid)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif node.PersistentDisk.IsAttached {\n\t\terr = rackhdapi.MakeDiskRequest(c, node, false)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tworkflowName, err := workflows.PublishDeprovisionNodeWorkflow(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = workflows.RunDeprovisionNodeWorkflow(c, node.ID, workflowName, cid)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor _, tag := range node.Tags {\n\t\tif strings.HasPrefix(tag, DiskCIDTagPrefix) {\n\t\t\treturn nil\n\t\t}\n\t}\n\n\terr = rackhdapi.ReleaseNode(c, node.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *TestClient) DetachDisk(project, zone, instance, disk string) error {\n\tif c.DetachDiskFn != nil {\n\t\treturn c.DetachDiskFn(project, zone, instance, disk)\n\t}\n\treturn c.client.DetachDisk(project, zone, instance, disk)\n}", "func (yad *yandexDisk) DeleteResource(path string, fields []string, forceAsync bool, md5 string, permanently bool) (l *Link, e error) {\n\tvalues := url.Values{}\n\tvalues.Add(\"path\", path)\n\tvalues.Add(\"fields\", strings.Join(fields, \",\"))\n\tvalues.Add(\"force_async\", strconv.FormatBool(forceAsync))\n\tvalues.Add(\"md5\", md5)\n\tvalues.Add(\"permanently\", strconv.FormatBool(permanently))\n\n\treq, e := yad.client.request(http.MethodDelete, \"/disk/resources?\"+values.Encode(), nil)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\n\tl = new(Link)\n\tri, e := yad.client.getResponse(req, &l)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\tif ri.StatusCode == 204 {\n\t\treturn nil, nil\n\t}\n\treturn\n}", "func DiskFactory() worker.Worker {\n\treturn &Disk{}\n}", "func (r *DiskDirectoryResource) Delete(id string) error {\n\tif err := r.c.ModQuery(\"DELETE\", BasePath+DiskDirectoryEndpoint+\"/\"+id, nil); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (rc *RequiredCapability) Delete() (error, error, int) {\n\tauthorized, err := rc.isTenantAuthorized()\n\tif !authorized {\n\t\treturn errors.New(\"not authorized on this tenant\"), nil, http.StatusForbidden\n\t} else if err != nil {\n\t\treturn nil, fmt.Errorf(\"checking authorization for existing DS ID: %s\" + err.Error()), http.StatusInternalServerError\n\t}\n\t_, cdnName, _, err := dbhelpers.GetDSNameAndCDNFromID(rc.ReqInfo.Tx.Tx, *rc.DeliveryServiceID)\n\tif err != nil {\n\t\treturn nil, err, http.StatusInternalServerError\n\t}\n\tuserErr, sysErr, errCode := dbhelpers.CheckIfCurrentUserCanModifyCDN(rc.ReqInfo.Tx.Tx, string(cdnName), rc.ReqInfo.User.UserName)\n\tif userErr != nil || sysErr != nil {\n\t\treturn userErr, sysErr, errCode\n\t}\n\treturn api.GenericDelete(rc)\n}", "func (s *Module) DiskCreate(name string, size gridtypes.Unit) (disk pkg.VDisk, err error) {\n\tpath, err := s.findDisk(name)\n\tif err == nil {\n\t\treturn disk, errors.Wrapf(os.ErrExist, \"disk with id '%s' already exists\", name)\n\t}\n\n\tbase, err := s.diskFindCandidate(size)\n\tif err != nil {\n\t\treturn disk, errors.Wrapf(err, \"failed to find a candidate to host vdisk of size '%d'\", size)\n\t}\n\n\tpath, err = s.safePath(base, name)\n\tif err != nil {\n\t\treturn disk, err\n\t}\n\n\tdefer func() {\n\t\t// clean up disk file if error\n\t\tif err != nil {\n\t\t\tos.RemoveAll(path)\n\t\t}\n\t}()\n\n\tdefer syscall.Sync()\n\n\tvar file *os.File\n\tfile, err = os.Create(path)\n\tif err != nil {\n\t\treturn disk, err\n\t}\n\n\tdefer file.Close()\n\tif err = chattr.SetAttr(file, chattr.FS_NOCOW_FL); err != nil {\n\t\treturn disk, err\n\t}\n\n\tif err = syscall.Fallocate(int(file.Fd()), 0, 0, int64(size)); err != nil {\n\t\treturn disk, errors.Wrap(err, \"failed to truncate disk to size\")\n\t}\n\n\treturn pkg.VDisk{Path: path, Size: int64(size)}, nil\n}", "func DeleteVdisk(vdiskID string, configSource config.Source) (bool, error) {\n\tstaticConfig, err := config.ReadVdiskStaticConfig(configSource, vdiskID)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tnbdConfig, err := config.ReadVdiskNBDConfig(configSource, vdiskID)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tclusterConfig, err := config.ReadStorageClusterConfig(configSource, nbdConfig.StorageClusterID)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\t// if slave cluster is configured, we'll want to delete the vdisk from it as well\n\tif nbdConfig.SlaveStorageClusterID != \"\" {\n\t\tslaveClusterCfg, err := config.ReadStorageClusterConfig(configSource, nbdConfig.SlaveStorageClusterID)\n\t\tif err != nil {\n\t\t\treturn false, err\n\t\t}\n\t\tclusterConfig.Servers = append(clusterConfig.Servers, slaveClusterCfg.Servers...)\n\t}\n\n\t// create a cluster of all primary (and slave) servers\n\tcluster, err := ardb.NewCluster(*clusterConfig, nil)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\t// delete all data for this vdisk found in primary (and slave) servers\n\treturn DeleteVdiskInCluster(vdiskID, staticConfig.Type, cluster)\n}", "func (client StorageGatewayClient) DeleteFileSystem(ctx context.Context, request DeleteFileSystemRequest) (response DeleteFileSystemResponse, err error) {\n\tvar ociResponse common.OCIResponse\n\tpolicy := common.NoRetryPolicy()\n\tif client.RetryPolicy() != nil {\n\t\tpolicy = *client.RetryPolicy()\n\t}\n\tif request.RetryPolicy() != nil {\n\t\tpolicy = *request.RetryPolicy()\n\t}\n\tociResponse, err = common.Retry(ctx, request, client.deleteFileSystem, policy)\n\tif err != nil {\n\t\tif ociResponse != nil {\n\t\t\tif httpResponse := ociResponse.HTTPResponse(); httpResponse != nil {\n\t\t\t\topcRequestId := httpResponse.Header.Get(\"opc-request-id\")\n\t\t\t\tresponse = DeleteFileSystemResponse{RawResponse: httpResponse, OpcRequestId: &opcRequestId}\n\t\t\t} else {\n\t\t\t\tresponse = DeleteFileSystemResponse{}\n\t\t\t}\n\t\t}\n\t\treturn\n\t}\n\tif convertedResponse, ok := ociResponse.(DeleteFileSystemResponse); ok {\n\t\tresponse = convertedResponse\n\t} else {\n\t\terr = fmt.Errorf(\"failed to convert OCIResponse into DeleteFileSystemResponse\")\n\t}\n\treturn\n}", "func (zk *dbZk) Delete(path string) error {\n\tvar failed bool\n\tvar existed bool\n\tvar err error\n\tstarted := time.Now()\n\n\texisted, err = zk.ZkCli.Exist(path)\n\tif err != nil {\n\t\tfailed = true\n\t}\n\n\tif existed {\n\t\terr = zk.ZkCli.Del(path, -1)\n\t\tif err != nil {\n\t\t\tfailed = true\n\t\t}\n\t}\n\n\tstore.ReportStorageOperatorMetrics(store.StoreOperatorDelete, started, failed)\n\treturn err\n}", "func (c *Compute) Disk(name string) (string, error) {\n\tdisk, err := c.Disks.Get(c.Project, c.Zone, name).Do()\n\tif err == nil {\n\t\tlog.Printf(\"found existing root disk: %q\", disk.SelfLink)\n\t\treturn disk.SelfLink, nil\n\t}\n\tlog.Printf(\"not found, creating new root disk: %q\", name)\n\top, err := c.Disks.Insert(c.Project, c.Zone, &compute.Disk{\n\t\tName: name,\n\t}).SourceImage(*image).Do()\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"disk insert api call failed: %v\", err)\n\t}\n\tif err := c.wait(op); err != nil {\n\t\treturn \"\", fmt.Errorf(\"disk insert operation failed: %v\", err)\n\t}\n\tlog.Printf(\"root disk created: %q\", op.TargetLink)\n\treturn op.TargetLink, nil\n}", "func (f *Fs) delete(ctx context.Context, isFile bool, id string, remote string, hardDelete bool) (err error) {\n\tif hardDelete {\n\t\topts := rest.Opts{\n\t\t\tMethod: \"DELETE\",\n\t\t\tRootURL: id,\n\t\t\tNoResponse: true,\n\t\t}\n\t\treturn f.pacer.Call(func() (bool, error) {\n\t\t\tresp, err := f.srv.Call(ctx, &opts)\n\t\t\treturn shouldRetry(ctx, resp, err)\n\t\t})\n\t}\n\t// Move file/dir to deleted files if not hard delete\n\tleaf := path.Base(remote)\n\tif isFile {\n\t\t_, err = f.moveFile(ctx, id, leaf, f.opt.DeletedID)\n\t} else {\n\t\terr = f.moveDir(ctx, id, leaf, f.opt.DeletedID)\n\t}\n\treturn err\n}", "func (d *DiskStorage) Delete(key string) error {\n\terr := d.db.Update(func(tx *bolt.Tx) error {\n\t\tb := tx.Bucket([]byte(d.name))\n\t\treturn b.Delete([]byte(key))\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn d.memory.Delete(key)\n}", "func (c *MockAzureCloud) Disk() azure.DisksClient {\n\treturn c.DisksClient\n}", "func (m *ClusterService) decommissionDisk(ctx context.Context, args struct {\n\tOffLineAddr string\n\tDiskPath string\n}) (*proto.GeneralResp, error) {\n\n\tnode, err := m.cluster.dataNode(args.OffLineAddr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbadPartitions := node.badPartitions(args.DiskPath, m.cluster)\n\tif len(badPartitions) == 0 {\n\t\terr = fmt.Errorf(\"node[%v] disk[%v] does not have any data partition\", node.Addr, args.DiskPath)\n\t\treturn nil, err\n\t}\n\n\tvar badPartitionIds []uint64\n\tfor _, bdp := range badPartitions {\n\t\tbadPartitionIds = append(badPartitionIds, bdp.PartitionID)\n\t}\n\trstMsg := fmt.Sprintf(\"receive decommissionDisk node[%v] disk[%v], badPartitionIds[%v] has offline successfully\",\n\t\tnode.Addr, args.DiskPath, badPartitionIds)\n\tif err = m.cluster.decommissionDisk(node, args.DiskPath, badPartitions); err != nil {\n\t\treturn nil, err\n\t}\n\tWarn(m.cluster.Name, rstMsg)\n\n\treturn proto.Success(\"success\"), nil\n\n}", "func (o *Partition) Delete(ctx context.Context, options map[string]dbus.Variant) (err error) {\n\terr = o.object.CallWithContext(ctx, InterfacePartition+\".Delete\", 0, options).Store()\n\treturn\n}", "func (s stage) partitionDisk(dev types.Disk, devAlias string) error {\n\tif cutil.IsTrue(dev.WipeTable) {\n\t\top := sgdisk.Begin(s.Logger, devAlias)\n\t\ts.Logger.Info(\"wiping partition table requested on %q\", devAlias)\n\t\top.WipeTable(true)\n\t\tif err := op.Commit(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Ensure all partitions with number 0 are last\n\tsort.Stable(PartitionList(dev.Partitions))\n\n\top := sgdisk.Begin(s.Logger, devAlias)\n\n\tdiskInfo, err := s.getPartitionMap(devAlias)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// get a list of parititions that have size and start 0 replaced with the real sizes\n\t// that would be used if all specified partitions were to be created anew.\n\t// Also calculate sectors for all of the start/size values.\n\tresolvedPartitions, err := s.getRealStartAndSize(dev, devAlias, diskInfo)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor _, part := range resolvedPartitions {\n\t\tshouldExist := partitionShouldExist(part)\n\t\tinfo, exists := diskInfo.GetPartition(part.Number)\n\t\tvar matchErr error\n\t\tif exists {\n\t\t\tmatchErr = partitionMatches(info, part)\n\t\t}\n\t\tmatches := exists && matchErr == nil\n\t\twipeEntry := cutil.IsTrue(part.WipePartitionEntry)\n\n\t\t// This is a translation of the matrix in the operator notes.\n\t\tswitch {\n\t\tcase !exists && !shouldExist:\n\t\t\ts.Logger.Info(\"partition %d specified as nonexistant and no partition was found. Success.\", part.Number)\n\t\tcase !exists && shouldExist:\n\t\t\top.CreatePartition(part)\n\t\tcase exists && !shouldExist && !wipeEntry:\n\t\t\treturn fmt.Errorf(\"partition %d exists but is specified as nonexistant and wipePartitionEntry is false\", part.Number)\n\t\tcase exists && !shouldExist && wipeEntry:\n\t\t\top.DeletePartition(part.Number)\n\t\tcase exists && shouldExist && matches:\n\t\t\ts.Logger.Info(\"partition %d found with correct specifications\", part.Number)\n\t\tcase exists && shouldExist && !wipeEntry && !matches:\n\t\t\tif partitionMatchesResize(info, part) {\n\t\t\t\ts.Logger.Info(\"resizing partition %d\", part.Number)\n\t\t\t\top.DeletePartition(part.Number)\n\t\t\t\tpart.Number = info.Number\n\t\t\t\tpart.GUID = &info.GUID\n\t\t\t\tpart.TypeGUID = &info.TypeGUID\n\t\t\t\tpart.Label = &info.Label\n\t\t\t\tpart.StartSector = &info.StartSector\n\t\t\t\top.CreatePartition(part)\n\t\t\t} else {\n\t\t\t\treturn fmt.Errorf(\"Partition %d didn't match: %v\", part.Number, matchErr)\n\t\t\t}\n\t\tcase exists && shouldExist && wipeEntry && !matches:\n\t\t\ts.Logger.Info(\"partition %d did not meet specifications, wiping partition entry and recreating\", part.Number)\n\t\t\top.DeletePartition(part.Number)\n\t\t\top.CreatePartition(part)\n\t\tdefault:\n\t\t\t// unfortunatey, golang doesn't check that all cases are handled exhaustively\n\t\t\treturn fmt.Errorf(\"Unreachable code reached when processing partition %d. golang--\", part.Number)\n\t\t}\n\t}\n\n\tif err := op.Commit(); err != nil {\n\t\treturn fmt.Errorf(\"commit failure: %v\", err)\n\t}\n\treturn nil\n}", "func (d *Disk) Kill() error {\n\treturn nil\n\n}", "func (mzk *MockZK) Delete(path string, version int32) error {\n\tmzk.Args = append(mzk.Args, []interface{}{\n\t\t\"delete\",\n\t\tpath,\n\t\tversion,\n\t})\n\treturn mzk.DeleteFn(path, version)\n}", "func (c *TestClient) GetDisk(project, zone, name string) (*compute.Disk, error) {\n\tif c.GetDiskFn != nil {\n\t\treturn c.GetDiskFn(project, zone, name)\n\t}\n\treturn c.client.GetDisk(project, zone, name)\n}", "func (*OktetoClusterHelper) Delete(_ string) error {\n\treturn ErrNotImplemented\n}", "func (c *MockFileStorageClient) DeleteFileSystem(ctx context.Context, id string) error {\n\treturn nil\n}", "func (s *Service) Delete(ctx context.Context, spec azure.Spec) error {\n\tnicSpec, ok := spec.(*Spec)\n\tif !ok {\n\t\treturn errors.New(\"invalid network interface Specification\")\n\t}\n\tklog.V(2).Infof(\"deleting nic %s\", nicSpec.Name)\n\tf, err := s.Client.Delete(ctx, s.Scope.ClusterConfig.ResourceGroup, nicSpec.Name)\n\tif err != nil && azure.ResourceNotFound(err) {\n\t\t// already deleted\n\t\treturn nil\n\t}\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"failed to delete network interface %s in resource group %s\", nicSpec.Name, s.Scope.ClusterConfig.ResourceGroup)\n\t}\n\n\terr = f.WaitForCompletionRef(ctx, s.Client.Client)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"cannot create, future response\")\n\t}\n\n\t_, err = f.Result(s.Client)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"result error\")\n\t}\n\tklog.V(2).Infof(\"successfully deleted nic %s\", nicSpec.Name)\n\treturn err\n}", "func (f *FakeInstance) Delete(_ context.Context, _ string) error {\n\tpanic(\"implement me\")\n}", "func (cv *CurveVolume) Delete(ctx context.Context) error {\n\targs := []string{\"delete\", \"--user\", cv.User, \"--filename\", cv.FilePath}\n\tklog.V(4).Infof(util.Log(ctx, \"starting curve %v\"), args)\n\toutput, err := util.ExecCommandHost(\"curve\", args)\n\tif err != nil {\n\t\tif strings.Contains(string(output), fmt.Sprintf(retFailFormat, retNotExist)) {\n\t\t\tklog.Warningf(util.Log(ctx, \"[curve] the file %s already deleted, ignore deleting it\"), cv.FilePath)\n\t\t\treturn nil\n\t\t}\n\t\treturn fmt.Errorf(\"failed to delete %s, err: %v, output: %v\", cv.FilePath, err, string(output))\n\t}\n\n\tklog.V(4).Infof(util.Log(ctx, \"[curve] successfully delete %v\"), cv.FilePath)\n\treturn nil\n}", "func (c *UDiskClient) NewDeleteUDiskRequest() *DeleteUDiskRequest {\n\treq := &DeleteUDiskRequest{}\n\n\t// setup request with client config\n\tc.Client.SetupRequest(req)\n\n\t// setup retryable with default retry policy (retry for non-create action and common error)\n\treq.SetRetryable(true)\n\treturn req\n}", "func (api *distributedservicecardAPI) Delete(obj *cluster.DistributedServiceCard) error {\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = apicl.ClusterV1().DistributedServiceCard().Delete(context.Background(), &obj.ObjectMeta)\n\t\treturn err\n\t}\n\n\tapi.ct.handleDistributedServiceCardEvent(&kvstore.WatchEvent{Object: obj, Type: kvstore.Deleted})\n\treturn nil\n}", "func (c *vmClient) Delete(uid meta.UID) error {\n\tlog.Debugf(\"Client.Delete; UID: %q, Kind: %s\", uid, api.KindVM)\n\treturn c.storage.Delete(api.KindVM, uid)\n}", "func (m *VirtualMachinesClientMock) Delete(ctx context.Context, resourceGroupName string, VMName string) *retry.Error {\n\targs := m.Called(resourceGroupName, VMName)\n\tif args.Error(1) != nil {\n\t\treturn &retry.Error{RawError: args.Error(1)}\n\t}\n\treturn nil\n}", "func Delete(path string) {\n\tres, err := deleteThis(path)\n\tif err != nil {\n\t\tlog.Println(err)\n\t}\n\tfmt.Printf(sep)\n\tfmt.Printf(\"DELETE result: %s\\n\", res)\n}", "func (util *AzUtil) DeleteVirtualMachine(rg string, name string) error {\n\tctx, cancel := getContextWithCancel()\n\tdefer cancel()\n\n\tvm, rerr := util.manager.azClient.virtualMachinesClient.Get(ctx, rg, name, \"\")\n\tif rerr != nil {\n\t\tif exists, _ := checkResourceExistsFromRetryError(rerr); !exists {\n\t\t\tklog.V(2).Infof(\"VirtualMachine %s/%s has already been removed\", rg, name)\n\t\t\treturn nil\n\t\t}\n\n\t\tklog.Errorf(\"failed to get VM: %s/%s: %s\", rg, name, rerr.Error())\n\t\treturn rerr.Error()\n\t}\n\n\tvhd := vm.VirtualMachineProperties.StorageProfile.OsDisk.Vhd\n\tmanagedDisk := vm.VirtualMachineProperties.StorageProfile.OsDisk.ManagedDisk\n\tif vhd == nil && managedDisk == nil {\n\t\tklog.Errorf(\"failed to get a valid os disk URI for VM: %s/%s\", rg, name)\n\t\treturn fmt.Errorf(\"os disk does not have a VHD URI\")\n\t}\n\n\tosDiskName := vm.VirtualMachineProperties.StorageProfile.OsDisk.Name\n\tvar nicName string\n\tvar err error\n\tnicID := (*vm.VirtualMachineProperties.NetworkProfile.NetworkInterfaces)[0].ID\n\tif nicID == nil {\n\t\tklog.Warningf(\"NIC ID is not set for VM (%s/%s)\", rg, name)\n\t} else {\n\t\tnicName, err = resourceName(*nicID)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tklog.Infof(\"found nic name for VM (%s/%s): %s\", rg, name, nicName)\n\t}\n\n\tklog.Infof(\"deleting VM: %s/%s\", rg, name)\n\tdeleteCtx, deleteCancel := getContextWithCancel()\n\tdefer deleteCancel()\n\n\tklog.Infof(\"waiting for VirtualMachine deletion: %s/%s\", rg, name)\n\trerr = util.manager.azClient.virtualMachinesClient.Delete(deleteCtx, rg, name)\n\t_, realErr := checkResourceExistsFromRetryError(rerr)\n\tif realErr != nil {\n\t\treturn realErr\n\t}\n\tklog.V(2).Infof(\"VirtualMachine %s/%s removed\", rg, name)\n\n\tif len(nicName) > 0 {\n\t\tklog.Infof(\"deleting nic: %s/%s\", rg, nicName)\n\t\tinterfaceCtx, interfaceCancel := getContextWithCancel()\n\t\tdefer interfaceCancel()\n\t\tklog.Infof(\"waiting for nic deletion: %s/%s\", rg, nicName)\n\t\tnicErr := util.manager.azClient.interfacesClient.Delete(interfaceCtx, rg, nicName)\n\t\t_, realErr := checkResourceExistsFromRetryError(nicErr)\n\t\tif realErr != nil {\n\t\t\treturn realErr\n\t\t}\n\t\tklog.V(2).Infof(\"interface %s/%s removed\", rg, nicName)\n\t}\n\n\tif vhd != nil {\n\t\taccountName, vhdContainer, vhdBlob, err := splitBlobURI(*vhd.URI)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tklog.Infof(\"found os disk storage reference: %s %s %s\", accountName, vhdContainer, vhdBlob)\n\n\t\tklog.Infof(\"deleting blob: %s/%s\", vhdContainer, vhdBlob)\n\t\tif err = util.DeleteBlob(accountName, vhdContainer, vhdBlob); err != nil {\n\t\t\t_, realErr := checkResourceExistsFromError(err)\n\t\t\tif realErr != nil {\n\t\t\t\treturn realErr\n\t\t\t}\n\t\t\tklog.V(2).Infof(\"Blob %s/%s removed\", rg, vhdBlob)\n\t\t}\n\t} else if managedDisk != nil {\n\t\tif osDiskName == nil {\n\t\t\tklog.Warningf(\"osDisk is not set for VM %s/%s\", rg, name)\n\t\t} else {\n\t\t\tklog.Infof(\"deleting managed disk: %s/%s\", rg, *osDiskName)\n\t\t\tdisksCtx, disksCancel := getContextWithCancel()\n\t\t\tdefer disksCancel()\n\t\t\tdiskErr := util.manager.azClient.disksClient.Delete(disksCtx, util.manager.config.SubscriptionID, rg, *osDiskName)\n\t\t\t_, realErr := checkResourceExistsFromRetryError(diskErr)\n\t\t\tif realErr != nil {\n\t\t\t\treturn realErr\n\t\t\t}\n\t\t\tklog.V(2).Infof(\"disk %s/%s removed\", rg, *osDiskName)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (d *DiskStore) Delete(hash string) error {\n\terr := d.initOnce()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\thas, err := d.Has(hash)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif !has {\n\t\treturn nil\n\t}\n\n\terr = os.Remove(d.path(hash))\n\treturn errors.Err(err)\n}", "func (d *driver) Delete(ctx context.Context, path string) error {\n\tdefer debugTime()()\n\td.rootlock.Lock()\n\tdefer d.rootlock.Unlock()\n\tlog.Error(\"roothash: \", d.roothash)\n\tnewParentHash, err := d.shell.Patch(d.roothash, \"rm-link\", path[1:])\n\tif err != nil {\n\t\tlog.Error(\"delete err: \", err)\n\t\tif err.Error() == \"merkledag: not found\" {\n\t\t\tfmt.Println(\"PATHNOTFOUND HAPPY HAPPY JOY JOY\")\n\t\t\treturn storagedriver.PathNotFoundError{Path: path}\n\t\t} else {\n\t\t\tfmt.Println(\"GOT A BAD ERROR: \", err)\n\t\t\treturn err\n\t\t}\n\t}\n\n\td.roothash = newParentHash\n\td.publishHash(newParentHash)\n\treturn nil\n}", "func (client StorageGatewayClient) deleteFileSystem(ctx context.Context, request common.OCIRequest) (common.OCIResponse, error) {\n\thttpRequest, err := request.HTTPRequest(http.MethodDelete, \"/storageGateways/{storageGatewayId}/fileSystems/{fileSystemName}\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar response DeleteFileSystemResponse\n\tvar httpResponse *http.Response\n\thttpResponse, err = client.Call(ctx, &httpRequest)\n\tdefer common.CloseBodyIfValid(httpResponse)\n\tresponse.RawResponse = httpResponse\n\tif err != nil {\n\t\treturn response, err\n\t}\n\n\terr = common.UnmarshalResponse(httpResponse, &response)\n\treturn response, err\n}", "func (c *TestClient) CreateDisk(project, zone string, d *compute.Disk) error {\n\tif c.CreateDiskFn != nil {\n\t\treturn c.CreateDiskFn(project, zone, d)\n\t}\n\treturn c.client.CreateDisk(project, zone, d)\n}", "func (s *azureMachineService) Delete() error {\n\tvmSpec := &virtualmachines.Spec{\n\t\tName: s.machineScope.Name(),\n\t}\n\n\terr := s.virtualMachinesSvc.Delete(s.clusterScope.Context, vmSpec)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"failed to delete machine\")\n\t}\n\n\tnetworkInterfaceSpec := &networkinterfaces.Spec{\n\t\tName: azure.GenerateNICName(s.machineScope.Name()),\n\t\tVnetName: azure.GenerateVnetName(s.clusterScope.Name()),\n\t}\n\n\terr = s.networkInterfacesSvc.Delete(s.clusterScope.Context, networkInterfaceSpec)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"Unable to delete network interface\")\n\t}\n\n\tpublicIPSpec := &publicips.Spec{\n\t\tName: azure.GenerateNICName(s.machineScope.Name()) + \"-public-ip\",\n\t}\n\n\terr = s.publicIPSvc.Delete(s.clusterScope.Context, publicIPSpec)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"unable to delete publicIP\")\n\t}\n\n\tOSDiskSpec := &disks.Spec{\n\t\tName: azure.GenerateOSDiskName(s.machineScope.Name()),\n\t}\n\terr = s.disksSvc.Delete(s.clusterScope.Context, OSDiskSpec)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"Failed to delete OS disk of machine %s\", s.machineScope.Name())\n\t}\n\n\treturn nil\n}", "func (qd *Qdisc) Delete(info *Object) error {\n\tif info == nil {\n\t\treturn ErrNoArg\n\t}\n\toptions, err := validateQdiscObject(rtmDelQdisc, info)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn qd.action(rtmDelQdisc, netlink.HeaderFlags(0), info, options)\n}", "func Delete(c *gophercloud.ServiceClient, networkID string) os.DeleteResult {\n\treturn os.Delete(c, networkID)\n}", "func (d *DefaultDriver) DeleteVolume(volumeID string) error {\n\treturn &errors.ErrNotSupported{\n\t\tType: \"Function\",\n\t\tOperation: \"DeleteVolume()\",\n\t}\n}", "func (api *distributedservicecardAPI) SyncDelete(obj *cluster.DistributedServiceCard) error {\n\tvar writeErr error\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t_, writeErr = apicl.ClusterV1().DistributedServiceCard().Delete(context.Background(), &obj.ObjectMeta)\n\t}\n\n\tif writeErr == nil {\n\t\tapi.ct.handleDistributedServiceCardEvent(&kvstore.WatchEvent{Object: obj, Type: kvstore.Deleted})\n\t}\n\n\treturn writeErr\n}", "func (c *Client) Delete(d core.Digest) error {\n\t_, err := httputil.Delete(fmt.Sprintf(\"http://%s/blobs/%s\", c.addr, d))\n\treturn err\n}", "func (in *Disk) DeepCopy() *Disk {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(Disk)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (in *Disk) DeepCopy() *Disk {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(Disk)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (m MockRepositoryStore) DeleteReplica(ctx context.Context, repositoryID int64, storage string) error {\n\tif m.DeleteReplicaFunc == nil {\n\t\treturn nil\n\t}\n\n\treturn m.DeleteReplicaFunc(ctx, repositoryID, storage)\n}", "func (c *TestClient) DeleteMachineImage(project, name string) error {\n\tif c.DeleteMachineImageFn != nil {\n\t\treturn c.DeleteMachineImageFn(project, name)\n\t}\n\treturn c.client.DeleteMachineImage(project, name)\n}", "func (d *common) Delete() error {\n\tisUsed, err := d.isUsed()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif isUsed {\n\t\treturn fmt.Errorf(\"Cannot delete an ACL that is in use\")\n\t}\n\n\treturn d.state.DB.Cluster.DeleteNetworkACL(d.id)\n}", "func (of OperatorFactory) DiskOrder() types.DiskOrder {\n\treturn operator{}.DiskOrder()\n}", "func (of OperatorFactory) DiskOrder() types.DiskOrder {\n\treturn Operator{}.DiskOrder()\n}", "func (c *TestClient) ResizeDisk(project, zone, disk string, drr *compute.DisksResizeRequest) error {\n\tif c.ResizeDiskFn != nil {\n\t\treturn c.ResizeDiskFn(project, zone, disk, drr)\n\t}\n\treturn c.client.ResizeDisk(project, zone, disk, drr)\n}", "func Delete() error {\n\n}", "func (client DatasetClient) Delete(ctx context.Context, datasetID string) (result autorest.Response, err error) {\n if tracing.IsEnabled() {\n ctx = tracing.StartSpan(ctx, fqdn + \"/DatasetClient.Delete\")\n defer func() {\n sc := -1\n if result.Response != nil {\n sc = result.Response.StatusCode\n }\n tracing.EndSpan(ctx, sc, err)\n }()\n }\n req, err := client.DeletePreparer(ctx, datasetID)\n if err != nil {\n err = autorest.NewErrorWithError(err, \"creator.DatasetClient\", \"Delete\", nil , \"Failure preparing request\")\n return\n }\n\n resp, err := client.DeleteSender(req)\n if err != nil {\n result.Response = resp\n err = autorest.NewErrorWithError(err, \"creator.DatasetClient\", \"Delete\", resp, \"Failure sending request\")\n return\n }\n\n result, err = client.DeleteResponder(resp)\n if err != nil {\n err = autorest.NewErrorWithError(err, \"creator.DatasetClient\", \"Delete\", resp, \"Failure responding to request\")\n return\n }\n\n return\n}", "func (s *storageImpl) Delete(ctx context.Context, module, version string) error {\n\tconst op errors.Op = \"fs.Delete\"\n\tctx, span := observ.StartSpan(ctx, op.String())\n\tdefer span.End()\n\tversionedPath := s.versionLocation(module, version)\n\texists, err := s.Exists(ctx, module, version)\n\tif err != nil {\n\t\treturn errors.E(op, err, errors.M(module), errors.V(version))\n\t}\n\tif !exists {\n\t\treturn errors.E(op, errors.M(module), errors.V(version), errors.KindNotFound)\n\t}\n\treturn s.filesystem.RemoveAll(versionedPath)\n}", "func (s *Module) DiskWrite(name string, image string) error {\n\tpath, err := s.findDisk(name)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"couldn't find disk with id: %s\", name)\n\t}\n\n\tif !s.isEmptyDisk(path) {\n\t\tlog.Debug().Str(\"disk\", path).Msg(\"disk already has a filesystem. no write\")\n\t\treturn nil\n\t}\n\n\tsource, err := os.Open(image)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to open image\")\n\t}\n\tdefer source.Close()\n\tfile, err := os.OpenFile(path, os.O_WRONLY, 0644)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer file.Close()\n\timgStat, err := source.Stat()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to stat image\")\n\t}\n\tfileStat, err := file.Stat()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to state disk\")\n\t}\n\n\tif imgStat.Size() > fileStat.Size() {\n\t\treturn fmt.Errorf(\"image size is bigger than disk\")\n\t}\n\n\t_, err = io.Copy(file, source)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to write disk image\")\n\t}\n\n\treturn nil\n}", "func (m *Module) InternalDelete(ctx context.Context, dbAlias, project, col string, req *model.DeleteRequest) error {\n\tm.RLock()\n\tdefer m.RUnlock()\n\n\tcrud, err := m.getCrudBlock(dbAlias)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err := crud.IsClientSafe(ctx); err != nil {\n\t\treturn err\n\t}\n\n\t// Adjust where clause\n\tdbType, err := m.getDBType(dbAlias)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif err := helpers.AdjustWhereClause(ctx, dbAlias, model.DBType(dbType), col, m.schemaDoc, req.Find); err != nil {\n\t\treturn err\n\t}\n\n\t// Perform the delete operation\n\tn, err := crud.Delete(ctx, col, req)\n\n\t// Invoke the metric hook if the operation was successful\n\tif err == nil {\n\t\tm.metricHook(m.project, dbAlias, col, n, model.Update)\n\t}\n\n\treturn err\n}", "func (o ResourcePolicySnapshotSchedulePolicyRetentionPolicyResponsePtrOutput) OnSourceDiskDelete() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ResourcePolicySnapshotSchedulePolicyRetentionPolicyResponse) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.OnSourceDiskDelete\n\t}).(pulumi.StringPtrOutput)\n}", "func (d *driver) Delete(ctx context.Context, path string) error {\n\treturn d.Bucket.Delete(ctx, path)\n}", "func deleteDiskTables(dbDir string, prefixes ...string) error {\n\tfor _, prefix := range prefixes {\n\t\tdataPath := path.Join(dbDir, prefix+diskTableDataFileName)\n\t\tif err := os.Remove(dataPath); err != nil {\n\t\t\treturn fmt.Errorf(\"failed to remove data file %s: %w\", dataPath, err)\n\t\t}\n\n\t\tindexPath := path.Join(dbDir, prefix+diskTableIndexFileName)\n\t\tif err := os.Remove(indexPath); err != nil {\n\t\t\treturn fmt.Errorf(\"failed to remove data file %s: %w\", indexPath, err)\n\t\t}\n\n\t\tsparseIndexPath := path.Join(dbDir, prefix+diskTableSparseIndexFileName)\n\t\tif err := os.Remove(sparseIndexPath); err != nil {\n\t\t\treturn fmt.Errorf(\"failed to remove data file %s: %w\", sparseIndexPath, err)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (m *FileRequestBuilder) Delete(ctx context.Context, requestConfiguration *FileRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.CreateDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": ia572726a95efa92ddd544552cd950653dc691023836923576b2f4bf716cf204a.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": ia572726a95efa92ddd544552cd950653dc691023836923576b2f4bf716cf204a.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.requestAdapter.SendNoContentAsync(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func (cli *FakeDatabaseClient) DeleteDir(ctx context.Context, in *dbdpb.DeleteDirRequest, opts ...grpc.CallOption) (*dbdpb.DeleteDirResponse, error) {\n\tpanic(\"implement me\")\n}", "func (m *GoMockDiskStorage) Delete(filePath string) error {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"Delete\", filePath)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func (s stack) DeleteVolume(ctx context.Context, ref string) fail.Error {\n\tif valid.IsNil(s) {\n\t\treturn fail.InvalidInstanceError()\n\t}\n\treturn fail.NotImplementedError(\"implement me\")\n}", "func (a DBFSAPI) Delete(path string, recursive bool) error {\n\tdeleteRequest := model.DBFSDeleteRequest{\n\t\tPath: path,\n\t\tRecursive: recursive,\n\t}\n\t_, err := a.Client.performQuery(http.MethodPost, \"/dbfs/delete\", \"2.0\", nil, deleteRequest, nil)\n\n\treturn err\n}", "func (a *Azure) DeleteNIC(ctx *lepton.Context, nic *network.Interface) error {\n\tlogger := ctx.Logger()\n\n\tnicClient := a.getNicClient()\n\n\tlogger.Infof(\"Deleting %s...\", *nic.ID)\n\tnicName := getAzureResourceNameFromID(*nic.ID)\n\tnicDeleteTask, err := nicClient.Delete(context.TODO(), a.groupName, nicName)\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn errors.New(\"error deleting network interface controller\")\n\t}\n\n\terr = nicDeleteTask.WaitForCompletionRef(context.TODO(), nicClient.Client)\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn errors.New(\"error waiting for network interface controller deleting\")\n\t}\n\n\treturn nil\n}", "func (c *ImageController) Delete(ctx *app.DeleteImageContext) error {\n\t// ImageController_Delete: start_implement\n\n\t// Put your logic here\n\n\t// ImageController_Delete: end_implement\n\treturn nil\n}", "func (s *Service) Delete(ctx context.Context) error {\n\tctx, _, done := tele.StartSpanWithLogger(ctx, \"privatedns.Service.Delete\")\n\tdefer done()\n\n\tctx, cancel := context.WithTimeout(ctx, reconciler.DefaultAzureServiceReconcileTimeout)\n\tdefer cancel()\n\n\tzoneSpec, links, _ := s.Scope.PrivateDNSSpec()\n\tif zoneSpec == nil {\n\t\treturn nil\n\t}\n\n\tmanaged, err := s.deleteLinks(ctx, links)\n\tif managed {\n\t\ts.Scope.UpdateDeleteStatus(infrav1.PrivateDNSLinkReadyCondition, serviceName, err)\n\t}\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tmanaged, err = s.deleteZone(ctx, zoneSpec)\n\tif managed {\n\t\ts.Scope.UpdateDeleteStatus(infrav1.PrivateDNSZoneReadyCondition, serviceName, err)\n\t\ts.Scope.UpdateDeleteStatus(infrav1.PrivateDNSRecordReadyCondition, serviceName, err)\n\t}\n\n\treturn err\n}", "func (o ResourcePolicySnapshotSchedulePolicyRetentionPolicyOutput) OnSourceDiskDelete() ResourcePolicySnapshotSchedulePolicyRetentionPolicyOnSourceDiskDeletePtrOutput {\n\treturn o.ApplyT(func(v ResourcePolicySnapshotSchedulePolicyRetentionPolicy) *ResourcePolicySnapshotSchedulePolicyRetentionPolicyOnSourceDiskDelete {\n\t\treturn v.OnSourceDiskDelete\n\t}).(ResourcePolicySnapshotSchedulePolicyRetentionPolicyOnSourceDiskDeletePtrOutput)\n}", "func (c *jsiiProxy_CfnPartition) AddDeletionOverride(path *string) {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"addDeletionOverride\",\n\t\t[]interface{}{path},\n\t)\n}", "func (o *ContentUnitDerivation) Delete(exec boil.Executor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"mdbmodels: no ContentUnitDerivation provided for delete\")\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), contentUnitDerivationPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"content_unit_derivations\\\" WHERE \\\"source_id\\\"=$1 AND \\\"derived_id\\\"=$2\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: unable to delete from content_unit_derivations\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: failed to get rows affected by delete for content_unit_derivations\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (r *regulator) Delete(ctx context.Context, path string) error {\n\tr.enter()\n\tdefer r.exit()\n\n\treturn r.StorageDriver.Delete(ctx, path)\n}", "func SimpleDelete() (pred func(x, y float64, e interface{}) bool) {\n\tpred = func(x, y float64, e interface{}) bool {\n\t\treturn true\n\t}\n\treturn\n}", "func (opts *deleteOpts) Delete(d interface{}, a ...string) error {\n\tif !opts.confirm {\n\t\tfmt.Println(opts.FailMessage())\n\t\treturn nil\n\t}\n\n\tvar err error\n\tswitch f := d.(type) {\n\tcase func(string) error:\n\t\terr = f(opts.entry)\n\tcase func(string, string) error:\n\t\terr = f(a[0], opts.entry)\n\tcase func(string, string, string) error:\n\t\terr = f(a[0], a[1], opts.entry)\n\tdefault:\n\t\treturn errors.New(\"invalid\")\n\t}\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfmt.Printf(opts.SuccessMessage(), opts.entry)\n\n\treturn nil\n}", "func (z *ZkPlus) Delete(path string, version int32) error {\n\tz.forPath(path).Log(logkey.ZkMethod, \"Delete\")\n\treturn z.blockOnConn().Delete(z.realPath(path), version)\n}", "func (s *segment) Delete() error {\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\n\tif s.f == nil {\n\t\treturn nil\n\t}\n\n\terr := s.close()\n\tif err != nil {\n\t\treturn err\n\t}\n\terrA := os.Remove(s.filePath)\n\tif errA != nil {\n\t\treturn errSegmentRemove(errA)\n\t}\n\n\t// do we need to do this?\n\ts.f = nil\n\ts = nil\n\n\treturn nil\n}", "func (o LocalCopyPtrOutput) Disk() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *LocalCopy) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Disk\n\t}).(pulumi.StringPtrOutput)\n}", "func DiskPercent() (string, error) {\n\tusage, err := Disk()\n\tif len(usage) > 0 {\n\t\tuseDisk := fmt.Sprintf(\"%.2f\", usage[0].UsedPercent)\n\t\treturn useDisk, err\n\t}\n\n\treturn \"0.00\", err\n}", "func (o ResourcePolicySnapshotSchedulePolicyRetentionPolicyResponseOutput) OnSourceDiskDelete() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ResourcePolicySnapshotSchedulePolicyRetentionPolicyResponse) string {\n\t\treturn v.OnSourceDiskDelete\n\t}).(pulumi.StringOutput)\n}", "func DiskVMVirtiofsdStop(socketPath string, pidPath string) error {\n\tif shared.PathExists(pidPath) {\n\t\tproc, err := subprocess.ImportProcess(pidPath)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\terr = proc.Stop()\n\t\t// The virtiofsd process will terminate automatically once the VM has stopped.\n\t\t// We therefore should only return an error if it's still running and fails to stop.\n\t\tif err != nil && err != subprocess.ErrNotRunning {\n\t\t\treturn err\n\t\t}\n\n\t\t// Remove PID file if needed.\n\t\tos.Remove(pidPath)\n\t}\n\n\t// Remove socket file if needed.\n\tos.Remove(socketPath)\n\n\treturn nil\n}", "func (v *ImageClient) Delete(imageName string) error {\n\n\t//Construct the composite key to select the entry\n\tkey := ImageKey{\n\t\t// Owner:\townerName,\n\t\t// ClusterName:\tclusterName,\n\t\tImageName: imageName,\n\t}\n\terr := v.util.DBDelete(v.storeName, key, v.tagMeta)\n\n\t//Delete image from FS\n\tfilePath, _, err := v.GetDirPath(imageName)\n\tif err != nil {\n\t\treturn pkgerrors.Wrap(err, \"Get file path\")\n\t}\n\terr = os.Remove(filePath)\n if err != nil {\n return pkgerrors.Wrap(err, \"Delete image file\")\n }\n\n\treturn nil\n}", "func (c *DriversController) DeleteDriver() {\n\tresult := models.DeleteDriver(c.Ctx.Input.Param(ID_PARAMETER))\n\tc.Ctx.Output.Body([]byte(result))\n}", "func (srv *VolumeService) Delete(ref string) error {\n\tvol, err := srv.Get(ref)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Volume '%s' does not exists\", ref)\n\t}\n\treturn srv.provider.DeleteVolume(vol.ID)\n}", "func (c *OperatorDNS) DeleteRecord(record SrvRecord) error {\n\treturn ErrNotImplemented\n}", "func (d *Double) Delete(stackName string) error {\n\treturn d.DeleteFn(stackName)\n}", "func (fs *Ipfs) Delete(path string) error {\n\t// Remoe file if on disk and unpinn\n\tif fname, err := fs.makeFilename(path); err == nil {\n\t\tos.Remove(fname)\n\t}\n\n\tipath := ipath.New(path)\n\treturn fs.coreAPI.Pin().Rm(context.Background(), ipath)\n}", "func DiskSizeBytes(disk string) uint64 {\n\tmsg := `\nThe DiskSizeBytes() function has been DEPRECATED and will be\nremoved in the 1.0 release of ghw. Please use the Disk.SizeBytes attribute.\n`\n\twarn(msg)\n\tctx := contextFromEnv()\n\treturn ctx.diskSizeBytes(disk)\n}", "func (s *Reconciler) Delete(ctx context.Context) error {\n\tvmSpec := &virtualmachines.Spec{\n\t\tName: s.scope.Machine.Name,\n\t}\n\n\t// Getting a vm object does not work here so let's assume\n\t// an instance is really being deleted\n\tif s.scope.Machine.Annotations == nil {\n\t\ts.scope.Machine.Annotations = make(map[string]string)\n\t}\n\ts.scope.Machine.Annotations[MachineInstanceStateAnnotationName] = string(machinev1.VMStateDeleting)\n\tvmStateDeleting := machinev1.VMStateDeleting\n\ts.scope.MachineStatus.VMState = &vmStateDeleting\n\n\terr := s.virtualMachinesSvc.Delete(ctx, vmSpec)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to delete machine: %w\", err)\n\t}\n\n\tosDiskSpec := &disks.Spec{\n\t\tName: azure.GenerateOSDiskName(s.scope.Machine.Name),\n\t}\n\terr = s.disksSvc.Delete(ctx, osDiskSpec)\n\tif err != nil {\n\t\tmetrics.RegisterFailedInstanceDelete(&metrics.MachineLabels{\n\t\t\tName: s.scope.Machine.Name,\n\t\t\tNamespace: s.scope.Machine.Namespace,\n\t\t\tReason: err.Error(),\n\t\t})\n\t\treturn fmt.Errorf(\"failed to delete OS disk: %w\", err)\n\t}\n\n\tif s.scope.MachineConfig.Vnet == \"\" {\n\t\treturn fmt.Errorf(\"MachineConfig vnet is missing on machine %s\", s.scope.Machine.Name)\n\t}\n\n\tnetworkInterfaceSpec := &networkinterfaces.Spec{\n\t\tName: azure.GenerateNetworkInterfaceName(s.scope.Machine.Name),\n\t\tVnetName: s.scope.MachineConfig.Vnet,\n\t}\n\n\terr = s.networkInterfacesSvc.Delete(ctx, networkInterfaceSpec)\n\tif err != nil {\n\t\tmetrics.RegisterFailedInstanceDelete(&metrics.MachineLabels{\n\t\t\tName: s.scope.Machine.Name,\n\t\t\tNamespace: s.scope.Machine.Namespace,\n\t\t\tReason: err.Error(),\n\t\t})\n\t\treturn fmt.Errorf(\"Unable to delete network interface: %w\", err)\n\t}\n\n\tif s.scope.MachineConfig.PublicIP {\n\t\tpublicIPName, err := azure.GenerateMachinePublicIPName(s.scope.MachineConfig.Name, s.scope.Machine.Name)\n\t\tif err != nil {\n\t\t\t// Only when the generated name is longer than allowed by the Azure portal\n\t\t\t// That can happen only when\n\t\t\t// - machine name is changed (can't happen without creating a new CR)\n\t\t\t// - cluster name is changed (could happen but then we will get different name anyway)\n\t\t\t// - machine CR was created with too long public ip name (in which case no instance was created)\n\t\t\tklog.Info(\"Generated public IP name was too long, skipping deletion of the resource\")\n\t\t\treturn nil\n\t\t}\n\n\t\terr = s.publicIPSvc.Delete(ctx, &publicips.Spec{\n\t\t\tName: publicIPName,\n\t\t})\n\t\tif err != nil {\n\t\t\tmetrics.RegisterFailedInstanceDelete(&metrics.MachineLabels{\n\t\t\t\tName: s.scope.Machine.Name,\n\t\t\t\tNamespace: s.scope.Machine.Namespace,\n\t\t\t\tReason: err.Error(),\n\t\t\t})\n\t\t\treturn fmt.Errorf(\"unable to delete Public IP: %w\", err)\n\t\t}\n\t}\n\n\t// Delete the availability set with the given name if no virtual machines are attached to it.\n\tif err := s.availabilitySetsSvc.Delete(ctx, &availabilitysets.Spec{\n\t\tName: s.getAvailibilitySetName(),\n\t}); err != nil {\n\t\treturn fmt.Errorf(\"failed to delete availability set: %w\", err)\n\t}\n\n\treturn nil\n}" ]
[ "0.7045807", "0.69736844", "0.69469583", "0.651236", "0.649518", "0.62397087", "0.62128055", "0.5853082", "0.57757497", "0.5718966", "0.571671", "0.5661196", "0.5647591", "0.5633581", "0.561698", "0.5580925", "0.5578185", "0.55739754", "0.5523698", "0.5505451", "0.5502442", "0.5400823", "0.5399816", "0.5386664", "0.53666043", "0.5352346", "0.53298765", "0.53237647", "0.53234506", "0.53169024", "0.52883625", "0.5280521", "0.52769244", "0.527339", "0.52726275", "0.52565086", "0.52477384", "0.5222851", "0.5159437", "0.51534593", "0.51461315", "0.51396394", "0.5115995", "0.51124805", "0.51030093", "0.50985914", "0.50838315", "0.5080545", "0.5072954", "0.50712305", "0.50574833", "0.5055079", "0.5054681", "0.5053159", "0.50412375", "0.5026715", "0.5025596", "0.5025596", "0.50242466", "0.50152886", "0.5013554", "0.5005891", "0.50017273", "0.49963608", "0.4995895", "0.49924475", "0.49786934", "0.49763826", "0.49761438", "0.49754593", "0.49734527", "0.4967314", "0.49637428", "0.49412853", "0.49382055", "0.49356905", "0.49333066", "0.49319002", "0.4926913", "0.4923811", "0.4919732", "0.49193844", "0.49183807", "0.4914588", "0.49108756", "0.49107817", "0.49076793", "0.4907233", "0.49063087", "0.4903968", "0.49013638", "0.4898213", "0.48968968", "0.4893892", "0.48891583", "0.48784286", "0.48761526", "0.48752165", "0.48723298", "0.48698258" ]
0.7246492
0
DeleteForwardingRule uses the override method DeleteForwardingRuleFn or the real implementation.
func (c *TestClient) DeleteForwardingRule(project, region, name string) error { if c.DeleteForwardingRuleFn != nil { return c.DeleteForwardingRuleFn(project, region, name) } return c.client.DeleteForwardingRule(project, region, name) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (f *fakeLB) DeleteForwardingRule(_ context.Context, _, _ string) error {\n\tpanic(\"implement me\")\n}", "func (m *MockFirewallServiceIface) DeletePortForwardingRule(p *DeletePortForwardingRuleParams) (*DeletePortForwardingRuleResponse, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"DeletePortForwardingRule\", p)\n\tret0, _ := ret[0].(*DeletePortForwardingRuleResponse)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func DeletePortForwarding(t *testing.T, client *gophercloud.ServiceClient, fipID string, pfID string) {\n\tt.Logf(\"Attempting to delete the port forwarding with ID %s for floating IP with ID %s\", pfID, fipID)\n\n\terr := portforwarding.Delete(client, fipID, pfID).ExtractErr()\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to delete Port forwarding with ID %s for floating IP with ID %s\", pfID, fipID)\n\t}\n\tt.Logf(\"Successfully deleted the port forwarding with ID %s for floating IP with ID %s\", pfID, fipID)\n\n}", "func DeleteRule(rule string) {\n\tdelete(customRuleFuncMap, rule)\n}", "func Delete(c *gophercloud.ServiceClient, floatingIpId string, pfId string) (r DeleteResult) {\n\t_, r.Err = c.Delete(singlePortForwardingUrl(c, floatingIpId, pfId), nil)\n\treturn\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRulesRule(ctx context.Context, name string, policyId string, sequenceId int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/frinx-openconfig-network-instance:rule/{sequence-id}/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"sequence-id\"+\"}\", fmt.Sprintf(\"%v\", sequenceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRulesRuleAction(ctx context.Context, name string, policyId string, sequenceId int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/frinx-openconfig-network-instance:rule/{sequence-id}/frinx-openconfig-network-instance:action/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"sequence-id\"+\"}\", fmt.Sprintf(\"%v\", sequenceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (_m *ComputeAPI) DeleteFirewallRule(project string, firewall string) {\n\t_m.Called(project, firewall)\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRulesRuleTransport(ctx context.Context, name string, policyId string, sequenceId int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/frinx-openconfig-network-instance:rule/{sequence-id}/frinx-openconfig-network-instance:transport/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"sequence-id\"+\"}\", fmt.Sprintf(\"%v\", sequenceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func DeleteRule(ctx context.Context, p *Protocol, routeID routing.RouteID) error {\n\tif err := p.WritePacket(PacketDeleteRules, []routing.RouteID{routeID}); err != nil {\n\t\treturn err\n\t}\n\tvar res []routing.RouteID\n\tif err := readAndDecodePacketWithTimeout(ctx, p, &res); err != nil {\n\t\treturn err\n\t}\n\tif len(res) == 0 {\n\t\treturn errors.New(\"empty response\")\n\t}\n\treturn nil\n}", "func (rh *ruleHandler) internalDelete(name string) {\n\t// deletes relevant discoverer delegate\n\tif delegate, exists := rh.d.delegates[name]; exists {\n\t\tdelegate.handler.DeleteMissing(nil)\n\t\tdelete(rh.d.delegates, name)\n\t}\n}", "func (s *ForwardingServiceOp) Delete(portForwardingConfig *PortForwardingConfig) error {\n\tportForwardingJSON, err := json.Marshal(*portForwardingConfig)\n\tif err != nil {\n\t\treturn err\n\t}\n\treq, err := http.NewRequest(\"POST\", s.client.ServerURL+\"/cloudapi/portforwarding/deleteByPort\", bytes.NewBuffer(portForwardingJSON))\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = s.client.Do(req)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (ruleset *DnsForwardingRuleset) ValidateDelete() (admission.Warnings, error) {\n\tvalidations := ruleset.deleteValidations()\n\tvar temp any = ruleset\n\tif runtimeValidator, ok := temp.(genruntime.Validator); ok {\n\t\tvalidations = append(validations, runtimeValidator.DeleteValidations()...)\n\t}\n\treturn genruntime.ValidateDelete(validations)\n}", "func (mr *MockFirewallServiceIfaceMockRecorder) DeletePortForwardingRule(p interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"DeletePortForwardingRule\", reflect.TypeOf((*MockFirewallServiceIface)(nil).DeletePortForwardingRule), p)\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRulesRuleIpv6(ctx context.Context, name string, policyId string, sequenceId int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/frinx-openconfig-network-instance:rule/{sequence-id}/frinx-openconfig-network-instance:ipv6/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"sequence-id\"+\"}\", fmt.Sprintf(\"%v\", sequenceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func DeleteJumpRule(v, tableName, srcChainName, dstChainName string) error {\n\tr, err := GetJumpRule(v, tableName, srcChainName, dstChainName)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif r == nil {\n\t\treturn nil\n\t}\n\n\tconn, err := initNftConn()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttb := &nftables.Table{\n\t\tName: tableName,\n\t}\n\tif v == \"4\" {\n\t\ttb.Family = nftables.TableFamilyIPv4\n\t} else {\n\t\ttb.Family = nftables.TableFamilyIPv6\n\t}\n\n\tch := &nftables.Chain{\n\t\tName: srcChainName,\n\t\tTable: tb,\n\t}\n\n\tconn.DelRule(&nftables.Rule{\n\t\tTable: tb,\n\t\tChain: ch,\n\t\t//&nftables.Chain{Name: r.Chain.Name, Type: r.Chain.Type},\n\t\tHandle: r.Handle,\n\t})\n\n\tif err := conn.Flush(); err != nil {\n\t\treturn fmt.Errorf(\n\t\t\t\"error deleting jump rule to %s chain found in chain %s in %s table: %s\",\n\t\t\tdstChainName, r.Chain.Name, r.Table.Name, err,\n\t\t)\n\t}\n\n\treturn nil\n}", "func (f *fakeLB) GetForwardingRule(_ context.Context, _, _ string) (*govultr.ForwardingRule, *http.Response, error) {\n\tpanic(\"implement me\")\n}", "func (probe *BridgeOfProbe) delRule(rule *Rule) {\n\tlogging.GetLogger().Infof(\"Rule %v deleted\", rule.UUID)\n\tg := probe.OvsOfProbe.Graph\n\tg.Lock()\n\tdefer g.Unlock()\n\n\truleNode := g.LookupFirstNode(graph.Metadata{\"UUID\": rule.UUID})\n\tif ruleNode != nil {\n\t\tg.DelNode(ruleNode)\n\t}\n}", "func DeleteIptableRule(version, tableName, chainName, match, target string) error {\n\tparams := fmt.Sprintf(\"-t %s -D %s %s -j %s\", tableName, chainName, match, target)\n\treturn RunCmd(version, params)\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRules(ctx context.Context, name string, policyId string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRulesRuleIpv4(ctx context.Context, name string, policyId string, sequenceId int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/frinx-openconfig-network-instance:rule/{sequence-id}/frinx-openconfig-network-instance:ipv4/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"sequence-id\"+\"}\", fmt.Sprintf(\"%v\", sequenceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func DeleteRule(table Table, chain Chain, args ...string) error {\n\tfullArgs := makeFullArgs(table, opDeleteRule, chain, args...)\n\tout, err := run(cmd, fullArgs...)\n\tif err != nil {\n\t\treturn trace.Wrap(err, \"failed to delete %v chain %v rule %v: %s\", table, chain, args, out)\n\t}\n\treturn nil\n}", "func (self *PolicyAgent) DelRule(rule *OfnetPolicyRule, ret *bool) error {\n\tlog.Infof(\"Received DelRule: %+v\", rule)\n\n\t// Gte the rule\n\tself.mutex.Lock()\n\tdefer self.mutex.Unlock()\n\tcache := self.Rules[rule.RuleId]\n\tif cache == nil {\n\t\tlog.Errorf(\"Could not find rule: %+v\", rule)\n\t\treturn errors.New(\"rule not found\")\n\t}\n\n\t// Delete the Flow\n\terr := cache.flow.Delete()\n\tif err != nil {\n\t\tlog.Errorf(\"Error deleting flow: %+v. Err: %v\", rule, err)\n\t}\n\n\t// Delete the rule from cache\n\tdelete(self.Rules, rule.RuleId)\n\n\treturn nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRulesRuleConfig(ctx context.Context, name string, policyId string, sequenceId int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/frinx-openconfig-network-instance:rule/{sequence-id}/frinx-openconfig-network-instance:config/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"sequence-id\"+\"}\", fmt.Sprintf(\"%v\", sequenceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRulesRuleActionConfig(ctx context.Context, name string, policyId string, sequenceId int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/frinx-openconfig-network-instance:rule/{sequence-id}/frinx-openconfig-network-instance:action/frinx-openconfig-network-instance:config/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"sequence-id\"+\"}\", fmt.Sprintf(\"%v\", sequenceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (s *LocalTests) deleteFwRule(c *gc.C, fwRuleId string) {\n\terr := s.testClient.DeleteFirewallRule(fwRuleId)\n\tc.Assert(err, gc.IsNil)\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRulesRuleTransportConfig(ctx context.Context, name string, policyId string, sequenceId int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/frinx-openconfig-network-instance:rule/{sequence-id}/frinx-openconfig-network-instance:transport/frinx-openconfig-network-instance:config/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"sequence-id\"+\"}\", fmt.Sprintf(\"%v\", sequenceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRulesRuleL2(ctx context.Context, name string, policyId string, sequenceId int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/frinx-openconfig-network-instance:rule/{sequence-id}/frinx-openconfig-network-instance:l2/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"sequence-id\"+\"}\", fmt.Sprintf(\"%v\", sequenceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwarding(ctx context.Context, name string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (se *StorageEndpoint) DeleteRule(ruleKey string) error {\n\treturn se.Remove(ruleKeyPath(ruleKey))\n}", "func DeleteRule(rule AuditRule) error {\n\tclient, err := libaudit.NewAuditClient(nil)\n\tdefer client.Close()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"Failed to initialize client\")\n\t}\n\tkr, _, _, err := rule.toKernelAuditRule()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"Failed to initialize client\")\n\t}\n\tif err := client.DeleteRule(kr.toWireFormat()); err != nil {\n\t\treturn errors.Wrap(err, \"Failed to delete audit rule\")\n\t}\n\treturn nil\n}", "func (f *FakeInstance) DeleteReverseIPv6(_ context.Context, _, _ string) error {\n\tpanic(\"implement me\")\n}", "func deleteRule(c *cli.Context) error {\n\n\truleid, err := hex.DecodeString(c.String(\"id\"))\n\tif err != nil {\n\t\treturn fmt.Errorf(\"[x] Incorrect ruleid format. \")\n\t}\n\n\terr = mapi.ExecuteMailRuleDelete(ruleid)\n\tif err == nil {\n\t\tfmt.Println(\"[*] Rule deleted. Fetching list of remaining rules...\")\n\t\trules, er := mapi.DisplayRules()\n\t\tif er != nil {\n\t\t\treturn er\n\t\t}\n\t\tfmt.Printf(\"[+] Found %d rules\\n\", len(rules))\n\t\tfor _, v := range rules {\n\t\t\tfmt.Printf(\"Rule: %s RuleID: %x\\n\", string(v.RuleName), v.RuleID)\n\t\t}\n\t\treturn nil\n\t}\n\treturn err\n}", "func (f *Forwarder) Delete(id userid.ID) {\n\tif _, loaded := f.transferAgents.LoadAndDelete(id); loaded {\n\t\tf.connections -= 1\n\t}\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicy(ctx context.Context, name string, policyId string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func NewPortForwardDelete(logger log.Logger, objectStore store.Store, portForwarder portforward.PortForwarder) *PortForwardDelete {\n\treturn &PortForwardDelete{\n\t\tlogger: logger,\n\t\tobjectStore: objectStore,\n\t\tportForwarder: portForwarder,\n\t}\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRulesRuleIpv6Config(ctx context.Context, name string, policyId string, sequenceId int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/frinx-openconfig-network-instance:rule/{sequence-id}/frinx-openconfig-network-instance:ipv6/frinx-openconfig-network-instance:config/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"sequence-id\"+\"}\", fmt.Sprintf(\"%v\", sequenceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (c *Client) DeleteRule(args *DeleteRuleArgs) error {\n\tjsonBytes, jsonErr := json.Marshal(args)\n\tif jsonErr != nil {\n\t\treturn jsonErr\n\t}\n\tbody, err := bce.NewBodyFromBytes(jsonBytes)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn DeleteRule(c, body)\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPolicies(ctx context.Context, name string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (f *fakeLB) CreateForwardingRule(_ context.Context, _ string, _ *govultr.ForwardingRule) (*govultr.ForwardingRule, *http.Response, error) {\n\tpanic(\"implement me\")\n}", "func (ruleset *DnsForwardingRuleset) deleteValidations() []func() (admission.Warnings, error) {\n\treturn nil\n}", "func Delete(c *golangsdk.ServiceClient, policyID, ruleID string) (r DeleteResult) {\n\treqOpt := &golangsdk.RequestOpts{\n\t\tMoreHeaders: RequestOpts.MoreHeaders,\n\t}\n\n\t_, r.Err = c.Delete(resourceURL(c, policyID, ruleID), reqOpt)\n\treturn\n}", "func (nsxtNat *NsxtNatRule) Delete() error {\n\tclient := nsxtNat.client\n\tendpoint := types.OpenApiPathVersion1_0_0 + types.OpenApiEndpointNsxtNatRules\n\tapiVersion, err := client.getOpenApiHighestElevatedVersion(endpoint)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif nsxtNat.NsxtNatRule.ID == \"\" {\n\t\treturn fmt.Errorf(\"cannot delete NSX-T NAT rule without ID\")\n\t}\n\n\turlRef, err := client.OpenApiBuildEndpoint(fmt.Sprintf(endpoint, nsxtNat.edgeGatewayId), nsxtNat.NsxtNatRule.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = client.OpenApiDeleteItem(apiVersion, urlRef, nil, nil)\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error deleting NSX-T NAT Rule: %s\", err)\n\t}\n\n\treturn nil\n}", "func DeleteACLRule(r *http.Request, target ACLTarget, actor ACLActor) (err error) {\n\n\t// get service\n\ts := service.Providers.MustService(r, \"ACLRule\")\n\n\t// delete the rules\n\tconds := service.NewConds()\n\tconds.Add(\"target\", string(target))\n\tconds.Add(\"actor\", string(actor))\n\ts.Delete(conds)\n\n\treturn\n}", "func (r *FirewallGlobalRulesStagedPolicyRulesResource) Delete(id string) error {\n\tif err := r.c.ModQuery(\"DELETE\", BasePath+FirewallGlobalRulesStagedPolicyRulesEndpoint+\"/\"+id, nil); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func TestFunnelDelete(t *testing.T) {\n\twithDB(func(db *DB) {\n\t\tdb.Do(func(tx *Tx) error {\n\t\t\t// Create account and funnel.\n\t\t\ta := &Account{}\n\t\t\tassert.NoError(t, tx.CreateAccount(a))\n\t\t\tf := &Funnel{Name: \"Funnel Y\", Steps: []*FunnelStep{{Condition: \"action == 'foo'\"}}}\n\t\t\tassert.NoError(t, a.CreateFunnel(f))\n\n\t\t\t// Delete the funnel.\n\t\t\tassert.NoError(t, f.Delete())\n\n\t\t\t// Retrieve the funnel again.\n\t\t\t_, err := tx.Funnel(1)\n\t\t\tassert.Equal(t, err, ErrFunnelNotFound)\n\t\t\treturn nil\n\t\t})\n\t})\n}", "func (o HttpRuleOutput) Delete() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v HttpRule) *string { return v.Delete }).(pulumi.StringPtrOutput)\n}", "func DeleteRule(id string) int {\n\tvar rule database.Rule\n\tdb.DB.Where(\"id = ? \", id).First(&rule)\n\tdb.DB.Where(\"id = ?\", id).Delete(&database.Rule{})\n\treturn rule.ReleaseID\n}", "func (hd *Datapath) DeleteRoute(rt *netproto.Route, vrf *netproto.Vrf) error {\n\t// This will ensure that only one datapath config will be active at a time. This is a temporary restriction\n\t// to ensure that HAL will use a single config thread , this will be removed prior to FCS to allow parallel configs to go through.\n\t// TODO Remove Global Locking\n\thd.Lock()\n\tdefer hd.Unlock()\n\tvrfKey := &halproto.VrfKeyHandle{\n\t\tKeyOrHandle: &halproto.VrfKeyHandle_VrfId{\n\t\t\tVrfId: vrf.Status.VrfID,\n\t\t},\n\t}\n\n\t// Build next hop key\n\tnextHopKey := &halproto.NexthopKeyHandle{\n\t\tKeyOrHandle: &halproto.NexthopKeyHandle_NexthopId{\n\t\t\tNexthopId: rt.Status.RouteID,\n\t\t},\n\t}\n\n\t// Build route key\n\tip, net, err := net.ParseCIDR(rt.Spec.IPPrefix)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error parsing the IP Prefix mask from %v. Err: %v\", rt.Spec.IPPrefix, err)\n\n\t}\n\tprefixLen, _ := net.Mask.Size()\n\tipPrefix := &halproto.IPPrefix{\n\t\tAddress: &halproto.IPAddress{\n\t\t\tIpAf: halproto.IPAddressFamily_IP_AF_INET,\n\t\t\tV4OrV6: &halproto.IPAddress_V4Addr{\n\t\t\t\tV4Addr: ipv4Touint32(ip),\n\t\t\t},\n\t\t},\n\t\tPrefixLen: uint32(prefixLen),\n\t}\n\n\trouteKey := &halproto.RouteKeyHandle{\n\t\tKeyOrHandle: &halproto.RouteKeyHandle_RouteKey{\n\t\t\tRouteKey: &halproto.RouteKey{\n\t\t\t\tVrfKeyHandle: vrfKey,\n\t\t\t\tIpPrefix: ipPrefix,\n\t\t\t},\n\t\t},\n\t}\n\n\tnhDelReq := &halproto.NexthopDeleteRequestMsg{\n\t\tRequest: []*halproto.NexthopDeleteRequest{\n\t\t\t{\n\t\t\t\tKeyOrHandle: nextHopKey,\n\t\t\t},\n\t\t},\n\t}\n\n\trtDelReq := &halproto.RouteDeleteRequestMsg{\n\t\tRequest: []*halproto.RouteDeleteRequest{\n\t\t\t{\n\t\t\t\tKeyOrHandle: routeKey,\n\t\t\t},\n\t\t},\n\t}\n\n\t// delete hal objects\n\tif hd.Kind == \"hal\" {\n\t\t// delete route\n\t\trtResp, err := hd.Hal.Netclient.RouteDelete(context.Background(), rtDelReq)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error deleting route. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\t\tif rtResp.Response[0].ApiStatus != halproto.ApiStatus_API_STATUS_OK {\n\t\t\tlog.Errorf(\"HAL returned non OK status. %v\", rtResp.Response[0].ApiStatus.String())\n\t\t\treturn fmt.Errorf(\"HAL returned non OK status. %v\", rtResp.Response[0].ApiStatus.String())\n\t\t}\n\n\t\t// delete next hop\n\t\tnhResp, err := hd.Hal.Netclient.NexthopDelete(context.Background(), nhDelReq)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error deleting next hop. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\t\tif nhResp.Response[0].ApiStatus != halproto.ApiStatus_API_STATUS_OK {\n\t\t\tlog.Errorf(\"HAL returned non OK status. %v\", nhResp.Response[0].ApiStatus.String())\n\t\t\treturn fmt.Errorf(\"HAL returned non OK status. %v\", nhResp.Response[0].ApiStatus.String())\n\t\t}\n\t} else {\n\t\t_, err := hd.Hal.Netclient.RouteDelete(context.Background(), rtDelReq)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error creating route. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRulesRuleIpv4Config(ctx context.Context, name string, policyId string, sequenceId int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/frinx-openconfig-network-instance:rule/{sequence-id}/frinx-openconfig-network-instance:ipv4/frinx-openconfig-network-instance:config/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"sequence-id\"+\"}\", fmt.Sprintf(\"%v\", sequenceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (p *DeleteService) ForwardDelete(request string, reply *string) error {\n\t_, index := Getip()\n\tmemid := election()\n\tif _, ok := fileList[request]; ok {\n\t\tnodelist := fileList[request].NodeList\n\t\tfor _, i := range nodelist {\n\t\t\tif i == index {\n\t\t\t\tos.Remove(request)\n\t\t\t} else {\n\t\t\t\tclient, err := rpc.Dial(\"tcp\", straddr[i]+\":4000\")\n\t\t\t\tif (err != nil) {\n\t\t\t\t\tfmt.Println(err)\n\t\t\t\t}\n\t\t\t\terr = client.Call(\"DeleteService.Delete\", request, &reply)\n\t\t\t\tif err != nil {\n\t\t\t\t\tfmt.Println(err)\n\t\t\t\t}\n\t\t\t\t*reply = \"file delete!\"\n\t\t\t\tdefer client.Close()\n\t\t\t}\n\t\t}\n\t\tfor _, index1 := range memid {\n\t\t\tif index1 == index {\n\t\t\t\tdelete(fileList, request)\n\t\t\t} else {\n\t\t\t\tvar temp = make(map[string]pair)\n\t\t\t\ttemp[request] = pair{\n\t\t\t\t\tTimeStamp: 0,\n\t\t\t\t\tNodeList: nil,\n\t\t\t\t}\n\t\t\t\tsendlist(temp, straddr[index1])\n\t\t\t}\n\t\t}\n\t} else {\n\t\t*reply = \"no such file!\"\n\t}\n\treturn nil\n}", "func HandleDeleteEventingTriggerRule(adminMan *admin.Manager, syncMan *syncman.Manager) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\n\t\t// Get the JWT token from header\n\t\ttoken := utils.GetTokenFromHeader(r)\n\n\t\tvars := mux.Vars(r)\n\t\truleName := vars[\"id\"]\n\t\tprojectID := vars[\"project\"]\n\n\t\tdefer utils.CloseTheCloser(r.Body)\n\n\t\tctx, cancel := context.WithTimeout(r.Context(), time.Duration(utils.DefaultContextTime)*time.Second)\n\t\tdefer cancel()\n\n\t\t// Check if the request is authorised\n\t\treqParams, err := adminMan.IsTokenValid(ctx, token, \"eventing-trigger\", \"modify\", map[string]string{\"project\": projectID, \"id\": ruleName})\n\t\tif err != nil {\n\t\t\t_ = helpers.Response.SendErrorResponse(ctx, w, http.StatusUnauthorized, err)\n\t\t\treturn\n\t\t}\n\n\t\treqParams = utils.ExtractRequestParams(r, reqParams, nil)\n\t\tstatus, err := syncMan.SetDeleteEventingRule(ctx, projectID, ruleName, reqParams)\n\t\tif err != nil {\n\t\t\t_ = helpers.Response.SendErrorResponse(ctx, w, status, err)\n\t\t\treturn\n\t\t}\n\n\t\t_ = helpers.Response.SendOkayResponse(ctx, status, w)\n\t}\n}", "func (c *TestClient) GetForwardingRule(project, region, name string) (*compute.ForwardingRule, error) {\n\tif c.GetForwardingRuleFn != nil {\n\t\treturn c.GetForwardingRuleFn(project, region, name)\n\t}\n\treturn c.client.GetForwardingRule(project, region, name)\n}", "func (gateway *EgressGateway) DeleteStaticIptablesRule(setName string, destinationIP, egressIP string) error {\n\n\t// delete rule in NAT postrouting to SNAT traffic\n\truleSpec := []string{\"-m\", \"set\", \"--match-set\", setName, \"src\", \"-d\", destinationIP, \"-j\", \"SNAT\", \"--to-source\", egressIP}\n\tif err := gateway.deleteRule(defaultNATIptable, egressGatewayNATChainName, ruleSpec...); err != nil {\n\t\treturn fmt.Errorf(\"failed to delete rule in chain %v err %v\", egressGatewayNATChainName, err)\n\t}\n\n\t// delete rule in FORWARD chain of filter table\n\truleSpec = []string{\"-m\", \"set\", \"--set\", setName, \"src\", \"-d\", destinationIP, \"-j\", \"ACCEPT\"}\n\thasRule, err := gateway.ipt.Exists(\"filter\", egressGatewayFWChainName, ruleSpec...)\n\tif err != nil {\n\t\treturn errors.New(\"Failed to verify rule exists in \" + egressGatewayFWChainName + \" chain of filter table\" + err.Error())\n\t}\n\tif hasRule {\n\t\terr = gateway.ipt.Delete(\"filter\", egressGatewayFWChainName, ruleSpec...)\n\t\tif err != nil {\n\t\t\treturn errors.New(\"Failed to delete iptables command to ACCEPT traffic from director nodes to get forwarded\" + err.Error())\n\t\t}\n\t}\n\n\tset, err := ipset.New(setName, \"hash:ip\", &ipset.Params{})\n\tif err != nil {\n\t\treturn errors.New(\"Failed to get ipset with name \" + setName + \" due to %\" + err.Error())\n\t}\n\terr = set.Destroy()\n\tif err != nil {\n\t\treturn errors.New(\"Failed to delete ipset due to \" + err.Error())\n\t}\n\n\treturn nil\n}", "func (a *SyncApiService) DeleteSyncRule(ctx context.Context, syncRuleId string) ( *http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/platform/3/sync/rules/{SyncRuleId}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"SyncRuleId\"+\"}\", fmt.Sprintf(\"%v\", syncRuleId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{ \"application/json\", }\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\n\t\t\"application/json\",\n\t\t}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\tdefer localVarHttpResponse.Body.Close()\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tbodyBytes, _ := ioutil.ReadAll(localVarHttpResponse.Body)\n\t\treturn localVarHttpResponse, reportError(\"Status: %v, Body: %s\", localVarHttpResponse.Status, bodyBytes)\n\t}\n\n\treturn localVarHttpResponse, err\n}", "func (fw *AzureSqlFirewallRuleManager) Delete(ctx context.Context, obj runtime.Object, opts ...resourcemanager.ConfigOption) (bool, error) {\n\tinstance, err := fw.convert(obj)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\tsubscriptionID := instance.Spec.SubscriptionID\n\tgroupName := instance.Spec.ResourceGroup\n\tserver := instance.Spec.Server\n\truleName := instance.ObjectMeta.Name\n\n\terr = fw.DeleteSQLFirewallRule(ctx, subscriptionID, groupName, server, ruleName)\n\tif err != nil {\n\t\tcatch := []string{\n\t\t\terrhelp.AsyncOpIncompleteError,\n\t\t}\n\t\tgone := []string{\n\t\t\terrhelp.ResourceGroupNotFoundErrorCode,\n\t\t\terrhelp.ParentNotFoundErrorCode,\n\t\t\terrhelp.NotFoundErrorCode,\n\t\t\terrhelp.ResourceNotFound,\n\t\t}\n\t\tazerr := errhelp.NewAzureError(err)\n\t\tif helpers.ContainsString(catch, azerr.Type) {\n\t\t\treturn true, nil\n\t\t} else if helpers.ContainsString(gone, azerr.Type) {\n\t\t\treturn false, nil\n\t\t}\n\t\tinstance.Status.Message = fmt.Sprintf(\"AzureSqlFirewallRule Delete failed with %s\", err.Error())\n\t\treturn true, err\n\t}\n\tinstance.Status.Message = fmt.Sprintf(\"Delete AzureSqlFirewallRule succeeded\")\n\treturn false, nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRulesRuleL2Config(ctx context.Context, name string, policyId string, sequenceId int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/frinx-openconfig-network-instance:rule/{sequence-id}/frinx-openconfig-network-instance:l2/frinx-openconfig-network-instance:config/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"sequence-id\"+\"}\", fmt.Sprintf(\"%v\", sequenceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (c *AuditClient) DeleteRule(rule []byte) error {\n\tmsg := syscall.NetlinkMessage{\n\t\tHeader: syscall.NlMsghdr{\n\t\t\tType: uint16(auparse.AUDIT_DEL_RULE),\n\t\t\tFlags: syscall.NLM_F_REQUEST | syscall.NLM_F_ACK,\n\t\t},\n\t\tData: rule,\n\t}\n\n\t// Send AUDIT_DEL_RULE message to the kernel.\n\tseq, err := c.Netlink.Send(msg)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed sending delete rule request: %w\", err)\n\t}\n\n\t_, err = c.getReply(seq)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to get ACK to rule delete request: %w\", err)\n\t}\n\n\treturn nil\n}", "func (s *socatManager) RemovePortForward(namespace string, host string, dest int) error {\n\ts.rm.Lock()\n\tdefer s.rm.Unlock()\n\tsrc, err := getSource(host)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\trule, ok := s.rules[src.port]\n\tif !ok {\n\t\treturn fmt.Errorf(\"no port forward from host port: %d\", src.port)\n\t}\n\n\tif rule.ns != forwardID(namespace, src.port, dest) {\n\t\treturn fmt.Errorf(\"permission denied\")\n\t}\n\n\tvar rs []nft.Rule\n\tfor _, r := range rule.Rules() {\n\t\trs = append(rs, nft.Rule{Body: r})\n\t}\n\n\tset := nft.Nft{\n\t\t\"nat\": nft.Table{\n\t\t\tFamily: nft.FamilyIP,\n\t\t\tChains: nft.Chains{\n\t\t\t\t\"pre\": nft.Chain{\n\t\t\t\t\tRules: rs,\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\n\tif err := s.nft().DropRules(set); err != nil {\n\t\treturn err\n\t}\n\n\tdelete(s.rules, src.port)\n\treturn nil\n}", "func (client FirewallPolicyRuleGroupsClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) {\n\terr = autorest.Respond(\n\t\tresp,\n\t\tazure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent),\n\t\tautorest.ByClosing())\n\tresult.Response = resp\n\treturn\n}", "func (fw *IPtables) deleteIPtablesRule(rule *IPtablesRule) error {\n\tif err := fw.Store.switchIPtablesRule(rule, setRuleInactive); err != nil {\n\t\tlog.Error(\"In deleteIPtablesRule() failed to deactivate the rule\", rule.GetBody())\n\t\treturn err\n\t}\n\n\tif err1 := fw.EnsureRule(rule, EnsureAbsent); err1 != nil {\n\t\tlog.Errorf(\"In deleteIPtablesRule() rule %s set inactive but failed to uninstall\", rule.GetBody())\n\t\treturn err1\n\t}\n\n\tif err2 := fw.Store.deleteIPtablesRule(rule); err2 != nil {\n\t\tlog.Errorf(\"In deleteIPtablesRule() rule %s set inactive and uninstalled but failed to delete DB record\", rule.GetBody())\n\t\treturn err2\n\t}\n\treturn nil\n}", "func (_m *ISession) AutoModerationRuleDelete(guildID string, ruleID string, options ...discordgo.RequestOption) error {\n\t_va := make([]interface{}, len(options))\n\tfor _i := range options {\n\t\t_va[_i] = options[_i]\n\t}\n\tvar _ca []interface{}\n\t_ca = append(_ca, guildID, ruleID)\n\t_ca = append(_ca, _va...)\n\tret := _m.Called(_ca...)\n\n\tvar r0 error\n\tif rf, ok := ret.Get(0).(func(string, string, ...discordgo.RequestOption) error); ok {\n\t\tr0 = rf(guildID, ruleID, options...)\n\t} else {\n\t\tr0 = ret.Error(0)\n\t}\n\n\treturn r0\n}", "func (m *MockFirewallServiceIface) DeleteIpv6FirewallRule(p *DeleteIpv6FirewallRuleParams) (*DeleteIpv6FirewallRuleResponse, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"DeleteIpv6FirewallRule\", p)\n\tret0, _ := ret[0].(*DeleteIpv6FirewallRuleResponse)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func NeighDel(ifIndex int, ip net.IP) error {\n\tvar family = netlink.FAMILY_V4\n\n\tif IsIPv4(ip) {\n\t\tfamily = netlink.FAMILY_V4\n\t}\n\n\tif IsIPv6(ip) {\n\t\tfamily = netlink.FAMILY_V6\n\t}\n\n\tif err := netlink.NeighDel(\n\t\t&netlink.Neigh{\n\t\t\tLinkIndex: ifIndex,\n\t\t\tFamily: family,\n\t\t\tState: netlink.NUD_NONE,\n\t\t\tType: netlink.NDA_UNSPEC,\n\t\t\tFlags: netlink.NTF_PROXY,\n\t\t\tIP: ip,\n\t\t},\n\t); err != nil {\n\t\treturn fmt.Errorf(\"netlink action error, ifIndex=%d, ip=%s: %w\", ifIndex, ip, err)\n\t}\n\n\treturn nil\n}", "func (m *MockInterface) DeleteRule(arg0 iptables.Protocol, arg1, arg2 string, arg3 []string) error {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"DeleteRule\", arg0, arg1, arg2, arg3)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func (a *BackendOptionsApiService) DeleteHTTPRequestRule(ctx _context.Context, id int32, parentName string, parentType string, localVarOptionals *DeleteHTTPRequestRuleOpts) (*_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/services/haproxy/configuration/http_request_rules/{id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"id\"+\"}\", _neturl.QueryEscape(fmt.Sprintf(\"%v\", id)), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tlocalVarQueryParams.Add(\"parent_name\", parameterToString(parentName, \"\"))\n\tlocalVarQueryParams.Add(\"parent_type\", parameterToString(parentType, \"\"))\n\tif localVarOptionals != nil && localVarOptionals.TransactionId.IsSet() {\n\t\tlocalVarQueryParams.Add(\"transaction_id\", parameterToString(localVarOptionals.TransactionId.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Version.IsSet() {\n\t\tlocalVarQueryParams.Add(\"version\", parameterToString(localVarOptionals.Version.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.ForceReload.IsSet() {\n\t\tlocalVarQueryParams.Add(\"force_reload\", parameterToString(localVarOptionals.ForceReload.Value(), \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tvar v ModelError\n\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\tif err != nil {\n\t\t\tnewErr.error = err.Error()\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tnewErr.model = v\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func (api *API) TeamsDeleteRule(ctx context.Context, accountID string, ruleId string) error {\n\turi := fmt.Sprintf(\"/accounts/%s/gateway/rules/%s\", accountID, ruleId)\n\n\t_, err := api.makeRequestContext(ctx, http.MethodDelete, uri, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (g *Google) getForwardingRule() (*compute.ForwardingRule, error) {\n\treturn g.computeService.GlobalForwardingRules.Get(g.project, forwardingRuleName).Do()\n}", "func (a *BackendOptionsApiService) DeleteServerSwitchingRule(ctx _context.Context, id int32, backend string, localVarOptionals *DeleteServerSwitchingRuleOpts) (*_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/services/haproxy/configuration/server_switching_rules/{id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"id\"+\"}\", _neturl.QueryEscape(fmt.Sprintf(\"%v\", id)), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tlocalVarQueryParams.Add(\"backend\", parameterToString(backend, \"\"))\n\tif localVarOptionals != nil && localVarOptionals.TransactionId.IsSet() {\n\t\tlocalVarQueryParams.Add(\"transaction_id\", parameterToString(localVarOptionals.TransactionId.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Version.IsSet() {\n\t\tlocalVarQueryParams.Add(\"version\", parameterToString(localVarOptionals.Version.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.ForceReload.IsSet() {\n\t\tlocalVarQueryParams.Add(\"force_reload\", parameterToString(localVarOptionals.ForceReload.Value(), \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tvar v ModelError\n\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\tif err != nil {\n\t\t\tnewErr.error = err.Error()\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tnewErr.model = v\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func (a *EventProcessingMatchingApiService) DeleteUsingDELETE18(ctx _context.Context, matchingRuleId string) ApiDeleteUsingDELETE18Request {\n\treturn ApiDeleteUsingDELETE18Request{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\tmatchingRuleId: matchingRuleId,\n\t}\n}", "func deleteOp(ctx context.Context, c *messaging.Client, fb wasabee.FirebaseCmd) error {\n\tdata := map[string]string{\n\t\t\"gid\": string(fb.Gid),\n\t\t\"msg\": fb.Msg,\n\t\t\"cmd\": fb.Cmd.String(),\n\t}\n\n\ttokens, err := fb.Gid.FirebaseTokens()\n\tif err != nil {\n\t\twasabee.Log.Error(err)\n\t\treturn err\n\t}\n\tgenericMulticast(ctx, c, data, tokens)\n\treturn nil\n}", "func forwardDeleteAccount(ctx context.Context, mux *runtime.ServeMux, marshaler runtime.Marshaler, w http.ResponseWriter, req *http.Request, resp proto.Message, opts ...func(context.Context, http.ResponseWriter, proto.Message) error) {\n\tw.WriteHeader(http.StatusNoContent)\n\truntime.ForwardResponseMessage(ctx, mux, marshaler, w, req, resp, opts...)\n}", "func (a *BackendOptionsApiService) DeleteStickRule(ctx _context.Context, id int32, backend string, localVarOptionals *DeleteStickRuleOpts) (*_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/services/haproxy/configuration/stick_rules/{id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"id\"+\"}\", _neturl.QueryEscape(fmt.Sprintf(\"%v\", id)), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tlocalVarQueryParams.Add(\"backend\", parameterToString(backend, \"\"))\n\tif localVarOptionals != nil && localVarOptionals.TransactionId.IsSet() {\n\t\tlocalVarQueryParams.Add(\"transaction_id\", parameterToString(localVarOptionals.TransactionId.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Version.IsSet() {\n\t\tlocalVarQueryParams.Add(\"version\", parameterToString(localVarOptionals.Version.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.ForceReload.IsSet() {\n\t\tlocalVarQueryParams.Add(\"force_reload\", parameterToString(localVarOptionals.ForceReload.Value(), \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tvar v ModelError\n\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\tif err != nil {\n\t\t\tnewErr.error = err.Error()\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tnewErr.model = v\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func (r *RouterGroup) DELETE(path string, fn Controller) {\n\tr.gin.DELETE(path, adapt(fn))\n}", "func (client DataFlowClient) DeleteDataFlowResponder(resp *http.Response) (result autorest.Response, err error) {\n\terr = autorest.Respond(\n\t\tresp,\n\t\tazure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent),\n\t\tautorest.ByClosing())\n\tresult.Response = resp\n\treturn\n}", "func (m *MockFirewallServiceIface) NewDeletePortForwardingRuleParams(id string) *DeletePortForwardingRuleParams {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"NewDeletePortForwardingRuleParams\", id)\n\tret0, _ := ret[0].(*DeletePortForwardingRuleParams)\n\treturn ret0\n}", "func (r *Router) DeleteFunc(pattern string, fn http.HandlerFunc) Route {\n\treturn r.Delete(pattern, http.HandlerFunc(fn))\n}", "func (m *ForwardingProfilesItemPoliciesPolicyLinkItemRequestBuilder) Delete(ctx context.Context, requestConfiguration *ForwardingProfilesItemPoliciesPolicyLinkItemRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func (hd *Datapath) DeleteTCPProxyPolicy(tcp *netproto.TCPProxyPolicy, vrf *netproto.Vrf) error {\n\t// This will ensure that only one datapath config will be active at a time. This is a temporary restriction\n\t// to ensure that HAL will use a single config thread , this will be removed prior to FCS to allow parallel configs to go through.\n\t// TODO Remove Global Locking\n\thd.Lock()\n\tdefer hd.Unlock()\n\tvrfKey := &halproto.VrfKeyHandle{\n\t\tKeyOrHandle: &halproto.VrfKeyHandle_VrfId{\n\t\t\tVrfId: vrf.Status.VrfID,\n\t\t},\n\t}\n\n\ttcpProxyPolicyDelReq := &halproto.TcpProxyRuleDeleteRequestMsg{\n\t\tRequest: []*halproto.TcpProxyRuleDeleteRequest{\n\t\t\t{\n\t\t\t\tKeyOrHandle: &halproto.TcpProxyRuleKeyHandle{\n\t\t\t\t\tKeyOrHandle: &halproto.TcpProxyRuleKeyHandle_RuleKey{\n\t\t\t\t\t\tRuleKey: &halproto.TcpProxyRuleKey{\n\t\t\t\t\t\t\tTcpProxyRuleId: tcp.Status.TCPProxyPolicyID,\n\t\t\t\t\t\t\tVrfKeyOrHandle: vrfKey,\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\tif hd.Kind == \"hal\" {\n\t\tresp, err := hd.Hal.TCPProxyPolicyClient.TcpProxyRuleDelete(context.Background(), tcpProxyPolicyDelReq)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error deleting TCPProxy Policy. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\t\tif resp.Response[0].ApiStatus != halproto.ApiStatus_API_STATUS_OK {\n\t\t\tlog.Errorf(\"HAL returned non OK status. %v\", resp.Response[0].ApiStatus.String())\n\t\t\treturn fmt.Errorf(\"HAL returned non OK status. %v\", resp.Response[0].ApiStatus.String())\n\t\t}\n\t} else {\n\t\t_, err := hd.Hal.TCPProxyPolicyClient.TcpProxyRuleDelete(context.Background(), tcpProxyPolicyDelReq)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error deleting TCPProxy Policy. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (a *BackendOptionsApiService) DeleteHTTPResponseRule(ctx _context.Context, id int32, parentName string, parentType string, localVarOptionals *DeleteHTTPResponseRuleOpts) (*_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/services/haproxy/configuration/http_response_rules/{id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"id\"+\"}\", _neturl.QueryEscape(fmt.Sprintf(\"%v\", id)), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tlocalVarQueryParams.Add(\"parent_name\", parameterToString(parentName, \"\"))\n\tlocalVarQueryParams.Add(\"parent_type\", parameterToString(parentType, \"\"))\n\tif localVarOptionals != nil && localVarOptionals.TransactionId.IsSet() {\n\t\tlocalVarQueryParams.Add(\"transaction_id\", parameterToString(localVarOptionals.TransactionId.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Version.IsSet() {\n\t\tlocalVarQueryParams.Add(\"version\", parameterToString(localVarOptionals.Version.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.ForceReload.IsSet() {\n\t\tlocalVarQueryParams.Add(\"force_reload\", parameterToString(localVarOptionals.ForceReload.Value(), \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tvar v ModelError\n\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\tif err != nil {\n\t\t\tnewErr.error = err.Error()\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tnewErr.model = v\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func (r *FirewallManagementIPRulesResource) Delete(id string) error {\n\tif err := r.c.ModQuery(\"DELETE\", BasePath+FirewallManagementIPRulesEndpoint+\"/\"+id, nil); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func DeleteNetworkSecurityGroupRule() {}", "func (g *Google) createForwardingRule(targetLink string) (string, error) {\n\tif rule, err := g.getForwardingRule(); err == nil {\n\t\tlog.Infof(\"found ForwardingRule %s: %s\", forwardingRuleName, rule.SelfLink)\n\t\treturn rule.SelfLink, nil\n\t}\n\n\top, err := g.computeService.GlobalForwardingRules.Insert(g.project,\n\t\t&compute.ForwardingRule{\n\t\t\tName: forwardingRuleName,\n\t\t\tIPProtocol: cockroachProtocol,\n\t\t\tPortRange: fmt.Sprintf(\"%d\", g.context.Port),\n\t\t\tTarget: targetLink,\n\t\t}).Do()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif err := g.waitForOperation(op); err != nil {\n\t\treturn \"\", err\n\t}\n\n\tlog.Infof(\"created ForwardingRule %s: %s\", forwardingRuleName, op.TargetLink)\n\treturn op.TargetLink, nil\n}", "func (sdk *MockGoSDKClient) DeleteSQLFirewallRule(ctx context.Context, resourceGroupName string, serverName string, ruleName string) (err error) {\n\treturn nil\n}", "func (m *Manager) DeleteRule(n string) bool {\n\tm.mu.Lock()\n\tdefer m.mu.Unlock()\n\tif _, ok := m.rules[n]; !ok {\n\t\treturn false\n\t}\n\tdelete(m.rules, n)\n\treturn true\n}", "func DeleteResource(r rest.GracefulDeleter, allowsOptions bool, scope *RequestScope, admit admission.Interface) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, req *http.Request) {\n\t\tctx := req.Context()\n\t\t// For performance tracking purposes.\n\t\tctx, span := tracing.Start(ctx, \"Delete\", traceFields(req)...)\n\t\tdefer span.End(500 * time.Millisecond)\n\n\t\tnamespace, name, err := scope.Namer.Name(req)\n\t\tif err != nil {\n\t\t\tscope.err(err, w, req)\n\t\t\treturn\n\t\t}\n\n\t\t// enforce a timeout of at most requestTimeoutUpperBound (34s) or less if the user-provided\n\t\t// timeout inside the parent context is lower than requestTimeoutUpperBound.\n\t\tctx, cancel := context.WithTimeout(ctx, requestTimeoutUpperBound)\n\t\tdefer cancel()\n\n\t\tctx = request.WithNamespace(ctx, namespace)\n\t\tadmit = admission.WithAudit(admit)\n\n\t\toutputMediaType, _, err := negotiation.NegotiateOutputMediaType(req, scope.Serializer, scope)\n\t\tif err != nil {\n\t\t\tscope.err(err, w, req)\n\t\t\treturn\n\t\t}\n\n\t\toptions := &metav1.DeleteOptions{}\n\t\tif allowsOptions {\n\t\t\tbody, err := limitedReadBodyWithRecordMetric(ctx, req, scope.MaxRequestBodyBytes, scope.Resource.GroupResource().String(), requestmetrics.Delete)\n\t\t\tif err != nil {\n\t\t\t\tspan.AddEvent(\"limitedReadBody failed\", attribute.Int(\"len\", len(body)), attribute.String(\"err\", err.Error()))\n\t\t\t\tscope.err(err, w, req)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tspan.AddEvent(\"limitedReadBody succeeded\", attribute.Int(\"len\", len(body)))\n\t\t\tif len(body) > 0 {\n\t\t\t\ts, err := negotiation.NegotiateInputSerializer(req, false, metainternalversionscheme.Codecs)\n\t\t\t\tif err != nil {\n\t\t\t\t\tscope.err(err, w, req)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\t// For backwards compatibility, we need to allow existing clients to submit per group DeleteOptions\n\t\t\t\t// It is also allowed to pass a body with meta.k8s.io/v1.DeleteOptions\n\t\t\t\tdefaultGVK := scope.MetaGroupVersion.WithKind(\"DeleteOptions\")\n\t\t\t\tobj, gvk, err := metainternalversionscheme.Codecs.DecoderToVersion(s.Serializer, defaultGVK.GroupVersion()).Decode(body, &defaultGVK, options)\n\t\t\t\tif err != nil {\n\t\t\t\t\tscope.err(err, w, req)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tif obj != options {\n\t\t\t\t\tscope.err(fmt.Errorf(\"decoded object cannot be converted to DeleteOptions\"), w, req)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tspan.AddEvent(\"Decoded delete options\")\n\n\t\t\t\tobjGV := gvk.GroupVersion()\n\t\t\t\taudit.LogRequestObject(req.Context(), obj, objGV, scope.Resource, scope.Subresource, metainternalversionscheme.Codecs)\n\t\t\t\tspan.AddEvent(\"Recorded the audit event\")\n\t\t\t} else {\n\t\t\t\tif err := metainternalversionscheme.ParameterCodec.DecodeParameters(req.URL.Query(), scope.MetaGroupVersion, options); err != nil {\n\t\t\t\t\terr = errors.NewBadRequest(err.Error())\n\t\t\t\t\tscope.err(err, w, req)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tif errs := validation.ValidateDeleteOptions(options); len(errs) > 0 {\n\t\t\terr := errors.NewInvalid(schema.GroupKind{Group: metav1.GroupName, Kind: \"DeleteOptions\"}, \"\", errs)\n\t\t\tscope.err(err, w, req)\n\t\t\treturn\n\t\t}\n\t\toptions.TypeMeta.SetGroupVersionKind(metav1.SchemeGroupVersion.WithKind(\"DeleteOptions\"))\n\n\t\tspan.AddEvent(\"About to delete object from database\")\n\t\twasDeleted := true\n\t\tuserInfo, _ := request.UserFrom(ctx)\n\t\tstaticAdmissionAttrs := admission.NewAttributesRecord(nil, nil, scope.Kind, namespace, name, scope.Resource, scope.Subresource, admission.Delete, options, dryrun.IsDryRun(options.DryRun), userInfo)\n\t\tresult, err := finisher.FinishRequest(ctx, func() (runtime.Object, error) {\n\t\t\tobj, deleted, err := r.Delete(ctx, name, rest.AdmissionToValidateObjectDeleteFunc(admit, staticAdmissionAttrs, scope), options)\n\t\t\twasDeleted = deleted\n\t\t\treturn obj, err\n\t\t})\n\t\tif err != nil {\n\t\t\tscope.err(err, w, req)\n\t\t\treturn\n\t\t}\n\t\tspan.AddEvent(\"Object deleted from database\")\n\n\t\tstatus := http.StatusOK\n\t\t// Return http.StatusAccepted if the resource was not deleted immediately and\n\t\t// user requested cascading deletion by setting OrphanDependents=false.\n\t\t// Note: We want to do this always if resource was not deleted immediately, but\n\t\t// that will break existing clients.\n\t\t// Other cases where resource is not instantly deleted are: namespace deletion\n\t\t// and pod graceful deletion.\n\t\t//nolint:staticcheck // SA1019 backwards compatibility\n\t\t//nolint: staticcheck\n\t\tif !wasDeleted && options.OrphanDependents != nil && !*options.OrphanDependents {\n\t\t\tstatus = http.StatusAccepted\n\t\t}\n\t\t// if the rest.Deleter returns a nil object, fill out a status. Callers may return a valid\n\t\t// object with the response.\n\t\tif result == nil {\n\t\t\tresult = &metav1.Status{\n\t\t\t\tStatus: metav1.StatusSuccess,\n\t\t\t\tCode: int32(status),\n\t\t\t\tDetails: &metav1.StatusDetails{\n\t\t\t\t\tName: name,\n\t\t\t\t\tKind: scope.Kind.Kind,\n\t\t\t\t},\n\t\t\t}\n\t\t}\n\n\t\tspan.AddEvent(\"About to write a response\")\n\t\tdefer span.AddEvent(\"Writing http response done\")\n\t\ttransformResponseObject(ctx, scope, req, w, status, outputMediaType, result)\n\t}\n}", "func (firewall *NsxtFirewall) DeleteRuleById(id string) error {\n\tif id == \"\" {\n\t\treturn fmt.Errorf(\"empty ID specified\")\n\t}\n\n\tendpoint := types.OpenApiPathVersion1_0_0 + types.OpenApiEndpointNsxtFirewallRules\n\tminimumApiVersion, err := firewall.client.checkOpenApiEndpointCompatibility(endpoint)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\turlRef, err := firewall.client.OpenApiBuildEndpoint(fmt.Sprintf(endpoint, firewall.edgeGatewayId), \"/\", id)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = firewall.client.OpenApiDeleteItem(minimumApiVersion, urlRef, nil, nil)\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error deleting NSX-T Firewall Rule with ID '%s': %s\", id, err)\n\t}\n\n\treturn nil\n}", "func (tr *Transport) Delete(url string, fn HandlerFunc, options ...HandlerOption) {\n\ttr.mux.Handler(net_http.MethodDelete, url, encapsulate(fn, tr.options, options))\n}", "func (r *DeviceComplianceScheduledActionForRuleRequest) Delete(ctx context.Context) error {\n\treturn r.JSONRequest(ctx, \"DELETE\", \"\", nil, nil)\n}", "func (c *TestClient) DeleteFirewallRule(project, name string) error {\n\tif c.DeleteFirewallRuleFn != nil {\n\t\treturn c.DeleteFirewallRuleFn(project, name)\n\t}\n\treturn c.client.DeleteFirewallRule(project, name)\n}", "func (a *BackendOptionsApiService) DeleteTCPRequestRule(ctx _context.Context, id int32, parentName string, parentType string, localVarOptionals *DeleteTCPRequestRuleOpts) (*_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/services/haproxy/configuration/tcp_request_rules/{id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"id\"+\"}\", _neturl.QueryEscape(fmt.Sprintf(\"%v\", id)), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tlocalVarQueryParams.Add(\"parent_name\", parameterToString(parentName, \"\"))\n\tlocalVarQueryParams.Add(\"parent_type\", parameterToString(parentType, \"\"))\n\tif localVarOptionals != nil && localVarOptionals.TransactionId.IsSet() {\n\t\tlocalVarQueryParams.Add(\"transaction_id\", parameterToString(localVarOptionals.TransactionId.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Version.IsSet() {\n\t\tlocalVarQueryParams.Add(\"version\", parameterToString(localVarOptionals.Version.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.ForceReload.IsSet() {\n\t\tlocalVarQueryParams.Add(\"force_reload\", parameterToString(localVarOptionals.ForceReload.Value(), \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tvar v ModelError\n\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\tif err != nil {\n\t\t\tnewErr.error = err.Error()\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tnewErr.model = v\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func (tr *Transport) DELETE(\n\turl string,\n\tfn Handler,\n\toptions ...HandlerOption,\n) {\n\ttr.mux.Handler(\n\t\tnet_http.MethodDelete,\n\t\turl,\n\t\tnewHandler(fn, append(tr.options, options...)...),\n\t)\n}", "func (c *client) DeleteHTTPResponseRule(id int64, parentType string, parentName string, transactionID string, version int64) error {\n\tp, t, err := c.loadDataForChange(transactionID, version)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar section parser.Section\n\tif parentType == \"backend\" {\n\t\tsection = parser.Backends\n\t} else if parentType == \"frontend\" {\n\t\tsection = parser.Frontends\n\t}\n\n\tif err := p.Delete(section, parentName, \"http-response\", int(id)); err != nil {\n\t\treturn c.HandleError(strconv.FormatInt(id, 10), parentType, parentName, t, transactionID == \"\", err)\n\t}\n\n\treturn c.SaveData(p, t, transactionID == \"\")\n}", "func (client FirewallPolicyRuleGroupsClient) Delete(ctx context.Context, resourceGroupName string, firewallPolicyName string, ruleGroupName string) (result FirewallPolicyRuleGroupsDeleteFuture, err error) {\n\tif tracing.IsEnabled() {\n\t\tctx = tracing.StartSpan(ctx, fqdn+\"/FirewallPolicyRuleGroupsClient.Delete\")\n\t\tdefer func() {\n\t\t\tsc := -1\n\t\t\tif result.FutureAPI != nil && result.FutureAPI.Response() != nil {\n\t\t\t\tsc = result.FutureAPI.Response().StatusCode\n\t\t\t}\n\t\t\ttracing.EndSpan(ctx, sc, err)\n\t\t}()\n\t}\n\treq, err := client.DeletePreparer(ctx, resourceGroupName, firewallPolicyName, ruleGroupName)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"network.FirewallPolicyRuleGroupsClient\", \"Delete\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresult, err = client.DeleteSender(req)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"network.FirewallPolicyRuleGroupsClient\", \"Delete\", result.Response(), \"Failure sending request\")\n\t\treturn\n\t}\n\n\treturn\n}", "func NewForwardingRule(ctx *pulumi.Context,\n\tname string, args *ForwardingRuleArgs, opts ...pulumi.ResourceOption) (*ForwardingRule, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.AcceleratorId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'AcceleratorId'\")\n\t}\n\tif args.ListenerId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'ListenerId'\")\n\t}\n\tif args.RuleActions == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'RuleActions'\")\n\t}\n\tif args.RuleConditions == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'RuleConditions'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource ForwardingRule\n\terr := ctx.RegisterResource(\"alicloud:ga/forwardingRule:ForwardingRule\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (d *ABFToInterfaceDescriptor) Delete(key string, emptyVal proto.Message, metadata api.Metadata) (err error) {\n\t// validate and get all required values\n\tisIPv6, abfIdx, ifIdx, priority, err := d.process(key)\n\tif err != nil {\n\t\td.log.Error(err)\n\t\treturn err\n\t}\n\n\t// detach interface to ABF policy\n\tif isIPv6 {\n\t\treturn d.abfHandler.AbfDetachInterfaceIPv6(abfIdx, ifIdx, priority)\n\t}\n\treturn d.abfHandler.AbfDetachInterfaceIPv4(abfIdx, ifIdx, priority)\n}", "func (api *API) DiscoveryRulesDelete(rules LLDRules) (err error) {\n\tvar ids []string\n\tfor _, rule := range rules {\n\t\tids = append(ids, rule.ItemID)\n\t}\n\n\terr = api.DiscoveryRulesDeletesByIDs(ids)\n\tif err == nil {\n\t\tfor i := range rules {\n\t\t\trules[i].ItemID = \"\"\n\t\t}\n\t}\n\treturn\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingInterfaces(ctx context.Context, name string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:interfaces/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingInterfacesInterfaceInterfaceRef(ctx context.Context, name string, interfaceId string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:interfaces/frinx-openconfig-network-instance:interface/{interface-id}/frinx-openconfig-network-instance:interface-ref/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"interface-id\"+\"}\", fmt.Sprintf(\"%v\", interfaceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (r *bitroute) DELETE(path string, f func(router.Control)) {\n\tr.register(\"DELETE\", path, f)\n}" ]
[ "0.8395745", "0.63098526", "0.62573814", "0.608502", "0.59710705", "0.5967293", "0.5838807", "0.57990897", "0.5739333", "0.5733386", "0.5693557", "0.5688699", "0.5687825", "0.56238353", "0.5586369", "0.55828404", "0.55754846", "0.5564328", "0.55586374", "0.5551078", "0.55271065", "0.5523465", "0.5419317", "0.5418269", "0.54110986", "0.54020995", "0.53967434", "0.5392264", "0.53679395", "0.53621817", "0.5334209", "0.5293554", "0.5271491", "0.52444464", "0.52369165", "0.5216752", "0.520581", "0.5200578", "0.5188862", "0.5187294", "0.5162968", "0.5139046", "0.5135195", "0.51316583", "0.512257", "0.51153505", "0.5114157", "0.51110786", "0.5110693", "0.51097244", "0.5107078", "0.5084663", "0.5081255", "0.506607", "0.506605", "0.50653726", "0.505968", "0.504559", "0.501545", "0.5001069", "0.49880624", "0.4982394", "0.49679825", "0.494744", "0.49444973", "0.49410903", "0.49409738", "0.49019173", "0.4901471", "0.4899641", "0.48929906", "0.48896995", "0.48871577", "0.48848996", "0.48841745", "0.48702765", "0.48632362", "0.48551252", "0.48532745", "0.48504376", "0.4839942", "0.4837996", "0.48347014", "0.48340684", "0.48274204", "0.4817982", "0.48145455", "0.48140806", "0.48134097", "0.48096308", "0.47966322", "0.47953802", "0.4787629", "0.47844777", "0.47775254", "0.47744587", "0.4773947", "0.4758297", "0.4752619", "0.4751485" ]
0.7395078
1
DeleteFirewallRule uses the override method DeleteFirewallRuleFn or the real implementation.
func (c *TestClient) DeleteFirewallRule(project, name string) error { if c.DeleteFirewallRuleFn != nil { return c.DeleteFirewallRuleFn(project, name) } return c.client.DeleteFirewallRule(project, name) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (_m *ComputeAPI) DeleteFirewallRule(project string, firewall string) {\n\t_m.Called(project, firewall)\n}", "func (sdk *MockGoSDKClient) DeleteSQLFirewallRule(ctx context.Context, resourceGroupName string, serverName string, ruleName string) (err error) {\n\treturn nil\n}", "func (fw *AzureSqlFirewallRuleManager) Delete(ctx context.Context, obj runtime.Object, opts ...resourcemanager.ConfigOption) (bool, error) {\n\tinstance, err := fw.convert(obj)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\tsubscriptionID := instance.Spec.SubscriptionID\n\tgroupName := instance.Spec.ResourceGroup\n\tserver := instance.Spec.Server\n\truleName := instance.ObjectMeta.Name\n\n\terr = fw.DeleteSQLFirewallRule(ctx, subscriptionID, groupName, server, ruleName)\n\tif err != nil {\n\t\tcatch := []string{\n\t\t\terrhelp.AsyncOpIncompleteError,\n\t\t}\n\t\tgone := []string{\n\t\t\terrhelp.ResourceGroupNotFoundErrorCode,\n\t\t\terrhelp.ParentNotFoundErrorCode,\n\t\t\terrhelp.NotFoundErrorCode,\n\t\t\terrhelp.ResourceNotFound,\n\t\t}\n\t\tazerr := errhelp.NewAzureError(err)\n\t\tif helpers.ContainsString(catch, azerr.Type) {\n\t\t\treturn true, nil\n\t\t} else if helpers.ContainsString(gone, azerr.Type) {\n\t\t\treturn false, nil\n\t\t}\n\t\tinstance.Status.Message = fmt.Sprintf(\"AzureSqlFirewallRule Delete failed with %s\", err.Error())\n\t\treturn true, err\n\t}\n\tinstance.Status.Message = fmt.Sprintf(\"Delete AzureSqlFirewallRule succeeded\")\n\treturn false, nil\n}", "func (s *LocalTests) deleteFwRule(c *gc.C, fwRuleId string) {\n\terr := s.testClient.DeleteFirewallRule(fwRuleId)\n\tc.Assert(err, gc.IsNil)\n}", "func (f *fakeLB) DeleteForwardingRule(_ context.Context, _, _ string) error {\n\tpanic(\"implement me\")\n}", "func DeleteRule(rule string) {\n\tdelete(customRuleFuncMap, rule)\n}", "func (c *TestClient) DeleteForwardingRule(project, region, name string) error {\n\tif c.DeleteForwardingRuleFn != nil {\n\t\treturn c.DeleteForwardingRuleFn(project, region, name)\n\t}\n\treturn c.client.DeleteForwardingRule(project, region, name)\n}", "func DeleteRule(rule AuditRule) error {\n\tclient, err := libaudit.NewAuditClient(nil)\n\tdefer client.Close()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"Failed to initialize client\")\n\t}\n\tkr, _, _, err := rule.toKernelAuditRule()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"Failed to initialize client\")\n\t}\n\tif err := client.DeleteRule(kr.toWireFormat()); err != nil {\n\t\treturn errors.Wrap(err, \"Failed to delete audit rule\")\n\t}\n\treturn nil\n}", "func DeleteACLRule(r *http.Request, target ACLTarget, actor ACLActor) (err error) {\n\n\t// get service\n\ts := service.Providers.MustService(r, \"ACLRule\")\n\n\t// delete the rules\n\tconds := service.NewConds()\n\tconds.Add(\"target\", string(target))\n\tconds.Add(\"actor\", string(actor))\n\ts.Delete(conds)\n\n\treturn\n}", "func (g *Cloud) DeleteFirewall(name string) error {\n\tctx, cancel := cloud.ContextWithCallTimeout()\n\tdefer cancel()\n\n\tmc := newFirewallMetricContext(\"delete\")\n\treturn mc.Observe(g.c.Firewalls().Delete(ctx, meta.GlobalKey(name)))\n}", "func (se *StorageEndpoint) DeleteRule(ruleKey string) error {\n\treturn se.Remove(ruleKeyPath(ruleKey))\n}", "func DeleteRule(ctx context.Context, p *Protocol, routeID routing.RouteID) error {\n\tif err := p.WritePacket(PacketDeleteRules, []routing.RouteID{routeID}); err != nil {\n\t\treturn err\n\t}\n\tvar res []routing.RouteID\n\tif err := readAndDecodePacketWithTimeout(ctx, p, &res); err != nil {\n\t\treturn err\n\t}\n\tif len(res) == 0 {\n\t\treturn errors.New(\"empty response\")\n\t}\n\treturn nil\n}", "func DeleteNetworkSecurityGroupRule() {}", "func DeleteRule(table Table, chain Chain, args ...string) error {\n\tfullArgs := makeFullArgs(table, opDeleteRule, chain, args...)\n\tout, err := run(cmd, fullArgs...)\n\tif err != nil {\n\t\treturn trace.Wrap(err, \"failed to delete %v chain %v rule %v: %s\", table, chain, args, out)\n\t}\n\treturn nil\n}", "func (firewall *NsxtFirewall) DeleteRuleById(id string) error {\n\tif id == \"\" {\n\t\treturn fmt.Errorf(\"empty ID specified\")\n\t}\n\n\tendpoint := types.OpenApiPathVersion1_0_0 + types.OpenApiEndpointNsxtFirewallRules\n\tminimumApiVersion, err := firewall.client.checkOpenApiEndpointCompatibility(endpoint)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\turlRef, err := firewall.client.OpenApiBuildEndpoint(fmt.Sprintf(endpoint, firewall.edgeGatewayId), \"/\", id)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = firewall.client.OpenApiDeleteItem(minimumApiVersion, urlRef, nil, nil)\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error deleting NSX-T Firewall Rule with ID '%s': %s\", id, err)\n\t}\n\n\treturn nil\n}", "func (m *MockFirewallServiceIface) DeleteFirewallRule(p *DeleteFirewallRuleParams) (*DeleteFirewallRuleResponse, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"DeleteFirewallRule\", p)\n\tret0, _ := ret[0].(*DeleteFirewallRuleResponse)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (c *Client) DeleteRule(args *DeleteRuleArgs) error {\n\tjsonBytes, jsonErr := json.Marshal(args)\n\tif jsonErr != nil {\n\t\treturn jsonErr\n\t}\n\tbody, err := bce.NewBodyFromBytes(jsonBytes)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn DeleteRule(c, body)\n}", "func (c *Client) DeleteFirewall(ctx context.Context, id int) error {\n\te, err := c.Firewalls.Endpoint()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treq := c.R(ctx)\n\n\te = fmt.Sprintf(\"%s/%d\", e, id)\n\t_, err = coupleAPIErrors(req.Delete(e))\n\treturn err\n}", "func (c *AuditClient) DeleteRule(rule []byte) error {\n\tmsg := syscall.NetlinkMessage{\n\t\tHeader: syscall.NlMsghdr{\n\t\t\tType: uint16(auparse.AUDIT_DEL_RULE),\n\t\t\tFlags: syscall.NLM_F_REQUEST | syscall.NLM_F_ACK,\n\t\t},\n\t\tData: rule,\n\t}\n\n\t// Send AUDIT_DEL_RULE message to the kernel.\n\tseq, err := c.Netlink.Send(msg)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed sending delete rule request: %w\", err)\n\t}\n\n\t_, err = c.getReply(seq)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to get ACK to rule delete request: %w\", err)\n\t}\n\n\treturn nil\n}", "func (m *MockFirewallServiceIface) DeleteIpv6FirewallRule(p *DeleteIpv6FirewallRuleParams) (*DeleteIpv6FirewallRuleResponse, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"DeleteIpv6FirewallRule\", p)\n\tret0, _ := ret[0].(*DeleteIpv6FirewallRuleResponse)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (self *PolicyAgent) DelRule(rule *OfnetPolicyRule, ret *bool) error {\n\tlog.Infof(\"Received DelRule: %+v\", rule)\n\n\t// Gte the rule\n\tself.mutex.Lock()\n\tdefer self.mutex.Unlock()\n\tcache := self.Rules[rule.RuleId]\n\tif cache == nil {\n\t\tlog.Errorf(\"Could not find rule: %+v\", rule)\n\t\treturn errors.New(\"rule not found\")\n\t}\n\n\t// Delete the Flow\n\terr := cache.flow.Delete()\n\tif err != nil {\n\t\tlog.Errorf(\"Error deleting flow: %+v. Err: %v\", rule, err)\n\t}\n\n\t// Delete the rule from cache\n\tdelete(self.Rules, rule.RuleId)\n\n\treturn nil\n}", "func (s *Service) DeleteFirewallRulePort(ruleID string) (string, error) {\n\tbody, err := s.deleteFirewallRulePortResponseBody(ruleID)\n\n\treturn body.Data.TaskID, err\n}", "func (mr *MockFirewallServiceIfaceMockRecorder) DeleteFirewallRule(p interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"DeleteFirewallRule\", reflect.TypeOf((*MockFirewallServiceIface)(nil).DeleteFirewallRule), p)\n}", "func (manager *NetworkPolicyManager) deleteFirewallManager(fwKey string) {\n\tmanager.lock.Lock()\n\tdefer manager.lock.Unlock()\n\n\t// The garbage collector will take care of destroying everything inside it\n\t// now that no one points to it anymore.\n\t// Goodbye!\n\tmanager.localFirewalls[fwKey].Destroy()\n\tdelete(manager.localFirewalls, fwKey)\n\tdelete(manager.flaggedForDeletion, fwKey)\n}", "func (manager *NetworkPolicyManager) deleteFirewallManager(fwKey string) {\n\tmanager.lock.Lock()\n\tdefer manager.lock.Unlock()\n\n\t// The garbage collector will take care of destroying everything inside it\n\t// now that no one points to it anymore.\n\t// Goodbye!\n\tmanager.localFirewalls[fwKey].Destroy()\n\tdelete(manager.localFirewalls, fwKey)\n\tdelete(manager.flaggedForDeletion, fwKey)\n}", "func (api *API) TeamsDeleteRule(ctx context.Context, accountID string, ruleId string) error {\n\turi := fmt.Sprintf(\"/accounts/%s/gateway/rules/%s\", accountID, ruleId)\n\n\t_, err := api.makeRequestContext(ctx, http.MethodDelete, uri, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (m *MockFirewallServiceIface) DeletePortForwardingRule(p *DeletePortForwardingRuleParams) (*DeletePortForwardingRuleResponse, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"DeletePortForwardingRule\", p)\n\tret0, _ := ret[0].(*DeletePortForwardingRuleResponse)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (probe *BridgeOfProbe) delRule(rule *Rule) {\n\tlogging.GetLogger().Infof(\"Rule %v deleted\", rule.UUID)\n\tg := probe.OvsOfProbe.Graph\n\tg.Lock()\n\tdefer g.Unlock()\n\n\truleNode := g.LookupFirstNode(graph.Metadata{\"UUID\": rule.UUID})\n\tif ruleNode != nil {\n\t\tg.DelNode(ruleNode)\n\t}\n}", "func (r *FirewallManagementIPRulesResource) Delete(id string) error {\n\tif err := r.c.ModQuery(\"DELETE\", BasePath+FirewallManagementIPRulesEndpoint+\"/\"+id, nil); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (client *FirewallRulesClient) deleteCreateRequest(ctx context.Context, resourceGroupName string, serverName string, firewallRuleName string, options *FirewallRulesBeginDeleteOptions) (*azcore.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DBforPostgreSQL/servers/{serverName}/firewallRules/{firewallRuleName}\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif serverName == \"\" {\n\t\treturn nil, errors.New(\"parameter serverName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{serverName}\", url.PathEscape(serverName))\n\tif firewallRuleName == \"\" {\n\t\treturn nil, errors.New(\"parameter firewallRuleName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{firewallRuleName}\", url.PathEscape(firewallRuleName))\n\treq, err := azcore.NewRequest(ctx, http.MethodDelete, azcore.JoinPaths(client.con.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq.Telemetry(telemetryInfo)\n\treqQP := req.URL.Query()\n\treqQP.Set(\"api-version\", \"2017-12-01\")\n\treq.URL.RawQuery = reqQP.Encode()\n\treq.Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func (az *Cloud) EnsureFirewallDeleted(ctx context.Context, service *apiv1.Service) error {\n\tserviceName := getServiceName(service)\n\n\tglog.V(2).Infof(\"delete(%s): START EnsureFirewallDeleted\", serviceName)\n\n\t// reconcile logic is capable of fully reconcile, so we can use this to delete\n\tservice.Spec.Ports = []apiv1.ServicePort{}\n\n\tsg, existsSg, err := az.getSecurityGroup(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif existsSg {\n\t\t// hack: We expect no new additions, so we can pass * as the destination address.\n\t\treconciledSg, sgNeedsUpdate, reconcileErr := az.reconcileFirewall(sg, service, \"*\")\n\t\tif reconcileErr != nil {\n\t\t\treturn reconcileErr\n\t\t}\n\t\tif sgNeedsUpdate {\n\t\t\tglog.V(3).Infof(\"delete(%s): sg(%s) - updating\", serviceName, az.SecurityGroupName)\n\t\t\t_, err := az.SecurityGroupsClient.CreateOrUpdate(ctx, az.ResourceGroup, *reconciledSg.Name, reconciledSg)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tglog.V(2).Infof(\"delete(%s): FINISH EnsureFirewallDeleted\", serviceName)\n\treturn nil\n}", "func DeleteRule(id string) int {\n\tvar rule database.Rule\n\tdb.DB.Where(\"id = ? \", id).First(&rule)\n\tdb.DB.Where(\"id = ?\", id).Delete(&database.Rule{})\n\treturn rule.ReleaseID\n}", "func HandleDeleteEventingSecurityRule(adminMan *admin.Manager, syncMan *syncman.Manager) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\n\t\t// Get the JWT token from header\n\t\ttoken := utils.GetTokenFromHeader(r)\n\t\tdefer utils.CloseTheCloser(r.Body)\n\n\t\tvars := mux.Vars(r)\n\t\tprojectID := vars[\"project\"]\n\t\tevType := vars[\"id\"]\n\n\t\tctx, cancel := context.WithTimeout(r.Context(), time.Duration(utils.DefaultContextTime)*time.Second)\n\t\tdefer cancel()\n\n\t\t// Check if the request is authorised\n\t\treqParams, err := adminMan.IsTokenValid(ctx, token, \"eventing-rule\", \"modify\", map[string]string{\"project\": projectID, \"id\": evType})\n\t\tif err != nil {\n\t\t\t_ = helpers.Logger.LogError(helpers.GetRequestID(ctx), \"Failed to validate token for delete eventing rules\", err, nil)\n\t\t\t_ = helpers.Response.SendErrorResponse(ctx, w, http.StatusUnauthorized, err)\n\t\t\treturn\n\t\t}\n\n\t\treqParams = utils.ExtractRequestParams(r, reqParams, nil)\n\t\tstatus, err := syncMan.SetDeleteEventingSecurityRules(ctx, projectID, evType, reqParams)\n\t\tif err != nil {\n\t\t\t_ = helpers.Logger.LogError(helpers.GetRequestID(ctx), \"Failed to delete eventing rules\", err, nil)\n\t\t\t_ = helpers.Response.SendErrorResponse(ctx, w, status, err)\n\t\t\treturn\n\t\t}\n\n\t\t_ = helpers.Response.SendOkayResponse(ctx, status, w)\n\t}\n}", "func (fw *IPtables) deleteIPtablesRule(rule *IPtablesRule) error {\n\tif err := fw.Store.switchIPtablesRule(rule, setRuleInactive); err != nil {\n\t\tlog.Error(\"In deleteIPtablesRule() failed to deactivate the rule\", rule.GetBody())\n\t\treturn err\n\t}\n\n\tif err1 := fw.EnsureRule(rule, EnsureAbsent); err1 != nil {\n\t\tlog.Errorf(\"In deleteIPtablesRule() rule %s set inactive but failed to uninstall\", rule.GetBody())\n\t\treturn err1\n\t}\n\n\tif err2 := fw.Store.deleteIPtablesRule(rule); err2 != nil {\n\t\tlog.Errorf(\"In deleteIPtablesRule() rule %s set inactive and uninstalled but failed to delete DB record\", rule.GetBody())\n\t\treturn err2\n\t}\n\treturn nil\n}", "func (c *Client) DeleteFirewall(ctx context.Context, params *DeleteFirewallInput, optFns ...func(*Options)) (*DeleteFirewallOutput, error) {\n\tif params == nil {\n\t\tparams = &DeleteFirewallInput{}\n\t}\n\n\tresult, metadata, err := c.invokeOperation(ctx, \"DeleteFirewall\", params, optFns, c.addOperationDeleteFirewallMiddlewares)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tout := result.(*DeleteFirewallOutput)\n\tout.ResultMetadata = metadata\n\treturn out, nil\n}", "func (a *SyncApiService) DeleteSyncRule(ctx context.Context, syncRuleId string) ( *http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/platform/3/sync/rules/{SyncRuleId}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"SyncRuleId\"+\"}\", fmt.Sprintf(\"%v\", syncRuleId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{ \"application/json\", }\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\n\t\t\"application/json\",\n\t\t}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\tdefer localVarHttpResponse.Body.Close()\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tbodyBytes, _ := ioutil.ReadAll(localVarHttpResponse.Body)\n\t\treturn localVarHttpResponse, reportError(\"Status: %v, Body: %s\", localVarHttpResponse.Status, bodyBytes)\n\t}\n\n\treturn localVarHttpResponse, err\n}", "func (m *MockFirewallServiceIface) DeletePaloAltoFirewall(p *DeletePaloAltoFirewallParams) (*DeletePaloAltoFirewallResponse, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"DeletePaloAltoFirewall\", p)\n\tret0, _ := ret[0].(*DeletePaloAltoFirewallResponse)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func NewDeleteRule(edgeId string, ruleId int) *DeleteEdgeFirewallRuleAPI {\n\tthis := new(DeleteEdgeFirewallRuleAPI)\n\tthis.BaseAPI = api.NewBaseAPI(http.MethodDelete, fmt.Sprintf(\n\t\t\"/api/4.0/edges/%s/firewall/config/rules/%d\", edgeId, ruleId,\n\t), nil, nil)\n\n\treturn this\n}", "func (client FirewallPolicyRuleGroupsClient) Delete(ctx context.Context, resourceGroupName string, firewallPolicyName string, ruleGroupName string) (result FirewallPolicyRuleGroupsDeleteFuture, err error) {\n\tif tracing.IsEnabled() {\n\t\tctx = tracing.StartSpan(ctx, fqdn+\"/FirewallPolicyRuleGroupsClient.Delete\")\n\t\tdefer func() {\n\t\t\tsc := -1\n\t\t\tif result.FutureAPI != nil && result.FutureAPI.Response() != nil {\n\t\t\t\tsc = result.FutureAPI.Response().StatusCode\n\t\t\t}\n\t\t\ttracing.EndSpan(ctx, sc, err)\n\t\t}()\n\t}\n\treq, err := client.DeletePreparer(ctx, resourceGroupName, firewallPolicyName, ruleGroupName)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"network.FirewallPolicyRuleGroupsClient\", \"Delete\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresult, err = client.DeleteSender(req)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"network.FirewallPolicyRuleGroupsClient\", \"Delete\", result.Response(), \"Failure sending request\")\n\t\treturn\n\t}\n\n\treturn\n}", "func (nsxtNat *NsxtNatRule) Delete() error {\n\tclient := nsxtNat.client\n\tendpoint := types.OpenApiPathVersion1_0_0 + types.OpenApiEndpointNsxtNatRules\n\tapiVersion, err := client.getOpenApiHighestElevatedVersion(endpoint)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif nsxtNat.NsxtNatRule.ID == \"\" {\n\t\treturn fmt.Errorf(\"cannot delete NSX-T NAT rule without ID\")\n\t}\n\n\turlRef, err := client.OpenApiBuildEndpoint(fmt.Sprintf(endpoint, nsxtNat.edgeGatewayId), nsxtNat.NsxtNatRule.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = client.OpenApiDeleteItem(apiVersion, urlRef, nil, nil)\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error deleting NSX-T NAT Rule: %s\", err)\n\t}\n\n\treturn nil\n}", "func deleteRule(c *cli.Context) error {\n\n\truleid, err := hex.DecodeString(c.String(\"id\"))\n\tif err != nil {\n\t\treturn fmt.Errorf(\"[x] Incorrect ruleid format. \")\n\t}\n\n\terr = mapi.ExecuteMailRuleDelete(ruleid)\n\tif err == nil {\n\t\tfmt.Println(\"[*] Rule deleted. Fetching list of remaining rules...\")\n\t\trules, er := mapi.DisplayRules()\n\t\tif er != nil {\n\t\t\treturn er\n\t\t}\n\t\tfmt.Printf(\"[+] Found %d rules\\n\", len(rules))\n\t\tfor _, v := range rules {\n\t\t\tfmt.Printf(\"Rule: %s RuleID: %x\\n\", string(v.RuleName), v.RuleID)\n\t\t}\n\t\treturn nil\n\t}\n\treturn err\n}", "func HandleDeleteEventingTriggerRule(adminMan *admin.Manager, syncMan *syncman.Manager) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\n\t\t// Get the JWT token from header\n\t\ttoken := utils.GetTokenFromHeader(r)\n\n\t\tvars := mux.Vars(r)\n\t\truleName := vars[\"id\"]\n\t\tprojectID := vars[\"project\"]\n\n\t\tdefer utils.CloseTheCloser(r.Body)\n\n\t\tctx, cancel := context.WithTimeout(r.Context(), time.Duration(utils.DefaultContextTime)*time.Second)\n\t\tdefer cancel()\n\n\t\t// Check if the request is authorised\n\t\treqParams, err := adminMan.IsTokenValid(ctx, token, \"eventing-trigger\", \"modify\", map[string]string{\"project\": projectID, \"id\": ruleName})\n\t\tif err != nil {\n\t\t\t_ = helpers.Response.SendErrorResponse(ctx, w, http.StatusUnauthorized, err)\n\t\t\treturn\n\t\t}\n\n\t\treqParams = utils.ExtractRequestParams(r, reqParams, nil)\n\t\tstatus, err := syncMan.SetDeleteEventingRule(ctx, projectID, ruleName, reqParams)\n\t\tif err != nil {\n\t\t\t_ = helpers.Response.SendErrorResponse(ctx, w, status, err)\n\t\t\treturn\n\t\t}\n\n\t\t_ = helpers.Response.SendOkayResponse(ctx, status, w)\n\t}\n}", "func (_m *FakeScheduleService) DeleteAlertRule(key models.AlertRuleKey) {\n\t_m.Called(key)\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRulesRule(ctx context.Context, name string, policyId string, sequenceId int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/frinx-openconfig-network-instance:rule/{sequence-id}/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"sequence-id\"+\"}\", fmt.Sprintf(\"%v\", sequenceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func DeleteJumpRule(v, tableName, srcChainName, dstChainName string) error {\n\tr, err := GetJumpRule(v, tableName, srcChainName, dstChainName)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif r == nil {\n\t\treturn nil\n\t}\n\n\tconn, err := initNftConn()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttb := &nftables.Table{\n\t\tName: tableName,\n\t}\n\tif v == \"4\" {\n\t\ttb.Family = nftables.TableFamilyIPv4\n\t} else {\n\t\ttb.Family = nftables.TableFamilyIPv6\n\t}\n\n\tch := &nftables.Chain{\n\t\tName: srcChainName,\n\t\tTable: tb,\n\t}\n\n\tconn.DelRule(&nftables.Rule{\n\t\tTable: tb,\n\t\tChain: ch,\n\t\t//&nftables.Chain{Name: r.Chain.Name, Type: r.Chain.Type},\n\t\tHandle: r.Handle,\n\t})\n\n\tif err := conn.Flush(); err != nil {\n\t\treturn fmt.Errorf(\n\t\t\t\"error deleting jump rule to %s chain found in chain %s in %s table: %s\",\n\t\t\tdstChainName, r.Chain.Name, r.Table.Name, err,\n\t\t)\n\t}\n\n\treturn nil\n}", "func (firewall *NsxtFirewall) DeleteAllRules() error {\n\n\tif firewall.edgeGatewayId == \"\" {\n\t\treturn fmt.Errorf(\"missing Edge Gateway ID\")\n\t}\n\n\tendpoint := types.OpenApiPathVersion1_0_0 + types.OpenApiEndpointNsxtFirewallRules\n\tminimumApiVersion, err := firewall.client.checkOpenApiEndpointCompatibility(endpoint)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\turlRef, err := firewall.client.OpenApiBuildEndpoint(fmt.Sprintf(endpoint, firewall.edgeGatewayId))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = firewall.client.OpenApiDeleteItem(minimumApiVersion, urlRef, nil, nil)\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error deleting all NSX-T Firewall Rules: %s\", err)\n\t}\n\n\treturn nil\n}", "func (client *Client) DeleteRule(request *DeleteRuleRequest) (_result *DeleteRuleResponse, _err error) {\n\truntime := &util.RuntimeOptions{}\n\t_result = &DeleteRuleResponse{}\n\t_body, _err := client.DeleteRuleWithOptions(request, runtime)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\t_result = _body\n\treturn _result, _err\n}", "func (r *FirewallGlobalRulesStagedPolicyRulesResource) Delete(id string) error {\n\tif err := r.c.ModQuery(\"DELETE\", BasePath+FirewallGlobalRulesStagedPolicyRulesEndpoint+\"/\"+id, nil); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (m *Manager) DeleteRule(n string) bool {\n\tm.mu.Lock()\n\tdefer m.mu.Unlock()\n\tif _, ok := m.rules[n]; !ok {\n\t\treturn false\n\t}\n\tdelete(m.rules, n)\n\treturn true\n}", "func (m *MockInterface) DeleteRule(arg0 iptables.Protocol, arg1, arg2 string, arg3 []string) error {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"DeleteRule\", arg0, arg1, arg2, arg3)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func DeleteL7Rule(t *testing.T, client *gophercloud.ServiceClient, lbID, policyID, ruleID string) {\n\tt.Logf(\"Attempting to delete l7 rule %s\", ruleID)\n\n\tif err := l7policies.DeleteRule(client, policyID, ruleID).ExtractErr(); err != nil {\n\t\tif _, ok := err.(gophercloud.ErrDefault404); !ok {\n\t\t\tt.Fatalf(\"Unable to delete l7 rule: %v\", err)\n\t\t}\n\t}\n\n\tif err := WaitForLoadBalancerState(client, lbID, \"ACTIVE\"); err != nil {\n\t\tt.Fatalf(\"Timed out waiting for loadbalancer to become active: %s\", err)\n\t}\n\n\tt.Logf(\"Successfully deleted l7 rule %s\", ruleID)\n}", "func (c *Firewall) Delete(e ...interface{}) error {\n\tnames, nErr := toNames(e)\n\treturn c.ns.Delete(\"\", \"\", c.pather(), names, nErr)\n}", "func (_m *PrometheusAlertClient) DeleteRule(filePrefix string, ruleName string) error {\n\tret := _m.Called(filePrefix, ruleName)\n\n\tvar r0 error\n\tif rf, ok := ret.Get(0).(func(string, string) error); ok {\n\t\tr0 = rf(filePrefix, ruleName)\n\t} else {\n\t\tr0 = ret.Error(0)\n\t}\n\n\treturn r0\n}", "func (hd *Datapath) DeleteTCPProxyPolicy(tcp *netproto.TCPProxyPolicy, vrf *netproto.Vrf) error {\n\t// This will ensure that only one datapath config will be active at a time. This is a temporary restriction\n\t// to ensure that HAL will use a single config thread , this will be removed prior to FCS to allow parallel configs to go through.\n\t// TODO Remove Global Locking\n\thd.Lock()\n\tdefer hd.Unlock()\n\tvrfKey := &halproto.VrfKeyHandle{\n\t\tKeyOrHandle: &halproto.VrfKeyHandle_VrfId{\n\t\t\tVrfId: vrf.Status.VrfID,\n\t\t},\n\t}\n\n\ttcpProxyPolicyDelReq := &halproto.TcpProxyRuleDeleteRequestMsg{\n\t\tRequest: []*halproto.TcpProxyRuleDeleteRequest{\n\t\t\t{\n\t\t\t\tKeyOrHandle: &halproto.TcpProxyRuleKeyHandle{\n\t\t\t\t\tKeyOrHandle: &halproto.TcpProxyRuleKeyHandle_RuleKey{\n\t\t\t\t\t\tRuleKey: &halproto.TcpProxyRuleKey{\n\t\t\t\t\t\t\tTcpProxyRuleId: tcp.Status.TCPProxyPolicyID,\n\t\t\t\t\t\t\tVrfKeyOrHandle: vrfKey,\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\tif hd.Kind == \"hal\" {\n\t\tresp, err := hd.Hal.TCPProxyPolicyClient.TcpProxyRuleDelete(context.Background(), tcpProxyPolicyDelReq)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error deleting TCPProxy Policy. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\t\tif resp.Response[0].ApiStatus != halproto.ApiStatus_API_STATUS_OK {\n\t\t\tlog.Errorf(\"HAL returned non OK status. %v\", resp.Response[0].ApiStatus.String())\n\t\t\treturn fmt.Errorf(\"HAL returned non OK status. %v\", resp.Response[0].ApiStatus.String())\n\t\t}\n\t} else {\n\t\t_, err := hd.Hal.TCPProxyPolicyClient.TcpProxyRuleDelete(context.Background(), tcpProxyPolicyDelReq)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error deleting TCPProxy Policy. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (r *DeviceComplianceScheduledActionForRuleRequest) Delete(ctx context.Context) error {\n\treturn r.JSONRequest(ctx, \"DELETE\", \"\", nil, nil)\n}", "func DeleteIptableRule(version, tableName, chainName, match, target string) error {\n\tparams := fmt.Sprintf(\"-t %s -D %s %s -j %s\", tableName, chainName, match, target)\n\treturn RunCmd(version, params)\n}", "func (client *ClientImpl) DeleteProcessWorkItemTypeRule(ctx context.Context, args DeleteProcessWorkItemTypeRuleArgs) error {\n\trouteValues := make(map[string]string)\n\tif args.ProcessId == nil {\n\t\treturn &azuredevops.ArgumentNilError{ArgumentName: \"args.ProcessId\"}\n\t}\n\trouteValues[\"processId\"] = (*args.ProcessId).String()\n\tif args.WitRefName == nil || *args.WitRefName == \"\" {\n\t\treturn &azuredevops.ArgumentNilOrEmptyError{ArgumentName: \"args.WitRefName\"}\n\t}\n\trouteValues[\"witRefName\"] = *args.WitRefName\n\tif args.RuleId == nil {\n\t\treturn &azuredevops.ArgumentNilError{ArgumentName: \"args.RuleId\"}\n\t}\n\trouteValues[\"ruleId\"] = (*args.RuleId).String()\n\n\tlocationId, _ := uuid.Parse(\"76fe3432-d825-479d-a5f6-983bbb78b4f3\")\n\t_, err := client.Client.Send(ctx, http.MethodDelete, locationId, \"6.0-preview.2\", routeValues, nil, nil, \"\", \"application/json\", nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (r *Router) DeleteFunc(pattern string, fn http.HandlerFunc) Route {\n\treturn r.Delete(pattern, http.HandlerFunc(fn))\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRulesRuleAction(ctx context.Context, name string, policyId string, sequenceId int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/frinx-openconfig-network-instance:rule/{sequence-id}/frinx-openconfig-network-instance:action/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"sequence-id\"+\"}\", fmt.Sprintf(\"%v\", sequenceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (a *Client) DeleteScopeRule(params *DeleteScopeRuleParams, authInfo runtime.ClientAuthInfoWriter) (*DeleteScopeRuleNoContent, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewDeleteScopeRuleParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"DeleteScopeRule\",\n\t\tMethod: \"DELETE\",\n\t\tPathPattern: \"/cdn/v1/stacks/{stack_id}/sites/{site_id}/scopes/{scope_id}/rules/{rule_id}\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &DeleteScopeRuleReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*DeleteScopeRuleNoContent), nil\n\n}", "func (mr *MockFirewallServiceIfaceMockRecorder) DeleteIpv6FirewallRule(p interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"DeleteIpv6FirewallRule\", reflect.TypeOf((*MockFirewallServiceIface)(nil).DeleteIpv6FirewallRule), p)\n}", "func (a *BackendOptionsApiService) DeleteServerSwitchingRule(ctx _context.Context, id int32, backend string, localVarOptionals *DeleteServerSwitchingRuleOpts) (*_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/services/haproxy/configuration/server_switching_rules/{id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"id\"+\"}\", _neturl.QueryEscape(fmt.Sprintf(\"%v\", id)), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tlocalVarQueryParams.Add(\"backend\", parameterToString(backend, \"\"))\n\tif localVarOptionals != nil && localVarOptionals.TransactionId.IsSet() {\n\t\tlocalVarQueryParams.Add(\"transaction_id\", parameterToString(localVarOptionals.TransactionId.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Version.IsSet() {\n\t\tlocalVarQueryParams.Add(\"version\", parameterToString(localVarOptionals.Version.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.ForceReload.IsSet() {\n\t\tlocalVarQueryParams.Add(\"force_reload\", parameterToString(localVarOptionals.ForceReload.Value(), \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tvar v ModelError\n\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\tif err != nil {\n\t\t\tnewErr.error = err.Error()\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tnewErr.model = v\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func TestFunnelDelete(t *testing.T) {\n\twithDB(func(db *DB) {\n\t\tdb.Do(func(tx *Tx) error {\n\t\t\t// Create account and funnel.\n\t\t\ta := &Account{}\n\t\t\tassert.NoError(t, tx.CreateAccount(a))\n\t\t\tf := &Funnel{Name: \"Funnel Y\", Steps: []*FunnelStep{{Condition: \"action == 'foo'\"}}}\n\t\t\tassert.NoError(t, a.CreateFunnel(f))\n\n\t\t\t// Delete the funnel.\n\t\t\tassert.NoError(t, f.Delete())\n\n\t\t\t// Retrieve the funnel again.\n\t\t\t_, err := tx.Funnel(1)\n\t\t\tassert.Equal(t, err, ErrFunnelNotFound)\n\t\t\treturn nil\n\t\t})\n\t})\n}", "func (a *BackendOptionsApiService) DeleteTCPRequestRule(ctx _context.Context, id int32, parentName string, parentType string, localVarOptionals *DeleteTCPRequestRuleOpts) (*_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/services/haproxy/configuration/tcp_request_rules/{id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"id\"+\"}\", _neturl.QueryEscape(fmt.Sprintf(\"%v\", id)), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tlocalVarQueryParams.Add(\"parent_name\", parameterToString(parentName, \"\"))\n\tlocalVarQueryParams.Add(\"parent_type\", parameterToString(parentType, \"\"))\n\tif localVarOptionals != nil && localVarOptionals.TransactionId.IsSet() {\n\t\tlocalVarQueryParams.Add(\"transaction_id\", parameterToString(localVarOptionals.TransactionId.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Version.IsSet() {\n\t\tlocalVarQueryParams.Add(\"version\", parameterToString(localVarOptionals.Version.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.ForceReload.IsSet() {\n\t\tlocalVarQueryParams.Add(\"force_reload\", parameterToString(localVarOptionals.ForceReload.Value(), \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tvar v ModelError\n\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\tif err != nil {\n\t\t\tnewErr.error = err.Error()\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tnewErr.model = v\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func (f *fakeLB) GetFirewallRule(_ context.Context, _, _ string) (*govultr.LBFirewallRule, *http.Response, error) {\n\treturn nil, nil, nil\n}", "func (gateway *EgressGateway) DeleteStaticIptablesRule(setName string, destinationIP, egressIP string) error {\n\n\t// delete rule in NAT postrouting to SNAT traffic\n\truleSpec := []string{\"-m\", \"set\", \"--match-set\", setName, \"src\", \"-d\", destinationIP, \"-j\", \"SNAT\", \"--to-source\", egressIP}\n\tif err := gateway.deleteRule(defaultNATIptable, egressGatewayNATChainName, ruleSpec...); err != nil {\n\t\treturn fmt.Errorf(\"failed to delete rule in chain %v err %v\", egressGatewayNATChainName, err)\n\t}\n\n\t// delete rule in FORWARD chain of filter table\n\truleSpec = []string{\"-m\", \"set\", \"--set\", setName, \"src\", \"-d\", destinationIP, \"-j\", \"ACCEPT\"}\n\thasRule, err := gateway.ipt.Exists(\"filter\", egressGatewayFWChainName, ruleSpec...)\n\tif err != nil {\n\t\treturn errors.New(\"Failed to verify rule exists in \" + egressGatewayFWChainName + \" chain of filter table\" + err.Error())\n\t}\n\tif hasRule {\n\t\terr = gateway.ipt.Delete(\"filter\", egressGatewayFWChainName, ruleSpec...)\n\t\tif err != nil {\n\t\t\treturn errors.New(\"Failed to delete iptables command to ACCEPT traffic from director nodes to get forwarded\" + err.Error())\n\t\t}\n\t}\n\n\tset, err := ipset.New(setName, \"hash:ip\", &ipset.Params{})\n\tif err != nil {\n\t\treturn errors.New(\"Failed to get ipset with name \" + setName + \" due to %\" + err.Error())\n\t}\n\terr = set.Destroy()\n\tif err != nil {\n\t\treturn errors.New(\"Failed to delete ipset due to \" + err.Error())\n\t}\n\n\treturn nil\n}", "func (mr *MockFirewallServiceIfaceMockRecorder) DeletePortForwardingRule(p interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"DeletePortForwardingRule\", reflect.TypeOf((*MockFirewallServiceIface)(nil).DeletePortForwardingRule), p)\n}", "func (s *Manager) SetDeleteEventingRule(ctx context.Context, project, ruleName string, params model.RequestParams) (int, error) {\n\t// Check if the request has been hijacked\n\thookResponse := s.integrationMan.InvokeHook(ctx, params)\n\tif hookResponse.CheckResponse() {\n\t\t// Check if an error occurred\n\t\tif err := hookResponse.Error(); err != nil {\n\t\t\treturn hookResponse.Status(), err\n\t\t}\n\n\t\t// Gracefully return\n\t\treturn hookResponse.Status(), nil\n\t}\n\n\t// Acquire a lock\n\ts.lock.Lock()\n\tdefer s.lock.Unlock()\n\n\tprojectConfig, err := s.getConfigWithoutLock(ctx, project)\n\tif err != nil {\n\t\treturn http.StatusBadRequest, err\n\t}\n\n\tresourceID := config.GenerateResourceID(s.clusterID, project, config.ResourceEventingTrigger, ruleName)\n\tdelete(projectConfig.EventingTriggers, resourceID)\n\n\tif err := s.modules.SetEventingTriggerConfig(ctx, project, projectConfig.EventingTriggers); err != nil {\n\t\treturn http.StatusInternalServerError, helpers.Logger.LogError(helpers.GetRequestID(ctx), \"error setting eventing config\", err, nil)\n\t}\n\n\tif err := s.store.DeleteResource(ctx, resourceID); err != nil {\n\t\treturn http.StatusInternalServerError, err\n\t}\n\n\treturn http.StatusOK, nil\n}", "func (r *FirewallScheduleResource) Delete(id string) error {\n\tif err := r.c.ModQuery(\"DELETE\", BasePath+FirewallScheduleEndpoint+\"/\"+id, nil); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRulesRuleTransport(ctx context.Context, name string, policyId string, sequenceId int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/frinx-openconfig-network-instance:rule/{sequence-id}/frinx-openconfig-network-instance:transport/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"sequence-id\"+\"}\", fmt.Sprintf(\"%v\", sequenceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (mr *MockFirewallServiceIfaceMockRecorder) DeletePaloAltoFirewall(p interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"DeletePaloAltoFirewall\", reflect.TypeOf((*MockFirewallServiceIface)(nil).DeletePaloAltoFirewall), p)\n}", "func (a *BackendOptionsApiService) DeleteStickRule(ctx _context.Context, id int32, backend string, localVarOptionals *DeleteStickRuleOpts) (*_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/services/haproxy/configuration/stick_rules/{id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"id\"+\"}\", _neturl.QueryEscape(fmt.Sprintf(\"%v\", id)), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tlocalVarQueryParams.Add(\"backend\", parameterToString(backend, \"\"))\n\tif localVarOptionals != nil && localVarOptionals.TransactionId.IsSet() {\n\t\tlocalVarQueryParams.Add(\"transaction_id\", parameterToString(localVarOptionals.TransactionId.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Version.IsSet() {\n\t\tlocalVarQueryParams.Add(\"version\", parameterToString(localVarOptionals.Version.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.ForceReload.IsSet() {\n\t\tlocalVarQueryParams.Add(\"force_reload\", parameterToString(localVarOptionals.ForceReload.Value(), \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tvar v ModelError\n\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\tif err != nil {\n\t\t\tnewErr.error = err.Error()\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tnewErr.model = v\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRulesRuleIpv6(ctx context.Context, name string, policyId string, sequenceId int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/frinx-openconfig-network-instance:rule/{sequence-id}/frinx-openconfig-network-instance:ipv6/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"sequence-id\"+\"}\", fmt.Sprintf(\"%v\", sequenceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (r *Route) DeleteFunc(f http.HandlerFunc) *Route {\n\treturn r.Delete(http.HandlerFunc(f))\n}", "func (hd *Datapath) DeleteNatPolicy(np *netproto.NatPolicy, vrf *netproto.Vrf) error {\n\t// This will ensure that only one datapath config will be active at a time. This is a temporary restriction\n\t// to ensure that HAL will use a single config thread , this will be removed prior to FCS to allow parallel configs to go through.\n\t// TODO Remove Global Locking\n\thd.Lock()\n\tdefer hd.Unlock()\n\tvrfKey := &halproto.VrfKeyHandle{\n\t\tKeyOrHandle: &halproto.VrfKeyHandle_VrfId{\n\t\t\tVrfId: vrf.Status.VrfID,\n\t\t},\n\t}\n\n\t// Build Nat Policy Key\n\tnatPolicyKey := &halproto.NatPolicyKeyHandle{\n\t\tKeyOrHandle: &halproto.NatPolicyKeyHandle_PolicyKey{\n\t\t\tPolicyKey: &halproto.NATPolicyKey{\n\t\t\t\tVrfKeyOrHandle: vrfKey,\n\t\t\t\tNatPolicyId: np.Status.NatPolicyID,\n\t\t\t},\n\t\t},\n\t}\n\n\tnpDelReq := &halproto.NatPolicyDeleteRequestMsg{\n\t\tRequest: []*halproto.NatPolicyDeleteRequest{\n\t\t\t{\n\t\t\t\tKeyOrHandle: natPolicyKey,\n\t\t\t},\n\t\t},\n\t}\n\n\t// delete hal objects\n\tif hd.Kind == \"hal\" {\n\t\t// delete route\n\t\tresp, err := hd.Hal.Natclient.NatPolicyDelete(context.Background(), npDelReq)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error deleting nat policy. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\t\tif resp.Response[0].ApiStatus != halproto.ApiStatus_API_STATUS_OK {\n\t\t\tlog.Errorf(\"HAL returned non OK status. %v\", resp.Response[0].ApiStatus.String())\n\t\t\treturn fmt.Errorf(\"HAL returned non OK status. %v\", resp.Response[0].ApiStatus.String())\n\t\t}\n\n\t} else {\n\t\t_, err := hd.Hal.Natclient.NatPolicyDelete(context.Background(), npDelReq)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error deleting nat policy. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (a *BackendOptionsApiService) DeleteHTTPRequestRule(ctx _context.Context, id int32, parentName string, parentType string, localVarOptionals *DeleteHTTPRequestRuleOpts) (*_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/services/haproxy/configuration/http_request_rules/{id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"id\"+\"}\", _neturl.QueryEscape(fmt.Sprintf(\"%v\", id)), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tlocalVarQueryParams.Add(\"parent_name\", parameterToString(parentName, \"\"))\n\tlocalVarQueryParams.Add(\"parent_type\", parameterToString(parentType, \"\"))\n\tif localVarOptionals != nil && localVarOptionals.TransactionId.IsSet() {\n\t\tlocalVarQueryParams.Add(\"transaction_id\", parameterToString(localVarOptionals.TransactionId.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Version.IsSet() {\n\t\tlocalVarQueryParams.Add(\"version\", parameterToString(localVarOptionals.Version.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.ForceReload.IsSet() {\n\t\tlocalVarQueryParams.Add(\"force_reload\", parameterToString(localVarOptionals.ForceReload.Value(), \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tvar v ModelError\n\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\tif err != nil {\n\t\t\tnewErr.error = err.Error()\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tnewErr.model = v\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func (s *Service) Delete(ctx context.Context) error {\n\tctx, _, done := tele.StartSpanWithLogger(ctx, \"inboundnatrules.Service.Delete\")\n\tdefer done()\n\n\tctx, cancel := context.WithTimeout(ctx, reconciler.DefaultAzureServiceReconcileTimeout)\n\tdefer cancel()\n\n\tspecs := s.Scope.InboundNatSpecs(make(map[int32]struct{}))\n\tif len(specs) == 0 {\n\t\treturn nil\n\t}\n\n\t// We go through the list of InboundNatSpecs to delete each one, independently of the result of the previous one.\n\t// If multiple errors occur, we return the most pressing one.\n\t// Order of precedence (highest -> lowest) is: error that is not an operationNotDoneError (i.e. error deleting) -> operationNotDoneError (i.e. deleting in progress) -> no error (i.e. deleted)\n\tvar result error\n\tfor _, natRule := range specs {\n\t\tif err := s.DeleteResource(ctx, natRule, serviceName); err != nil {\n\t\t\tif !azure.IsOperationNotDoneError(err) || result == nil {\n\t\t\t\tresult = err\n\t\t\t}\n\t\t}\n\t}\n\n\ts.Scope.UpdateDeleteStatus(infrav1.InboundNATRulesReadyCondition, serviceName, result)\n\treturn result\n}", "func (c *Client) DeleteAclRule(aclRuleId, clientToken string) error {\n\treturn bce.NewRequestBuilder(c).\n\t\tWithURL(getURLForAclRuleId(aclRuleId)).\n\t\tWithMethod(http.DELETE).\n\t\tWithQueryParamFilter(\"clientToken\", clientToken).\n\t\tDo()\n}", "func (c *Client) DeleteEscalationRule(escID string, id string) error {\n\t_, err := c.delete(escPath + \"/\" + escID + \"/escalation_rules/\" + id)\n\treturn err\n}", "func (client *FirewallRulesClient) deleteCreateRequest(ctx context.Context, resourceGroupName string, cacheName string, ruleName string, options *FirewallRulesClientDeleteOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Cache/redis/{cacheName}/firewallRules/{ruleName}\"\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif cacheName == \"\" {\n\t\treturn nil, errors.New(\"parameter cacheName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{cacheName}\", url.PathEscape(cacheName))\n\tif ruleName == \"\" {\n\t\treturn nil, errors.New(\"parameter ruleName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{ruleName}\", url.PathEscape(ruleName))\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodDelete, runtime.JoinPaths(client.host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-06-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func Delete(c *golangsdk.ServiceClient, policyID, ruleID string) (r DeleteResult) {\n\treqOpt := &golangsdk.RequestOpts{\n\t\tMoreHeaders: RequestOpts.MoreHeaders,\n\t}\n\n\t_, r.Err = c.Delete(resourceURL(c, policyID, ruleID), reqOpt)\n\treturn\n}", "func (hd *Datapath) DeleteRoute(rt *netproto.Route, vrf *netproto.Vrf) error {\n\t// This will ensure that only one datapath config will be active at a time. This is a temporary restriction\n\t// to ensure that HAL will use a single config thread , this will be removed prior to FCS to allow parallel configs to go through.\n\t// TODO Remove Global Locking\n\thd.Lock()\n\tdefer hd.Unlock()\n\tvrfKey := &halproto.VrfKeyHandle{\n\t\tKeyOrHandle: &halproto.VrfKeyHandle_VrfId{\n\t\t\tVrfId: vrf.Status.VrfID,\n\t\t},\n\t}\n\n\t// Build next hop key\n\tnextHopKey := &halproto.NexthopKeyHandle{\n\t\tKeyOrHandle: &halproto.NexthopKeyHandle_NexthopId{\n\t\t\tNexthopId: rt.Status.RouteID,\n\t\t},\n\t}\n\n\t// Build route key\n\tip, net, err := net.ParseCIDR(rt.Spec.IPPrefix)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error parsing the IP Prefix mask from %v. Err: %v\", rt.Spec.IPPrefix, err)\n\n\t}\n\tprefixLen, _ := net.Mask.Size()\n\tipPrefix := &halproto.IPPrefix{\n\t\tAddress: &halproto.IPAddress{\n\t\t\tIpAf: halproto.IPAddressFamily_IP_AF_INET,\n\t\t\tV4OrV6: &halproto.IPAddress_V4Addr{\n\t\t\t\tV4Addr: ipv4Touint32(ip),\n\t\t\t},\n\t\t},\n\t\tPrefixLen: uint32(prefixLen),\n\t}\n\n\trouteKey := &halproto.RouteKeyHandle{\n\t\tKeyOrHandle: &halproto.RouteKeyHandle_RouteKey{\n\t\t\tRouteKey: &halproto.RouteKey{\n\t\t\t\tVrfKeyHandle: vrfKey,\n\t\t\t\tIpPrefix: ipPrefix,\n\t\t\t},\n\t\t},\n\t}\n\n\tnhDelReq := &halproto.NexthopDeleteRequestMsg{\n\t\tRequest: []*halproto.NexthopDeleteRequest{\n\t\t\t{\n\t\t\t\tKeyOrHandle: nextHopKey,\n\t\t\t},\n\t\t},\n\t}\n\n\trtDelReq := &halproto.RouteDeleteRequestMsg{\n\t\tRequest: []*halproto.RouteDeleteRequest{\n\t\t\t{\n\t\t\t\tKeyOrHandle: routeKey,\n\t\t\t},\n\t\t},\n\t}\n\n\t// delete hal objects\n\tif hd.Kind == \"hal\" {\n\t\t// delete route\n\t\trtResp, err := hd.Hal.Netclient.RouteDelete(context.Background(), rtDelReq)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error deleting route. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\t\tif rtResp.Response[0].ApiStatus != halproto.ApiStatus_API_STATUS_OK {\n\t\t\tlog.Errorf(\"HAL returned non OK status. %v\", rtResp.Response[0].ApiStatus.String())\n\t\t\treturn fmt.Errorf(\"HAL returned non OK status. %v\", rtResp.Response[0].ApiStatus.String())\n\t\t}\n\n\t\t// delete next hop\n\t\tnhResp, err := hd.Hal.Netclient.NexthopDelete(context.Background(), nhDelReq)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error deleting next hop. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\t\tif nhResp.Response[0].ApiStatus != halproto.ApiStatus_API_STATUS_OK {\n\t\t\tlog.Errorf(\"HAL returned non OK status. %v\", nhResp.Response[0].ApiStatus.String())\n\t\t\treturn fmt.Errorf(\"HAL returned non OK status. %v\", nhResp.Response[0].ApiStatus.String())\n\t\t}\n\t} else {\n\t\t_, err := hd.Hal.Netclient.RouteDelete(context.Background(), rtDelReq)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error creating route. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRulesRuleL2(ctx context.Context, name string, policyId string, sequenceId int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/frinx-openconfig-network-instance:rule/{sequence-id}/frinx-openconfig-network-instance:l2/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"sequence-id\"+\"}\", fmt.Sprintf(\"%v\", sequenceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (f *FirewallRule) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &f.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &f.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &f.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &f.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (api *API) DiscoveryRulesDelete(rules LLDRules) (err error) {\n\tvar ids []string\n\tfor _, rule := range rules {\n\t\tids = append(ids, rule.ItemID)\n\t}\n\n\terr = api.DiscoveryRulesDeletesByIDs(ids)\n\tif err == nil {\n\t\tfor i := range rules {\n\t\t\trules[i].ItemID = \"\"\n\t\t}\n\t}\n\treturn\n}", "func (_m *DAO) DeleteImmutableRule(ctx context.Context, id int64) error {\n\tret := _m.Called(ctx, id)\n\n\tvar r0 error\n\tif rf, ok := ret.Get(0).(func(context.Context, int64) error); ok {\n\t\tr0 = rf(ctx, id)\n\t} else {\n\t\tr0 = ret.Error(0)\n\t}\n\n\treturn r0\n}", "func (client FirewallPolicyRuleGroupsClient) DeletePreparer(ctx context.Context, resourceGroupName string, firewallPolicyName string, ruleGroupName string) (*http.Request, error) {\n\tpathParameters := map[string]interface{}{\n\t\t\"firewallPolicyName\": autorest.Encode(\"path\", firewallPolicyName),\n\t\t\"resourceGroupName\": autorest.Encode(\"path\", resourceGroupName),\n\t\t\"ruleGroupName\": autorest.Encode(\"path\", ruleGroupName),\n\t\t\"subscriptionId\": autorest.Encode(\"path\", client.SubscriptionID),\n\t}\n\n\tconst APIVersion = \"2019-06-01\"\n\tqueryParameters := map[string]interface{}{\n\t\t\"api-version\": APIVersion,\n\t}\n\n\tpreparer := autorest.CreatePreparer(\n\t\tautorest.AsDelete(),\n\t\tautorest.WithBaseURL(client.BaseURI),\n\t\tautorest.WithPathParameters(\"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/firewallPolicies/{firewallPolicyName}/ruleGroups/{ruleGroupName}\", pathParameters),\n\t\tautorest.WithQueryParameters(queryParameters))\n\treturn preparer.Prepare((&http.Request{}).WithContext(ctx))\n}", "func (s Service) DeleteTransitionRule(ctx context.Context, docID, ruleID []byte) error {\n\treturn s.pendingDocSrv.DeleteTransitionRule(ctx, docID, ruleID)\n}", "func (m *AuthorizationServerPolicyRuleResource) DeleteAuthorizationServerPolicyRule(ctx context.Context, authServerId string, policyId string, ruleId string) (*Response, error) {\n\turl := fmt.Sprintf(\"/api/v1/authorizationServers/%v/policies/%v/rules/%v\", authServerId, policyId, ruleId)\n\n\trq := m.client.CloneRequestExecutor()\n\n\treq, err := rq.WithAccept(\"application/json\").WithContentType(\"application/json\").NewRequest(\"DELETE\", url, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresp, err := m.client.requestExecutor.Do(ctx, req, nil)\n\tif err != nil {\n\t\treturn resp, err\n\t}\n\n\treturn resp, nil\n}", "func (m *MockFirewallServiceIface) DeleteEgressFirewallRule(p *DeleteEgressFirewallRuleParams) (*DeleteEgressFirewallRuleResponse, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"DeleteEgressFirewallRule\", p)\n\tret0, _ := ret[0].(*DeleteEgressFirewallRuleResponse)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (client *Client) DeleteRuleWithOptions(request *DeleteRuleRequest, runtime *util.RuntimeOptions) (_result *DeleteRuleResponse, _err error) {\n\t_err = util.ValidateModel(request)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\t_result = &DeleteRuleResponse{}\n\t_body, _err := client.DoRequest(tea.String(\"deleteRule\"), tea.String(\"HTTP\"), tea.String(\"POST\"), tea.String(\"/openapi/deleteRule\"), nil, tea.ToMap(request), runtime)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\t_err = tea.Convert(_body, &_result)\n\treturn _result, _err\n}", "func (ks *kuiperService) RemoveRule(ctx context.Context, token string, id string) error {\n\tres, err := ks.auth.Identify(ctx, &mainflux.Token{Value: token})\n\tif err != nil {\n\t\treturn ErrUnauthorizedAccess\n\t}\n\t// Retrieve the rule with specified token and id\n\tr, err := ks.rules.RetrieveByID(ctx, token, id)\n\tif err != nil {\n\t\treturn err\n\t}\n\trule, err := ks.ruleManager.getRuleByJson(r.Name, r.SQL)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif err := ks.ruleManager.deleteRule(rule); err != nil {\n\t\treturn err\n\t}\n\treturn ks.rules.Remove(ctx, res.GetValue(), id)\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPoliciesPolicyRulesRuleIpv4(ctx context.Context, name string, policyId string, sequenceId int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/frinx-openconfig-network-instance:policy/{policy-id}/frinx-openconfig-network-instance:rules/frinx-openconfig-network-instance:rule/{sequence-id}/frinx-openconfig-network-instance:ipv4/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"policy-id\"+\"}\", fmt.Sprintf(\"%v\", policyId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"sequence-id\"+\"}\", fmt.Sprintf(\"%v\", sequenceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (s *LocalTests) createFirewallRule(c *gc.C) *cloudapi.FirewallRule {\n\tfwRule, err := s.testClient.CreateFirewallRule(cloudapi.CreateFwRuleOpts{Enabled: false, Rule: testFwRule})\n\tc.Assert(err, gc.IsNil)\n\tc.Assert(fwRule, gc.NotNil)\n\tc.Assert(fwRule.Rule, gc.Equals, testFwRule)\n\tc.Assert(fwRule.Enabled, gc.Equals, false)\n\ttime.Sleep(10 * time.Second)\n\n\treturn fwRule\n}", "func (o *AutomodRuleDatum) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no AutomodRuleDatum provided for delete\")\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), automodRuleDatumPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"automod_rule_data\\\" WHERE \\\"id\\\"=$1\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete from automod_rule_data\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by delete for automod_rule_data\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (ruleset *DnsForwardingRuleset) ValidateDelete() (admission.Warnings, error) {\n\tvalidations := ruleset.deleteValidations()\n\tvar temp any = ruleset\n\tif runtimeValidator, ok := temp.(genruntime.Validator); ok {\n\t\tvalidations = append(validations, runtimeValidator.DeleteValidations()...)\n\t}\n\treturn genruntime.ValidateDelete(validations)\n}", "func (hd *Datapath) DeleteNatBinding(nb *netproto.NatBinding, vrf *netproto.Vrf) error {\n\t// This will ensure that only one datapath config will be active at a time. This is a temporary restriction\n\t// to ensure that HAL will use a single config thread , this will be removed prior to FCS to allow parallel configs to go through.\n\t// TODO Remove Global Locking\n\thd.Lock()\n\tdefer hd.Unlock()\n\tvrfKey := &halproto.VrfKeyHandle{\n\t\tKeyOrHandle: &halproto.VrfKeyHandle_VrfId{\n\t\t\tVrfId: vrf.Status.VrfID,\n\t\t},\n\t}\n\n\t// Build Nat Binding Key\n\tnatBindingIP := net.ParseIP(nb.Spec.IPAddress)\n\tif len(natBindingIP) == 0 {\n\t\tlog.Errorf(\"could not parse IP from {%v}\", natBindingIP)\n\t\treturn ErrIPParse\n\t}\n\n\tipAddr := &halproto.IPAddress{\n\t\tIpAf: halproto.IPAddressFamily_IP_AF_INET,\n\t\tV4OrV6: &halproto.IPAddress_V4Addr{\n\t\t\tV4Addr: ipv4Touint32(natBindingIP),\n\t\t},\n\t}\n\n\tnatBindingKey := &halproto.NatMappingKeyHandle{\n\t\tKeyOrHandle: &halproto.NatMappingKeyHandle_Svc{\n\t\t\tSvc: &halproto.Svc{\n\t\t\t\tVrfKh: vrfKey,\n\t\t\t\tIpAddr: ipAddr,\n\t\t\t},\n\t\t},\n\t}\n\n\tnbDelReq := &halproto.NatMappingDeleteRequestMsg{\n\t\tRequest: []*halproto.NatMappingDeleteRequest{\n\t\t\t{\n\t\t\t\tKeyOrHandle: natBindingKey,\n\t\t\t},\n\t\t},\n\t}\n\n\t// delete hal objects\n\tif hd.Kind == \"hal\" {\n\t\t// delete route\n\t\tresp, err := hd.Hal.Natclient.NatMappingDelete(context.Background(), nbDelReq)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error deleting nat binding. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\t\tif resp.Response[0].ApiStatus != halproto.ApiStatus_API_STATUS_OK {\n\t\t\tlog.Errorf(\"HAL returned non OK status. %v\", resp.Response[0].ApiStatus.String())\n\t\t\treturn fmt.Errorf(\"HAL returned non OK status. %v\", resp.Response[0].ApiStatus.String())\n\t\t}\n\n\t} else {\n\t\t_, err := hd.Hal.Natclient.NatMappingDelete(context.Background(), nbDelReq)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Error deleting nat binding. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (mr *MockInterfaceMockRecorder) DeleteRule(arg0, arg1, arg2, arg3 interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"DeleteRule\", reflect.TypeOf((*MockInterface)(nil).DeleteRule), arg0, arg1, arg2, arg3)\n}", "func Delete(c *gophercloud.ServiceClient, floatingIpId string, pfId string) (r DeleteResult) {\n\t_, r.Err = c.Delete(singlePortForwardingUrl(c, floatingIpId, pfId), nil)\n\treturn\n}", "func (f *FakeInstance) DeleteReverseIPv6(_ context.Context, _, _ string) error {\n\tpanic(\"implement me\")\n}" ]
[ "0.80311286", "0.71021074", "0.7009589", "0.693023", "0.6863643", "0.6640745", "0.6509416", "0.64293075", "0.6252599", "0.6220564", "0.6204096", "0.61772496", "0.6094996", "0.6079349", "0.60738516", "0.60570604", "0.60272664", "0.6006264", "0.5977528", "0.5897861", "0.58776677", "0.58538544", "0.58434105", "0.58386105", "0.58386105", "0.58360183", "0.57233495", "0.5689908", "0.56807715", "0.56571484", "0.56306416", "0.5616039", "0.5604032", "0.5599181", "0.5591767", "0.55851716", "0.55784994", "0.55693483", "0.55676925", "0.55554396", "0.5553706", "0.55353105", "0.55030286", "0.547229", "0.5452209", "0.5429436", "0.5425878", "0.53957343", "0.53856945", "0.5372597", "0.5366245", "0.5353586", "0.53278965", "0.5322812", "0.5307662", "0.5292889", "0.528822", "0.5276125", "0.52757573", "0.52685034", "0.52508265", "0.52337104", "0.52221894", "0.5212442", "0.52006954", "0.51949114", "0.5153419", "0.51521504", "0.51151913", "0.5114551", "0.51142895", "0.51129586", "0.5109157", "0.5107553", "0.5102749", "0.50920004", "0.50725836", "0.5069266", "0.5062121", "0.5056645", "0.5051849", "0.5048367", "0.5046298", "0.50452644", "0.50399303", "0.50364643", "0.50226974", "0.50125295", "0.50040966", "0.49992356", "0.49823573", "0.4970977", "0.49709734", "0.49703506", "0.49557894", "0.49549595", "0.4953746", "0.49454686", "0.494447", "0.494318" ]
0.7126576
1
DeleteImage uses the override method DeleteImageFn or the real implementation.
func (c *TestClient) DeleteImage(project, name string) error { if c.DeleteImageFn != nil { return c.DeleteImageFn(project, name) } return c.client.DeleteImage(project, name) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (ref ostreeReference) DeleteImage(ctx context.Context, sys *types.SystemContext) error {\n\treturn errors.New(\"Deleting images not implemented for ostree: images\")\n}", "func (s *Server) DeleteImage(w http.ResponseWriter, r *http.Request, params httprouter.Params) {\n\tUUID := params.ByName(\"UUID\")\n\tdeleteKey := params.ByName(\"key\")\n\timage, err := s.imageDao.Load(UUID)\n\tif err != nil {\n\t\tw.WriteHeader(http.StatusNotFound)\n\t\treturn\n\t}\n\tif image.Delete != deleteKey {\n\t\tw.WriteHeader(http.StatusForbidden)\n\t\treturn\n\t}\n\terr = s.imageDao.Delete(image)\n\tif err != nil {\n\t\ts.logger.Println(err)\n\t\tw.WriteHeader(http.StatusNotFound)\n\t\treturn\n\t}\n\terr = s.fs.Delete(image)\n\tif err != nil {\n\t\tw.WriteHeader(http.StatusNotFound)\n\t\treturn\n\t}\n\n}", "func (d *driverMock) DeleteImage(ctx context.Context, id string) error {\n\tif d.DeleteImageErr != nil {\n\t\treturn d.DeleteImageErr\n\t}\n\n\td.DeleteImageID = id\n\n\treturn nil\n}", "func (v *IBM) DeleteImage(ctx *lepton.Context, snapshotID string) error {\n\treturn nil\n}", "func DeleteImage(c *fiber.Ctx) {\n\tShopID := c.Params(\"shop_id\")\n\n\tvar DeleteImage DataDeleteImage\n\n\tif errorParse := c.BodyParser(&DeleteImage); errorParse != nil {\n\t\tfmt.Println(\"Error parsing data\", errorParse)\n\t\tc.JSON(ErrorResponse{MESSAGE: \"Error al parsear información\"})\n\t\tc.Status(400)\n\t\treturn\n\t}\n\n\t_, ErrorDelete := sq.Delete(\"images_shop\").\n\t\tWhere(\"url_image = ? AND shop_id = ?\", DeleteImage.URLImage, ShopID).\n\t\tRunWith(database).\n\t\tExec()\n\n\tif ErrorDelete != nil {\n\t\tfmt.Println(\"Error to delete image\", ErrorDelete)\n\t\tc.JSON(ErrorResponse{MESSAGE: \"Error to delete image\"})\n\t\tc.Status(400)\n\t}\n\n\tc.JSON(SuccessResponse{MESSAGE: \"Imagen eliminada\"})\n}", "func (c *Client) DeleteImage(lib, source string) (*api.BaseResponse, error) {\n\treturn api.DeleteImage(c, lib, source)\n}", "func DeleteImage(imageid string) Resp {\n\tpath := image_path(imageid)\n\treturn is_delete(path)\n}", "func DeleteImage(cli bce.Client, imageId string) error {\n\t// Build the request\n\treq := &bce.BceRequest{}\n\treq.SetUri(getImageUriWithId(imageId))\n\treq.SetMethod(http.DELETE)\n\n\t// Send request and get response\n\tresp := &bce.BceResponse{}\n\tif err := cli.SendRequest(req, resp); err != nil {\n\t\treturn err\n\t}\n\tif resp.IsFail() {\n\t\treturn resp.ServiceError()\n\t}\n\n\tdefer func() { resp.Body().Close() }()\n\treturn nil\n}", "func (c *Client) ImageDelete(imgID int) error {\n\targs := make(map[string]interface{})\n\targs[\"ImageID\"] = imgID\n\n\t_, err := c.apiCall(\"image.delete\", args)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (h *Handler) DeleteImage(w http.ResponseWriter, r *http.Request) {\n\timageName := r.URL.Query().Get(\"name\")\n\timagePool := r.URL.Query().Get(\"pool\")\n\n\tdeleteImageReq := model.BlockImage{\n\t\tName: imageName,\n\t\tPoolName: imagePool,\n\t}\n\n\tif deleteImageReq.Name == \"\" || deleteImageReq.PoolName == \"\" {\n\t\tlogger.Errorf(\"image missing required fields: %+v\", deleteImageReq)\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\treturn\n\t}\n\n\terr := ceph.DeleteImage(h.context, h.config.clusterInfo.Name, deleteImageReq.Name, deleteImageReq.PoolName)\n\tif err != nil {\n\t\tlogger.Errorf(\"failed to delete image %+v: %+v\", deleteImageReq, err)\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tw.Write([]byte(fmt.Sprintf(\"succeeded deleting image %s\", deleteImageReq.Name)))\n}", "func (i *PowerVSImageScope) DeleteImage() error {\n\tif err := i.IBMPowerVSClient.DeleteImage(i.IBMPowerVSImage.Status.ImageID); err != nil {\n\t\trecord.Warnf(i.IBMPowerVSImage, \"FailedDeleteImage\", \"Failed image deletion - %v\", err)\n\t\treturn err\n\t}\n\trecord.Eventf(i.IBMPowerVSImage, \"SuccessfulDeleteImage\", \"Deleted Image %q\", i.IBMPowerVSImage.Status.ImageID)\n\treturn nil\n}", "func (c *ImageController) Delete(ctx *app.DeleteImageContext) error {\n\t// ImageController_Delete: start_implement\n\n\t// Put your logic here\n\n\t// ImageController_Delete: end_implement\n\treturn nil\n}", "func DeleteImage(c * gin.Context){\n\tdb := database.DBConn()\n\tid:= c.Param(\"id\")\n\t_, err := db.Query(\"Delete FROM images WHERE id = \" + id)\n\tif err != nil{\n\t\tc.JSON(500, gin.H{\n\t\t\t\"messages\" : \"Story not found\",\n\t\t});\n\t\tpanic(\"error delte clothes\")\n\t}\n\tc.JSON(200, gin.H{\n\t\t\"messages\": \"deleted\",\n\t})\n\tdefer db.Close()\n}", "func DeleteImage(t *testing.T, projectID string, imageID string) {\n\terr := DeleteImageE(t, projectID, imageID)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n}", "func DeleteImage(t *testing.T, client *gophercloud.ServiceClient, image *images.Image) {\n\terr := images.Delete(client, image.ID).ExtractErr()\n\tif err != nil {\n\t\tt.Fatalf(\"Unable to delete image %s: %v\", image.ID, err)\n\t}\n\n\tt.Logf(\"Deleted image: %s\", image.ID)\n}", "func (g Goba) DeleteImage(typ DatabaseType, name string) error {\n\tfor _, handler := range g.handlers {\n\t\tif handler.Type() == typ {\n\t\t\treturn handler.DeleteImage(name)\n\t\t}\n\t}\n\treturn ErrNoSuchHandler\n}", "func deleteImageResource(ctx context.Context, d *schema.ResourceData, meta interface{}) diag.Diagnostics {\n\t// Warning or errors can be collected in a slice type\n\tvar diags diag.Diagnostics\n\tclient := (meta.(Client)).Client\n\tname := rdEntryStr(d, \"name\")\n\tid := rdEntryStr(d, \"id\")\n\terrMsgPrefix := getErrMsgPrefix(\"Image\", name, id, \"Delete\")\n\tcfg, err := getImage(client, name, id)\n\tif err != nil {\n\t\treturn diag.Errorf(\"%s Failed to get Image. err: %s\", errMsgPrefix, err.Error())\n\t}\n\tif cfg == nil {\n\t\tlog.Printf(\"%s Unexpected Error. nil config\", errMsgPrefix)\n\t\treturn diags\n\t}\n\tclient.XRequestIdPrefix = \"TF-image-delete\"\n\turlExtension := getImageUrl(name, id, \"delete\")\n\trspData := &swagger_models.ZsrvResponse{}\n\t_, err = client.SendReq(\"DELETE\", urlExtension, nil, rspData)\n\tif err != nil {\n\t\treturn diag.Errorf(\"%s. Request Failed. err: %s\", errMsgPrefix, err.Error())\n\t}\n\tlog.Printf(\"[INFO] Image %s(id:%s) Delete Successful.\", name, cfg.ID)\n\treturn diags\n}", "func (i *ImagesModel) DeleteImage(imageID string) error {\n\tfound, err := i.GetImage(imageID)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"Getting image metadata\")\n\t}\n\n\tif found == nil {\n\t\treturn controller.ErrImageMetaNotFound\n\t}\n\n\tinUse, err := i.deployments.ImageUsedInActiveDeployment(imageID)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"Checking if image is used in active deployment\")\n\t}\n\n\t// Image is in use, not allowed to delete\n\tif inUse {\n\t\treturn controller.ErrModelImageInActiveDeployment\n\t}\n\n\t// Delete image file (call to external service)\n\t// Noop for not existing file\n\tif err := i.fileStorage.Delete(imageID); err != nil {\n\t\treturn errors.Wrap(err, \"Deleting image file\")\n\t}\n\n\t// Delete metadata\n\tif err := i.imagesStorage.Delete(imageID); err != nil {\n\t\treturn errors.Wrap(err, \"Deleting image metadata\")\n\t}\n\n\treturn nil\n}", "func (v *ImageClient) Delete(imageName string) error {\n\n\t//Construct the composite key to select the entry\n\tkey := ImageKey{\n\t\t// Owner:\townerName,\n\t\t// ClusterName:\tclusterName,\n\t\tImageName: imageName,\n\t}\n\terr := v.util.DBDelete(v.storeName, key, v.tagMeta)\n\n\t//Delete image from FS\n\tfilePath, _, err := v.GetDirPath(imageName)\n\tif err != nil {\n\t\treturn pkgerrors.Wrap(err, \"Get file path\")\n\t}\n\terr = os.Remove(filePath)\n if err != nil {\n return pkgerrors.Wrap(err, \"Delete image file\")\n }\n\n\treturn nil\n}", "func DeleteImage(albName, imgName string) *utils.ApplicationError {\n\treturn model.DeleteImage(albName, imgName)\n}", "func (c CRImage) DeleteImg() {\n\t_, err := dbmap.Delete(&c)\n\tif err != nil {\n\t\tlog.Println(\"Delete failed\", err)\n\t\treturn\n\t}\n\tcf := new(CRFork)\n\terr = dbmap.SelectOne(&cf, \"select fork_id from cr_fork where user_id = ? and image_id = ?\", c.UserId, c.ImageId)\n\tif err != nil {\n\t\treturn\n\t}\n\t_, err = dbmap.Delete(&cf)\n\tif err != nil {\n\t\tlog.Println(\"Delete failed\", err)\n\t\treturn\n\t}\n}", "func DeleteImage(w http.ResponseWriter, r *http.Request) {\n\n\t//Get current Session\n\tsession, _ := store.Get(r, \"session\")\n\tname := session.Values[\"username\"].(string)\n\n\t//Pathparameter\n\tvars := mux.Vars(r)\n\timageID := vars[\"imageID\"]\n\n\t//Get User Commenting Image\n\tuser, err := model.GetUserByUsername(name)\n\tif err != nil {\n\n\t\tw.WriteHeader(http.StatusConflict)\n\t\tw.Write([]byte(err.Error()))\n\t\treturn\n\n\t}\n\n\t//Delete Image from Database\n\tuser.DeleteImage(imageID)\n\n\t//Make Response JSON\n\tresponseModel := struct {\n\t\tMessage string\n\t}{Message: \"Image was deleted Sussesfully!\"}\n\tresponseJSON, err := json.Marshal(responseModel)\n\tif err != nil {\n\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\tw.Write([]byte(err.Error()))\n\t\treturn\n\n\t}\n\n\t//Write response\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=UTF-8\")\n\tw.WriteHeader(http.StatusAccepted)\n\tw.Write(responseJSON)\n}", "func deleteImage(c *cli.Context) error {\n\terr := checkArgCount(c, 1)\n\tif err != nil {\n\t\treturn err\n\t}\n\tid := c.Args().First()\n\n\tif confirmed(c) {\n\t\tclient.Photonclient, err = client.GetClient(c)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tdeleteTask, err := client.Photonclient.Images.Delete(id)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = waitOnTaskOperation(deleteTask.ID, c)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\tfmt.Println(\"OK, canceled\")\n\t}\n\n\treturn nil\n}", "func CtrDeleteImage(reference string) error {\n\tif err := verifyCtr(); err != nil {\n\t\treturn fmt.Errorf(\"CtrDeleteImage: exception while verifying ctrd client: %s\", err.Error())\n\t}\n\treturn CtrdClient.ImageService().Delete(ctrdCtx, reference)\n}", "func (client *Client) DeleteImage(request *DeleteImageRequest) (_result *DeleteImageResponse, _err error) {\n\truntime := &util.RuntimeOptions{}\n\t_result = &DeleteImageResponse{}\n\t_body, _err := client.DeleteImageWithOptions(request, runtime)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\t_result = _body\n\treturn _result, _err\n}", "func (m *Manager) DeleteImage(hash string) error {\n\tif hash == \"\" {\n\t\treturn nil\n\t}\n\tm.imagesLock.Lock()\n\tdelete(m.images, hash)\n\tm.imagesLock.Unlock()\n\treturn os.RemoveAll(filepath.Join(m.Options.Directory, hash))\n}", "func (a *ImagesApiService) DeleteImage(ctx context.Context, imageDigest string) ApiDeleteImageRequest {\n\treturn ApiDeleteImageRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\timageDigest: imageDigest,\n\t}\n}", "func DeleteImageService(imageid string) bool {\n\tsuccess := domain.UserDeleteItem(imageid)\n\treturn success\n}", "func deleteImage(t *testing.T, projectID string, imageName string) {\n\t// Load the Image ID saved by the earlier build_image stage\n\timage := gcp.FetchImage(t, projectID, imageName)\n\timage.DeleteImage(t)\n}", "func (in *Database) DeleteImage(img *types.Image) error {\n\treturn in.delete(\"image\", img)\n}", "func (c *PostImageClient) Delete() *PostImageDelete {\n\tmutation := newPostImageMutation(c.config, OpDelete)\n\treturn &PostImageDelete{config: c.config, hooks: c.Hooks(), mutation: mutation}\n}", "func delImage(w http.ResponseWriter, req *http.Request) {\n\n\t// Manage Cors\n\tsetCors(&w)\n\tif req.Method == \"OPTIONS\" {\n\t\treturn\n\t}\n\n\t// Authenticate user\n\tclaims, err := authRequest(req)\n\tif err != nil {\n\t\tlogger.Error(\"Unauthorized request to upload sending 401: %v\", err)\n\t\tw.WriteHeader(http.StatusUnauthorized)\n\t\tw.Write([]byte(\"401 - Unauthorized request, ensure you sign in and obtain the jwt auth token\"))\n\t\treturn\n\t}\n\n\tvars := mux.Vars(req)\n\t// validate url parameters and retrieve imageMeta\n\timageMeta, err := validateVars(vars)\n\tif err != nil {\n\t\tlogger.Error(\"Failed to validate vars sending 400: %v\", err)\n\t\tif strings.Contains(err.Error(), \"404 - Not found\") {\n\t\t\tw.WriteHeader(http.StatusNotFound)\n\t\t\tw.Write([]byte(\"404 - Not found, no image with that information available\"))\n\t\t\treturn\n\t\t}\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\tw.Write([]byte(\"400 - Bad request unable to parse url parameters\"))\n\t\treturn\n\t}\n\n\t// Ensure there is no uid miss match\n\tuidVal, err := strconv.Atoi(vars[\"uid\"])\n\tif uidVal != int(imageMeta.Uid) {\n\t\tlogger.Error(\"uid miss match when attempting to delete image sending 400\")\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\tw.Write([]byte(\"400 - Uid mismatch ensure you are using the correct image reference\"))\n\t\treturn\n\t}\n\n\t// Ensure user has access permissions\n\tif claims.Uid != int(imageMeta.Uid) {\n\t\tlogger.Error(\"unauthorized user attempting to delete image\")\n\t\tw.WriteHeader(http.StatusUnauthorized)\n\t\tw.Write([]byte(\"401 - Unauthorized, you do not have permissions to modify this image\"))\n\t\treturn\n\t}\n\n\t// Delete meta from database\n\terr = DeleteImageData(imageMeta)\n\tif err != nil {\n\t\tlogger.Error(\"failed to delete image from database sending 500: %v\", err)\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\tw.Write([]byte(\"500 - Unable to delete image from database, try again later\"))\n\t\treturn\n\t}\n\n\t// Delete file from storage\n\tfileRef := fmt.Sprintf(\"./%s/%s/%s\", IMAGE_DIR, vars[\"uid\"], vars[\"fileId\"])\n\terr = os.Remove(fileRef)\n\t// Orphaned file is ok to leave as database entry is already deleted\n\t// Automated data integrity checks or manual removal is recommended\n\t// This will look like a successfull deletion from the users perspective\n\tif err != nil {\n\t\tlogger.Error(\"failed to delete image data, clean orphaned files via automated data integrity check: %v\", err)\n\t} else {\n\t\tlogger.Info(\"Successfully deleted image: %v\", imageMeta.Id)\n\t}\n\n\treturn\n}", "func (s *API) DeleteImage(req *DeleteImageRequest, opts ...scw.RequestOption) (*Image, error) {\n\tvar err error\n\n\tif req.Region == \"\" {\n\t\tdefaultRegion, _ := s.client.GetDefaultRegion()\n\t\treq.Region = defaultRegion\n\t}\n\n\tif fmt.Sprint(req.Region) == \"\" {\n\t\treturn nil, errors.New(\"field Region cannot be empty in request\")\n\t}\n\n\tif fmt.Sprint(req.ImageID) == \"\" {\n\t\treturn nil, errors.New(\"field ImageID cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"DELETE\",\n\t\tPath: \"/registry/v1/regions/\" + fmt.Sprint(req.Region) + \"/images/\" + fmt.Sprint(req.ImageID) + \"\",\n\t\tHeaders: http.Header{},\n\t}\n\n\tvar resp Image\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func (f *FakeImagesClient) Delete(ctx context.Context, deleteOpts *images.DeleteRequest, opts ...grpc.CallOption) (*googleprotobuf.Empty, error) {\n\tf.Lock()\n\tdefer f.Unlock()\n\tf.appendCalled(\"delete\", deleteOpts)\n\tif err := f.getError(\"delete\"); err != nil {\n\t\treturn nil, err\n\t}\n\t_, ok := f.ImageList[deleteOpts.Name]\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"image does not exist\")\n\t}\n\tdelete(f.ImageList, deleteOpts.Name)\n\treturn &googleprotobuf.Empty{}, nil\n}", "func (g GCPClient) DeleteImage(name string) error {\n\tvar notFound bool\n\top, err := g.compute.Images.Delete(g.projectName, name).Do()\n\tif err != nil {\n\t\tif _, ok := err.(*googleapi.Error); !ok {\n\t\t\treturn err\n\t\t}\n\t\tif err.(*googleapi.Error).Code != 404 {\n\t\t\treturn err\n\t\t}\n\t\tnotFound = true\n\t}\n\tif !notFound {\n\t\tlog.Infof(\"Deleting existing image...\")\n\t\tif err := g.pollOperationStatus(op.Name); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tlog.Infof(\"Image %s deleted\", name)\n\t}\n\treturn nil\n}", "func (c *Client) DeleteImage(imageId string) error {\n\treturn DeleteImage(c, imageId)\n}", "func (a *API) DeleteImage(id string, force bool) error {\n\trequest := ecs.CreateDeleteImageRequest()\n\trequest.Scheme = \"https\"\n\trequest.ImageId = id\n\trequest.Force = requests.NewBoolean(force)\n\n\timages, err := a.GetImagesByID(id)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"getting image: %v\", err)\n\t}\n\n\t_, err = a.ecs.DeleteImage(request)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"deleting image: %v\", err)\n\t}\n\n\tvar errs multierror.Error\n\tfor _, img := range images.Images.Image {\n\t\tfor _, mapping := range img.DiskDeviceMappings.DiskDeviceMapping {\n\t\t\terr = a.DeleteSnapshot(mapping.SnapshotId, force)\n\t\t\tif err != nil {\n\t\t\t\terrs = append(errs, fmt.Errorf(\"deleting snapshot %v: %v\", mapping.SnapshotId, err))\n\t\t\t}\n\t\t}\n\t}\n\treturn errs.AsError()\n}", "func (client ListManagementImageClient) DeleteImage(ctx context.Context, listID string, imageID string) (result String, err error) {\n\tif tracing.IsEnabled() {\n\t\tctx = tracing.StartSpan(ctx, fqdn+\"/ListManagementImageClient.DeleteImage\")\n\t\tdefer func() {\n\t\t\tsc := -1\n\t\t\tif result.Response.Response != nil {\n\t\t\t\tsc = result.Response.Response.StatusCode\n\t\t\t}\n\t\t\ttracing.EndSpan(ctx, sc, err)\n\t\t}()\n\t}\n\treq, err := client.DeleteImagePreparer(ctx, listID, imageID)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"contentmoderator.ListManagementImageClient\", \"DeleteImage\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresp, err := client.DeleteImageSender(req)\n\tif err != nil {\n\t\tresult.Response = autorest.Response{Response: resp}\n\t\terr = autorest.NewErrorWithError(err, \"contentmoderator.ListManagementImageClient\", \"DeleteImage\", resp, \"Failure sending request\")\n\t\treturn\n\t}\n\n\tresult, err = client.DeleteImageResponder(resp)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"contentmoderator.ListManagementImageClient\", \"DeleteImage\", resp, \"Failure responding to request\")\n\t}\n\n\treturn\n}", "func DeleteImageE(t *testing.T, projectID string, imageID string) error {\n\tlogger.Logf(t, \"Destroying Image %s\", imageID)\n\n\tctx := context.Background()\n\n\tservice, err := NewComputeServiceE(t)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif _, err := service.Images.Delete(projectID, imageID).Context(ctx).Do(); err != nil {\n\t\treturn fmt.Errorf(\"Images.Delete(%s) got error: %v\", imageID, err)\n\t}\n\n\treturn err\n}", "func (jb *JobBoard) DeleteImage(ctx context.Context, image string) error {\n\tv := url.Values{}\n\tv.Set(\"infra\", \"jupiterbrain\")\n\tv.Set(\"name\", image)\n\n\tu := \"/images?\" + v.Encode()\n\treq, err := jb.newRequest(\"DELETE\", u, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = jb.client.Do(req)\n\treturn err\n}", "func (a *ImagesApiService) DeleteImageExecute(r ApiDeleteImageRequest) (*DeleteImageResponse, *http.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = http.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tformFiles []formFile\n\t\tlocalVarReturnValue *DeleteImageResponse\n\t)\n\n\tlocalBasePath, err := a.client.cfg.ServerURLWithContext(r.ctx, \"ImagesApiService.DeleteImage\")\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, &GenericOpenAPIError{error: err.Error()}\n\t}\n\n\tlocalVarPath := localBasePath + \"/images/{imageDigest}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"imageDigest\"+\"}\", url.PathEscape(parameterToString(r.imageDigest, \"\")), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\tif r.force != nil {\n\t\tlocalVarQueryParams.Add(\"force\", parameterToString(*r.force, \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tif r.xAnchoreAccount != nil {\n\t\tlocalVarHeaderParams[\"x-anchore-account\"] = parameterToString(*r.xAnchoreAccount, \"\")\n\t}\n\treq, err := a.client.prepareRequest(r.ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, formFiles)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(req)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tlocalVarHTTPResponse.Body = ioutil.NopCloser(bytes.NewBuffer(localVarBody))\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := &GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := &GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func (client *Client) DeleteDynamicImage(request *DeleteDynamicImageRequest) (_result *DeleteDynamicImageResponse, _err error) {\n\truntime := &util.RuntimeOptions{}\n\t_result = &DeleteDynamicImageResponse{}\n\t_body, _err := client.DeleteDynamicImageWithOptions(request, runtime)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\t_result = _body\n\treturn _result, _err\n}", "func deleteAPIImageHandler(w http.ResponseWriter, r *http.Request, _ map[string]string) {\n\tuserName := sessionHandler.GetUserName(r)\n\tif userName != \"\" { // TODO: Check if the user has permissions to delete the image\n\t\t// Get the file name from the json data\n\t\tdecoder := json.NewDecoder(r.Body)\n\t\tvar json JSONImage\n\t\terr := decoder.Decode(&json)\n\t\tif err != nil {\n\t\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\terr = filepath.Walk(filenames.ImagesFilepath, func(filePath string, info os.FileInfo, err error) error {\n\t\t\tif !info.IsDir() && filepath.Base(filePath) == filepath.Base(json.Filename) {\n\t\t\t\terr := os.Remove(filePath)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\t\tif err != nil {\n\t\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\tw.WriteHeader(http.StatusOK)\n\t\tw.Write([]byte(\"Image deleted!\"))\n\t\treturn\n\t}\n\thttp.Error(w, \"Not logged in!\", http.StatusInternalServerError)\n}", "func (me imageStore) Delete(imageID uint) (bool, error) {\n\tif err := me.Db.Unscoped().Exec(\"DELETE FROM images WHERE id = ? \", imageID).Error; err != nil {\n\t\treturn false, err\n\t}\n\n\treturn true, nil\n}", "func (client ListManagementImageClient) DeleteImageResponder(resp *http.Response) (result String, err error) {\n\terr = autorest.Respond(\n\t\tresp,\n\t\tclient.ByInspecting(),\n\t\tazure.WithErrorUnlessStatusCode(http.StatusOK),\n\t\tautorest.ByUnmarshallingJSON(&result.Value),\n\t\tautorest.ByClosing())\n\tresult.Response = autorest.Response{Response: resp}\n\treturn\n}", "func (m Config) DelImage(name string, isForce bool) {\n\t//check exists image\n\tif _, ok := m.Images[name]; ok {\n\t\t//insert image\n\t\tdelete(m.Images, name)\n\t} else {\n\t\tfmt.Println(\"Image not found.:\", name)\n\t\texit(1)\n\t\treturn\n\t}\n\treturn\n}", "func qemuImgDelete(sysOS *sys.OS, imgPath string) error {\n\treturn deleteProfile(sysOS, qemuImgProfileName(imgPath), qemuImgProfileFilename(imgPath))\n}", "func (c *UnsavedPostImageClient) Delete() *UnsavedPostImageDelete {\n\tmutation := newUnsavedPostImageMutation(c.config, OpDelete)\n\treturn &UnsavedPostImageDelete{config: c.config, hooks: c.Hooks(), mutation: mutation}\n}", "func (client *Client) DeleteImageWithOptions(request *DeleteImageRequest, runtime *util.RuntimeOptions) (_result *DeleteImageResponse, _err error) {\n\t_err = util.ValidateModel(request)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\tquery := map[string]interface{}{}\n\tif !tea.BoolValue(util.IsUnset(request.DeleteImageType)) {\n\t\tquery[\"DeleteImageType\"] = request.DeleteImageType\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.ImageIds)) {\n\t\tquery[\"ImageIds\"] = request.ImageIds\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.ImageType)) {\n\t\tquery[\"ImageType\"] = request.ImageType\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.ImageURLs)) {\n\t\tquery[\"ImageURLs\"] = request.ImageURLs\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.VideoId)) {\n\t\tquery[\"VideoId\"] = request.VideoId\n\t}\n\n\treq := &openapi.OpenApiRequest{\n\t\tQuery: openapiutil.Query(query),\n\t}\n\tparams := &openapi.Params{\n\t\tAction: tea.String(\"DeleteImage\"),\n\t\tVersion: tea.String(\"2017-03-21\"),\n\t\tProtocol: tea.String(\"HTTPS\"),\n\t\tPathname: tea.String(\"/\"),\n\t\tMethod: tea.String(\"POST\"),\n\t\tAuthType: tea.String(\"AK\"),\n\t\tStyle: tea.String(\"RPC\"),\n\t\tReqBodyType: tea.String(\"formData\"),\n\t\tBodyType: tea.String(\"json\"),\n\t}\n\t_result = &DeleteImageResponse{}\n\t_body, _err := client.CallApi(params, req, runtime)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\t_err = tea.Convert(_body, &_result)\n\treturn _result, _err\n}", "func (p *AWS) DeleteImage(ctx *lepton.Context, imagename string) error {\n\t// delete ami by ami name\n\timage, err := p.findImageByName(imagename)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Error running deregister image operation: %s\", err)\n\t}\n\n\tamiID := aws.StringValue(image.ImageId)\n\tsnapID := aws.StringValue(image.BlockDeviceMappings[0].Ebs.SnapshotId)\n\n\t// grab snapshotid && grab image id\n\n\tparams := &ec2.DeregisterImageInput{\n\t\tImageId: aws.String(amiID),\n\t\tDryRun: aws.Bool(false),\n\t}\n\t_, err = p.ec2.DeregisterImage(params)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Error running deregister image operation: %s\", err)\n\t}\n\n\t// DeleteSnapshot\n\tparams2 := &ec2.DeleteSnapshotInput{\n\t\tSnapshotId: aws.String(snapID),\n\t\tDryRun: aws.Bool(false),\n\t}\n\t_, err = p.ec2.DeleteSnapshot(params2)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Error running snapshot delete: %s\", err)\n\t}\n\n\treturn nil\n}", "func (a *ImageApiService) DeleteUserImage(ctx _context.Context, userId string, imageType ImageType, index int32, itemType string) (*_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/Users/{userId}/Images/{itemType}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"userId\"+\"}\", _neturl.QueryEscape(parameterToString(userId, \"\")) , -1)\n\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"imageType\"+\"}\", _neturl.QueryEscape(parameterToString(imageType, \"\")) , -1)\n\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", _neturl.QueryEscape(parameterToString(index, \"\")) , -1)\n\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"itemType\"+\"}\", _neturl.QueryEscape(parameterToString(itemType, \"\")) , -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\", \"application/json; profile=CamelCase\", \"application/json; profile=PascalCase\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tif ctx != nil {\n\t\t// API Key Authentication\n\t\tif auth, ok := ctx.Value(ContextAPIKey).(APIKey); ok {\n\t\t\tvar key string\n\t\t\tif auth.Prefix != \"\" {\n\t\t\t\tkey = auth.Prefix + \" \" + auth.Key\n\t\t\t} else {\n\t\t\t\tkey = auth.Key\n\t\t\t}\n\t\t\tlocalVarHeaderParams[\"X-Emby-Authorization\"] = key\n\t\t}\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 403 {\n\t\t\tvar v ProblemDetails\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t}\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func (client ArtifactsClient) deleteContainerImage(ctx context.Context, request common.OCIRequest, binaryReqBody *common.OCIReadSeekCloser, extraHeaders map[string]string) (common.OCIResponse, error) {\n\n\thttpRequest, err := request.HTTPRequest(http.MethodDelete, \"/container/images/{imageId}\", binaryReqBody, extraHeaders)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar response DeleteContainerImageResponse\n\tvar httpResponse *http.Response\n\thttpResponse, err = client.Call(ctx, &httpRequest)\n\tdefer common.CloseBodyIfValid(httpResponse)\n\tresponse.RawResponse = httpResponse\n\tif err != nil {\n\t\tapiReferenceLink := \"https://docs.oracle.com/iaas/api/#/en/registry/20160918/ContainerImage/DeleteContainerImage\"\n\t\terr = common.PostProcessServiceError(err, \"Artifacts\", \"DeleteContainerImage\", apiReferenceLink)\n\t\treturn response, err\n\t}\n\n\terr = common.UnmarshalResponse(httpResponse, &response)\n\treturn response, err\n}", "func (c *TestClient) DeleteMachineImage(project, name string) error {\n\tif c.DeleteMachineImageFn != nil {\n\t\treturn c.DeleteMachineImageFn(project, name)\n\t}\n\treturn c.client.DeleteMachineImage(project, name)\n}", "func (client ListManagementImageClient) DeleteImagePreparer(ctx context.Context, listID string, imageID string) (*http.Request, error) {\n\turlParameters := map[string]interface{}{\n\t\t\"Endpoint\": client.Endpoint,\n\t}\n\n\tpathParameters := map[string]interface{}{\n\t\t\"ImageId\": autorest.Encode(\"path\", imageID),\n\t\t\"listId\": autorest.Encode(\"path\", listID),\n\t}\n\n\tpreparer := autorest.CreatePreparer(\n\t\tautorest.AsDelete(),\n\t\tautorest.WithCustomBaseURL(\"{Endpoint}\", urlParameters),\n\t\tautorest.WithPathParameters(\"/contentmoderator/lists/v1.0/imagelists/{listId}/images/{ImageId}\", pathParameters))\n\treturn preparer.Prepare((&http.Request{}).WithContext(ctx))\n}", "func (cl Client) DeleteAnonymousUploadedImg(deleteHash string) (DeleteResponse, error) {\n\tvar err error\n\tir := DeleteResponse{}\n\trequest, _ := cl.PrepareAuthRequest(\"DELETE\", ImageBase+\"/\"+deleteHash)\n\tresponse, err := cl.Do(request)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\tdefer response.Body.Close()\n\tbody, err := ioutil.ReadAll(response.Body)\n\n\terr = json.Unmarshal(body, &ir)\n\tif err != nil {\n\t\treturn ir, err\n\t}\n\treturn ir, err\n}", "func (client ListManagementImageClient) DeleteImageSender(req *http.Request) (*http.Response, error) {\n\treturn autorest.SendWithSender(client, req,\n\t\tautorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n}", "func (a *ImagesApiService) DeleteImageByImageIdExecute(r ApiDeleteImageByImageIdRequest) (*DeleteImageResponse, *http.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = http.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tformFiles []formFile\n\t\tlocalVarReturnValue *DeleteImageResponse\n\t)\n\n\tlocalBasePath, err := a.client.cfg.ServerURLWithContext(r.ctx, \"ImagesApiService.DeleteImageByImageId\")\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, &GenericOpenAPIError{error: err.Error()}\n\t}\n\n\tlocalVarPath := localBasePath + \"/images/by_id/{imageId}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"imageId\"+\"}\", url.PathEscape(parameterToString(r.imageId, \"\")), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\tif r.force != nil {\n\t\tlocalVarQueryParams.Add(\"force\", parameterToString(*r.force, \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tif r.xAnchoreAccount != nil {\n\t\tlocalVarHeaderParams[\"x-anchore-account\"] = parameterToString(*r.xAnchoreAccount, \"\")\n\t}\n\treq, err := a.client.prepareRequest(r.ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, formFiles)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(req)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tlocalVarHTTPResponse.Body = ioutil.NopCloser(bytes.NewBuffer(localVarBody))\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := &GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 500 {\n\t\t\tvar v ApiErrorResponse\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := &GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func (c *GalleryImageClient) Delete(ctx context.Context, location, name string) error {\n\treturn c.internal.Delete(ctx, location, name)\n}", "func (c *UploadController) Delete() {\n\timg := struct {\n\t\tFileName string `json:\"fileName\"`\n\t}{}\n\terr := json.Unmarshal(c.Ctx.Input.RequestBody, &img)\n\tif err != nil {\n\t\tlog.Println(err)\n\t}\n\n\t// remove thumbnail\n\terr = os.Remove(thumbnailsFolder + img.FileName)\n\tif err != nil {\n\t\tlog.Println(err)\n\t}\n\t// remove main image\n\terr = os.Remove(imagesFolder + img.FileName)\n\tif err != nil {\n\t\tlog.Println(err)\n\t}\n\tc.Data[\"json\"] = img\n\tc.ServeJSON()\n}", "func (s *Service) SignaturesDeleteImage(imageType string, signatureID string, userID string) *SignaturesDeleteImageOp {\n\treturn &SignaturesDeleteImageOp{\n\t\tCredential: s.credential,\n\t\tMethod: \"DELETE\",\n\t\tPath: strings.Join([]string{\"users\", userID, \"signatures\", signatureID, imageType}, \"/\"),\n\t\tAccept: \"application/json\",\n\t\tQueryOpts: make(url.Values),\n\t\tVersion: esign.APIv21,\n\t}\n}", "func (a *ImageApiService) ImageDeletePage(ctx _context.Context, imageDeletePageParameters ImageDeletePageParameters) (ImageDeletePageResponse, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodPost\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue ImageDeletePageResponse\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/api/image/ImageDeletePage\"\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{\"application/json-patch+json\", \"application/json\", \"text/json\", \"application/_*+json\"}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"text/plain\", \"application/json\", \"text/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &imageDeletePageParameters\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 200 {\n\t\t\tvar v ImageDeletePageResponse\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func (client *Client) DeleteVirtualImage(id int64, req *Request) (*Response, error) {\n\treturn client.Execute(&Request{\n\t\tMethod: \"DELETE\",\n\t\tPath: fmt.Sprintf(\"%s/%d\", VirtualImagesPath, id),\n\t\tQueryParams: req.QueryParams,\n\t\tBody: req.Body,\n\t\tResult: &DeleteVirtualImageResult{},\n\t})\n}", "func (es *etcdStore) Delete(imageID string) error {\n\tkey := path.Join(es.prefix, imageID)\n\tif _, err := es.client.Delete(key, true); err != nil {\n\t\tetcdErr := err.(*etcd.EtcdError)\n\t\tif etcdErr.ErrorCode != etcderr.EcodeKeyNotFound {\n\t\t\tlog.WithFields(etcdLogFields).WithFields(log.Fields{\n\t\t\t\t\"error\": err,\n\t\t\t\t\"key\": key,\n\t\t\t}).Error(\"failed to delete image\")\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func deleteImage(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\",\"application/json\")\n\tparam := mux.Vars(r)\n\t//CQL Operation\n\tif err:= Session.Query(`UPDATE albumtable SET imagelist=imagelist-['?'] WHERE albname=?;`,param[\"image\"],param[\"album\"]).Exec();err!=nil {\n\t\tfmt.Println(err)\n\t} else {\n\t\tfmt.Fprintf(w, \"New image added\")\n\t}\n}", "func (a *ImageApiService) DeleteUserImage2(ctx _context.Context, userId string, imageType ImageType, index int32, itemType string) (*_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/Users/{userId}/Images/{itemType}/{index}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"userId\"+\"}\", _neturl.QueryEscape(parameterToString(userId, \"\")) , -1)\n\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"imageType\"+\"}\", _neturl.QueryEscape(parameterToString(imageType, \"\")) , -1)\n\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", _neturl.QueryEscape(parameterToString(index, \"\")) , -1)\n\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"itemType\"+\"}\", _neturl.QueryEscape(parameterToString(itemType, \"\")) , -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\", \"application/json; profile=CamelCase\", \"application/json; profile=PascalCase\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tif ctx != nil {\n\t\t// API Key Authentication\n\t\tif auth, ok := ctx.Value(ContextAPIKey).(APIKey); ok {\n\t\t\tvar key string\n\t\t\tif auth.Prefix != \"\" {\n\t\t\t\tkey = auth.Prefix + \" \" + auth.Key\n\t\t\t} else {\n\t\t\t\tkey = auth.Key\n\t\t\t}\n\t\t\tlocalVarHeaderParams[\"X-Emby-Authorization\"] = key\n\t\t}\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 403 {\n\t\t\tvar v ProblemDetails\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t}\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func (a *Client) DeleteContentLibraryImage(params *DeleteContentLibraryImageParams, opts ...ClientOption) (*DeleteContentLibraryImageOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewDeleteContentLibraryImageParams()\n\t}\n\top := &runtime.ClientOperation{\n\t\tID: \"DeleteContentLibraryImage\",\n\t\tMethod: \"POST\",\n\t\tPathPattern: \"/delete-content-library-image\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\"},\n\t\tParams: params,\n\t\tReader: &DeleteContentLibraryImageReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t}\n\tfor _, opt := range opts {\n\t\topt(op)\n\t}\n\n\tresult, err := a.transport.Submit(op)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*DeleteContentLibraryImageOK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\t// safeguard: normally, absent a default response, unknown success responses return an error above: so this is a codegen issue\n\tmsg := fmt.Sprintf(\"unexpected success response for DeleteContentLibraryImage: API contract not enforced by server. Client expected to get an error, but got: %T\", result)\n\tpanic(msg)\n}", "func (c Client) DestroyImage(image models.Image) error {\n\turl := fmt.Sprintf(\"/images/%d\", image.ID)\n\tresp, err := c.delete(url)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif resp.StatusCode != http.StatusNoContent {\n\t\treturn parseError(resp.Body)\n\t}\n\n\treturn nil\n}", "func (a *ImageApiService) DeleteItemImage(ctx _context.Context, itemId string, imageType ImageType, imageIndex int32) (*_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/Items/{itemId}/Images/{imageType}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"itemId\"+\"}\", _neturl.QueryEscape(parameterToString(itemId, \"\")) , -1)\n\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"imageType\"+\"}\", _neturl.QueryEscape(parameterToString(imageType, \"\")) , -1)\n\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"imageIndex\"+\"}\", _neturl.QueryEscape(parameterToString(imageIndex, \"\")) , -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\", \"application/json; profile=CamelCase\", \"application/json; profile=PascalCase\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tif ctx != nil {\n\t\t// API Key Authentication\n\t\tif auth, ok := ctx.Value(ContextAPIKey).(APIKey); ok {\n\t\t\tvar key string\n\t\t\tif auth.Prefix != \"\" {\n\t\t\t\tkey = auth.Prefix + \" \" + auth.Key\n\t\t\t} else {\n\t\t\t\tkey = auth.Key\n\t\t\t}\n\t\t\tlocalVarHeaderParams[\"X-Emby-Authorization\"] = key\n\t\t}\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v ProblemDetails\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t}\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func (db *ImageDB) DeleteImageByID(id primitive.ObjectID) error {\n\tbucket, _ := gridfs.NewBucket(\n\t\tdb.database,\n\t)\n\treturn bucket.Delete(id)\n}", "func (qs *QueryServer) DeleteContainerImage(id uint64) error {\n\tciKey := qs.containerImageKey(id)\n\terr := qs.SoftDelete(ciKey)\n\treturn err\n}", "func DeleteImages(c echo.Context) error {\n\terr := model.DeleteImages()\n\tif err != nil {\n\t\treturn c.String(http.StatusExpectationFailed, \"failed\")\n\t}\n\treturn c.String(http.StatusOK, \"deleted\")\n}", "func (c *Client) DeleteImagesWithContext(ctx context.Context, request *DeleteImagesRequest) (response *DeleteImagesResponse, err error) {\n if request == nil {\n request = NewDeleteImagesRequest()\n }\n \n if c.GetCredential() == nil {\n return nil, errors.New(\"DeleteImages require credential\")\n }\n\n request.SetContext(ctx)\n \n response = NewDeleteImagesResponse()\n err = c.Send(request, response)\n return\n}", "func (c *FakeImageSignatures) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error {\n\t_, err := c.Fake.\n\t\tInvokes(testing.NewRootDeleteActionWithOptions(imagesignaturesResource, name, opts), &v1.ImageSignature{})\n\treturn err\n}", "func (client ArtifactsClient) deleteContainerImageSignature(ctx context.Context, request common.OCIRequest, binaryReqBody *common.OCIReadSeekCloser, extraHeaders map[string]string) (common.OCIResponse, error) {\n\n\thttpRequest, err := request.HTTPRequest(http.MethodDelete, \"/container/imageSignatures/{imageSignatureId}\", binaryReqBody, extraHeaders)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar response DeleteContainerImageSignatureResponse\n\tvar httpResponse *http.Response\n\thttpResponse, err = client.Call(ctx, &httpRequest)\n\tdefer common.CloseBodyIfValid(httpResponse)\n\tresponse.RawResponse = httpResponse\n\tif err != nil {\n\t\tapiReferenceLink := \"https://docs.oracle.com/iaas/api/#/en/registry/20160918/ContainerImageSignature/DeleteContainerImageSignature\"\n\t\terr = common.PostProcessServiceError(err, \"Artifacts\", \"DeleteContainerImageSignature\", apiReferenceLink)\n\t\treturn response, err\n\t}\n\n\terr = common.UnmarshalResponse(httpResponse, &response)\n\treturn response, err\n}", "func (o *Picture) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Picture provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), picturePrimaryKeyMapping)\n\tsql := \"DELETE FROM `pictures` WHERE `id`=?\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete from pictures\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (client *GalleryImageVersionsClient) delete(ctx context.Context, resourceGroupName string, galleryName string, galleryImageName string, galleryImageVersionName string, options *GalleryImageVersionsBeginDeleteOptions) (*azcore.Response, error) {\n\treq, err := client.deleteCreateRequest(ctx, resourceGroupName, galleryName, galleryImageName, galleryImageVersionName, options)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresp, err := client.con.Pipeline().Do(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif !resp.HasStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent) {\n\t\treturn nil, client.deleteHandleError(resp)\n\t}\n\treturn resp, nil\n}", "func (r *SoftwareImageResource) Delete(id string) error {\n\tif err := r.c.ModQuery(\"DELETE\", BasePath+SoftwareImageEndpoint+\"/\"+id, nil); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (_m *MockECRAPI) BatchDeleteImage(_param0 *ecr.BatchDeleteImageInput) (*ecr.BatchDeleteImageOutput, error) {\n\tret := _m.ctrl.Call(_m, \"BatchDeleteImage\", _param0)\n\tret0, _ := ret[0].(*ecr.BatchDeleteImageOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func DeleteImageWithDelay(msg *discordgo.Message, delay time.Duration) {\n\tif msg == nil {\n\t\treturn\n\t}\n\n\ttime.Sleep(delay)\n\tcache.GetDiscordSession().ChannelMessageDelete(msg.ChannelID, msg.ID)\n}", "func RunImagesDelete(ns string, config doit.Config, out io.Writer, args []string) error {\n\tclient := config.GetGodoClient()\n\n\tif len(args) != 1 {\n\t\treturn doit.NewMissingArgsErr(ns)\n\t}\n\n\tid, err := strconv.Atoi(args[0])\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = client.Images.Delete(id)\n\treturn err\n}", "func NewDeleteImageDefault(code int) *DeleteImageDefault {\n\tif code <= 0 {\n\t\tcode = 500\n\t}\n\n\treturn &DeleteImageDefault{\n\t\t_statusCode: code,\n\t}\n}", "func DeleteImageMeta(key string) error {\n\terr := imageDb.Delete([]byte(key), pebble.Sync)\n\n\treturn err\n}", "func (c *Client) DeleteImages(request *DeleteImagesRequest) (response *DeleteImagesResponse, err error) {\n return c.DeleteImagesWithContext(context.Background(), request)\n}", "func (m *ItemPhotoRequestBuilder) Delete(ctx context.Context, requestConfiguration *ItemPhotoRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": ia572726a95efa92ddd544552cd950653dc691023836923576b2f4bf716cf204a.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": ia572726a95efa92ddd544552cd950653dc691023836923576b2f4bf716cf204a.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func (s *DataStore) DeleteBackingImage(name string) error {\n\tpropagation := metav1.DeletePropagationForeground\n\treturn s.lhClient.LonghornV1beta2().BackingImages(s.namespace).Delete(context.TODO(), name, metav1.DeleteOptions{PropagationPolicy: &propagation})\n}", "func (client ArtifactsClient) DeleteContainerImage(ctx context.Context, request DeleteContainerImageRequest) (response DeleteContainerImageResponse, err error) {\n\tvar ociResponse common.OCIResponse\n\tpolicy := common.NoRetryPolicy()\n\tif client.RetryPolicy() != nil {\n\t\tpolicy = *client.RetryPolicy()\n\t}\n\tif request.RetryPolicy() != nil {\n\t\tpolicy = *request.RetryPolicy()\n\t}\n\tociResponse, err = common.Retry(ctx, request, client.deleteContainerImage, policy)\n\tif err != nil {\n\t\tif ociResponse != nil {\n\t\t\tif httpResponse := ociResponse.HTTPResponse(); httpResponse != nil {\n\t\t\t\topcRequestId := httpResponse.Header.Get(\"opc-request-id\")\n\t\t\t\tresponse = DeleteContainerImageResponse{RawResponse: httpResponse, OpcRequestId: &opcRequestId}\n\t\t\t} else {\n\t\t\t\tresponse = DeleteContainerImageResponse{}\n\t\t\t}\n\t\t}\n\t\treturn\n\t}\n\tif convertedResponse, ok := ociResponse.(DeleteContainerImageResponse); ok {\n\t\tresponse = convertedResponse\n\t} else {\n\t\terr = fmt.Errorf(\"failed to convert OCIResponse into DeleteContainerImageResponse\")\n\t}\n\treturn\n}", "func (m *DeleteImageTagModel) Delete() (err error) {\n\n\t// check model validity\n\tif !m.IsValid() {\n\t\treturn errors.New(\"DeleteImageTagModel is not valid\")\n\t}\n\n\t// Get Database handle\n\tdbase, err := db.GetDb()\n\tif err != nil {\n\t\treturn\n\t}\n\n\tps1, err := dbase.Prepare(`DELETE tm FROM tagmap AS tm\n INNER JOIN tags ON tm.tag_id = tags.tag_id\n WHERE image_id = ? AND tm.tag_id = ? AND ib_id = ?`)\n\tif err != nil {\n\t\treturn\n\t}\n\tdefer ps1.Close()\n\n\t_, err = ps1.Exec(m.Image, m.Tag, m.Ib)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n\n}", "func adminImagesDelete(w http.ResponseWriter, r *http.Request) {\n\ttype response struct {\n\t\tOk bool\n\t}\n\tid := r.URL.Path\n\tresp := response{}\n\n\t_, err := db.Exec(`DELETE FROM images WHERE appointment_id = $1`, id)\n\tif err != nil {\n\t\tjson.NewEncoder(w).Encode(resp)\n\t\treturn\n\t}\n\n\t// set ok and send\n\tresp.Ok = true\n\terr = json.NewEncoder(w).Encode(resp)\n\tif err != nil {\n\t\tlog.Println(\"appointment img delete:\", err)\n\t}\n}", "func (s *DataStore) DeleteEngineImage(name string) error {\n\tpropagation := metav1.DeletePropagationForeground\n\treturn s.lhClient.LonghornV1beta2().EngineImages(s.namespace).Delete(context.TODO(), name, metav1.DeleteOptions{PropagationPolicy: &propagation})\n}", "func (cmd *imagesCmd) RunCleanupImages(f factory.Factory, cobraCmd *cobra.Command, args []string) error {\n\t// Set config root\n\tlog := f.GetLog()\n\tconfigLoader := f.NewConfigLoader(cmd.ToConfigOptions(), log)\n\tkubeConfigLoader := f.NewKubeConfigLoader()\n\tconfigExists, err := configLoader.SetDevSpaceRoot()\n\tif err != nil {\n\t\treturn err\n\t}\n\tif !configExists {\n\t\treturn errors.New(message.ConfigNotFound)\n\t}\n\n\t// Get active context\n\tkubeContext, err := kubeConfigLoader.GetCurrentContext()\n\tif err != nil {\n\t\treturn err\n\t}\n\tif cmd.KubeContext != \"\" {\n\t\tkubeContext = cmd.KubeContext\n\t}\n\n\t// Create docker client\n\tclient, err := docker.NewClientWithMinikube(kubeContext, true, log)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Load config\n\tconfig, err := configLoader.Load()\n\tif err != nil {\n\t\treturn err\n\t}\n\tif config.Images == nil || len(config.Images) == 0 {\n\t\tlog.Done(\"No images found in config to delete\")\n\t\treturn nil\n\t}\n\n\t_, err = client.Ping(context.Background())\n\tif err != nil {\n\t\treturn errors.Errorf(\"Docker seems to be not running: %v\", err)\n\t}\n\n\tdefer log.StopWait()\n\n\t// Delete all images\n\tfor _, imageConfig := range config.Images {\n\t\tlog.StartWait(\"Deleting local image \" + imageConfig.Image)\n\n\t\tresponse, err := client.DeleteImageByName(imageConfig.Image, log)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tfor _, t := range response {\n\t\t\tif t.Deleted != \"\" {\n\t\t\t\tlog.Donef(\"Deleted %s\", t.Deleted)\n\t\t\t} else if t.Untagged != \"\" {\n\t\t\t\tlog.Donef(\"Untagged %s\", t.Untagged)\n\t\t\t}\n\t\t}\n\t}\n\n\tlog.StartWait(\"Deleting local dangling images\")\n\n\t// Cleanup dangling images aswell\n\tfor {\n\t\tresponse, err := client.DeleteImageByFilter(filters.NewArgs(filters.Arg(\"dangling\", \"true\")), log)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tfor _, t := range response {\n\t\t\tif t.Deleted != \"\" {\n\t\t\t\tlog.Donef(\"Deleted %s\", t.Deleted)\n\t\t\t} else if t.Untagged != \"\" {\n\t\t\t\tlog.Donef(\"Untagged %s\", t.Untagged)\n\t\t\t}\n\t\t}\n\n\t\tif len(response) == 0 {\n\t\t\tbreak\n\t\t}\n\t}\n\n\tlog.StopWait()\n\tlog.Donef(\"Successfully cleaned up images\")\n\treturn nil\n}", "func (m *CompaniesItemSalesCreditMemosItemCustomerPicturePictureItemRequestBuilder) Delete(ctx context.Context, requestConfiguration *CompaniesItemSalesCreditMemosItemCustomerPicturePictureItemRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func (api *Api) RemoveImage(ctx *gin.Context) {\n\tcraneContext, _ := ctx.Get(\"craneContext\")\n\timageID := ctx.Param(\"image_id\")\n\tif err := api.GetDockerClient().RemoveImage(craneContext.(context.Context), imageID); err != nil {\n\t\tlog.Error(\"RemoveImage got error: \", err)\n\t\thttpresponse.Error(ctx, err)\n\t\treturn\n\t}\n\n\thttpresponse.Ok(ctx, \"success\")\n\treturn\n}", "func (s *Service) DeleteProfileImage(userID string) *DeleteProfileImageOp {\n\treturn &DeleteProfileImageOp{\n\t\tCredential: s.credential,\n\t\tMethod: \"DELETE\",\n\t\tPath: strings.Join([]string{\"users\", userID, \"profile\", \"image\"}, \"/\"),\n\t\tQueryOpts: make(url.Values),\n\t\tVersion: esign.APIv21,\n\t}\n}", "func (a *ImageApiService) DeleteItemImage2(ctx _context.Context, itemId string, imageType ImageType, imageIndex int32) (*_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/Items/{itemId}/Images/{imageType}/{imageIndex}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"itemId\"+\"}\", _neturl.QueryEscape(parameterToString(itemId, \"\")) , -1)\n\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"imageType\"+\"}\", _neturl.QueryEscape(parameterToString(imageType, \"\")) , -1)\n\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"imageIndex\"+\"}\", _neturl.QueryEscape(parameterToString(imageIndex, \"\")) , -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\", \"application/json; profile=CamelCase\", \"application/json; profile=PascalCase\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tif ctx != nil {\n\t\t// API Key Authentication\n\t\tif auth, ok := ctx.Value(ContextAPIKey).(APIKey); ok {\n\t\t\tvar key string\n\t\t\tif auth.Prefix != \"\" {\n\t\t\t\tkey = auth.Prefix + \" \" + auth.Key\n\t\t\t} else {\n\t\t\t\tkey = auth.Key\n\t\t\t}\n\t\t\tlocalVarHeaderParams[\"X-Emby-Authorization\"] = key\n\t\t}\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v ProblemDetails\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t}\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func (s *AvatarsService) Delete (ctx context.Context, entityType string, owningObjectID string, avatarID int64) (*http.Response, error) {\n\tendpoint := fmt.Sprintf(\"universal_avatar/type/%v/owner/%v/avatar/%v\", entityType, owningObjectID, avatarID)\n\treq, err := s.client.NewRequest(\"DELETE\", endpoint, nil, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresp, err := s.client.Do(ctx, req, nil)\n\tif err != nil {\n\t\treturn resp, err\n\t}\n\n\treturn resp, err\n}", "func (c *Client) DeleteImageSpriteTemplate(request *DeleteImageSpriteTemplateRequest) (response *DeleteImageSpriteTemplateResponse, err error) {\n if request == nil {\n request = NewDeleteImageSpriteTemplateRequest()\n }\n response = NewDeleteImageSpriteTemplateResponse()\n err = c.Send(request, response)\n return\n}", "func (handle *DBHandle) DeleteImageCategory(id uint) bool {\n\treturn handle.deleteWithID(id, ImageCategory{})\n}", "func (d *Driver) Delete(ctx context.Context, f *functions.Function) error {\n\tspan, ctx := trace.Trace(ctx, \"\")\n\tdefer span.Finish()\n\tif err := d.deleteContainers(ctx, f, true); err != nil {\n\t\t// no error wrapping, delete already does that\n\t\treturn err\n\t}\n\treturn nil\n}", "func (mock *RequestsServiceMock) DeleteImageCalls() []struct {\n\tCtx context.Context\n\tR *requests.Image\n} {\n\tvar calls []struct {\n\t\tCtx context.Context\n\t\tR *requests.Image\n\t}\n\tlockRequestsServiceMockDeleteImage.RLock()\n\tcalls = mock.calls.DeleteImage\n\tlockRequestsServiceMockDeleteImage.RUnlock()\n\treturn calls\n}", "func (a *ImagesApiService) DeleteImageByImageId(ctx context.Context, imageId string) ApiDeleteImageByImageIdRequest {\n\treturn ApiDeleteImageByImageIdRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\timageId: imageId,\n\t}\n}" ]
[ "0.74038446", "0.73479116", "0.73390186", "0.72878873", "0.7150099", "0.7147888", "0.7128131", "0.7102309", "0.7073419", "0.70429873", "0.6996273", "0.69926614", "0.69794863", "0.6975652", "0.6958761", "0.69291633", "0.6914108", "0.68879896", "0.68820465", "0.6870168", "0.6868291", "0.68612295", "0.6848216", "0.6828671", "0.68087435", "0.67714435", "0.67695266", "0.6750623", "0.6741595", "0.6674301", "0.6646622", "0.663669", "0.66359156", "0.66142297", "0.6610889", "0.66084063", "0.6603519", "0.6542015", "0.6489898", "0.6481269", "0.64095604", "0.6372039", "0.6370643", "0.63424736", "0.63116676", "0.6287503", "0.6202354", "0.61830646", "0.61645514", "0.6162834", "0.61429334", "0.61376154", "0.6093238", "0.60620487", "0.60593265", "0.6040361", "0.60268277", "0.6006125", "0.5964864", "0.59558463", "0.5950708", "0.5940145", "0.5899155", "0.58475494", "0.58474684", "0.584745", "0.5839369", "0.5836055", "0.582964", "0.5824855", "0.5817679", "0.5815235", "0.58092564", "0.5795232", "0.5795031", "0.5792226", "0.5762828", "0.57601243", "0.57317644", "0.5724371", "0.57204753", "0.570733", "0.56570965", "0.56556726", "0.56414735", "0.56368035", "0.5629495", "0.5628049", "0.5621342", "0.561829", "0.5615644", "0.5614857", "0.5614015", "0.5612536", "0.55943114", "0.55838186", "0.5563482", "0.5553599", "0.55505395", "0.5549382" ]
0.707854
8
DeleteInstance uses the override method DeleteInstanceFn or the real implementation.
func (c *TestClient) DeleteInstance(project, zone, name string) error { if c.DeleteInstanceFn != nil { return c.DeleteInstanceFn(project, zone, name) } return c.client.DeleteInstance(project, zone, name) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (p *OnPrem) DeleteInstance(ctx *Context, instancename string) error {\n\treturn fmt.Errorf(\"Operation not supported\")\n}", "func (p *OnPrem) DeleteInstance(ctx *Context, instancename string) error {\n\n\tpid, err := strconv.Atoi(instancename)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\n\t// yolo\n\terr = sysKill(pid)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\n\topshome := GetOpsHome()\n\tipath := path.Join(opshome, \"instances\", instancename)\n\terr = os.Remove(ipath)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *InstanceManagerClient) InstanceDelete(backendStoreDriver longhorn.BackendStoreDriverType, name, kind, diskUUID string, cleanupRequired bool) (err error) {\n\tif c.GetAPIVersion() < 4 {\n\t\t/* Fall back to the old way of deleting process */\n\t\t_, err = c.processManagerGrpcClient.ProcessDelete(name)\n\t} else {\n\t\t_, err = c.instanceServiceGrpcClient.InstanceDelete(string(backendStoreDriver), name, kind, diskUUID, cleanupRequired)\n\t}\n\n\treturn err\n}", "func (c *MockAzureCloud) DeleteInstance(i *cloudinstances.CloudInstance) error {\n\treturn errors.New(\"DeleteInstance not implemented on azureCloud\")\n}", "func (c *Client) DeleteInstance(name string) error {\n\tid, err := c.getDeviceID(name)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = c.client.Devices.Delete(id, true)\n\treturn err\n}", "func deleteInstance(params martini.Params, r render.Render) {\n\tinstanceName := params[\"name\"]\n\n\tif !instanceExists(instanceName) {\n\t\tfmt.Println(\"Instance with specified name does not exist in provision table\")\n\t\tr.Text(400, \"Bad Request\")\n\t\treturn\n\t}\n\n\tsvc := neptune.New(session.New(&aws.Config{\n\t\tRegion: aws.String(os.Getenv(\"REGION\")),\n\t}))\n\n\tinstanceParamsDelete := &neptune.DeleteDBInstanceInput{\n\t\tDBInstanceIdentifier: aws.String(instanceName),\n\t\tSkipFinalSnapshot: aws.Bool(true),\n\t}\n\n\tclusterParamsDelete := &neptune.DeleteDBClusterInput{\n\t\tDBClusterIdentifier: aws.String(instanceName),\n\t\tSkipFinalSnapshot: aws.Bool(true),\n\t}\n\n\tinstanceResp, instanceErr := svc.DeleteDBInstance(instanceParamsDelete)\n\tname := *instanceParamsDelete.DBInstanceIdentifier\n\tif instanceErr != nil {\n\t\tfmt.Println(instanceErr.Error())\n\t\toutput500Error(r, instanceErr)\n\t\treturn\n\t}\n\tfmt.Println(\"Deletion in progress for instance \" + *instanceResp.DBInstance.DBInstanceIdentifier)\n\n\tclusterResp, clusterErr := svc.DeleteDBCluster(clusterParamsDelete)\n\tif clusterErr != nil {\n\t\tfmt.Println(instanceErr.Error())\n\t\toutput500Error(r, clusterErr)\n\t\treturn\n\t}\n\tfmt.Println(\"Deletion in progress for cluster \" + *clusterResp.DBCluster.DBClusterIdentifier)\n\n\t_, err := pool.Exec(\"DELETE FROM provision WHERE name=$1\", name)\n\tif err != nil {\n\t\tfmt.Println(err.Error())\n\t\toutput500Error(r, err)\n\t\treturn\n\t}\n\n\tr.JSON(200, map[string]string{\"Response\": \"Instance deletion in progress\"})\n\n\tusername := *instanceParamsDelete.DBInstanceIdentifier\n\tdeleteUserPolicy(username)\n\tdeleteAccessKey(username)\n\tdeleteUser(username)\n\n}", "func (c *GRPCClient) DeleteInstance(ctx context.Context, cloudID string) error {\n\treq := &pb.DeleteInstanceRequest{\n\t\tCloudId: cloudID,\n\t}\n\t_, err := c.client.DeleteInstance(ctx, req)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to DeleteInstance: %w\", err)\n\t}\n\n\treturn nil\n}", "func TestDeleteInstance(t *testing.T) {\n\tvar reason payloads.StartFailureReason\n\n\tclient, instances := testStartWorkload(t, 1, false, reason)\n\tdefer client.Shutdown()\n\n\tsendStatsCmd(client, t)\n\n\tserverCh := server.AddCmdChan(ssntp.DELETE)\n\n\terr := ctl.deleteInstance(instances[0].ID)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tresult, err := server.GetCmdChanResult(serverCh, ssntp.DELETE)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\tif result.InstanceUUID != instances[0].ID {\n\t\tt.Fatal(\"Did not get correct Instance ID\")\n\t}\n}", "func (i *InstanceServiceHandler) Delete(ctx context.Context, instanceID string) error {\n\turi := fmt.Sprintf(\"%s/%s\", instancePath, instanceID)\n\n\treq, err := i.client.NewRequest(ctx, http.MethodDelete, uri, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn i.client.DoWithContext(ctx, req, nil)\n}", "func (s *API) DeleteInstance(req *DeleteInstanceRequest, opts ...scw.RequestOption) (*Instance, error) {\n\tvar err error\n\n\tif req.Region == \"\" {\n\t\tdefaultRegion, _ := s.client.GetDefaultRegion()\n\t\treq.Region = defaultRegion\n\t}\n\n\tif fmt.Sprint(req.Region) == \"\" {\n\t\treturn nil, errors.New(\"field Region cannot be empty in request\")\n\t}\n\n\tif fmt.Sprint(req.InstanceID) == \"\" {\n\t\treturn nil, errors.New(\"field InstanceID cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"DELETE\",\n\t\tPath: \"/rdb/v1/regions/\" + fmt.Sprint(req.Region) + \"/instances/\" + fmt.Sprint(req.InstanceID) + \"\",\n\t\tHeaders: http.Header{},\n\t}\n\n\tvar resp Instance\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func (p *ProxMox) DeleteInstance(ctx *lepton.Context, instanceID string) error {\n\n\treq, err := http.NewRequest(\"DELETE\", p.apiURL+\"/api2/json/nodes/\"+p.nodeNAME+\"/qemu/\"+instanceID, nil)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\ttr := &http.Transport{\n\t\tTLSClientConfig: &tls.Config{InsecureSkipVerify: true},\n\t}\n\tclient := &http.Client{Transport: tr}\n\n\treq.Header.Add(\"Authorization\", \"PVEAPIToken=\"+p.tokenID+\"=\"+p.secret)\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\t_, err = io.ReadAll(resp.Body)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\n\t}\n\n\treturn err\n\n}", "func (tx *Tx) Delete(instance Instance) error {\n\ttx.Process(func(iter *Iter) {\n\t\tif iter.MatchInstance(tx.resource, instance) {\n\t\t\titer.Delete()\n\t\t}\n\t})\n\treturn nil\n}", "func (m *InstancesManager) DeleteInstance(instanceID string) {\n\tm.mutex.Lock()\n\tdefer m.mutex.Unlock()\n\tm.instances.Delete(instanceID)\n}", "func (client *Client) DeleteInstance(request *DeleteInstanceRequest) (_result *DeleteInstanceResponse, _err error) {\n\truntime := &util.RuntimeOptions{}\n\t_result = &DeleteInstanceResponse{}\n\t_body, _err := client.DeleteInstanceWithOptions(request, runtime)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\t_result = _body\n\treturn _result, _err\n}", "func (awsI *Ec2Instance) DeleteInstance() error {\n\tinstanceWorkingDir := path.Join(awsI.DeployDir, \"etc\", \"terraform\", \"instance\")\n\tinstanceConfPath := path.Join(instanceWorkingDir, \"instance.json\")\n\tif !sdutils.PathExists(instanceConfPath) {\n\t\treturn errors.New(\"There is no configured instance\")\n\t}\n\tterraformPath, err := GetTerraformPath(awsI.Ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\tcmdArray := []string{terraformPath, \"destroy\", \"-force\", \"-var-file\", instanceConfPath}\n\tcmd := exec.Cmd{\n\t\tPath: cmdArray[0],\n\t\tArgs: cmdArray,\n\t\tDir: instanceWorkingDir,\n\t}\n\tawsI.Ctx.Logf(sdutils.INFO, \"Running terraform...\\n\")\n\tspin := sdutils.NewSpinner(awsI.Ctx, 1, \"Deleting the instance VMs\")\n\t_, err = sdutils.RunCommand(awsI.Ctx, cmd, volumeLineScanner, spin)\n\tif err != nil {\n\t\treturn err\n\t}\n\tos.Remove(instanceConfPath)\n\tawsI.Ctx.ConsoleLog(1, \"Successfully destroyed the instance.\\n\")\n\treturn nil\n}", "func DeleteInstance(c GCECloud, instanceSelfLink string) error {\n\tklog.V(2).Infof(\"Deleting GCE Instance %s\", instanceSelfLink)\n\tu, err := ParseGoogleCloudURL(instanceSelfLink)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\top, err := c.Compute().Instances().Delete(u.Project, u.Zone, u.Name)\n\tif err != nil {\n\t\tif IsNotFound(err) {\n\t\t\tklog.Infof(\"Instance not found, assuming deleted: %q\", instanceSelfLink)\n\t\t\treturn nil\n\t\t}\n\t\treturn fmt.Errorf(\"error deleting Instance %s: %v\", instanceSelfLink, err)\n\t}\n\n\treturn c.WaitForOp(op)\n}", "func (g GCPClient) DeleteInstance(instance, zone string, wait bool) error {\n\tvar notFound bool\n\top, err := g.compute.Instances.Delete(g.projectName, zone, instance).Do()\n\tif err != nil {\n\t\tif _, ok := err.(*googleapi.Error); !ok {\n\t\t\treturn err\n\t\t}\n\t\tif err.(*googleapi.Error).Code != 404 {\n\t\t\treturn err\n\t\t}\n\t\tnotFound = true\n\t}\n\tif !notFound && wait {\n\t\tlog.Infof(\"Deleting existing instance...\")\n\t\tif err := g.pollZoneOperationStatus(op.Name, zone); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tlog.Infof(\"Instance %s deleted\", instance)\n\t}\n\treturn nil\n}", "func (s *Service) Delete(ctx context.Context) error {\n\tlog := log.FromContext(ctx)\n\tlog.Info(\"Deleting instance resources\")\n\tinstanceSpec := s.scope.InstanceSpec(log)\n\tinstanceName := instanceSpec.Name\n\tinstanceKey := meta.ZonalKey(instanceName, s.scope.Zone())\n\tlog.V(2).Info(\"Looking for instance before deleting\", \"name\", instanceName, \"zone\", s.scope.Zone())\n\tinstance, err := s.instances.Get(ctx, instanceKey)\n\tif err != nil {\n\t\tif !gcperrors.IsNotFound(err) {\n\t\t\tlog.Error(err, \"Error looking for instance before deleting\", \"name\", instanceName)\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}\n\n\tif s.scope.IsControlPlane() {\n\t\tif err := s.deregisterControlPlaneInstance(ctx, instance); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tlog.V(2).Info(\"Deleting instance\", \"name\", instanceName, \"zone\", s.scope.Zone())\n\treturn gcperrors.IgnoreNotFound(s.instances.Delete(ctx, instanceKey))\n}", "func (f *FakeInstance) Delete(_ context.Context, _ string) error {\n\tpanic(\"implement me\")\n}", "func DeleteInstance(instance *spotcluster.Instance,\n\tpool *spotcluster.Pool) error {\n\n\tif pool == nil {\n\t\treturn errors.New(\"got nil pool object\")\n\t}\n\n\tif instance == nil {\n\t\treturn errors.New(\"got nil instance object\")\n\t}\n\n\tclient := godo.NewFromToken(pool.ProviderSpec.DigitalOcean.APIKey)\n\tif client == nil {\n\t\treturn errors.New(\"got nil godo client\")\n\t}\n\n\tdoc := Client{\n\t\tProvider: client,\n\t}\n\n\treturn doc.Delete(string(instance.GetUID()))\n}", "func (m *InstanceManager) Delete(ctx context.Context, obj runtime.Object) (bool, error) {\n\tinstance, err := convertInstance(obj)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\tregion := scw.Region(instance.Spec.Region)\n\n\tresourceID := instance.Spec.InstanceID\n\tif resourceID == \"\" {\n\t\treturn true, nil\n\t}\n\n\t_, err = m.API.DeleteInstance(&rdb.DeleteInstanceRequest{\n\t\tRegion: region,\n\t\tInstanceID: resourceID,\n\t})\n\tif err != nil {\n\t\tif _, ok := err.(*scw.ResourceNotFoundError); ok {\n\t\t\treturn true, nil\n\t\t}\n\t\treturn false, err\n\t}\n\n\t//instance.Status.Status = strcase.ToCamel(instanceResp.Status.String())\n\n\treturn false, nil\n}", "func (client *Client) DeleteInstanceWithOptions(request *DeleteInstanceRequest, runtime *util.RuntimeOptions) (_result *DeleteInstanceResponse, _err error) {\n\t_err = util.ValidateModel(request)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\tquery := map[string]interface{}{}\n\tif !tea.BoolValue(util.IsUnset(request.GlobalInstanceId)) {\n\t\tquery[\"GlobalInstanceId\"] = request.GlobalInstanceId\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.InstanceId)) {\n\t\tquery[\"InstanceId\"] = request.InstanceId\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.OwnerAccount)) {\n\t\tquery[\"OwnerAccount\"] = request.OwnerAccount\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.OwnerId)) {\n\t\tquery[\"OwnerId\"] = request.OwnerId\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.ResourceOwnerAccount)) {\n\t\tquery[\"ResourceOwnerAccount\"] = request.ResourceOwnerAccount\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.ResourceOwnerId)) {\n\t\tquery[\"ResourceOwnerId\"] = request.ResourceOwnerId\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.SecurityToken)) {\n\t\tquery[\"SecurityToken\"] = request.SecurityToken\n\t}\n\n\treq := &openapi.OpenApiRequest{\n\t\tQuery: openapiutil.Query(query),\n\t}\n\tparams := &openapi.Params{\n\t\tAction: tea.String(\"DeleteInstance\"),\n\t\tVersion: tea.String(\"2015-01-01\"),\n\t\tProtocol: tea.String(\"HTTPS\"),\n\t\tPathname: tea.String(\"/\"),\n\t\tMethod: tea.String(\"POST\"),\n\t\tAuthType: tea.String(\"AK\"),\n\t\tStyle: tea.String(\"RPC\"),\n\t\tReqBodyType: tea.String(\"formData\"),\n\t\tBodyType: tea.String(\"json\"),\n\t}\n\t_result = &DeleteInstanceResponse{}\n\t_body, _err := client.CallApi(params, req, runtime)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\t_err = tea.Convert(_body, &_result)\n\treturn _result, _err\n}", "func (a ProcessInstanceApi) DeleteInstance(processId string, instanceId string) (*ResultSuccess, *APIResponse, error) {\n\n\tvar localVarHttpMethod = strings.ToUpper(\"Delete\")\n\t// create path and map variables\n\tlocalVarPath := a.Configuration.BasePath + \"/processes/{process_id}/instances/{instance_id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"process_id\"+\"}\", fmt.Sprintf(\"%v\", processId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"instance_id\"+\"}\", fmt.Sprintf(\"%v\", instanceId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := make(map[string]string)\n\tvar localVarPostBody interface{}\n\tvar localVarFileName string\n\tvar localVarFileBytes []byte\n\t// authentication '(PasswordGrant)' required\n\t// set key with prefix in header\n\tlocalVarHeaderParams[\"Authorization\"] = a.Configuration.GetAPIKeyWithPrefix(\"Authorization\")\n\t// add default headers if any\n\tfor key := range a.Configuration.DefaultHeader {\n\t\tlocalVarHeaderParams[key] = a.Configuration.DefaultHeader[key]\n\t}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{ \"application/json\", }\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := a.Configuration.APIClient.SelectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\n\t\t\"application/json\",\n\t\t}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := a.Configuration.APIClient.SelectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tvar successPayload = new(ResultSuccess)\n\tlocalVarHttpResponse, err := a.Configuration.APIClient.CallAPI(localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\n\tvar localVarURL, _ = url.Parse(localVarPath)\n\tlocalVarURL.RawQuery = localVarQueryParams.Encode()\n\tvar localVarAPIResponse = &APIResponse{Operation: \"DeleteInstance\", Method: localVarHttpMethod, RequestURL: localVarURL.String()}\n\tif localVarHttpResponse != nil {\n\t\tlocalVarAPIResponse.Response = localVarHttpResponse.RawResponse\n\t\tlocalVarAPIResponse.Payload = localVarHttpResponse.Body()\n\t}\n\n\tif err != nil {\n\t\treturn successPayload, localVarAPIResponse, err\n\t}\n\terr = json.Unmarshal(localVarHttpResponse.Body(), &successPayload)\n\treturn successPayload, localVarAPIResponse, err\n}", "func (a *AssembliesApiService) DeleteInstance(ctx _context.Context, did string, eid string, wid string, nid string) (*_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/api/assemblies/d/{did}/w/{wid}/e/{eid}/instance/nodeid/{nid}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"did\"+\"}\", _neturl.QueryEscape(parameterToString(did, \"\")) , -1)\n\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"eid\"+\"}\", _neturl.QueryEscape(parameterToString(eid, \"\")) , -1)\n\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"wid\"+\"}\", _neturl.QueryEscape(parameterToString(wid, \"\")) , -1)\n\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"nid\"+\"}\", _neturl.QueryEscape(parameterToString(nid, \"\")) , -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/vnd.onshape.v1+json;charset=UTF-8;qs=0.1\", \"application/json;charset=UTF-8; qs=0.09\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func (c *Client) DeleteInstance(instanceId string) error {\n\treturn DeleteInstance(c, instanceId)\n}", "func (client BaseClient) DeleteFeatureInstanceResponder(resp *http.Response) (result FeatureInstance, err error) {\n\terr = autorest.Respond(\n\t\tresp,\n\t\tclient.ByInspecting(),\n\t\tazure.WithErrorUnlessStatusCode(http.StatusOK),\n\t\tautorest.ByUnmarshallingJSON(&result),\n\t\tautorest.ByClosing())\n\tresult.Response = autorest.Response{Response: resp}\n\treturn\n}", "func (client BaseClient) DeleteFeatureInstance(ctx context.Context, featureName string, featureVersion string, instanceName string) (result FeatureInstance, err error) {\n\treq, err := client.DeleteFeatureInstancePreparer(ctx, featureName, featureVersion, instanceName)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"DeleteFeatureInstance\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresp, err := client.DeleteFeatureInstanceSender(req)\n\tif err != nil {\n\t\tresult.Response = autorest.Response{Response: resp}\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"DeleteFeatureInstance\", resp, \"Failure sending request\")\n\t\treturn\n\t}\n\n\tresult, err = client.DeleteFeatureInstanceResponder(resp)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"DeleteFeatureInstance\", resp, \"Failure responding to request\")\n\t}\n\n\treturn\n}", "func (c *TestClient) DeleteTargetInstance(project, zone, name string) error {\n\tif c.DeleteTargetInstanceFn != nil {\n\t\treturn c.DeleteTargetInstanceFn(project, zone, name)\n\t}\n\treturn c.client.DeleteTargetInstance(project, zone, name)\n}", "func TestDelete(t *testing.T) {\n\tRunWithInstance(func(instance *Instance) {\n\t\tInsertFixtures(instance, []EntryFixture{\n\t\t\t{Name: \"int\", Value: \"2891\", ValueType: 1},\n\t\t\t{Name: \"string\", Value: \"hello world!\", ValueType: 3},\n\t\t})\n\n\t\tif err := instance.Delete(\"int\"); err != nil {\n\t\t\tt.Error(\"Instance.Delete: got error:\\n\", err)\n\t\t}\n\n\t\tif err := panicked(func() { instance.MustDelete(\"string\") }); err != nil {\n\t\t\tt.Error(\"Instance.MustDelete: got panic:\\n\", err)\n\t\t}\n\n\t\tif err := instance.Delete(\"foo\"); err == nil {\n\t\t\tt.Error(\"Instance.Delete: expected error with non-existent entry\")\n\t\t} else if _, ok := err.(*ErrNoEntry); !ok {\n\t\t\tt.Error(\"Instance.Delete: expected error of type *ErrNoEntry\")\n\t\t}\n\n\t\tif err := panicked(func() { instance.MustDelete(\"foo\") }); err == nil {\n\t\t\tt.Error(\"Instance.MustDelete: expected panic with non-existent entry\")\n\t\t}\n\t})\n}", "func (c *FakeClient) Delete(instanceID string) error {\n\tc.DeleteCalled = true\n\treturn c.DeleteErr\n}", "func (client AccessGovernanceCPClient) deleteGovernanceInstance(ctx context.Context, request common.OCIRequest, binaryReqBody *common.OCIReadSeekCloser, extraHeaders map[string]string) (common.OCIResponse, error) {\n\n\thttpRequest, err := request.HTTPRequest(http.MethodDelete, \"/governanceInstances/{governanceInstanceId}\", binaryReqBody, extraHeaders)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar response DeleteGovernanceInstanceResponse\n\tvar httpResponse *http.Response\n\thttpResponse, err = client.Call(ctx, &httpRequest)\n\tdefer common.CloseBodyIfValid(httpResponse)\n\tresponse.RawResponse = httpResponse\n\tif err != nil {\n\t\tapiReferenceLink := \"https://docs.oracle.com/iaas/api/#/en/access-governance-cp/20220518/GovernanceInstance/DeleteGovernanceInstance\"\n\t\terr = common.PostProcessServiceError(err, \"AccessGovernanceCP\", \"DeleteGovernanceInstance\", apiReferenceLink)\n\t\treturn response, err\n\t}\n\n\terr = common.UnmarshalResponse(httpResponse, &response)\n\treturn response, err\n}", "func (client BaseClient) DeleteFeatureInstanceSender(req *http.Request) (*http.Response, error) {\n\treturn autorest.SendWithSender(client, req,\n\t\tautorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n}", "func DeleteInstanceTemplate(c GCECloud, selfLink string) error {\n\tklog.V(2).Infof(\"Deleting GCE InstanceTemplate %s\", selfLink)\n\tu, err := ParseGoogleCloudURL(selfLink)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\top, err := c.Compute().InstanceTemplates().Delete(u.Project, u.Name)\n\tif err != nil {\n\t\tif IsNotFound(err) {\n\t\t\tklog.Infof(\"instancetemplate not found, assuming deleted: %q\", selfLink)\n\t\t\treturn nil\n\t\t}\n\t\treturn fmt.Errorf(\"error deleting InstanceTemplate %s: %v\", selfLink, err)\n\t}\n\n\treturn c.WaitForOp(op)\n}", "func TestIPAddress_Instance_Delete(t *testing.T) {\n\tclient, instance, _, teardown, err := setupInstanceWithoutDisks(t, \"fixtures/TestIPAddress_Instance_Delete\")\n\tdefer teardown()\n\tif err != nil {\n\t\tt.Error(err)\n\t}\n\n\tip, err := client.AddInstanceIPAddress(context.TODO(), instance.ID, true)\n\tif err != nil {\n\t\tt.Fatalf(\"failed to allocate public IPv4 for instance (%d): %s\", instance.ID, err)\n\t}\n\n\ti, err := client.GetInstanceIPAddresses(context.TODO(), instance.ID)\n\tif err != nil {\n\t\tt.Fatalf(\"failed to get instance (%d) IP addresses: %s\", instance.ID, err)\n\t}\n\tif len(i.IPv4.Public) != 2 {\n\t\tt.Errorf(\"expected instance (%d) to have 2 public IPv4 addresses; got %d\", instance.ID, len(i.IPv4.Public))\n\t}\n\n\tif err := client.DeleteInstanceIPAddress(context.TODO(), instance.ID, ip.Address); err != nil {\n\t\tt.Fatalf(\"failed to delete instance (%d) public IPv4 address (%s): %s\", instance.ID, ip.Address, err)\n\t}\n\n\tif i, err = client.GetInstanceIPAddresses(context.TODO(), instance.ID); err != nil {\n\t\tt.Fatalf(\"failed to get instance (%d) IP addresses: %s\", instance.ID, err)\n\t}\n\tif len(i.IPv4.Public) != 1 {\n\t\tt.Errorf(\"expected instance (%d) to have 1 public IPv4 address; got %d\", instance.ID, len(i.IPv4.Public))\n\t}\n}", "func (c *Client) CloudDeleteInstance(projectID, instanceID string) error {\n\terr := c.Delete(queryEscape(\"/cloud/project/%s/instance/%s\", projectID, instanceID), nil)\n\tif apierror, ok := err.(*APIError); ok && apierror.Code == 404 {\n\t\terr = nil\n\t}\n\treturn err\n}", "func (client BaseClient) DeleteFeatureInstancePreparer(ctx context.Context, featureName string, featureVersion string, instanceName string) (*http.Request, error) {\n\tpathParameters := map[string]interface{}{\n\t\t\"featureName\": autorest.Encode(\"path\", featureName),\n\t\t\"featureVersion\": autorest.Encode(\"path\", featureVersion),\n\t\t\"instanceName\": autorest.Encode(\"path\", instanceName),\n\t}\n\n\tpreparer := autorest.CreatePreparer(\n\t\tautorest.AsDelete(),\n\t\tautorest.WithBaseURL(client.BaseURI),\n\t\tautorest.WithPathParameters(\"/api/features/instances/{featureName}/{featureVersion}/{instanceName}\", pathParameters))\n\treturn preparer.Prepare((&http.Request{}).WithContext(ctx))\n}", "func (e *EurekaConnection) DeregisterInstance(ins *Instance) error {\n\tslug := fmt.Sprintf(\"%s/%s/%s\", EurekaURLSlugs[\"Apps\"], ins.App, ins.Id())\n\treqURL := e.generateURL(slug)\n\tlog.Debugf(\"Deregistering instance with url %s\", reqURL)\n\n\trcode, err := deleteReq(reqURL)\n\tif err != nil {\n\t\tlog.Errorf(\"Could not complete deregistration, error: %s\", err.Error())\n\t\treturn err\n\t}\n\t// Eureka promises to return HTTP status code upon deregistration success, but fargo used to accept status code 204\n\t// here instead. Accommodate both for backward compatibility with any fake or proxy Eureka stand-ins.\n\tif rcode != http.StatusOK && rcode != http.StatusNoContent {\n\t\tlog.Warningf(\"HTTP returned %d deregistering Instance=%s App=%s\", rcode, ins.Id(), ins.App)\n\t\treturn &unsuccessfulHTTPResponse{rcode, \"possible failure deregistering instance\"}\n\t}\n\n\treturn nil\n}", "func (c *Client) DeleteInstances(args *DeleteInstanceArgs) error {\n\tjsonBytes, jsonErr := json.Marshal(args)\n\tif jsonErr != nil {\n\t\treturn jsonErr\n\t}\n\tbody, err := bce.NewBodyFromBytes(jsonBytes)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn DeleteInstances(c, body)\n}", "func (w *WorkflowInstance) Delete(trs *db.Transaction) error {\n\tif err := db.DeleteStructTx(trs.Tx, constants.TableCoreGroups, &db.Options{\n\t\tConditions: builder.Equal(\"code\", w.ID),\n\t}); err != nil {\n\t\treturn customerror.New(http.StatusInternalServerError, \"workflow instance delete\", err.Error())\n\t}\n\treturn nil\n}", "func (s *AutoscalerSuite) TestInstanceTerminate(c *check.C) {\n\tclusterName := \"bob\"\n\tinstance := &gaws.Instance{\n\t\tID: \"instance-1\",\n\t}\n\tec := newMockEC2(&ec2.Instance{\n\t\tInstanceId: aws.String(\"instance-1\"),\n\t})\n\tqueue := newMockQueue(\"queue-1\")\n\ta, err := New(Config{\n\t\tClusterName: clusterName,\n\t\tNewLocalInstance: func() (*gaws.Instance, error) {\n\t\t\treturn instance, nil\n\t\t},\n\t\tQueue: queue,\n\t\tCloud: ec,\n\t})\n\tc.Assert(err, check.IsNil)\n\tc.Assert(a, check.NotNil)\n\n\tctx, cancel := context.WithCancel(context.TODO())\n\tdefer cancel()\n\tserver := storage.Server{\n\t\tInstanceID: \"instance-to-delete\",\n\t\tHostname: \"instance-to-delete.hostname\",\n\t}\n\top := newMockOperator(ops.Site{\n\t\tAccountID: \"1\",\n\t\tDomain: \"example.com\",\n\t\tClusterState: storage.ClusterState{\n\t\t\tServers: []storage.Server{server},\n\t\t},\n\t})\n\tgo a.ProcessEvents(ctx, queue.url, op)\n\n\t// send terminated event\n\tmsg := &message{\n\t\treceipt: \"message-1\",\n\t\tbody: mustMarshalHook(HookEvent{\n\t\t\tInstanceID: server.InstanceID,\n\t\t\tType: InstanceTerminating,\n\t\t}),\n\t}\n\tselect {\n\tcase <-time.After(time.Second):\n\t\tc.Fatalf(\"timeout\")\n\tcase queue.messagesC <- msg:\n\t}\n\n\t// expect the shrink operation to arrive and the message to be deleted\n\tselect {\n\tcase op := <-op.shrinksC:\n\t\tc.Assert(op.Servers, check.DeepEquals, []string{server.Hostname})\n\tcase <-time.After(time.Second):\n\t\tc.Fatalf(\"timeout\")\n\t}\n\n\t// expect the message to be scheduled for deletion\n\tselect {\n\tcase m := <-queue.deletedC:\n\t\tc.Assert(aws.StringValue(m.ReceiptHandle), check.DeepEquals, msg.receipt)\n\tcase <-time.After(time.Second):\n\t\tc.Fatalf(\"timeout\")\n\t}\n}", "func (a Access) DeleteDBInstance(instanceID string) error {\n\turl := fmt.Sprintf(\"%s%s/instances/%s\", RDB_URL, a.TenantID,\n\t\tinstanceID)\n\t_, err := a.baseRequest(url, \"DELETE\", nil)\n\treturn err\n}", "func (d *DestinationClient) DeleteInstanceDestination(name string) (AffectedRecords, error) {\n\n\tvar retval AffectedRecords\n\tvar errResponse ErrorMessage\n\n\tresponse, err := d.restyClient.R().\n\t\tSetResult(&retval).\n\t\tSetError(&errResponse).\n\t\tSetPathParams(map[string]string{\n\t\t\t\"name\": name,\n\t\t}).\n\t\tDelete(\"/instanceDestinations/{name}\")\n\n\tif err != nil {\n\t\treturn retval, err\n\t}\n\tif response.StatusCode() != 200 {\n\t\terrResponse.statusCode = response.StatusCode()\n\t\treturn retval, errResponse\n\t}\n\treturn retval, nil\n}", "func (mr *MockECSClientMockRecorder) DeleteInstance(arg0 interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"DeleteInstance\", reflect.TypeOf((*MockECSClient)(nil).DeleteInstance), arg0)\n}", "func DeleteInstanceIP(\n\tctx context.Context,\n\ttx *sql.Tx,\n\trequest *models.DeleteInstanceIPRequest) error {\n\tdeleteQuery := deleteInstanceIPQuery\n\tselectQuery := \"select count(uuid) from instance_ip where uuid = ?\"\n\tvar err error\n\tvar count int\n\tuuid := request.ID\n\tauth := common.GetAuthCTX(ctx)\n\tif auth.IsAdmin() {\n\t\trow := tx.QueryRowContext(ctx, selectQuery, uuid)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"not found\")\n\t\t}\n\t\trow.Scan(&count)\n\t\tif count == 0 {\n\t\t\treturn errors.New(\"Not found\")\n\t\t}\n\t\t_, err = tx.ExecContext(ctx, deleteQuery, uuid)\n\t} else {\n\t\tdeleteQuery += \" and owner = ?\"\n\t\tselectQuery += \" and owner = ?\"\n\t\trow := tx.QueryRowContext(ctx, selectQuery, uuid, auth.ProjectID())\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"not found\")\n\t\t}\n\t\trow.Scan(&count)\n\t\tif count == 0 {\n\t\t\treturn errors.New(\"Not found\")\n\t\t}\n\t\t_, err = tx.ExecContext(ctx, deleteQuery, uuid, auth.ProjectID())\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"delete failed\")\n\t}\n\n\terr = common.DeleteMetaData(tx, uuid)\n\tlog.WithFields(log.Fields{\n\t\t\"uuid\": uuid,\n\t}).Debug(\"deleted\")\n\treturn err\n}", "func deleteRDSInstance(rdsClientSess *rds.RDS, restoreParams map[string]string) error {\n\trdsClusterName := restoreParams[\"restoreRDS\"]\n\trdsInstanceName := rdsClusterName + \"-0\" // TODO: this should be handled better\n\n\tinput := &rds.DeleteDBInstanceInput{\n\t\tDBInstanceIdentifier: aws.String(rdsInstanceName),\n\t\tSkipFinalSnapshot: aws.Bool(true),\n\t}\n\n\t_, err := rdsClientSess.DeleteDBInstance(input)\n\tif err != nil {\n\t\tif aerr, ok := err.(awserr.Error); ok {\n\t\t\tswitch aerr.Code() {\n\t\t\tcase rds.ErrCodeDBInstanceNotFoundFault:\n\t\t\t\tfmt.Println(rds.ErrCodeDBInstanceNotFoundFault, aerr.Error())\n\t\t\t\treturn fmt.Errorf(\"Error deleting RDS instance [%v] in RDS cluster [%v]\", rdsInstanceName, rdsClusterName)\n\t\t\tcase rds.ErrCodeInvalidDBInstanceStateFault:\n\t\t\t\tfmt.Println(rds.ErrCodeInvalidDBInstanceStateFault, aerr.Error())\n\t\t\t\treturn fmt.Errorf(\"Error deleting RDS instance [%v] in RDS cluster [%v]\", rdsInstanceName, rdsClusterName)\n\t\t\tcase rds.ErrCodeDBSnapshotAlreadyExistsFault:\n\t\t\t\tfmt.Println(rds.ErrCodeDBSnapshotAlreadyExistsFault, aerr.Error())\n\t\t\t\treturn fmt.Errorf(\"Error deleting RDS instance [%v] in RDS cluster [%v]\", rdsInstanceName, rdsClusterName)\n\t\t\tcase rds.ErrCodeSnapshotQuotaExceededFault:\n\t\t\t\tfmt.Println(rds.ErrCodeSnapshotQuotaExceededFault, aerr.Error())\n\t\t\t\treturn fmt.Errorf(\"Error deleting RDS instance [%v] in RDS cluster [%v]\", rdsInstanceName, rdsClusterName)\n\t\t\tcase rds.ErrCodeInvalidDBClusterStateFault:\n\t\t\t\tfmt.Println(rds.ErrCodeInvalidDBClusterStateFault, aerr.Error())\n\t\t\t\treturn fmt.Errorf(\"Error deleting RDS instance [%v] in RDS cluster [%v]\", rdsInstanceName, rdsClusterName)\n\t\t\tcase rds.ErrCodeDBInstanceAutomatedBackupQuotaExceededFault:\n\t\t\t\tfmt.Println(rds.ErrCodeDBInstanceAutomatedBackupQuotaExceededFault, aerr.Error())\n\t\t\t\treturn fmt.Errorf(\"Error deleting RDS instance [%v] in RDS cluster [%v]\", rdsInstanceName, rdsClusterName)\n\t\t\tdefault:\n\t\t\t\tfmt.Println(aerr.Error())\n\t\t\t\treturn fmt.Errorf(\"Error deleting RDS instance [%v] in RDS cluster [%v]\", rdsInstanceName, rdsClusterName)\n\t\t\t}\n\t\t} else {\n\t\t\t// Print the error, cast err to awserr.Error to get the Code and\n\t\t\t// Message from an error.\n\t\t\tfmt.Println(err.Error())\n\t\t\treturn fmt.Errorf(\"Error deleting RDS instance [%v] in RDS cluster [%v]\", rdsInstanceName, rdsClusterName)\n\t\t}\n\t}\n\n\t// TODO: DEBUG - fmt.Println(result)\n\tfmt.Printf(\"Deleting RDS instance [%v] in RDS cluster [%v]\\n\", rdsClusterName, rdsClusterName)\n\treturn nil\n}", "func (backRepoTask *BackRepoTaskStruct) CommitDeleteInstance(id uint) (Error error) {\n\n\ttask := (*backRepoTask.Map_TaskDBID_TaskPtr)[id]\n\n\t// task is not staged anymore, remove taskDB\n\ttaskDB := (*backRepoTask.Map_TaskDBID_TaskDB)[id]\n\tquery := backRepoTask.db.Unscoped().Delete(&taskDB)\n\tif query.Error != nil {\n\t\treturn query.Error\n\t}\n\n\t// update stores\n\tdelete((*backRepoTask.Map_TaskPtr_TaskDBID), task)\n\tdelete((*backRepoTask.Map_TaskDBID_TaskPtr), id)\n\tdelete((*backRepoTask.Map_TaskDBID_TaskDB), id)\n\n\treturn\n}", "func (backRepoFoo *BackRepoFooStruct) CommitDeleteInstance(id uint) (Error error) {\n\n\tfoo := (*backRepoFoo.Map_FooDBID_FooPtr)[id]\n\n\t// foo is not staged anymore, remove fooDB\n\tfooDB := (*backRepoFoo.Map_FooDBID_FooDB)[id]\n\tquery := backRepoFoo.db.Unscoped().Delete(&fooDB)\n\tif query.Error != nil {\n\t\treturn query.Error\n\t}\n\n\t// update stores\n\tdelete((*backRepoFoo.Map_FooPtr_FooDBID), foo)\n\tdelete((*backRepoFoo.Map_FooDBID_FooPtr), id)\n\tdelete((*backRepoFoo.Map_FooDBID_FooDB), id)\n\n\treturn\n}", "func (c Client) DestroyInstance(instance models.Instance) error {\n\turl := fmt.Sprintf(\"/instances/%d\", instance.ID)\n\tresp, err := c.delete(url)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif resp.StatusCode != http.StatusNoContent {\n\t\treturn parseError(resp.Body)\n\t}\n\n\treturn nil\n}", "func DeleteRelationshipInstance(ctx context.Context, deploymentID, nodeName, instanceName string) error {\n\trelInstancePath := path.Join(consulutil.DeploymentKVPrefix, deploymentID, \"topology/relationship_instances\")\n\tnodeRelInstancePath := path.Join(relInstancePath, nodeName)\n\treqIndices, err := consulutil.GetKeys(nodeRelInstancePath)\n\tif err != nil {\n\t\treturn errors.Wrap(err, consulutil.ConsulGenericErrMsg)\n\t}\n\tfor _, reqindex := range reqIndices {\n\t\terr = consulutil.Delete(path.Join(reqindex, instanceName)+\"/\", true)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, consulutil.ConsulGenericErrMsg)\n\t\t}\n\t}\n\n\t// now delete from targets in relationships instances\n\treturn addOrRemoveInstanceFromTargetRelationship(ctx, deploymentID, nodeName, instanceName, false)\n}", "func (o InstanceOutput) ForceDelete() pulumi.BoolPtrOutput {\n\treturn o.ApplyT(func(v *Instance) pulumi.BoolPtrOutput { return v.ForceDelete }).(pulumi.BoolPtrOutput)\n}", "func (adm Admin) DropInstance(zkSvr string, cluster string, instance string) {\n\tconn := newConnection(adm.ZkSvr)\n\terr := conn.Connect()\n\tif err != nil {\n\t\tfmt.Println(\"Failed to connect to zookeeper.\")\n\t}\n\tdefer conn.Disconnect()\n\n\tkb := KeyBuilder{cluster}\n\tinstanceKey := kb.instance(instance)\n\terr = conn.Delete(instanceKey)\n\tmust(err)\n\n\tfmt.Printf(\"/%s/%s deleted from zookeeper.\\n\", cluster, instance)\n}", "func (p *OnPrem) StopInstance(ctx *Context, instancename string) error {\n\treturn fmt.Errorf(\"Operation not supported\")\n}", "func (p *OnPrem) StopInstance(ctx *Context, instancename string) error {\n\treturn fmt.Errorf(\"Operation not supported\")\n}", "func (imc *inMemoryCatalog) delete(instanceID string) *ServiceInstance {\n\tinstance, exists := imc.instances[instanceID]\n\tif !exists {\n\t\treturn nil\n\t}\n\tserviceName := instance.ServiceName\n\n\tdelete(imc.services[serviceName], instanceID)\n\tif len(imc.services[serviceName]) == 0 {\n\t\tdelete(imc.services, serviceName)\n\t}\n\n\tdelete(imc.instances, instanceID)\n\n\tlifetime := time.Now().Sub(instance.RegistrationTime)\n\timc.lifetimeMetric.Update(int64(lifetime))\n\n\thadMetadata := len(instance.Metadata) > 0\n\thadTags := len(instance.Tags) > 0\n\n\tif hadMetadata {\n\t\timc.metadataInstancesMetric.Dec(1)\n\t}\n\tif hadTags {\n\t\timc.tagsInstancesMetric.Dec(1)\n\t}\n\n\timc.instancesMetric.Dec(1)\n\treturn instance\n}", "func Delete(client *gophercloud.ServiceClient, instanceID, dbName string) (r DeleteResult) {\n\tresp, err := client.Delete(dbURL(client, instanceID, dbName), nil)\n\t_, r.Header, r.Err = gophercloud.ParseResponse(resp, err)\n\treturn\n}", "func (client AccessGovernanceCPClient) DeleteGovernanceInstance(ctx context.Context, request DeleteGovernanceInstanceRequest) (response DeleteGovernanceInstanceResponse, err error) {\n\tvar ociResponse common.OCIResponse\n\tpolicy := common.DefaultRetryPolicy()\n\tif client.RetryPolicy() != nil {\n\t\tpolicy = *client.RetryPolicy()\n\t}\n\tif request.RetryPolicy() != nil {\n\t\tpolicy = *request.RetryPolicy()\n\t}\n\n\tif !(request.OpcRetryToken != nil && *request.OpcRetryToken != \"\") {\n\t\trequest.OpcRetryToken = common.String(common.RetryToken())\n\t}\n\n\tociResponse, err = common.Retry(ctx, request, client.deleteGovernanceInstance, policy)\n\tif err != nil {\n\t\tif ociResponse != nil {\n\t\t\tif httpResponse := ociResponse.HTTPResponse(); httpResponse != nil {\n\t\t\t\topcRequestId := httpResponse.Header.Get(\"opc-request-id\")\n\t\t\t\tresponse = DeleteGovernanceInstanceResponse{RawResponse: httpResponse, OpcRequestId: &opcRequestId}\n\t\t\t} else {\n\t\t\t\tresponse = DeleteGovernanceInstanceResponse{}\n\t\t\t}\n\t\t}\n\t\treturn\n\t}\n\tif convertedResponse, ok := ociResponse.(DeleteGovernanceInstanceResponse); ok {\n\t\tresponse = convertedResponse\n\t} else {\n\t\terr = fmt.Errorf(\"failed to convert OCIResponse into DeleteGovernanceInstanceResponse\")\n\t}\n\treturn\n}", "func (c *TestClient) StopInstance(project, zone, name string) error {\n\tif c.StopInstanceFn != nil {\n\t\treturn c.StopInstanceFn(project, zone, name)\n\t}\n\treturn c.client.StopInstance(project, zone, name)\n}", "func handleAppInstanceStatusDelete(ctxArg interface{}, key string,\n\tstatusArg interface{}) {\n\tctx := ctxArg.(*zedmanagerContext)\n\tpublishAppInstanceSummary(ctx)\n}", "func (r DeleteInstanceRequest) Send(ctx context.Context) (*DeleteInstanceResponse, error) {\n\tr.Request.SetContext(ctx)\n\terr := r.Request.Send()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresp := &DeleteInstanceResponse{\n\t\tDeleteInstanceOutput: r.Request.Data.(*DeleteInstanceOutput),\n\t\tresponse: &aws.Response{Request: r.Request},\n\t}\n\n\treturn resp, nil\n}", "func (r DeleteInstanceRequest) Send(ctx context.Context) (*DeleteInstanceResponse, error) {\n\tr.Request.SetContext(ctx)\n\terr := r.Request.Send()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresp := &DeleteInstanceResponse{\n\t\tDeleteInstanceOutput: r.Request.Data.(*DeleteInstanceOutput),\n\t\tresponse: &aws.Response{Request: r.Request},\n\t}\n\n\treturn resp, nil\n}", "func OVNInstanceDevicePortDelete(network Network, instanceID int, deviceName string, internalRoutes []*net.IPNet, externalRoutes []*net.IPNet) error {\n\t// Check network is of type OVN.\n\tn, ok := network.(*ovn)\n\tif !ok {\n\t\treturn fmt.Errorf(\"Network is not OVN type\")\n\t}\n\n\treturn n.instanceDevicePortDelete(instanceID, deviceName, internalRoutes, externalRoutes)\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstance(ctx context.Context, name string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstanceFdb(ctx context.Context, name string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:fdb/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (m *AccessReviewsDefinitionsItemInstancesItemStagesItemDecisionsItemInstanceRequestBuilder) Delete(ctx context.Context, requestConfiguration *AccessReviewsDefinitionsItemInstancesItemStagesItemDecisionsItemInstanceRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func (group *NodeGroup) deleteInstances(ips []string) error {\n\tgroupID := group.nodeGroupID\n\tklog.V(4).Infof(\"Start remove nodes %v\", ips)\n\treturn group.client.RemoveNodes(groupID, ips)\n}", "func (_m *ISession) StageInstanceDelete(channelID string, options ...discordgo.RequestOption) error {\n\t_va := make([]interface{}, len(options))\n\tfor _i := range options {\n\t\t_va[_i] = options[_i]\n\t}\n\tvar _ca []interface{}\n\t_ca = append(_ca, channelID)\n\t_ca = append(_ca, _va...)\n\tret := _m.Called(_ca...)\n\n\tvar r0 error\n\tif rf, ok := ret.Get(0).(func(string, ...discordgo.RequestOption) error); ok {\n\t\tr0 = rf(channelID, options...)\n\t} else {\n\t\tr0 = ret.Error(0)\n\t}\n\n\treturn r0\n}", "func (adm Admin) DropInstance(cluster string, instance string) error {\n\tkb := KeyBuilder{cluster}\n\tinstanceKey := kb.instance(instance)\n\terr := adm.zkClient.DeleteTree(instanceKey)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfmt.Printf(\"/%s/%s deleted from zookeeper.\\n\", cluster, instance)\n\n\treturn err\n}", "func (s *StepTeardownInstance) Run(state multistep.StateBag) multistep.StepAction {\n\tconfig := state.Get(\"config\").(*Config)\n\tdriver := state.Get(\"driver\").(Driver)\n\tui := state.Get(\"ui\").(packer.Ui)\n\n\tname := config.InstanceName\n\tif name == \"\" {\n\t\treturn multistep.ActionHalt\n\t}\n\n\tui.Say(\"Deleting instance...\")\n\tinstanceLog, _ := driver.GetSerialPortOutput(config.Zone, name)\n\tstate.Put(\"instance_log\", instanceLog)\n\terrCh, err := driver.DeleteInstance(config.Zone, name)\n\tif err == nil {\n\t\tselect {\n\t\tcase err = <-errCh:\n\t\tcase <-time.After(config.stateTimeout):\n\t\t\terr = errors.New(\"time out while waiting for instance to delete\")\n\t\t}\n\t}\n\n\tif err != nil {\n\t\tui.Error(fmt.Sprintf(\n\t\t\t\"Error deleting instance. Please delete it manually.\\n\\n\"+\n\t\t\t\t\"Name: %s\\n\"+\n\t\t\t\t\"Error: %s\", name, err))\n\t\treturn multistep.ActionHalt\n\t}\n\tui.Message(\"Instance has been deleted!\")\n\tstate.Put(\"instance_name\", \"\")\n\n\treturn multistep.ActionContinue\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwarding(ctx context.Context, name string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (m Manager) DeleteServiceInstanceDetails(tx *pg.Tx, instanceID string) error {\n\treturn tx.Delete(&ServiceInstance{ID: instanceID})\n}", "func (d *driverMock) TerminateInstance(ctx context.Context, id string) error {\n\tif d.TerminateInstanceErr != nil {\n\t\treturn d.TerminateInstanceErr\n\t}\n\n\td.TerminateInstanceID = id\n\n\treturn nil\n}", "func (a *Client) DeleteInstances(params *DeleteInstancesParams) (*DeleteInstancesOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewDeleteInstancesParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"deleteInstances\",\n\t\tMethod: \"DELETE\",\n\t\tPathPattern: \"/instances\",\n\t\tProducesMediaTypes: []string{\"\"},\n\t\tConsumesMediaTypes: []string{\"\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &DeleteInstancesReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*DeleteInstancesOK), nil\n\n}", "func (i instanceHandler) deleteHandler(w http.ResponseWriter, r *http.Request) {\n\tvars := mux.Vars(r)\n\tid := vars[\"instID\"]\n\n\terr := i.client.Delete(id)\n\tif err != nil {\n\t\tlog.Error(\"Error Deleting Instance\", log.Fields{\n\t\t\t\"error\": err,\n\t\t})\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(http.StatusAccepted)\n}", "func deleteInstanceConfig(w io.Writer, projectID, userConfigID string) error {\n\t// projectID := \"my-project-id\"\n\t// userConfigID := \"custom-config\", custom config names must start with the prefix “custom-”.\n\n\tctx := context.Background()\n\tadminClient, err := instance.NewInstanceAdminClient(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer adminClient.Close()\n\terr = adminClient.DeleteInstanceConfig(ctx, &instancepb.DeleteInstanceConfigRequest{\n\t\tName: fmt.Sprintf(\"projects/%s/instanceConfigs/%s\", projectID, userConfigID),\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\tfmt.Fprintf(w, \"Deleted instance configuration [%s]\\n\", userConfigID)\n\treturn nil\n}", "func (client *VirtualMachineScaleSetsClientMock) DeleteInstances(ctx context.Context, resourceGroupName string, vmScaleSetName string, vmInstanceIDs compute.VirtualMachineScaleSetVMInstanceRequiredIDs) *retry.Error {\n\targs := client.Called(resourceGroupName, vmScaleSetName, vmInstanceIDs)\n\tif args.Error(1) != nil {\n\t\treturn &retry.Error{RawError: args.Error(1)}\n\t}\n\treturn nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstanceProtocolsProtocolBgpNeighborsNeighborAfiSafisAfiSafiApplyPolicy(ctx context.Context, name string, identifier string, protocolName string, neighborAddress string, afiSafiName string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:protocols/frinx-openconfig-network-instance:protocol/{identifier}/{protocol-name}/frinx-openconfig-network-instance:bgp/frinx-openconfig-network-instance:neighbors/frinx-openconfig-network-instance:neighbor/{neighbor-address}/frinx-openconfig-network-instance:afi-safis/frinx-openconfig-network-instance:afi-safi/{afi-safi-name}/frinx-openconfig-network-instance:apply-policy/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"identifier\"+\"}\", fmt.Sprintf(\"%v\", identifier), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"protocol-name\"+\"}\", fmt.Sprintf(\"%v\", protocolName), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"neighbor-address\"+\"}\", fmt.Sprintf(\"%v\", neighborAddress), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"afi-safi-name\"+\"}\", fmt.Sprintf(\"%v\", afiSafiName), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (r *ProjectsBrokersV2ServiceInstancesService) Delete(parent string, instanceId string) *ProjectsBrokersV2ServiceInstancesDeleteCall {\n\tc := &ProjectsBrokersV2ServiceInstancesDeleteCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.parent = parent\n\tc.instanceId = instanceId\n\treturn c\n}", "func (i *pdfiumInstance) Kill() (err error) {\n\t// Kill should not be protected by a lock, since Kill is a last-effort\n\t// to \"recover\" a broken process.\n\tif i.closed {\n\t\treturn errors.New(\"instance is already closed\")\n\t}\n\n\tdefer func() {\n\t\tif panicError := recover(); panicError != nil {\n\t\t\terr = fmt.Errorf(\"panic occurred in %s: %v\", \"Close\", panicError)\n\t\t}\n\t}()\n\n\tdefer func() {\n\t\ti.pool.workerPool.ReturnObject(goctx.Background(), i.worker)\n\t\ti.worker = nil\n\t\ti.pool.lock.Lock()\n\t\tdelete(i.pool.instanceRefs, i.instanceRef)\n\t\ti.pool.lock.Unlock()\n\t\ti.pool = nil\n\t\ti.closed = true\n\t}()\n\n\ti.worker.pluginClient.Kill()\n\treturn\n}", "func (instance *Host) Delete(ctx context.Context) (ferr fail.Error) {\n\tdefer fail.OnPanic(&ferr)\n\n\tdefer func() {\n\t\t// drop the cache when we are done creating the cluster\n\t\tif ka, err := instance.Service().GetCache(context.Background()); err == nil {\n\t\t\tif ka != nil {\n\t\t\t\t_ = ka.Clear(context.Background())\n\t\t\t}\n\t\t}\n\t}()\n\n\tif valid.IsNil(instance) {\n\t\treturn fail.InvalidInstanceError()\n\t}\n\tif ctx == nil {\n\t\treturn fail.InvalidParameterCannotBeNilError(\"ctx\")\n\t}\n\n\txerr := instance.Inspect(ctx, func(clonable data.Clonable, props *serialize.JSONProperties) fail.Error {\n\t\t// Do not remove a Host that is a gateway\n\t\treturn props.Inspect(hostproperty.NetworkV2, func(clonable data.Clonable) fail.Error {\n\t\t\thostNetworkV2, ok := clonable.(*propertiesv2.HostNetworking)\n\t\t\tif !ok {\n\t\t\t\treturn fail.InconsistentError(\"'*propertiesv2.HostNetworking' expected, '%s' provided\", reflect.TypeOf(clonable).String())\n\t\t\t}\n\n\t\t\tif hostNetworkV2.IsGateway {\n\t\t\t\treturn fail.NotAvailableError(\"cannot delete Host, it's a gateway that can only be deleted through its Subnet\")\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\t})\n\txerr = debug.InjectPlannedFail(xerr)\n\tif xerr != nil {\n\t\treturn xerr\n\t}\n\n\txerr = instance.RelaxedDeleteHost(cleanupContextFrom(ctx))\n\treturn xerr\n}", "func (p *FuncPool) RemoveInstance(fID, imageName string, isSync bool) (string, error) {\n\tf := p.getFunction(fID, imageName)\n\n\treturn f.RemoveInstance(isSync)\n}", "func (c *nodePoolCache) removeInstance(nodePoolID, instanceID string) error {\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\n\tklog.Infof(\"Deleting instance %q from node pool %q\", instanceID, nodePoolID)\n\t// always try to remove the instance. This call is idempotent\n\tscaleDown := true\n\tresp, err := c.okeClient.DeleteNode(context.Background(), oke.DeleteNodeRequest{\n\t\tNodePoolId: &nodePoolID,\n\t\tNodeId: &instanceID,\n\t\tIsDecrementSize: &scaleDown,\n\t})\n\n\tklog.Infof(\"Delete Node API returned response: %v, err: %v\", resp, err)\n\thttpResp := resp.HTTPResponse()\n\tvar success bool\n\tif httpResp != nil {\n\t\tstatusCode := httpResp.StatusCode\n\t\t// status returned should be a 202, but let's accept any 2XX codes anyway\n\t\tstatusSuccess := statusCode >= 200 && statusCode < 300\n\t\tsuccess = statusSuccess ||\n\t\t\t// 409 means the instance is already going to be processed for deletion\n\t\t\tstatusCode == http.StatusConflict ||\n\t\t\t// 404 means it is probably already deleted and our cache may be stale\n\t\t\tstatusCode == http.StatusNotFound\n\t\tif !success {\n\t\t\tstatus := httpResp.Status\n\t\t\tklog.Infof(\"Received error status %s while deleting node %q\", status, instanceID)\n\n\t\t\t// statuses that we might expect but are still errors:\n\t\t\t// 400s (if cluster still uses TA or is v1 based)\n\t\t\t// 401 unauthorized\n\t\t\t// 412 etag mismatch\n\t\t\t// 429 too many requests\n\t\t\t// 500 internal server errors\n\t\t\treturn errors.Errorf(\"received error status %s while deleting node %q\", status, instanceID)\n\t\t} else if statusSuccess {\n\t\t\t// since delete node endpoint scales down by 1, we need to update the cache's target size by -1 too\n\t\t\tc.targetSize[nodePoolID]--\n\t\t}\n\t}\n\n\tif !success && err != nil {\n\t\treturn err\n\t}\n\n\tnodePool := c.cache[nodePoolID]\n\t// theoretical max number of nodes inside a cluster is 1000\n\t// so at most we'll be copying 1000 nodes\n\tnewNodeSlice := make([]oke.Node, 0, len(nodePool.Nodes))\n\tfor _, node := range nodePool.Nodes {\n\t\tif *node.Id != instanceID {\n\t\t\tnewNodeSlice = append(newNodeSlice, node)\n\t\t} else {\n\t\t\tklog.Infof(\"Deleting instance %q from cache\", instanceID)\n\t\t}\n\t}\n\tnodePool.Nodes = newNodeSlice\n\n\treturn nil\n}", "func Delete(client *golangsdk.ServiceClient, instanceId, authId string) (r DeleteResult) {\n\t_, r.Err = client.Delete(resourceURL(client, instanceId, authId), nil)\n\treturn\n}", "func (pluginSPI *PluginSPIImpl) NewDeleteInstanceRequest(instanceID string, force bool) (*ecs.DeleteInstanceRequest, error) {\n\trequest := ecs.CreateDeleteInstanceRequest()\n\n\trequest.InstanceId = instanceID\n\trequest.Force = requests.NewBoolean(force)\n\n\treturn request, nil\n}", "func (f *FakeInstanceGroups) DeleteInstanceGroup(name, zone string) error {\n\tig, err := f.getInstanceGroup(name, zone)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdelete(f.zonesToIGsToInstances[zone], ig)\n\treturn nil\n}", "func (m *MockECSClient) DeleteInstance(arg0 *ecs.DeleteInstanceRequest) (*ecs.DeleteInstanceResponse, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"DeleteInstance\", arg0)\n\tret0, _ := ret[0].(*ecs.DeleteInstanceResponse)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstanceInterInstancePolicies(ctx context.Context, name string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:inter-instance-policies/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func TerminateInstance() {\n\tlog.Println(\"Starting to run terminate instance process\")\n\t// Get instance id and region from metadata\n\tinstanceId, instanceRegion := getInstanceID()\n\tlog.Printf(\"Working on %v in %v region\", instanceId, instanceRegion)\n\n\t// Init aws session\n\tawsSession,_ := initSession(instanceRegion)\n\tlog.Println(\"Initialized aws session\")\n\n\t// Init Aws auto scaling session\n\tinitAutoScalingAwsSession(awsSession)\n\tlog.Println(\"Initialized auto scaling session\")\n\n\t// Get auto scaling group name\n\tinstanceAutoScaleGroupName := getAutoScalingName(instanceId)\n\tlog.Printf(\"Instance %v auto scaling group name is: %v\", instanceId, instanceAutoScaleGroupName)\n\n\t// Set instance scale in policy to false\n\tsuccess := setScaleInProtectionToInstance(instanceAutoScaleGroupName, instanceId)\n\n\t// Terminate ec2 instance after setting scale in policy to false\n\tif success{\n\t\tterminateInstance(instanceId)\n\t}\n}", "func (o *FeatureCvtermDbxref) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no FeatureCvtermDbxref provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), featureCvtermDbxrefPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"feature_cvterm_dbxref\\\" WHERE \\\"feature_cvterm_dbxref_id\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to delete from feature_cvterm_dbxref\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (s *server) Delete(ctx context.Context, body *pb.NameHolder) (*pb.DeletionResponse, error) {\n\tappName := body.GetName()\n\tfilter := types.M{\n\t\tmongo.NameKey: appName,\n\t\tmongo.InstanceTypeKey: mongo.AppInstance,\n\t}\n\n\tnode, _ := redis.FetchAppNode(appName)\n\tgo redis.DecrementServiceLoad(ServiceName, node)\n\tgo redis.RemoveApp(appName)\n\tgo diskCleanup(appName)\n\n\tif configs.CloudflareConfig.PlugIn {\n\t\tgo cloudflare.DeleteRecord(appName, mongo.AppInstance)\n\t}\n\n\t_, err := mongo.DeleteInstance(filter)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &pb.DeletionResponse{Success: true}, nil\n}", "func (f *Function) RemoveInstance(isSync bool) (string, error) {\n\tf.Lock()\n\tdefer f.Unlock()\n\n\tlogger := log.WithFields(log.Fields{\"fID\": f.fID, \"isSync\": isSync})\n\n\tlogger.Debug(\"Removing instance\")\n\n\tvar (\n\t\tr string\n\t\terr error\n\t)\n\n\tf.OnceAddInstance = new(sync.Once)\n\n\tif orch.GetSnapshotsEnabled() {\n\t\tf.OffloadInstance()\n\t\tr = \"Successfully offloaded instance \" + f.vmID\n\t} else {\n\t\tif isSync {\n\t\t\terr = orch.StopSingleVM(context.Background(), f.vmID)\n\t\t} else {\n\t\t\tf.RemoveInstanceAsync()\n\t\t\tr = \"Successfully removed (async) instance \" + f.vmID\n\t\t}\n\t}\n\n\treturn r, err\n}", "func (client *Client) DeleteResourceInstancesWithCallback(request *DeleteResourceInstancesRequest, callback func(response *DeleteResourceInstancesResponse, err error)) <-chan int {\n\tresult := make(chan int, 1)\n\terr := client.AddAsyncTask(func() {\n\t\tvar response *DeleteResourceInstancesResponse\n\t\tvar err error\n\t\tdefer close(result)\n\t\tresponse, err = client.DeleteResourceInstances(request)\n\t\tcallback(response, err)\n\t\tresult <- 1\n\t})\n\tif err != nil {\n\t\tdefer close(result)\n\t\tcallback(nil, err)\n\t\tresult <- 0\n\t}\n\treturn result\n}", "func (c *EC2) deleteAWSEC2Instance(id string) (err error) {\n\n\tterminateinstancesrequest := &TerminateInstancesInput{}\n\tterminateinstancesrequest.InstanceIds = append(terminateinstancesrequest.InstanceIds, &id)\n\tterminateinstancesresult, err := TerminateInstances(terminateinstancesrequest)\n\tif err == nil {\n\t\tdescribeinstancesrequest := &DescribeInstancesInput{}\n\t\tif err := awsutil.CopyValue(describeinstancesrequest, \"InstanceIds\", terminateinstancesrequest, \"InstanceIds[]\"); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif err := WaitUntilInstanceTerminated(describeinstancesrequest); err != nil {\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (a *CertificatesApiService) CertificatesDeleteInstance(ctx _context.Context, id string) (*_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/v1/certificates/{id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"id\"+\"}\", _neturl.PathEscape(parameterToString(id, \"\")), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 400 {\n\t\t\tvar v ErrorResponse\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 403 {\n\t\t\tvar v ErrorResponse\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v ErrorResponse\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 409 {\n\t\t\tvar v ErrorResponse\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func (d *DestinationClient) DeleteInstanceCertificate(name string) (AffectedRecords, error) {\n\n\tvar retval AffectedRecords\n\tvar errResponse ErrorMessage\n\n\tresponse, err := d.restyClient.R().\n\t\tSetResult(&retval).\n\t\tSetError(&errResponse).\n\t\tSetPathParams(map[string]string{\n\t\t\t\"name\": name,\n\t\t}).\n\t\tDelete(\"/instanceCertificate/{name}\")\n\n\tif err != nil {\n\t\treturn retval, err\n\t}\n\tif response.StatusCode() != 200 {\n\t\terrResponse.statusCode = response.StatusCode()\n\t\treturn retval, errResponse\n\t}\n\treturn retval, nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstanceInterInstancePoliciesApplyPolicy(ctx context.Context, name string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:inter-instance-policies/frinx-openconfig-network-instance:apply-policy/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func TestDeleteInstanceIPAddress(t *testing.T) {\n\tclient, instance, _, teardown, err := setupInstanceWithoutDisks(t, \"fixtures/TestDeleteInstanceIPAddress\")\n\tdefer teardown()\n\tif err != nil {\n\t\tt.Error(err)\n\t}\n\n\tip, err := client.AddInstanceIPAddress(context.TODO(), instance.ID, true)\n\tif err != nil {\n\t\tt.Fatalf(\"failed to allocate public IPv4 for instance (%d): %s\", instance.ID, err)\n\t}\n\n\ti, err := client.GetInstanceIPAddresses(context.TODO(), instance.ID)\n\tif err != nil {\n\t\tt.Fatalf(\"failed to get instance (%d) IP addresses: %s\", instance.ID, err)\n\t}\n\tif len(i.IPv4.Public) != 2 {\n\t\tt.Errorf(\"expected instance (%d) to have 2 public IPv4 addresses; got %d\", instance.ID, len(i.IPv4.Public))\n\t}\n\n\tif err := client.DeleteInstanceIPAddress(context.TODO(), instance.ID, ip.Address); err != nil {\n\t\tt.Fatalf(\"failed to delete instance (%d) public IPv4 address (%s): %s\", instance.ID, ip.Address, err)\n\t}\n\n\tif i, err = client.GetInstanceIPAddresses(context.TODO(), instance.ID); err != nil {\n\t\tt.Fatalf(\"failed to get instance (%d) IP addresses: %s\", instance.ID, err)\n\t}\n\tif len(i.IPv4.Public) != 1 {\n\t\tt.Errorf(\"expected instance (%d) to have 1 public IPv4 address; got %d\", instance.ID, len(i.IPv4.Public))\n\t}\n}", "func (s *storage) DeleteInstances(instances []*pb.SyncInstance) {\n\tfor _, val := range instances {\n\t\ts.deleteInstance(val.InstanceId)\n\t}\n}", "func (client *VirtualMachineScaleSetsClient) deleteInstances(ctx context.Context, resourceGroupName string, vmScaleSetName string, vmInstanceIDs VirtualMachineScaleSetVMInstanceRequiredIDs, options *VirtualMachineScaleSetsBeginDeleteInstancesOptions) (*http.Response, error) {\n\treq, err := client.deleteInstancesCreateRequest(ctx, resourceGroupName, vmScaleSetName, vmInstanceIDs, options)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresp, err := client.pl.Do(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif !runtime.HasStatusCode(resp, http.StatusOK, http.StatusAccepted) {\n\t\treturn nil, client.deleteInstancesHandleError(resp)\n\t}\n\treturn resp, nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwardingPolicies(ctx context.Context, name string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/frinx-openconfig-network-instance:policies/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstanceProtocolsProtocolBgpPeerGroupsPeerGroupAfiSafisAfiSafiApplyPolicy(ctx context.Context, name string, identifier string, protocolName string, peerGroupName string, afiSafiName string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:protocols/frinx-openconfig-network-instance:protocol/{identifier}/{protocol-name}/frinx-openconfig-network-instance:bgp/frinx-openconfig-network-instance:peer-groups/frinx-openconfig-network-instance:peer-group/{peer-group-name}/frinx-openconfig-network-instance:afi-safis/frinx-openconfig-network-instance:afi-safi/{afi-safi-name}/frinx-openconfig-network-instance:apply-policy/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"identifier\"+\"}\", fmt.Sprintf(\"%v\", identifier), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"protocol-name\"+\"}\", fmt.Sprintf(\"%v\", protocolName), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"peer-group-name\"+\"}\", fmt.Sprintf(\"%v\", peerGroupName), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"afi-safi-name\"+\"}\", fmt.Sprintf(\"%v\", afiSafiName), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}" ]
[ "0.76150215", "0.7271618", "0.7231779", "0.7048726", "0.6929907", "0.69071984", "0.6847707", "0.6841172", "0.6795391", "0.67552304", "0.6753494", "0.66945577", "0.662324", "0.6621943", "0.6571608", "0.6565294", "0.65361947", "0.6464905", "0.6432096", "0.64298874", "0.64264876", "0.6374261", "0.63412875", "0.61396354", "0.6115972", "0.60910445", "0.6066322", "0.6040317", "0.60302573", "0.5965695", "0.5927336", "0.59042805", "0.589639", "0.58849835", "0.58787477", "0.5865624", "0.5853978", "0.5850186", "0.5827372", "0.5803044", "0.5794551", "0.5793022", "0.5790364", "0.5753526", "0.5731126", "0.57306707", "0.5715245", "0.57101923", "0.57065415", "0.56493545", "0.56491023", "0.56351095", "0.56351095", "0.5627055", "0.5625381", "0.5573644", "0.55537605", "0.5525064", "0.55234444", "0.55234444", "0.55139714", "0.5506782", "0.5487105", "0.54706955", "0.5469369", "0.54685646", "0.5467265", "0.5465245", "0.5422004", "0.5418378", "0.5417186", "0.5409653", "0.54078984", "0.54066026", "0.540507", "0.5387043", "0.5377907", "0.5367945", "0.53469723", "0.53395504", "0.5336855", "0.5328051", "0.53210765", "0.5319019", "0.5298471", "0.5293549", "0.52932245", "0.52927405", "0.52911687", "0.52860063", "0.5285195", "0.5283353", "0.52773637", "0.527616", "0.526171", "0.5242098", "0.5236202", "0.52196795", "0.52068657", "0.5205153" ]
0.7125871
3
DeleteNetwork uses the override method DeleteNetworkFn or the real implementation.
func (c *TestClient) DeleteNetwork(project, name string) error { if c.DeleteNetworkFn != nil { return c.DeleteNetworkFn(project, name) } return c.client.DeleteNetwork(project, name) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *Stack) DeleteNetwork(ref string) (err error) {\n\ttheNetwork, err := s.GetNetwork(ref)\n\tif err != nil {\n\t\tif gerr, ok := err.(*googleapi.Error); ok {\n\t\t\tif gerr.Code != 404 {\n\t\t\t\treturn err\n\t\t\t}\n\t\t} else {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif theNetwork == nil {\n\t\treturn fail.Errorf(\n\t\t\tfmt.Sprintf(\"delete network failed: unexpected nil network when looking for [%s]\", ref), err,\n\t\t)\n\t}\n\n\tif !theNetwork.OK() {\n\t\tlogrus.Warnf(\"Missing data in network: %s\", spew.Sdump(theNetwork))\n\t}\n\n\tcompuService := s.ComputeService\n\tsubnetwork, err := compuService.Subnetworks.Get(s.GcpConfig.ProjectID, s.GcpConfig.Region, theNetwork.Name).Do()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\topp, err := compuService.Subnetworks.Delete(s.GcpConfig.ProjectID, s.GcpConfig.Region, subnetwork.Name).Do()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\toco := OpContext{\n\t\tOperation: opp,\n\t\tProjectID: s.GcpConfig.ProjectID,\n\t\tService: compuService,\n\t\tDesiredState: \"DONE\",\n\t}\n\n\terr = waitUntilOperationIsSuccessfulOrTimeout(oco, temporal.GetMinDelay(), temporal.GetHostCleanupTimeout())\n\tif err != nil {\n\t\tswitch err.(type) {\n\t\tcase fail.ErrTimeout:\n\t\t\tlogrus.Warnf(\"Timeout waiting for subnetwork deletion\")\n\t\t\treturn err\n\t\tdefault:\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Delete routes and firewall\n\tfirewallRuleName := fmt.Sprintf(\"%s-%s-all-in\", s.GcpConfig.NetworkName, subnetwork.Name)\n\tfws, err := compuService.Firewalls.Get(s.GcpConfig.ProjectID, firewallRuleName).Do()\n\tif fws != nil && err == nil {\n\t\topp, operr := compuService.Firewalls.Delete(s.GcpConfig.ProjectID, firewallRuleName).Do()\n\t\tif operr == nil {\n\t\t\toco := OpContext{\n\t\t\t\tOperation: opp,\n\t\t\t\tProjectID: s.GcpConfig.ProjectID,\n\t\t\t\tService: compuService,\n\t\t\t\tDesiredState: \"DONE\",\n\t\t\t}\n\n\t\t\toperr = waitUntilOperationIsSuccessfulOrTimeout(\n\t\t\t\toco, temporal.GetMinDelay(), temporal.GetHostCleanupTimeout(),\n\t\t\t)\n\t\t\tif operr != nil {\n\t\t\t\tlogrus.Warn(operr)\n\t\t\t}\n\t\t}\n\t}\n\tif err != nil {\n\t\tlogrus.Warn(err)\n\t}\n\n\tnatRuleName := fmt.Sprintf(\"%s-%s-nat-allowed\", s.GcpConfig.NetworkName, subnetwork.Name)\n\tnws, err := compuService.Routes.Get(s.GcpConfig.ProjectID, natRuleName).Do()\n\tif nws != nil && err == nil {\n\t\topp, operr := compuService.Routes.Delete(s.GcpConfig.ProjectID, natRuleName).Do()\n\t\tif operr == nil {\n\t\t\toco := OpContext{\n\t\t\t\tOperation: opp,\n\t\t\t\tProjectID: s.GcpConfig.ProjectID,\n\t\t\t\tService: compuService,\n\t\t\t\tDesiredState: \"DONE\",\n\t\t\t}\n\n\t\t\toperr = waitUntilOperationIsSuccessfulOrTimeout(\n\t\t\t\toco, temporal.GetMinDelay(), temporal.GetHostCleanupTimeout(),\n\t\t\t)\n\t\t\tif operr != nil {\n\t\t\t\tlogrus.Warn(operr)\n\t\t\t}\n\t\t}\n\t}\n\tif err != nil {\n\t\tlogrus.Warn(err)\n\t}\n\n\treturn nil\n}", "func NetworkDelete(state *state.State, n network) error {\n\treturn deleteProfile(state, dnsmasqProfileFilename(n))\n}", "func DelNetwork(c *check.C, cname string) (*http.Response, error) {\n\treturn request.Delete(\"/networks/\" + cname)\n}", "func NewDeleteNetwork(objRef string) *DeleteNetAPI {\n\tthis := new(DeleteNetAPI)\n\tthis.BaseAPI = api.NewBaseAPI(http.MethodDelete, fmt.Sprintf(\"%s/%s\", wapiVersion, objRef), nil, new(string))\n\treturn this\n}", "func (d *Driver) DeleteNetwork(r *pluginNet.DeleteNetworkRequest) error {\n\tdefer osl.InitOSContext()()\n\tnid := r.NetworkID\n\tlogrus.Infof(\"DeleteNetwork macvlan nid=%s\", nid)\n\tif nid == \"\" {\n\t\treturn fmt.Errorf(\"invalid network id\")\n\t}\n\n\tn := d.network(nid)\n\tif n == nil {\n\t\treturn fmt.Errorf(\"network id %s not found\", nid)\n\t}\n\t// if the driver created the slave interface, delete it, otherwise leave it\n\tif ok := n.config.CreatedSlaveLink; ok {\n\t\t// if the interface exists, only delete if it matches iface.vlan or dummy.net_id naming\n\t\tif ok := parentExists(n.config.Parent); ok {\n\t\t\t// only delete the link if it is named the net_id\n\t\t\tif n.config.Parent == getDummyName(stringid.TruncateID(nid)) {\n\t\t\t\terr := delDummyLink(n.config.Parent)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlogrus.Errorf(\"link %s was not deleted, continuing the delete network operation: %v\",\n\t\t\t\t\t\tn.config.Parent, err)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\t// only delete the link if it matches iface.vlan naming\n\t\t\t\terr := delVlanLink(n.config.Parent)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlogrus.Errorf(\"link %s was not deleted, continuing the delete network operation: %v\",\n\t\t\t\t\t\tn.config.Parent, err)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tfor _, ep := range n.endpoints {\n\t\tif link, err := ns.NlHandle().LinkByName(ep.srcName); err == nil {\n\t\t\tns.NlHandle().LinkDel(link)\n\t\t\tlogrus.Infof(\"DeleteNetwork delete macvlan link %s\", ep.srcName)\n\t\t}\n\n\t\tif err := d.store.StoreDelete(ep); err != nil {\n\t\t\tlogrus.Warnf(\"Failed to remove macvlan endpoint %s from store: %v\", ep.id[0:7], err)\n\t\t}\n\t}\n\t// delete the *network\n\td.deleteNetwork(nid)\n\treturn nil\n}", "func (client MeshNetworkClient) Delete(ctx context.Context, networkResourceName string) (result autorest.Response, err error) {\n\tif tracing.IsEnabled() {\n\t\tctx = tracing.StartSpan(ctx, fqdn+\"/MeshNetworkClient.Delete\")\n\t\tdefer func() {\n\t\t\tsc := -1\n\t\t\tif result.Response != nil {\n\t\t\t\tsc = result.Response.StatusCode\n\t\t\t}\n\t\t\ttracing.EndSpan(ctx, sc, err)\n\t\t}()\n\t}\n\treq, err := client.DeletePreparer(ctx, networkResourceName)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"servicefabric.MeshNetworkClient\", \"Delete\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresp, err := client.DeleteSender(req)\n\tif err != nil {\n\t\tresult.Response = resp\n\t\terr = autorest.NewErrorWithError(err, \"servicefabric.MeshNetworkClient\", \"Delete\", resp, \"Failure sending request\")\n\t\treturn\n\t}\n\n\tresult, err = client.DeleteResponder(resp)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"servicefabric.MeshNetworkClient\", \"Delete\", resp, \"Failure responding to request\")\n\t}\n\n\treturn\n}", "func TestNetworkCreateDelete(t *testing.T) {\n\t// create netagent\n\tag, _, _ := createNetAgent(t)\n\tAssert(t, ag != nil, \"Failed to create agent %#v\", ag)\n\tdefer ag.Stop()\n\n\t// network message\n\tnt := netproto.Network{\n\t\tTypeMeta: api.TypeMeta{Kind: \"Network\"},\n\t\tObjectMeta: api.ObjectMeta{\n\t\t\tTenant: \"default\",\n\t\t\tName: \"default\",\n\t\t\tNamespace: \"default\",\n\t\t},\n\t\tSpec: netproto.NetworkSpec{\n\t\t\tVlanID: 42,\n\t\t},\n\t}\n\n\t// make create network call\n\terr := ag.CreateNetwork(&nt)\n\tAssertOk(t, err, \"Error creating network\")\n\ttnt, err := ag.FindNetwork(nt.ObjectMeta)\n\tAssertOk(t, err, \"Network was not found in DB\")\n\tAssert(t, tnt.Spec.VlanID == 42, \"Network VLAN didn't match\", tnt)\n\n\t// verify duplicate network creations succeed\n\terr = ag.CreateNetwork(&nt)\n\tAssertOk(t, err, \"Error creating duplicate network\")\n\n\t// verify duplicate network name with different content does not succeed\n\tnnt := netproto.Network{\n\t\tTypeMeta: api.TypeMeta{Kind: \"Network\"},\n\t\tObjectMeta: api.ObjectMeta{\n\t\t\tTenant: \"default\",\n\t\t\tName: \"default\",\n\t\t\tNamespace: \"default\",\n\t\t},\n\t\tSpec: netproto.NetworkSpec{\n\t\t\tVlanID: 84,\n\t\t},\n\t}\n\terr = ag.CreateNetwork(&nnt)\n\tAssert(t, (err != nil), \"conflicting network creation succeeded\")\n\n\t// verify list api works\n\tnetList := ag.ListNetwork()\n\tAssert(t, len(netList) == 2, \"Incorrect number of networks\")\n\n\t// delete the network and verify its gone from db\n\terr = ag.DeleteNetwork(nt.Tenant, nt.Namespace, nt.Name)\n\tAssertOk(t, err, \"Error deleting network\")\n\t_, err = ag.FindNetwork(nt.ObjectMeta)\n\tAssert(t, err != nil, \"Network was still found in database after deleting\", ag)\n\n\t// verify you can not delete non-existing network\n\terr = ag.DeleteNetwork(nt.Tenant, nt.Namespace, nt.Name)\n\tAssert(t, err != nil, \"deleting non-existing network succeeded\", ag)\n}", "func (c *Client) DeleteNetwork(id string) error {\n\tendpoint := fmt.Sprintf(\"%snetworks/%s\", baseAddr, id)\n\tr, err := http.NewRequest(\"DELETE\", endpoint, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\tresp, err := c.http.Do(r)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn statusCode(resp.StatusCode, http.StatusNoContent)\n}", "func (n *NetworkServiceHandler) Delete(ctx context.Context, networkID string) error {\n\turi := \"/v1/network/destroy\"\n\n\tvalues := url.Values{\n\t\t\"NETWORKID\": {networkID},\n\t}\n\n\treq, err := n.client.NewRequest(ctx, http.MethodPost, uri, values)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = n.client.DoWithContext(ctx, req, nil)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (h networkHandler) deleteNetworkHandler(w http.ResponseWriter, r *http.Request) {\n\tvars := mux.Vars(r)\n\tclusterProvider := vars[\"provider-name\"]\n\tcluster := vars[\"cluster-name\"]\n\tname := vars[\"name\"]\n\n\terr := h.client.DeleteNetwork(name, clusterProvider, cluster)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tw.WriteHeader(http.StatusNoContent)\n}", "func (d *Driver) DeleteNetwork(r *sdk.DeleteNetworkRequest) error {\n\tlog.Debugf(\"Delete network request: %+v\", &r)\n\tn := d.network(r.NetworkID)\n\tif n == nil {\n\t\treturn nil\n\t}\n\n\tlog.Debugf(\"Delete network name: %v\", n.name)\n\n\tif err := d.nlh.LinkDel(n.bridge.Link); err != nil {\n\t\tlog.Warnf(\"Failed to remove bridge interface %s on network %s delete: %v\", n.name, r.NetworkID, err)\n\t}\n\n\td.deleteNetwork(r.NetworkID)\n\treturn nil\n}", "func (d Docker) DeleteNetwork(ctx context.Context, ID string) error {\n\t// (0) create new docker client\n\tdocker, err := client.NewClientWithOpts(client.FromEnv, client.WithAPIVersionNegotiation())\n\tif err != nil {\n\t\tlog.Errorln(\"Failed to create docker client\")\n\t\treturn err\n\t}\n\tdefer docker.Close()\n\n\t// (3) delete network\n\treturn docker.NetworkRemove(ctx, ID)\n}", "func (nb *NetBuilder) DeleteNetwork(nw *Network) error {\n\t// Find the HNS network.\n\thnsNetwork, err := hcsshim.GetHNSNetworkByName(nw.Name)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Delete the HNS network.\n\tlog.Infof(\"Deleting HNS network name: %s ID: %s\", nw.Name, hnsNetwork.Id)\n\t_, err = hnsNetwork.Delete()\n\tif err != nil {\n\t\tlog.Errorf(\"Failed to delete HNS network: %v.\", err)\n\t}\n\n\treturn err\n}", "func (s *GRPCServerHandler) DelNetwork(context context.Context, in *rpc.DelNetworkRequest) (*rpc.DelNetworkReply, error) {\n\tklog.V(1).Infof(\"Received DelNetwork for IP %s, Pod %s, Namespace %s, Container %s\",\n\t\tin.IPv4Addr, in.K8S_POD_NAME, in.K8S_POD_NAMESPACE, in.K8S_POD_INFRA_CONTAINER_ID)\n\n\tip, deviceNumber, err := s.ipamd.dataStore.UnassignPodIPv4Address(&k8sapi.K8SPodInfo{\n\t\tName: in.K8S_POD_NAME,\n\t\tNamespace: in.K8S_POD_NAMESPACE,\n\t\tContainer: in.K8S_POD_INFRA_CONTAINER_ID})\n\n\tif err != nil && err == datastore.ErrUnknownPod {\n\t\t// If L-IPAMD restarts, the pod's IP address are assigned by only pod's name and namespace due to kubelet's introspection.\n\t\tip, deviceNumber, err = s.ipamd.dataStore.UnassignPodIPv4Address(&k8sapi.K8SPodInfo{\n\t\t\tName: in.K8S_POD_NAME,\n\t\t\tNamespace: in.K8S_POD_NAMESPACE})\n\t\tif err == datastore.ErrUnknownPod {\n\t\t\tklog.Warningf(\"Detect unhealthy pod %s/%s\", in.K8S_POD_NAME, in.K8S_POD_NAMESPACE)\n\t\t\treturn &rpc.DelNetworkReply{Success: true}, nil\n\t\t}\n\t}\n\tklog.V(1).Infof(\"Send DelNetworkReply: IPv4Addr %s, DeviceNumber: %d, err: %v\", ip, deviceNumber, err)\n\tresp := &rpc.DelNetworkReply{Success: err == nil, IPv4Addr: ip, DeviceNumber: int32(deviceNumber)}\n\tif err != nil {\n\t\tresp.Message = err.Error()\n\t}\n\ts.ipamd.trigCh <- udevNotify{}\n\treturn resp, nil\n}", "func (in *Database) DeleteNetwork(netw *types.Network) error {\n\treturn in.delete(\"network\", netw)\n}", "func NetworkResourceDelete(w http.ResponseWriter, r *http.Request) {\n\tnrID := mux.Vars(r)[\"nr_id\"]\n\tcp := r.URL.Query().Get(\"cloud_provider\")\n\tif nrID == \"\" || cp == \"\" {\n\t\thandleError(w, http.StatusBadRequest,\n\t\t\t\"url params 'nr_id' or 'cloud_provider' is missing\")\n\t\treturn\n\t}\n\n\tif err := ctr.DeleteNetworkResource(nrID); err != nil {\n\t\thandleError(w, http.StatusInternalServerError, err.Error())\n\t\treturn\n\t}\n\n\thandleResult(w, http.StatusOK, emptyBody)\n\treturn\n}", "func checkDeleteNetwork(t *testing.T, expError bool, tenant, network string) {\n\terr := contivClient.NetworkDelete(tenant, network)\n\tif err != nil && !expError {\n\t\tt.Fatalf(\"Error deleting network %s/%s. Err: %v\", tenant, network, err)\n\t} else if err == nil && expError {\n\t\tt.Fatalf(\"Delete network %s/%s succeded while expecing error\", tenant, network)\n\t} else if err == nil {\n\t\t// verify network is gone\n\t\t_, err := contivClient.NetworkGet(tenant, network)\n\t\tif err == nil {\n\t\t\tt.Fatalf(\"Network %s/%s not deleted\", tenant, network)\n\t\t}\n\n\t\t// verify network state is gone too\n\t\tnetworkID := network + \".\" + tenant\n\t\tnwCfg := &mastercfg.CfgNetworkState{}\n\t\tnwCfg.StateDriver = stateStore\n\t\terr = nwCfg.Read(networkID)\n\t\tif err == nil {\n\t\t\tt.Fatalf(\"Network state %s not deleted\", networkID)\n\t\t}\n\t}\n}", "func (a *ASA) DeleteNetworkService(n interface{}) error {\n\tvar err error\n\tvar objectID string\n\n\tswitch v := n.(type) {\n\tcase *ReferenceObject:\n\t\tobjectID = v.ObjectID\n\tcase *NetworkService:\n\t\tobjectID = v.ObjectID\n\tcase string:\n\t\tobjectID = v\n\tdefault:\n\t\treturn fmt.Errorf(\"unknown type\")\n\t}\n\n\tif objectID == \"\" {\n\t\treturn fmt.Errorf(\"error objectid is null\")\n\t}\n\n\terr = a.Delete(fmt.Sprintf(\"%s/%s\", apiNetworkServicesEndpoint, objectID))\n\tif err != nil {\n\t\tif a.debug {\n\t\t\tglog.Errorf(\"Error: %s\\n\", err)\n\t\t}\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (f *IBMPICloudConnectionClient) DeleteNetwork(pcloudnetworkdef *p_cloud_cloud_connections.PcloudCloudconnectionsNetworksDeleteParams) (*models.CloudConnection, error) {\n\n\tparams := p_cloud_cloud_connections.NewPcloudCloudconnectionsNetworksDeleteParamsWithTimeout(helpers.PIDeleteTimeOut).WithCloudInstanceID(pcloudnetworkdef.CloudInstanceID).WithCloudConnectionID(pcloudnetworkdef.CloudConnectionID).WithNetworkID(pcloudnetworkdef.NetworkID)\n\t_, _, err := f.session.Power.PCloudCloudConnections.PcloudCloudconnectionsNetworksDelete(params, ibmpisession.NewAuth(f.session, pcloudnetworkdef.CloudInstanceID))\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to perform the delete operation... %s\", err)\n\t}\n\treturn nil, nil\n}", "func TestCreateModifyDeleteNetwork(t *testing.T) {\n\trecord(t, \"createmodifydeletenetwork\", func(t *testing.T, svc *Service) {\n\t\tnetwork, err := svc.CreateNetwork(&request.CreateNetworkRequest{\n\t\t\tName: \"test private network (test)\",\n\t\t\tZone: \"fi-hel2\",\n\t\t\tIPNetworks: []upcloud.IPNetwork{\n\t\t\t\t{\n\t\t\t\t\tAddress: \"172.16.0.0/22\",\n\t\t\t\t\tDHCP: upcloud.True,\n\t\t\t\t\tDHCPDefaultRoute: upcloud.False,\n\t\t\t\t\tDHCPDns: []string{\n\t\t\t\t\t\t\"172.16.0.10\",\n\t\t\t\t\t\t\"172.16.1.10\",\n\t\t\t\t\t},\n\t\t\t\t\tFamily: upcloud.IPAddressFamilyIPv4,\n\t\t\t\t\tGateway: \"172.16.0.1\",\n\t\t\t\t},\n\t\t\t},\n\t\t})\n\t\trequire.NoError(t, err)\n\t\tassert.NotEmpty(t, network.UUID)\n\t\tassert.Equal(t, \"test private network (test)\", network.Name)\n\n\t\tpostModifyNetwork, err := svc.ModifyNetwork(&request.ModifyNetworkRequest{\n\t\t\tUUID: network.UUID,\n\t\t\tName: \"modified private network (test)\",\n\t\t})\n\t\trequire.NoError(t, err)\n\t\tassert.Equal(t, \"modified private network (test)\", postModifyNetwork.Name)\n\n\t\tserverDetails, err := createServer(svc, \"TestCreateModifyDeleteNetwork\")\n\t\trequire.NoError(t, err)\n\n\t\terr = stopServer(svc, serverDetails.UUID)\n\t\trequire.NoError(t, err)\n\n\t\tiface, err := svc.CreateNetworkInterface(&request.CreateNetworkInterfaceRequest{\n\t\t\tServerUUID: serverDetails.UUID,\n\t\t\tNetworkUUID: postModifyNetwork.UUID,\n\t\t\tType: postModifyNetwork.Type,\n\t\t\tIPAddresses: []request.CreateNetworkInterfaceIPAddress{\n\t\t\t\t{\n\t\t\t\t\tFamily: upcloud.IPAddressFamilyIPv4,\n\t\t\t\t},\n\t\t\t},\n\t\t})\n\t\trequire.NoError(t, err)\n\t\tassert.NotEmpty(t, iface.IPAddresses)\n\t\tassert.NotEmpty(t, iface.IPAddresses[0].Address)\n\n\t\tmodifyIface, err := svc.ModifyNetworkInterface(&request.ModifyNetworkInterfaceRequest{\n\t\t\tServerUUID: serverDetails.UUID,\n\t\t\tCurrentIndex: iface.Index,\n\t\t\tNewIndex: iface.Index + 1,\n\t\t})\n\t\trequire.NoError(t, err)\n\t\tassert.Equal(t, iface.Index+1, modifyIface.Index)\n\n\t\terr = svc.DeleteNetworkInterface(&request.DeleteNetworkInterfaceRequest{\n\t\t\tServerUUID: serverDetails.UUID,\n\t\t\tIndex: modifyIface.Index,\n\t\t})\n\t\trequire.NoError(t, err)\n\n\t\terr = svc.DeleteNetwork(&request.DeleteNetworkRequest{\n\t\t\tUUID: network.UUID,\n\t\t})\n\t\trequire.NoError(t, err)\n\n\t\tnetworks, err := svc.GetNetworksInZone(&request.GetNetworksInZoneRequest{\n\t\t\tZone: network.Zone,\n\t\t})\n\t\trequire.NoError(t, err)\n\n\t\tvar found bool\n\t\tfor _, n := range networks.Networks {\n\t\t\tif n.UUID == network.UUID {\n\t\t\t\tfound = true\n\t\t\t}\n\t\t}\n\t\tassert.False(t, found)\n\t})\n}", "func (f *IBMPICloudConnectionClient) DeleteNetwork(pcloudnetworkdef *p_cloud_cloud_connections.PcloudCloudconnectionsNetworksDeleteParams) (*models.CloudConnection, error) {\n\n\tparams := p_cloud_cloud_connections.NewPcloudCloudconnectionsNetworksDeleteParams().WithCloudInstanceID(pcloudnetworkdef.CloudInstanceID).WithCloudConnectionID(pcloudnetworkdef.CloudConnectionID).WithNetworkID(pcloudnetworkdef.NetworkID)\n\tresp, err := f.session.Power.PCloudCloudConnections.PcloudCloudconnectionsNetworksDelete(params, ibmpisession.NewAuth(f.session, pcloudnetworkdef.CloudInstanceID))\n\n\tif err != nil || resp.Payload == nil {\n\t\treturn nil, fmt.Errorf(\"Failed to perform the delete operation... %s\", err)\n\t}\n\treturn resp.Payload, nil\n}", "func dockerNetDelete(subnetID string) error {\n\t_, err := exec.Command(\"docker\", \"network\", \"rm\", subnetID).CombinedOutput()\n\treturn err\n}", "func (c *client) Delete(ctx context.Context, group, name string) error {\n\tvnet, err := c.Get(ctx, group, name)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif len(*vnet) == 0 {\n\t\treturn fmt.Errorf(\"Virtual Network [%s] not found\", name)\n\t}\n\n\trequest, err := getVirtualNetworkRequest(wssdcloudcommon.Operation_DELETE, group, name, &(*vnet)[0])\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = c.VirtualNetworkAgentClient.Invoke(ctx, request)\n\n\treturn err\n}", "func DeleteHnsNetwork(networkName string) error {\n\treturn fmt.Errorf(\"DeleteHnsNetwork shouldn't be called for linux platform\")\n}", "func (vns *VirtualNetworkService) Delete(ctx context.Context, vn resources.VirtualNetwork) error {\n\tvnID, err := vn.ID()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = vns.call(ctx, \"one.vn.delete\", vnID)\n\n\treturn err\n}", "func (m *InterfacesClientMock) Delete(ctx context.Context, resourceGroupName string, networkInterfaceName string) *retry.Error {\n\targs := m.Called(resourceGroupName, networkInterfaceName)\n\tif args.Error(1) != nil {\n\t\treturn &retry.Error{RawError: args.Error(1)}\n\t}\n\treturn nil\n}", "func (a *AllApiService) NetworkDeleteNetworkGatewayPool(ctx _context.Context, body NetworkDeleteNetworkGatewayPool) (NetworkDeleteNetworkGatewayPoolResult, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = _nethttp.MethodPost\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue NetworkDeleteNetworkGatewayPoolResult\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/network/deleteNetworkGatewayPool\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = &body\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t}\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\tif localVarHttpResponse.StatusCode == 200 {\n\t\t\tvar v NetworkDeleteNetworkGatewayPoolResult\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t}\n\t\tif localVarHttpResponse.StatusCode == 400 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t}\n\t\tif localVarHttpResponse.StatusCode == 500 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t}\n\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHttpResponse, nil\n}", "func Delete(c *gophercloud.ServiceClient, networkID string) os.DeleteResult {\n\treturn os.Delete(c, networkID)\n}", "func DeleteNetworkSecurityGroupRule() {}", "func (plugin *Plugin) Del(args *cniSkel.CmdArgs) error {\n\t// Parse network configuration.\n\tnetConfig, err := config.New(args)\n\tif err != nil {\n\t\tlog.Errorf(\"Failed to parse netconfig from args: %v.\", err)\n\t\treturn err\n\t}\n\n\tlog.Infof(\"Executing DEL with netconfig: %+v ContainerID:%v Netns:%v IfName:%v Args:%v.\",\n\t\tnetConfig, args.ContainerID, args.Netns, args.IfName, args.Args)\n\n\tvar vpcENI *eni.ENI\n\t// If existing network is to be used then ENI is not required.\n\tif !netConfig.UseExistingNetwork {\n\t\t// Find the ENI.\n\t\tvpcENI, err = eni.NewENI(args.IfName, netConfig.ENIMACAddress)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Failed to find ENI %s: %v.\", netConfig.ENIName, err)\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Call operating system specific handler.\n\tnb := plugin.nb\n\n\tnw := network.Network{\n\t\tName: netConfig.Name,\n\t\tENI: vpcENI,\n\t\tUseExisting: netConfig.UseExistingNetwork,\n\t}\n\n\tep := network.Endpoint{\n\t\tContainerID: args.ContainerID,\n\t\tNetNSName: args.Netns,\n\t\tIPAddresses: netConfig.ENIIPAddresses,\n\t\tMACAddress: netConfig.ENIMACAddress,\n\t}\n\n\terr = nb.DeleteEndpoint(&nw, &ep)\n\tif err != nil {\n\t\t// DEL is best-effort. Log and ignore the failure.\n\t\tlog.Errorf(\"Failed to delete endpoint, ignoring: %v.\", err)\n\t}\n\n\t// Do not delete pre-existing networks.\n\tif !nw.UseExisting {\n\t\terr = nb.DeleteNetwork(&nw)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Failed to delete network: %v.\", err)\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func TestNetworkAddDelete(t *testing.T) {\n\t// Basic vlan network\n\tcheckCreateNetwork(t, false, \"default\", \"contiv\", \"vlan\", \"10.1.1.1/24\", \"10.1.1.254\", 1)\n\tverifyNetworkState(t, \"default\", \"contiv\", \"vlan\", \"10.1.1.1\", \"10.1.1.254\", 24, 1, 0)\n\tcheckDeleteNetwork(t, false, \"default\", \"contiv\")\n\n\t// Basic Vxlan network\n\tcheckCreateNetwork(t, false, \"default\", \"contiv\", \"vxlan\", \"10.1.1.1/16\", \"10.1.1.254\", 1)\n\tverifyNetworkState(t, \"default\", \"contiv\", \"vxlan\", \"10.1.1.1\", \"10.1.1.254\", 16, 1, 1)\n\tcheckDeleteNetwork(t, false, \"default\", \"contiv\")\n\n\t// Basic IP range network checks\n\tcheckCreateNetwork(t, false, \"default\", \"contiv\", \"vxlan\", \"10.1.1.10-20/24\", \"10.1.1.254\", 1)\n\tverifyNetworkState(t, \"default\", \"contiv\", \"vxlan\", \"10.1.1.10\", \"10.1.1.254\", 24, 1, 1)\n\tcheckDeleteNetwork(t, false, \"default\", \"contiv\")\n\n\t// Try network create with invalid network range\n\tcheckCreateNetwork(t, true, \"default\", \"contiv\", \"vxlan\", \"10.1.1.1-70/26\", \"10.1.1.63\", 1)\n\n\t// Try network create with invalid subnet length\n\tcheckCreateNetwork(t, true, \"default\", \"contiv\", \"vxlan\", \"10.1.1.1/32\", \"10.1.1.1\", 1)\n\n\t// try creating network without tenant\n\tcheckCreateNetwork(t, true, \"tenant1\", \"contiv\", \"vxlan\", \"10.1.1.1/24\", \"10.1.1.254\", 1)\n\n\t// try invalid encap\n\tcheckCreateNetwork(t, true, \"default\", \"contiv\", \"vvvv\", \"10.1.1.1/24\", \"10.1.1.254\", 1)\n\n\t// try invalid pkt tags\n\tcheckCreateNetwork(t, true, \"default\", \"contiv\", \"vlan\", \"10.1.1.1/24\", \"10.1.1.254\", 5000)\n\tcheckCreateNetwork(t, true, \"default\", \"contiv\", \"vxlan\", \"10.1.1.1/24\", \"10.1.1.254\", 20000)\n\n\t// Try gateway outside the network\n\tcheckCreateNetwork(t, true, \"default\", \"contiv\", \"vxlan\", \"10.1.1.1/24\", \"10.1.2.254\", 1)\n\tcheckCreateNetwork(t, true, \"default\", \"contiv\", \"vxlan\", \"10.1.1.65-70/26\", \"10.1.1.1\", 2)\n\n\t// Try deleting a non-existing network\n\tcheckDeleteNetwork(t, true, \"default\", \"contiv\")\n}", "func networkRemoveExample() string {\n\treturn `$ pouch network remove pouch-net\nRemoved: pouch-net`\n}", "func (client MeshNetworkClient) DeletePreparer(ctx context.Context, networkResourceName string) (*http.Request, error) {\n\tpathParameters := map[string]interface{}{\n\t\t\"networkResourceName\": networkResourceName,\n\t}\n\n\tconst APIVersion = \"6.4-preview\"\n\tqueryParameters := map[string]interface{}{\n\t\t\"api-version\": APIVersion,\n\t}\n\n\tpreparer := autorest.CreatePreparer(\n\t\tautorest.AsDelete(),\n\t\tautorest.WithBaseURL(client.BaseURI),\n\t\tautorest.WithPathParameters(\"/Resources/Networks/{networkResourceName}\", pathParameters),\n\t\tautorest.WithQueryParameters(queryParameters))\n\treturn preparer.Prepare((&http.Request{}).WithContext(ctx))\n}", "func (r *DOSProfileDOSNetworkResource) Delete(id string) error {\n\tif err := r.c.ModQuery(\"DELETE\", BasePath+DOSProfileDOSNetworkEndpoint+\"/\"+id, nil); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (client IdentityClient) DeleteNetworkSource(ctx context.Context, request DeleteNetworkSourceRequest) (response DeleteNetworkSourceResponse, err error) {\n\tvar ociResponse common.OCIResponse\n\tpolicy := common.NoRetryPolicy()\n\tif client.RetryPolicy() != nil {\n\t\tpolicy = *client.RetryPolicy()\n\t}\n\tif request.RetryPolicy() != nil {\n\t\tpolicy = *request.RetryPolicy()\n\t}\n\tociResponse, err = common.Retry(ctx, request, client.deleteNetworkSource, policy)\n\tif err != nil {\n\t\tif ociResponse != nil {\n\t\t\tif httpResponse := ociResponse.HTTPResponse(); httpResponse != nil {\n\t\t\t\topcRequestId := httpResponse.Header.Get(\"opc-request-id\")\n\t\t\t\tresponse = DeleteNetworkSourceResponse{RawResponse: httpResponse, OpcRequestId: &opcRequestId}\n\t\t\t} else {\n\t\t\t\tresponse = DeleteNetworkSourceResponse{}\n\t\t\t}\n\t\t}\n\t\treturn\n\t}\n\tif convertedResponse, ok := ociResponse.(DeleteNetworkSourceResponse); ok {\n\t\tresponse = convertedResponse\n\t} else {\n\t\terr = fmt.Errorf(\"failed to convert OCIResponse into DeleteNetworkSourceResponse\")\n\t}\n\treturn\n}", "func (d *common) Delete() error {\n\tisUsed, err := d.isUsed()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif isUsed {\n\t\treturn fmt.Errorf(\"Cannot delete an ACL that is in use\")\n\t}\n\n\treturn d.state.DB.Cluster.DeleteNetworkACL(d.id)\n}", "func DeleteVirtualNetworkSubnet() {}", "func DeleteVirtualNetworkSubnet() {}", "func (pool *PoolHandler) Delete(request *restful.Request, response *restful.Response) {\n\tstarted := time.Now()\n\tnetKey := request.PathParameter(\"net\")\n\tnetCluster := request.PathParameter(\"cluster\")\n\tnetRes := &types.NetResponse{\n\t\tType: types.ResponseType_POOL,\n\t}\n\tif len(netKey) == 0 || len(netCluster) == 0 {\n\t\tnetRes.Code = 1\n\t\tnetRes.Message = \"Lost param needed\"\n\t\tresponse.WriteEntity(netRes)\n\t\treportMetrics(\"deleteIPPool\", \"4xx\", started)\n\t\treturn\n\t}\n\tif err := pool.netSvr.DeletePool(netCluster + \"/\" + netKey); err != nil {\n\t\tblog.Errorf(\"NetPool Delete %s/%s request err: %s\", netCluster, netKey, err.Error())\n\t\tnetRes.Code = 1\n\t\tnetRes.Message = err.Error()\n\t\tresponse.WriteEntity(netRes)\n\t\treportMetrics(\"deleteIPPool\", \"5xx\", started)\n\t\treturn\n\t}\n\tnetRes.Code = 0\n\tnetRes.Message = SUCCESS\n\tif err := response.WriteEntity(netRes); err != nil {\n\t\tblog.Errorf(\"PoolHandler reply client DELETE request Err: %v\", err)\n\t}\n\treportMetrics(\"deleteIPPool\", \"2xx\", started)\n}", "func DelNetworkOk(c *check.C, cname string) {\n\tresp, err := DelNetwork(c, cname)\n\tc.Assert(err, check.IsNil)\n\n\tCheckRespStatus(c, resp, 204)\n}", "func DeleteVNet(vNetName string, rgroupName string) {\n\ta := wow.New(os.Stdout, spin.Get(spin.Dots), \"Deleting virtual network : \"+vNetName)\n\ta.Start()\n\ttime.Sleep(2 * time.Second)\n\ta.Text(\"This would take a few minutes...\").Spinner(spin.Get(spin.Dots))\n\tcmd := exec.Command(\"az\", \"network\", \"vnet\", \"delete\", \"--name\",\n\t\tvNetName, \"--resource-group\", rgroupName)\n\tvar out bytes.Buffer\n\tvar stderr bytes.Buffer\n\tcmd.Stdout = &out\n\tcmd.Stderr = &stderr\n\terr := cmd.Run()\n\tif err != nil {\n\t\tfmt.Println(fmt.Sprint(err) + \": \" + stderr.String())\n\t\treturn\n\t}\n\ta.PersistWith(spin.Spinner{}, \"....\")\n\tfmt.Println(\"Result: \" + out.String())\n}", "func (r *ProtocolIncus) DeleteNetworkACL(name string) error {\n\tif !r.HasExtension(\"network_acl\") {\n\t\treturn fmt.Errorf(`The server is missing the required \"network_acl\" API extension`)\n\t}\n\n\t// Send the request.\n\t_, _, err := r.query(\"DELETE\", fmt.Sprintf(\"/network-acls/%s\", url.PathEscape(name)), nil, \"\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (gc *Cfg) UnassignNetwork(networkName string) error {\n\tif networkName == \"\" {\n\t\treturn nil\n\t}\n\n\tg := &Oper{}\n\tg.StateDriver = gc.StateDriver\n\tif err := g.Read(gc.Tenant); core.ErrIfKeyExists(err) != nil {\n\t\treturn err\n\t}\n\n\tif networkName == g.DefaultNetwork {\n\t\tg.DefaultNetwork = \"\"\n\t\tif err := g.Write(); err != nil {\n\t\t\tlog.Errorf(\"error '%s' updating goper state %v \\n\", err, g)\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (l *Libvirt) NetworkPortDelete(Port NetworkPort, Flags uint32) (err error) {\n\tvar buf []byte\n\n\targs := NetworkPortDeleteArgs {\n\t\tPort: Port,\n\t\tFlags: Flags,\n\t}\n\n\tbuf, err = encode(&args)\n\tif err != nil {\n\t\treturn\n\t}\n\n\n\t_, err = l.requestStream(410, constants.Program, buf, nil, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func (a *NetworkServerAPI) Delete(ctx context.Context, req *pb.DeleteNetworkServerRequest) (*pb.DeleteNetworkServerResponse, error) {\n\tif err := a.validator.Validate(ctx,\n\t\tauth.ValidateNetworkServerAccess(auth.Delete, req.Id),\n\t); err != nil {\n\t\treturn nil, grpc.Errorf(codes.Unauthenticated, \"authentication failed: %s\", err)\n\t}\n\n\terr := storage.Transaction(config.C.PostgreSQL.DB, func(tx sqlx.Ext) error {\n\t\treturn storage.DeleteNetworkServer(tx, req.Id)\n\t})\n\tif err != nil {\n\t\treturn nil, errToRPCError(err)\n\t}\n\n\treturn &pb.DeleteNetworkServerResponse{}, nil\n}", "func (n *Network) ValidateDelete() error {\n\tnetworklog.Info(\"validate delete\", \"name\", n.Name)\n\n\t// TODO(user): fill in your validation logic upon object deletion.\n\treturn nil\n}", "func (client *Client) DeleteNetworkPool(id int64, req *Request) (*Response, error) {\n\treturn client.Execute(&Request{\n\t\tMethod: \"DELETE\",\n\t\tPath: fmt.Sprintf(\"%s/%d\", NetworkPoolsPath, id),\n\t\tQueryParams: req.QueryParams,\n\t\tBody: req.Body,\n\t\tResult: &DeleteNetworkPoolResult{},\n\t})\n}", "func (c *FakePodNetworkings) Delete(ctx context.Context, name string, opts v1.DeleteOptions) error {\n\t_, err := c.Fake.\n\t\tInvokes(testing.NewRootDeleteActionWithOptions(podnetworkingsResource, name, opts), &v1beta1.PodNetworking{})\n\treturn err\n}", "func (liqoIPAM *IPAM) RemoveNetworkPool(network string) error {\n\t// Get resource\n\tipamPools, err := liqoIPAM.ipamStorage.getPools()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"cannot get Ipam config: %w\", err)\n\t}\n\t// Get cluster subnets\n\tclusterSubnet, err := liqoIPAM.ipamStorage.getClusterSubnet()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"cannot get cluster subnets: %w\", err)\n\t}\n\t// Check existence\n\tif exists := slice.ContainsString(ipamPools, network, nil); !exists {\n\t\treturn fmt.Errorf(\"network %s is not a network pool\", network)\n\t}\n\t// Cannot remove a default one\n\tif contains := slice.ContainsString(Pools, network, nil); contains {\n\t\treturn fmt.Errorf(\"cannot remove a default network pool\")\n\t}\n\t// Check overlapping with cluster networks\n\tcluster, overlaps, err := liqoIPAM.overlapsWithCluster(network)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"cannot check if network pool %s overlaps with cluster networks:%w\", network, err)\n\t}\n\tif overlaps {\n\t\treturn fmt.Errorf(\"cannot remove network pool %s because it overlaps with network %s of cluster %s\", network, clusterSubnet[cluster], cluster)\n\t}\n\t// Release it\n\t_, err = liqoIPAM.ipam.DeletePrefix(network)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"cannot remove network pool %s:%w\", network, err)\n\t}\n\t// Delete it\n\tvar i int\n\tfor index, value := range ipamPools {\n\t\tif value == network {\n\t\t\ti = index\n\t\t\tbreak\n\t\t}\n\t}\n\tif i == (len(ipamPools) - 1) {\n\t\tipamPools = ipamPools[:len(ipamPools)-1]\n\t} else {\n\t\tcopy(ipamPools[i:], ipamPools[i+1:])\n\t\tipamPools = ipamPools[:len(ipamPools)-1]\n\t}\n\terr = liqoIPAM.ipamStorage.updatePools(ipamPools)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"cannot update Ipam configuration:%w\", err)\n\t}\n\tklog.Infof(\"Network pool %s has just been removed\", network)\n\treturn nil\n}", "func (n *Network) Close() error {\n\tscopes.Framework.Infof(\"Closing Docker network %s (ID=%s)\", n.Name, n.id)\n\treturn n.dockerClient.NetworkRemove(context.Background(), n.id)\n}", "func (sm *Statemgr) OnNetworkSecurityPolicyOperDelete(nodeID string, objinfo *netproto.NetworkSecurityPolicy) error {\n\treturn nil\n}", "func (client IdentityClient) deleteNetworkSource(ctx context.Context, request common.OCIRequest, binaryReqBody *common.OCIReadSeekCloser, extraHeaders map[string]string) (common.OCIResponse, error) {\n\n\thttpRequest, err := request.HTTPRequest(http.MethodDelete, \"/networkSources/{networkSourceId}\", binaryReqBody, extraHeaders)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar response DeleteNetworkSourceResponse\n\tvar httpResponse *http.Response\n\thttpResponse, err = client.Call(ctx, &httpRequest)\n\tdefer common.CloseBodyIfValid(httpResponse)\n\tresponse.RawResponse = httpResponse\n\tif err != nil {\n\t\treturn response, err\n\t}\n\n\terr = common.UnmarshalResponse(httpResponse, &response)\n\treturn response, err\n}", "func (p Pipeline) RemoveNetwork() error {\n\treturn p.localRunner.NetworkRemover(p.Network)()\n}", "func DeleteProspectNetwork(id int) bool {\n\tlog.Printf(\"debug: deleting network %d\", id)\n\trow := connection.QueryRow(`delete from prospect_network where id = $1`, id)\n\treturn row != nil\n}", "func (c *client) RemoveNetwork(ctx context.Context, b *pipeline.Build) error {\n\tlogrus.Tracef(\"removing network for pipeline %s\", b.ID)\n\n\t// remove the network definition from the pod spec\n\t//\n\t// https://pkg.go.dev/k8s.io/api/core/v1?tab=doc#PodSpec\n\tc.Pod.Spec.HostAliases = []v1.HostAlias{}\n\n\treturn nil\n}", "func (l *Libvirt) NetworkDestroy(Net Network) (err error) {\n\tvar buf []byte\n\n\targs := NetworkDestroyArgs {\n\t\tNet: Net,\n\t}\n\n\tbuf, err = encode(&args)\n\tif err != nil {\n\t\treturn\n\t}\n\n\n\t_, err = l.requestStream(42, constants.Program, buf, nil, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func DeleteNetns(name string) error {\n\tif err := netns.DeleteNamed(name); err != nil && !errors.IsError(err, unix.ENOENT) {\n\t\tklog.Errorf(\"an error occurred while removing network namespace with name %s: %v\", name, err)\n\t\treturn err\n\t}\n\treturn nil\n}", "func (client *NetworkToNetworkInterconnectsClient) deleteCreateRequest(ctx context.Context, resourceGroupName string, networkFabricName string, networkToNetworkInterconnectName string, options *NetworkToNetworkInterconnectsClientBeginDeleteOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.ManagedNetworkFabric/networkFabrics/{networkFabricName}/networkToNetworkInterconnects/{networkToNetworkInterconnectName}\"\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif networkFabricName == \"\" {\n\t\treturn nil, errors.New(\"parameter networkFabricName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{networkFabricName}\", url.PathEscape(networkFabricName))\n\tif networkToNetworkInterconnectName == \"\" {\n\t\treturn nil, errors.New(\"parameter networkToNetworkInterconnectName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{networkToNetworkInterconnectName}\", url.PathEscape(networkToNetworkInterconnectName))\n\treq, err := runtime.NewRequest(ctx, http.MethodDelete, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2023-06-15\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func (service *ContrailService) DeleteNetworkPolicy(ctx context.Context, request *models.DeleteNetworkPolicyRequest) (*models.DeleteNetworkPolicyResponse, error) {\n\tif err := common.DoInTransaction(\n\t\tservice.DB,\n\t\tfunc(tx *sql.Tx) error {\n\t\t\treturn db.DeleteNetworkPolicy(ctx, tx, request)\n\t\t}); err != nil {\n\t\tlog.WithField(\"err\", err).Debug(\"error deleting a resource\")\n\t\treturn nil, common.ErrorInternal\n\t}\n\treturn &models.DeleteNetworkPolicyResponse{\n\t\tID: request.ID,\n\t}, nil\n}", "func DeleteOperator(namespace string) error {\n\tif namespace == \"\" {\n\t\tnamespace = DEFAULT_NAMESPACE\n\t}\n\tif err := sh.RunV(\"helm\", \"delete\", \"-n\", namespace, \"kedahttp\"); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func DeleteDefaultExtNetwork() error {\n\treturn fmt.Errorf(\"DeleteDefaultExtNetwork shouldn't be called for linux platform\")\n}", "func DeleteCNINetwork(ctx context.Context, cni gocni.CNI, client *containerd.Client, name string) error {\n\tcontainer, containerErr := client.LoadContainer(ctx, name)\n\tif containerErr == nil {\n\t\ttask, err := container.Task(ctx, nil)\n\t\tif err != nil {\n\t\t\tlog.Printf(\"[Delete] unable to find task for container: %s\\n\", name)\n\t\t\treturn nil\n\t\t}\n\n\t\tlog.Printf(\"[Delete] removing CNI network for: %s\\n\", task.ID())\n\n\t\tid := NetID(task)\n\t\tnetns := NetNamespace(task)\n\n\t\tif err := cni.Remove(ctx, id, netns); err != nil {\n\t\t\treturn errors.Wrapf(err, \"Failed to remove network for task: %q, %v\", id, err)\n\t\t}\n\t\tlog.Printf(\"[Delete] removed: %s from namespace: %s, ID: %s\\n\", name, netns, id)\n\n\t\treturn nil\n\t}\n\n\treturn errors.Wrapf(containerErr, \"Unable to find container: %s, error: %s\", name, containerErr)\n}", "func (n *NetworkRemoveCommand) runNetworkRemove(args []string) error {\n\tname := args[0]\n\n\tctx := context.Background()\n\tapiClient := n.cli.Client()\n\tif err := apiClient.NetworkRemove(ctx, name); err != nil {\n\t\treturn err\n\t}\n\tfmt.Printf(\"Removed: %s\\n\", name)\n\treturn nil\n}", "func (c *MockVirtualNetworksClient) Delete(ctx context.Context, resourceGroupName, vnetName string) error {\n\t// Ignore resourceGroupName for simplicity.\n\tif _, ok := c.VNets[vnetName]; !ok {\n\t\treturn fmt.Errorf(\"%s does not exist\", vnetName)\n\t}\n\tdelete(c.VNets, vnetName)\n\treturn nil\n}", "func (d *driver) CreateNetwork(id string, option map[string]interface{}, nInfo driverapi.NetworkInfo, ipV4Data, ipV6Data []driverapi.IPAMData) error {\n\tif len(ipV4Data) == 0 || ipV4Data[0].Pool.String() == \"0.0.0.0/0\" {\n\t\treturn types.InvalidParameterErrorf(\"ipv4 pool is empty\")\n\t}\n\t// Sanity checks\n\td.Lock()\n\tif _, ok := d.networks[id]; ok {\n\t\td.Unlock()\n\t\treturn types.ForbiddenErrorf(\"network %s exists\", id)\n\t}\n\td.Unlock()\n\n\t// Parse and validate the config. It should not be conflict with existing networks' config\n\tconfig, err := parseNetworkOptions(id, option)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err = config.processIPAM(id, ipV4Data, ipV6Data); err != nil {\n\t\treturn err\n\t}\n\n\t// start the critical section, from this point onward we are dealing with the list of networks\n\t// so to be consistent we cannot allow that the list changes\n\td.configNetwork.Lock()\n\tdefer d.configNetwork.Unlock()\n\n\t// check network conflicts\n\tif err = d.checkConflict(config); err != nil {\n\t\tnerr, ok := err.(defaultBridgeNetworkConflict)\n\t\tif !ok {\n\t\t\treturn err\n\t\t}\n\t\t// Got a conflict with a stale default network, clean that up and continue\n\t\tlog.G(context.TODO()).Warn(nerr)\n\t\tif err := d.deleteNetwork(nerr.ID); err != nil {\n\t\t\tlog.G(context.TODO()).WithError(err).Debug(\"Error while cleaning up network on conflict\")\n\t\t}\n\t}\n\n\t// there is no conflict, now create the network\n\tif err = d.createNetwork(config); err != nil {\n\t\treturn err\n\t}\n\n\treturn d.storeUpdate(config)\n}", "func (c *WifiClient) Delete() *WifiDelete {\n\tmutation := newWifiMutation(c.config, OpDelete)\n\treturn &WifiDelete{config: c.config, hooks: c.Hooks(), mutation: mutation}\n}", "func (s *Service) DeleteNetworkRule(ruleID string) (string, error) {\n\tbody, err := s.deleteNetworkRuleResponseBody(ruleID)\n\n\treturn body.Data.TaskID, err\n}", "func (c *localController) DisconnectNetwork(name string, containerPid int) error {\n\tlogrus.Debugf(\"disconnecting %d from networks %s\", containerPid, name)\n\n\tpeer, err := c.ds.GetNetworkPeer(name, containerPid)\n\tif err != nil {\n\t\tif err == ds.ErrNetworkPeerDoesNotExist {\n\t\t\treturn fmt.Errorf(\"container %d is not connected to network %s\", containerPid, name)\n\t\t}\n\n\t\treturn err\n\t}\n\ttmpConfDir, err := ioutil.TempDir(\"\", \"circuit-\")\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer os.RemoveAll(tmpConfDir)\n\n\tcninet, nc, rt, err := c.getCniConfig(name, tmpConfDir, containerPid, peer.IfaceName)\n\tif err != nil {\n\t\tlogrus.Warnf(\"unable to detect peer: %s\", err)\n\t}\n\n\tif err := cninet.DelNetwork(nc, rt); err != nil {\n\t\tlogrus.Warnf(\"unable to disconnect: %s\", err)\n\t}\n\n\tif err := c.ds.DeleteNetworkPeer(name, containerPid); err != nil {\n\t\tlogrus.Fatal(err)\n\t}\n\n\treturn nil\n}", "func (s *Service) Delete(ctx context.Context, spec azure.Spec) error {\n\tnicSpec, ok := spec.(*Spec)\n\tif !ok {\n\t\treturn errors.New(\"invalid network interface Specification\")\n\t}\n\tklog.V(2).Infof(\"deleting nic %s\", nicSpec.Name)\n\tf, err := s.Client.Delete(ctx, s.Scope.ClusterConfig.ResourceGroup, nicSpec.Name)\n\tif err != nil && azure.ResourceNotFound(err) {\n\t\t// already deleted\n\t\treturn nil\n\t}\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"failed to delete network interface %s in resource group %s\", nicSpec.Name, s.Scope.ClusterConfig.ResourceGroup)\n\t}\n\n\terr = f.WaitForCompletionRef(ctx, s.Client.Client)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"cannot create, future response\")\n\t}\n\n\t_, err = f.Result(s.Client)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"result error\")\n\t}\n\tklog.V(2).Infof(\"successfully deleted nic %s\", nicSpec.Name)\n\treturn err\n}", "func (m *Setup) DelNetworks(ctx context.Context) error {\n\treturn m.command(ctx, \"DEL\")\n}", "func (plugin *Plugin) Del(args *cniSkel.CmdArgs) error {\n\t// Parse network configuration.\n\tnetConfig, err := config.New(args, false)\n\tif err != nil {\n\t\tlog.Errorf(\"Failed to parse netconfig from args: %v.\", err)\n\t\treturn err\n\t}\n\n\tlog.Infof(\"Executing DEL with netconfig: %+v.\", netConfig)\n\n\t// Derive names from CNI network config.\n\tpatNetNSName := fmt.Sprintf(patNetNSNameFormat, netConfig.BranchVlanID)\n\ttapBridgeName := fmt.Sprintf(tapBridgeNameFormat, netConfig.BranchVlanID)\n\ttapLinkName := args.IfName\n\ttargetNetNSName := args.Netns\n\n\t// Delete the tap link and veth pair from the target netns.\n\tplugin.deleteTapVethLinks(targetNetNSName, tapLinkName, tapBridgeName)\n\n\t// Search for the PAT network namespace.\n\tpatNetNS, err := netns.GetNetNSByName(patNetNSName)\n\tif err != nil {\n\t\t// Log and ignore the failure. DEL can be called multiple times and thus must be idempotent.\n\t\tlog.Errorf(\"Failed to find netns %s, ignoring: %v.\", patNetNSName, err)\n\t\treturn nil\n\t}\n\tlastVethLinkDeleted := false\n\n\t// In PAT network namespace...\n\terr = patNetNS.Run(func() error {\n\t\t// Check whether there are any remaining veth links connected to this bridge.\n\t\tifaces, _ := net.Interfaces()\n\t\tlog.Infof(\"Number of remaining links: %v.\", len(ifaces))\n\t\tif len(ifaces) == 4 {\n\t\t\t// Only VLAN link, bridge, dummy and loopback remain.\n\t\t\tlastVethLinkDeleted = true\n\t\t}\n\n\t\treturn nil\n\t})\n\n\t// If all veth links connected to this PAT bridge are deleted, clean up the PAT network\n\t// namespace and all virtual interfaces in it. Otherwise, leave it running.\n\tif lastVethLinkDeleted && netConfig.CleanupPATNetNS {\n\t\tlog.Infof(\"Deleting PAT network namespace: %v.\", patNetNSName)\n\t\terr = patNetNS.Close()\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Failed to delete netns: %v.\", err)\n\t\t}\n\t} else {\n\t\tlog.Infof(\"Skipping PAT netns deletion. Last veth link deleted: %t, cleanup PAT netns: %t.\",\n\t\t\tlastVethLinkDeleted, netConfig.CleanupPATNetNS)\n\t}\n\n\treturn nil\n}", "func (c *MockNetworkSecurityGroupsClient) Delete(ctx context.Context, resourceGroupName, nsgName string) error {\n\t// Ignore resourceGroupName for simplicity.\n\tif _, ok := c.NSGs[nsgName]; !ok {\n\t\treturn fmt.Errorf(\"%s does not exist\", nsgName)\n\t}\n\tdelete(c.NSGs, nsgName)\n\treturn nil\n}", "func newNetwork(networkType NetworkModel) network {\n\tswitch networkType {\n\tcase NoopNetworkModel:\n\t\treturn &noopNetwork{}\n\tcase DefaultNetworkModel:\n\t\treturn &defNetwork{}\n\tdefault:\n\t\treturn &noopNetwork{}\n\t}\n}", "func (c *TestClient) DeleteSubnetwork(project, region, name string) error {\n\tif c.DeleteSubnetworkFn != nil {\n\t\treturn c.DeleteSubnetworkFn(project, region, name)\n\t}\n\treturn c.client.DeleteSubnetwork(project, region, name)\n}", "func (v *Network) Remove() error {\n\treturn v.Client.NetworkRemove(context.Background(), v.Name)\n}", "func deleteHCNEndpoint(ep *hcn.HostComputeEndpoint, netns string) (string, error) {\n\tdefer func() { _ = recover() }()\n\tif ep == nil {\n\t\treturn \"\", nil\n\t}\n\n\t// remove endpoint from namespace\n\tif netns == \"\" {\n\t\tnetns = ep.HostComputeNamespace\n\t}\n\t_ = hcn.RemoveNamespaceEndpoint(netns, ep.Id)\n\n\t// delete endpoint\n\tif err := ep.Delete(); err != nil {\n\t\tif !hcn.IsNotFoundError(err) {\n\t\t\treturn ep.HostComputeNetwork, err\n\t\t}\n\t}\n\treturn ep.HostComputeNetwork, nil\n}", "func (c *Client) DisconnectNetwork(nwid string, cid string) error {\n\tendpoint := fmt.Sprintf(\"%snetworks/%s/disconnect\", baseAddr, nwid)\n\n\tmin := struct {\n\t\tContainer string `json:\"Container\"`\n\t}{\n\t\tContainer: cid,\n\t}\n\tb, err := json.Marshal(&min)\n\tif err != nil {\n\t\treturn err\n\t}\n\tr, err := c.http.Post(endpoint, \"application/json\", bytes.NewReader(b))\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn statusCode(r.StatusCode, http.StatusOK)\n}", "func (group *NodeGroup) Delete() error {\n\treturn cloudprovider.ErrNotImplemented\n}", "func (c *MockNatGatewaysClient) Delete(ctx context.Context, resourceGroupName, ngwName string) error {\n\t// Ignore resourceGroupName for simplicity.\n\tif _, ok := c.NGWs[ngwName]; !ok {\n\t\treturn fmt.Errorf(\"%s does not exist\", ngwName)\n\t}\n\tdelete(c.NGWs, ngwName)\n\treturn nil\n}", "func Delete(strinput2 string, tc net.Conn) {\n\tnetwork.SendDataMessage(&tc, 4, 6, 1, strinput2)\n}", "func (c *Client) DeleteNetconf(url string) error {\n\t_, err := c.httpRequest(url, \"DELETE\", bytes.Buffer{})\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func GetPartialDeleteNetworkHandler(path string, key string) obsidian.Handler {\n\treturn obsidian.Handler{\n\t\tPath: path,\n\t\tMethods: obsidian.DELETE,\n\t\tHandlerFunc: func(c echo.Context) error {\n\t\t\tnetworkID, nerr := obsidian.GetNetworkId(c)\n\t\t\tif nerr != nil {\n\t\t\t\treturn nerr\n\t\t\t}\n\t\t\tupdate := configurator.NetworkUpdateCriteria{\n\t\t\t\tID: networkID,\n\t\t\t\tConfigsToDelete: []string{key},\n\t\t\t}\n\t\t\terr := configurator.UpdateNetworks([]configurator.NetworkUpdateCriteria{update})\n\t\t\tif err != nil {\n\t\t\t\treturn obsidian.HttpError(err, http.StatusInternalServerError)\n\t\t\t}\n\t\t\treturn c.NoContent(http.StatusNoContent)\n\t\t},\n\t}\n}", "func (orgVdcNet *OpenApiOrgVdcNetwork) DeletNetworkDhcp() error {\n\tendpoint := types.OpenApiPathVersion1_0_0 + types.OpenApiEndpointOrgVdcNetworksDhcp\n\tapiVersion, err := orgVdcNet.client.getOpenApiHighestElevatedVersion(endpoint)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif orgVdcNet.OpenApiOrgVdcNetwork.ID == \"\" {\n\t\treturn fmt.Errorf(\"cannot delete Org VDC network DHCP configuration without ID\")\n\t}\n\n\turlRef, err := orgVdcNet.client.OpenApiBuildEndpoint(fmt.Sprintf(endpoint, orgVdcNet.OpenApiOrgVdcNetwork.ID))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = orgVdcNet.client.OpenApiDeleteItem(apiVersion, urlRef, nil, nil)\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error deleting Org VDC network DHCP configuration: %s\", err)\n\t}\n\n\treturn nil\n}", "func (s *Service) DeleteNetworkInterface(resourceGroup string, networkInterfaceName string) (armada.InterfacesDeleteFuture, error) {\n\t//JEB return s.scope.AirshipClients.Interfaces.Delete(s.scope.Context, resourceGroup, networkInterfaceName)\n\treturn armada.InterfacesDeleteFuture{}, nil\n}", "func (c *sonicwallNetworkPolicies) Delete(ctx context.Context, name string, opts v1.DeleteOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"sonicwallnetworkpolicies\").\n\t\tName(name).\n\t\tBody(&opts).\n\t\tDo(ctx).\n\t\tError()\n}", "func (a *HyperflexApiService) DeleteHyperflexClusterNetworkPolicyExecute(r ApiDeleteHyperflexClusterNetworkPolicyRequest) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = http.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tformFiles []formFile\n\t)\n\n\tlocalBasePath, err := a.client.cfg.ServerURLWithContext(r.ctx, \"HyperflexApiService.DeleteHyperflexClusterNetworkPolicy\")\n\tif err != nil {\n\t\treturn nil, &GenericOpenAPIError{error: err.Error()}\n\t}\n\n\tlocalVarPath := localBasePath + \"/api/v1/hyperflex/ClusterNetworkPolicies/{Moid}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"Moid\"+\"}\", url.PathEscape(parameterToString(r.moid, \"\")), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\treq, err := a.client.prepareRequest(r.ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, formFiles)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(req)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tlocalVarHTTPResponse.Body = ioutil.NopCloser(bytes.NewBuffer(localVarBody))\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := &GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 400 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 401 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 403 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tvar v Error\n\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\tif err != nil {\n\t\t\tnewErr.error = err.Error()\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tnewErr.model = v\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func DeleteMultiHop(exec executor.Executor, execnet, tonet, routingTable string, ref map[string]NetworkSettings) error {\n\tlocalNetGW, ok := ref[execnet]\n\tif !ok {\n\t\treturn fmt.Errorf(\"network %s not found in %v\", execnet, ref)\n\t}\n\n\texternalNet, ok := ref[tonet]\n\tif !ok {\n\t\treturn fmt.Errorf(\"network %s not found in %v\", tonet, ref)\n\t}\n\n\terr := routes.Delete(exec, fmt.Sprintf(\"%s/%d\", externalNet.IPAddress, externalNet.IPPrefixLen), localNetGW.IPAddress, routingTable)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = routes.Delete(exec, fmt.Sprintf(\"%s/%d\", externalNet.GlobalIPv6Address, externalNet.GlobalIPv6PrefixLen), localNetGW.GlobalIPv6Address, routingTable)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (nb *NetBuilder) DeleteEndpoint(nw *Network, ep *Endpoint) error {\n\t// Generate network name here as endpoint name is dependent upon network name.\n\tnw.Name = nb.generateHNSNetworkName(nw)\n\t// Query the namespace identifier.\n\tnsType, namespaceIdentifier := nb.getNamespaceIdentifier(ep)\n\n\t// Find the HNS endpoint ID.\n\tendpointName := nb.generateHNSEndpointName(nw.Name, namespaceIdentifier)\n\thnsEndpoint, err := hcsshim.GetHNSEndpointByName(endpointName)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Detach the HNS endpoint from the container's network namespace.\n\tlog.Infof(\"Detaching HNS endpoint %s from container %s netns.\", hnsEndpoint.Id, ep.ContainerID)\n\tif nsType == hcsNamespace {\n\t\t// Detach the HNS endpoint from the namespace, if we can.\n\t\t// HCN Namespace and HNS Endpoint have a 1-1 relationship, therefore,\n\t\t// even if detachment of endpoint from namespace fails, we can still proceed to delete it.\n\t\terr = hcn.RemoveNamespaceEndpoint(namespaceIdentifier, hnsEndpoint.Id)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Failed to detach endpoint, ignoring: %v\", err)\n\t\t}\n\t} else {\n\t\terr = hcsshim.HotDetachEndpoint(ep.ContainerID, hnsEndpoint.Id)\n\t\tif err != nil && err != hcsshim.ErrComputeSystemDoesNotExist {\n\t\t\treturn err\n\t\t}\n\n\t\t// The rest of the delete logic applies to infrastructure container only.\n\t\tif nsType == nonInfraContainerNS {\n\t\t\t// For non-infra containers, the network must not be deleted.\n\t\t\tnw.UseExisting = true\n\t\t\treturn nil\n\t\t}\n\t}\n\n\t// Delete the HNS endpoint.\n\tlog.Infof(\"Deleting HNS endpoint name: %s ID: %s\", endpointName, hnsEndpoint.Id)\n\t_, err = hnsEndpoint.Delete()\n\tif err != nil {\n\t\tlog.Errorf(\"Failed to delete HNS endpoint: %v.\", err)\n\t}\n\n\treturn err\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstance(ctx context.Context, name string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (t *KCPTransport) GetNetwork() string {\n\treturn \"udp\"\n}", "func NetworkDeinit() int {\n\treturn int(C.avformat_network_deinit())\n}", "func (client *VirtualNetworkLinksClient) deleteCreateRequest(ctx context.Context, resourceGroupName string, privateZoneName string, virtualNetworkLinkName string, options *VirtualNetworkLinksBeginDeleteOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/privateDnsZones/{privateZoneName}/virtualNetworkLinks/{virtualNetworkLinkName}\"\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif privateZoneName == \"\" {\n\t\treturn nil, errors.New(\"parameter privateZoneName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{privateZoneName}\", url.PathEscape(privateZoneName))\n\tif virtualNetworkLinkName == \"\" {\n\t\treturn nil, errors.New(\"parameter virtualNetworkLinkName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{virtualNetworkLinkName}\", url.PathEscape(virtualNetworkLinkName))\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodDelete, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2020-06-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\tif options != nil && options.IfMatch != nil {\n\t\treq.Raw().Header.Set(\"If-Match\", *options.IfMatch)\n\t}\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func (manager *syncerManager) ensureDeleteNetworkEndpointGroup(name, zone string) error {\n\t_, err := manager.cloud.GetNetworkEndpointGroup(name, zone, meta.VersionGA)\n\tif err != nil {\n\t\t// Assume error is caused by not existing\n\t\treturn nil\n\t}\n\tklog.V(2).Infof(\"Deleting NEG %q in %q.\", name, zone)\n\treturn manager.cloud.DeleteNetworkEndpointGroup(name, zone, meta.VersionGA)\n}", "func (c *Client) StopNetwork() error {\n\treturn c.Post(\"/stop\", nil, nil)\n}", "func NetworkClear(protocol string, comment string, table string) error {\n\treturn iptablesClear(protocol, fmt.Sprintf(\"LXD network %s\", comment),\n\t\ttable)\n}", "func (n *cni) remove(pod Pod, networkNS NetworkNamespace) error {\n\tif err := removeNetworkCommon(networkNS); err != nil {\n\t\treturn err\n\t}\n\n\tif err := n.deleteVirtInterfaces(networkNS); err != nil {\n\t\treturn err\n\t}\n\n\treturn deleteNetNS(networkNS.NetNsPath, true)\n}", "func detachNetwork(appModel *models.App, componentModel *models.Component) error {\n\tdisplay.StartTask(\"Releasing IPs\")\n\tdefer display.StopTask()\n\n\tif componentModel.IPAddr() == \"\" {\n\t\treturn nil\n\t}\n\n\t// return the external IP\n\t// don't return the external IP if this is portal\n\tif componentModel.Name != \"portal\" && appModel.LocalIPs[componentModel.Name] == \"\" {\n\t\tip := net.ParseIP(componentModel.IPAddr())\n\t\tif err := dhcp.ReturnIP(ip); err != nil {\n\t\t\tlumber.Error(\"component:detachNetwork:dhcp.ReturnIP(%s): %s\", ip, err.Error())\n\t\t\tdisplay.ErrorTask()\n\t\t\treturn util.ErrorAppend(err, \"failed to release IP back to pool\")\n\t\t}\n\t}\n\n\treturn nil\n}", "func (a *HyperflexApiService) DeleteHyperflexClusterNetworkPolicy(ctx context.Context, moid string) ApiDeleteHyperflexClusterNetworkPolicyRequest {\n\treturn ApiDeleteHyperflexClusterNetworkPolicyRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\tmoid: moid,\n\t}\n}", "func (s *BoltState) NetworkDisconnect(ctr *Container, network string) error {\n\tif !s.valid {\n\t\treturn define.ErrDBClosed\n\t}\n\n\tif !ctr.valid {\n\t\treturn define.ErrCtrRemoved\n\t}\n\n\tif network == \"\" {\n\t\treturn fmt.Errorf(\"network names must not be empty: %w\", define.ErrInvalidArg)\n\t}\n\n\tctrID := []byte(ctr.ID())\n\n\tdb, err := s.getDBCon()\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer s.deferredCloseDBCon(db)\n\n\treturn db.Update(func(tx *bolt.Tx) error {\n\t\tctrBucket, err := getCtrBucket(tx)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tdbCtr := ctrBucket.Bucket(ctrID)\n\t\tif dbCtr == nil {\n\t\t\tctr.valid = false\n\t\t\treturn fmt.Errorf(\"container %s does not exist in database: %w\", ctr.ID(), define.ErrNoSuchCtr)\n\t\t}\n\n\t\tctrAliasesBkt := dbCtr.Bucket(aliasesBkt)\n\t\tctrNetworksBkt := dbCtr.Bucket(networksBkt)\n\t\tif ctrNetworksBkt == nil {\n\t\t\treturn fmt.Errorf(\"container %s is not connected to any networks, so cannot disconnect: %w\", ctr.ID(), define.ErrNoSuchNetwork)\n\t\t}\n\t\tnetConnected := ctrNetworksBkt.Get([]byte(network))\n\t\tif netConnected == nil {\n\t\t\treturn fmt.Errorf(\"container %s is not connected to network %q: %w\", ctr.ID(), network, define.ErrNoSuchNetwork)\n\t\t}\n\n\t\tif err := ctrNetworksBkt.Delete([]byte(network)); err != nil {\n\t\t\treturn fmt.Errorf(\"removing container %s from network %s: %w\", ctr.ID(), network, err)\n\t\t}\n\n\t\tif ctrAliasesBkt != nil {\n\t\t\tbktExists := ctrAliasesBkt.Bucket([]byte(network))\n\t\t\tif bktExists == nil {\n\t\t\t\treturn nil\n\t\t\t}\n\n\t\t\tif err := ctrAliasesBkt.DeleteBucket([]byte(network)); err != nil {\n\t\t\t\treturn fmt.Errorf(\"removing container %s network aliases for network %s: %w\", ctr.ID(), network, err)\n\t\t\t}\n\t\t}\n\n\t\treturn nil\n\t})\n}", "func (client *IPPoolClient) Delete(pool *hcipamTypes.NSIPPool) error {\n\te := client.Cli\n\t_, err := concurrency.NewSTM(e, func(s concurrency.STM) error {\n\t\thcPoolKey := fmt.Sprintf(hcipamTypes.NS_POOL_Key, pool.Namespace, pool.Name)\n\t\ts.Del(hcPoolKey)\n\t\treturn nil\n\t})\n\treturn err\n}" ]
[ "0.7176192", "0.69172025", "0.6913784", "0.6838265", "0.68334675", "0.68241584", "0.6823366", "0.6820862", "0.6820698", "0.68127763", "0.6747861", "0.66906804", "0.6568566", "0.6511086", "0.6478214", "0.64192456", "0.6379316", "0.63729864", "0.62730604", "0.6255611", "0.6218766", "0.61195534", "0.61024374", "0.6018943", "0.60107446", "0.59829855", "0.5981828", "0.5974403", "0.5966099", "0.5923801", "0.5876423", "0.58396995", "0.5800201", "0.574496", "0.574327", "0.5719466", "0.56747335", "0.56747335", "0.566433", "0.5649748", "0.5640617", "0.5624475", "0.5623869", "0.56016594", "0.5577182", "0.5576579", "0.55694944", "0.55446887", "0.5536818", "0.55297196", "0.5522184", "0.5513847", "0.5497154", "0.549249", "0.54871917", "0.5481159", "0.5464982", "0.54604244", "0.54401195", "0.5439132", "0.5422209", "0.538016", "0.53602034", "0.5352372", "0.53350633", "0.5319037", "0.530211", "0.52970326", "0.52820665", "0.52806437", "0.5261812", "0.5250379", "0.52335465", "0.52141404", "0.5210464", "0.5206305", "0.5187348", "0.51765686", "0.51749367", "0.51650447", "0.51351535", "0.5134688", "0.51291794", "0.5129059", "0.5115759", "0.510273", "0.5088449", "0.5088257", "0.50851554", "0.50816", "0.5069336", "0.5068083", "0.5067454", "0.5058759", "0.505663", "0.5055293", "0.50491744", "0.50489724", "0.5048508", "0.50432277" ]
0.7206242
0
DeleteSubnetwork uses the override method DeleteSubnetworkFn or the real implementation.
func (c *TestClient) DeleteSubnetwork(project, region, name string) error { if c.DeleteSubnetworkFn != nil { return c.DeleteSubnetworkFn(project, region, name) } return c.client.DeleteSubnetwork(project, region, name) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (a *Azure) DeleteSubnetwork(ctx *lepton.Context, subnetID string) error {\n\tlogger := ctx.Logger()\n\n\tsubnetsClient, err := a.getSubnetsClient()\n\tif err != nil {\n\t\treturn err\n\t}\n\tvnetClient, err := a.getVnetClient()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tsubnetName := getAzureResourceNameFromID(subnetID)\n\tvnName := getAzureVirtualNetworkFromID(subnetID)\n\n\tsubnet, err := subnetsClient.Get(context.TODO(), a.groupName, vnName, subnetName, \"\")\n\tif err != nil {\n\t\tctx.Logger().Error(err)\n\t\treturn fmt.Errorf(\"failed getting subnet\")\n\t}\n\n\tvirtualNetwork, err := vnetClient.Get(context.TODO(), a.groupName, vnName, \"\")\n\tif err != nil {\n\t\tctx.Logger().Error(err)\n\t\treturn errors.New(\"failed getting virtual network\")\n\t}\n\n\tif hasAzureOpsTags(virtualNetwork.Tags) && (subnet.IPConfigurations == nil || len(*subnet.IPConfigurations) == 0) {\n\t\tlogger.Infof(\"Deleting %s...\", *subnet.ID)\n\t\tsubnetDeleteTask, err := subnetsClient.Delete(context.TODO(), a.groupName, subnetName, subnetName)\n\t\tif err != nil {\n\t\t\tlogger.Error(err)\n\t\t\treturn errors.New(\"error deleting subnet\")\n\t\t}\n\n\t\terr = subnetDeleteTask.WaitForCompletionRef(context.TODO(), subnetsClient.Client)\n\t\tif err != nil {\n\t\t\tlogger.Error(err)\n\t\t\treturn errors.New(\"error waiting for subnet deletion\")\n\t\t}\n\n\t\tlogger.Infof(\"Deleting virtualNetworks/%s\", vnName)\n\t\tvnDeleteTask, err := vnetClient.Delete(context.TODO(), a.groupName, vnName)\n\t\tif err != nil {\n\t\t\tlogger.Error(err)\n\t\t\treturn errors.New(\"error deleting virtual network\")\n\t\t}\n\n\t\terr = vnDeleteTask.WaitForCompletionRef(context.TODO(), vnetClient.Client)\n\t\tif err != nil {\n\t\t\tlogger.Error(err)\n\t\t\treturn errors.New(\"error waiting for virtual network deletion\")\n\t\t}\n\t} else {\n\t\treturn errors.New(\"other devices are connected to the same subnet\")\n\t}\n\n\treturn nil\n}", "func DeleteVirtualNetworkSubnet() {}", "func DeleteVirtualNetworkSubnet() {}", "func (s *Stack) DeleteNetwork(ref string) (err error) {\n\ttheNetwork, err := s.GetNetwork(ref)\n\tif err != nil {\n\t\tif gerr, ok := err.(*googleapi.Error); ok {\n\t\t\tif gerr.Code != 404 {\n\t\t\t\treturn err\n\t\t\t}\n\t\t} else {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif theNetwork == nil {\n\t\treturn fail.Errorf(\n\t\t\tfmt.Sprintf(\"delete network failed: unexpected nil network when looking for [%s]\", ref), err,\n\t\t)\n\t}\n\n\tif !theNetwork.OK() {\n\t\tlogrus.Warnf(\"Missing data in network: %s\", spew.Sdump(theNetwork))\n\t}\n\n\tcompuService := s.ComputeService\n\tsubnetwork, err := compuService.Subnetworks.Get(s.GcpConfig.ProjectID, s.GcpConfig.Region, theNetwork.Name).Do()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\topp, err := compuService.Subnetworks.Delete(s.GcpConfig.ProjectID, s.GcpConfig.Region, subnetwork.Name).Do()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\toco := OpContext{\n\t\tOperation: opp,\n\t\tProjectID: s.GcpConfig.ProjectID,\n\t\tService: compuService,\n\t\tDesiredState: \"DONE\",\n\t}\n\n\terr = waitUntilOperationIsSuccessfulOrTimeout(oco, temporal.GetMinDelay(), temporal.GetHostCleanupTimeout())\n\tif err != nil {\n\t\tswitch err.(type) {\n\t\tcase fail.ErrTimeout:\n\t\t\tlogrus.Warnf(\"Timeout waiting for subnetwork deletion\")\n\t\t\treturn err\n\t\tdefault:\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Delete routes and firewall\n\tfirewallRuleName := fmt.Sprintf(\"%s-%s-all-in\", s.GcpConfig.NetworkName, subnetwork.Name)\n\tfws, err := compuService.Firewalls.Get(s.GcpConfig.ProjectID, firewallRuleName).Do()\n\tif fws != nil && err == nil {\n\t\topp, operr := compuService.Firewalls.Delete(s.GcpConfig.ProjectID, firewallRuleName).Do()\n\t\tif operr == nil {\n\t\t\toco := OpContext{\n\t\t\t\tOperation: opp,\n\t\t\t\tProjectID: s.GcpConfig.ProjectID,\n\t\t\t\tService: compuService,\n\t\t\t\tDesiredState: \"DONE\",\n\t\t\t}\n\n\t\t\toperr = waitUntilOperationIsSuccessfulOrTimeout(\n\t\t\t\toco, temporal.GetMinDelay(), temporal.GetHostCleanupTimeout(),\n\t\t\t)\n\t\t\tif operr != nil {\n\t\t\t\tlogrus.Warn(operr)\n\t\t\t}\n\t\t}\n\t}\n\tif err != nil {\n\t\tlogrus.Warn(err)\n\t}\n\n\tnatRuleName := fmt.Sprintf(\"%s-%s-nat-allowed\", s.GcpConfig.NetworkName, subnetwork.Name)\n\tnws, err := compuService.Routes.Get(s.GcpConfig.ProjectID, natRuleName).Do()\n\tif nws != nil && err == nil {\n\t\topp, operr := compuService.Routes.Delete(s.GcpConfig.ProjectID, natRuleName).Do()\n\t\tif operr == nil {\n\t\t\toco := OpContext{\n\t\t\t\tOperation: opp,\n\t\t\t\tProjectID: s.GcpConfig.ProjectID,\n\t\t\t\tService: compuService,\n\t\t\t\tDesiredState: \"DONE\",\n\t\t\t}\n\n\t\t\toperr = waitUntilOperationIsSuccessfulOrTimeout(\n\t\t\t\toco, temporal.GetMinDelay(), temporal.GetHostCleanupTimeout(),\n\t\t\t)\n\t\t\tif operr != nil {\n\t\t\t\tlogrus.Warn(operr)\n\t\t\t}\n\t\t}\n\t}\n\tif err != nil {\n\t\tlogrus.Warn(err)\n\t}\n\n\treturn nil\n}", "func dockerNetDelete(subnetID string) error {\n\t_, err := exec.Command(\"docker\", \"network\", \"rm\", subnetID).CombinedOutput()\n\treturn err\n}", "func (o RegionNetworkEndpointGroupOutput) Subnetwork() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *RegionNetworkEndpointGroup) pulumi.StringPtrOutput { return v.Subnetwork }).(pulumi.StringPtrOutput)\n}", "func (o NetworkInterfaceOutput) Subnetwork() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v NetworkInterface) *string { return v.Subnetwork }).(pulumi.StringPtrOutput)\n}", "func RemoveSubnet(connectionName string, vpcName string, nameID string, force string) (bool, error) {\n\tcblog.Info(\"call RemoveSubnet()\")\n\n\t// check empty and trim user inputs\n connectionName, err := EmptyCheckAndTrim(\"connectionName\", connectionName)\n if err != nil {\n return false, err\n\t\tcblog.Error(err)\n }\n\n vpcName, err = EmptyCheckAndTrim(\"vpcName\", vpcName)\n if err != nil {\n return false, err\n\t\tcblog.Error(err)\n }\n\n nameID, err = EmptyCheckAndTrim(\"nameID\", nameID)\n if err != nil {\n return false, err\n\t\tcblog.Error(err)\n }\n\n\tcldConn, err := ccm.GetCloudConnection(connectionName)\n\tif err != nil {\n\t\tcblog.Error(err)\n\t\treturn false, err\n\t}\n\n\thandler, err := cldConn.CreateVPCHandler()\n\tif err != nil {\n\t\tcblog.Error(err)\n\t\treturn false, err\n\t}\n\n\tvpcSPLock.Lock(connectionName, vpcName)\n\tdefer vpcSPLock.Unlock(connectionName, vpcName)\n\n\t// (1) get spiderIID for creating driverIID\n\tiidInfo, err := iidRWLock.GetIID(iidm.SUBNETGROUP, connectionName, vpcName, cres.IID{nameID, \"\"})\n\tif err != nil {\n\t\tcblog.Error(err)\n\t\treturn false, err\n\t}\n\n\t// (2) delete Resource(SystemId)\n\tdriverIId := getDriverIID(iidInfo.IId)\n\tresult := false\n\n\n\tiidVPCInfo, err := iidRWLock.GetIID(iidm.IIDSGROUP, connectionName, rsVPC, cres.IID{vpcName, \"\"})\n\tif err != nil {\n\t\tcblog.Error(err)\n\t\treturn false, err\n\t}\n\tresult, err = handler.(cres.VPCHandler).RemoveSubnet(getDriverIID(iidVPCInfo.IId), driverIId)\n\tif err != nil {\n\t\tcblog.Error(err)\n\t\tif force != \"true\" {\n\t\t\treturn false, err\n\t\t}\n\t}\n\n\tif force != \"true\" {\n\t\tif result == false {\n\t\t\treturn result, nil\n\t\t}\n\t}\n\n\t// (3) delete IID\n\t_, err = iidRWLock.DeleteIID(iidm.SUBNETGROUP, connectionName, vpcName, cres.IID{nameID, \"\"})\n\tif err != nil {\n\t\tcblog.Error(err)\n\t\tif force != \"true\" {\n\t\t\treturn false, err\n\t\t}\n\t}\n\n\n\treturn result, nil\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv6(ctx context.Context, name string, index int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv6/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (o RouterInterfaceOutput) Subnetwork() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v RouterInterface) *string { return v.Subnetwork }).(pulumi.StringPtrOutput)\n}", "func (o NetworkInterfaceResponseOutput) Subnetwork() pulumi.StringOutput {\n\treturn o.ApplyT(func(v NetworkInterfaceResponse) string { return v.Subnetwork }).(pulumi.StringOutput)\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv6RouterAdvertisement(ctx context.Context, name string, index int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv6/frinx-openconfig-if-ip:router-advertisement/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (o LookupRegionNetworkEndpointGroupResultOutput) Subnetwork() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupRegionNetworkEndpointGroupResult) string { return v.Subnetwork }).(pulumi.StringOutput)\n}", "func (c *SubresourceClient) Delete(namespace, name string) (e error) {\n\tif c.Error != \"\" {\n\t\te = fmt.Errorf(c.Error)\n\t}\n\treturn\n}", "func (c *TestClient) GetSubnetwork(project, region, name string) (*compute.Subnetwork, error) {\n\tif c.GetSubnetworkFn != nil {\n\t\treturn c.GetSubnetworkFn(project, region, name)\n\t}\n\treturn c.client.GetSubnetwork(project, region, name)\n}", "func (m *InterfacesClientMock) Delete(ctx context.Context, resourceGroupName string, networkInterfaceName string) *retry.Error {\n\targs := m.Called(resourceGroupName, networkInterfaceName)\n\tif args.Error(1) != nil {\n\t\treturn &retry.Error{RawError: args.Error(1)}\n\t}\n\treturn nil\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv4(ctx context.Context, name string, index int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv4/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func SubDelete(w http.ResponseWriter, r *http.Request) {\n\n\t// Init output\n\toutput := []byte(\"\")\n\n\t// Add content type header to the response\n\tcontentType := \"application/json\"\n\tcharset := \"utf-8\"\n\tw.Header().Add(\"Content-Type\", fmt.Sprintf(\"%s; charset=%s\", contentType, charset))\n\n\t// Grab url path variables\n\turlVars := mux.Vars(r)\n\n\t// Grab context references\n\trefStr := gorillaContext.Get(r, \"str\").(stores.Store)\n\tprojectUUID := gorillaContext.Get(r, \"auth_project_uuid\").(string)\n\n\t// Get Result Object\n\tresults, err := subscriptions.Find(projectUUID, \"\", urlVars[\"subscription\"], \"\", 0, refStr)\n\tif err != nil {\n\t\terr := APIErrGenericBackend()\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// If not found\n\tif results.Empty() {\n\t\terr := APIErrorNotFound(\"Subscription\")\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\terr = subscriptions.RemoveSub(projectUUID, urlVars[\"subscription\"], refStr)\n\tif err != nil {\n\t\tif err.Error() == \"not found\" {\n\t\t\terr := APIErrorNotFound(\"Subscription\")\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\t\terr := APIErrGenericInternal(err.Error())\n\t\trespondErr(w, err)\n\t\treturn\n\t}\n\n\t// if it is a push sub and it is also has a verified push endpoint, deactivate it\n\tif results.Subscriptions[0].PushCfg != (subscriptions.PushConfig{}) {\n\t\tif results.Subscriptions[0].PushCfg.Verified {\n\t\t\tpr := make(map[string]string)\n\t\t\tapsc := gorillaContext.Get(r, \"apsc\").(push.Client)\n\t\t\tpr[\"message\"] = apsc.DeactivateSubscription(context.TODO(), results.Subscriptions[0].FullName).Result()\n\t\t\tb, _ := json.Marshal(pr)\n\t\t\toutput = b\n\t\t}\n\t}\n\trespondOK(w, output)\n}", "func (o RouterInterfaceResponseOutput) Subnetwork() pulumi.StringOutput {\n\treturn o.ApplyT(func(v RouterInterfaceResponse) string { return v.Subnetwork }).(pulumi.StringOutput)\n}", "func DeleteNetworkSecurityGroupRule() {}", "func (m *MockClient) DeleteSubnet(arg0 *ec2.DeleteSubnetInput) (*ec2.DeleteSubnetOutput, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"DeleteSubnet\", arg0)\n\tret0, _ := ret[0].(*ec2.DeleteSubnetOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (d *Driver) DeleteNetwork(r *pluginNet.DeleteNetworkRequest) error {\n\tdefer osl.InitOSContext()()\n\tnid := r.NetworkID\n\tlogrus.Infof(\"DeleteNetwork macvlan nid=%s\", nid)\n\tif nid == \"\" {\n\t\treturn fmt.Errorf(\"invalid network id\")\n\t}\n\n\tn := d.network(nid)\n\tif n == nil {\n\t\treturn fmt.Errorf(\"network id %s not found\", nid)\n\t}\n\t// if the driver created the slave interface, delete it, otherwise leave it\n\tif ok := n.config.CreatedSlaveLink; ok {\n\t\t// if the interface exists, only delete if it matches iface.vlan or dummy.net_id naming\n\t\tif ok := parentExists(n.config.Parent); ok {\n\t\t\t// only delete the link if it is named the net_id\n\t\t\tif n.config.Parent == getDummyName(stringid.TruncateID(nid)) {\n\t\t\t\terr := delDummyLink(n.config.Parent)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlogrus.Errorf(\"link %s was not deleted, continuing the delete network operation: %v\",\n\t\t\t\t\t\tn.config.Parent, err)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\t// only delete the link if it matches iface.vlan naming\n\t\t\t\terr := delVlanLink(n.config.Parent)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlogrus.Errorf(\"link %s was not deleted, continuing the delete network operation: %v\",\n\t\t\t\t\t\tn.config.Parent, err)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tfor _, ep := range n.endpoints {\n\t\tif link, err := ns.NlHandle().LinkByName(ep.srcName); err == nil {\n\t\t\tns.NlHandle().LinkDel(link)\n\t\t\tlogrus.Infof(\"DeleteNetwork delete macvlan link %s\", ep.srcName)\n\t\t}\n\n\t\tif err := d.store.StoreDelete(ep); err != nil {\n\t\t\tlogrus.Warnf(\"Failed to remove macvlan endpoint %s from store: %v\", ep.id[0:7], err)\n\t\t}\n\t}\n\t// delete the *network\n\td.deleteNetwork(nid)\n\treturn nil\n}", "func NewDelSubDeviceWithCoreRequestWithoutParam() *DelSubDeviceWithCoreRequest {\n\n return &DelSubDeviceWithCoreRequest{\n JDCloudRequest: core.JDCloudRequest{\n URL: \"/regions/{regionId}/instances/{instanceId}/products/{productKey}/edges/{edgeName}:delSubDevice\",\n Method: \"POST\",\n Header: nil,\n Version: \"v2\",\n },\n }\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv6Config(ctx context.Context, name string, index int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv6/frinx-openconfig-if-ip:config/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv6Unnumbered(ctx context.Context, name string, index int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv6/frinx-openconfig-if-ip:unnumbered/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (f *FakeInstance) DeleteReverseIPv6(_ context.Context, _, _ string) error {\n\tpanic(\"implement me\")\n}", "func (c *MockSubnetsClient) Delete(ctx context.Context, resourceGroupName, vnetName, subnetName string) error {\n\t// Ignore resourceGroupName and virtualNetworkName for simplicity.\n\tif _, ok := c.Subnets[subnetName]; !ok {\n\t\treturn fmt.Errorf(\"%s does not exist\", subnetName)\n\t}\n\tdelete(c.Subnets, subnetName)\n\treturn nil\n}", "func DeleteVNet(vNetName string, rgroupName string) {\n\ta := wow.New(os.Stdout, spin.Get(spin.Dots), \"Deleting virtual network : \"+vNetName)\n\ta.Start()\n\ttime.Sleep(2 * time.Second)\n\ta.Text(\"This would take a few minutes...\").Spinner(spin.Get(spin.Dots))\n\tcmd := exec.Command(\"az\", \"network\", \"vnet\", \"delete\", \"--name\",\n\t\tvNetName, \"--resource-group\", rgroupName)\n\tvar out bytes.Buffer\n\tvar stderr bytes.Buffer\n\tcmd.Stdout = &out\n\tcmd.Stderr = &stderr\n\terr := cmd.Run()\n\tif err != nil {\n\t\tfmt.Println(fmt.Sprint(err) + \": \" + stderr.String())\n\t\treturn\n\t}\n\ta.PersistWith(spin.Spinner{}, \"....\")\n\tfmt.Println(\"Result: \" + out.String())\n}", "func (o ClusterOutput) Subnetwork() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Cluster) pulumi.StringOutput { return v.Subnetwork }).(pulumi.StringOutput)\n}", "func (o *DeleteSubnetParams) WithHTTPClient(client *http.Client) *DeleteSubnetParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv6RouterAdvertisementConfig(ctx context.Context, name string, index int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv6/frinx-openconfig-if-ip:router-advertisement/frinx-openconfig-if-ip:config/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func TestCreateModifyDeleteNetwork(t *testing.T) {\n\trecord(t, \"createmodifydeletenetwork\", func(t *testing.T, svc *Service) {\n\t\tnetwork, err := svc.CreateNetwork(&request.CreateNetworkRequest{\n\t\t\tName: \"test private network (test)\",\n\t\t\tZone: \"fi-hel2\",\n\t\t\tIPNetworks: []upcloud.IPNetwork{\n\t\t\t\t{\n\t\t\t\t\tAddress: \"172.16.0.0/22\",\n\t\t\t\t\tDHCP: upcloud.True,\n\t\t\t\t\tDHCPDefaultRoute: upcloud.False,\n\t\t\t\t\tDHCPDns: []string{\n\t\t\t\t\t\t\"172.16.0.10\",\n\t\t\t\t\t\t\"172.16.1.10\",\n\t\t\t\t\t},\n\t\t\t\t\tFamily: upcloud.IPAddressFamilyIPv4,\n\t\t\t\t\tGateway: \"172.16.0.1\",\n\t\t\t\t},\n\t\t\t},\n\t\t})\n\t\trequire.NoError(t, err)\n\t\tassert.NotEmpty(t, network.UUID)\n\t\tassert.Equal(t, \"test private network (test)\", network.Name)\n\n\t\tpostModifyNetwork, err := svc.ModifyNetwork(&request.ModifyNetworkRequest{\n\t\t\tUUID: network.UUID,\n\t\t\tName: \"modified private network (test)\",\n\t\t})\n\t\trequire.NoError(t, err)\n\t\tassert.Equal(t, \"modified private network (test)\", postModifyNetwork.Name)\n\n\t\tserverDetails, err := createServer(svc, \"TestCreateModifyDeleteNetwork\")\n\t\trequire.NoError(t, err)\n\n\t\terr = stopServer(svc, serverDetails.UUID)\n\t\trequire.NoError(t, err)\n\n\t\tiface, err := svc.CreateNetworkInterface(&request.CreateNetworkInterfaceRequest{\n\t\t\tServerUUID: serverDetails.UUID,\n\t\t\tNetworkUUID: postModifyNetwork.UUID,\n\t\t\tType: postModifyNetwork.Type,\n\t\t\tIPAddresses: []request.CreateNetworkInterfaceIPAddress{\n\t\t\t\t{\n\t\t\t\t\tFamily: upcloud.IPAddressFamilyIPv4,\n\t\t\t\t},\n\t\t\t},\n\t\t})\n\t\trequire.NoError(t, err)\n\t\tassert.NotEmpty(t, iface.IPAddresses)\n\t\tassert.NotEmpty(t, iface.IPAddresses[0].Address)\n\n\t\tmodifyIface, err := svc.ModifyNetworkInterface(&request.ModifyNetworkInterfaceRequest{\n\t\t\tServerUUID: serverDetails.UUID,\n\t\t\tCurrentIndex: iface.Index,\n\t\t\tNewIndex: iface.Index + 1,\n\t\t})\n\t\trequire.NoError(t, err)\n\t\tassert.Equal(t, iface.Index+1, modifyIface.Index)\n\n\t\terr = svc.DeleteNetworkInterface(&request.DeleteNetworkInterfaceRequest{\n\t\t\tServerUUID: serverDetails.UUID,\n\t\t\tIndex: modifyIface.Index,\n\t\t})\n\t\trequire.NoError(t, err)\n\n\t\terr = svc.DeleteNetwork(&request.DeleteNetworkRequest{\n\t\t\tUUID: network.UUID,\n\t\t})\n\t\trequire.NoError(t, err)\n\n\t\tnetworks, err := svc.GetNetworksInZone(&request.GetNetworksInZoneRequest{\n\t\t\tZone: network.Zone,\n\t\t})\n\t\trequire.NoError(t, err)\n\n\t\tvar found bool\n\t\tfor _, n := range networks.Networks {\n\t\t\tif n.UUID == network.UUID {\n\t\t\t\tfound = true\n\t\t\t}\n\t\t}\n\t\tassert.False(t, found)\n\t})\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv4Unnumbered(ctx context.Context, name string, index int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv4/frinx-openconfig-if-ip:unnumbered/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (o LookupWorkstationClusterResultOutput) Subnetwork() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupWorkstationClusterResult) string { return v.Subnetwork }).(pulumi.StringOutput)\n}", "func (a *ASA) DeleteNetworkService(n interface{}) error {\n\tvar err error\n\tvar objectID string\n\n\tswitch v := n.(type) {\n\tcase *ReferenceObject:\n\t\tobjectID = v.ObjectID\n\tcase *NetworkService:\n\t\tobjectID = v.ObjectID\n\tcase string:\n\t\tobjectID = v\n\tdefault:\n\t\treturn fmt.Errorf(\"unknown type\")\n\t}\n\n\tif objectID == \"\" {\n\t\treturn fmt.Errorf(\"error objectid is null\")\n\t}\n\n\terr = a.Delete(fmt.Sprintf(\"%s/%s\", apiNetworkServicesEndpoint, objectID))\n\tif err != nil {\n\t\tif a.debug {\n\t\t\tglog.Errorf(\"Error: %s\\n\", err)\n\t\t}\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (sm *Statemgr) OnNetworkSecurityPolicyOperDelete(nodeID string, objinfo *netproto.NetworkSecurityPolicy) error {\n\treturn nil\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv6UnnumberedConfig(ctx context.Context, name string, index int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv6/frinx-openconfig-if-ip:unnumbered/frinx-openconfig-if-ip:config/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv4Config(ctx context.Context, name string, index int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv4/frinx-openconfig-if-ip:config/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (s *SubnetworkServer) DeleteComputeSubnetwork(ctx context.Context, request *computepb.DeleteComputeSubnetworkRequest) (*emptypb.Empty, error) {\n\n\tcl, err := createConfigSubnetwork(ctx, request.GetServiceAccountFile())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &emptypb.Empty{}, cl.DeleteSubnetwork(ctx, ProtoToSubnetwork(request.GetResource()))\n\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv6Addresses(ctx context.Context, name string, index int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv6/frinx-openconfig-if-ip:addresses/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (s *Service) Delete(ctx context.Context, spec azure.Spec) error {\n\tnicSpec, ok := spec.(*Spec)\n\tif !ok {\n\t\treturn errors.New(\"invalid network interface Specification\")\n\t}\n\tklog.V(2).Infof(\"deleting nic %s\", nicSpec.Name)\n\tf, err := s.Client.Delete(ctx, s.Scope.ClusterConfig.ResourceGroup, nicSpec.Name)\n\tif err != nil && azure.ResourceNotFound(err) {\n\t\t// already deleted\n\t\treturn nil\n\t}\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"failed to delete network interface %s in resource group %s\", nicSpec.Name, s.Scope.ClusterConfig.ResourceGroup)\n\t}\n\n\terr = f.WaitForCompletionRef(ctx, s.Client.Client)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"cannot create, future response\")\n\t}\n\n\t_, err = f.Result(s.Client)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"result error\")\n\t}\n\tklog.V(2).Infof(\"successfully deleted nic %s\", nicSpec.Name)\n\treturn err\n}", "func (de *DockerEngine) getSubNet() (string, error) {\n\tde.subNetMu.Lock()\n\tdefer de.subNetMu.Unlock()\n\n\taddrs, err := net.InterfaceAddrs()\n\n\tif err != nil {\n\t\treturn \"\", errors.Wrap(err, \"Error getting network addresses\")\n\t}\n\n\tvar nets []*net.IPNet\n\n\tfor _, addr := range addrs {\n\t\tdockerLog.Debugf(\"Inspecting interface %s\", addr.String())\n\n\t\t_, n, err := net.ParseCIDR(addr.String())\n\n\t\tif err != nil {\n\t\t\tdockerLog.Warningf(\"Error parsing address: %s\", addr.String())\n\n\t\t\tcontinue\n\t\t}\n\n\t\tnets = append(nets, n)\n\t}\n\n\tnetaddr := func() string {\n\t\ttpl := \"10.%d.%d.0/24\"\n\n\t\treturn fmt.Sprintf(tpl, de.subNetOct1, de.subNetOct2)\n\t}\n\n\t_, pnet, _ := net.ParseCIDR(netaddr())\n\n\tfor {\n\t\t// Find non-overlapping network\n\t\toverlap := false\n\n\t\tfor _, n := range nets {\n\t\t\tif lib.NetsOverlap(pnet, n) {\n\t\t\t\toverlap = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\n\t\tif overlap {\n\t\t\tde.subNetOct2 += 1\n\n\t\t\tif de.subNetOct2 > 255 {\n\t\t\t\tde.subNetOct1 += 1\n\t\t\t\tde.subNetOct2 = 0\n\t\t\t}\n\n\t\t\t_, pnet, _ = net.ParseCIDR(netaddr())\n\t\t} else {\n\t\t\tbreak\n\t\t}\n\t}\n\n\treturn netaddr(), nil\n}", "func (c *client) Delete(ctx context.Context, group, name string) error {\n\tvnet, err := c.Get(ctx, group, name)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif len(*vnet) == 0 {\n\t\treturn fmt.Errorf(\"Virtual Network [%s] not found\", name)\n\t}\n\n\trequest, err := getVirtualNetworkRequest(wssdcloudcommon.Operation_DELETE, group, name, &(*vnet)[0])\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = c.VirtualNetworkAgentClient.Invoke(ctx, request)\n\n\treturn err\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv6UnnumberedInterfaceRef(ctx context.Context, name string, index int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv6/frinx-openconfig-if-ip:unnumbered/frinx-openconfig-if-ip:interface-ref/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (c *Control) sendE2TSubscriptionDeleteRequest(subs *Subscription, trans *TransactionSubs, parentTrans *TransactionXapp) interface{} {\n\tvar err error\n\tvar event interface{}\n\tvar timedOut bool\n\n\tsubDelReqMsg := &e2ap.E2APSubscriptionDeleteRequest{}\n\tsubDelReqMsg.RequestId = subs.GetReqId().RequestId\n\tsubDelReqMsg.FunctionId = subs.SubReqMsg.FunctionId\n\ttrans.Mtype, trans.Payload, err = c.e2ap.PackSubscriptionDeleteRequest(subDelReqMsg)\n\tif err != nil {\n\t\txapp.Logger.Error(\"SUBS-SubDelReq: %s\", idstring(err, trans, subs, parentTrans))\n\t\treturn event\n\t}\n\n\tfor retries := uint64(0); retries < e2tMaxSubDelReqTryCount; retries++ {\n\t\tdesc := fmt.Sprintf(\"(retry %d)\", retries)\n\t\tc.msgSendToE2T(desc, subs, trans)\n\t\tevent, timedOut = trans.WaitEvent(e2tSubDelReqTime)\n\t\tif timedOut {\n\t\t\tcontinue\n\t\t}\n\t\tbreak\n\t}\n\txapp.Logger.Debug(\"SUBS-SubDelReq: Response handling event(%s) %s\", typeofSubsMessage(event), idstring(nil, trans, subs, parentTrans))\n\treturn event\n}", "func NetworkDelete(state *state.State, n network) error {\n\treturn deleteProfile(state, dnsmasqProfileFilename(n))\n}", "func checkDeleteNetwork(t *testing.T, expError bool, tenant, network string) {\n\terr := contivClient.NetworkDelete(tenant, network)\n\tif err != nil && !expError {\n\t\tt.Fatalf(\"Error deleting network %s/%s. Err: %v\", tenant, network, err)\n\t} else if err == nil && expError {\n\t\tt.Fatalf(\"Delete network %s/%s succeded while expecing error\", tenant, network)\n\t} else if err == nil {\n\t\t// verify network is gone\n\t\t_, err := contivClient.NetworkGet(tenant, network)\n\t\tif err == nil {\n\t\t\tt.Fatalf(\"Network %s/%s not deleted\", tenant, network)\n\t\t}\n\n\t\t// verify network state is gone too\n\t\tnetworkID := network + \".\" + tenant\n\t\tnwCfg := &mastercfg.CfgNetworkState{}\n\t\tnwCfg.StateDriver = stateStore\n\t\terr = nwCfg.Read(networkID)\n\t\tif err == nil {\n\t\t\tt.Fatalf(\"Network state %s not deleted\", networkID)\n\t\t}\n\t}\n}", "func (ec2 *EC2) DeleteNetworkInterface(id string) (resp *SimpleResp, err error) {\n\tparams := makeParams(\"DeleteNetworkInterface\")\n\tparams[\"NetworkInterfaceId\"] = id\n\tresp = &SimpleResp{}\n\terr = ec2.query(params, resp)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn resp, nil\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv6Neighbors(ctx context.Context, name string, index int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv6/frinx-openconfig-if-ip:neighbors/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (c *TestClient) CreateSubnetwork(project, region string, n *compute.Subnetwork) error {\n\tif c.CreateSubnetworkFn != nil {\n\t\treturn c.CreateSubnetworkFn(project, region, n)\n\t}\n\treturn c.client.CreateSubnetwork(project, region, n)\n}", "func (client WorkloadNetworksClient) DeleteSegment(ctx context.Context, resourceGroupName string, privateCloudName string, segmentID string) (result WorkloadNetworksDeleteSegmentFuture, err error) {\n\tif tracing.IsEnabled() {\n\t\tctx = tracing.StartSpan(ctx, fqdn+\"/WorkloadNetworksClient.DeleteSegment\")\n\t\tdefer func() {\n\t\t\tsc := -1\n\t\t\tif result.Response() != nil {\n\t\t\t\tsc = result.Response().StatusCode\n\t\t\t}\n\t\t\ttracing.EndSpan(ctx, sc, err)\n\t\t}()\n\t}\n\tif err := validation.Validate([]validation.Validation{\n\t\t{TargetValue: client.SubscriptionID,\n\t\t\tConstraints: []validation.Constraint{{Target: \"client.SubscriptionID\", Name: validation.MinLength, Rule: 1, Chain: nil}}},\n\t\t{TargetValue: resourceGroupName,\n\t\t\tConstraints: []validation.Constraint{{Target: \"resourceGroupName\", Name: validation.MaxLength, Rule: 90, Chain: nil},\n\t\t\t\t{Target: \"resourceGroupName\", Name: validation.MinLength, Rule: 1, Chain: nil},\n\t\t\t\t{Target: \"resourceGroupName\", Name: validation.Pattern, Rule: `^[-\\w\\._\\(\\)]+$`, Chain: nil}}}}); err != nil {\n\t\treturn result, validation.NewError(\"avs.WorkloadNetworksClient\", \"DeleteSegment\", err.Error())\n\t}\n\n\treq, err := client.DeleteSegmentPreparer(ctx, resourceGroupName, privateCloudName, segmentID)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"avs.WorkloadNetworksClient\", \"DeleteSegment\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresult, err = client.DeleteSegmentSender(req)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"avs.WorkloadNetworksClient\", \"DeleteSegment\", nil, \"Failure sending request\")\n\t\treturn\n\t}\n\n\treturn\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv6AddressesAddress(ctx context.Context, name string, index int64, ip string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv6/frinx-openconfig-if-ip:addresses/frinx-openconfig-if-ip:address/{ip}/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"ip\"+\"}\", fmt.Sprintf(\"%v\", ip), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv6AddressesAddressVrrp(ctx context.Context, name string, index int64, ip string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv6/frinx-openconfig-if-ip:addresses/frinx-openconfig-if-ip:address/{ip}/frinx-openconfig-if-ip:vrrp/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"ip\"+\"}\", fmt.Sprintf(\"%v\", ip), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (client *VirtualNetworkTapsClient) delete(ctx context.Context, resourceGroupName string, tapName string, options *VirtualNetworkTapsBeginDeleteOptions) (*azcore.Response, error) {\n\treq, err := client.deleteCreateRequest(ctx, resourceGroupName, tapName, options)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresp, err := client.con.Pipeline().Do(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif !resp.HasStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent) {\n\t\treturn nil, client.deleteHandleError(resp)\n\t}\n\treturn resp, nil\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv6NeighborsNeighbor(ctx context.Context, name string, index int64, ip string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv6/frinx-openconfig-if-ip:neighbors/frinx-openconfig-if-ip:neighbor/{ip}/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"ip\"+\"}\", fmt.Sprintf(\"%v\", ip), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv4UnnumberedConfig(ctx context.Context, name string, index int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv4/frinx-openconfig-if-ip:unnumbered/frinx-openconfig-if-ip:config/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func deleteSubscriber(id string) (retErr error) {\n\n\tvar err error = nil\n\n\t//Ottengo l'URL della coda\n\tqueueUrl, err := getQueueUrl(id)\n\tif err != nil {\n\t\tcommon.Warning(\"[BROKER] Errore nell'ottenere l'URL della coda. \" + err.Error())\n\t\t//return errors.New(\"error obtaining queue\")\n\t}\n\n\tcommon.Info(\"[BROKER] La coda da eliminare per il subscriber: \" + id + \" è: \" + queueUrl)\n\n\t//Eliminazione della coda\n\terr = deleteQueue(queueUrl)\n\tif err != nil {\n\t\tcommon.Warning(\"[BROKER] Errore nell'eliminazione della coda. \" + err.Error())\n\t} else {\n\t\tcommon.Info(\"[BROKER] Coda rimossa con successo\")\n\t}\n\n\t//Eliminazione della entry dal DB\n\terr = removeEntryDB(id)\n\tif err != nil {\n\t\tcommon.Warning(\"[BROKER] Errore nell'eliminazione della entry sul DB. \" + err.Error())\n\t\treturn errors.New(\"error in removing item in dynamodb\")\n\t} else {\n\t\tcommon.Info(\"[BROKER] Rimozione subscriber \" + id + \" effettuata con successo.\")\n\t}\n\n\treturn nil\n\n}", "func (c *TestClient) DeleteNetwork(project, name string) error {\n\tif c.DeleteNetworkFn != nil {\n\t\treturn c.DeleteNetworkFn(project, name)\n\t}\n\treturn c.client.DeleteNetwork(project, name)\n}", "func (plugin *Plugin) Del(args *cniSkel.CmdArgs) error {\n\t// Parse network configuration.\n\tnetConfig, err := config.New(args)\n\tif err != nil {\n\t\tlog.Errorf(\"Failed to parse netconfig from args: %v.\", err)\n\t\treturn err\n\t}\n\n\tlog.Infof(\"Executing DEL with netconfig: %+v ContainerID:%v Netns:%v IfName:%v Args:%v.\",\n\t\tnetConfig, args.ContainerID, args.Netns, args.IfName, args.Args)\n\n\tvar vpcENI *eni.ENI\n\t// If existing network is to be used then ENI is not required.\n\tif !netConfig.UseExistingNetwork {\n\t\t// Find the ENI.\n\t\tvpcENI, err = eni.NewENI(args.IfName, netConfig.ENIMACAddress)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Failed to find ENI %s: %v.\", netConfig.ENIName, err)\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Call operating system specific handler.\n\tnb := plugin.nb\n\n\tnw := network.Network{\n\t\tName: netConfig.Name,\n\t\tENI: vpcENI,\n\t\tUseExisting: netConfig.UseExistingNetwork,\n\t}\n\n\tep := network.Endpoint{\n\t\tContainerID: args.ContainerID,\n\t\tNetNSName: args.Netns,\n\t\tIPAddresses: netConfig.ENIIPAddresses,\n\t\tMACAddress: netConfig.ENIMACAddress,\n\t}\n\n\terr = nb.DeleteEndpoint(&nw, &ep)\n\tif err != nil {\n\t\t// DEL is best-effort. Log and ignore the failure.\n\t\tlog.Errorf(\"Failed to delete endpoint, ignoring: %v.\", err)\n\t}\n\n\t// Do not delete pre-existing networks.\n\tif !nw.UseExisting {\n\t\terr = nb.DeleteNetwork(&nw)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Failed to delete network: %v.\", err)\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv6AddressesAddressVrrpVrrpGroup(ctx context.Context, name string, index int64, ip string, virtualRouterId int32, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv6/frinx-openconfig-if-ip:addresses/frinx-openconfig-if-ip:address/{ip}/frinx-openconfig-if-ip:vrrp/frinx-openconfig-if-ip:vrrp-group/{virtual-router-id}/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"ip\"+\"}\", fmt.Sprintf(\"%v\", ip), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"virtual-router-id\"+\"}\", fmt.Sprintf(\"%v\", virtualRouterId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func NetworkResourceDelete(w http.ResponseWriter, r *http.Request) {\n\tnrID := mux.Vars(r)[\"nr_id\"]\n\tcp := r.URL.Query().Get(\"cloud_provider\")\n\tif nrID == \"\" || cp == \"\" {\n\t\thandleError(w, http.StatusBadRequest,\n\t\t\t\"url params 'nr_id' or 'cloud_provider' is missing\")\n\t\treturn\n\t}\n\n\tif err := ctr.DeleteNetworkResource(nrID); err != nil {\n\t\thandleError(w, http.StatusInternalServerError, err.Error())\n\t\treturn\n\t}\n\n\thandleResult(w, http.StatusOK, emptyBody)\n\treturn\n}", "func RunNamespacesDelete(c *CmdConfig) error {\n\terr := ensureOneArg(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\targ := c.Args[0]\n\tss := c.Serverless()\n\tctx := context.TODO()\n\t// Since arg may be either a label or an id, match against existing namespaces\n\tvar (\n\t\tid string\n\t\tlabel string\n\t)\n\tmatches, err := getMatchingNamespaces(ctx, ss, arg)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif len(matches) > 0 {\n\t\tid = matches[0].Namespace\n\t\tlabel = matches[0].Label\n\t}\n\t// Must be an exact match though (avoids errors).\n\tif len(matches) != 1 || (arg != label && arg != id) {\n\t\treturn fmt.Errorf(\"'%s' does not exactly match the label or id of any of your namespaces\", arg)\n\t}\n\tforce, _ := c.Doit.GetBool(c.NS, \"force\")\n\tif !force {\n\t\tfmt.Fprintf(c.Out, \"Deleting namespace '%s' with label '%s'.\\n\", id, label)\n\t\tif AskForConfirmDelete(\"namespace\", 1) != nil {\n\t\t\treturn fmt.Errorf(\"deletion of '%s' not confirmed, doing nothing\", id)\n\t\t}\n\t}\n\terr = ss.DeleteNamespace(ctx, id)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfmt.Fprintln(c.Out, \"Namespace successfully deleted\")\n\treturn nil\n}", "func (c *MockNetworkSecurityGroupsClient) Delete(ctx context.Context, resourceGroupName, nsgName string) error {\n\t// Ignore resourceGroupName for simplicity.\n\tif _, ok := c.NSGs[nsgName]; !ok {\n\t\treturn fmt.Errorf(\"%s does not exist\", nsgName)\n\t}\n\tdelete(c.NSGs, nsgName)\n\treturn nil\n}", "func (sdk *MockGoSDKClient) DeleteFailoverGroup(ctx context.Context, resourceGroupName string, serverName string, failoverGroupName string) (result autorest.Response, err error) {\n\n\treturn helpers.GetRestResponse(http.StatusOK), nil\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv4UnnumberedInterfaceRef(ctx context.Context, name string, index int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv4/frinx-openconfig-if-ip:unnumbered/frinx-openconfig-if-ip:interface-ref/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv4Neighbors(ctx context.Context, name string, index int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv4/frinx-openconfig-if-ip:neighbors/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv6NeighborsNeighborConfig(ctx context.Context, name string, index int64, ip string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv6/frinx-openconfig-if-ip:neighbors/frinx-openconfig-if-ip:neighbor/{ip}/frinx-openconfig-if-ip:config/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"ip\"+\"}\", fmt.Sprintf(\"%v\", ip), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (client *DedicatedHostsClient) deleteOperation(ctx context.Context, resourceGroupName string, hostGroupName string, hostName string, options *DedicatedHostsBeginDeleteOptions) (*http.Response, error) {\n\treq, err := client.deleteCreateRequest(ctx, resourceGroupName, hostGroupName, hostName, options)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresp, err := client.pl.Do(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif !runtime.HasStatusCode(resp, http.StatusOK, http.StatusAccepted, http.StatusNoContent) {\n\t\treturn nil, client.deleteHandleError(resp)\n\t}\n\treturn resp, nil\n}", "func Delete(strinput2 string, tc net.Conn) {\n\tnetwork.SendDataMessage(&tc, 4, 6, 1, strinput2)\n}", "func (d *DestinationClient) DeleteSubaccountCertificate(name string) (AffectedRecords, error) {\n\n\tvar retval AffectedRecords\n\tvar errResponse ErrorMessage\n\n\tresponse, err := d.restyClient.R().\n\t\tSetResult(&retval).\n\t\tSetError(&errResponse).\n\t\tSetPathParams(map[string]string{\n\t\t\t\"name\": name,\n\t\t}).\n\t\tDelete(\"/subaccountCertificate/{name}\")\n\n\tif err != nil {\n\t\treturn retval, err\n\t}\n\tif response.StatusCode() != 200 {\n\t\terrResponse.statusCode = response.StatusCode()\n\t\treturn retval, errResponse\n\t}\n\treturn retval, nil\n}", "func (n *cni) remove(pod Pod, networkNS NetworkNamespace) error {\n\tif err := removeNetworkCommon(networkNS); err != nil {\n\t\treturn err\n\t}\n\n\tif err := n.deleteVirtInterfaces(networkNS); err != nil {\n\t\treturn err\n\t}\n\n\treturn deleteNetNS(networkNS.NetNsPath, true)\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv6UnnumberedInterfaceRefConfig(ctx context.Context, name string, index int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv6/frinx-openconfig-if-ip:unnumbered/frinx-openconfig-if-ip:interface-ref/frinx-openconfig-if-ip:config/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv4NeighborsNeighbor(ctx context.Context, name string, index int64, ip string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv4/frinx-openconfig-if-ip:neighbors/frinx-openconfig-if-ip:neighbor/{ip}/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"ip\"+\"}\", fmt.Sprintf(\"%v\", ip), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (client WorkloadNetworksClient) DeleteSegmentPreparer(ctx context.Context, resourceGroupName string, privateCloudName string, segmentID string) (*http.Request, error) {\n\tpathParameters := map[string]interface{}{\n\t\t\"privateCloudName\": autorest.Encode(\"path\", privateCloudName),\n\t\t\"resourceGroupName\": autorest.Encode(\"path\", resourceGroupName),\n\t\t\"segmentId\": autorest.Encode(\"path\", segmentID),\n\t\t\"subscriptionId\": autorest.Encode(\"path\", client.SubscriptionID),\n\t}\n\n\tconst APIVersion = \"2020-07-17-preview\"\n\tqueryParameters := map[string]interface{}{\n\t\t\"api-version\": APIVersion,\n\t}\n\n\tpreparer := autorest.CreatePreparer(\n\t\tautorest.AsDelete(),\n\t\tautorest.WithBaseURL(client.BaseURI),\n\t\tautorest.WithPathParameters(\"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AVS/privateClouds/{privateCloudName}/workloadNetworks/default/segments/{segmentId}\", pathParameters),\n\t\tautorest.WithQueryParameters(queryParameters))\n\treturn preparer.Prepare((&http.Request{}).WithContext(ctx))\n}", "func (s *Service) Delete(ctx context.Context) error {\n\tfor _, subnetSpec := range s.Scope.SubnetSpecs() {\n\t\tif !s.Scope.Vnet().IsManaged(s.Scope.ClusterName()) {\n\t\t\ts.Scope.V(4).Info(\"Skipping subnets deletion in custom vnet mode\")\n\t\t\tcontinue\n\t\t}\n\t\ts.Scope.V(2).Info(\"deleting subnet in vnet\", \"subnet\", subnetSpec.Name, \"vnet\", subnetSpec.VNetName)\n\t\terr := s.Client.Delete(ctx, s.Scope.Vnet().ResourceGroup, subnetSpec.VNetName, subnetSpec.Name)\n\t\tif err != nil && azure.ResourceNotFound(err) {\n\t\t\t// already deleted\n\t\t\tcontinue\n\t\t}\n\t\tif err != nil {\n\t\t\treturn errors.Wrapf(err, \"failed to delete subnet %s in resource group %s\", subnetSpec.Name, s.Scope.Vnet().ResourceGroup)\n\t\t}\n\n\t\ts.Scope.V(2).Info(\"successfully deleted subnet in vnet\", \"subnet\", subnetSpec.Name, \"vnet\", subnetSpec.VNetName)\n\t}\n\treturn nil\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv6AddressesAddressVrrpVrrpGroupConfig(ctx context.Context, name string, index int64, ip string, virtualRouterId int32, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv6/frinx-openconfig-if-ip:addresses/frinx-openconfig-if-ip:address/{ip}/frinx-openconfig-if-ip:vrrp/frinx-openconfig-if-ip:vrrp-group/{virtual-router-id}/frinx-openconfig-if-ip:config/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"ip\"+\"}\", fmt.Sprintf(\"%v\", ip), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"virtual-router-id\"+\"}\", fmt.Sprintf(\"%v\", virtualRouterId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func deleteHNSEndpoint(ep *hcsshim.HNSEndpoint, containerID string) (string, error) {\n\tdefer func() { _ = recover() }()\n\tif ep == nil {\n\t\treturn \"\", nil\n\t}\n\n\t// detach container from endpoint\n\tif containerID != \"\" {\n\t\t_ = hcsshim.HotDetachEndpoint(containerID, ep.Id)\n\t}\n\n\t// delete endpoint\n\t_ = ep.HostDetach()\n\tif _, err := ep.Delete(); err != nil {\n\t\tif !hcsshim.IsNotExist(err) {\n\t\t\treturn ep.VirtualNetwork, err\n\t\t}\n\t}\n\treturn ep.VirtualNetwork, nil\n}", "func (_Flytrap *FlytrapTransactor) RevokeSub(opts *bind.TransactOpts, person common.Address, topic string, reason string) (*types.Transaction, error) {\n\treturn _Flytrap.contract.Transact(opts, \"revokeSub\", person, topic, reason)\n}", "func (c *Client) DeleteNetwork(id string) error {\n\tendpoint := fmt.Sprintf(\"%snetworks/%s\", baseAddr, id)\n\tr, err := http.NewRequest(\"DELETE\", endpoint, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\tresp, err := c.http.Do(r)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn statusCode(resp.StatusCode, http.StatusNoContent)\n}", "func (c *OvnClusterController) deleteNamespace(namespaceName string, namespaceSubnet *net.IPNet) error {\n\n\tif namespaceSubnet != nil {\n\t\tif err := c.deleteNamespaceSubnet(namespaceName, namespaceSubnet); err != nil {\n\t\t\tlogrus.Errorf(\"Error deleting namespace %s HostSubnet: %v\", namespaceName, err)\n\t\t}\n\t}\n\n\tif err := c.deleteNamespaceLogicalNetwork(namespaceName); err != nil {\n\t\tlogrus.Errorf(\"Error deleting namespace %s logical network: %v\", namespaceName, err)\n\t}\n\n\treturn nil\n\n}", "func (client ModelClient) DeleteSubListSender(req *http.Request) (*http.Response, error) {\n\treturn autorest.SendWithSender(client, req,\n\t\tautorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n}", "func (client MeshNetworkClient) Delete(ctx context.Context, networkResourceName string) (result autorest.Response, err error) {\n\tif tracing.IsEnabled() {\n\t\tctx = tracing.StartSpan(ctx, fqdn+\"/MeshNetworkClient.Delete\")\n\t\tdefer func() {\n\t\t\tsc := -1\n\t\t\tif result.Response != nil {\n\t\t\t\tsc = result.Response.StatusCode\n\t\t\t}\n\t\t\ttracing.EndSpan(ctx, sc, err)\n\t\t}()\n\t}\n\treq, err := client.DeletePreparer(ctx, networkResourceName)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"servicefabric.MeshNetworkClient\", \"Delete\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresp, err := client.DeleteSender(req)\n\tif err != nil {\n\t\tresult.Response = resp\n\t\terr = autorest.NewErrorWithError(err, \"servicefabric.MeshNetworkClient\", \"Delete\", resp, \"Failure sending request\")\n\t\treturn\n\t}\n\n\tresult, err = client.DeleteResponder(resp)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"servicefabric.MeshNetworkClient\", \"Delete\", resp, \"Failure responding to request\")\n\t}\n\n\treturn\n}", "func (c *Client) SubRemove(id string) error {\n\tres, err := c.Exec(\"sub-remove\", id)\n\tif res == nil {\n\t\treturn err\n\t}\n\treturn err\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv6AddressesAddressConfig(ctx context.Context, name string, index int64, ip string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv6/frinx-openconfig-if-ip:addresses/frinx-openconfig-if-ip:address/{ip}/frinx-openconfig-if-ip:config/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"ip\"+\"}\", fmt.Sprintf(\"%v\", ip), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func deleteHCNEndpoint(ep *hcn.HostComputeEndpoint, netns string) (string, error) {\n\tdefer func() { _ = recover() }()\n\tif ep == nil {\n\t\treturn \"\", nil\n\t}\n\n\t// remove endpoint from namespace\n\tif netns == \"\" {\n\t\tnetns = ep.HostComputeNamespace\n\t}\n\t_ = hcn.RemoveNamespaceEndpoint(netns, ep.Id)\n\n\t// delete endpoint\n\tif err := ep.Delete(); err != nil {\n\t\tif !hcn.IsNotFoundError(err) {\n\t\t\treturn ep.HostComputeNetwork, err\n\t\t}\n\t}\n\treturn ep.HostComputeNetwork, nil\n}", "func (c *TestClient) ListSubnetworks(project, region string, opts ...ListCallOption) ([]*compute.Subnetwork, error) {\n\tif c.ListSubnetworksFn != nil {\n\t\treturn c.ListSubnetworksFn(project, region, opts...)\n\t}\n\treturn c.client.ListSubnetworks(project, region, opts...)\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstanceMplsSignalingProtocolsRsvpTeInterfaceAttributesInterfaceSubscription(ctx context.Context, name string, interfaceId string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:mpls/frinx-openconfig-network-instance:signaling-protocols/frinx-openconfig-network-instance:rsvp-te/frinx-openconfig-network-instance:interface-attributes/frinx-openconfig-network-instance:interface/{interface-id}/frinx-openconfig-network-instance:subscription/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"interface-id\"+\"}\", fmt.Sprintf(\"%v\", interfaceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (plugin *Plugin) Del(args *cniSkel.CmdArgs) error {\n\t// Parse network configuration.\n\tnetConfig, err := config.New(args, false)\n\tif err != nil {\n\t\tlog.Errorf(\"Failed to parse netconfig from args: %v.\", err)\n\t\treturn err\n\t}\n\n\tlog.Infof(\"Executing DEL with netconfig: %+v.\", netConfig)\n\n\t// Derive names from CNI network config.\n\tpatNetNSName := fmt.Sprintf(patNetNSNameFormat, netConfig.BranchVlanID)\n\ttapBridgeName := fmt.Sprintf(tapBridgeNameFormat, netConfig.BranchVlanID)\n\ttapLinkName := args.IfName\n\ttargetNetNSName := args.Netns\n\n\t// Delete the tap link and veth pair from the target netns.\n\tplugin.deleteTapVethLinks(targetNetNSName, tapLinkName, tapBridgeName)\n\n\t// Search for the PAT network namespace.\n\tpatNetNS, err := netns.GetNetNSByName(patNetNSName)\n\tif err != nil {\n\t\t// Log and ignore the failure. DEL can be called multiple times and thus must be idempotent.\n\t\tlog.Errorf(\"Failed to find netns %s, ignoring: %v.\", patNetNSName, err)\n\t\treturn nil\n\t}\n\tlastVethLinkDeleted := false\n\n\t// In PAT network namespace...\n\terr = patNetNS.Run(func() error {\n\t\t// Check whether there are any remaining veth links connected to this bridge.\n\t\tifaces, _ := net.Interfaces()\n\t\tlog.Infof(\"Number of remaining links: %v.\", len(ifaces))\n\t\tif len(ifaces) == 4 {\n\t\t\t// Only VLAN link, bridge, dummy and loopback remain.\n\t\t\tlastVethLinkDeleted = true\n\t\t}\n\n\t\treturn nil\n\t})\n\n\t// If all veth links connected to this PAT bridge are deleted, clean up the PAT network\n\t// namespace and all virtual interfaces in it. Otherwise, leave it running.\n\tif lastVethLinkDeleted && netConfig.CleanupPATNetNS {\n\t\tlog.Infof(\"Deleting PAT network namespace: %v.\", patNetNSName)\n\t\terr = patNetNS.Close()\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Failed to delete netns: %v.\", err)\n\t\t}\n\t} else {\n\t\tlog.Infof(\"Skipping PAT netns deletion. Last veth link deleted: %t, cleanup PAT netns: %t.\",\n\t\t\tlastVethLinkDeleted, netConfig.CleanupPATNetNS)\n\t}\n\n\treturn nil\n}", "func (g *smartContractGW) deleteStreamOrSub(res http.ResponseWriter, req *http.Request, params httprouter.Params) {\n\tlog.Infof(\"--> %s %s\", req.Method, req.URL)\n\n\tif g.sm == nil {\n\t\tg.gatewayErrReply(res, req, errors.New(errEventSupportMissing), 405)\n\t\treturn\n\t}\n\n\tvar err error\n\tif strings.HasPrefix(req.URL.Path, events.SubPathPrefix) {\n\t\terr = g.sm.DeleteSubscription(req.Context(), params.ByName(\"id\"))\n\t} else {\n\t\terr = g.sm.DeleteStream(req.Context(), params.ByName(\"id\"))\n\t}\n\tif err != nil {\n\t\tg.gatewayErrReply(res, req, err, 500)\n\t\treturn\n\t}\n\n\tstatus := 204\n\tlog.Infof(\"<-- %s %s [%d]\", req.Method, req.URL, status)\n\tres.Header().Set(\"Content-Type\", \"application/json\")\n\tres.WriteHeader(status)\n}", "func ExampleDelegatedSubnetServiceClient_BeginDeleteDetails() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclientFactory, err := armdelegatednetwork.NewClientFactory(\"<subscription-id>\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tpoller, err := clientFactory.NewDelegatedSubnetServiceClient().BeginDeleteDetails(ctx, \"TestRG\", \"delegated1\", &armdelegatednetwork.DelegatedSubnetServiceClientBeginDeleteDetailsOptions{ForceDelete: nil})\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\t_, err = poller.PollUntilDone(ctx, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to pull the result: %v\", err)\n\t}\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv4Addresses(ctx context.Context, name string, index int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv4/frinx-openconfig-if-ip:addresses/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func btrfsSubVolumesDelete(subvol string) error {\n\t// Delete subsubvols.\n\tsubsubvols, err := btrfsSubVolumesGet(subvol)\n\tif err != nil {\n\t\treturn err\n\t}\n\tsort.Sort(sort.Reverse(sort.StringSlice(subsubvols)))\n\n\tfor _, subsubvol := range subsubvols {\n\t\terr := btrfsSubVolumeDelete(path.Join(subvol, subsubvol))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Delete the subvol itself\n\terr = btrfsSubVolumeDelete(subvol)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (client *NetworkToNetworkInterconnectsClient) deleteCreateRequest(ctx context.Context, resourceGroupName string, networkFabricName string, networkToNetworkInterconnectName string, options *NetworkToNetworkInterconnectsClientBeginDeleteOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.ManagedNetworkFabric/networkFabrics/{networkFabricName}/networkToNetworkInterconnects/{networkToNetworkInterconnectName}\"\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif networkFabricName == \"\" {\n\t\treturn nil, errors.New(\"parameter networkFabricName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{networkFabricName}\", url.PathEscape(networkFabricName))\n\tif networkToNetworkInterconnectName == \"\" {\n\t\treturn nil, errors.New(\"parameter networkToNetworkInterconnectName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{networkToNetworkInterconnectName}\", url.PathEscape(networkToNetworkInterconnectName))\n\treq, err := runtime.NewRequest(ctx, http.MethodDelete, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2023-06-15\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func RunKubeDelete(log *LoggingConfig, args []string, dryrun bool) (string, error) {\n\tlog.Info.log(\"Attempting to delete resource from Kubernetes ...\")\n\tkargs := []string{\"delete\"}\n\tkargs = append(kargs, args...)\n\treturn RunKube(log, kargs, dryrun)\n}", "func TestNetworkCreateDelete(t *testing.T) {\n\t// create netagent\n\tag, _, _ := createNetAgent(t)\n\tAssert(t, ag != nil, \"Failed to create agent %#v\", ag)\n\tdefer ag.Stop()\n\n\t// network message\n\tnt := netproto.Network{\n\t\tTypeMeta: api.TypeMeta{Kind: \"Network\"},\n\t\tObjectMeta: api.ObjectMeta{\n\t\t\tTenant: \"default\",\n\t\t\tName: \"default\",\n\t\t\tNamespace: \"default\",\n\t\t},\n\t\tSpec: netproto.NetworkSpec{\n\t\t\tVlanID: 42,\n\t\t},\n\t}\n\n\t// make create network call\n\terr := ag.CreateNetwork(&nt)\n\tAssertOk(t, err, \"Error creating network\")\n\ttnt, err := ag.FindNetwork(nt.ObjectMeta)\n\tAssertOk(t, err, \"Network was not found in DB\")\n\tAssert(t, tnt.Spec.VlanID == 42, \"Network VLAN didn't match\", tnt)\n\n\t// verify duplicate network creations succeed\n\terr = ag.CreateNetwork(&nt)\n\tAssertOk(t, err, \"Error creating duplicate network\")\n\n\t// verify duplicate network name with different content does not succeed\n\tnnt := netproto.Network{\n\t\tTypeMeta: api.TypeMeta{Kind: \"Network\"},\n\t\tObjectMeta: api.ObjectMeta{\n\t\t\tTenant: \"default\",\n\t\t\tName: \"default\",\n\t\t\tNamespace: \"default\",\n\t\t},\n\t\tSpec: netproto.NetworkSpec{\n\t\t\tVlanID: 84,\n\t\t},\n\t}\n\terr = ag.CreateNetwork(&nnt)\n\tAssert(t, (err != nil), \"conflicting network creation succeeded\")\n\n\t// verify list api works\n\tnetList := ag.ListNetwork()\n\tAssert(t, len(netList) == 2, \"Incorrect number of networks\")\n\n\t// delete the network and verify its gone from db\n\terr = ag.DeleteNetwork(nt.Tenant, nt.Namespace, nt.Name)\n\tAssertOk(t, err, \"Error deleting network\")\n\t_, err = ag.FindNetwork(nt.ObjectMeta)\n\tAssert(t, err != nil, \"Network was still found in database after deleting\", ag)\n\n\t// verify you can not delete non-existing network\n\terr = ag.DeleteNetwork(nt.Tenant, nt.Namespace, nt.Name)\n\tAssert(t, err != nil, \"deleting non-existing network succeeded\", ag)\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv4NeighborsNeighborConfig(ctx context.Context, name string, index int64, ip string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv4/frinx-openconfig-if-ip:neighbors/frinx-openconfig-if-ip:neighbor/{ip}/frinx-openconfig-if-ip:config/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"ip\"+\"}\", fmt.Sprintf(\"%v\", ip), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (s *SubnetworkServer) applySubnetwork(ctx context.Context, c *compute.Client, request *computepb.ApplyComputeSubnetworkRequest) (*computepb.ComputeSubnetwork, error) {\n\tp := ProtoToSubnetwork(request.GetResource())\n\tres, err := c.ApplySubnetwork(ctx, p)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tr := SubnetworkToProto(res)\n\treturn r, nil\n}", "func (s *GRPCServerHandler) DelNetwork(context context.Context, in *rpc.DelNetworkRequest) (*rpc.DelNetworkReply, error) {\n\tklog.V(1).Infof(\"Received DelNetwork for IP %s, Pod %s, Namespace %s, Container %s\",\n\t\tin.IPv4Addr, in.K8S_POD_NAME, in.K8S_POD_NAMESPACE, in.K8S_POD_INFRA_CONTAINER_ID)\n\n\tip, deviceNumber, err := s.ipamd.dataStore.UnassignPodIPv4Address(&k8sapi.K8SPodInfo{\n\t\tName: in.K8S_POD_NAME,\n\t\tNamespace: in.K8S_POD_NAMESPACE,\n\t\tContainer: in.K8S_POD_INFRA_CONTAINER_ID})\n\n\tif err != nil && err == datastore.ErrUnknownPod {\n\t\t// If L-IPAMD restarts, the pod's IP address are assigned by only pod's name and namespace due to kubelet's introspection.\n\t\tip, deviceNumber, err = s.ipamd.dataStore.UnassignPodIPv4Address(&k8sapi.K8SPodInfo{\n\t\t\tName: in.K8S_POD_NAME,\n\t\t\tNamespace: in.K8S_POD_NAMESPACE})\n\t\tif err == datastore.ErrUnknownPod {\n\t\t\tklog.Warningf(\"Detect unhealthy pod %s/%s\", in.K8S_POD_NAME, in.K8S_POD_NAMESPACE)\n\t\t\treturn &rpc.DelNetworkReply{Success: true}, nil\n\t\t}\n\t}\n\tklog.V(1).Infof(\"Send DelNetworkReply: IPv4Addr %s, DeviceNumber: %d, err: %v\", ip, deviceNumber, err)\n\tresp := &rpc.DelNetworkReply{Success: err == nil, IPv4Addr: ip, DeviceNumber: int32(deviceNumber)}\n\tif err != nil {\n\t\tresp.Message = err.Error()\n\t}\n\ts.ipamd.trigCh <- udevNotify{}\n\treturn resp, nil\n}", "func (a *FrinxOpenconfigIfIpApiService) DeleteFrinxOpenconfigInterfacesInterfacesInterfaceSubinterfacesSubinterfaceIpv4UnnumberedInterfaceRefConfig(ctx context.Context, name string, index int64, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-interfaces:interfaces/frinx-openconfig-interfaces:interface/{name}/frinx-openconfig-interfaces:subinterfaces/frinx-openconfig-interfaces:subinterface/{index}/frinx-openconfig-if-ip:ipv4/frinx-openconfig-if-ip:unnumbered/frinx-openconfig-if-ip:interface-ref/frinx-openconfig-if-ip:config/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"index\"+\"}\", fmt.Sprintf(\"%v\", index), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstanceMplsSignalingProtocolsRsvpTeInterfaceAttributesInterfaceSubscriptionConfig(ctx context.Context, name string, interfaceId string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:mpls/frinx-openconfig-network-instance:signaling-protocols/frinx-openconfig-network-instance:rsvp-te/frinx-openconfig-network-instance:interface-attributes/frinx-openconfig-network-instance:interface/{interface-id}/frinx-openconfig-network-instance:subscription/frinx-openconfig-network-instance:config/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"interface-id\"+\"}\", fmt.Sprintf(\"%v\", interfaceId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}" ]
[ "0.6693976", "0.6683054", "0.6683054", "0.6585151", "0.6103286", "0.5706947", "0.56175166", "0.56010973", "0.544689", "0.54297084", "0.5403982", "0.53904927", "0.53755575", "0.5363284", "0.53612095", "0.53451115", "0.53301173", "0.532158", "0.5317167", "0.5313285", "0.52845067", "0.52675754", "0.5261594", "0.52585584", "0.5252118", "0.52452356", "0.5212113", "0.5167041", "0.5162542", "0.51368904", "0.513028", "0.51269865", "0.5122496", "0.5109602", "0.51065814", "0.51000935", "0.5097318", "0.50935906", "0.5074776", "0.50603884", "0.5059343", "0.5058613", "0.505808", "0.50411725", "0.50377667", "0.50353307", "0.50315785", "0.5022292", "0.50206256", "0.5013515", "0.5003796", "0.500335", "0.49941945", "0.49789718", "0.4978258", "0.49717188", "0.49680892", "0.49602315", "0.49534646", "0.49461526", "0.49433696", "0.49406558", "0.49399704", "0.4936303", "0.49317718", "0.49064493", "0.490424", "0.49031475", "0.48993683", "0.4888103", "0.48844692", "0.48819247", "0.48756567", "0.48755065", "0.4875007", "0.48738915", "0.48512036", "0.48508808", "0.4848155", "0.48466286", "0.48378727", "0.4837237", "0.48337615", "0.48328134", "0.48283556", "0.48194668", "0.4805982", "0.48059177", "0.48053828", "0.48033765", "0.48028547", "0.47934824", "0.4790301", "0.47882935", "0.47756314", "0.477133", "0.47678682", "0.4765017", "0.47642535", "0.4761247" ]
0.740523
0
DeleteTargetInstance uses the override method DeleteTargetInstanceFn or the real implementation.
func (c *TestClient) DeleteTargetInstance(project, zone, name string) error { if c.DeleteTargetInstanceFn != nil { return c.DeleteTargetInstanceFn(project, zone, name) } return c.client.DeleteTargetInstance(project, zone, name) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (p *OnPrem) DeleteInstance(ctx *Context, instancename string) error {\n\treturn fmt.Errorf(\"Operation not supported\")\n}", "func (p *ProxMox) DeleteInstance(ctx *lepton.Context, instanceID string) error {\n\n\treq, err := http.NewRequest(\"DELETE\", p.apiURL+\"/api2/json/nodes/\"+p.nodeNAME+\"/qemu/\"+instanceID, nil)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\ttr := &http.Transport{\n\t\tTLSClientConfig: &tls.Config{InsecureSkipVerify: true},\n\t}\n\tclient := &http.Client{Transport: tr}\n\n\treq.Header.Add(\"Authorization\", \"PVEAPIToken=\"+p.tokenID+\"=\"+p.secret)\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\t_, err = io.ReadAll(resp.Body)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\n\t}\n\n\treturn err\n\n}", "func (p *OnPrem) DeleteInstance(ctx *Context, instancename string) error {\n\n\tpid, err := strconv.Atoi(instancename)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\n\t// yolo\n\terr = sysKill(pid)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\n\topshome := GetOpsHome()\n\tipath := path.Join(opshome, \"instances\", instancename)\n\terr = os.Remove(ipath)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func TestDeleteInstance(t *testing.T) {\n\tvar reason payloads.StartFailureReason\n\n\tclient, instances := testStartWorkload(t, 1, false, reason)\n\tdefer client.Shutdown()\n\n\tsendStatsCmd(client, t)\n\n\tserverCh := server.AddCmdChan(ssntp.DELETE)\n\n\terr := ctl.deleteInstance(instances[0].ID)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tresult, err := server.GetCmdChanResult(serverCh, ssntp.DELETE)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\tif result.InstanceUUID != instances[0].ID {\n\t\tt.Fatal(\"Did not get correct Instance ID\")\n\t}\n}", "func (d *Dao) DeleteTarget(c context.Context, id int64) (res int64, err error) {\n\tvar (\n\t\tr sql.Result\n\t\tnow = time.Now()\n\t)\n\tif r, err = d.db.Exec(c, _deleteTargetSQL, now, id); err != nil {\n\t\tlog.Error(\"d.UpdateGroup.Exec error(%+v), target(%d)\", err, id)\n\t\treturn\n\t}\n\tif res, err = r.RowsAffected(); err != nil {\n\t\tlog.Error(\"d.UpdateGroup.RowsAffected error(%+v), target(%d)\", err, id)\n\t}\n\treturn\n}", "func (c *InstanceManagerClient) InstanceDelete(backendStoreDriver longhorn.BackendStoreDriverType, name, kind, diskUUID string, cleanupRequired bool) (err error) {\n\tif c.GetAPIVersion() < 4 {\n\t\t/* Fall back to the old way of deleting process */\n\t\t_, err = c.processManagerGrpcClient.ProcessDelete(name)\n\t} else {\n\t\t_, err = c.instanceServiceGrpcClient.InstanceDelete(string(backendStoreDriver), name, kind, diskUUID, cleanupRequired)\n\t}\n\n\treturn err\n}", "func (m *Client) DeleteTarget(arg0 context.Context, arg1 int64) error {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"DeleteTarget\", arg0, arg1)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func deleteInstance(params martini.Params, r render.Render) {\n\tinstanceName := params[\"name\"]\n\n\tif !instanceExists(instanceName) {\n\t\tfmt.Println(\"Instance with specified name does not exist in provision table\")\n\t\tr.Text(400, \"Bad Request\")\n\t\treturn\n\t}\n\n\tsvc := neptune.New(session.New(&aws.Config{\n\t\tRegion: aws.String(os.Getenv(\"REGION\")),\n\t}))\n\n\tinstanceParamsDelete := &neptune.DeleteDBInstanceInput{\n\t\tDBInstanceIdentifier: aws.String(instanceName),\n\t\tSkipFinalSnapshot: aws.Bool(true),\n\t}\n\n\tclusterParamsDelete := &neptune.DeleteDBClusterInput{\n\t\tDBClusterIdentifier: aws.String(instanceName),\n\t\tSkipFinalSnapshot: aws.Bool(true),\n\t}\n\n\tinstanceResp, instanceErr := svc.DeleteDBInstance(instanceParamsDelete)\n\tname := *instanceParamsDelete.DBInstanceIdentifier\n\tif instanceErr != nil {\n\t\tfmt.Println(instanceErr.Error())\n\t\toutput500Error(r, instanceErr)\n\t\treturn\n\t}\n\tfmt.Println(\"Deletion in progress for instance \" + *instanceResp.DBInstance.DBInstanceIdentifier)\n\n\tclusterResp, clusterErr := svc.DeleteDBCluster(clusterParamsDelete)\n\tif clusterErr != nil {\n\t\tfmt.Println(instanceErr.Error())\n\t\toutput500Error(r, clusterErr)\n\t\treturn\n\t}\n\tfmt.Println(\"Deletion in progress for cluster \" + *clusterResp.DBCluster.DBClusterIdentifier)\n\n\t_, err := pool.Exec(\"DELETE FROM provision WHERE name=$1\", name)\n\tif err != nil {\n\t\tfmt.Println(err.Error())\n\t\toutput500Error(r, err)\n\t\treturn\n\t}\n\n\tr.JSON(200, map[string]string{\"Response\": \"Instance deletion in progress\"})\n\n\tusername := *instanceParamsDelete.DBInstanceIdentifier\n\tdeleteUserPolicy(username)\n\tdeleteAccessKey(username)\n\tdeleteUser(username)\n\n}", "func (c *TestClient) DeleteInstance(project, zone, name string) error {\n\tif c.DeleteInstanceFn != nil {\n\t\treturn c.DeleteInstanceFn(project, zone, name)\n\t}\n\treturn c.client.DeleteInstance(project, zone, name)\n}", "func (c *MockAzureCloud) DeleteInstance(i *cloudinstances.CloudInstance) error {\n\treturn errors.New(\"DeleteInstance not implemented on azureCloud\")\n}", "func (i *InstanceServiceHandler) Delete(ctx context.Context, instanceID string) error {\n\turi := fmt.Sprintf(\"%s/%s\", instancePath, instanceID)\n\n\treq, err := i.client.NewRequest(ctx, http.MethodDelete, uri, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn i.client.DoWithContext(ctx, req, nil)\n}", "func (awsI *Ec2Instance) DeleteInstance() error {\n\tinstanceWorkingDir := path.Join(awsI.DeployDir, \"etc\", \"terraform\", \"instance\")\n\tinstanceConfPath := path.Join(instanceWorkingDir, \"instance.json\")\n\tif !sdutils.PathExists(instanceConfPath) {\n\t\treturn errors.New(\"There is no configured instance\")\n\t}\n\tterraformPath, err := GetTerraformPath(awsI.Ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\tcmdArray := []string{terraformPath, \"destroy\", \"-force\", \"-var-file\", instanceConfPath}\n\tcmd := exec.Cmd{\n\t\tPath: cmdArray[0],\n\t\tArgs: cmdArray,\n\t\tDir: instanceWorkingDir,\n\t}\n\tawsI.Ctx.Logf(sdutils.INFO, \"Running terraform...\\n\")\n\tspin := sdutils.NewSpinner(awsI.Ctx, 1, \"Deleting the instance VMs\")\n\t_, err = sdutils.RunCommand(awsI.Ctx, cmd, volumeLineScanner, spin)\n\tif err != nil {\n\t\treturn err\n\t}\n\tos.Remove(instanceConfPath)\n\tawsI.Ctx.ConsoleLog(1, \"Successfully destroyed the instance.\\n\")\n\treturn nil\n}", "func (s *AutoscalerSuite) TestInstanceTerminate(c *check.C) {\n\tclusterName := \"bob\"\n\tinstance := &gaws.Instance{\n\t\tID: \"instance-1\",\n\t}\n\tec := newMockEC2(&ec2.Instance{\n\t\tInstanceId: aws.String(\"instance-1\"),\n\t})\n\tqueue := newMockQueue(\"queue-1\")\n\ta, err := New(Config{\n\t\tClusterName: clusterName,\n\t\tNewLocalInstance: func() (*gaws.Instance, error) {\n\t\t\treturn instance, nil\n\t\t},\n\t\tQueue: queue,\n\t\tCloud: ec,\n\t})\n\tc.Assert(err, check.IsNil)\n\tc.Assert(a, check.NotNil)\n\n\tctx, cancel := context.WithCancel(context.TODO())\n\tdefer cancel()\n\tserver := storage.Server{\n\t\tInstanceID: \"instance-to-delete\",\n\t\tHostname: \"instance-to-delete.hostname\",\n\t}\n\top := newMockOperator(ops.Site{\n\t\tAccountID: \"1\",\n\t\tDomain: \"example.com\",\n\t\tClusterState: storage.ClusterState{\n\t\t\tServers: []storage.Server{server},\n\t\t},\n\t})\n\tgo a.ProcessEvents(ctx, queue.url, op)\n\n\t// send terminated event\n\tmsg := &message{\n\t\treceipt: \"message-1\",\n\t\tbody: mustMarshalHook(HookEvent{\n\t\t\tInstanceID: server.InstanceID,\n\t\t\tType: InstanceTerminating,\n\t\t}),\n\t}\n\tselect {\n\tcase <-time.After(time.Second):\n\t\tc.Fatalf(\"timeout\")\n\tcase queue.messagesC <- msg:\n\t}\n\n\t// expect the shrink operation to arrive and the message to be deleted\n\tselect {\n\tcase op := <-op.shrinksC:\n\t\tc.Assert(op.Servers, check.DeepEquals, []string{server.Hostname})\n\tcase <-time.After(time.Second):\n\t\tc.Fatalf(\"timeout\")\n\t}\n\n\t// expect the message to be scheduled for deletion\n\tselect {\n\tcase m := <-queue.deletedC:\n\t\tc.Assert(aws.StringValue(m.ReceiptHandle), check.DeepEquals, msg.receipt)\n\tcase <-time.After(time.Second):\n\t\tc.Fatalf(\"timeout\")\n\t}\n}", "func (client BaseClient) DeleteFeatureInstanceResponder(resp *http.Response) (result FeatureInstance, err error) {\n\terr = autorest.Respond(\n\t\tresp,\n\t\tclient.ByInspecting(),\n\t\tazure.WithErrorUnlessStatusCode(http.StatusOK),\n\t\tautorest.ByUnmarshallingJSON(&result),\n\t\tautorest.ByClosing())\n\tresult.Response = autorest.Response{Response: resp}\n\treturn\n}", "func (client BaseClient) DeleteFeatureInstancePreparer(ctx context.Context, featureName string, featureVersion string, instanceName string) (*http.Request, error) {\n\tpathParameters := map[string]interface{}{\n\t\t\"featureName\": autorest.Encode(\"path\", featureName),\n\t\t\"featureVersion\": autorest.Encode(\"path\", featureVersion),\n\t\t\"instanceName\": autorest.Encode(\"path\", instanceName),\n\t}\n\n\tpreparer := autorest.CreatePreparer(\n\t\tautorest.AsDelete(),\n\t\tautorest.WithBaseURL(client.BaseURI),\n\t\tautorest.WithPathParameters(\"/api/features/instances/{featureName}/{featureVersion}/{instanceName}\", pathParameters))\n\treturn preparer.Prepare((&http.Request{}).WithContext(ctx))\n}", "func (client BaseClient) DeleteFeatureInstanceSender(req *http.Request) (*http.Response, error) {\n\treturn autorest.SendWithSender(client, req,\n\t\tautorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n}", "func (c *Client) DeleteInstance(name string) error {\n\tid, err := c.getDeviceID(name)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = c.client.Devices.Delete(id, true)\n\treturn err\n}", "func (tx *Tx) Delete(instance Instance) error {\n\ttx.Process(func(iter *Iter) {\n\t\tif iter.MatchInstance(tx.resource, instance) {\n\t\t\titer.Delete()\n\t\t}\n\t})\n\treturn nil\n}", "func (client *Client) DeleteInstanceWithOptions(request *DeleteInstanceRequest, runtime *util.RuntimeOptions) (_result *DeleteInstanceResponse, _err error) {\n\t_err = util.ValidateModel(request)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\tquery := map[string]interface{}{}\n\tif !tea.BoolValue(util.IsUnset(request.GlobalInstanceId)) {\n\t\tquery[\"GlobalInstanceId\"] = request.GlobalInstanceId\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.InstanceId)) {\n\t\tquery[\"InstanceId\"] = request.InstanceId\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.OwnerAccount)) {\n\t\tquery[\"OwnerAccount\"] = request.OwnerAccount\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.OwnerId)) {\n\t\tquery[\"OwnerId\"] = request.OwnerId\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.ResourceOwnerAccount)) {\n\t\tquery[\"ResourceOwnerAccount\"] = request.ResourceOwnerAccount\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.ResourceOwnerId)) {\n\t\tquery[\"ResourceOwnerId\"] = request.ResourceOwnerId\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.SecurityToken)) {\n\t\tquery[\"SecurityToken\"] = request.SecurityToken\n\t}\n\n\treq := &openapi.OpenApiRequest{\n\t\tQuery: openapiutil.Query(query),\n\t}\n\tparams := &openapi.Params{\n\t\tAction: tea.String(\"DeleteInstance\"),\n\t\tVersion: tea.String(\"2015-01-01\"),\n\t\tProtocol: tea.String(\"HTTPS\"),\n\t\tPathname: tea.String(\"/\"),\n\t\tMethod: tea.String(\"POST\"),\n\t\tAuthType: tea.String(\"AK\"),\n\t\tStyle: tea.String(\"RPC\"),\n\t\tReqBodyType: tea.String(\"formData\"),\n\t\tBodyType: tea.String(\"json\"),\n\t}\n\t_result = &DeleteInstanceResponse{}\n\t_body, _err := client.CallApi(params, req, runtime)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\t_err = tea.Convert(_body, &_result)\n\treturn _result, _err\n}", "func (f *FakeInstance) Delete(_ context.Context, _ string) error {\n\tpanic(\"implement me\")\n}", "func (a *AdminApiService) DeleteTarget(ctx _context.Context, id string) (*_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/admin/target/{id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"id\"+\"}\", _neturl.QueryEscape(parameterToString(id, \"\")) , -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 401 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 403 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 500 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t}\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func (r *kRegister) DeleteTargetByService(svc *register.Service) error {\n\treturn fmt.Errorf(\"not implemented\")\n}", "func DeleteInstance(instance *spotcluster.Instance,\n\tpool *spotcluster.Pool) error {\n\n\tif pool == nil {\n\t\treturn errors.New(\"got nil pool object\")\n\t}\n\n\tif instance == nil {\n\t\treturn errors.New(\"got nil instance object\")\n\t}\n\n\tclient := godo.NewFromToken(pool.ProviderSpec.DigitalOcean.APIKey)\n\tif client == nil {\n\t\treturn errors.New(\"got nil godo client\")\n\t}\n\n\tdoc := Client{\n\t\tProvider: client,\n\t}\n\n\treturn doc.Delete(string(instance.GetUID()))\n}", "func (w *Watcher) DeleteTarget(targetName string) (error) {\n\tmutableMutex.Lock()\n\tdefer mutableMutex.Unlock()\n\tif w.TargetMap == nil {\n\t\tw.TargetMap = make(map[string]*Target)\n\t}\n\t_, ok := w.TargetMap[targetName]\n\tif !ok {\n\t\treturn errors.Errorf(\"not exist domain\")\n\t}\n\tdelete(w.TargetMap, targetName)\n\treturn nil\n}", "func (c *GRPCClient) DeleteInstance(ctx context.Context, cloudID string) error {\n\treq := &pb.DeleteInstanceRequest{\n\t\tCloudId: cloudID,\n\t}\n\t_, err := c.client.DeleteInstance(ctx, req)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to DeleteInstance: %w\", err)\n\t}\n\n\treturn nil\n}", "func (r *apiRegister) DeleteTargetByService(svc *register.Service) error {\n\treturn fmt.Errorf(\"Not Implemented\")\n}", "func DeleteInstance(c GCECloud, instanceSelfLink string) error {\n\tklog.V(2).Infof(\"Deleting GCE Instance %s\", instanceSelfLink)\n\tu, err := ParseGoogleCloudURL(instanceSelfLink)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\top, err := c.Compute().Instances().Delete(u.Project, u.Zone, u.Name)\n\tif err != nil {\n\t\tif IsNotFound(err) {\n\t\t\tklog.Infof(\"Instance not found, assuming deleted: %q\", instanceSelfLink)\n\t\t\treturn nil\n\t\t}\n\t\treturn fmt.Errorf(\"error deleting Instance %s: %v\", instanceSelfLink, err)\n\t}\n\n\treturn c.WaitForOp(op)\n}", "func DeleteInstanceTemplate(c GCECloud, selfLink string) error {\n\tklog.V(2).Infof(\"Deleting GCE InstanceTemplate %s\", selfLink)\n\tu, err := ParseGoogleCloudURL(selfLink)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\top, err := c.Compute().InstanceTemplates().Delete(u.Project, u.Name)\n\tif err != nil {\n\t\tif IsNotFound(err) {\n\t\t\tklog.Infof(\"instancetemplate not found, assuming deleted: %q\", selfLink)\n\t\t\treturn nil\n\t\t}\n\t\treturn fmt.Errorf(\"error deleting InstanceTemplate %s: %v\", selfLink, err)\n\t}\n\n\treturn c.WaitForOp(op)\n}", "func TestDelete(t *testing.T) {\n\tRunWithInstance(func(instance *Instance) {\n\t\tInsertFixtures(instance, []EntryFixture{\n\t\t\t{Name: \"int\", Value: \"2891\", ValueType: 1},\n\t\t\t{Name: \"string\", Value: \"hello world!\", ValueType: 3},\n\t\t})\n\n\t\tif err := instance.Delete(\"int\"); err != nil {\n\t\t\tt.Error(\"Instance.Delete: got error:\\n\", err)\n\t\t}\n\n\t\tif err := panicked(func() { instance.MustDelete(\"string\") }); err != nil {\n\t\t\tt.Error(\"Instance.MustDelete: got panic:\\n\", err)\n\t\t}\n\n\t\tif err := instance.Delete(\"foo\"); err == nil {\n\t\t\tt.Error(\"Instance.Delete: expected error with non-existent entry\")\n\t\t} else if _, ok := err.(*ErrNoEntry); !ok {\n\t\t\tt.Error(\"Instance.Delete: expected error of type *ErrNoEntry\")\n\t\t}\n\n\t\tif err := panicked(func() { instance.MustDelete(\"foo\") }); err == nil {\n\t\t\tt.Error(\"Instance.MustDelete: expected panic with non-existent entry\")\n\t\t}\n\t})\n}", "func (a ProcessInstanceApi) DeleteInstance(processId string, instanceId string) (*ResultSuccess, *APIResponse, error) {\n\n\tvar localVarHttpMethod = strings.ToUpper(\"Delete\")\n\t// create path and map variables\n\tlocalVarPath := a.Configuration.BasePath + \"/processes/{process_id}/instances/{instance_id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"process_id\"+\"}\", fmt.Sprintf(\"%v\", processId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"instance_id\"+\"}\", fmt.Sprintf(\"%v\", instanceId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := make(map[string]string)\n\tvar localVarPostBody interface{}\n\tvar localVarFileName string\n\tvar localVarFileBytes []byte\n\t// authentication '(PasswordGrant)' required\n\t// set key with prefix in header\n\tlocalVarHeaderParams[\"Authorization\"] = a.Configuration.GetAPIKeyWithPrefix(\"Authorization\")\n\t// add default headers if any\n\tfor key := range a.Configuration.DefaultHeader {\n\t\tlocalVarHeaderParams[key] = a.Configuration.DefaultHeader[key]\n\t}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{ \"application/json\", }\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := a.Configuration.APIClient.SelectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\n\t\t\"application/json\",\n\t\t}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := a.Configuration.APIClient.SelectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tvar successPayload = new(ResultSuccess)\n\tlocalVarHttpResponse, err := a.Configuration.APIClient.CallAPI(localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\n\tvar localVarURL, _ = url.Parse(localVarPath)\n\tlocalVarURL.RawQuery = localVarQueryParams.Encode()\n\tvar localVarAPIResponse = &APIResponse{Operation: \"DeleteInstance\", Method: localVarHttpMethod, RequestURL: localVarURL.String()}\n\tif localVarHttpResponse != nil {\n\t\tlocalVarAPIResponse.Response = localVarHttpResponse.RawResponse\n\t\tlocalVarAPIResponse.Payload = localVarHttpResponse.Body()\n\t}\n\n\tif err != nil {\n\t\treturn successPayload, localVarAPIResponse, err\n\t}\n\terr = json.Unmarshal(localVarHttpResponse.Body(), &successPayload)\n\treturn successPayload, localVarAPIResponse, err\n}", "func (s *TargetCRUD) Delete(arg ...crud.Arg) (crud.Arg, error) {\n\tevent := eventFromArg(arg[0])\n\ttarget := targetFromStuct(event)\n\tprint.DeletePrintln(\"deleting target\", *target.Target.Target,\n\t\t\"from upstream\", *target.Upstream.ID)\n\treturn target, nil\n}", "func (m *Manager) Delete(target string) error {\n\targs := []string{\n\t\t\"delete\", \"node\", target,\n\t}\n\n\t// Delete the node\n\toutput, err := m.Command.Run(args...)\n\tlog.Println(string(output))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (a *AssembliesApiService) DeleteInstance(ctx _context.Context, did string, eid string, wid string, nid string) (*_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/api/assemblies/d/{did}/w/{wid}/e/{eid}/instance/nodeid/{nid}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"did\"+\"}\", _neturl.QueryEscape(parameterToString(did, \"\")) , -1)\n\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"eid\"+\"}\", _neturl.QueryEscape(parameterToString(eid, \"\")) , -1)\n\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"wid\"+\"}\", _neturl.QueryEscape(parameterToString(wid, \"\")) , -1)\n\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"nid\"+\"}\", _neturl.QueryEscape(parameterToString(nid, \"\")) , -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/vnd.onshape.v1+json;charset=UTF-8;qs=0.1\", \"application/json;charset=UTF-8; qs=0.09\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func (m *InstancesManager) DeleteInstance(instanceID string) {\n\tm.mutex.Lock()\n\tdefer m.mutex.Unlock()\n\tm.instances.Delete(instanceID)\n}", "func (s *Service) Delete(ctx context.Context) error {\n\tlog := log.FromContext(ctx)\n\tlog.Info(\"Deleting instance resources\")\n\tinstanceSpec := s.scope.InstanceSpec(log)\n\tinstanceName := instanceSpec.Name\n\tinstanceKey := meta.ZonalKey(instanceName, s.scope.Zone())\n\tlog.V(2).Info(\"Looking for instance before deleting\", \"name\", instanceName, \"zone\", s.scope.Zone())\n\tinstance, err := s.instances.Get(ctx, instanceKey)\n\tif err != nil {\n\t\tif !gcperrors.IsNotFound(err) {\n\t\t\tlog.Error(err, \"Error looking for instance before deleting\", \"name\", instanceName)\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}\n\n\tif s.scope.IsControlPlane() {\n\t\tif err := s.deregisterControlPlaneInstance(ctx, instance); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tlog.V(2).Info(\"Deleting instance\", \"name\", instanceName, \"zone\", s.scope.Zone())\n\treturn gcperrors.IgnoreNotFound(s.instances.Delete(ctx, instanceKey))\n}", "func (s *StepTeardownInstance) Run(state multistep.StateBag) multistep.StepAction {\n\tconfig := state.Get(\"config\").(*Config)\n\tdriver := state.Get(\"driver\").(Driver)\n\tui := state.Get(\"ui\").(packer.Ui)\n\n\tname := config.InstanceName\n\tif name == \"\" {\n\t\treturn multistep.ActionHalt\n\t}\n\n\tui.Say(\"Deleting instance...\")\n\tinstanceLog, _ := driver.GetSerialPortOutput(config.Zone, name)\n\tstate.Put(\"instance_log\", instanceLog)\n\terrCh, err := driver.DeleteInstance(config.Zone, name)\n\tif err == nil {\n\t\tselect {\n\t\tcase err = <-errCh:\n\t\tcase <-time.After(config.stateTimeout):\n\t\t\terr = errors.New(\"time out while waiting for instance to delete\")\n\t\t}\n\t}\n\n\tif err != nil {\n\t\tui.Error(fmt.Sprintf(\n\t\t\t\"Error deleting instance. Please delete it manually.\\n\\n\"+\n\t\t\t\t\"Name: %s\\n\"+\n\t\t\t\t\"Error: %s\", name, err))\n\t\treturn multistep.ActionHalt\n\t}\n\tui.Message(\"Instance has been deleted!\")\n\tstate.Put(\"instance_name\", \"\")\n\n\treturn multistep.ActionContinue\n}", "func (m *InstanceManager) Delete(ctx context.Context, obj runtime.Object) (bool, error) {\n\tinstance, err := convertInstance(obj)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\tregion := scw.Region(instance.Spec.Region)\n\n\tresourceID := instance.Spec.InstanceID\n\tif resourceID == \"\" {\n\t\treturn true, nil\n\t}\n\n\t_, err = m.API.DeleteInstance(&rdb.DeleteInstanceRequest{\n\t\tRegion: region,\n\t\tInstanceID: resourceID,\n\t})\n\tif err != nil {\n\t\tif _, ok := err.(*scw.ResourceNotFoundError); ok {\n\t\t\treturn true, nil\n\t\t}\n\t\treturn false, err\n\t}\n\n\t//instance.Status.Status = strcase.ToCamel(instanceResp.Status.String())\n\n\treturn false, nil\n}", "func (_m *ISession) StageInstanceDelete(channelID string, options ...discordgo.RequestOption) error {\n\t_va := make([]interface{}, len(options))\n\tfor _i := range options {\n\t\t_va[_i] = options[_i]\n\t}\n\tvar _ca []interface{}\n\t_ca = append(_ca, channelID)\n\t_ca = append(_ca, _va...)\n\tret := _m.Called(_ca...)\n\n\tvar r0 error\n\tif rf, ok := ret.Get(0).(func(string, ...discordgo.RequestOption) error); ok {\n\t\tr0 = rf(channelID, options...)\n\t} else {\n\t\tr0 = ret.Error(0)\n\t}\n\n\treturn r0\n}", "func (c *MockFileStorageClient) DeleteMountTarget(ctx context.Context, id string) error {\n\treturn nil\n}", "func TestIPAddress_Instance_Delete(t *testing.T) {\n\tclient, instance, _, teardown, err := setupInstanceWithoutDisks(t, \"fixtures/TestIPAddress_Instance_Delete\")\n\tdefer teardown()\n\tif err != nil {\n\t\tt.Error(err)\n\t}\n\n\tip, err := client.AddInstanceIPAddress(context.TODO(), instance.ID, true)\n\tif err != nil {\n\t\tt.Fatalf(\"failed to allocate public IPv4 for instance (%d): %s\", instance.ID, err)\n\t}\n\n\ti, err := client.GetInstanceIPAddresses(context.TODO(), instance.ID)\n\tif err != nil {\n\t\tt.Fatalf(\"failed to get instance (%d) IP addresses: %s\", instance.ID, err)\n\t}\n\tif len(i.IPv4.Public) != 2 {\n\t\tt.Errorf(\"expected instance (%d) to have 2 public IPv4 addresses; got %d\", instance.ID, len(i.IPv4.Public))\n\t}\n\n\tif err := client.DeleteInstanceIPAddress(context.TODO(), instance.ID, ip.Address); err != nil {\n\t\tt.Fatalf(\"failed to delete instance (%d) public IPv4 address (%s): %s\", instance.ID, ip.Address, err)\n\t}\n\n\tif i, err = client.GetInstanceIPAddresses(context.TODO(), instance.ID); err != nil {\n\t\tt.Fatalf(\"failed to get instance (%d) IP addresses: %s\", instance.ID, err)\n\t}\n\tif len(i.IPv4.Public) != 1 {\n\t\tt.Errorf(\"expected instance (%d) to have 1 public IPv4 address; got %d\", instance.ID, len(i.IPv4.Public))\n\t}\n}", "func (client AccessGovernanceCPClient) deleteGovernanceInstance(ctx context.Context, request common.OCIRequest, binaryReqBody *common.OCIReadSeekCloser, extraHeaders map[string]string) (common.OCIResponse, error) {\n\n\thttpRequest, err := request.HTTPRequest(http.MethodDelete, \"/governanceInstances/{governanceInstanceId}\", binaryReqBody, extraHeaders)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar response DeleteGovernanceInstanceResponse\n\tvar httpResponse *http.Response\n\thttpResponse, err = client.Call(ctx, &httpRequest)\n\tdefer common.CloseBodyIfValid(httpResponse)\n\tresponse.RawResponse = httpResponse\n\tif err != nil {\n\t\tapiReferenceLink := \"https://docs.oracle.com/iaas/api/#/en/access-governance-cp/20220518/GovernanceInstance/DeleteGovernanceInstance\"\n\t\terr = common.PostProcessServiceError(err, \"AccessGovernanceCP\", \"DeleteGovernanceInstance\", apiReferenceLink)\n\t\treturn response, err\n\t}\n\n\terr = common.UnmarshalResponse(httpResponse, &response)\n\treturn response, err\n}", "func (g GCPClient) DeleteInstance(instance, zone string, wait bool) error {\n\tvar notFound bool\n\top, err := g.compute.Instances.Delete(g.projectName, zone, instance).Do()\n\tif err != nil {\n\t\tif _, ok := err.(*googleapi.Error); !ok {\n\t\t\treturn err\n\t\t}\n\t\tif err.(*googleapi.Error).Code != 404 {\n\t\t\treturn err\n\t\t}\n\t\tnotFound = true\n\t}\n\tif !notFound && wait {\n\t\tlog.Infof(\"Deleting existing instance...\")\n\t\tif err := g.pollZoneOperationStatus(op.Name, zone); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tlog.Infof(\"Instance %s deleted\", instance)\n\t}\n\treturn nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstanceMplsSignalingProtocolsLdpTargeted(ctx context.Context, name string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:mpls/frinx-openconfig-network-instance:signaling-protocols/frinx-openconfig-network-instance:ldp/frinx-openconfig-network-instance:targeted/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (s *API) DeleteInstance(req *DeleteInstanceRequest, opts ...scw.RequestOption) (*Instance, error) {\n\tvar err error\n\n\tif req.Region == \"\" {\n\t\tdefaultRegion, _ := s.client.GetDefaultRegion()\n\t\treq.Region = defaultRegion\n\t}\n\n\tif fmt.Sprint(req.Region) == \"\" {\n\t\treturn nil, errors.New(\"field Region cannot be empty in request\")\n\t}\n\n\tif fmt.Sprint(req.InstanceID) == \"\" {\n\t\treturn nil, errors.New(\"field InstanceID cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"DELETE\",\n\t\tPath: \"/rdb/v1/regions/\" + fmt.Sprint(req.Region) + \"/instances/\" + fmt.Sprint(req.InstanceID) + \"\",\n\t\tHeaders: http.Header{},\n\t}\n\n\tvar resp Instance\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func (mr *ClientMockRecorder) DeleteTarget(arg0, arg1 interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"DeleteTarget\", reflect.TypeOf((*Client)(nil).DeleteTarget), arg0, arg1)\n}", "func (c *FakeClient) Delete(instanceID string) error {\n\tc.DeleteCalled = true\n\treturn c.DeleteErr\n}", "func (c *TestClient) CreateTargetInstance(project, zone string, ti *compute.TargetInstance) error {\n\tif c.CreateTargetInstanceFn != nil {\n\t\treturn c.CreateTargetInstanceFn(project, zone, ti)\n\t}\n\treturn c.client.CreateTargetInstance(project, zone, ti)\n}", "func (client *Client) DeleteInstance(request *DeleteInstanceRequest) (_result *DeleteInstanceResponse, _err error) {\n\truntime := &util.RuntimeOptions{}\n\t_result = &DeleteInstanceResponse{}\n\t_body, _err := client.DeleteInstanceWithOptions(request, runtime)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\t_result = _body\n\treturn _result, _err\n}", "func (c *Client) CloudDeleteInstance(projectID, instanceID string) error {\n\terr := c.Delete(queryEscape(\"/cloud/project/%s/instance/%s\", projectID, instanceID), nil)\n\tif apierror, ok := err.(*APIError); ok && apierror.Code == 404 {\n\t\terr = nil\n\t}\n\treturn err\n}", "func DeleteRelationshipInstance(ctx context.Context, deploymentID, nodeName, instanceName string) error {\n\trelInstancePath := path.Join(consulutil.DeploymentKVPrefix, deploymentID, \"topology/relationship_instances\")\n\tnodeRelInstancePath := path.Join(relInstancePath, nodeName)\n\treqIndices, err := consulutil.GetKeys(nodeRelInstancePath)\n\tif err != nil {\n\t\treturn errors.Wrap(err, consulutil.ConsulGenericErrMsg)\n\t}\n\tfor _, reqindex := range reqIndices {\n\t\terr = consulutil.Delete(path.Join(reqindex, instanceName)+\"/\", true)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, consulutil.ConsulGenericErrMsg)\n\t\t}\n\t}\n\n\t// now delete from targets in relationships instances\n\treturn addOrRemoveInstanceFromTargetRelationship(ctx, deploymentID, nodeName, instanceName, false)\n}", "func DeleteInstanceIP(\n\tctx context.Context,\n\ttx *sql.Tx,\n\trequest *models.DeleteInstanceIPRequest) error {\n\tdeleteQuery := deleteInstanceIPQuery\n\tselectQuery := \"select count(uuid) from instance_ip where uuid = ?\"\n\tvar err error\n\tvar count int\n\tuuid := request.ID\n\tauth := common.GetAuthCTX(ctx)\n\tif auth.IsAdmin() {\n\t\trow := tx.QueryRowContext(ctx, selectQuery, uuid)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"not found\")\n\t\t}\n\t\trow.Scan(&count)\n\t\tif count == 0 {\n\t\t\treturn errors.New(\"Not found\")\n\t\t}\n\t\t_, err = tx.ExecContext(ctx, deleteQuery, uuid)\n\t} else {\n\t\tdeleteQuery += \" and owner = ?\"\n\t\tselectQuery += \" and owner = ?\"\n\t\trow := tx.QueryRowContext(ctx, selectQuery, uuid, auth.ProjectID())\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"not found\")\n\t\t}\n\t\trow.Scan(&count)\n\t\tif count == 0 {\n\t\t\treturn errors.New(\"Not found\")\n\t\t}\n\t\t_, err = tx.ExecContext(ctx, deleteQuery, uuid, auth.ProjectID())\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"delete failed\")\n\t}\n\n\terr = common.DeleteMetaData(tx, uuid)\n\tlog.WithFields(log.Fields{\n\t\t\"uuid\": uuid,\n\t}).Debug(\"deleted\")\n\treturn err\n}", "func (backRepoTask *BackRepoTaskStruct) CommitDeleteInstance(id uint) (Error error) {\n\n\ttask := (*backRepoTask.Map_TaskDBID_TaskPtr)[id]\n\n\t// task is not staged anymore, remove taskDB\n\ttaskDB := (*backRepoTask.Map_TaskDBID_TaskDB)[id]\n\tquery := backRepoTask.db.Unscoped().Delete(&taskDB)\n\tif query.Error != nil {\n\t\treturn query.Error\n\t}\n\n\t// update stores\n\tdelete((*backRepoTask.Map_TaskPtr_TaskDBID), task)\n\tdelete((*backRepoTask.Map_TaskDBID_TaskPtr), id)\n\tdelete((*backRepoTask.Map_TaskDBID_TaskDB), id)\n\n\treturn\n}", "func (p *OnPrem) StopInstance(ctx *Context, instancename string) error {\n\treturn fmt.Errorf(\"Operation not supported\")\n}", "func (p *OnPrem) StopInstance(ctx *Context, instancename string) error {\n\treturn fmt.Errorf(\"Operation not supported\")\n}", "func (client StorageTargetsClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) {\n\terr = autorest.Respond(\n\t\tresp,\n\t\tazure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent),\n\t\tautorest.ByClosing())\n\tresult.Response = resp\n\treturn\n}", "func (d *DestinationClient) DeleteInstanceDestination(name string) (AffectedRecords, error) {\n\n\tvar retval AffectedRecords\n\tvar errResponse ErrorMessage\n\n\tresponse, err := d.restyClient.R().\n\t\tSetResult(&retval).\n\t\tSetError(&errResponse).\n\t\tSetPathParams(map[string]string{\n\t\t\t\"name\": name,\n\t\t}).\n\t\tDelete(\"/instanceDestinations/{name}\")\n\n\tif err != nil {\n\t\treturn retval, err\n\t}\n\tif response.StatusCode() != 200 {\n\t\terrResponse.statusCode = response.StatusCode()\n\t\treturn retval, errResponse\n\t}\n\treturn retval, nil\n}", "func (c *TestClient) StopInstance(project, zone, name string) error {\n\tif c.StopInstanceFn != nil {\n\t\treturn c.StopInstanceFn(project, zone, name)\n\t}\n\treturn c.client.StopInstance(project, zone, name)\n}", "func (instance *Host) Delete(ctx context.Context) (ferr fail.Error) {\n\tdefer fail.OnPanic(&ferr)\n\n\tdefer func() {\n\t\t// drop the cache when we are done creating the cluster\n\t\tif ka, err := instance.Service().GetCache(context.Background()); err == nil {\n\t\t\tif ka != nil {\n\t\t\t\t_ = ka.Clear(context.Background())\n\t\t\t}\n\t\t}\n\t}()\n\n\tif valid.IsNil(instance) {\n\t\treturn fail.InvalidInstanceError()\n\t}\n\tif ctx == nil {\n\t\treturn fail.InvalidParameterCannotBeNilError(\"ctx\")\n\t}\n\n\txerr := instance.Inspect(ctx, func(clonable data.Clonable, props *serialize.JSONProperties) fail.Error {\n\t\t// Do not remove a Host that is a gateway\n\t\treturn props.Inspect(hostproperty.NetworkV2, func(clonable data.Clonable) fail.Error {\n\t\t\thostNetworkV2, ok := clonable.(*propertiesv2.HostNetworking)\n\t\t\tif !ok {\n\t\t\t\treturn fail.InconsistentError(\"'*propertiesv2.HostNetworking' expected, '%s' provided\", reflect.TypeOf(clonable).String())\n\t\t\t}\n\n\t\t\tif hostNetworkV2.IsGateway {\n\t\t\t\treturn fail.NotAvailableError(\"cannot delete Host, it's a gateway that can only be deleted through its Subnet\")\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\t})\n\txerr = debug.InjectPlannedFail(xerr)\n\tif xerr != nil {\n\t\treturn xerr\n\t}\n\n\txerr = instance.RelaxedDeleteHost(cleanupContextFrom(ctx))\n\treturn xerr\n}", "func (m *AccessReviewsDefinitionsItemInstancesItemStagesItemDecisionsItemInstanceRequestBuilder) Delete(ctx context.Context, requestConfiguration *AccessReviewsDefinitionsItemInstancesItemStagesItemDecisionsItemInstanceRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func (c *TestClient) GetTargetInstance(project, zone, name string) (*compute.TargetInstance, error) {\n\tif c.GetTargetInstanceFn != nil {\n\t\treturn c.GetTargetInstanceFn(project, zone, name)\n\t}\n\treturn c.client.GetTargetInstance(project, zone, name)\n}", "func (ts *TargetSyncer) NewTarget(session sarama.ConsumerGroupSession, claim sarama.ConsumerGroupClaim) (RunnableTarget, error) {\n\tdiscoveredLabels := model.LabelSet{\n\t\t\"__meta_kafka_topic\": model.LabelValue(claim.Topic()),\n\t\t\"__meta_kafka_partition\": model.LabelValue(fmt.Sprintf(\"%d\", claim.Partition())),\n\t\t\"__meta_kafka_member_id\": model.LabelValue(session.MemberID()),\n\t\t\"__meta_kafka_group_id\": model.LabelValue(ts.cfg.KafkaConfig.GroupID),\n\t}\n\tdetails := newDetails(session, claim)\n\tlabelMap := make(map[string]string)\n\tfor k, v := range discoveredLabels.Clone().Merge(ts.cfg.KafkaConfig.Labels) {\n\t\tlabelMap[string(k)] = string(v)\n\t}\n\tlabelOut := format(labels.FromMap(labelMap), ts.cfg.RelabelConfigs)\n\tif len(labelOut) == 0 {\n\t\tlevel.Warn(ts.logger).Log(\"msg\", \"dropping target\", \"reason\", \"no labels\", \"details\", details, \"discovered_labels\", discoveredLabels.String())\n\t\treturn &runnableDroppedTarget{\n\t\t\tTarget: target.NewDroppedTarget(\"dropping target, no labels\", discoveredLabels),\n\t\t\trunFn: func() {\n\t\t\t\tfor range claim.Messages() {\n\t\t\t\t}\n\t\t\t},\n\t\t}, nil\n\t}\n\tt := NewKafkaTarget(\n\t\tts.logger,\n\t\tsession,\n\t\tclaim,\n\t\tdiscoveredLabels,\n\t\tlabelOut,\n\t\tts.cfg.RelabelConfigs,\n\t\tts.client,\n\t\tts.cfg.KafkaConfig.UseIncomingTimestamp,\n\t\tts.messageParser,\n\t)\n\n\treturn t, nil\n}", "func (t *TargetToExtent) Delete(server *Server) (*http.Response, error) {\n\tendpoint := fmt.Sprintf(\"/api/v1.0/services/iscsi/targettoextent/%d/\", t.ID)\n\tresp, err := server.getSlingConnection().Delete(endpoint).Receive(nil, nil)\n\tif err != nil {\n\t\tglog.Warningln(err)\n\t}\n\treturn resp, err\n}", "func OVNInstanceDevicePortDelete(network Network, instanceID int, deviceName string, internalRoutes []*net.IPNet, externalRoutes []*net.IPNet) error {\n\t// Check network is of type OVN.\n\tn, ok := network.(*ovn)\n\tif !ok {\n\t\treturn fmt.Errorf(\"Network is not OVN type\")\n\t}\n\n\treturn n.instanceDevicePortDelete(instanceID, deviceName, internalRoutes, externalRoutes)\n}", "func (a *AdminApiService) DeleteTargetSegment(ctx _context.Context, id string) (*_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/admin/target-segment/{id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"id\"+\"}\", _neturl.QueryEscape(parameterToString(id, \"\")) , -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 401 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 403 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 500 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t}\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func (m Manager) DeleteServiceInstanceDetails(tx *pg.Tx, instanceID string) error {\n\treturn tx.Delete(&ServiceInstance{ID: instanceID})\n}", "func (client BaseClient) DeleteFeatureInstance(ctx context.Context, featureName string, featureVersion string, instanceName string) (result FeatureInstance, err error) {\n\treq, err := client.DeleteFeatureInstancePreparer(ctx, featureName, featureVersion, instanceName)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"DeleteFeatureInstance\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresp, err := client.DeleteFeatureInstanceSender(req)\n\tif err != nil {\n\t\tresult.Response = autorest.Response{Response: resp}\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"DeleteFeatureInstance\", resp, \"Failure sending request\")\n\t\treturn\n\t}\n\n\tresult, err = client.DeleteFeatureInstanceResponder(resp)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"DeleteFeatureInstance\", resp, \"Failure responding to request\")\n\t}\n\n\treturn\n}", "func (adm Admin) DropInstance(zkSvr string, cluster string, instance string) {\n\tconn := newConnection(adm.ZkSvr)\n\terr := conn.Connect()\n\tif err != nil {\n\t\tfmt.Println(\"Failed to connect to zookeeper.\")\n\t}\n\tdefer conn.Disconnect()\n\n\tkb := KeyBuilder{cluster}\n\tinstanceKey := kb.instance(instance)\n\terr = conn.Delete(instanceKey)\n\tmust(err)\n\n\tfmt.Printf(\"/%s/%s deleted from zookeeper.\\n\", cluster, instance)\n}", "func (p *ProxMox) StopInstance(ctx *lepton.Context, instanceID string) error {\n\n\treq, err := http.NewRequest(\"POST\", p.apiURL+\"/api2/json/nodes/\"+p.nodeNAME+\"/qemu/\"+instanceID+\"/status/stop\", nil)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\ttr := &http.Transport{\n\t\tTLSClientConfig: &tls.Config{InsecureSkipVerify: true},\n\t}\n\tclient := &http.Client{Transport: tr}\n\n\treq.Header.Add(\"Authorization\", \"PVEAPIToken=\"+p.tokenID+\"=\"+p.secret)\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\t_, err = io.ReadAll(resp.Body)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (d *driverMock) TerminateInstance(ctx context.Context, id string) error {\n\tif d.TerminateInstanceErr != nil {\n\t\treturn d.TerminateInstanceErr\n\t}\n\n\td.TerminateInstanceID = id\n\n\treturn nil\n}", "func DeleteDeviceTargetTemp(c *gin.Context) {\n\tdeviceName := c.Param(\"device_name\")\n\terr := database.DeleteDeviceTargetTemp(c, deviceName)\n\tif err != nil {\n\t\tc.Status(http.StatusNotFound)\n\t} else {\n\t\tc.Status(http.StatusOK)\n\t}\n}", "func (c *nodePoolCache) removeInstance(nodePoolID, instanceID string) error {\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\n\tklog.Infof(\"Deleting instance %q from node pool %q\", instanceID, nodePoolID)\n\t// always try to remove the instance. This call is idempotent\n\tscaleDown := true\n\tresp, err := c.okeClient.DeleteNode(context.Background(), oke.DeleteNodeRequest{\n\t\tNodePoolId: &nodePoolID,\n\t\tNodeId: &instanceID,\n\t\tIsDecrementSize: &scaleDown,\n\t})\n\n\tklog.Infof(\"Delete Node API returned response: %v, err: %v\", resp, err)\n\thttpResp := resp.HTTPResponse()\n\tvar success bool\n\tif httpResp != nil {\n\t\tstatusCode := httpResp.StatusCode\n\t\t// status returned should be a 202, but let's accept any 2XX codes anyway\n\t\tstatusSuccess := statusCode >= 200 && statusCode < 300\n\t\tsuccess = statusSuccess ||\n\t\t\t// 409 means the instance is already going to be processed for deletion\n\t\t\tstatusCode == http.StatusConflict ||\n\t\t\t// 404 means it is probably already deleted and our cache may be stale\n\t\t\tstatusCode == http.StatusNotFound\n\t\tif !success {\n\t\t\tstatus := httpResp.Status\n\t\t\tklog.Infof(\"Received error status %s while deleting node %q\", status, instanceID)\n\n\t\t\t// statuses that we might expect but are still errors:\n\t\t\t// 400s (if cluster still uses TA or is v1 based)\n\t\t\t// 401 unauthorized\n\t\t\t// 412 etag mismatch\n\t\t\t// 429 too many requests\n\t\t\t// 500 internal server errors\n\t\t\treturn errors.Errorf(\"received error status %s while deleting node %q\", status, instanceID)\n\t\t} else if statusSuccess {\n\t\t\t// since delete node endpoint scales down by 1, we need to update the cache's target size by -1 too\n\t\t\tc.targetSize[nodePoolID]--\n\t\t}\n\t}\n\n\tif !success && err != nil {\n\t\treturn err\n\t}\n\n\tnodePool := c.cache[nodePoolID]\n\t// theoretical max number of nodes inside a cluster is 1000\n\t// so at most we'll be copying 1000 nodes\n\tnewNodeSlice := make([]oke.Node, 0, len(nodePool.Nodes))\n\tfor _, node := range nodePool.Nodes {\n\t\tif *node.Id != instanceID {\n\t\t\tnewNodeSlice = append(newNodeSlice, node)\n\t\t} else {\n\t\t\tklog.Infof(\"Deleting instance %q from cache\", instanceID)\n\t\t}\n\t}\n\tnodePool.Nodes = newNodeSlice\n\n\treturn nil\n}", "func (o *FeatureCvtermDbxref) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no FeatureCvtermDbxref provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), featureCvtermDbxrefPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"feature_cvterm_dbxref\\\" WHERE \\\"feature_cvterm_dbxref_id\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to delete from feature_cvterm_dbxref\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (adm Admin) DropInstance(cluster string, instance string) error {\n\tkb := KeyBuilder{cluster}\n\tinstanceKey := kb.instance(instance)\n\terr := adm.zkClient.DeleteTree(instanceKey)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfmt.Printf(\"/%s/%s deleted from zookeeper.\\n\", cluster, instance)\n\n\treturn err\n}", "func (s *JobLogEventData) SetTargetInstanceID(v string) *JobLogEventData {\n\ts.TargetInstanceID = &v\n\treturn s\n}", "func Delete(client *gophercloud.ServiceClient, instanceID, dbName string) (r DeleteResult) {\n\tresp, err := client.Delete(dbURL(client, instanceID, dbName), nil)\n\t_, r.Header, r.Err = gophercloud.ParseResponse(resp, err)\n\treturn\n}", "func (client Client) DeleteResponder(resp *http.Response) (result VolumeInstancesDeleteResponse, err error) {\n\terr = autorest.Respond(\n\t\tresp,\n\t\tazure.WithErrorUnlessStatusCode(http.StatusOK),\n\t\tautorest.ByUnmarshallingJSON(&result),\n\t\tautorest.ByClosing())\n\tresult.Response = autorest.Response{Response: resp}\n\treturn\n}", "func (client *VirtualMachineScaleSetsClientMock) DeleteInstances(ctx context.Context, resourceGroupName string, vmScaleSetName string, vmInstanceIDs compute.VirtualMachineScaleSetVMInstanceRequiredIDs) *retry.Error {\n\targs := client.Called(resourceGroupName, vmScaleSetName, vmInstanceIDs)\n\tif args.Error(1) != nil {\n\t\treturn &retry.Error{RawError: args.Error(1)}\n\t}\n\treturn nil\n}", "func (a *BackendOptionsApiService) DeleteLogTarget(ctx _context.Context, id int32, parentName string, parentType string, localVarOptionals *DeleteLogTargetOpts) (*_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/services/haproxy/configuration/log_targets/{id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"id\"+\"}\", _neturl.QueryEscape(fmt.Sprintf(\"%v\", id)), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tlocalVarQueryParams.Add(\"parent_name\", parameterToString(parentName, \"\"))\n\tlocalVarQueryParams.Add(\"parent_type\", parameterToString(parentType, \"\"))\n\tif localVarOptionals != nil && localVarOptionals.TransactionId.IsSet() {\n\t\tlocalVarQueryParams.Add(\"transaction_id\", parameterToString(localVarOptionals.TransactionId.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Version.IsSet() {\n\t\tlocalVarQueryParams.Add(\"version\", parameterToString(localVarOptionals.Version.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.ForceReload.IsSet() {\n\t\tlocalVarQueryParams.Add(\"force_reload\", parameterToString(localVarOptionals.ForceReload.Value(), \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v ModelError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tvar v ModelError\n\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\tif err != nil {\n\t\t\tnewErr.error = err.Error()\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tnewErr.model = v\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstancePolicyForwarding(ctx context.Context, name string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:policy-forwarding/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstance(ctx context.Context, name string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (client AccessGovernanceCPClient) DeleteGovernanceInstance(ctx context.Context, request DeleteGovernanceInstanceRequest) (response DeleteGovernanceInstanceResponse, err error) {\n\tvar ociResponse common.OCIResponse\n\tpolicy := common.DefaultRetryPolicy()\n\tif client.RetryPolicy() != nil {\n\t\tpolicy = *client.RetryPolicy()\n\t}\n\tif request.RetryPolicy() != nil {\n\t\tpolicy = *request.RetryPolicy()\n\t}\n\n\tif !(request.OpcRetryToken != nil && *request.OpcRetryToken != \"\") {\n\t\trequest.OpcRetryToken = common.String(common.RetryToken())\n\t}\n\n\tociResponse, err = common.Retry(ctx, request, client.deleteGovernanceInstance, policy)\n\tif err != nil {\n\t\tif ociResponse != nil {\n\t\t\tif httpResponse := ociResponse.HTTPResponse(); httpResponse != nil {\n\t\t\t\topcRequestId := httpResponse.Header.Get(\"opc-request-id\")\n\t\t\t\tresponse = DeleteGovernanceInstanceResponse{RawResponse: httpResponse, OpcRequestId: &opcRequestId}\n\t\t\t} else {\n\t\t\t\tresponse = DeleteGovernanceInstanceResponse{}\n\t\t\t}\n\t\t}\n\t\treturn\n\t}\n\tif convertedResponse, ok := ociResponse.(DeleteGovernanceInstanceResponse); ok {\n\t\tresponse = convertedResponse\n\t} else {\n\t\terr = fmt.Errorf(\"failed to convert OCIResponse into DeleteGovernanceInstanceResponse\")\n\t}\n\treturn\n}", "func DeleteServiceInstanceDestination(serviceURL string, accessToken string, destinationName string) error {\n\tvar err error\n\tvar request *http.Request\n\tvar response *http.Response\n\tvar destinationsURL string\n\tvar payload = []byte{}\n\tvar body []byte\n\n\tdestinationsURL = serviceURL + \"/destination-configuration/v1/instanceDestinations/\" + destinationName\n\tlog.Tracef(\"Making request to: %s\\n\", destinationsURL)\n\trequest, err = http.NewRequest(\"DELETE\", destinationsURL, bytes.NewBuffer(payload))\n\tif err != nil {\n\t\treturn err\n\t}\n\trequest.Header.Set(\"Content-Type\", \"application/json\")\n\trequest.Header.Set(\"Authorization\", \"Bearer \"+accessToken)\n\n\tclient := &http.Client{}\n\tresponse, err = client.Do(request)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer response.Body.Close()\n\n\tif response.StatusCode > 201 {\n\t\tbody, err = ioutil.ReadAll(response.Body)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Could not delete destination: [%s] %+v\", response.Status, body)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (w *WorkflowInstance) Delete(trs *db.Transaction) error {\n\tif err := db.DeleteStructTx(trs.Tx, constants.TableCoreGroups, &db.Options{\n\t\tConditions: builder.Equal(\"code\", w.ID),\n\t}); err != nil {\n\t\treturn customerror.New(http.StatusInternalServerError, \"workflow instance delete\", err.Error())\n\t}\n\treturn nil\n}", "func (i instanceHandler) deleteHandler(w http.ResponseWriter, r *http.Request) {\n\tvars := mux.Vars(r)\n\tid := vars[\"instID\"]\n\n\terr := i.client.Delete(id)\n\tif err != nil {\n\t\tlog.Error(\"Error Deleting Instance\", log.Fields{\n\t\t\t\"error\": err,\n\t\t})\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(http.StatusAccepted)\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstanceFdb(ctx context.Context, name string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:fdb/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (f *FakeInstance) DeleteReverseIPv6(_ context.Context, _, _ string) error {\n\tpanic(\"implement me\")\n}", "func (c *TestClient) ListTargetInstances(project, zone string, opts ...ListCallOption) ([]*compute.TargetInstance, error) {\n\tif c.ListTargetInstancesFn != nil {\n\t\treturn c.ListTargetInstancesFn(project, zone, opts...)\n\t}\n\treturn c.client.ListTargetInstances(project, zone, opts...)\n}", "func (c *EC2) deleteAWSEC2Instance(id string) (err error) {\n\n\tterminateinstancesrequest := &TerminateInstancesInput{}\n\tterminateinstancesrequest.InstanceIds = append(terminateinstancesrequest.InstanceIds, &id)\n\tterminateinstancesresult, err := TerminateInstances(terminateinstancesrequest)\n\tif err == nil {\n\t\tdescribeinstancesrequest := &DescribeInstancesInput{}\n\t\tif err := awsutil.CopyValue(describeinstancesrequest, \"InstanceIds\", terminateinstancesrequest, \"InstanceIds[]\"); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif err := WaitUntilInstanceTerminated(describeinstancesrequest); err != nil {\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (backRepoFoo *BackRepoFooStruct) CommitDeleteInstance(id uint) (Error error) {\n\n\tfoo := (*backRepoFoo.Map_FooDBID_FooPtr)[id]\n\n\t// foo is not staged anymore, remove fooDB\n\tfooDB := (*backRepoFoo.Map_FooDBID_FooDB)[id]\n\tquery := backRepoFoo.db.Unscoped().Delete(&fooDB)\n\tif query.Error != nil {\n\t\treturn query.Error\n\t}\n\n\t// update stores\n\tdelete((*backRepoFoo.Map_FooPtr_FooDBID), foo)\n\tdelete((*backRepoFoo.Map_FooDBID_FooPtr), id)\n\tdelete((*backRepoFoo.Map_FooDBID_FooDB), id)\n\n\treturn\n}", "func (c *Client) DeleteInstance(instanceId string) error {\n\treturn DeleteInstance(c, instanceId)\n}", "func (client DataFlowClient) DeleteDataFlowResponder(resp *http.Response) (result autorest.Response, err error) {\n\terr = autorest.Respond(\n\t\tresp,\n\t\tazure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent),\n\t\tautorest.ByClosing())\n\tresult.Response = resp\n\treturn\n}", "func handleAppInstanceStatusDelete(ctxArg interface{}, key string,\n\tstatusArg interface{}) {\n\tctx := ctxArg.(*zedmanagerContext)\n\tpublishAppInstanceSummary(ctx)\n}", "func (c *FakeAWSSNSTargets) Delete(ctx context.Context, name string, opts v1.DeleteOptions) error {\n\t_, err := c.Fake.\n\t\tInvokes(testing.NewDeleteActionWithOptions(awssnstargetsResource, c.ns, name, opts), &v1alpha1.AWSSNSTarget{})\n\n\treturn err\n}", "func TestDeleteInstanceIPAddress(t *testing.T) {\n\tclient, instance, _, teardown, err := setupInstanceWithoutDisks(t, \"fixtures/TestDeleteInstanceIPAddress\")\n\tdefer teardown()\n\tif err != nil {\n\t\tt.Error(err)\n\t}\n\n\tip, err := client.AddInstanceIPAddress(context.TODO(), instance.ID, true)\n\tif err != nil {\n\t\tt.Fatalf(\"failed to allocate public IPv4 for instance (%d): %s\", instance.ID, err)\n\t}\n\n\ti, err := client.GetInstanceIPAddresses(context.TODO(), instance.ID)\n\tif err != nil {\n\t\tt.Fatalf(\"failed to get instance (%d) IP addresses: %s\", instance.ID, err)\n\t}\n\tif len(i.IPv4.Public) != 2 {\n\t\tt.Errorf(\"expected instance (%d) to have 2 public IPv4 addresses; got %d\", instance.ID, len(i.IPv4.Public))\n\t}\n\n\tif err := client.DeleteInstanceIPAddress(context.TODO(), instance.ID, ip.Address); err != nil {\n\t\tt.Fatalf(\"failed to delete instance (%d) public IPv4 address (%s): %s\", instance.ID, ip.Address, err)\n\t}\n\n\tif i, err = client.GetInstanceIPAddresses(context.TODO(), instance.ID); err != nil {\n\t\tt.Fatalf(\"failed to get instance (%d) IP addresses: %s\", instance.ID, err)\n\t}\n\tif len(i.IPv4.Public) != 1 {\n\t\tt.Errorf(\"expected instance (%d) to have 1 public IPv4 address; got %d\", instance.ID, len(i.IPv4.Public))\n\t}\n}", "func (r DeleteInstanceRequest) Send(ctx context.Context) (*DeleteInstanceResponse, error) {\n\tr.Request.SetContext(ctx)\n\terr := r.Request.Send()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresp := &DeleteInstanceResponse{\n\t\tDeleteInstanceOutput: r.Request.Data.(*DeleteInstanceOutput),\n\t\tresponse: &aws.Response{Request: r.Request},\n\t}\n\n\treturn resp, nil\n}", "func (r DeleteInstanceRequest) Send(ctx context.Context) (*DeleteInstanceResponse, error) {\n\tr.Request.SetContext(ctx)\n\terr := r.Request.Send()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresp := &DeleteInstanceResponse{\n\t\tDeleteInstanceOutput: r.Request.Data.(*DeleteInstanceOutput),\n\t\tresponse: &aws.Response{Request: r.Request},\n\t}\n\n\treturn resp, nil\n}", "func deleteRDSInstance(rdsClientSess *rds.RDS, restoreParams map[string]string) error {\n\trdsClusterName := restoreParams[\"restoreRDS\"]\n\trdsInstanceName := rdsClusterName + \"-0\" // TODO: this should be handled better\n\n\tinput := &rds.DeleteDBInstanceInput{\n\t\tDBInstanceIdentifier: aws.String(rdsInstanceName),\n\t\tSkipFinalSnapshot: aws.Bool(true),\n\t}\n\n\t_, err := rdsClientSess.DeleteDBInstance(input)\n\tif err != nil {\n\t\tif aerr, ok := err.(awserr.Error); ok {\n\t\t\tswitch aerr.Code() {\n\t\t\tcase rds.ErrCodeDBInstanceNotFoundFault:\n\t\t\t\tfmt.Println(rds.ErrCodeDBInstanceNotFoundFault, aerr.Error())\n\t\t\t\treturn fmt.Errorf(\"Error deleting RDS instance [%v] in RDS cluster [%v]\", rdsInstanceName, rdsClusterName)\n\t\t\tcase rds.ErrCodeInvalidDBInstanceStateFault:\n\t\t\t\tfmt.Println(rds.ErrCodeInvalidDBInstanceStateFault, aerr.Error())\n\t\t\t\treturn fmt.Errorf(\"Error deleting RDS instance [%v] in RDS cluster [%v]\", rdsInstanceName, rdsClusterName)\n\t\t\tcase rds.ErrCodeDBSnapshotAlreadyExistsFault:\n\t\t\t\tfmt.Println(rds.ErrCodeDBSnapshotAlreadyExistsFault, aerr.Error())\n\t\t\t\treturn fmt.Errorf(\"Error deleting RDS instance [%v] in RDS cluster [%v]\", rdsInstanceName, rdsClusterName)\n\t\t\tcase rds.ErrCodeSnapshotQuotaExceededFault:\n\t\t\t\tfmt.Println(rds.ErrCodeSnapshotQuotaExceededFault, aerr.Error())\n\t\t\t\treturn fmt.Errorf(\"Error deleting RDS instance [%v] in RDS cluster [%v]\", rdsInstanceName, rdsClusterName)\n\t\t\tcase rds.ErrCodeInvalidDBClusterStateFault:\n\t\t\t\tfmt.Println(rds.ErrCodeInvalidDBClusterStateFault, aerr.Error())\n\t\t\t\treturn fmt.Errorf(\"Error deleting RDS instance [%v] in RDS cluster [%v]\", rdsInstanceName, rdsClusterName)\n\t\t\tcase rds.ErrCodeDBInstanceAutomatedBackupQuotaExceededFault:\n\t\t\t\tfmt.Println(rds.ErrCodeDBInstanceAutomatedBackupQuotaExceededFault, aerr.Error())\n\t\t\t\treturn fmt.Errorf(\"Error deleting RDS instance [%v] in RDS cluster [%v]\", rdsInstanceName, rdsClusterName)\n\t\t\tdefault:\n\t\t\t\tfmt.Println(aerr.Error())\n\t\t\t\treturn fmt.Errorf(\"Error deleting RDS instance [%v] in RDS cluster [%v]\", rdsInstanceName, rdsClusterName)\n\t\t\t}\n\t\t} else {\n\t\t\t// Print the error, cast err to awserr.Error to get the Code and\n\t\t\t// Message from an error.\n\t\t\tfmt.Println(err.Error())\n\t\t\treturn fmt.Errorf(\"Error deleting RDS instance [%v] in RDS cluster [%v]\", rdsInstanceName, rdsClusterName)\n\t\t}\n\t}\n\n\t// TODO: DEBUG - fmt.Println(result)\n\tfmt.Printf(\"Deleting RDS instance [%v] in RDS cluster [%v]\\n\", rdsClusterName, rdsClusterName)\n\treturn nil\n}", "func DeleteOperator(namespace string) error {\n\tif namespace == \"\" {\n\t\tnamespace = DEFAULT_NAMESPACE\n\t}\n\tif err := sh.RunV(\"helm\", \"delete\", \"-n\", namespace, \"kedahttp\"); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (d *Driver) Delete(ctx context.Context, f *functions.Function) error {\n\tspan, ctx := trace.Trace(ctx, \"\")\n\tdefer span.Finish()\n\tif err := d.deleteContainers(ctx, f, true); err != nil {\n\t\t// no error wrapping, delete already does that\n\t\treturn err\n\t}\n\treturn nil\n}", "func (o *ContentUnitDerivation) Delete(exec boil.Executor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"mdbmodels: no ContentUnitDerivation provided for delete\")\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), contentUnitDerivationPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"content_unit_derivations\\\" WHERE \\\"source_id\\\"=$1 AND \\\"derived_id\\\"=$2\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: unable to delete from content_unit_derivations\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: failed to get rows affected by delete for content_unit_derivations\")\n\t}\n\n\treturn rowsAff, nil\n}" ]
[ "0.62708145", "0.6218664", "0.61142814", "0.6080617", "0.60300386", "0.59788084", "0.59749955", "0.59245163", "0.59131694", "0.583875", "0.58271617", "0.58113635", "0.57535976", "0.57502824", "0.5738181", "0.5693608", "0.5684962", "0.5676946", "0.5616006", "0.56058383", "0.5589527", "0.5565557", "0.5563521", "0.55350375", "0.5483614", "0.54617625", "0.54612684", "0.53764206", "0.5367666", "0.5360982", "0.53430665", "0.53403354", "0.5302848", "0.5295139", "0.52926", "0.52803785", "0.5257297", "0.5255792", "0.5250679", "0.52444625", "0.5229466", "0.5206922", "0.519567", "0.5182933", "0.5174808", "0.51538867", "0.5152623", "0.51373386", "0.51159745", "0.5080573", "0.50772035", "0.5062476", "0.5057244", "0.5057244", "0.50256", "0.5021513", "0.50193846", "0.5005031", "0.50003403", "0.500025", "0.49915802", "0.49881893", "0.4986567", "0.49579883", "0.49514467", "0.4930749", "0.49023873", "0.48868454", "0.4877547", "0.48706713", "0.48659423", "0.48637676", "0.4848449", "0.48211142", "0.48190227", "0.48133996", "0.48104987", "0.47957033", "0.4789985", "0.47870898", "0.47797516", "0.47783294", "0.47720185", "0.47661558", "0.47640803", "0.4752633", "0.47512197", "0.47378194", "0.47376838", "0.473495", "0.4729293", "0.4727782", "0.47236362", "0.4720937", "0.47098967", "0.47098967", "0.4707991", "0.4693237", "0.4682791", "0.46803477" ]
0.7660001
0
DeprecateImage uses the override method DeprecateImageFn or the real implementation.
func (c *TestClient) DeprecateImage(project, name string, deprecationstatus *compute.DeprecationStatus) error { if c.DeprecateImageFn != nil { return c.DeprecateImageFn(project, name, deprecationstatus) } return c.client.DeprecateImage(project, name, deprecationstatus) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func newUpdateImageDeprecateRequest(ctx context.Context, f *Image, c *Client) (map[string]interface{}, error) {\n\treq := map[string]interface{}{}\n\n\tif v, err := expandImageDeprecated(c, f.Deprecated); err != nil {\n\t\treturn nil, fmt.Errorf(\"error expanding Deprecated into deprecated: %w\", err)\n\t} else if !dcl.IsEmptyValueIndirect(v) {\n\t\treq[\"deprecated\"] = v\n\t}\n\treturn req, nil\n}", "func (op *updateImageDeprecateOperation) do(ctx context.Context, r *Image, c *Client) error {\n\t_, err := c.GetImage(ctx, r.urlNormalized())\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tu, err := r.updateURL(c.Config.BasePath, \"deprecate\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treq, err := newUpdateImageDeprecateRequest(ctx, r, c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tc.Config.Logger.Infof(\"Created update: %#v\", req)\n\tbody, err := marshalUpdateImageDeprecateRequest(c, req)\n\tif err != nil {\n\t\treturn err\n\t}\n\tresp, err := dcl.SendRequest(ctx, c.Config, \"POST\", u, bytes.NewBuffer(body), c.Config.RetryProvider)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar o operations.ComputeOperation\n\tif err := dcl.ParseResponse(resp.Response, &o); err != nil {\n\t\treturn err\n\t}\n\terr = o.Wait(ctx, c.Config, \"https://www.googleapis.com/compute/beta/\", \"GET\")\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func expandImageDeprecated(c *Client, f *ImageDeprecated) (map[string]interface{}, error) {\n\tif dcl.IsEmptyValueIndirect(f) {\n\t\treturn nil, nil\n\t}\n\n\tm := make(map[string]interface{})\n\tif v := f.State; !dcl.IsEmptyValueIndirect(v) {\n\t\tm[\"state\"] = v\n\t}\n\tif v := f.Replacement; !dcl.IsEmptyValueIndirect(v) {\n\t\tm[\"replacement\"] = v\n\t}\n\tif v := f.Deprecated; !dcl.IsEmptyValueIndirect(v) {\n\t\tm[\"deprecated\"] = v\n\t}\n\tif v := f.Obsolete; !dcl.IsEmptyValueIndirect(v) {\n\t\tm[\"obsolete\"] = v\n\t}\n\tif v := f.Deleted; !dcl.IsEmptyValueIndirect(v) {\n\t\tm[\"deleted\"] = v\n\t}\n\n\treturn m, nil\n}", "func flattenImageDeprecated(c *Client, i interface{}) *ImageDeprecated {\n\tm, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn nil\n\t}\n\n\tr := &ImageDeprecated{}\n\tr.State = flattenImageDeprecatedStateEnum(m[\"state\"])\n\tr.Replacement = dcl.FlattenString(m[\"replacement\"])\n\tr.Deprecated = dcl.FlattenString(m[\"deprecated\"])\n\tr.Obsolete = dcl.FlattenString(m[\"obsolete\"])\n\tr.Deleted = dcl.FlattenString(m[\"deleted\"])\n\n\treturn r\n}", "func marshalUpdateImageDeprecateRequest(c *Client, m map[string]interface{}) ([]byte, error) {\n\n\tm = EncodeImageDeprecateRequest(m)\n\treturn json.Marshal(m)\n}", "func (l *Logger) Deprecate(a ...interface{}) {\n\tif l.Verbosity <= VerbosityLevelQuiet {\n\t\treturn\n\t}\n\n\tif len(a) > 0 {\n\t\tmessage := fmt.Sprint(a[0])\n\t\tfor _, d := range l.deprecations {\n\t\t\tif message == d {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\tl.deprecations = append(l.deprecations, message)\n\t}\n\n\tl.logInStyle(deprecatedString, yellow, a...)\n\tfmt.Fprintln(l.Stderr)\n}", "func (o *Operation) Deprecate() *Operation {\n\to.Operation.Deprecate()\n\treturn o\n}", "func (o LookupImageResultOutput) Deprecated() DeprecationStatusResponseOutput {\n\treturn o.ApplyT(func(v LookupImageResult) DeprecationStatusResponse { return v.Deprecated }).(DeprecationStatusResponseOutput)\n}", "func (i ImageIndexer) DeprecateFromIndex(request DeprecateFromIndexRequest) error {\n\tbuildDir, outDockerfile, cleanup, err := buildContext(request.Generate, request.OutDockerfile)\n\tdefer cleanup()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdatabasePath, err := i.ExtractDatabase(buildDir, request.FromIndex, request.CaFile, request.SkipTLSVerify, request.PlainHTTP)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdeprecateFromRegistryReq := registry.DeprecateFromRegistryRequest{\n\t\tBundles: request.Bundles,\n\t\tInputDatabase: databasePath,\n\t\tPermissive: request.Permissive,\n\t\tAllowPackageRemoval: request.AllowPackageRemoval,\n\t}\n\n\t// Deprecate the bundles from the registry\n\terr = i.RegistryDeprecator.DeprecateFromRegistry(deprecateFromRegistryReq)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// generate the dockerfile\n\tdockerfile := i.DockerfileGenerator.GenerateIndexDockerfile(request.BinarySourceImage, databasePath)\n\terr = write(dockerfile, outDockerfile, i.Logger)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif request.Generate {\n\t\treturn nil\n\t}\n\n\t// build the dockerfile with requested tooling\n\terr = build(outDockerfile, request.Tag, i.CommandRunner, i.Logger)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (o *Operation) Undeprecate() *Operation {\n\to.Operation.Undeprecate()\n\treturn o\n}", "func ValidateImage(image string, metadata, labels map[string]string, watchByDef bool) (img model.Image, err error) {\n\tif i := strings.Index(image, \"@sha256:\"); i > 0 {\n\t\timage = image[:i]\n\t}\n\timg = model.Image{\n\t\tName: image,\n\t\tMetadata: metadata,\n\t\tNotifyOn: model.NotifyOnDefaults,\n\t\tSortTags: registry.SortTagReverse,\n\t}\n\n\tif enableStr, ok := labels[\"diun.enable\"]; ok {\n\t\tenable, err := strconv.ParseBool(enableStr)\n\t\tif err != nil {\n\t\t\treturn img, fmt.Errorf(\"cannot parse %s value of label diun.enable\", enableStr)\n\t\t}\n\t\tif !enable {\n\t\t\treturn model.Image{}, nil\n\t\t}\n\t} else if !watchByDef {\n\t\treturn model.Image{}, nil\n\t}\n\n\tfor key, value := range labels {\n\t\tswitch {\n\t\tcase key == \"diun.regopt\":\n\t\t\timg.RegOpt = value\n\t\tcase key == \"diun.watch_repo\":\n\t\t\tif img.WatchRepo, err = strconv.ParseBool(value); err != nil {\n\t\t\t\treturn img, fmt.Errorf(\"cannot parse %s value of label %s\", value, key)\n\t\t\t}\n\t\tcase key == \"diun.notify_on\":\n\t\t\tif len(value) == 0 {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\timg.NotifyOn = []model.NotifyOn{}\n\t\t\tfor _, no := range strings.Split(value, \";\") {\n\t\t\t\tnotifyOn := model.NotifyOn(no)\n\t\t\t\tif !notifyOn.Valid() {\n\t\t\t\t\treturn img, fmt.Errorf(\"unknown notify status %q\", value)\n\t\t\t\t}\n\t\t\t\timg.NotifyOn = append(img.NotifyOn, notifyOn)\n\t\t\t}\n\t\tcase key == \"diun.sort_tags\":\n\t\t\tif value == \"\" {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tsortTags := registry.SortTag(value)\n\t\t\tif !sortTags.Valid() {\n\t\t\t\treturn img, fmt.Errorf(\"unknown sort tags type %q\", value)\n\t\t\t}\n\t\t\timg.SortTags = sortTags\n\t\tcase key == \"diun.max_tags\":\n\t\t\tif img.MaxTags, err = strconv.Atoi(value); err != nil {\n\t\t\t\treturn img, fmt.Errorf(\"cannot parse %s value of label %s\", value, key)\n\t\t\t}\n\t\tcase key == \"diun.include_tags\":\n\t\t\timg.IncludeTags = strings.Split(value, \";\")\n\t\tcase key == \"diun.exclude_tags\":\n\t\t\timg.ExcludeTags = strings.Split(value, \";\")\n\t\tcase key == \"diun.hub_tpl\":\n\t\t\timg.HubTpl = value\n\t\tcase key == \"diun.hub_link\":\n\t\t\timg.HubLink = value\n\t\tcase key == \"diun.platform\":\n\t\t\tplatform, err := platforms.Parse(value)\n\t\t\tif err != nil {\n\t\t\t\treturn img, fmt.Errorf(\"cannot parse %s platform of label %s\", value, key)\n\t\t\t}\n\t\t\timg.Platform = model.ImagePlatform{\n\t\t\t\tOS: platform.OS,\n\t\t\t\tArch: platform.Architecture,\n\t\t\t\tVariant: platform.Variant,\n\t\t\t}\n\t\tcase strings.HasPrefix(key, \"diun.metadata.\"):\n\t\t\tmkey := strings.TrimPrefix(key, \"diun.metadata.\")\n\t\t\tif len(mkey) == 0 || len(value) == 0 {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tif err := validateMetadataKey(mkey); err != nil {\n\t\t\t\treturn img, errors.Wrapf(err, \"invalid metadata key %q\", mkey)\n\t\t\t}\n\t\t\tif img.Metadata == nil {\n\t\t\t\timg.Metadata = map[string]string{}\n\t\t\t}\n\t\t\timg.Metadata[mkey] = value\n\t\t}\n\t}\n\n\treturn img, nil\n}", "func (dfs *DistributedFilesystem) Override(newimg, oldimg string) error {\n\n\t// make sure the old image exists\n\toldImage, err := dfs.index.FindImage(oldimg)\n\tif err != nil {\n\t\tglog.Errorf(\"Could not find image %s in registry: %s\", oldimg, err)\n\t\treturn err\n\t}\n\n\t// make sure the new image exists\n\tnewImage, err := dfs.docker.FindImage(newimg)\n\tif err != nil {\n\t\tglog.Errorf(\"Could not find replacement image %s: %s\", newimg, err)\n\t\treturn err\n\t}\n\n\t// push the image into the registry\n\thash, err := dfs.docker.GetImageHash(newImage.ID)\n\tif err != nil {\n\t\tglog.Errorf(\"Could not get hash for image %s: %s\", newimg, err)\n\t\treturn err\n\t}\n\n\tif err := dfs.index.PushImage(oldImage.String(), newImage.ID, hash); err != nil {\n\t\tglog.Errorf(\"Could not replace image %s with %s (%s): %s\", oldImage, newimg, newImage.ID, err)\n\t\treturn err\n\t}\n\treturn nil\n}", "func fn2() { // want fn2:`Deprecated: Don't use this\\.`\n}", "func (v *Virt) ImageRemove(ctx context.Context, image string, force, prune bool) (names []string, err error) {\n\tlog.Warnf(ctx, \"does not implement\")\n\treturn\n}", "func (m Config) DelImage(name string, isForce bool) {\n\t//check exists image\n\tif _, ok := m.Images[name]; ok {\n\t\t//insert image\n\t\tdelete(m.Images, name)\n\t} else {\n\t\tfmt.Println(\"Image not found.:\", name)\n\t\texit(1)\n\t\treturn\n\t}\n\treturn\n}", "func (r *Reconciler) isProxyPluginDeprecated() bool {\n\timageParts := strings.Split(utils.PointerToString(r.Config.Spec.IstioCoreDNS.Image), \":\")\n\ttag := imageParts[1]\n\n\tv140 := semver.New(\"1.4.0\")\n\tvCoreDNSTag := semver.New(tag)\n\n\tif v140.LessThan(*vCoreDNSTag) {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (b *RouteBuilder) Deprecate() *RouteBuilder {\n\tb.deprecated = true\n\treturn b\n}", "func expandImageDeprecatedSlice(c *Client, f []ImageDeprecated) ([]map[string]interface{}, error) {\n\tif f == nil {\n\t\treturn nil, nil\n\t}\n\n\titems := []map[string]interface{}{}\n\tfor _, item := range f {\n\t\ti, err := expandImageDeprecated(c, &item)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\titems = append(items, i)\n\t}\n\n\treturn items, nil\n}", "func fn3() { // want fn3:`Deprecated: Don't use this\\.`\n}", "func flattenImageDeprecatedSlice(c *Client, i interface{}) []ImageDeprecated {\n\ta, ok := i.([]interface{})\n\tif !ok {\n\t\treturn []ImageDeprecated{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn []ImageDeprecated{}\n\t}\n\n\titems := make([]ImageDeprecated, 0, len(a))\n\tfor _, item := range a {\n\t\titems = append(items, *flattenImageDeprecated(c, item.(map[string]interface{})))\n\t}\n\n\treturn items\n}", "func Deprecated(depricated ...interface{}) {\n\tIssue(append([]interface{}{\"Deprecated: \"}, depricated...))\n}", "func DecompileImage(vorteilImage *vdecompiler.IO, outputPath string, skipNotTouched bool) (DecompileReport, error) {\n\treport := DecompileReport{\n\t\tImageFiles: make([]DecompiledFile, 0),\n\t\tSkipNotTouched: skipNotTouched,\n\t}\n\n\tfi, err := os.Stat(outputPath)\n\tif err != nil && !os.IsNotExist(err) {\n\t\treturn report, err\n\t}\n\tvar into bool\n\tif !os.IsNotExist(err) && fi.IsDir() {\n\t\tinto = true\n\t}\n\n\tfpath := \"/\"\n\tdpath := outputPath\n\tif into {\n\t\tdpath = filepath.ToSlash(filepath.Join(outputPath, filepath.Base(fpath)))\n\t}\n\n\tsymlinkCallbacks := make([]func() error, 0)\n\n\tino, err := vorteilImage.ResolvePathToInodeNo(fpath)\n\tif err != nil {\n\t\treturn report, err\n\t}\n\treport, symlinkCallbacks, err = decompileImageRecursive(vorteilImage, report, symlinkCallbacks, ino, filepath.ToSlash(filepath.Base(fpath)), dpath)\n\tif err != nil {\n\t\treturn report, err\n\t}\n\n\tfor _, fn := range symlinkCallbacks {\n\t\terr = fn()\n\t\tif err != nil {\n\t\t\tbreak\n\t\t}\n\t}\n\n\treturn report, err\n}", "func expandImageDeprecatedMap(c *Client, f map[string]ImageDeprecated) (map[string]interface{}, error) {\n\tif f == nil {\n\t\treturn nil, nil\n\t}\n\n\titems := make(map[string]interface{})\n\tfor k, item := range f {\n\t\ti, err := expandImageDeprecated(c, &item)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif i != nil {\n\t\t\titems[k] = i\n\t\t}\n\t}\n\n\treturn items, nil\n}", "func flattenImageDeprecatedMap(c *Client, i interface{}) map[string]ImageDeprecated {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]ImageDeprecated{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]ImageDeprecated{}\n\t}\n\n\titems := make(map[string]ImageDeprecated)\n\tfor k, item := range a {\n\t\titems[k] = *flattenImageDeprecated(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func ServiceUse(serviceID string, imageName string, registry string, noOp bool) (string, error) {\n\t// If noOp is True, then replace the 'real' functions that talk to Docker with\n\t// no-op functions (for dry run purposes)\n\tpullImage := PullImage\n\tfindImage := FindImage\n\ttagImage := TagImage\n\tif noOp {\n\t\tpullImage = noOpPullImage\n\t\tfindImage = noOpFindImage\n\t\ttagImage = noOpTagImage\n\t}\n\n\t// imageName is the new image to pull, eg. \"zenoss/resmgr-unstable:1.2.3.4\"\n\tglog.V(0).Infof(\"preparing to use image: %s\", imageName)\n\timageID, err := commons.ParseImageID(imageName)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif imageID.Tag == \"\" {\n\t\timageID.Tag = \"latest\"\n\t}\n\tglog.Infof(\"pulling image %s, this may take a while...\", imageID)\n\tif err := pullImage(imageID.String()); err != nil {\n\t\tglog.Warningf(\"unable to pull image %s\", imageID)\n\t}\n\n\t//verify image has been pulled\n\timg, err := findImage(imageID.String(), false)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"could not look up image %s: %s. Check your docker login and retry service deployment.\", imageID, err)\n\t\treturn \"\", err\n\t}\n\n\t//Tag images to latest all images\n\tvar newTag *commons.ImageID\n\n\tnewTag, err = commons.RenameImageID(registry, serviceID, imageID.String(), \"latest\")\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tglog.Infof(\"tagging image %s to %s \", imageName, newTag)\n\tif _, err = tagImage(img, newTag.String(), true); err != nil {\n\t\tglog.Errorf(\"could not tag image: %s (%v)\", imageName, err)\n\t\treturn \"\", err\n\t}\n\treturn newTag.String(), nil\n}", "func (*DeleteImageMetadata) Descriptor() ([]byte, []int) {\n\treturn file_yandex_cloud_compute_v1_image_service_proto_rawDescGZIP(), []int{9}\n}", "func Deprecated(options DeprecationOptions) OptionFunc {\n\treturn func(c *ControllerActivator) {\n\t\tc.Use(func(ctx *context.Context) {\n\t\t\tversioning.WriteDeprecated(ctx, options)\n\t\t\tctx.Next()\n\t\t})\n\t}\n}", "func resolveConformanceImage(imageVersion string) string {\n\t// TODO(johnschnake): This logic should be temporary and is only\n\t// required as we phase in the use of the upstream k8s kube-conformance\n\t// image instead of our own heptio/kube-conformance one. They started\n\t// publishing it for v1.14.1. (https://github.com/kubernetes/kubernetes/pull/76101)\n\tvar imageURL string\n\tswitch {\n\tcase imageVersion == imagepkg.ConformanceImageVersionLatest:\n\t\timageURL = config.UpstreamKubeConformanceImageURL\n\tcase imageVersion < \"v1.14.1\":\n\t\timageURL = config.DefaultKubeConformanceImageURL\n\tdefault:\n\t\timageURL = config.UpstreamKubeConformanceImageURL\n\t}\n\treturn fmt.Sprintf(\"%v:%v\", imageURL, imageVersion)\n\n}", "func (ref ostreeReference) DeleteImage(ctx context.Context, sys *types.SystemContext) error {\n\treturn errors.New(\"Deleting images not implemented for ostree: images\")\n}", "func ImageReplacement(t *testing.T) {\n\tctx := test.NewTestCtx(t)\n\tdefer ctx.Cleanup()\n\n\terr := deployOperator(t, ctx)\n\thelpers.AssertNoError(t, err)\n\n\tt.Run(\"create-cr\", testsuites.ValidateManualInstall)\n}", "func releaseImageData(img image.Image) {\n\tswitch raw := img.(type) {\n\tcase *image.Alpha:\n\t\traw.Pix = nil\n\tcase *image.Alpha16:\n\t\traw.Pix = nil\n\tcase *image.Gray:\n\t\traw.Pix = nil\n\tcase *image.Gray16:\n\t\traw.Pix = nil\n\tcase *image.NRGBA:\n\t\traw.Pix = nil\n\tcase *image.NRGBA64:\n\t\traw.Pix = nil\n\tcase *image.Paletted:\n\t\traw.Pix = nil\n\tcase *image.RGBA:\n\t\traw.Pix = nil\n\tcase *image.RGBA64:\n\t\traw.Pix = nil\n\tdefault:\n\t\treturn\n\t}\n}", "func UnmanagedImage(id, ref string, hasAnnotations bool, annotation, value string) imageapi.Image {\n\timage := ImageWithLayers(id, ref, nil)\n\tif !hasAnnotations {\n\t\timage.Annotations = nil\n\t} else {\n\t\tdelete(image.Annotations, imageapi.ManagedByOpenShiftAnnotation)\n\t\timage.Annotations[annotation] = value\n\t}\n\treturn image\n}", "func (o DocumentationRuleOutput) DeprecationDescription() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v DocumentationRule) *string { return v.DeprecationDescription }).(pulumi.StringPtrOutput)\n}", "func (a *Agent) DeprecatePlugin(plugin ids.PluginID) {\n\ta.oldPlugins = append(a.oldPlugins, plugin)\n}", "func deprecateHandler(rw http.ResponseWriter, r *http.Request, rule *controlplane.Rule) {\n\t// count hits\n\tinst.Counter(1.0, deprecate, 1)\n\n\tif rule.Payload == nil {\n\t\trw.WriteHeader(410)\n\t\trw.Write(deprecatePayload)\n\t} else {\n\t\tfor k, v := range rule.Payload.Headers {\n\t\t\trw.Header().Add(k, v)\n\t\t}\n\t\trw.WriteHeader(rule.Payload.HttpStatus)\n\t\trw.Write([]byte(rule.Payload.Body))\n\t}\n}", "func (_DelegateProfile *DelegateProfileTransactor) DeprecateField(opts *bind.TransactOpts, _name string) (*types.Transaction, error) {\n\treturn _DelegateProfile.contract.Transact(opts, \"deprecateField\", _name)\n}", "func PatchServiceImage(t *testing.T, clients *Clients, svc *v1alpha1.Service, imagePath string) (*v1alpha1.Service, error) {\n\tnewSvc := svc.DeepCopy()\n\tif svc.Spec.DeprecatedRunLatest != nil {\n\t\tnewSvc.Spec.DeprecatedRunLatest.Configuration.GetTemplate().Spec.GetContainer().Image = imagePath\n\t} else if svc.Spec.DeprecatedRelease != nil {\n\t\tnewSvc.Spec.DeprecatedRelease.Configuration.GetTemplate().Spec.GetContainer().Image = imagePath\n\t} else if svc.Spec.DeprecatedPinned != nil {\n\t\tnewSvc.Spec.DeprecatedPinned.Configuration.GetTemplate().Spec.GetContainer().Image = imagePath\n\t} else {\n\t\tnewSvc.Spec.ConfigurationSpec.GetTemplate().Spec.GetContainer().Image = imagePath\n\t}\n\tLogResourceObject(t, ResourceObjects{Service: newSvc})\n\tpatchBytes, err := createPatch(svc, newSvc)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn clients.ServingClient.Services.Patch(svc.ObjectMeta.Name, types.JSONPatchType, patchBytes, \"\")\n}", "func (d *namespaceHandlerImpl) DeprecateNamespace(\n\tctx context.Context,\n\tdeprecateRequest *workflowservice.DeprecateNamespaceRequest,\n) (*workflowservice.DeprecateNamespaceResponse, error) {\n\n\tclusterMetadata := d.clusterMetadata\n\t// TODO remove the IsGlobalNamespaceEnabled check once cross DC is public\n\tif clusterMetadata.IsGlobalNamespaceEnabled() && !clusterMetadata.IsMasterCluster() {\n\t\treturn nil, errNotMasterCluster\n\t}\n\n\t// must get the metadata (notificationVersion) first\n\t// this version can be regarded as the lock on the v2 namespace table\n\t// and since we do not know which table will return the namespace afterwards\n\t// this call has to be made\n\tmetadata, err := d.metadataMgr.GetMetadata(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tnotificationVersion := metadata.NotificationVersion\n\tgetResponse, err := d.metadataMgr.GetNamespace(ctx, &persistence.GetNamespaceRequest{Name: deprecateRequest.GetNamespace()})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tgetResponse.Namespace.ConfigVersion = getResponse.Namespace.ConfigVersion + 1\n\tgetResponse.Namespace.Info.State = enumspb.NAMESPACE_STATE_DEPRECATED\n\tupdateReq := &persistence.UpdateNamespaceRequest{\n\t\tNamespace: &persistencespb.NamespaceDetail{\n\t\t\tInfo: getResponse.Namespace.Info,\n\t\t\tConfig: getResponse.Namespace.Config,\n\t\t\tReplicationConfig: getResponse.Namespace.ReplicationConfig,\n\t\t\tConfigVersion: getResponse.Namespace.ConfigVersion,\n\t\t\tFailoverVersion: getResponse.Namespace.FailoverVersion,\n\t\t\tFailoverNotificationVersion: getResponse.Namespace.FailoverNotificationVersion,\n\t\t},\n\t\tNotificationVersion: notificationVersion,\n\t\tIsGlobalNamespace: getResponse.IsGlobalNamespace,\n\t}\n\terr = d.metadataMgr.UpdateNamespace(ctx, updateReq)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn nil, nil\n}", "func (d *Desc) IsDeprecated() bool {\n\treturn d.isDeprecated\n}", "func Deprecated(handler http.Handler, options DeprecationOptions) http.Handler {\n\tif options.WarnMessage == \"\" {\n\t\toptions.WarnMessage = DefaultDeprecationOptions.WarnMessage\n\t}\n\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tw.Header().Set(\"X-API-Warn\", options.WarnMessage)\n\n\t\tif !options.DeprecationDate.IsZero() {\n\t\t\tw.Header().Set(\"X-API-Deprecation-Date\", options.DeprecationDate.Format(HeaderTimeFormat))\n\t\t}\n\n\t\tif options.DeprecationInfo != \"\" {\n\t\t\tw.Header().Set(\"X-API-Deprecation-Info\", options.DeprecationInfo)\n\t\t}\n\n\t\thandler.ServeHTTP(w, r)\n\t})\n}", "func (i *LibpodAPI) RemoveImage(call ioprojectatomicpodman.VarlinkCall, name string, force bool) error {\n\truntime, err := libpodruntime.GetRuntime(i.Cli)\n\tif err != nil {\n\t\treturn call.ReplyRuntimeError(err.Error())\n\t}\n\tnewImage, err := runtime.ImageRuntime().NewFromLocal(name)\n\tif err != nil {\n\t\treturn call.ReplyImageNotFound(name)\n\t}\n\timageID, err := runtime.RemoveImage(newImage, force)\n\tif err != nil {\n\t\treturn call.ReplyErrorOccurred(err.Error())\n\t}\n\treturn call.ReplyRemoveImage(imageID)\n}", "func (r *Factory) NewDeprecatedLogger(name string) DeprecatedLogger {\n\tr.Lock()\n\tdefer r.Unlock()\n\tinfo := r.getLoggerInfo(name)\n\treturn &logger{\n\t\tatomicLogger: &info.atomicLogger,\n\t}\n}", "func (c MockDockerClient) ImageInspect(ctx context.Context, imageName string) (dockertypes.ImageInspect, error) {\n\tif c.ImageInspectFn != nil {\n\t\tfmt.Println(\"[MockDockerClient] In \", utils.CurrentFunctionName())\n\t\tfmt.Println(\"[MockDockerClient] - ctx: \", ctx)\n\t\tfmt.Println(\"[MockDockerClient] - imageName: \", imageName)\n\t\treturn c.ImageInspectFn(ctx, imageName)\n\t}\n\tpanic(fmt.Sprintf(\"No function defined for: %s\", utils.CurrentFunctionName()))\n}", "func TestRemoveObsoleteDesignDocsErrors(t *testing.T) {\n\n\tDesignDocPreviousVersions = []string{\"test\"}\n\n\tleakyBucketConfig := base.LeakyBucketConfig{\n\t\tDDocGetErrorCount: 1,\n\t\tDDocDeleteErrorCount: 1,\n\t}\n\n\tbucket := base.NewLeakyBucket(base.GetTestBucket(t), leakyBucketConfig)\n\tdefer bucket.Close()\n\n\tmapFunction := `function (doc, meta){ emit(); }`\n\n\terr := bucket.PutDDoc(DesignDocSyncGatewayPrefix+\"_test\", &sgbucket.DesignDoc{\n\t\tViews: sgbucket.ViewMap{\n\t\t\t\"channels\": sgbucket.ViewDef{Map: mapFunction},\n\t\t},\n\t})\n\tassert.NoError(t, err)\n\terr = bucket.PutDDoc(DesignDocSyncHousekeepingPrefix+\"_test\", &sgbucket.DesignDoc{\n\t\tViews: sgbucket.ViewMap{\n\t\t\t\"channels\": sgbucket.ViewDef{Map: mapFunction},\n\t\t},\n\t})\n\n\tremovedDDocsPreview, _ := removeObsoleteDesignDocs(bucket, true, false)\n\tremovedDDocsNonPreview, _ := removeObsoleteDesignDocs(bucket, false, false)\n\n\tassert.Equal(t, removedDDocsPreview, removedDDocsNonPreview)\n\n}", "func ImageStreamReplenishmentUpdateFunc(options *kresourcequota.ReplenishmentControllerOptions) func(oldObj, newObj interface{}) {\n\treturn func(oldObj, newObj interface{}) {\n\t\toldIS := oldObj.(*imageapi.ImageStream)\n\t\tnewIS := newObj.(*imageapi.ImageStream)\n\t\tif !reflect.DeepEqual(oldIS.Status.Tags, newIS.Status.Tags) {\n\t\t\toptions.ReplenishmentFunc(options.GroupKind, newIS.Namespace, newIS)\n\t\t}\n\t}\n}", "func Deprecated() {\n\tfmt.Println(deprecated)\n}", "func (v *IBM) DeleteImage(ctx *lepton.Context, snapshotID string) error {\n\treturn nil\n}", "func diffImage(c *Client, desired, actual *Image, opts ...dcl.ApplyOption) ([]imageDiff, error) {\n\tif desired == nil || actual == nil {\n\t\treturn nil, fmt.Errorf(\"nil resource passed to diff - always a programming error: %#v, %#v\", desired, actual)\n\t}\n\n\tvar diffs []imageDiff\n\tif !dcl.IsZeroValue(desired.Description) && !dcl.StringCanonicalize(desired.Description, actual.Description) {\n\t\tc.Config.Logger.Infof(\"Detected diff in Description.\\nDESIRED: %v\\nACTUAL: %v\", desired.Description, actual.Description)\n\t\tdiffs = append(diffs, imageDiff{\n\t\t\tRequiresRecreate: true,\n\t\t\tFieldName: \"Description\",\n\t\t})\n\t}\n\tif !reflect.DeepEqual(desired.DiskSizeGb, actual.DiskSizeGb) {\n\t\tc.Config.Logger.Infof(\"Detected diff in DiskSizeGb.\\nDESIRED: %v\\nACTUAL: %v\", desired.DiskSizeGb, actual.DiskSizeGb)\n\t\tdiffs = append(diffs, imageDiff{\n\t\t\tRequiresRecreate: true,\n\t\t\tFieldName: \"DiskSizeGb\",\n\t\t})\n\t}\n\tif !dcl.IsZeroValue(desired.Family) && !dcl.StringCanonicalize(desired.Family, actual.Family) {\n\t\tc.Config.Logger.Infof(\"Detected diff in Family.\\nDESIRED: %v\\nACTUAL: %v\", desired.Family, actual.Family)\n\t\tdiffs = append(diffs, imageDiff{\n\t\t\tRequiresRecreate: true,\n\t\t\tFieldName: \"Family\",\n\t\t})\n\t}\n\tif compareImageGuestOsFeatureSlice(c, desired.GuestOsFeature, actual.GuestOsFeature) {\n\t\tc.Config.Logger.Infof(\"Detected diff in GuestOsFeature.\\nDESIRED: %v\\nACTUAL: %v\", desired.GuestOsFeature, actual.GuestOsFeature)\n\t\ttoAdd, toRemove := compareImageGuestOsFeatureSets(c, desired.GuestOsFeature, actual.GuestOsFeature)\n\t\tif len(toAdd) > 0 {\n\t\t\tdiffs = append(diffs, imageDiff{\n\t\t\t\tRequiresRecreate: true,\n\t\t\t\tFieldName: \"GuestOsFeature\",\n\t\t\t})\n\t\t}\n\t\tif len(toRemove) > 0 {\n\t\t\tdiffs = append(diffs, imageDiff{\n\t\t\t\tRequiresRecreate: true,\n\t\t\t\tFieldName: \"GuestOsFeature\",\n\t\t\t})\n\t\t}\n\t}\n\tif compareImageImageEncryptionKey(c, desired.ImageEncryptionKey, actual.ImageEncryptionKey) {\n\t\tc.Config.Logger.Infof(\"Detected diff in ImageEncryptionKey.\\nDESIRED: %v\\nACTUAL: %v\", desired.ImageEncryptionKey, actual.ImageEncryptionKey)\n\t\tdiffs = append(diffs, imageDiff{\n\t\t\tRequiresRecreate: true,\n\t\t\tFieldName: \"ImageEncryptionKey\",\n\t\t})\n\t}\n\tif !dcl.MapEquals(desired.Labels, actual.Labels, []string(nil)) {\n\t\tc.Config.Logger.Infof(\"Detected diff in Labels.\\nDESIRED: %v\\nACTUAL: %v\", desired.Labels, actual.Labels)\n\n\t\tdiffs = append(diffs, imageDiff{\n\t\t\tUpdateOp: &updateImageSetLabelsOperation{},\n\t\t\tFieldName: \"Labels\",\n\t\t})\n\n\t}\n\tif !dcl.StringSliceEquals(desired.License, actual.License) {\n\t\tc.Config.Logger.Infof(\"Detected diff in License.\\nDESIRED: %v\\nACTUAL: %v\", desired.License, actual.License)\n\t\tdiffs = append(diffs, imageDiff{\n\t\t\tRequiresRecreate: true,\n\t\t\tFieldName: \"License\",\n\t\t})\n\t}\n\tif !dcl.IsZeroValue(desired.Name) && !dcl.StringCanonicalize(desired.Name, actual.Name) {\n\t\tc.Config.Logger.Infof(\"Detected diff in Name.\\nDESIRED: %v\\nACTUAL: %v\", desired.Name, actual.Name)\n\t\tdiffs = append(diffs, imageDiff{\n\t\t\tRequiresRecreate: true,\n\t\t\tFieldName: \"Name\",\n\t\t})\n\t}\n\tif compareImageShieldedInstanceInitialState(c, desired.ShieldedInstanceInitialState, actual.ShieldedInstanceInitialState) {\n\t\tc.Config.Logger.Infof(\"Detected diff in ShieldedInstanceInitialState.\\nDESIRED: %v\\nACTUAL: %v\", desired.ShieldedInstanceInitialState, actual.ShieldedInstanceInitialState)\n\t\tdiffs = append(diffs, imageDiff{\n\t\t\tRequiresRecreate: true,\n\t\t\tFieldName: \"ShieldedInstanceInitialState\",\n\t\t})\n\t}\n\tif !dcl.IsZeroValue(desired.SourceDisk) && !dcl.StringCanonicalize(desired.SourceDisk, actual.SourceDisk) {\n\t\tc.Config.Logger.Infof(\"Detected diff in SourceDisk.\\nDESIRED: %v\\nACTUAL: %v\", desired.SourceDisk, actual.SourceDisk)\n\t\tdiffs = append(diffs, imageDiff{\n\t\t\tRequiresRecreate: true,\n\t\t\tFieldName: \"SourceDisk\",\n\t\t})\n\t}\n\tif compareImageSourceDiskEncryptionKey(c, desired.SourceDiskEncryptionKey, actual.SourceDiskEncryptionKey) {\n\t\tc.Config.Logger.Infof(\"Detected diff in SourceDiskEncryptionKey.\\nDESIRED: %v\\nACTUAL: %v\", desired.SourceDiskEncryptionKey, actual.SourceDiskEncryptionKey)\n\t\tdiffs = append(diffs, imageDiff{\n\t\t\tRequiresRecreate: true,\n\t\t\tFieldName: \"SourceDiskEncryptionKey\",\n\t\t})\n\t}\n\tif !dcl.IsZeroValue(desired.SourceImage) && !dcl.StringCanonicalize(desired.SourceImage, actual.SourceImage) {\n\t\tc.Config.Logger.Infof(\"Detected diff in SourceImage.\\nDESIRED: %v\\nACTUAL: %v\", desired.SourceImage, actual.SourceImage)\n\t\tdiffs = append(diffs, imageDiff{\n\t\t\tRequiresRecreate: true,\n\t\t\tFieldName: \"SourceImage\",\n\t\t})\n\t}\n\tif compareImageSourceImageEncryptionKey(c, desired.SourceImageEncryptionKey, actual.SourceImageEncryptionKey) {\n\t\tc.Config.Logger.Infof(\"Detected diff in SourceImageEncryptionKey.\\nDESIRED: %v\\nACTUAL: %v\", desired.SourceImageEncryptionKey, actual.SourceImageEncryptionKey)\n\t\tdiffs = append(diffs, imageDiff{\n\t\t\tRequiresRecreate: true,\n\t\t\tFieldName: \"SourceImageEncryptionKey\",\n\t\t})\n\t}\n\tif !dcl.IsZeroValue(desired.SourceImageId) && !dcl.StringCanonicalize(desired.SourceImageId, actual.SourceImageId) {\n\t\tc.Config.Logger.Infof(\"Detected diff in SourceImageId.\\nDESIRED: %v\\nACTUAL: %v\", desired.SourceImageId, actual.SourceImageId)\n\t\tdiffs = append(diffs, imageDiff{\n\t\t\tRequiresRecreate: true,\n\t\t\tFieldName: \"SourceImageId\",\n\t\t})\n\t}\n\tif !dcl.IsZeroValue(desired.SourceSnapshot) && !dcl.StringCanonicalize(desired.SourceSnapshot, actual.SourceSnapshot) {\n\t\tc.Config.Logger.Infof(\"Detected diff in SourceSnapshot.\\nDESIRED: %v\\nACTUAL: %v\", desired.SourceSnapshot, actual.SourceSnapshot)\n\t\tdiffs = append(diffs, imageDiff{\n\t\t\tRequiresRecreate: true,\n\t\t\tFieldName: \"SourceSnapshot\",\n\t\t})\n\t}\n\tif compareImageSourceSnapshotEncryptionKey(c, desired.SourceSnapshotEncryptionKey, actual.SourceSnapshotEncryptionKey) {\n\t\tc.Config.Logger.Infof(\"Detected diff in SourceSnapshotEncryptionKey.\\nDESIRED: %v\\nACTUAL: %v\", desired.SourceSnapshotEncryptionKey, actual.SourceSnapshotEncryptionKey)\n\t\tdiffs = append(diffs, imageDiff{\n\t\t\tRequiresRecreate: true,\n\t\t\tFieldName: \"SourceSnapshotEncryptionKey\",\n\t\t})\n\t}\n\tif !dcl.IsZeroValue(desired.SourceSnapshotId) && !dcl.StringCanonicalize(desired.SourceSnapshotId, actual.SourceSnapshotId) {\n\t\tc.Config.Logger.Infof(\"Detected diff in SourceSnapshotId.\\nDESIRED: %v\\nACTUAL: %v\", desired.SourceSnapshotId, actual.SourceSnapshotId)\n\t\tdiffs = append(diffs, imageDiff{\n\t\t\tRequiresRecreate: true,\n\t\t\tFieldName: \"SourceSnapshotId\",\n\t\t})\n\t}\n\tif !reflect.DeepEqual(desired.SourceType, actual.SourceType) {\n\t\tc.Config.Logger.Infof(\"Detected diff in SourceType.\\nDESIRED: %v\\nACTUAL: %v\", desired.SourceType, actual.SourceType)\n\t\tdiffs = append(diffs, imageDiff{\n\t\t\tRequiresRecreate: true,\n\t\t\tFieldName: \"SourceType\",\n\t\t})\n\t}\n\tif !dcl.StringSliceEquals(desired.StorageLocation, actual.StorageLocation) {\n\t\tc.Config.Logger.Infof(\"Detected diff in StorageLocation.\\nDESIRED: %v\\nACTUAL: %v\", desired.StorageLocation, actual.StorageLocation)\n\t\tdiffs = append(diffs, imageDiff{\n\t\t\tRequiresRecreate: true,\n\t\t\tFieldName: \"StorageLocation\",\n\t\t})\n\t}\n\tif compareImageDeprecated(c, desired.Deprecated, actual.Deprecated) {\n\t\tc.Config.Logger.Infof(\"Detected diff in Deprecated.\\nDESIRED: %v\\nACTUAL: %v\", desired.Deprecated, actual.Deprecated)\n\n\t\tdiffs = append(diffs, imageDiff{\n\t\t\tUpdateOp: &updateImageDeprecateOperation{},\n\t\t\tFieldName: \"Deprecated\",\n\t\t})\n\n\t}\n\t// We need to ensure that this list does not contain identical operations *most of the time*.\n\t// There may be some cases where we will need multiple copies of the same operation - for instance,\n\t// if a resource has multiple prerequisite-containing fields. For now, we don't know of any\n\t// such examples and so we deduplicate unconditionally.\n\n\t// The best way for us to do this is to iterate through the list\n\t// and remove any copies of operations which are identical to a previous operation.\n\t// This is O(n^2) in the number of operations, but n will always be very small,\n\t// even 10 would be an extremely high number.\n\tvar opTypes []string\n\tvar deduped []imageDiff\n\tfor _, d := range diffs {\n\t\t// Two operations are considered identical if they have the same type.\n\t\t// The type of an operation is derived from the name of the update method.\n\t\tif !dcl.StringSliceContains(fmt.Sprintf(\"%T\", d.UpdateOp), opTypes) {\n\t\t\tdeduped = append(deduped, d)\n\t\t\topTypes = append(opTypes, fmt.Sprintf(\"%T\", d.UpdateOp))\n\t\t} else {\n\t\t\tc.Config.Logger.Infof(\"Omitting planned operation of type %T since once is already scheduled.\", d.UpdateOp)\n\t\t}\n\t}\n\n\treturn deduped, nil\n}", "func (b *OperationMutator) Deprecated(v bool) *OperationMutator {\n\tb.proxy.deprecated = v\n\treturn b\n}", "func (o DocumentationRuleResponseOutput) DeprecationDescription() pulumi.StringOutput {\n\treturn o.ApplyT(func(v DocumentationRuleResponse) string { return v.DeprecationDescription }).(pulumi.StringOutput)\n}", "func (_DelegateProfile *DelegateProfileTransactorSession) DeprecateField(_name string) (*types.Transaction, error) {\n\treturn _DelegateProfile.Contract.DeprecateField(&_DelegateProfile.TransactOpts, _name)\n}", "func (i *ImageService) ExportImage(names []string, outStream io.Writer) error {\n\tpanic(\"not implemented\")\n}", "func (s imageRepositoryStrategy) Decorate(obj runtime.Object) error {\n\tir := obj.(*api.ImageRepository)\n\tir.Status.DockerImageRepository = s.dockerImageRepository(ir)\n\treturn nil\n}", "func (g *Group) Deprecated(options DeprecationOptions) *Group {\n\t// store it for future use, e.g. collect all deprecated APIs and notify the developer.\n\tg.deprecation = options\n\n\tg.API.UseOnce(func(ctx *context.Context) {\n\t\tWriteDeprecated(ctx, options)\n\t\tctx.Next()\n\t})\n\treturn g\n}", "func (c MockDockerClient) ImagePull(ctx context.Context, imageName string) error {\n\tif c.ImagePullFn != nil {\n\t\tfmt.Println(\"[MockDockerClient] In \", utils.CurrentFunctionName())\n\t\tfmt.Println(\"[MockDockerClient] - ctx: \", ctx)\n\t\tfmt.Println(\"[MockDockerClient] - imageName: \", imageName)\n\t\treturn c.ImagePullFn(ctx, imageName)\n\t}\n\tpanic(fmt.Sprintf(\"No function defined for: %s\", utils.CurrentFunctionName()))\n}", "func newImage(vres *artworksviews.ImageView) *Image {\n\tres := &Image{}\n\tif vres.ImageID != nil {\n\t\tres.ImageID = *vres.ImageID\n\t}\n\tif vres.ExpiresIn != nil {\n\t\tres.ExpiresIn = *vres.ExpiresIn\n\t}\n\treturn res\n}", "func Deprecated(w io.Writer, format string, args ...any) {\n\tformat = strings.TrimRight(format, \"\\r\\n\") + \"\\n\"\n\tfmt.Fprintf(w, \"\\n\"+Wrap(BoldRed(\"DEPRECATED: \")+format, DefaultTextWidth)+\"\\n\", args...)\n}", "func (d *DeprecatedProcessor) Process(field protoreflect.FieldDescriptor, msg protoreflect.Message) (data ResultData, applied bool) {\n\tif d.ClearField {\n\t\tmsg.Clear(field)\n\t}\n\treturn ResultData{Message: \"deprecated\"}, true\n}", "func registerDeprecated(global *flag.FlagSet, local *pflag.FlagSet, globalName, deprecated string) {\n\tregister(global, local, globalName)\n\tlocal.Lookup(normalize(globalName)).Deprecated = deprecated\n}", "func newImageWithoutInit(width, height int, volatile bool) *Image {\n\ti := &Image{\n\t\timage: graphics.NewImage(width, height),\n\t\tvolatile: volatile,\n\t}\n\ttheImages.add(i)\n\treturn i\n}", "func (*UpdateImageMetadata) Descriptor() ([]byte, []int) {\n\treturn file_yandex_cloud_compute_v1_image_service_proto_rawDescGZIP(), []int{7}\n}", "func (d *ConnectionDetails) ReplaceWithInternalRegistry(imageURL string) (string, error) {\n\tprivateURL, err := d.PrivateRegistryURL()\n\tif err != nil {\n\t\treturn imageURL, err\n\t}\n\tif privateURL == \"\" {\n\t\treturn imageURL, nil // no-op\n\t}\n\n\tpublicURL, err := d.PublicRegistryURL()\n\tif err != nil {\n\t\treturn imageURL, err\n\t}\n\n\timageRegistryURL, _, err := ExtractImageParts(imageURL)\n\tif err != nil {\n\t\treturn imageURL, err\n\t}\n\n\tif imageRegistryURL == publicURL {\n\t\treturn strings.Replace(imageURL, imageRegistryURL, privateURL, -1), nil\n\t}\n\n\treturn imageURL, nil\n}", "func (sc *BaseClient) AddJobToDeprecate(jobid string) {\n\tsc.deprecationChannels[jobid] = make(chan bool)\n}", "func (montage *Montage) UpdateImage(source string, payload []byte) {\n\tlogrus.Debugf(\"montage.UpdateImage: source=%s for montage %s\", source, montage.Config.Name)\n\n\tfor _, placement := range montage.actualPlacement {\n\t\t// Finish the loop.\n\t\t// It is possible that multiple layouts use the same source, for example one image is zoomed in.\n\t\tif placement.Source == source {\n\t\t\tif montage.useLibJpeg {\n\t\t\t\t_ = montage.DrawImageIntoLayout(&placement, payload)\n\t\t\t} else {\n\t\t\t\t_ = montage.DrawImageIntoLayout(&placement, payload)\n\t\t\t}\n\t\t}\n\t}\n}", "func uxImage(cmd cli.Command) cli.Command {\n\tcmd.Flags = append(cmd.Flags, cli.StringFlag{\n\t\tName: \"image\",\n\t\tUsage: \"OCI image URI of the form 'path[:tag]'\",\n\t})\n\n\toldBefore := cmd.Before\n\tcmd.Before = func(ctx *cli.Context) error {\n\t\t// Verify and parse --image.\n\t\tif ctx.IsSet(\"image\") {\n\t\t\timage := ctx.String(\"image\")\n\n\t\t\tvar dir, tag string\n\t\t\tsep := strings.Index(image, \":\")\n\t\t\tif sep == -1 {\n\t\t\t\tdir = image\n\t\t\t\ttag = \"latest\"\n\t\t\t} else {\n\t\t\t\tdir = image[:sep]\n\t\t\t\ttag = image[sep+1:]\n\t\t\t}\n\n\t\t\t// Verify directory value.\n\t\t\tif dir == \"\" {\n\t\t\t\treturn errors.Wrap(fmt.Errorf(\"path is empty\"), \"invalid --image\")\n\t\t\t}\n\n\t\t\t// Verify tag value.\n\t\t\tif !casext.IsValidReferenceName(tag) {\n\t\t\t\treturn errors.Wrap(fmt.Errorf(\"tag contains invalid characters: '%s'\", tag), \"invalid --image\")\n\t\t\t}\n\t\t\tif tag == \"\" {\n\t\t\t\treturn errors.Wrap(fmt.Errorf(\"tag is empty\"), \"invalid --image\")\n\t\t\t}\n\n\t\t\tctx.App.Metadata[\"--image-path\"] = dir\n\t\t\tctx.App.Metadata[\"--image-tag\"] = tag\n\t\t}\n\n\t\tif oldBefore != nil {\n\t\t\treturn oldBefore(ctx)\n\t\t}\n\t\treturn nil\n\t}\n\n\treturn cmd\n}", "func newReplaceDecorator(value interface{}) (interface{}, reflect.Type) {\n\tswitch value.(type) {\n\tcase nil:\n\t\tpanic(\"untyped nil passed to fx.Replace\")\n\tcase error:\n\t\tpanic(\"error value passed to fx.Replace\")\n\t}\n\n\ttyp := reflect.TypeOf(value)\n\treturnTypes := []reflect.Type{typ}\n\treturnValues := []reflect.Value{reflect.ValueOf(value)}\n\n\tft := reflect.FuncOf([]reflect.Type{}, returnTypes, false)\n\tfv := reflect.MakeFunc(ft, func([]reflect.Value) []reflect.Value {\n\t\treturn returnValues\n\t})\n\n\treturn fv.Interface(), typ\n}", "func NewDeleteImageDefault(code int) *DeleteImageDefault {\n\tif code <= 0 {\n\t\tcode = 500\n\t}\n\n\treturn &DeleteImageDefault{\n\t\t_statusCode: code,\n\t}\n}", "func (*DeleteImageRequest) Descriptor() ([]byte, []int) {\n\treturn file_yandex_cloud_compute_v1_image_service_proto_rawDescGZIP(), []int{8}\n}", "func OverwriteImage(buff draw.Image, img image.Image, x, y int) {\n\tdraw.Draw(buff, buff.Bounds(),\n\t\timg, image.Point{-x, -y}, draw.Src)\n}", "func ImageService() {\n\t//deleteThumbnailId(2);\n\t//return;\n\n/*\tif flags.mode == \"fetchNewsSourceIcons\" {\n\t\tfor newsSource, imageUrl := range newsSourceIcons {\n\t\t\tcheck(downsampleImage(imageUrl, \"newsSourceIcons\", newsSource, \"png\", 16, 16))\n\t\t}\n\t\treturn\n\t}*/\n\tfetchImagesToDownsampleQuery := [NUM_GEN_THUMBS_PASSES]string {\n\t\t`SELECT Id, UrlToImage, ThumbnailStatus\n\t\t FROM $$LinkPost\n\t\t WHERE 0 <= ThumbnailStatus AND ThumbnailStatus < %d\n\t\t AND UrlToImage <> ''\n\t\t ORDER BY Created DESC\n\t\t LIMIT %d;`,\n\n\t\t`SELECT Id, UrlToImage, ThumbnailStatus\n\t\t FROM $$NewsPost\n\t\t WHERE 0 <= ThumbnailStatus AND ThumbnailStatus < %d\n\t\t AND UrlToImage <> ''\n\t\t AND Created > now() - interval '2 weeks'\n\t\t ORDER BY COALESCE(PublishedAt, Created) DESC\n\t\t LIMIT %d;`,\n\n\t\t`SELECT Id, UrlToImage, ThumbnailStatus\n\t\t FROM $$PollPost\n\t\t WHERE 0 <= ThumbnailStatus AND ThumbnailStatus < %d\n\t\t AND UrlToImage <> ''\n\t\t ORDER BY Created DESC\n\t\t LIMIT %d;`,\n\t}\n\tfor i := 0; i < NUM_GEN_THUMBS_PASSES; i++ {\n\t\tfetchImagesToDownsampleQuery[i] = fmt.Sprintf(\n\t\t\tfetchImagesToDownsampleQuery[i],\n\t\t\timage_DownsampleVersionTarget,\n\t\t\tkImageBatchSize)\n\t}\n\tprVal(\"fetchImagesToDownsampleQuery[0]\", fetchImagesToDownsampleQuery[0])\n\tprVal(\"fetchImagesToDownsampleQuery[1]\", fetchImagesToDownsampleQuery[1])\n\tprVal(\"fetchImagesToDownsampleQuery[2]\", fetchImagesToDownsampleQuery[2])\n\n\n\tpr(\"========================================\")\n\tpr(\"======== STARTING IMAGE SERVICE ========\")\n\tpr(\"========================================\\n\")\n\n\tfor { // Infinite loop\n\n\t\t// \tDelete news thumbnails more than 2 weeks old, so we don't run out of hard disk space.\n\t\tpr(\"Deleting old news thumbnails pass\");\n\t\tfor {\n\t\t\tpr(\"Next Image Deletion Loop\");\n\n\t\t\tnumImagesDeleted := 0\n\n\t\t\tDoQuery(\n\t\t\t\tfunc(rows *sql.Rows) {\n\t\t\t\t\tvar id int\n\n\t\t\t\t\terr := rows.Scan(&id)\n\t\t\t\t\tcheck(err)\n\n\t\t\t\t\tprVal(\"id\", id);\n\n\t\t\t\t\tdeleteThumbnailId(id);\n\n\t\t\t\t\tDbExec(\"UPDATE $$NewsPost SET ThumbnailStatus=0 WHERE Id = $1::bigint\", id)\n\n\t\t\t\t\tnumImagesDeleted++\n\n\t\t\t\t},\n\t\t\t\t`SELECT Id FROM $$NewsPost\n\t\t\t\t WHERE ThumbnailStatus > 0\n\t\t\t\t AND UrlToImage <> ''\n\t\t\t\t AND COALESCE(PublishedAt, Created) <= now() - interval '2 weeks'\n\t\t\t\t ORDER BY COALESCE(PublishedAt, Created)\n\t\t\t\t LIMIT 100000;`,\n\t\t\t)\n\n\t\t\tprVal(\"numImagesDeleted\", numImagesDeleted);\n\n\t\t\tif numImagesDeleted == 0 {\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\n\t\tnumImageProcessAttempts := 0\n\n\t\t// Downsample news images\n\t\tfor pass := 0; pass < NUM_GEN_THUMBS_PASSES; pass++ {\n\t\t\tpr(\"========================================\")\n\t\t\tprf(\"======= FETCHING IMAGES PASS: %d =======\", pass)\n\t\t\tpr(\"========================================\\n\")\n\n\n\t\t\t// Grab a batch of images to downsample from new news posts.\n\t\t\tids2urls := fetchPostIds2Urls(fetchImagesToDownsampleQuery[pass])\n\t\t\tprVal(\"len(ids2urls)\", len(ids2urls))\n\n\t\t\tif len(ids2urls) == 0 {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// Download and downsample the images in parallel.\n\t\t\tc := make(chan DownsampleResult)\n\t\t\ttimeout := time.After(30 * time.Second)\n\n\t\t\tfor id, urlStatus := range ids2urls {\n\t\t\t\tnumImageProcessAttempts++\n\n\t\t\t\tprf(\"trying to create channel to downsample id %d url %s status %d\", id, urlStatus.url, urlStatus.status)\n\t\t\t\tgo downsamplePostImage(urlStatus.url, urlStatus.status, id, c)\n\t\t\t}\n\n\t\t\t// TODO: Generalize this code. Can use fn callbacks for the main and timeout cases.\n\t\t\tdownsampleImagesLoop: for {\n\t\t\t\tselect {\n\t\t\t\t\tcase downsampleResult := <-c:\n\t\t\t\t\t\tnewThumbnailStatus := ternary_int(\n\t\t\t\t\t\t\tdownsampleResult.err == nil,\n\t\t\t\t\t\t\timage_DownsampleVersionTarget,\n\t\t\t\t\t\t\timage_DownsampleError)\n\n\t\t\t\t\t\tprVal(\"downsampleResult\", downsampleResult)\n\t\t\t\t\t\tprVal(\" newThumbnailStatus\", newThumbnailStatus)\n\n\t\t\t\t\t\tswitch pass {\n\t\t\t\t\t\t\tcase genThumbPass_LinkPost:\n\t\t\t\t\t\t\t\tDbExec(\n\t\t\t\t\t\t\t\t\t`UPDATE $$LinkPost\n\t\t\t\t\t\t\t\t\t SET ThumbnailStatus = $1\n\t\t\t\t\t\t\t\t\t WHERE Id = $2::bigint`,\n\t\t\t\t\t\t\t\t\tnewThumbnailStatus,\n\t\t\t\t\t\t\t\t\tdownsampleResult.postId)\n\t\t\t\t\t\t\tcase genThumbPass_NewsPost:\n\t\t\t\t\t\t\t\tDbExec(\n\t\t\t\t\t\t\t\t\t`UPDATE $$NewsPost\n\t\t\t\t\t\t\t\t\t SET ThumbnailStatus = $1\n\t\t\t\t\t\t\t\t\t WHERE Id = $2::bigint`,\n\t\t\t\t\t\t\t\t\tnewThumbnailStatus,\n\t\t\t\t\t\t\t\t\tdownsampleResult.postId)\n\t\t\t\t\t\t\tcase genThumbPass_PollPost:\n\t\t\t\t\t\t\t\tDbExec(\n\t\t\t\t\t\t\t\t\t`UPDATE $$PollPost\n\t\t\t\t\t\t\t\t\t SET ThumbnailStatus = $1\n\t\t\t\t\t\t\t\t\t WHERE Id = $2::bigint`,\n\t\t\t\t\t\t\t\t\tnewThumbnailStatus,\n\t\t\t\t\t\t\t\t\tdownsampleResult.postId)\n\t\t\t\t\t\t\tdefault:\n\t\t\t\t\t\t\t\tassert(false)\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\t// Remove this from the list of ids, so we can tell which ids were never processed.\n\t\t\t\t\t\tdelete(ids2urls, downsampleResult.postId)\n\n\t\t\t\t\t\tif len(ids2urls) == 0 {\n\t\t\t\t\t\t\tpr(\"Processed all images!\")\n\t\t\t\t\t\t\tbreak downsampleImagesLoop\n\t\t\t\t\t\t}\n\t\t\t\t\tcase <- timeout:\n\t\t\t\t\t\tpr(\"Timeout!\")\n\n\t\t\t\t\t\t// Set status to -1 for any images that timed out.\n\t\t\t\t\t\tfor id, urlStatus := range ids2urls {\n\t\t\t\t\t\t\tprf(\"Removing timed out id %d url %s prevStatus %d\", id, urlStatus.url, urlStatus.status)\n\n\t\t\t\t\t\t\tswitch pass {\n\t\t\t\t\t\t\t\tcase genThumbPass_LinkPost:\n\t\t\t\t\t\t\t\t\tDbExec(\n\t\t\t\t\t\t\t\t\t\t`UPDATE $$LinkPost\n\t\t\t\t\t\t\t\t\t\t SET ThumbnailStatus = -1\n\t\t\t\t\t\t\t\t\t\t WHERE Id = $1::bigint`,\n\t\t\t\t\t\t\t\t\t\tid)\n\t\t\t\t\t\t\t\tcase genThumbPass_NewsPost:\n\t\t\t\t\t\t\t\t\tDbExec(\n\t\t\t\t\t\t\t\t\t\t`UPDATE $$NewsPost\n\t\t\t\t\t\t\t\t\t\t SET ThumbnailStatus = -1\n\t\t\t\t\t\t\t\t\t\t WHERE Id = $1::bigint`,\n\t\t\t\t\t\t\t\t\t\tid)\n\t\t\t\t\t\t\t\tcase genThumbPass_PollPost:\n\t\t\t\t\t\t\t\t\tDbExec(\n\t\t\t\t\t\t\t\t\t\t`UPDATE $$PollPost\n\t\t\t\t\t\t\t\t\t\t SET ThumbnailStatus = -1\n\t\t\t\t\t\t\t\t\t\t WHERE Id = $1::bigint`,\n\t\t\t\t\t\t\t\t\t\tid)\n\t\t\t\t\t\t\t\tdefault:\n\t\t\t\t\t\t\t\t\tassert(false)\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\tbreak downsampleImagesLoop\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tDbTrackOpenConnections()\n\t\t}\n\n\t\t// Sleep when there are no records to process.\n\t\tif numImageProcessAttempts == 0 {\n\t\t\tpr(\"Sleep 10 seconds\")\n\t\t\ttime.Sleep(10 * time.Second)\n\t\t}\n\t}\n}", "func (c *CRIImageService) RemoveImage(ctx context.Context, r *runtime.RemoveImageRequest) (*runtime.RemoveImageResponse, error) {\n\tspan := tracing.SpanFromContext(ctx)\n\timage, err := c.LocalResolve(r.GetImage().GetImage())\n\tif err != nil {\n\t\tif errdefs.IsNotFound(err) {\n\t\t\tspan.AddEvent(err.Error())\n\t\t\t// return empty without error when image not found.\n\t\t\treturn &runtime.RemoveImageResponse{}, nil\n\t\t}\n\t\treturn nil, fmt.Errorf(\"can not resolve %q locally: %w\", r.GetImage().GetImage(), err)\n\t}\n\tspan.SetAttributes(tracing.Attribute(\"image.id\", image.ID))\n\t// Remove all image references.\n\tfor i, ref := range image.References {\n\t\tvar opts []images.DeleteOpt\n\t\tif i == len(image.References)-1 {\n\t\t\t// Delete the last image reference synchronously to trigger garbage collection.\n\t\t\t// This is best effort. It is possible that the image reference is deleted by\n\t\t\t// someone else before this point.\n\t\t\topts = []images.DeleteOpt{images.SynchronousDelete()}\n\t\t}\n\t\terr = c.client.ImageService().Delete(ctx, ref, opts...)\n\t\tif err == nil || errdefs.IsNotFound(err) {\n\t\t\t// Update image store to reflect the newest state in containerd.\n\t\t\tif err := c.imageStore.Update(ctx, ref); err != nil {\n\t\t\t\treturn nil, fmt.Errorf(\"failed to update image reference %q for %q: %w\", ref, image.ID, err)\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\t\treturn nil, fmt.Errorf(\"failed to delete image reference %q for %q: %w\", ref, image.ID, err)\n\t}\n\treturn &runtime.RemoveImageResponse{}, nil\n}", "func (d *DaemonSetPrepuller) DeleteFunc(component string) error {\n\tdsName := addPrepullPrefix(component)\n\tif err := apiclient.DeleteDaemonSetForeground(d.client, metav1.NamespaceSystem, dsName); err != nil {\n\t\treturn fmt.Errorf(\"unable to cleanup the DaemonSet used for prepulling %s: %v\", component, err)\n\t}\n\tfmt.Printf(\"[upgrade/prepull] Prepulled image for component %s.\\n\", component)\n\treturn nil\n}", "func (api *Api) RemoveImage(ctx *gin.Context) {\n\tcraneContext, _ := ctx.Get(\"craneContext\")\n\timageID := ctx.Param(\"image_id\")\n\tif err := api.GetDockerClient().RemoveImage(craneContext.(context.Context), imageID); err != nil {\n\t\tlog.Error(\"RemoveImage got error: \", err)\n\t\thttpresponse.Error(ctx, err)\n\t\treturn\n\t}\n\n\thttpresponse.Ok(ctx, \"success\")\n\treturn\n}", "func DeprecateFlags(fs *flag.FlagSet, names ...string) {\n\tfs.Visit(func(f *flag.Flag) {\n\t\tfor _, name := range names {\n\t\t\tif f.Name == name {\n\t\t\t\tlog.Warnf(\"the --%s flag is deprecated and will be removed in a future release\", name)\n\t\t\t}\n\t\t}\n\t})\n}", "func (_DelegateProfile *DelegateProfileSession) DeprecateField(_name string) (*types.Transaction, error) {\n\treturn _DelegateProfile.Contract.DeprecateField(&_DelegateProfile.TransactOpts, _name)\n}", "func (i *LibpodAPI) ExportImage(call ioprojectatomicpodman.VarlinkCall, name, destination string, compress bool) error {\n\truntime, err := libpodruntime.GetRuntime(i.Cli)\n\tif err != nil {\n\t\treturn call.ReplyRuntimeError(err.Error())\n\t}\n\tnewImage, err := runtime.ImageRuntime().NewFromLocal(name)\n\tif err != nil {\n\t\treturn call.ReplyImageNotFound(name)\n\t}\n\tif err := newImage.PushImage(getContext(), destination, \"\", \"\", \"\", nil, compress, image.SigningOptions{}, &image.DockerRegistryOptions{}, false); err != nil {\n\t\treturn call.ReplyErrorOccurred(err.Error())\n\t}\n\treturn call.ReplyExportImage(newImage.ID())\n}", "func (o DeprecationStatusOutput) Deprecated() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v DeprecationStatus) *string { return v.Deprecated }).(pulumi.StringPtrOutput)\n}", "func (DBConnection *MariaDBPlugin) ReplaceImageTags(OldTagID uint64, NewTagID uint64, LinkerID uint64) error {\n\tquery := `UPDATE ImageTags\n\tSET TagID = ? , LinkerID=?\n\tWHERE TagID=? AND ImageID NOT IN\n\t(\n\t\tSELECT ImageID from ImageTags WHERE TagID=?\n\t);`\n\t_, err := DBConnection.DBHandle.Exec(query, NewTagID, LinkerID, OldTagID, NewTagID)\n\tif err != nil {\n\t\tlogging.WriteLog(logging.LogLevelError, \"MariaDBPlugin/ReplaceImageTags\", strconv.FormatUint(LinkerID, 10), logging.ResultFailure, []string{\"Failed to update imagetags\", err.Error()})\n\t\treturn err\n\t}\n\t//Remove any instances of old tag, first query replaces the old tag on all images, but does not allow duplicates. This query will remove the old tag that would have been replaced if it would not have lead to a duplicate.\n\t_, err = DBConnection.DBHandle.Exec(\"DELETE FROM ImageTags WHERE TagID=?;\", OldTagID)\n\tif err != nil {\n\t\tlogging.WriteLog(logging.LogLevelError, \"MariaDBPlugin/ReplaceImageTags\", strconv.FormatUint(LinkerID, 10), logging.ResultFailure, []string{\"Failed to remove old instances of tag\", err.Error()})\n\t\treturn err\n\t}\n\treturn nil\n}", "func (i *imageHandler) resolveTag() string {\n\tif len(i.image.Tag) == 0 {\n\t\treturn infrastructure.GetRuntimeImageVersion()\n\t}\n\treturn i.image.Tag\n}", "func deleteImageResource(ctx context.Context, d *schema.ResourceData, meta interface{}) diag.Diagnostics {\n\t// Warning or errors can be collected in a slice type\n\tvar diags diag.Diagnostics\n\tclient := (meta.(Client)).Client\n\tname := rdEntryStr(d, \"name\")\n\tid := rdEntryStr(d, \"id\")\n\terrMsgPrefix := getErrMsgPrefix(\"Image\", name, id, \"Delete\")\n\tcfg, err := getImage(client, name, id)\n\tif err != nil {\n\t\treturn diag.Errorf(\"%s Failed to get Image. err: %s\", errMsgPrefix, err.Error())\n\t}\n\tif cfg == nil {\n\t\tlog.Printf(\"%s Unexpected Error. nil config\", errMsgPrefix)\n\t\treturn diags\n\t}\n\tclient.XRequestIdPrefix = \"TF-image-delete\"\n\turlExtension := getImageUrl(name, id, \"delete\")\n\trspData := &swagger_models.ZsrvResponse{}\n\t_, err = client.SendReq(\"DELETE\", urlExtension, nil, rspData)\n\tif err != nil {\n\t\treturn diag.Errorf(\"%s. Request Failed. err: %s\", errMsgPrefix, err.Error())\n\t}\n\tlog.Printf(\"[INFO] Image %s(id:%s) Delete Successful.\", name, cfg.ID)\n\treturn diags\n}", "func (i *imageHandler) resolveRegistryImage() string {\n\tdomain := i.image.Domain\n\tif len(domain) == 0 {\n\t\tdomain = defaultImageDomain\n\t}\n\tns := i.image.Namespace\n\tif len(ns) == 0 {\n\t\tns = defaultImageNamespace\n\t}\n\treturn fmt.Sprintf(\"%s/%s/%s\", domain, ns, i.resolveImageNameTag())\n}", "func (d *Dry) RemoveImage(position int, force bool) {\n\tif image, err := d.dockerDaemon.ImageAt(position); err == nil {\n\t\tid := drydocker.ImageID(image.ID)\n\t\tshortID := drydocker.TruncateID(id)\n\t\td.appmessage(fmt.Sprintf(\"<red>Removing image:</> <white>%s</>\", shortID))\n\t\tif _, err = d.dockerDaemon.Rmi(id, force); err == nil {\n\t\t\td.doRefresh()\n\t\t\td.appmessage(fmt.Sprintf(\"<red>Removed image:</> <white>%s</>\", shortID))\n\t\t} else {\n\t\t\td.appmessage(fmt.Sprintf(\"<red>Error removing image </><white>%s: %s</>\", shortID, err.Error()))\n\t\t}\n\t} else {\n\t\td.appmessage(fmt.Sprintf(\"<red>Error removing image</>: %s\", err.Error()))\n\t}\n}", "func ExportImage(ctx context.Contexter) (err error) {\n\toutputDir := ctx.Get(\"output\").(string)\n\tsources := ctx.Get(\"sources\").([]string)\n\n\tif len(sources) == 0 {\n\t\treturn fmt.Errorf(\"source file/directory of function required\")\n\t}\n\tif len(sources) == 1 &&\n\t\tutils.IsDir(sources[0]) &&\n\t\tutils.HasDockerfile(sources[0]) {\n\t\tif err := copy.Copy(sources[0], outputDir); err != nil {\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\tif err := packer.Pack(outputDir, sources...); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tlog.Infof(\"exported to %v: %v\", outputDir, constants.CheckedSymbol)\n\treturn nil\n}", "func CtrDeleteImage(reference string) error {\n\tif err := verifyCtr(); err != nil {\n\t\treturn fmt.Errorf(\"CtrDeleteImage: exception while verifying ctrd client: %s\", err.Error())\n\t}\n\treturn CtrdClient.ImageService().Delete(ctrdCtx, reference)\n}", "func decodeImg(data []byte, blockSize bool) *image.RGBA {\n\tsrc, err := png.Decode(bytes.NewBuffer(data))\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\t// Convert to image.RGBA, also make sure result image has zero Min point\n\tb := src.Bounds()\n\tif blockSize && (b.Dx() != engine.BlockSize || b.Dy() != engine.BlockSize) {\n\t\tpanic(\"Invalid image size!\")\n\t}\n\n\timg := image.NewRGBA(image.Rect(0, 0, b.Dx(), b.Dy()))\n\tdraw.Draw(img, src.Bounds(), src, b.Min, draw.Src)\n\n\treturn img\n}", "func (d *DefaultDriver) UpdateStorageClusterImage(string) error {\n\treturn &errors.ErrNotSupported{\n\t\tType: \"Function\",\n\t\tOperation: \"UpdateStorageClusterImage()\",\n\t}\n}", "func componentImageFor(component v1.ComponentKind) (types.Image, error) {\n\tenvVarFor := map[v1.ComponentKind]string{\n\t\tv1.ComponentQuay: componentImagePrefix + \"QUAY\",\n\t\tv1.ComponentClair: componentImagePrefix + \"CLAIR\",\n\t\tv1.ComponentRedis: componentImagePrefix + \"REDIS\",\n\t\tv1.ComponentPostgres: componentImagePrefix + \"POSTGRES\",\n\t}\n\tdefaultImagesFor := map[v1.ComponentKind]string{\n\t\tv1.ComponentQuay: \"quay.io/projectquay/quay\",\n\t\tv1.ComponentClair: \"quay.io/projectquay/clair\",\n\t\tv1.ComponentRedis: \"centos/redis-32-centos7\",\n\t\tv1.ComponentPostgres: \"centos/postgresql-10-centos7\",\n\t}\n\n\timageOverride := types.Image{\n\t\tName: defaultImagesFor[component],\n\t}\n\n\timage := os.Getenv(envVarFor[component])\n\tif image == \"\" {\n\t\treturn imageOverride, nil\n\t}\n\n\tif len(strings.Split(image, \"@\")) == 2 {\n\t\timageOverride.NewName = strings.Split(image, \"@\")[0]\n\t\timageOverride.Digest = strings.Split(image, \"@\")[1]\n\t} else if len(strings.Split(image, \":\")) == 2 {\n\t\timageOverride.NewName = strings.Split(image, \":\")[0]\n\t\timageOverride.NewTag = strings.Split(image, \":\")[1]\n\t} else {\n\t\treturn types.Image{}, fmt.Errorf(\n\t\t\t\"image override must be reference by tag or digest: %s\", image,\n\t\t)\n\t}\n\n\treturn imageOverride, nil\n}", "func decompileImageRecursive(vorteilImage *vdecompiler.IO, report DecompileReport, symlinkCallbacks []func() error, ino int, rpath string, dpath string) (DecompileReport, []func() error, error) {\n\tvar entries []*vdecompiler.DirectoryEntry\n\n\tinode, err := vorteilImage.ResolveInode(ino)\n\tif err != nil {\n\t\treturn report, nil, err\n\t}\n\n\tif report.SkipNotTouched && inode.LastAccessTime == 0 && !vdecompiler.InodeIsDirectory(inode) && rpath != \"/\" {\n\t\treport.ImageFiles = append(report.ImageFiles, DecompiledFile{\n\t\t\tPath: rpath,\n\t\t\tResult: SkippedNotTouched,\n\t\t})\n\t\tgoto DONE\n\t}\n\n\tif vdecompiler.InodeIsSymlink(inode) {\n\t\tsymlinkCallbacks = append(symlinkCallbacks, createSymlinkCallback(vorteilImage, inode, dpath))\n\t\treport.ImageFiles = append(report.ImageFiles, DecompiledFile{\n\t\t\tPath: rpath,\n\t\t\tResult: CopiedSymlink,\n\t\t})\n\t\tgoto DONE\n\t}\n\n\tif vdecompiler.InodeIsRegularFile(inode) {\n\t\terr = copyInodeToRegularFile(vorteilImage, inode, dpath)\n\t\tif err == nil {\n\t\t\treport.ImageFiles = append(report.ImageFiles, DecompiledFile{\n\t\t\t\tPath: rpath,\n\t\t\t\tResult: CopiedRegularFile,\n\t\t\t})\n\t\t}\n\t\tgoto DONE\n\t}\n\n\tif !vdecompiler.InodeIsDirectory(inode) {\n\t\treport.ImageFiles = append(report.ImageFiles, DecompiledFile{\n\t\t\tPath: rpath,\n\t\t\tResult: SkippedAbnormalFile,\n\t\t})\n\t\tgoto DONE\n\t}\n\n\t// INODE IS DIR\n\terr = utilFileNotExists(dpath)\n\tif err == nil {\n\t\terr = os.MkdirAll(dpath, 0777)\n\t\tif err == nil {\n\t\t\treport.ImageFiles = append(report.ImageFiles, DecompiledFile{\n\t\t\t\tPath: rpath,\n\t\t\t\tResult: CopiedMkDir,\n\t\t\t})\n\t\t\tentries, err = vorteilImage.Readdir(inode)\n\t\t}\n\t}\n\n\tif err != nil {\n\t\treturn report, nil, err\n\t}\n\n\tfor _, entry := range entries {\n\t\tif entry.Name == \".\" || entry.Name == \"..\" {\n\t\t\tcontinue\n\t\t}\n\t\treport, symlinkCallbacks, err = decompileImageRecursive(vorteilImage, report, symlinkCallbacks, entry.Inode, filepath.ToSlash(filepath.Join(rpath, entry.Name)), filepath.Join(dpath, entry.Name))\n\t\tif err != nil {\n\t\t\treturn report, nil, err\n\t\t}\n\t}\n\nDONE:\n\treturn report, symlinkCallbacks, err\n}", "func (c MockDockerClient) ImageExists(ctx context.Context, imageName string) bool {\n\tif c.ImageExistsFn != nil {\n\t\tfmt.Println(\"[MockDockerClient] In \", utils.CurrentFunctionName())\n\t\tfmt.Println(\"[MockDockerClient] - ctx: \", ctx)\n\t\tfmt.Println(\"[MockDockerClient] - imageName: \", imageName)\n\t\treturn c.ImageExistsFn(ctx, imageName)\n\t}\n\tpanic(fmt.Sprintf(\"No function defined for: %s\", utils.CurrentFunctionName()))\n}", "func (l Path) writeImageWithoutLayers(img v1.Image, annotations map[string]string) error {\n\tif err := l.writeImage(img); err != nil {\n\t\treturn err\n\t}\n\n\tmt, err := img.MediaType()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\td, err := img.Digest()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tmanifest, err := img.RawManifest()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdesc := v1.Descriptor{\n\t\tMediaType: mt,\n\t\tSize: int64(len(manifest)),\n\t\tDigest: d,\n\t\tAnnotations: annotations,\n\t}\n\treturn l.AppendDescriptor(desc)\n}", "func (daemon *Daemon) ImageService() ImageService {\n\treturn daemon.imageService\n}", "func supportsImageUpgrade(uc *upgradev1alpha1.UpgradeConfig) bool {\n\treturn !empty(uc.Spec.Desired.Image) && !empty(uc.Spec.Desired.Version) && empty(uc.Spec.Desired.Channel)\n}", "func (m *HealthCheck_HttpHealthCheck) GetHiddenEnvoyDeprecatedServiceName() string {\n\tif m != nil {\n\t\treturn m.HiddenEnvoyDeprecatedServiceName\n\t}\n\treturn \"\"\n}", "func NewImageReplacer() *ImageReplacer {\n\tir := ImageReplacer{}\n\n\treturn &ir\n}", "func newImageDestination(ctx context.Context, sys *types.SystemContext, ref daemonReference) (private.ImageDestination, error) {\n\tif ref.ref == nil {\n\t\treturn nil, fmt.Errorf(\"Invalid destination docker-daemon:%s: a destination must be a name:tag\", ref.StringWithinTransport())\n\t}\n\tnamedTaggedRef, ok := ref.ref.(reference.NamedTagged)\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"Invalid destination docker-daemon:%s: a destination must be a name:tag\", ref.StringWithinTransport())\n\t}\n\n\tvar mustMatchRuntimeOS = true\n\tif sys != nil && sys.DockerDaemonHost != client.DefaultDockerHost {\n\t\tmustMatchRuntimeOS = false\n\t}\n\n\tc, err := newDockerClient(sys)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"initializing docker engine client: %w\", err)\n\t}\n\n\treader, writer := io.Pipe()\n\tarchive := tarfile.NewWriter(writer)\n\t// Commit() may never be called, so we may never read from this channel; so, make this buffered to allow imageLoadGoroutine to write status and terminate even if we never read it.\n\tstatusChannel := make(chan error, 1)\n\n\tgoroutineContext, goroutineCancel := context.WithCancel(ctx)\n\tgo imageLoadGoroutine(goroutineContext, c, reader, statusChannel)\n\n\treturn &daemonImageDestination{\n\t\tref: ref,\n\t\tmustMatchRuntimeOS: mustMatchRuntimeOS,\n\t\tDestination: tarfile.NewDestination(sys, archive, ref.Transport().Name(), namedTaggedRef),\n\t\tarchive: archive,\n\t\tgoroutineCancel: goroutineCancel,\n\t\tstatusChannel: statusChannel,\n\t\twriter: writer,\n\t\tcommitted: false,\n\t}, nil\n}", "func (x *Place_OpeningHours_OpeningHoursPeriod_OpeningHoursPoint) GetDateDeprecated() string {\n\tif x != nil {\n\t\treturn x.DateDeprecated\n\t}\n\treturn \"\"\n}", "func (p *Policy) MutateImage(image string) (string, bool) {\n\tvar msg string\n\tfor _, rule := range p.Rules {\n\t\tif rule.re.MatchString(image) {\n\t\t\timage := image\n\t\t\tif rule.Replacement != \"\" {\n\t\t\t\timage = rule.re.ReplaceAllString(image, rule.Replacement)\n\t\t\t}\n\t\t\tif rule.Condition == \"Exists\" && !imageExists(image) {\n\t\t\t\tmsg = fmt.Sprintf(\"%s does not exist in private registry\", image)\n\t\t\t\tlog.Debug(msg)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\treturn image, true\n\t\t}\n\t}\n\tif msg != \"\" {\n\t\tlog.Print(msg)\n\t\tSendSlackNotification(msg)\n\t}\n\treturn image, false\n}", "func DeleteImageService(imageid string) bool {\n\tsuccess := domain.UserDeleteItem(imageid)\n\treturn success\n}", "func fallback(src image.Image) *image.RGBA {\n\tstart := time.Now()\n\n\t// Create a new RGBA image and draw the src image onto it.\n\tbounds := src.Bounds()\n\tdr := image.Rect(0, 0, bounds.Dx(), bounds.Dy())\n\tdst := image.NewRGBA(dr)\n\tdraw.Draw(dst, dr, src, bounds.Min, draw.Src)\n\n\tlog.Printf(\"texture.fallback: fallback conversion for non-RGBA image (%T) finished in: %v\", src, time.Since(start))\n\n\treturn dst\n}", "func resolveCustomImageStreamName(tagImageName string) string {\n\treturn fmt.Sprintf(\"custom-%s\", tagImageName)\n}" ]
[ "0.6556271", "0.5815334", "0.57859814", "0.57609904", "0.56974554", "0.56365556", "0.56276387", "0.55553997", "0.530765", "0.5110216", "0.5105804", "0.50867426", "0.50865674", "0.5020373", "0.49845996", "0.49688843", "0.4942152", "0.48918223", "0.48170793", "0.4805322", "0.4791222", "0.47827047", "0.47800207", "0.47603402", "0.46475676", "0.46458855", "0.46337977", "0.462443", "0.46160224", "0.46131253", "0.45951673", "0.45812827", "0.45581752", "0.4555108", "0.45257387", "0.45255446", "0.45140222", "0.4498226", "0.4466163", "0.4465933", "0.44609603", "0.4427841", "0.44216436", "0.4407003", "0.44064415", "0.44044223", "0.44029218", "0.43995273", "0.43943083", "0.43643498", "0.43641812", "0.43607286", "0.43472835", "0.43333206", "0.43315458", "0.43088734", "0.42980903", "0.42955264", "0.4284065", "0.42796367", "0.42718995", "0.42682976", "0.42673254", "0.42641765", "0.42545432", "0.42521623", "0.42421165", "0.42404622", "0.42386094", "0.42371026", "0.42360967", "0.42259687", "0.42226136", "0.422248", "0.4211665", "0.42096514", "0.42056718", "0.41859958", "0.41850668", "0.4180659", "0.41733032", "0.41718826", "0.41710743", "0.41705716", "0.41688475", "0.41648158", "0.41550153", "0.41523597", "0.41322446", "0.41314912", "0.4126649", "0.4126576", "0.4120257", "0.4118307", "0.41065174", "0.4099092", "0.40974042", "0.40972054", "0.40891472", "0.4088012" ]
0.7522792
0
GetProject uses the override method GetProjectFn or the real implementation.
func (c *TestClient) GetProject(project string) (*compute.Project, error) { if c.GetProjectFn != nil { return c.GetProjectFn(project) } return c.client.GetProject(project) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (_BREMFactory *BREMFactoryCaller) GetProject(opts *bind.CallOpts, _index *big.Int) (common.Address, error) {\n\tvar (\n\t\tret0 = new(common.Address)\n\t)\n\tout := ret0\n\terr := _BREMFactory.contract.Call(opts, out, \"getProject\", _index)\n\treturn *ret0, err\n}", "func (_BREM *BREMCaller) GetProject(opts *bind.CallOpts, _index *big.Int) (common.Address, error) {\n\tvar (\n\t\tret0 = new(common.Address)\n\t)\n\tout := ret0\n\terr := _BREM.contract.Call(opts, out, \"getProject\", _index)\n\treturn *ret0, err\n}", "func (_BREMFactory *BREMFactoryCallerSession) GetProject(_index *big.Int) (common.Address, error) {\n\treturn _BREMFactory.Contract.GetProject(&_BREMFactory.CallOpts, _index)\n}", "func (p *Provider) GetProject() string {\n\to := p.opts\n\tif len(o.projects) > 1 {\n\t\tpanic(fmt.Sprintf(\n\t\t\t\"multiple projects not supported (%d specified)\", len(o.projects)))\n\t}\n\treturn o.projects[0]\n}", "func (db *DB) GetProject(ctx context.Context, projectName string) (*cluster.Project, error) {\n\tvar err error\n\tvar p *cluster.Project\n\terr = db.Cluster.Transaction(context.TODO(), func(ctx context.Context, tx *ClusterTx) error {\n\t\tp, err = cluster.GetProject(ctx, tx.Tx(), projectName)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn p, nil\n}", "func GetProject(r *http.Request) *model.Project {\n\tp := r.Context().Value(model.ApiProjectKey)\n\tif p == nil {\n\t\treturn nil\n\t}\n\n\treturn p.(*model.Project)\n}", "func (f *FakeProjectProvider) Get(userInfo *provider.UserInfo, projectInternalName string, options *provider.ProjectGetOptions) (*kubermaticapiv1.Project, error) {\n\tif NoExistingFakeProjectID == projectInternalName || ForbiddenFakeProjectID == projectInternalName {\n\t\treturn nil, createError(http.StatusForbidden, ImpersonatedClientErrorMsg)\n\t}\n\n\treturn GenProject(ExistingFakeProject, kubermaticapiv1.ProjectActive, DefaultCreationTimestamp().Add(2*time.Minute)), nil\n}", "func (_BREM *BREMCallerSession) GetProject(_index *big.Int) (common.Address, error) {\n\treturn _BREM.Contract.GetProject(&_BREM.CallOpts, _index)\n}", "func (test *Test) GetProject(projectName string) (models.Project, error) {\n\tps, err := test.GetProjects()\n\tvar projects []models.Project\n\tprojects = append([]models.Project{}, ps...)\n\n\tif err != nil {\n\t\treturn models.Project{}, errors.New(\"Could not get projects\" + err.Error())\n\t}\n\n\tfor _, p := range projects {\n\t\tif p.Name == projectName {\n\t\t\treturn p, nil\n\t\t}\n\t}\n\treturn models.Project{}, errors.New(\"Could not get project \" + projectName)\n}", "func (_BREMFactory *BREMFactorySession) GetProject(_index *big.Int) (common.Address, error) {\n\treturn _BREMFactory.Contract.GetProject(&_BREMFactory.CallOpts, _index)\n}", "func (c *Cluster) GetProject(projectName string) (*Project, error) {\n\tvar err error\n\tvar p *Project\n\terr = c.Transaction(func(tx *ClusterTx) error {\n\t\tp, err = tx.GetProject(projectName)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn p, nil\n}", "func GetProject(w http.ResponseWriter, r *http.Request) {\n\t// Get item params\n\t// Perform get, db n' stuff.\n\t// render.JSON(w, r)\n}", "func (p *ProjectProvider) Get(userInfo *provider.UserInfo, projectInternalName string, options *provider.ProjectGetOptions) (*kubermaticapiv1.Project, error) {\n\tif userInfo == nil {\n\t\treturn nil, errors.New(\"a user is missing but required\")\n\t}\n\tif options == nil {\n\t\toptions = &provider.ProjectGetOptions{IncludeUninitialized: true}\n\t}\n\tmasterImpersonatedClient, err := createImpersonationClientWrapperFromUserInfo(userInfo, p.createMasterImpersonatedClient)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\texistingProject := &kubermaticapiv1.Project{}\n\tif err := masterImpersonatedClient.Get(context.Background(), ctrlruntimeclient.ObjectKey{Name: projectInternalName}, existingProject); err != nil {\n\t\treturn nil, err\n\t}\n\tif !options.IncludeUninitialized && existingProject.Status.Phase != kubermaticapiv1.ProjectActive {\n\t\treturn nil, kerrors.NewServiceUnavailable(\"Project is not initialized yet\")\n\t}\n\n\treturn existingProject, nil\n}", "func GetProject(props *OvfImportTestProperties, testProjectConfig *testconfig.Project) string {\n\tif props.Project != \"\" {\n\t\treturn props.Project\n\t}\n\treturn testProjectConfig.TestProjectID\n}", "func (s *StubTodoStore) GetProject(name string) model.Project {\n\tproject := model.Project{}\n\tif _, exists := s.Projects[name]; exists {\n\t\tproject.Name = name\n\t\treturn project\n\t} else {\n\t\treturn project\n\t}\n}", "func (c *client) GetProject(\n\tid interface{},\n\topt *glab.GetProjectOptions,\n\toptions ...glab.RequestOptionFunc,\n) (*glab.Project, *glab.Response, error) {\n\treturn c.c.Projects.GetProject(id, opt, options...)\n}", "func (w *ServerInterfaceWrapper) GetProject(ctx echo.Context) error {\n\tvar err error\n\t// ------------- Path parameter \"id\" -------------\n\tvar id string\n\n\terr = runtime.BindStyledParameter(\"simple\", false, \"id\", ctx.Param(\"id\"), &id)\n\tif err != nil {\n\t\treturn echo.NewHTTPError(http.StatusBadRequest, fmt.Sprintf(\"Invalid format for parameter id: %s\", err))\n\t}\n\n\t// HasSecurity is set\n\n\tctx.Set(\"OpenId.Scopes\", []string{\"exitus/project.read\"})\n\n\t// Invoke the callback with all the unmarshalled arguments\n\terr = w.Handler.GetProject(ctx, id)\n\treturn err\n}", "func GetProject(n string) Project {\n\tvar p Project\n\tn = strings.TrimSpace(n)\n\tDB.Where(&Project{Name: n}).First(&p)\n\treturn p\n}", "func (g *GitLab) getProject(ctx context.Context, client *gitlab.Client, owner, name string) (*gitlab.Project, error) {\n\trepo, _, err := client.Projects.GetProject(fmt.Sprintf(\"%s/%s\", owner, name), nil, gitlab.WithContext(ctx))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn repo, nil\n}", "func (a GetSLITriggeredAdapter) GetProject() string {\n\treturn a.event.Project\n}", "func (d *Driver) ProjectGet(partitionID string) (*ProjectGetResponse, error) {\n\tresponse := &ProjectGetResponse{}\n\tgetProject := project.NewFindProjectParams()\n\tgetProject.ID = partitionID\n\tresp, err := d.project.FindProject(getProject, d.auth)\n\tif err != nil {\n\t\treturn response, err\n\t}\n\tresponse.Project = resp.Payload\n\treturn response, nil\n}", "func (_BREM *BREMSession) GetProject(_index *big.Int) (common.Address, error) {\n\treturn _BREM.Contract.GetProject(&_BREM.CallOpts, _index)\n}", "func GetProject(w http.ResponseWriter, r *http.Request) {\n\tvars := mux.Vars(r)\n\n\tp := models.Project{\n\t\tKey: vars[\"key\"],\n\t}\n\n\terr := Store.Projects().Get(&p)\n\tif err != nil {\n\t\tw.WriteHeader(500)\n\t\tw.Write(apiError(err.Error()))\n\t\tlog.Println(err)\n\t\treturn\n\t}\n\n\tsendJSON(w, p)\n}", "func GetProject(name Fixture) (*model.Project, error) {\n\twd, err := os.Getwd()\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn model.NewProject(string(name), path.Join(wd, string(name))), nil\n}", "func (s *Server) Get(ctx context.Context, q *ProjectQuery) (*v1alpha1.AppProject, error) {\n\tif !s.enf.EnforceClaims(ctx.Value(\"claims\"), \"projects\", \"get\", q.Name) {\n\t\treturn nil, grpc.ErrPermissionDenied\n\t}\n\treturn s.appclientset.ArgoprojV1alpha1().AppProjects(s.ns).Get(q.Name, metav1.GetOptions{})\n}", "func (a ProblemAdapter) GetProject() string {\n\treturn a.event.KeptnProject\n}", "func GetProject(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *ProjectState, opts ...pulumi.ResourceOption) (*Project, error) {\n\tvar resource Project\n\terr := ctx.ReadResource(\"google-native:cloudresourcemanager/v3:Project\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func GetProject(pid string, token string) (*Project, error) {\n\t// declarations\n\tproject := Project{}\n\tvar name, clone_url, fs_path sql.NullString\n\n\t// fetch project and verify token\n\tif err := db.QueryRow(\"SELECT projects.*, users.token FROM projects\"+\n\t\t\" INNER JOIN members ON projects.id=members.pid\"+\n\t\t\" INNER JOIN users ON members.uid=users.id\"+\n\t\t\" WHERE projects.id=$1 AND users.token=$2\", pid, token).\n\t\tScan(&project.Id, &project.GH_Id, &name, &clone_url, &fs_path,\n\t\t&token); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// set remaining fields\n\tif name.Valid {\n\t\tproject.Name = name.String\n\t}\n\tif clone_url.Valid {\n\t\tproject.Clone_url = clone_url.String\n\t}\n\tif fs_path.Valid {\n\t\tproject.Fs_path = fs_path.String\n\t}\n\n\treturn &project, nil\n}", "func Project(c context.Context) luciConfig.ProjectName {\n\tns, _ := info.Get(c).GetNamespace()\n\tproject := ProjectFromNamespace(ns)\n\tif project != \"\" {\n\t\treturn project\n\t}\n\tpanic(fmt.Errorf(\"current namespace %q does not begin with project namespace prefix (%q)\", ns, projectNamespacePrefix))\n}", "func GetProject(w http.ResponseWriter, r *http.Request) {\n\tvars := mux.Vars(r)\n\tname := vars[\"name\"]\n\tid := vars[\"short_id\"]\n\n\t_, err := database.GetSiteByName(name)\n\tif err != nil {\n\t\tutils.RespondWithJSON(w, http.StatusNotFound, \"not_found\", nil)\n\t\treturn\n\t}\n\n\tproject, err := database.GetProjectByShortID(id)\n\tif err != nil {\n\t\tutils.RespondWithJSON(w, http.StatusNotFound, err.Error(), nil)\n\t\treturn\n\t}\n\tutils.RespondWithJSON(w, http.StatusOK, \"success\", project)\n\treturn\n}", "func (p *ProjectHandler) GetProject(ctx context.Context,\n\treq *proto.GetProjectRequest, resp *proto.ProjectResponse) error {\n\t// 查询项目信息\n\tga := project.NewGetAction(p.model)\n\tprojectInfo, err := ga.Do(ctx, req)\n\tif err != nil {\n\t\treturn err\n\t}\n\tbusinessName := \"\"\n\tif projectInfo.BusinessID != \"\" && projectInfo.BusinessID != \"0\" {\n\t\tbusiness, err := cmdb.GetBusinessByID(projectInfo.BusinessID, true)\n\t\tif err != nil {\n\t\t\treturn errorx.NewRequestCMDBErr(err.Error())\n\t\t}\n\t\tbusinessName = business.BKBizName\n\t}\n\t// 处理返回数据及权限\n\tsetResp(resp, projectInfo)\n\tresp.Data.BusinessName = businessName\n\treturn nil\n}", "func (cfg Config) GetProject(projectName string) (*Project, error) {\n\tproject, ok := cfg.Projects[projectName]\n\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"Could not find project with name %q\", projectName)\n\t}\n\n\treturn project, nil\n}", "func GetProject(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *ProjectState, opts ...pulumi.ResourceOption) (*Project, error) {\n\tvar resource Project\n\terr := ctx.ReadResource(\"azure-nextgen:migrate/latest:Project\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (p *Projects) GetProject(rw http.ResponseWriter, r *http.Request) {\n\n\tvars := mux.Vars(r)\n\tid, ok := vars[\"projectID\"]\n\tif !ok {\n\t\tio.WriteString(rw, `{{\"error\": \"id not found\"}}`)\n\t\treturn\n\t}\n\n\tproject, err := data.FindProjectByID(id)\n\n\tif err != nil {\n\t\tio.WriteString(rw, `{{\"error\": \"user not found\"}}`)\n\t\treturn\n\t}\n\n\terr = data.ToJSON(project, rw)\n}", "func (v *ProjectClient) GetProject(name string) (Project, error) {\n\n\t//Construct the composite key to select the entry\n\tkey := ProjectKey{\n\t\tProjectName: name,\n\t}\n\tvalue, err := db.DBconn.Find(v.storeName, key, v.tagMeta)\n\tif err != nil {\n\t\treturn Project{}, pkgerrors.Wrap(err, \"Get Project\")\n\t}\n\n\t//value is a byte array\n\tif value != nil {\n\t\tproj := Project{}\n\t\terr = db.DBconn.Unmarshal(value[0], &proj)\n\t\tif err != nil {\n\t\t\treturn Project{}, pkgerrors.Wrap(err, \"Unmarshaling Value\")\n\t\t}\n\t\treturn proj, nil\n\t}\n\n\treturn Project{}, pkgerrors.New(\"Error getting Project\")\n}", "func (s *Stub) GetProject() string {\n\treturn \"\"\n}", "func GetProject() string {\n\tvar (\n\t\tproject string\n\t\terr error\n\t)\n\n\tproject, err = metadata.ProjectID()\n\tif err != nil {\n\t\tif project = os.Getenv(\"GOOGLE_PROJECT\"); project == \"\" {\n\t\t\tlog.Println(\"project id can't be empty\")\n\t\t\tos.Exit(1)\n\t\t}\n\t}\n\n\treturn project\n}", "func (mockProvider) GetProjectConfig(c context.Context, p string) (*tricium.ProjectConfig, error) {\n\treturn &tricium.ProjectConfig{}, nil\n}", "func GetProject() string {\n\tvar (\n\t\tproject string\n\t\terr error\n\t)\n\n\tproject, err = metadata.ProjectID()\n\tif err != nil {\n\t\tif project = os.Getenv(\"GOOGLE_PROJECT\"); project == \"\" {\n\t\t\tlog.Fatal(\"project id can't be empty\")\n\t\t}\n\t}\n\n\treturn project\n}", "func (d *common) Project() string {\n\treturn d.projectName\n}", "func GetProject(host string, verifyTLS bool, apiKey string, project string) (models.ProjectInfo, Error) {\n\tvar params []queryParam\n\tparams = append(params, queryParam{Key: \"project\", Value: project})\n\n\turl, err := generateURL(host, \"/v3/projects/project\", params)\n\tif err != nil {\n\t\treturn models.ProjectInfo{}, Error{Err: err, Message: \"Unable to generate url\"}\n\t}\n\n\tstatusCode, _, response, err := GetRequest(url, verifyTLS, apiKeyHeader(apiKey))\n\tif err != nil {\n\t\treturn models.ProjectInfo{}, Error{Err: err, Message: \"Unable to fetch project\", Code: statusCode}\n\t}\n\n\tvar result map[string]interface{}\n\terr = json.Unmarshal(response, &result)\n\tif err != nil {\n\t\treturn models.ProjectInfo{}, Error{Err: err, Message: \"Unable to parse API response\", Code: statusCode}\n\t}\n\n\tresultProject, ok := result[\"project\"].(map[string]interface{})\n\tif !ok {\n\t\treturn models.ProjectInfo{}, Error{Err: fmt.Errorf(\"Unexpected type for project, expected map[string]interface{}, got %T\", result[\"project\"]), Message: \"Unable to parse API response\", Code: statusCode}\n\t}\n\tprojectInfo := models.ParseProjectInfo(resultProject)\n\treturn projectInfo, Error{}\n}", "func (c *ClientImpl) GetProject(ctx context.Context, hcpHostURL, projectID string) (hcpModels.Tenant, error) {\n\tspan, _ := opentracing.StartSpanFromContext(ctx, \"Get HCP Project\")\n\tdefer span.Finish()\n\n\tsession, err := c.getSession(ctx, hcpHostURL, hcpUserName, hcpPassword)\n\tif err != nil {\n\t\treturn hcpModels.Tenant{}, err\n\t}\n\n\tstatus = Failure\n\tmonitor := metrics.StartExternalCall(externalSvcName, \"Get Project from HCP\")\n\tdefer func() { monitor.RecordWithStatus(status) }()\n\n\tresp, err := mlopsHttp.ExecuteHTTPRequest(\n\t\tctx,\n\t\tc.client,\n\t\thcpHostURL+projectPathV1+projectID,\n\t\thttp.MethodGet,\n\t\tmap[string]string{sessionHeader: session},\n\t\tbytes.NewReader(nil),\n\t)\n\tif err != nil {\n\t\treturn hcpModels.Tenant{}, errors.Wrapf(err, \"while fetching project from MLOps controller platform.\")\n\t}\n\n\tstatus = Success\n\n\terr = c.deleteSession(ctx, hcpHostURL, session)\n\tif err != nil {\n\t\treturn hcpModels.Tenant{}, err\n\t}\n\n\tvar tenants hcpModels.Tenant\n\t_, parseRespErr := common.ParseResponse(resp, &tenants)\n\tif parseRespErr != nil {\n\t\tlog.Errorf(\"Failed to fetch project from HCP: %v\", parseRespErr)\n\t\treturn hcpModels.Tenant{}, errors.Wrapf(parseRespErr, \"Failed to fetch project from HCP\")\n\t}\n\n\treturn tenants, nil\n}", "func (p *ProjectEvent) GetProject() *Project {\n\tif p == nil {\n\t\treturn nil\n\t}\n\treturn p.Project\n}", "func GetProject(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *ProjectState, opts ...pulumi.ResourceOption) (*Project, error) {\n\tvar resource Project\n\terr := ctx.ReadResource(\"aws:codebuild/project:Project\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (p *Project) GetProject(name string) (project *domain.Project, err error) {\n\tprojects := &domain.Projects{}\n\tif err := toml.Unmarshal(p.bytes, projects); err != nil {\n\t\treturn nil, err\n\t}\n\tfor _, v := range projects.Projects {\n\t\tif v.Name == name {\n\t\t\tproject = v\n\t\t}\n\t}\n\tif project == nil {\n\t\treturn nil, fmt.Errorf(\"%s is not found\", name)\n\t}\n\n\treturn\n}", "func Get(collectiom Getter) func(ctx context.Context, id string) (interface{}, error) {\n\treturn func(ctx context.Context, id string) (interface{}, error) {\n\t\tproject := &Project{}\n\t\tif err := collectiom.Get(ctx, id, project); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn project, nil\n\t}\n}", "func (s *Server) Get(ctx context.Context, q *project.ProjectQuery) (*v1alpha1.AppProject, error) {\n\tif err := s.enf.EnforceErr(ctx.Value(\"claims\"), rbacpolicy.ResourceProjects, rbacpolicy.ActionGet, q.Name); err != nil {\n\t\treturn nil, err\n\t}\n\tproj, err := s.appclientset.ArgoprojV1alpha1().AppProjects(s.ns).Get(ctx, q.Name, metav1.GetOptions{})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tproj.NormalizeJWTTokens()\n\treturn proj, err\n}", "func (_BREMFactory *BREMFactoryCaller) GetProjectByName(opts *bind.CallOpts, _projectName string) (common.Address, error) {\n\tvar (\n\t\tret0 = new(common.Address)\n\t)\n\tout := ret0\n\terr := _BREMFactory.contract.Call(opts, out, \"getProjectByName\", _projectName)\n\treturn *ret0, err\n}", "func (repo *repo) GetExternalProject(projectExternalID string) (*models.Project, error) {\n\tlog.Debugf(\"GetExternalProject - projectID: %s\", projectExternalID)\n\ttableName := fmt.Sprintf(\"cla-%s-projects\", repo.stage)\n\t// This is the key we want to match\n\tcondition := expression.Key(\"project_external_id\").Equal(expression.Value(projectExternalID))\n\n\t// Use the builder to create the expression\n\texpr, err := expression.NewBuilder().WithKeyCondition(condition).WithProjection(buildProjection()).Build()\n\tif err != nil {\n\t\tlog.Warnf(\"error building expression for Project query, projectExternalID: %s, error: %v\",\n\t\t\tprojectExternalID, err)\n\t\treturn nil, err\n\t}\n\n\t// Assemble the query input parameters\n\tqueryInput := &dynamodb.QueryInput{\n\t\tExpressionAttributeNames: expr.Names(),\n\t\tExpressionAttributeValues: expr.Values(),\n\t\tKeyConditionExpression: expr.KeyCondition(),\n\t\tProjectionExpression: expr.Projection(),\n\t\tTableName: aws.String(tableName),\n\t\tIndexName: aws.String(\"external-project-index\"),\n\t}\n\n\t// Make the DynamoDB Query API call\n\tresults, queryErr := repo.dynamoDBClient.Query(queryInput)\n\tif queryErr != nil {\n\t\tlog.Warnf(\"error retrieving project by projectExternalID: %s, error: %v\", projectExternalID, queryErr)\n\t\treturn nil, queryErr\n\t}\n\n\t// No match, didn't find it\n\tif *results.Count == 0 {\n\t\treturn nil, nil\n\t}\n\n\t// Should only have one result\n\tif *results.Count > 1 {\n\t\tlog.Warnf(\"Project query returned more than one result using projectExternalID: %s\", projectExternalID)\n\t}\n\n\tvar dbModel DBProjectModel\n\terr = dynamodbattribute.UnmarshalMap(results.Items[0], &dbModel)\n\tif err != nil {\n\t\tlog.Warnf(\"error unmarshalling db project model, error: %+v\", err)\n\t\treturn nil, err\n\t}\n\n\t// Convert the database model to an API response model\n\treturn repo.buildProjectModel(dbModel), nil\n}", "func (o *Vm) GetProject() string {\n\tif o == nil || o.Project == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Project\n}", "func getProjectClient(ctx iris.Context) {\n\tclientID := ctx.Params().Get(\"clientID\")\n\tif clientID == \"\" {\n\t\thandleError(ctx, apiErrors.ErrBadRequest)\n\t\treturn\n\t}\n\tprojectID := ctx.Params().Get(\"projectID\")\n\tif projectID == \"\" {\n\t\thandleError(ctx, apiErrors.ErrBadRequest)\n\t\treturn\n\t}\n\n\tresult, err := store.GetProjectClient(projectID, clientID)\n\tif err != nil {\n\t\thandleError(ctx, err)\n\t\treturn\n\t}\n\n\trender.JSON(ctx, iris.StatusOK, result)\n}", "func (_BREM *BREMCaller) GetProjectByName(opts *bind.CallOpts, _projectName string) (common.Address, error) {\n\tvar (\n\t\tret0 = new(common.Address)\n\t)\n\tout := ret0\n\terr := _BREM.contract.Call(opts, out, \"getProjectByName\", _projectName)\n\treturn *ret0, err\n}", "func (m *ModelProject) GetProject(ctx context.Context, projectIDOrCode string) (*Project, error) {\n\t// query project info by the `or` operation\n\tprojectIDCond := operator.NewLeafCondition(operator.Eq, operator.M{projectIDField: projectIDOrCode})\n\tprojectCodeCond := operator.NewLeafCondition(operator.Eq, operator.M{projectCodeField: projectIDOrCode})\n\tcond := operator.NewBranchCondition(operator.Or, projectIDCond, projectCodeCond)\n\n\tretProject := &Project{}\n\tif err := m.db.Table(m.tableName).Find(cond).One(ctx, retProject); err != nil {\n\t\treturn nil, err\n\t}\n\treturn retProject, nil\n}", "func (m *migration) project(endpoint *gitlab.Client, name, which string) (*gitlab.Project, error) {\n\tproj, resp, err := endpoint.Projects.GetProject(name)\n\tif resp == nil {\n\t\treturn nil, errors.New(\"network error: \" + err.Error())\n\t}\n\tif resp.StatusCode == http.StatusNotFound {\n\t\treturn nil, fmt.Errorf(\"%s project '%s' not found\", which, name)\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn proj, nil\n}", "func (node *Node) GetProject() string {\n\treturn node.project\n}", "func GetProject(id int) Project {\n\tvar project Project\n\tGetDB().First(&project, id)\n\treturn project\n}", "func (s *ResourceScope) GetProject() string {\n\treturn s.project\n}", "func (k *Keystone) GetProjectAPI(c echo.Context) error {\n\tclusterID := c.Request().Header.Get(xClusterIDKey)\n\tif ke := getKeystoneEndpoints(clusterID, k.endpointStore); len(ke) > 0 {\n\t\tc.Request().URL.Path = path.Join(c.Request().URL.Path, c.Param(\"id\"))\n\t\treturn k.proxyRequest(c)\n\t}\n\n\ttoken, err := k.validateToken(c.Request())\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// TODO(dfurman): prevent panic: use fields without pointers in models and/or provide getters with nil checks\n\tfor _, role := range token.User.Roles {\n\t\tif role.Project.ID == c.Param(\"id\") {\n\t\t\treturn c.JSON(http.StatusOK, &asfkeystone.ProjectResponse{\n\t\t\t\tProject: role.Project,\n\t\t\t})\n\t\t}\n\t}\n\n\treturn c.JSON(http.StatusNotFound, nil)\n}", "func (o *ProformaArray) GetProject() string {\n\tif o == nil || o.Project == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Project\n}", "func (pg *MongoDb) GetProject(ctx context.Context, pID string) (*prpb.Project, error) {\n\tpName := name.FormatProject(pID)\n\tcount, err := pg.Collection(\"Project\").CountDocuments(context.TODO(), bson.M{\"name\": pName})\n\tif err != nil {\n\t\treturn nil, status.Error(codes.Internal, \"Failed to query Project from database\")\n\t}\n\tif count > 0 {\n\t\treturn nil, status.Errorf(codes.NotFound, \"Project with name %q does not Exist\", pName)\n\t}\n\treturn &prpb.Project{Name: pName}, nil\n}", "func (k *Keystone) GetProjectAPI(c echo.Context) error {\n\tclusterID := c.Request().Header.Get(xClusterIDKey)\n\tif ke := getKeystoneEndpoints(clusterID, k.endpointStore); len(ke) > 0 {\n\t\tc.Request().URL.Path = path.Join(c.Request().URL.Path, c.Param(\"id\"))\n\t\treturn k.proxyRequest(c, ke)\n\t}\n\n\ttoken, err := k.validateToken(c.Request())\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// TODO(dfurman): prevent panic: use fields without pointers in models and/or provide getters with nil checks\n\tfor _, role := range token.User.Roles {\n\t\tif role.Project.ID == c.Param(\"id\") {\n\t\t\treturn c.JSON(http.StatusOK, &ProjectResponse{\n\t\t\t\tProject: role.Project,\n\t\t\t})\n\t\t}\n\t}\n\n\treturn c.JSON(http.StatusNotFound, nil)\n}", "func (o *Invoice) GetProject() string {\n\tif o == nil || o.Project == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Project\n}", "func (s *ProjectsService) GetProject(projectName string) (*ProjectInfo, *Response, error) {\n\tu := fmt.Sprintf(\"projects/%s\", url.QueryEscape(projectName))\n\n\tv := new(ProjectInfo)\n\tresp, err := s.client.Call(\"GET\", u, nil, v)\n\treturn v, resp, err\n}", "func (control *project) GetProject(ctx context.Context, projectCode string) (*bcsproject.Project, error) {\n\t// get information from project-manager\n\treq := &bcsproject.GetProjectRequest{ProjectIDOrCode: projectCode}\n\t// setting auth info\n\theader := metadata.New(map[string]string{\"Authorization\": fmt.Sprintf(\"Bearer %s\", control.option.APIToken)})\n\toutCxt := metadata.NewOutgoingContext(ctx, header)\n\tresp, err := control.client.GetProject(outCxt, req)\n\tif err != nil {\n\t\tblog.Errorf(\"get project %s details from project-manager failed, %s\", projectCode, err.Error())\n\t\treturn nil, fmt.Errorf(\"request to project-manager failure %s\", err.Error())\n\t}\n\tif resp.Code != 0 {\n\t\tblog.Errorf(\"request project-manager for %s failed, %s\", projectCode, resp.Message)\n\t\treturn nil, fmt.Errorf(\"project-manager response failure: %s\", resp.Message)\n\t}\n\tif resp.Data == nil {\n\t\tblog.Warnf(\"no project %s in bcs-project-manager\", projectCode)\n\t\treturn nil, nil\n\t}\n\tblog.V(5).Infof(\"project %s request details: %+v\", projectCode, resp.Data)\n\treturn resp.Data, nil\n}", "func GetProjectHandler(c *gin.Context) {\r\n\tuuid := c.Param(\"uuid\")\r\n\tres, err := GetProjectCore(uuid)\r\n\tif err != nil {\r\n\t\tc.JSON(404, ErrorResponse{ErrorCode: 404, ErrorMessage: err.Error()})\r\n\t} else {\r\n\t\tc.JSON(200, res)\r\n\t}\r\n}", "func Get(name string) (Project, error) {\n\tproj, ok := projects[name]\n\tif !ok {\n\t\treturn nil, ErrProjectNotFound\n\t}\n\treturn proj, nil\n}", "func (s *ProjectService) Get(project *models.Project) *ProjectService {\n\tif s.err != nil {\n\t\treturn s\n\t}\n\n\tproject.UUID = s.ksfile.ProjectId\n\tproject.Name = s.ksfile.ProjectName\n\n\t// But environment data is still on the server\n\tenvironments, err := s.cli.Project(project.UUID).GetAccessibleEnvironments()\n\tif err != nil {\n\t\ts.err = err\n\t\treturn s\n\t}\n\n\tproject.Environments = environments\n\n\treturn s\n}", "func ProjectGet(w http.ResponseWriter, r *http.Request) {\n\tdb := utils.GetDB()\n\tdefer db.Close()\n\n\tvar projects []models.Project\n\terr := db.Find(&projects).Error\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), 400)\n\t\tutils.LOG.Println(err)\n\t\treturn\n\t}\n\n\terr = json.NewEncoder(w).Encode(projects)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), 500)\n\t\tutils.LOG.Println(err)\n\t\treturn\n\t}\n\n\tw.WriteHeader(http.StatusOK)\n\tw.Write([]byte(`success`))\n\n}", "func Get(ctx context.Context, client *selvpcclient.ServiceClient, id string) (*Project, *selvpcclient.ResponseResult, error) {\n\turl := strings.Join([]string{client.Endpoint, resourceURL, id}, \"/\")\n\tresponseResult, err := client.DoRequest(ctx, http.MethodGet, url, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tif responseResult.Err != nil {\n\t\treturn nil, responseResult, responseResult.Err\n\t}\n\n\t// Extract a project from the response body.\n\tvar result struct {\n\t\tProject *Project `json:\"project\"`\n\t}\n\terr = responseResult.ExtractResult(&result)\n\tif err != nil {\n\t\treturn nil, responseResult, err\n\t}\n\n\treturn result.Project, responseResult, nil\n}", "func (b *ProjectModels) GetProject(id string) (ProjectAll, error) {\n\tvar result ProjectAll\n\tjoin := fmt.Sprintf(\"join %s on %s.id = %s.project_id \", TblProjectDetail, PROJECT, TblProjectDetail)\n\twhere := fmt.Sprintf(\"%s.id = ?\", PROJECT)\n\tselects := fmt.Sprintf(\"%s.*,%s.*\", PROJECT, TblProjectDetail)\n\n\terr := configs.GetDB.Table(PROJECT).Select(selects).Joins(join).Where(where, id).Find(&result).Error\n\treturn result, err\n}", "func (e *BcsDataManager) GetProjectInfo(ctx context.Context,\n\treq *bcsdatamanager.GetProjectInfoRequest, rsp *bcsdatamanager.GetProjectInfoResponse) error {\n\tblog.Infof(\"Received GetProjectInfo.Call request. Project id: %s, dimension:%s, startTime=%s, endTime=%s\",\n\t\treq.GetProject(), req.GetDimension(), time.Unix(req.GetStartTime(), 0),\n\t\ttime.Unix(req.GetEndTime(), 0))\n\tstart := time.Now()\n\tif req.GetProject() == \"\" && req.GetProjectCode() == \"\" && req.GetBusiness() == \"\" {\n\t\trsp.Message = fmt.Sprintf(\"get project info error, projectId, businessID or projectCode is required\")\n\t\trsp.Code = bcsCommon.AdditionErrorCode + 500\n\t\tblog.Errorf(rsp.Message)\n\t\tprom.ReportAPIRequestMetric(\"GetProjectInfo\", \"grpc\", prom.StatusErr, start)\n\t\treturn nil\n\t}\n\tif req.GetProject() == \"\" && req.GetBusiness() == \"\" && req.GetProjectCode() != \"\" {\n\t\tproject, err := e.resourceGetter.GetProjectInfo(ctx, \"\", req.GetProjectCode(), nil)\n\t\tif err != nil {\n\t\t\trsp.Message = fmt.Sprintf(\"get project info err:%v\", err)\n\t\t\trsp.Code = bcsCommon.AdditionErrorCode + 500\n\t\t\tblog.Errorf(rsp.Message)\n\t\t\tprom.ReportAPIRequestMetric(\"GetProjectInfo\", \"grpc\", prom.StatusErr, start)\n\t\t\treturn nil\n\t\t}\n\t\tif project == nil {\n\t\t\trsp.Message = fmt.Sprintf(\"cannot get project info by project code:%s\", req.GetProjectCode())\n\t\t\trsp.Code = bcsCommon.AdditionErrorCode + 500\n\t\t\tblog.Errorf(rsp.Message)\n\t\t\tprom.ReportAPIRequestMetric(\"GetProjectInfo\", \"grpc\", prom.StatusErr, start)\n\t\t\treturn nil\n\t\t}\n\t\treq.Project = project.ProjectID\n\t}\n\tresult, err := e.model.GetProjectInfo(ctx, req)\n\tif err != nil {\n\t\trsp.Message = fmt.Sprintf(\"get project info error: %v\", err)\n\t\trsp.Code = bcsCommon.AdditionErrorCode + 500\n\t\tblog.Errorf(rsp.Message)\n\t\tprom.ReportAPIRequestMetric(\"GetProjectInfo\", \"grpc\", prom.StatusErr, start)\n\t\treturn nil\n\t}\n\trsp.Data = result\n\trsp.Message = bcsCommon.BcsSuccessStr\n\trsp.Code = bcsCommon.BcsSuccess\n\tprom.ReportAPIRequestMetric(\"GetProjectInfo\", \"grpc\", prom.StatusOK, start)\n\treturn nil\n}", "func GetProject(projectID interface{}) (*gitlab.Project, error) {\n\tvar opts gitlab.GetProjectOptions\n\ttarget, resp, err := lab.Projects.GetProject(projectID, &opts)\n\tif resp != nil && resp.StatusCode == http.StatusNotFound {\n\t\treturn nil, ErrProjectNotFound\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn target, nil\n}", "func NewGetProjectDefault(code int) *GetProjectDefault {\n\treturn &GetProjectDefault{\n\t\t_statusCode: code,\n\t}\n}", "func (configProvider) GetProjectConfig(c context.Context, p string) (*tricium.ProjectConfig, error) {\n\treturn getProjectConfig(c, p)\n}", "func (o LookupTargetProjectResultOutput) Project() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupTargetProjectResult) string { return v.Project }).(pulumi.StringOutput)\n}", "func (o GetProjectHooksResultOutput) Project() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetProjectHooksResult) string { return v.Project }).(pulumi.StringOutput)\n}", "func Project(ctx context.Context, project string) (*configpb.ProjectConfig, error) {\n\tconfigs, err := Projects(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif c, ok := configs[project]; ok {\n\t\treturn c, nil\n\t}\n\treturn nil, ErrNotFoundProjectConfig\n}", "func (r commonResult) ExtractProject() (*Project, error) {\n\tvar s struct {\n\t\tProject *Project `json:\"project\"`\n\t}\n\terr := r.ExtractInto(&s)\n\treturn s.Project, err\n}", "func (o LookupServiceIntegrationResultOutput) Project() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupServiceIntegrationResult) string { return v.Project }).(pulumi.StringOutput)\n}", "func (o PipelineTriggerOutput) Project() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *PipelineTrigger) pulumi.StringOutput { return v.Project }).(pulumi.StringOutput)\n}", "func (o LookupPipelineResultOutput) Project() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupPipelineResult) string { return v.Project }).(pulumi.StringOutput)\n}", "func (c *Client) GetProject(p GetProjectParam) (*dto.Project, error) {\n\tvar project *dto.Project\n\n\tr, err := c.NewRequest(\n\t\t\"GET\",\n\t\tfmt.Sprintf(\n\t\t\t\"workspaces/%s/projects/%s\",\n\t\t\tp.Workspace,\n\t\t\tp.ProjectID,\n\t\t),\n\t\tnil,\n\t)\n\n\tif err != nil {\n\t\treturn project, err\n\t}\n\n\t_, err = c.Do(r, &project)\n\treturn project, err\n}", "func (o LookupServiceAccountResultOutput) Project() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupServiceAccountResult) string { return v.Project }).(pulumi.StringOutput)\n}", "func (o *ViewPortfolioCard) GetProject() ViewRelationship {\n\tif o == nil || o.Project == nil {\n\t\tvar ret ViewRelationship\n\t\treturn ret\n\t}\n\treturn *o.Project\n}", "func (o *FileProjectUsage) GetProject() ViewRelationship {\n\tif o == nil || o.Project == nil {\n\t\tvar ret ViewRelationship\n\t\treturn ret\n\t}\n\treturn *o.Project\n}", "func (a *IamProjectApiService) IamProjectGet(ctx context.Context, projectId string) ApiIamProjectGetRequest {\n\treturn ApiIamProjectGetRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\tprojectId: projectId,\n\t}\n}", "func (r *ResourceHandler) GetProjectResource(project string, resourceURI string) (*models.Resource, error) {\n\tr.ensureHandlerIsSet()\n\tbuildURI := r.Scheme + \"://\" + r.BaseURL + v1ProjectPath + \"/\" + project + pathToResource + \"/\" + url.QueryEscape(resourceURI)\n\treturn r.resourceHandler.GetResourceByURI(context.TODO(), buildURI)\n}", "func (o M3DbOutput) Project() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *M3Db) pulumi.StringOutput { return v.Project }).(pulumi.StringOutput)\n}", "func (c *ProjectService) Get(id string) (*Project, *http.Response, error) {\n\tproject := new(Project)\n\tapiError := new(APIError)\n\tpath := fmt.Sprintf(\"%s\", id)\n\tresp, err := c.sling.New().Get(path).Receive(project, apiError)\n\treturn project, resp, relevantError(err, *apiError)\n}", "func (_BREMFactory *BREMFactoryCallerSession) GetProjectByName(_projectName string) (common.Address, error) {\n\treturn _BREMFactory.Contract.GetProjectByName(&_BREMFactory.CallOpts, _projectName)\n}", "func GetProject(path string) (*cfg.Project, error) {\n\traw, err := ioutil.ReadFile(path)\n\tif err != nil {\n\t\tif os.IsNotExist(err) {\n\t\t\treturn nil, errors.New(\"project config file doesn't exist - try running 'inertia init'\")\n\t\t}\n\t\treturn nil, err\n\t}\n\n\tvar project cfg.Project\n\tif err = toml.Unmarshal(raw, &project); err != nil {\n\t\treturn nil, err\n\t}\n\treturn &project, nil\n}", "func (s *ProjectService) Get(projectsQuery ProjectsQuery) (*resources.Resources[*Project], error) {\n\tv, _ := query.Values(projectsQuery)\n\tpath := s.BasePath\n\tencodedQueryString := v.Encode()\n\tif len(encodedQueryString) > 0 {\n\t\tpath += \"?\" + encodedQueryString\n\t}\n\n\tresp, err := api.ApiGet(s.GetClient(), new(resources.Resources[*Project]), path)\n\tif err != nil {\n\t\treturn &resources.Resources[*Project]{}, err\n\t}\n\n\treturn resp.(*resources.Resources[*Project]), nil\n}", "func (o LookupOrganizationResultOutput) Project() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupOrganizationResult) string { return v.Project }).(pulumi.StringOutput)\n}", "func (dp *DummyProject) Get(name string) *project.Project {\n\tfor _, p := range projects {\n\t\tif p.Name == name {\n\t\t\treturn p\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o FlinkApplicationOutput) Project() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *FlinkApplication) pulumi.StringOutput { return v.Project }).(pulumi.StringOutput)\n}", "func GetProjectReport(cluster *core.Cluster, dbDomain db.Domain, dbProject db.Project, dbi db.Interface, filter reports.Filter, withSubresources bool) (*limes.ProjectReport, error) {\n\tprojectReports, err := reports.GetProjects(cluster, dbDomain.ID, &dbProject.ID, dbi, filter, withSubresources)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif len(projectReports) == 0 {\n\t\treturn nil, errors.New(\"no resource data found for project\")\n\t}\n\treturn projectReports[0], nil\n}", "func (o LookupMirrorMakerReplicationFlowResultOutput) Project() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupMirrorMakerReplicationFlowResult) string { return v.Project }).(pulumi.StringOutput)\n}", "func (s *projectService) getProjectByName(projectName, owner, user string) (*schema.Project, error) {\n\tproj, err := s.projects.ByName(projectName, owner)\n\tswitch {\n\tcase err != nil:\n\t\treturn nil, err\n\tcase !s.access.AllowedByOwner(proj.ID, user):\n\t\treturn nil, app.ErrNoAccess\n\tdefault:\n\t\treturn proj, nil\n\t}\n}", "func (s *ProjectsService) Get(ctx context.Context, projectName string) (*Project, error) {\n\tquery := url.Values{\n\t\t\"name\": []string{projectName},\n\t}\n\treq, err := s.Client.NewRequest(ctx, http.MethodGet, newURI(projectsURI), WithQuery(query))\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"get project request creation failed: %w\", err)\n\t}\n\tres, resp, err := s.Client.Do(req)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"get project failed: %w\", err)\n\t}\n\n\tif resp != nil && resp.StatusCode == http.StatusNotFound {\n\t\treturn nil, ErrNotFound\n\t}\n\n\tp := &ProjectsList{\n\t\tProjects: []*Project{},\n\t}\n\tif err := json.Unmarshal(res, &p); err != nil {\n\t\treturn nil, fmt.Errorf(\"get project failed, unable to unmarshal repository list json: %w\", err)\n\t}\n\n\tp.Projects[0].Session.set(resp)\n\treturn p.Projects[0], nil\n\n}", "func (o LaunchOutput) Project() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Launch) pulumi.StringOutput { return v.Project }).(pulumi.StringOutput)\n}" ]
[ "0.7660287", "0.7618937", "0.7389755", "0.7385537", "0.72930443", "0.7269164", "0.7258649", "0.72377634", "0.7146087", "0.71349865", "0.7119145", "0.71064484", "0.7099731", "0.70990634", "0.703752", "0.6977766", "0.6954682", "0.6931504", "0.6921696", "0.6904593", "0.6877311", "0.6870546", "0.6858693", "0.684964", "0.68473756", "0.6846279", "0.6837989", "0.68372643", "0.6789423", "0.678328", "0.6782017", "0.67593366", "0.67431825", "0.6735865", "0.67288065", "0.6723993", "0.67150426", "0.67032236", "0.6702301", "0.66971904", "0.6684082", "0.66794205", "0.6647854", "0.6607171", "0.6606337", "0.6593453", "0.6588639", "0.6553295", "0.65414846", "0.65262806", "0.6513647", "0.6503321", "0.64930725", "0.6487612", "0.6485601", "0.648337", "0.64654464", "0.6458356", "0.6451074", "0.64432585", "0.64220315", "0.64122295", "0.64079815", "0.6399948", "0.6382996", "0.63796526", "0.6376228", "0.63467854", "0.6344704", "0.63406104", "0.6309592", "0.62903535", "0.62843204", "0.62805194", "0.62662005", "0.6236846", "0.6235062", "0.62298757", "0.62228507", "0.62214595", "0.619718", "0.61904037", "0.61596894", "0.6139427", "0.61373305", "0.6134365", "0.612629", "0.6119194", "0.6110621", "0.61099356", "0.61060065", "0.61040586", "0.6099883", "0.60984224", "0.60776657", "0.6077018", "0.60658497", "0.60535663", "0.60493255", "0.6045336" ]
0.75150627
2
GetMachineType uses the override method GetMachineTypeFn or the real implementation.
func (c *TestClient) GetMachineType(project, zone, machineType string) (*compute.MachineType, error) { if c.GetMachineTypeFn != nil { return c.GetMachineTypeFn(project, zone, machineType) } return c.client.GetMachineType(project, zone, machineType) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o LookupInstanceResultOutput) MachineType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupInstanceResult) string { return v.MachineType }).(pulumi.StringOutput)\n}", "func MachineType(name string) string {\n\tif IsKIC(name) {\n\t\treturn \"container\"\n\t}\n\n\tif IsSSH(name) {\n\t\treturn \"bare metal machine\"\n\t}\n\n\tif IsVM(name) {\n\t\treturn \"VM\"\n\t}\n\n\t// none or mock\n\treturn \"bare metal machine\"\n}", "func (o AiEndpointDeployedModelDedicatedResourceMachineSpecOutput) MachineType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v AiEndpointDeployedModelDedicatedResourceMachineSpec) *string { return v.MachineType }).(pulumi.StringPtrOutput)\n}", "func (o InstanceFromTemplateOutput) MachineType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *InstanceFromTemplate) pulumi.StringOutput { return v.MachineType }).(pulumi.StringOutput)\n}", "func (o InstanceOutput) MachineType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Instance) pulumi.StringOutput { return v.MachineType }).(pulumi.StringOutput)\n}", "func (o AllocationSpecificSKUAllocationReservedInstancePropertiesResponseOutput) MachineType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v AllocationSpecificSKUAllocationReservedInstancePropertiesResponse) string { return v.MachineType }).(pulumi.StringOutput)\n}", "func (o SourceInstancePropertiesResponseOutput) MachineType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v SourceInstancePropertiesResponse) string { return v.MachineType }).(pulumi.StringOutput)\n}", "func (o AllocationSpecificSKUAllocationReservedInstancePropertiesOutput) MachineType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v AllocationSpecificSKUAllocationReservedInstanceProperties) *string { return v.MachineType }).(pulumi.StringPtrOutput)\n}", "func (o TriggerBuildOptionsPtrOutput) MachineType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *TriggerBuildOptions) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.MachineType\n\t}).(pulumi.StringPtrOutput)\n}", "func (o TriggerBuildOptionsOutput) MachineType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v TriggerBuildOptions) *string { return v.MachineType }).(pulumi.StringPtrOutput)\n}", "func (o AllocationSpecificSKUAllocationReservedInstancePropertiesPtrOutput) MachineType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *AllocationSpecificSKUAllocationReservedInstanceProperties) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.MachineType\n\t}).(pulumi.StringPtrOutput)\n}", "func (o AllocationSpecificSKUAllocationReservedInstancePropertiesResponsePtrOutput) MachineType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *AllocationSpecificSKUAllocationReservedInstancePropertiesResponse) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.MachineType\n\t}).(pulumi.StringPtrOutput)\n}", "func (o InstancePropertiesOutput) MachineType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v InstanceProperties) *string { return v.MachineType }).(pulumi.StringPtrOutput)\n}", "func (a *AzureInfo) GetMachineTypes() (map[string]components.MachineType, error) {\n\treturn nil, constants.ErrorRequiredZone\n}", "func (o InstancePropertiesResponseOutput) MachineType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v InstancePropertiesResponse) string { return v.MachineType }).(pulumi.StringOutput)\n}", "func (o SourceInstancePropertiesResponsePtrOutput) MachineType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *SourceInstancePropertiesResponse) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.MachineType\n\t}).(pulumi.StringPtrOutput)\n}", "func (machine *VirtualMachine) GetType() string {\n\treturn \"Microsoft.Compute/virtualMachines\"\n}", "func (machine *VirtualMachine) GetType() string {\n\treturn \"Microsoft.Compute/virtualMachines\"\n}", "func (o InstancePropertiesPtrOutput) MachineType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *InstanceProperties) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.MachineType\n\t}).(pulumi.StringPtrOutput)\n}", "func (gc *GceCache) GetMachine(machineTypeName string, zone string) (MachineType, bool) {\n\tgc.cacheMutex.Lock()\n\tdefer gc.cacheMutex.Unlock()\n\n\tm, found := gc.machinesCache[MachineTypeKey{zone, machineTypeName}]\n\treturn m, found\n}", "func GetMachineTypes(orgId uint, secretId, location string) (response map[string]components.MachineType, err error) {\n\tclient, err := getAKSClient(orgId, secretId)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresponse = make(map[string]components.MachineType)\n\tresponse[location], err = azureClient.GetVmSizes(client, location)\n\n\treturn\n\n}", "func (o InstancePropertiesResponsePtrOutput) MachineType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *InstancePropertiesResponse) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.MachineType\n\t}).(pulumi.StringPtrOutput)\n}", "func LoadMachineForType(id int, t Type) Machine {\n\treturn nil\n}", "func GetMachineID() (m MachineID, err error) {\n\t// First grab all the interfaces\n\tifaces, err := net.Interfaces()\n\n\tif err != nil {\n\t\treturn\n\t}\n\n\tif len(ifaces) == 0 {\n\t\terr = fmt.Errorf(\"No network interfaces found!\")\n\t\treturn\n\t}\n\n\t// Build up list of all mac addresses\n\tmacs := make([]string, 0, len(ifaces))\n\n\tfor _, iface := range ifaces {\n\t\tstr := iface.HardwareAddr.String()\n\t\tif len(str) > 0 {\n\t\t\tmacs = append(macs, str)\n\t\t}\n\t}\n\n\t// Sort then and pick the first one\n\tsort.Strings(macs)\n\n\tm = MachineID(macs[0])\n\n\treturn\n}", "func (m *MachineRPC) MachineOfType(data shared.MachineRPCData, machines *[]shared.Machine) error {\n\tstart := time.Now()\n\n\tconn := Connections.Get(data.Channel)\n\n\t// Read the machines for the given site\n\terr := DB.SQL(MachinesOfType, data.ID).QueryStructs(machines)\n\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t}\n\n\t// For each machine, fetch all components\n\tfor k, m := range *machines {\n\t\t// fetch all components\n\t\tDB.SQL(`select\n\t\t* from component c\n\t\tleft join machine_type_tool x on x.id=c.mtt_id\n\t\twhere c.machine_id=$1\n\t\torder by x.position,c.zindex,lower(c.name)`, m.ID).\n\t\t\tQueryStructs(&(*machines)[k].Components)\n\n\t\t// err = DB.Select(\"*\").\n\t\t// \tFrom(\"component\").\n\t\t// \tWhere(\"machine_id = $1\", m.ID).\n\t\t// \tOrderBy(\"position,zindex,lower(name)\").\n\t\t// \tQueryStructs(&(*machines)[k].Components)\n\t}\n\n\tlogger(start, \"Machine.MachinesOfType\",\n\t\tfmt.Sprintf(\"Channel %d, Type %d, User %d %s %s\",\n\t\t\tdata.Channel, data.ID, conn.UserID, conn.Username, conn.UserRole),\n\t\tfmt.Sprintf(\"%d machines\", len(*machines)),\n\t\tdata.Channel, conn.UserID, \"machine\", 0, false)\n\n\treturn nil\n}", "func (s *OnDiskStateMachine) StateMachineType() pb.StateMachineType {\n\treturn pb.OnDiskStateMachine\n}", "func (tencentcloud *tencentCloudProvider) GetAvailableMachineTypes() ([]string, error) {\n\treturn []string{}, nil\n}", "func (aws *awsCloudProvider) GetAvailableMachineTypes() ([]string, error) {\n\treturn []string{}, nil\n}", "func (wmid *WzMachineIDUtil) GetMachineId() string {\n\treturn _MachineIDUtil.machineid\n}", "func GetMachine() (*Machine, error) {\n\t// dockerClient, err := client.NewClient(\"unix:///var/run/docker.sock\", \"1.13.1\", &http.Client{}, make(map[string]string))\n\tdockerClient, err := client.NewEnvClient()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Error creating Docker API Client :\\n%s\", err)\n\t}\n\tnewMachine := &Machine{dockerClient}\n\n\treturn newMachine, nil\n}", "func (i *Idb) GetMachine(fqdn string) (*machine.Machine, error) {\n\tfqdn = url.QueryEscape(fqdn)\n\tu := i.joinBaseURL(\"machines\")\n\n\tquery := url.Values{}\n\tquery.Add(\"fqdn\",fqdn)\n\tu.RawQuery = query.Encode()\n\n\trequest, err := http.NewRequest(\"GET\", u.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresponse, err := i.request(request)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar newMachine machine.Machine\n\tnewMachine.Fqdn = fqdn\n\n\tif response.StatusCode != http.StatusOK {\n\t\treturn nil, newErrStatus(response.StatusCode, http.StatusOK, &newMachine)\n\t}\n\n\terr = i.decodeResponse(&newMachine, response)\n\n\treturn &newMachine, err\n}", "func (u *Uname) Machine() string {\n\treturn charsToString(u.ub.Machine[:])\n}", "func (u *Uname) Machine() string {\n\treturn toString(u.Utsname.Machine[:])\n}", "func (ds *RegularStateMachineWrapper) StateMachineType() pb.StateMachineType {\n\treturn pb.RegularStateMachine\n}", "func (provider *cloudStackCloudProvider) GetAvailableMachineTypes() ([]string, error) {\n\treturn availableMachineTypes, nil\n}", "func (gce *Connection) ListMachineTypes(zone string) ([]MachineType, error) {\n machines, err := gce.raw.ListMachineTypes(gce.projectID, zone) <========================\n if err != nil {\n return nil, errors.Trace(err)\n }\n res := make([]MachineType, len(machines.Items))\n for i, machine := range machines.Items {\n deprecated := false\n if machine.Deprecated != nil {\n deprecated = machine.Deprecated.State != \"\"\n }\n res[i] = MachineType{\n CreationTimestamp: machine.CreationTimestamp,\n Deprecated: deprecated,\n Description: machine.Description,\n GuestCpus: machine.GuestCpus,\n Id: machine.Id,\n ImageSpaceGb: machine.ImageSpaceGb,\n Kind: machine.Kind,\n MaximumPersistentDisks: machine.MaximumPersistentDisks,\n MaximumPersistentDisksSizeGb: machine.MaximumPersistentDisksSizeGb,\n MemoryMb: machine.MemoryMb,\n Name: machine.Name,\n }\n }\n return res, nil\n}", "func (d *Driver) GetMachineName() string {\n\treturn d.MachineName\n}", "func (k *kamateraCloudProvider) GetAvailableMachineTypes() ([]string, error) {\n\treturn []string{}, cloudprovider.ErrNotImplemented\n}", "func GetMachineForName(envContainer app.EnvContainer, name string) (_ Machine, retErr error) {\n\tfilePath, err := GetFilePath(envContainer)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn GetMachineForNameAndFilePath(name, filePath)\n}", "func (c *Conn) GetMachine(name string) (dbus.ObjectPath, error) {\n\treturn c.getPath(\"GetMachine\", name)\n}", "func (k *Kind) MachineName() string {\n\treturn \"team\"\n}", "func (s *RegularStateMachine) StateMachineType() pb.StateMachineType {\n\treturn pb.RegularStateMachine\n}", "func (c *TestClient) GetMachineImage(project, name string) (*compute.MachineImage, error) {\n\tif c.GetMachineImageFn != nil {\n\t\treturn c.GetMachineImageFn(project, name)\n\t}\n\treturn c.client.GetMachineImage(project, name)\n}", "func (c *MachineImagesClient) GetMachineImage(getInput *GetMachineImageInput) (*MachineImage, error) {\n\tgetInput.Name = c.getQualifiedName(getInput.Name)\n\n\tvar machineImage MachineImage\n\tif err := c.getResource(getInput.Name, &machineImage); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn c.success(&machineImage)\n}", "func GetMachineId() (uint16, error) {\n\tmachineID = cast.ToUint16(conf.GetInt(\"SNOWFLAKE_MACHINE_ID\"))\n\treturn machineID, nil\n}", "func (s *ConcurrentStateMachine) StateMachineType() pb.StateMachineType {\n\treturn pb.ConcurrentStateMachine\n}", "func GetBuiltinProcessorType(programHash hashing.HashValue) (string, bool) {\n\tif _, err := core.GetProcessor(programHash); err == nil {\n\t\treturn core.VMType, true\n\t}\n\tif _, ok := native.GetProcessor(programHash); ok {\n\t\treturn native.VMType, true\n\t}\n\treturn \"\", false\n}", "func platformType() (string, error) {\n\taksNet, _ := hcsshim.GetHNSNetworkByName(\"azure\")\n\tif aksNet != nil {\n\t\treturn \"aks\", nil\n\t}\n\n\teksNet, _ := hcsshim.GetHNSNetworkByName(\"vpcbr*\")\n\tif eksNet != nil {\n\t\treturn \"eks\", nil\n\t}\n\n\t// EC2\n\tec2Resp, err := http.Get(\"http://169.254.169.254/latest/meta-data/local-hostname\")\n\tif err != nil && hasTimedOut(err) {\n\t\treturn \"\", err\n\t}\n\tif ec2Resp != nil {\n\t\tdefer ec2Resp.Body.Close()\n\t\tif ec2Resp.StatusCode == http.StatusOK {\n\t\t\treturn \"ec2\", nil\n\t\t}\n\t}\n\n\t// GCE\n\tclient := &http.Client{}\n\treq, err := http.NewRequest(\"GET\", \"http://metadata.google.internal/computeMetadata/v1/instance/hostname\", nil)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treq.Header.Add(\"Metadata-Flavor\", \"Google\")\n\tgceResp, err := client.Do(req)\n\tif err != nil && hasTimedOut(err) {\n\t\treturn \"\", err\n\t}\n\tif gceResp != nil {\n\t\tdefer gceResp.Body.Close()\n\t\tif gceResp.StatusCode == http.StatusOK {\n\t\t\treturn \"gce\", nil\n\t\t}\n\t}\n\n\treturn \"bare-metal\", nil\n}", "func getMachineID() string {\n\tmachineIDOnce.Do(func() {\n\t\tid, err := machineid.ProtectedID(\"dolt\")\n\n\t\tif err == nil {\n\t\t\tmachineID = id\n\t\t}\n\t})\n\n\treturn machineID\n}", "func (d *HetznerCloudProvider) GetAvailableMachineTypes() ([]string, error) {\n\tserverTypes, err := d.manager.cachedServerType.getAllServerTypes()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttypes := make([]string, len(serverTypes))\n\tfor _, server := range serverTypes {\n\t\ttypes = append(types, server.Name)\n\t}\n\n\treturn types, nil\n}", "func (x XID) Machine() []byte {\n\treturn x[4:7]\n}", "func (kc *KubeletClient) GetMachineCpuFrequency(host string) (uint64, error) {\n\tminfo, err := kc.GetMachineInfo(host)\n\tif err != nil {\n\t\tglog.Errorf(\"failed to get machine[%s] cpu.frequency: %v\", host, err)\n\t\treturn 0, err\n\t}\n\n\treturn minfo.CpuFrequency, nil\n}", "func (client *XenClient) SMGetType(self string) (result string, err error) {\n\tobj, err := client.APICall(\"SM.get_type\", self)\n\tif err != nil {\n\t\treturn\n\t}\n\tresult = obj.(string)\n\treturn\n}", "func GetMachineExtension(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *MachineExtensionState, opts ...pulumi.ResourceOption) (*MachineExtension, error) {\n\tvar resource MachineExtension\n\terr := ctx.ReadResource(\"azure-native:hybridcompute:MachineExtension\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func getMachineGUID() (string, error) {\n\tconst key = registry.LOCAL_MACHINE\n\tconst path = `SOFTWARE\\Microsoft\\Cryptography`\n\tconst name = \"MachineGuid\"\n\n\tk, err := registry.OpenKey(key, path, registry.READ|registry.WOW64_64KEY)\n\tif err != nil {\n\t\treturn \"\", errors.Wrapf(err, `failed to open HKLM\\%v`, path)\n\t}\n\n\tguid, _, err := k.GetStringValue(name)\n\tif err != nil {\n\t\treturn \"\", errors.Wrapf(err, `failed to get value of HKLM\\%v\\%v`, path, name)\n\t}\n\n\treturn guid, nil\n}", "func Machine(id ID) []byte {\n\treturn dgen.Machine(id)\n}", "func (process *process) getStateMachine(stateName string, tmpRow []interface{}) *statemachine.MachineState {\n\nStart:\n\tfor name, machine := range process.machine {\n\t\tif name == stateName {\n\t\t\treturn machine\n\t\t}\n\t}\n\tprocess.findStateAndAddToMachine(stateName, tmpRow)\n\tgoto Start\n\n}", "func (id ID) Machine() []byte {\n\treturn id[4:7]\n}", "func (rc *rawConn) ListMachineTypes(projectID, zone string) (*compute.MachineTypeList, error) { \n op := rc.MachineTypes.List(projectID, zone) <==============\n machines, err := op.Do() <==============\n if err != nil {\n return nil, errors.Annotatef(err, \"listing machine types for project %q and zone %q\", projectID, zone)\n }\n return machines, nil\n}", "func (dm *DockerMachine) Machine(name string) (Machine, error) {\n\tmachines, err := dm.Machines()\n\tif err != nil {\n\t\treturn Machine{}, err\n\t}\n\tfor _, machine := range machines {\n\t\tif machine.Name == name {\n\t\t\treturn machine, nil\n\t\t}\n\t}\n\treturn Machine{}, fmt.Errorf(\"machine %s not found\", name)\n}", "func (o *ShowSystem) GetType() string {\n\tif o == nil || IsNil(o.Type) {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Type\n}", "func GetMachine(c *gin.Context) {\n\n\tsn := c.Param(\"sn\")\n\tdata, err := service.GetMachine(sn)\n\tif err != nil {\n\t\tapi.Error(c, http.StatusInternalServerError, \"server error\")\n\t\treturn\n\t}\n\tif data.SerialNumber == \"\" { //没找到\n\t\tapi.Fail(c, \"machine not found\")\n\t\treturn\n\t}\n\tapi.Success(c, data, \"\")\n}", "func (s *MultipassServer) GetAvailableMachineTypes(ctx context.Context, request *apigrpc.CloudProviderServiceRequest) (*apigrpc.AvailableMachineTypesReply, error) {\n\tglog.V(5).Infof(\"Call server GetAvailableMachineTypes: %v\", request)\n\n\tif s.Configuration.Optionals.GetAvailableMachineTypes {\n\t\treturn nil, fmt.Errorf(errNotImplemented)\n\t}\n\n\tif request.GetProviderID() != s.Configuration.ProviderID {\n\t\tglog.Errorf(errMismatchingProvider)\n\t\treturn nil, fmt.Errorf(errMismatchingProvider)\n\t}\n\n\tmachineTypes := make([]string, 0, len(s.Configuration.Machines))\n\n\tfor n := range s.Configuration.Machines {\n\t\tmachineTypes = append(machineTypes, n)\n\t}\n\n\treturn &apigrpc.AvailableMachineTypesReply{\n\t\tResponse: &apigrpc.AvailableMachineTypesReply_AvailableMachineTypes{\n\t\t\tAvailableMachineTypes: &apigrpc.AvailableMachineTypes{\n\t\t\t\tMachineType: machineTypes,\n\t\t\t},\n\t\t},\n\t}, nil\n}", "func (*Tfdt) GetType() BoxType {\n\treturn BoxTypeTfdt()\n}", "func (gc *GceCache) GetMachineFromCache(machineType string, zone string) *gce.MachineType {\n\tgc.cacheMutex.Lock()\n\tdefer gc.cacheMutex.Unlock()\n\n\treturn gc.machinesCache[MachineTypeKey{zone, machineType}]\n}", "func (m *WindowsInformationProtectionDeviceRegistration) GetDeviceType()(*string) {\n val, err := m.GetBackingStore().Get(\"deviceType\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (c *TestClient) ListMachineTypes(project, zone string, opts ...ListCallOption) ([]*compute.MachineType, error) {\n\tif c.ListMachineTypesFn != nil {\n\t\treturn c.ListMachineTypesFn(project, zone, opts...)\n\t}\n\treturn c.client.ListMachineTypes(project, zone, opts...)\n}", "func Win32LobAppMsiPackageTypePPerMachine() *Win32LobAppMsiPackageType {\n\tv := Win32LobAppMsiPackageTypeVPerMachine\n\treturn &v\n}", "func (r Virtual_Guest) GetType() (resp datatypes.Virtual_Guest_Type, err error) {\n\terr = r.Session.DoRequest(\"SoftLayer_Virtual_Guest\", \"getType\", nil, &r.Options, &resp)\n\treturn\n}", "func IsValidMachineType(machineType byte) bool {\n\treturn machineType == Rolling || machineType == Cutting || machineType == Folding || machineType == Packing\n}", "func (a *AzureInfo) GetMachineTypesWithFilter(filter *components.InstanceFilter) (map[string]components.MachineType, error) {\n\n\tif len(filter.Zone) == 0 {\n\t\treturn nil, constants.ErrorRequiredZone\n\t}\n\n\treturn cluster.GetMachineTypes(a.OrgId, a.SecretId, filter.Zone)\n}", "func GetMachineFromNode(client runtimeclient.Client, node *corev1.Node) (*mapiv1beta1.Machine, error) {\n\tmachineNamespaceKey, ok := node.Annotations[controllernode.MachineAnnotationKey]\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"node %q does not have a MachineAnnotationKey %q\", node.Name, controllernode.MachineAnnotationKey)\n\t}\n\tnamespace, machineName, err := cache.SplitMetaNamespaceKey(machineNamespaceKey)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"machine annotation format is incorrect %v: %v\", machineNamespaceKey, err)\n\t}\n\n\tif namespace != TestContext.MachineApiNamespace {\n\t\treturn nil, fmt.Errorf(\"Machine %q is forbidden to live outside of default %v namespace\", machineNamespaceKey, TestContext.MachineApiNamespace)\n\t}\n\n\tmachine, err := GetMachine(context.TODO(), client, machineName)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error querying api for machine object: %v\", err)\n\t}\n\n\treturn machine, nil\n}", "func (g *Generator) Machine(id ID) []byte {\n\tif g.mode == Sequential {\n\t\treturn id[4:7]\n\t} else {\n\t\treturn id[3:6]\n\t}\n}", "func HostOsTypeEnumFromValue(value string) HostOsTypeEnum {\r\n switch value {\r\n case \"kLinux\":\r\n return HostOsType_KLINUX\r\n case \"kWindows\":\r\n return HostOsType_KWINDOWS\r\n case \"kAix\":\r\n return HostOsType_KAIX\r\n case \"kSolaris\":\r\n return HostOsType_KSOLARIS\r\n default:\r\n return HostOsType_KLINUX\r\n }\r\n}", "func (m *MachineScope) GetVMImage(ctx context.Context) (*infrav1.Image, error) {\n\tctx, log, done := tele.StartSpanWithLogger(ctx, \"scope.MachineScope.GetVMImage\")\n\tdefer done()\n\n\t// Use custom Marketplace image, Image ID or a Shared Image Gallery image if provided\n\tif m.AzureMachine.Spec.Image != nil {\n\t\treturn m.AzureMachine.Spec.Image, nil\n\t}\n\n\tsvc := virtualmachineimages.New(m)\n\n\tif m.AzureMachine.Spec.OSDisk.OSType == azure.WindowsOS {\n\t\truntime := m.AzureMachine.Annotations[\"runtime\"]\n\t\twindowsServerVersion := m.AzureMachine.Annotations[\"windowsServerVersion\"]\n\t\tlog.Info(\"No image specified for machine, using default Windows Image\", \"machine\", m.AzureMachine.GetName(), \"runtime\", runtime, \"windowsServerVersion\", windowsServerVersion)\n\t\treturn svc.GetDefaultWindowsImage(ctx, m.Location(), ptr.Deref(m.Machine.Spec.Version, \"\"), runtime, windowsServerVersion)\n\t}\n\n\tlog.Info(\"No image specified for machine, using default Linux Image\", \"machine\", m.AzureMachine.GetName())\n\treturn svc.GetDefaultUbuntuImage(ctx, m.Location(), ptr.Deref(m.Machine.Spec.Version, \"\"))\n}", "func GetMachineIfExists(c client.Client, namespace, name string) (*clusterv1.Machine, error) {\n\tif c == nil {\n\t\t// Being called before k8s is setup as part of control plane VM creation\n\t\treturn nil, nil\n\t}\n\n\t// Machines are identified by name\n\tmachine := &clusterv1.Machine{}\n\terr := c.Get(context.Background(), client.ObjectKey{Namespace: namespace, Name: name}, machine)\n\tif err != nil {\n\t\tif apierrors.IsNotFound(err) {\n\t\t\treturn nil, nil\n\t\t}\n\t\treturn nil, err\n\t}\n\n\treturn machine, nil\n}", "func (a *Actuator) getNodeForMachine(c *clusterv1.Cluster, m *clusterv1.Machine) (string, string, error) {\n\tmasterSSHClient, err := a.getMasterSSHClient(c, m)\n\tif err != nil {\n\t\tglog.Error(\"Error getting master sshClient\")\n\t\treturn \"\", \"\", err\n\t}\n\tnodeCmd := getNodeCmd + \" | grep \" + m.Namespace + \"/\" + m.Name\n\tglog.Infof(\"nodeCmd = %s\", nodeCmd)\n\toutput, err := masterSSHClient.ProcessCMDWithOutput(nodeCmd)\n\tif err != nil {\n\t\tglog.Errorf(\"Error getting node: cmd = %s, error = %s\", nodeCmd, err)\n\t\treturn \"\", \"\", err\n\t}\n\tstrs := strings.Split(string(output), \":\")\n\tif len(strs) < 2 {\n\t\treturn \"\", \"\", errors.New(\"Error getting node name for machine\")\n\t}\n\tnode := strs[0]\n\tversion := a.semanticVersion(strs[1])\n\treturn node, version, nil\n}", "func (myOperatingSystemType *OperatingSystemType) Type() (param string) {\n\treturn myOperatingSystemType.Typevar\n}", "func (m *RegistryKeyState) GetValueType()(*RegistryValueType) {\n return m.valueType\n}", "func GetType(entropyName string) uint32 {\n\tswitch strings.ToUpper(entropyName) {\n\n\tcase \"HUFFMAN\":\n\t\treturn HUFFMAN_TYPE\n\n\tcase \"ANS0\":\n\t\treturn ANS0_TYPE\n\n\tcase \"ANS1\":\n\t\treturn ANS1_TYPE\n\n\tcase \"RANGE\":\n\t\treturn RANGE_TYPE\n\n\tcase \"FPAQ\":\n\t\treturn FPAQ_TYPE\n\n\tcase \"CM\":\n\t\treturn CM_TYPE\n\n\tcase \"TPAQ\":\n\t\treturn TPAQ_TYPE\n\n\tcase \"TPAQX\":\n\t\treturn TPAQX_TYPE\n\n\tcase \"NONE\":\n\t\treturn NONE_TYPE\n\n\tdefault:\n\t\tpanic(fmt.Errorf(\"Unsupported entropy codec type: '%s'\", entropyName))\n\t}\n}", "func (nmd *NetMethodDispatch) GetType(key reflect.Type) NetMethodFun {\n\tnmd.sync.RLock()\n\tdefer nmd.sync.RUnlock()\n\tf, success := nmd.m[key]\n\tif !success {\n\t\treturn nil\n\t}\n\treturn f\n}", "func (ms *MachinePlugin) GetMachineStatus(ctx context.Context, req *cmi.GetMachineStatusRequest) (*cmi.GetMachineStatusResponse, error) {\n\t// Log messages to track start and end of request\n\tglog.V(2).Infof(\"Get request has been recieved for %q\", req.MachineName)\n\n\tproviderSpec, secrets, err := decodeProviderSpecAndSecret(req.ProviderSpec, req.Secrets, false)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tinstances, err := ms.getInstancesFromMachineName(req.MachineName, providerSpec, secrets)\n\tif err != nil {\n\t\treturn nil, err\n\t} else if len(instances) > 1 {\n\t\tinstanceIDs := []string{}\n\t\tfor _, instance := range instances {\n\t\t\tinstanceIDs = append(instanceIDs, *instance.InstanceId)\n\t\t}\n\n\t\terrMessage := fmt.Sprintf(\"AWS plugin is returning multiple VM instances backing this machine object. IDs for all backing VMs - %v \", instanceIDs)\n\t\treturn nil, status.Error(codes.OutOfRange, errMessage)\n\t}\n\n\trequiredInstance := instances[0]\n\n\tresponse := &cmi.GetMachineStatusResponse{\n\t\tNodeName: *requiredInstance.PrivateDnsName,\n\t\tProviderID: encodeProviderID(providerSpec.Region, *requiredInstance.InstanceId),\n\t}\n\n\tglog.V(2).Infof(\"Machine get request has been processed successfully for %q\", req.MachineName)\n\treturn response, nil\n}", "func LookupStateMachine(ctx *pulumi.Context, args *LookupStateMachineArgs, opts ...pulumi.InvokeOption) (*LookupStateMachineResult, error) {\n\topts = internal.PkgInvokeDefaultOpts(opts)\n\tvar rv LookupStateMachineResult\n\terr := ctx.Invoke(\"aws-native:stepfunctions:getStateMachine\", args, &rv, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &rv, nil\n}", "func GetMachineStatus() (Machine, error) {\n\tdb, err := sql.Open(\"sqlite3\", config.GetPath(config.AppConfig.DBFile)+\"?mode=ro\")\n\tif err != nil {\n\t\treturn Machine{}, err\n\t}\n\tdefer db.Close()\n\n\tlatitude, longitude, err := getCoordinates(db)\n\tif err != nil {\n\t\treturn Machine{}, err\n\t}\n\n\tmeasures, err := getMeasures(db)\n\tif err != nil {\n\t\treturn Machine{}, err\n\t}\n\n\tsensors, err := getSensors(db)\n\tif err != nil {\n\t\treturn Machine{}, err\n\t}\n\n\tboards, err := getBoards(db)\n\tif err != nil {\n\t\treturn Machine{}, err\n\t}\n\n\tfor i, s := range sensors {\n\t\tsensors[i].Measures, _ = measures[s.Channel]\n\n\t\tfor j, b := range boards {\n\t\t\tif b.ID == s.BoardID {\n\t\t\t\tboards[j].Sensors = append(boards[j].Sensors, sensors[i])\n\t\t\t}\n\t\t}\n\t}\n\n\treturn Machine{\n\t\tLatitude: latitude,\n\t\tLongitude: longitude,\n\t\tBoards: boards,\n\t}, nil\n}", "func NewMachine(node node.Inf, resource resource.Inf, logger *log.Logger) Inf {\n\treturn &machine{node: node, resource: resource, logger: logger}\n}", "func WindowsType() {\n\tk, err := registry.OpenKey(registry.LOCAL_MACHINE, `SOFTWARE\\Microsoft\\Windows NT\\CurrentVersion`, registry.QUERY_VALUE)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdefer k.Close()\n\n\ts, _, err := k.GetStringValue(\"EditionID\")\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tif strings.Contains(s, \"Server\") {\n\t\tWindowsServerChecks()\n\t} else {\n\t\tWindowsWorkstationChecks()\n\t}\n}", "func (m *Win32LobAppRegistryDetection) GetDetectionType()(*Win32LobAppRegistryDetectionType) {\n val, err := m.GetBackingStore().Get(\"detectionType\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*Win32LobAppRegistryDetectionType)\n }\n return nil\n}", "func (m *MicrosoftManagedDesktop) GetTypeEscaped()(*MicrosoftManagedDesktopType) {\n val, err := m.GetBackingStore().Get(\"typeEscaped\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*MicrosoftManagedDesktopType)\n }\n return nil\n}", "func (c *UPHostClient) NewDescribePHostMachineTypeRequest() *DescribePHostMachineTypeRequest {\n\treq := &DescribePHostMachineTypeRequest{}\n\n\t// setup request with client config\n\tc.Client.SetupRequest(req)\n\n\t// setup retryable with default retry policy (retry for non-create action and common error)\n\treq.SetRetryable(true)\n\treturn req\n}", "func MachineCRDCreator() reconciling.NamedCustomResourceDefinitionCreatorGetter {\n\treturn func() (string, reconciling.CustomResourceDefinitionCreator) {\n\t\treturn resources.MachineCRDName, func(crd *apiextensionsv1beta1.CustomResourceDefinition) (*apiextensionsv1beta1.CustomResourceDefinition, error) {\n\t\t\tcrd.Spec.Group = clusterAPIGroup\n\t\t\tcrd.Spec.Version = clusterAPIVersion\n\t\t\tcrd.Spec.Scope = apiextensionsv1beta1.NamespaceScoped\n\t\t\tcrd.Spec.Names.Kind = \"Machine\"\n\t\t\tcrd.Spec.Names.ListKind = \"MachineList\"\n\t\t\tcrd.Spec.Names.Plural = \"machines\"\n\t\t\tcrd.Spec.Names.Singular = \"machine\"\n\t\t\tcrd.Spec.Names.ShortNames = []string{\"ma\"}\n\t\t\tcrd.Spec.AdditionalPrinterColumns = []apiextensionsv1beta1.CustomResourceColumnDefinition{\n\t\t\t\t{\n\t\t\t\t\tName: \"Age\",\n\t\t\t\t\tType: \"date\",\n\t\t\t\t\tJSONPath: \".metadata.creationTimestamp\",\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tName: \"Deleted\",\n\t\t\t\t\tType: \"date\",\n\t\t\t\t\tJSONPath: \".metadata.deletionTimestamp\",\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tName: \"MachineSet\",\n\t\t\t\t\tType: \"string\",\n\t\t\t\t\tJSONPath: \".metadata.ownerReferences[0].name\",\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tName: \"Address\",\n\t\t\t\t\tType: \"string\",\n\t\t\t\t\tJSONPath: \".status.addresses[0].address\",\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tName: \"Node\",\n\t\t\t\t\tType: \"string\",\n\t\t\t\t\tJSONPath: \".status.nodeRef.name\",\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tName: \"Provider\",\n\t\t\t\t\tType: \"string\",\n\t\t\t\t\tJSONPath: \".spec.providerSpec.value.cloudProvider\",\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tName: \"OS\",\n\t\t\t\t\tType: \"string\",\n\t\t\t\t\tJSONPath: \".spec.providerSpec.value.operatingSystem\",\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tName: \"Version\",\n\t\t\t\t\tType: \"string\",\n\t\t\t\t\tJSONPath: \".spec.versions.kubelet\",\n\t\t\t\t},\n\t\t\t}\n\n\t\t\treturn crd, nil\n\t\t}\n\t}\n\n}", "func PhysicalServerHostTypesEnumFromValue(value string) PhysicalServerHostTypesEnum {\r\n switch value {\r\n case \"kLinux\":\r\n return PhysicalServerHostTypes_KLINUX\r\n case \"kWindows\":\r\n return PhysicalServerHostTypes_KWINDOWS\r\n case \"kAix\":\r\n return PhysicalServerHostTypes_KAIX\r\n case \"kSolaris\":\r\n return PhysicalServerHostTypes_KSOLARIS\r\n default:\r\n return PhysicalServerHostTypes_KLINUX\r\n }\r\n}", "func GetMachineFromNode(ctx context.Context, c client.Client, nodeName string) (*clusterv1.Machine, error) {\n\tmachineList := &clusterv1.MachineList{}\n\tif err := c.List(\n\t\tctx,\n\t\tmachineList,\n\t\tclient.MatchingFields{clusterv1.MachineNodeNameIndex: nodeName},\n\t); err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed getting machine list\")\n\t}\n\t// TODO(vincepri): Remove this loop once controller runtime fake client supports\n\t// adding indexes on objects.\n\titems := []*clusterv1.Machine{}\n\tfor i := range machineList.Items {\n\t\tmachine := &machineList.Items[i]\n\t\tif machine.Status.NodeRef != nil && machine.Status.NodeRef.Name == nodeName {\n\t\t\titems = append(items, machine)\n\t\t}\n\t}\n\tif len(items) != 1 {\n\t\treturn nil, errors.Errorf(\"expecting one machine for node %v, got %v\", nodeName, machineNames(items))\n\t}\n\treturn items[0], nil\n}", "func (o AzureMachineLearningServiceFunctionBindingOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v AzureMachineLearningServiceFunctionBinding) string { return v.Type }).(pulumi.StringOutput)\n}", "func GetVirtualGuestBootParameterTypeService(sess *session.Session) Virtual_Guest_Boot_Parameter_Type {\n\treturn Virtual_Guest_Boot_Parameter_Type{Session: sess}\n}", "func (o AzureMachineLearningStudioFunctionBindingOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v AzureMachineLearningStudioFunctionBinding) string { return v.Type }).(pulumi.StringOutput)\n}", "func (m *MdiaBox) Type() string {\n\treturn \"mdia\"\n}", "func (m *kubeGenericRuntimeManager) Type() string {\n\tif runtime, err := m.runtimeRegistry.GetPrimaryRuntimeService(); err == nil {\n\t\treturn m.RuntimeType(runtime.ServiceApi)\n\t}\n\treturn \"unknownType\"\n}", "func (c *ControlPlaneContract) MachineTemplate() *ControlPlaneMachineTemplate {\n\treturn &ControlPlaneMachineTemplate{}\n}", "func (wzid WzMachineIDUtilConsumer) GetMachineIdUtil() *WzMachineIDUtil {\n\tif _MachineIDUtil == nil {\n\t\tpanic(\"MachineID utility was not properly initialised yet\")\n\t}\n\n\treturn _MachineIDUtil\n}", "func GetMachineByName(ctx context.Context, c client.Client, namespace, name string) (*machinev1.Machine, error) {\n\tm := &machinev1.Machine{}\n\tkey := client.ObjectKey{Name: name, Namespace: namespace}\n\tif err := c.Get(ctx, key, m); err != nil {\n\t\treturn nil, err\n\t}\n\treturn m, nil\n}" ]
[ "0.70354706", "0.6988988", "0.68864304", "0.670869", "0.6683729", "0.66089046", "0.6592785", "0.65553164", "0.65475166", "0.65143734", "0.64995205", "0.64739525", "0.6431266", "0.64054203", "0.63845634", "0.63837296", "0.6315042", "0.6315042", "0.62513", "0.6221095", "0.61845624", "0.61602604", "0.6142231", "0.61119413", "0.60704404", "0.60616547", "0.60164165", "0.59963405", "0.5959108", "0.59391856", "0.59362257", "0.59328336", "0.59212804", "0.59069383", "0.58690625", "0.5857934", "0.58105487", "0.5809074", "0.5720063", "0.56966615", "0.5647513", "0.56418544", "0.5548468", "0.5510037", "0.5509027", "0.54942566", "0.54700726", "0.54630226", "0.54471135", "0.5447037", "0.5440751", "0.5406918", "0.5384342", "0.5375728", "0.5373552", "0.5354237", "0.53425515", "0.53416014", "0.5321115", "0.53019893", "0.52871156", "0.5254507", "0.5239465", "0.52087396", "0.5178405", "0.5172884", "0.51695263", "0.51657534", "0.51547086", "0.51272017", "0.51236784", "0.5122867", "0.5117464", "0.5107725", "0.50894207", "0.5061663", "0.5050389", "0.504511", "0.50359577", "0.50354445", "0.5026673", "0.5018827", "0.50037676", "0.49813998", "0.49772137", "0.4974148", "0.4969381", "0.49460664", "0.49330637", "0.4926355", "0.49255058", "0.49249873", "0.49073896", "0.4900487", "0.48671892", "0.48630473", "0.48515567", "0.48455325", "0.48223075", "0.48168457" ]
0.75291127
0
ListMachineTypes uses the override method ListMachineTypesFn or the real implementation.
func (c *TestClient) ListMachineTypes(project, zone string, opts ...ListCallOption) ([]*compute.MachineType, error) { if c.ListMachineTypesFn != nil { return c.ListMachineTypesFn(project, zone, opts...) } return c.client.ListMachineTypes(project, zone, opts...) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (gce *Connection) ListMachineTypes(zone string) ([]MachineType, error) {\n machines, err := gce.raw.ListMachineTypes(gce.projectID, zone) <========================\n if err != nil {\n return nil, errors.Trace(err)\n }\n res := make([]MachineType, len(machines.Items))\n for i, machine := range machines.Items {\n deprecated := false\n if machine.Deprecated != nil {\n deprecated = machine.Deprecated.State != \"\"\n }\n res[i] = MachineType{\n CreationTimestamp: machine.CreationTimestamp,\n Deprecated: deprecated,\n Description: machine.Description,\n GuestCpus: machine.GuestCpus,\n Id: machine.Id,\n ImageSpaceGb: machine.ImageSpaceGb,\n Kind: machine.Kind,\n MaximumPersistentDisks: machine.MaximumPersistentDisks,\n MaximumPersistentDisksSizeGb: machine.MaximumPersistentDisksSizeGb,\n MemoryMb: machine.MemoryMb,\n Name: machine.Name,\n }\n }\n return res, nil\n}", "func (rc *rawConn) ListMachineTypes(projectID, zone string) (*compute.MachineTypeList, error) { \n op := rc.MachineTypes.List(projectID, zone) <==============\n machines, err := op.Do() <==============\n if err != nil {\n return nil, errors.Annotatef(err, \"listing machine types for project %q and zone %q\", projectID, zone)\n }\n return machines, nil\n}", "func (tencentcloud *tencentCloudProvider) GetAvailableMachineTypes() ([]string, error) {\n\treturn []string{}, nil\n}", "func (aws *awsCloudProvider) GetAvailableMachineTypes() ([]string, error) {\n\treturn []string{}, nil\n}", "func (k *kamateraCloudProvider) GetAvailableMachineTypes() ([]string, error) {\n\treturn []string{}, cloudprovider.ErrNotImplemented\n}", "func (provider *cloudStackCloudProvider) GetAvailableMachineTypes() ([]string, error) {\n\treturn availableMachineTypes, nil\n}", "func (a *AzureInfo) GetMachineTypes() (map[string]components.MachineType, error) {\n\treturn nil, constants.ErrorRequiredZone\n}", "func GetMachineTypes(orgId uint, secretId, location string) (response map[string]components.MachineType, err error) {\n\tclient, err := getAKSClient(orgId, secretId)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresponse = make(map[string]components.MachineType)\n\tresponse[location], err = azureClient.GetVmSizes(client, location)\n\n\treturn\n\n}", "func (d *HetznerCloudProvider) GetAvailableMachineTypes() ([]string, error) {\n\tserverTypes, err := d.manager.cachedServerType.getAllServerTypes()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttypes := make([]string, len(serverTypes))\n\tfor _, server := range serverTypes {\n\t\ttypes = append(types, server.Name)\n\t}\n\n\treturn types, nil\n}", "func (s *MultipassServer) GetAvailableMachineTypes(ctx context.Context, request *apigrpc.CloudProviderServiceRequest) (*apigrpc.AvailableMachineTypesReply, error) {\n\tglog.V(5).Infof(\"Call server GetAvailableMachineTypes: %v\", request)\n\n\tif s.Configuration.Optionals.GetAvailableMachineTypes {\n\t\treturn nil, fmt.Errorf(errNotImplemented)\n\t}\n\n\tif request.GetProviderID() != s.Configuration.ProviderID {\n\t\tglog.Errorf(errMismatchingProvider)\n\t\treturn nil, fmt.Errorf(errMismatchingProvider)\n\t}\n\n\tmachineTypes := make([]string, 0, len(s.Configuration.Machines))\n\n\tfor n := range s.Configuration.Machines {\n\t\tmachineTypes = append(machineTypes, n)\n\t}\n\n\treturn &apigrpc.AvailableMachineTypesReply{\n\t\tResponse: &apigrpc.AvailableMachineTypesReply_AvailableMachineTypes{\n\t\t\tAvailableMachineTypes: &apigrpc.AvailableMachineTypes{\n\t\t\t\tMachineType: machineTypes,\n\t\t\t},\n\t\t},\n\t}, nil\n}", "func (a *AzureInfo) GetMachineTypesWithFilter(filter *components.InstanceFilter) (map[string]components.MachineType, error) {\n\n\tif len(filter.Zone) == 0 {\n\t\treturn nil, constants.ErrorRequiredZone\n\t}\n\n\treturn cluster.GetMachineTypes(a.OrgId, a.SecretId, filter.Zone)\n}", "func (m *MachineRPC) MachineOfType(data shared.MachineRPCData, machines *[]shared.Machine) error {\n\tstart := time.Now()\n\n\tconn := Connections.Get(data.Channel)\n\n\t// Read the machines for the given site\n\terr := DB.SQL(MachinesOfType, data.ID).QueryStructs(machines)\n\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t}\n\n\t// For each machine, fetch all components\n\tfor k, m := range *machines {\n\t\t// fetch all components\n\t\tDB.SQL(`select\n\t\t* from component c\n\t\tleft join machine_type_tool x on x.id=c.mtt_id\n\t\twhere c.machine_id=$1\n\t\torder by x.position,c.zindex,lower(c.name)`, m.ID).\n\t\t\tQueryStructs(&(*machines)[k].Components)\n\n\t\t// err = DB.Select(\"*\").\n\t\t// \tFrom(\"component\").\n\t\t// \tWhere(\"machine_id = $1\", m.ID).\n\t\t// \tOrderBy(\"position,zindex,lower(name)\").\n\t\t// \tQueryStructs(&(*machines)[k].Components)\n\t}\n\n\tlogger(start, \"Machine.MachinesOfType\",\n\t\tfmt.Sprintf(\"Channel %d, Type %d, User %d %s %s\",\n\t\t\tdata.Channel, data.ID, conn.UserID, conn.Username, conn.UserRole),\n\t\tfmt.Sprintf(\"%d machines\", len(*machines)),\n\t\tdata.Channel, conn.UserID, \"machine\", 0, false)\n\n\treturn nil\n}", "func (fc *FleetClient) ListMachines() ([]machine.MachineState, error) {\n\treturn fc.api.Machines()\n}", "func (c *TestClient) ListMachineImages(project string, opts ...ListCallOption) ([]*compute.MachineImage, error) {\n\tif c.ListMachineImagesFn != nil {\n\t\treturn c.ListMachineImagesFn(project, opts...)\n\t}\n\treturn c.client.ListMachineImages(project, opts...)\n}", "func (s *LocalTests) listMachines(c *gc.C, filter *cloudapi.Filter) {\n\tvar contains bool\n\ttestMachine := s.createMachine(c)\n\tdefer s.deleteMachine(c, testMachine.Id)\n\n\tmachines, err := s.testClient.ListMachines(filter)\n\tc.Assert(err, gc.IsNil)\n\tc.Assert(machines, gc.NotNil)\n\tfor _, m := range machines {\n\t\tif m.Id == testMachine.Id {\n\t\t\tcontains = true\n\t\t\tbreak\n\t\t}\n\t}\n\n\t// result\n\tif !contains {\n\t\tc.Fatalf(\"Obtained machines [%v] do not contain test machine [%v]\", machines, *testMachine)\n\t}\n}", "func (client GroupClient) ListTypesResponder(resp *http.Response) (result USQLTypeList, err error) {\n\terr = autorest.Respond(\n\t\tresp,\n\t\tclient.ByInspecting(),\n\t\tazure.WithErrorUnlessStatusCode(http.StatusOK),\n\t\tautorest.ByUnmarshallingJSON(&result),\n\t\tautorest.ByClosing())\n\tresult.Response = autorest.Response{Response: resp}\n\treturn\n}", "func (ms *MachinePlugin) ListMachines(ctx context.Context, req *cmi.ListMachinesRequest) (*cmi.ListMachinesResponse, error) {\n\t// Log messages to track start and end of request\n\tglog.V(2).Infof(\"List machines request has been recieved for %q\", req.ProviderSpec)\n\n\tproviderSpec, secrets, err := decodeProviderSpecAndSecret(req.ProviderSpec, req.Secrets, true)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tclusterName := \"\"\n\tnodeRole := \"\"\n\n\tfor key := range providerSpec.Tags {\n\t\tif strings.Contains(key, \"kubernetes.io/cluster/\") {\n\t\t\tclusterName = key\n\t\t} else if strings.Contains(key, \"kubernetes.io/role/\") {\n\t\t\tnodeRole = key\n\t\t}\n\t}\n\n\tsvc, err := ms.createSVC(secrets, providerSpec.Region)\n\tif err != nil {\n\t\treturn nil, status.Error(codes.Internal, err.Error())\n\t}\n\n\tinput := ec2.DescribeInstancesInput{\n\t\tFilters: []*ec2.Filter{\n\t\t\t&ec2.Filter{\n\t\t\t\tName: aws.String(\"tag-key\"),\n\t\t\t\tValues: []*string{\n\t\t\t\t\t&clusterName,\n\t\t\t\t},\n\t\t\t},\n\t\t\t&ec2.Filter{\n\t\t\t\tName: aws.String(\"tag-key\"),\n\t\t\t\tValues: []*string{\n\t\t\t\t\t&nodeRole,\n\t\t\t\t},\n\t\t\t},\n\t\t\t&ec2.Filter{\n\t\t\t\tName: aws.String(\"instance-state-name\"),\n\t\t\t\tValues: []*string{\n\t\t\t\t\taws.String(\"pending\"),\n\t\t\t\t\taws.String(\"running\"),\n\t\t\t\t\taws.String(\"stopping\"),\n\t\t\t\t\taws.String(\"stopped\"),\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\n\trunResult, err := svc.DescribeInstances(&input)\n\tif err != nil {\n\t\tglog.Errorf(\"AWS plugin is returning error while describe instances request is sent: %s\", err)\n\t\treturn nil, status.Error(codes.Internal, err.Error())\n\t}\n\n\tlistOfVMs := make(map[string]string)\n\tfor _, reservation := range runResult.Reservations {\n\t\tfor _, instance := range reservation.Instances {\n\n\t\t\tmachineName := \"\"\n\t\t\tfor _, tag := range instance.Tags {\n\t\t\t\tif *tag.Key == \"Name\" {\n\t\t\t\t\tmachineName = *tag.Value\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\tlistOfVMs[encodeProviderID(providerSpec.Region, *instance.InstanceId)] = machineName\n\t\t}\n\t}\n\n\tglog.V(2).Infof(\"List machines request has been processed successfully\")\n\t// Core logic ends here.\n\tResp := &cmi.ListMachinesResponse{\n\t\tMachineList: listOfVMs,\n\t}\n\treturn Resp, nil\n}", "func (o LookupInstanceResultOutput) MachineType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupInstanceResult) string { return v.MachineType }).(pulumi.StringOutput)\n}", "func MachineListCommand(c *cli.Context, log logging.Logger, _ string) (int, error) {\n\topts := &machine.ListOptions{\n\t\tLog: log.New(\"machine:list\"),\n\t}\n\n\tinfos, err := machine.List(opts)\n\tif err != nil {\n\t\treturn 1, err\n\t}\n\n\ttabFormatter(os.Stdout, infos)\n\treturn 0, nil\n}", "func MachineType(name string) string {\n\tif IsKIC(name) {\n\t\treturn \"container\"\n\t}\n\n\tif IsSSH(name) {\n\t\treturn \"bare metal machine\"\n\t}\n\n\tif IsVM(name) {\n\t\treturn \"VM\"\n\t}\n\n\t// none or mock\n\treturn \"bare metal machine\"\n}", "func ListTypes(c echo.Context) error {\n\tlogger.Info.Info(\"Preparing to list types\")\n\tresult := new(struct {\n\t\tData interface{} `json:\"data\"`\n\t\tTotal int `json:\"total\"`\n\t})\n\ttypes, total, err := controller.ListTypes()\n\tif err != nil {\n\t\tlogger.Error.Error(\"Error while list of types\", err)\n\t\treturn echo.NewHTTPError(http.StatusNotAcceptable, \"Error while list of types\")\n\t}\n\tresult.Data = types\n\tresult.Total = total\n\tlogger.Success.Info(\"Types listed successfully\")\n\treturn c.JSON(http.StatusOK, result)\n}", "func (a *Client) ListResetTypes(params *ListResetTypesParams) (*ListResetTypesOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewListResetTypesParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"listResetTypes\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/Systems/{identifier}/Actions/ComputerSystem.Reset\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &ListResetTypesReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*ListResetTypesOK), nil\n\n}", "func addKnownTypes(scheme *runtime.Scheme) error {\n\tscheme.AddKnownTypes(SchemeGroupVersion,\n\t\t&mapi.Machine{},\n\t\t&mapi.MachineList{},\n\t)\n\tmeta.AddToGroupVersion(scheme, SchemeGroupVersion)\n\treturn nil\n}", "func (c *WorkitemtypeController) List(ctx *app.ListWorkitemtypeContext) error {\n\tstart, limit, err := parseLimit(ctx.Page)\n\tif err != nil {\n\t\tjerrors, _ := jsonapi.ErrorToJSONAPIErrors(goa.ErrBadRequest(fmt.Sprintf(\"could not parse paging: %s\", err.Error())))\n\t\treturn ctx.BadRequest(jerrors)\n\t}\n\treturn application.Transactional(c.db, func(appl application.Application) error {\n\t\tresult, err := appl.WorkItemTypes().List(ctx.Context, start, &limit)\n\t\tif err != nil {\n\t\t\tjerrors, _ := jsonapi.ErrorToJSONAPIErrors(goa.ErrBadRequest(fmt.Sprintf(\"Error listing work item types: %s\", err.Error())))\n\t\t\treturn ctx.BadRequest(jerrors)\n\t\t}\n\t\treturn ctx.OK(result)\n\t})\n}", "func (client GroupClient) ListTypes(accountName string, databaseName string, schemaName string, filter string, top *int32, skip *int32, expand string, selectParameter string, orderby string, count *bool) (result USQLTypeList, err error) {\n\treq, err := client.ListTypesPreparer(accountName, databaseName, schemaName, filter, top, skip, expand, selectParameter, orderby, count)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"catalog.GroupClient\", \"ListTypes\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresp, err := client.ListTypesSender(req)\n\tif err != nil {\n\t\tresult.Response = autorest.Response{Response: resp}\n\t\terr = autorest.NewErrorWithError(err, \"catalog.GroupClient\", \"ListTypes\", resp, \"Failure sending request\")\n\t\treturn\n\t}\n\n\tresult, err = client.ListTypesResponder(resp)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"catalog.GroupClient\", \"ListTypes\", resp, \"Failure responding to request\")\n\t}\n\n\treturn\n}", "func PhysicalServerHostTypesEnumFromValue(value string) PhysicalServerHostTypesEnum {\r\n switch value {\r\n case \"kLinux\":\r\n return PhysicalServerHostTypes_KLINUX\r\n case \"kWindows\":\r\n return PhysicalServerHostTypes_KWINDOWS\r\n case \"kAix\":\r\n return PhysicalServerHostTypes_KAIX\r\n case \"kSolaris\":\r\n return PhysicalServerHostTypes_KSOLARIS\r\n default:\r\n return PhysicalServerHostTypes_KLINUX\r\n }\r\n}", "func (dm *DockerMachine) Machines() ([]Machine, error) {\n\tfilter := `{\"Name\":\"{{.Name}}\",\"State\":\"{{.State}}\",\"URL\":\"{{.URL}}\"}`\n\tbytes, err := dm.RawCommandResult(\"ls\", \"-f\", filter)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar machines []Machine\n\tfor _, s := range strings.Split(string(bytes), \"\\n\") {\n\t\ts = strings.TrimSpace(s)\n\t\tif s == \"\" {\n\t\t\tcontinue\n\t\t}\n\t\tvar machine Machine\n\t\tif err = json.Unmarshal([]byte(s), &machine); err != nil {\n\t\t\treturn nil, fmt.Errorf(\"error while parsing \\\"%s\\\": %v\", s, err)\n\t\t}\n\t\tmachines = append(machines, machine)\n\t}\n\treturn machines, nil\n}", "func (client GroupClient) ListTableTypesResponder(resp *http.Response) (result USQLTableTypeList, err error) {\n\terr = autorest.Respond(\n\t\tresp,\n\t\tclient.ByInspecting(),\n\t\tazure.WithErrorUnlessStatusCode(http.StatusOK),\n\t\tautorest.ByUnmarshallingJSON(&result),\n\t\tautorest.ByClosing())\n\tresult.Response = autorest.Response{Response: resp}\n\treturn\n}", "func List(_ machine.ListOptions) ([]*machine.ListResponse, error) {\n\treturn GetVMInfos()\n}", "func (a *DefaultApiService) ListApplicationTypes(ctx _context.Context) ApiListApplicationTypesRequest {\n\treturn ApiListApplicationTypesRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t}\n}", "func (*ListAffinityGroupTypes) name() string {\n\treturn \"listAffinityGroupTypes\"\n}", "func (o SourceInstancePropertiesResponseOutput) MachineType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v SourceInstancePropertiesResponse) string { return v.MachineType }).(pulumi.StringOutput)\n}", "func (o AiEndpointDeployedModelDedicatedResourceMachineSpecOutput) MachineType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v AiEndpointDeployedModelDedicatedResourceMachineSpec) *string { return v.MachineType }).(pulumi.StringPtrOutput)\n}", "func (o InstanceOutput) MachineType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Instance) pulumi.StringOutput { return v.MachineType }).(pulumi.StringOutput)\n}", "func addKnownTypes(scheme *runtime.Scheme) error {\n\tscheme.AddKnownTypes(SchemeGroupVersion,\n\t\t&MachineConfig{},\n\t\t&MachineConfigList{},\n\t\t&MachineConfigPool{},\n\t\t&MachineConfigPoolList{},\n\t)\n\n\tmetav1.AddToGroupVersion(scheme, SchemeGroupVersion)\n\n\treturn nil\n}", "func typeListHandler(w http.ResponseWriter, r *http.Request) {\n\n\tresponder(w, r, data.Types)\n\n}", "func TestFleetctlListMachines(t *testing.T) {\n\tstdout, stderr, err := Run(fleetctlBinPath, \"list-machines\", \"--no-legend\")\n\tif err != nil {\n\t\tt.Fatalf(\"fleetctl list-machines failed with error: %v\\nstdout: %s\\nstderr: %s\", err, stdout, stderr)\n\t}\n\n\tstdout = strings.TrimSpace(stdout)\n\tif len(strings.Split(stdout, \"\\n\")) == 0 {\n\t\tt.Fatalf(\"Failed listing out at least one machine\\nstdout: %s\", stdout)\n\t}\n}", "func (o InstanceFromTemplateOutput) MachineType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *InstanceFromTemplate) pulumi.StringOutput { return v.MachineType }).(pulumi.StringOutput)\n}", "func (o AllocationSpecificSKUAllocationReservedInstancePropertiesResponseOutput) MachineType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v AllocationSpecificSKUAllocationReservedInstancePropertiesResponse) string { return v.MachineType }).(pulumi.StringOutput)\n}", "func ListTypes(options ...Option) []reflect.Type {\n\treturn registry.ListTypes(options...)\n}", "func LoadMachineForType(id int, t Type) Machine {\n\treturn nil\n}", "func (o TriggerBuildOptionsOutput) MachineType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v TriggerBuildOptions) *string { return v.MachineType }).(pulumi.StringPtrOutput)\n}", "func (t *TaskRPC) ListMachineSched(data shared.MachineRPCData, tasks *[]shared.SchedTask) error {\n\tstart := time.Now()\n\n\tconn := Connections.Get(data.Channel)\n\n\t// Read the sites that this user has access to\n\terr := DB.SQL(`select * from sched_task where machine_id=$1 order by id`, data.ID).QueryStructs(tasks)\n\n\tif err != nil {\n\t\tlog.Println(err.Error())\n\t}\n\n\t// Get the latest thumbnails for this task, if present\n\tfor i, v := range *tasks {\n\n\t\tphotos := []shared.Photo{}\n\t\tDB.SQL(`select id,thumb \n\t\t\tfrom photo \n\t\t\twhere (entity='sched' and entity_id=$1) \n\t\t\torder by type, id desc`, v.ID).\n\t\t\tQueryStructs(&photos)\n\t\t(*tasks)[i].Photos = photos\n\t}\n\n\tlogger(start, \"Task.ListMachineSched\",\n\t\tfmt.Sprintf(\"Machine %d\", data.ID),\n\t\tfmt.Sprintf(\"%d tasks\", len(*tasks)),\n\t\tdata.Channel, conn.UserID, \"machine\", 0, false)\n\n\treturn nil\n}", "func (o InstancePropertiesResponseOutput) MachineType() pulumi.StringOutput {\n\treturn o.ApplyT(func(v InstancePropertiesResponse) string { return v.MachineType }).(pulumi.StringOutput)\n}", "func (platforms *platformList) Type() string {\n\treturn \"platformList\"\n}", "func (a *Client) GetVMTypesByCredential(params *GetVMTypesByCredentialParams) (*GetVMTypesByCredentialOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetVMTypesByCredentialParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"getVmTypesByCredential\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/v1/platform_resources/machine_types\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &GetVMTypesByCredentialReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*GetVMTypesByCredentialOK), nil\n\n}", "func (o InstancePropertiesOutput) MachineType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v InstanceProperties) *string { return v.MachineType }).(pulumi.StringPtrOutput)\n}", "func (h *Handler) GetMachines(c echo.Context) (err error) {\n\treturn nil\n}", "func (o AllocationSpecificSKUAllocationReservedInstancePropertiesOutput) MachineType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v AllocationSpecificSKUAllocationReservedInstanceProperties) *string { return v.MachineType }).(pulumi.StringPtrOutput)\n}", "func native(name string) []Type {}", "func (o SourceInstancePropertiesResponsePtrOutput) MachineType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *SourceInstancePropertiesResponse) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.MachineType\n\t}).(pulumi.StringPtrOutput)\n}", "func nativeToTypeList(valueList []interface{}) []LogValueType {\n\tt := TypeByteData\n\ttypeList := make([]LogValueType, len(valueList))\n\tfor _, val := range valueList {\n\t\tswitch val.(type) {\n\t\tcase uint8:\n\t\t\tt = TypeUint8\n\t\tcase int8:\n\t\t\tt = TypeInt8\n\t\tcase uint32:\n\t\t\tt = TypeUint32\n\t\tcase int32:\n\t\t\tt = TypeInt32\n\t\tcase uint64:\n\t\t\tt = TypeUint64\n\t\tcase int64:\n\t\t\tt = TypeInt64\n\t\tcase bool:\n\t\t\tt = TypeBoolean\n\t\tcase string:\n\t\t\tt = TypeString\n\t\tcase float32:\n\t\t\tt = TypeFloat32\n\t\tcase float64:\n\t\t\tt = TypeFloat64\n\t\t}\n\t\ttypeList = append(typeList, t)\n\t}\n\treturn typeList\n}", "func (o TriggerBuildOptionsPtrOutput) MachineType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *TriggerBuildOptions) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.MachineType\n\t}).(pulumi.StringPtrOutput)\n}", "func (l Leftovers) ListByType(filter, rType string, regex bool) {\n\tl.List(filter, regex)\n}", "func (l Leftovers) ListByType(filter, rType string, regex bool) {\n\tl.List(filter, regex)\n}", "func (c *TestClient) GetMachineType(project, zone, machineType string) (*compute.MachineType, error) {\n\tif c.GetMachineTypeFn != nil {\n\t\treturn c.GetMachineTypeFn(project, zone, machineType)\n\t}\n\treturn c.client.GetMachineType(project, zone, machineType)\n}", "func (ha *hostedActors) ListActorTypes() []string {\n\tha.lock.RLock()\n\tdefer ha.lock.RUnlock()\n\treturn maps.Keys(ha.actors)\n}", "func (client WorkloadNetworksClient) ListVirtualMachinesResponder(resp *http.Response) (result WorkloadNetworkVirtualMachinesList, err error) {\n\terr = autorest.Respond(\n\t\tresp,\n\t\tazure.WithErrorUnlessStatusCode(http.StatusOK),\n\t\tautorest.ByUnmarshallingJSON(&result),\n\t\tautorest.ByClosing())\n\tresult.Response = autorest.Response{Response: resp}\n\treturn\n}", "func (machine *VirtualMachine) GetType() string {\n\treturn \"Microsoft.Compute/virtualMachines\"\n}", "func (machine *VirtualMachine) GetType() string {\n\treturn \"Microsoft.Compute/virtualMachines\"\n}", "func (*ValidatorSet) EnumTypes() []lcs.EnumVariant { return vsSchemeEnumDef }", "func (me *XsdGoPkgHasElems_ListItemType) Walk() (err error) {\r\n\tif fn := WalkHandlers.XsdGoPkgHasElems_ListItemType; me != nil {\r\n\t\tif fn != nil {\r\n\t\t\tif err = fn(me, true); xsdt.OnWalkError(&err, &WalkErrors, WalkContinueOnError, WalkOnError) {\r\n\t\t\t\treturn\r\n\t\t\t}\r\n\t\t}\r\n\t\tif fn != nil {\r\n\t\t\tif err = fn(me, false); xsdt.OnWalkError(&err, &WalkErrors, WalkContinueOnError, WalkOnError) {\r\n\t\t\t\treturn\r\n\t\t\t}\r\n\t\t}\r\n\t}\r\n\treturn\r\n}", "func (m *List) GetContentTypes()([]ContentTypeable) {\n return m.contentTypes\n}", "func DeviceTypesPWinCE() *DeviceTypes {\n\tv := DeviceTypesVWinCE\n\treturn &v\n}", "func DeviceTypesPWinCE() *DeviceTypes {\n\tv := DeviceTypesVWinCE\n\treturn &v\n}", "func (s *TypeRegistry) TypeListNames() []string {\n\tvalues := make([]string, len(s.typesByName))\n\n\ti := 0\n\tfor k := range s.typesByName {\n\t\tvalues[i] = k\n\t\ti++\n\t}\n\n\treturn values\n}", "func (o AllocationSpecificSKUAllocationReservedInstancePropertiesResponsePtrOutput) MachineType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *AllocationSpecificSKUAllocationReservedInstancePropertiesResponse) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.MachineType\n\t}).(pulumi.StringPtrOutput)\n}", "func (o InstancePropertiesPtrOutput) MachineType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *InstanceProperties) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.MachineType\n\t}).(pulumi.StringPtrOutput)\n}", "func (o InstancePropertiesResponsePtrOutput) MachineType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *InstancePropertiesResponse) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.MachineType\n\t}).(pulumi.StringPtrOutput)\n}", "func Machines(clusterID string, config *types.InstallConfig, pool *types.MachinePool, role, userDataSecret string) ([]machineapi.Machine, *machinev1.ControlPlaneMachineSet, error) {\n\tif configPlatform := config.Platform.Name(); configPlatform != powervs.Name {\n\t\treturn nil, nil, fmt.Errorf(\"non-PowerVS configuration: %q\", configPlatform)\n\t}\n\tif poolPlatform := pool.Platform.Name(); poolPlatform != powervs.Name {\n\t\treturn nil, nil, fmt.Errorf(\"non-PowerVS machine-pool: %q\", poolPlatform)\n\t}\n\tplatform := config.Platform.PowerVS\n\tmpool := pool.Platform.PowerVS\n\n\t// Only the service instance is guaranteed to exist and be passed via the install config\n\t// The other two, we should standardize a name including the cluster id.\n\timage := fmt.Sprintf(\"rhcos-%s\", clusterID)\n\tvar network string\n\tif platform.ClusterOSImage != \"\" {\n\t\timage = platform.ClusterOSImage\n\t}\n\tif platform.PVSNetworkName != \"\" {\n\t\tnetwork = platform.PVSNetworkName\n\t}\n\n\ttotal := int64(1)\n\tif pool.Replicas != nil {\n\t\ttotal = *pool.Replicas\n\t}\n\tvar machines []machineapi.Machine\n\tmachineProvider, err := provider(clusterID, platform, mpool, userDataSecret, image, network)\n\tif err != nil {\n\t\treturn nil, nil, errors.Wrap(err, \"failed to create provider\")\n\t}\n\tfor idx := int64(0); idx < total; idx++ {\n\t\tmachine := machineapi.Machine{\n\t\t\tTypeMeta: metav1.TypeMeta{\n\t\t\t\tAPIVersion: \"machine.openshift.io/v1beta1\",\n\t\t\t\tKind: \"Machine\",\n\t\t\t},\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tNamespace: \"openshift-machine-api\",\n\t\t\t\tName: fmt.Sprintf(\"%s-%s-%d\", clusterID, pool.Name, idx),\n\t\t\t\tLabels: map[string]string{\n\t\t\t\t\t\"machine.openshift.io/cluster-api-cluster\": clusterID,\n\t\t\t\t\t\"machine.openshift.io/cluster-api-machine-role\": role,\n\t\t\t\t\t\"machine.openshift.io/cluster-api-machine-type\": role,\n\t\t\t\t},\n\t\t\t},\n\t\t\tSpec: machineapi.MachineSpec{\n\t\t\t\tProviderSpec: machineapi.ProviderSpec{\n\t\t\t\t\tValue: &runtime.RawExtension{Object: machineProvider},\n\t\t\t\t},\n\t\t\t},\n\t\t}\n\t\tmachines = append(machines, machine)\n\t}\n\treplicas := int32(total)\n\tcontrolPlaneMachineSet := &machinev1.ControlPlaneMachineSet{\n\t\tTypeMeta: metav1.TypeMeta{\n\t\t\tAPIVersion: \"machine.openshift.io/v1\",\n\t\t\tKind: \"ControlPlaneMachineSet\",\n\t\t},\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tNamespace: \"openshift-machine-api\",\n\t\t\tName: \"cluster\",\n\t\t\tLabels: map[string]string{\n\t\t\t\t\"machine.openshift.io/cluster-api-cluster\": clusterID,\n\t\t\t},\n\t\t},\n\t\tSpec: machinev1.ControlPlaneMachineSetSpec{\n\t\t\tReplicas: &replicas,\n\t\t\tState: machinev1.ControlPlaneMachineSetStateInactive,\n\t\t\tSelector: metav1.LabelSelector{\n\t\t\t\tMatchLabels: map[string]string{\n\t\t\t\t\t\"machine.openshift.io/cluster-api-machine-role\": role,\n\t\t\t\t\t\"machine.openshift.io/cluster-api-machine-type\": role,\n\t\t\t\t\t\"machine.openshift.io/cluster-api-cluster\": clusterID,\n\t\t\t\t},\n\t\t\t},\n\t\t\tTemplate: machinev1.ControlPlaneMachineSetTemplate{\n\t\t\t\tMachineType: machinev1.OpenShiftMachineV1Beta1MachineType,\n\t\t\t\tOpenShiftMachineV1Beta1Machine: &machinev1.OpenShiftMachineV1Beta1MachineTemplate{\n\t\t\t\t\tObjectMeta: machinev1.ControlPlaneMachineSetTemplateObjectMeta{\n\t\t\t\t\t\tLabels: map[string]string{\n\t\t\t\t\t\t\t\"machine.openshift.io/cluster-api-cluster\": clusterID,\n\t\t\t\t\t\t\t\"machine.openshift.io/cluster-api-machine-role\": role,\n\t\t\t\t\t\t\t\"machine.openshift.io/cluster-api-machine-type\": role,\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t\tSpec: machineapi.MachineSpec{\n\t\t\t\t\t\tProviderSpec: machineapi.ProviderSpec{\n\t\t\t\t\t\t\tValue: &runtime.RawExtension{Object: machineProvider},\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\treturn machines, controlPlaneMachineSet, nil\n}", "func (r *PlatformTypesService) List(profileId int64) *PlatformTypesListCall {\n\tc := &PlatformTypesListCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.profileId = profileId\n\treturn c\n}", "func (me *XsdGoPkgHasElem_ListItemType) Walk() (err error) {\r\n\tif fn := WalkHandlers.XsdGoPkgHasElem_ListItemType; me != nil {\r\n\t\tif fn != nil {\r\n\t\t\tif err = fn(me, true); xsdt.OnWalkError(&err, &WalkErrors, WalkContinueOnError, WalkOnError) {\r\n\t\t\t\treturn\r\n\t\t\t}\r\n\t\t}\r\n\t\tif fn != nil {\r\n\t\t\tif err = fn(me, false); xsdt.OnWalkError(&err, &WalkErrors, WalkContinueOnError, WalkOnError) {\r\n\t\t\t\treturn\r\n\t\t\t}\r\n\t\t}\r\n\t}\r\n\treturn\r\n}", "func (o AllocationSpecificSKUAllocationReservedInstancePropertiesPtrOutput) MachineType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *AllocationSpecificSKUAllocationReservedInstanceProperties) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.MachineType\n\t}).(pulumi.StringPtrOutput)\n}", "func (s *fsStore) List(typ namespace.Type) []string {\n\tout := []string{}\n\tdir := filepath.Join(s.root, typ.StringLower())\n\tfl, err := ioutil.ReadDir(dir)\n\tif err != nil {\n\t\treturn out\n\t}\n\tfor _, inf := range fl {\n\t\tout = append(out, inf.Name())\n\t}\n\treturn out\n}", "func DeviceTypesPWindowsPhone() *DeviceTypes {\n\tv := DeviceTypesVWindowsPhone\n\treturn &v\n}", "func DeviceTypesPWindowsPhone() *DeviceTypes {\n\tv := DeviceTypesVWindowsPhone\n\treturn &v\n}", "func (client *ClientImpl) ListEventTypes(ctx context.Context, args ListEventTypesArgs) (*[]EventTypeDescriptor, error) {\n\trouteValues := make(map[string]string)\n\tif args.PublisherId == nil || *args.PublisherId == \"\" {\n\t\treturn nil, &azuredevops.ArgumentNilOrEmptyError{ArgumentName: \"args.PublisherId\"}\n\t}\n\trouteValues[\"publisherId\"] = *args.PublisherId\n\n\tlocationId, _ := uuid.Parse(\"db4777cd-8e08-4a84-8ba3-c974ea033718\")\n\tresp, err := client.Client.Send(ctx, http.MethodGet, locationId, \"7.1-preview.1\", routeValues, nil, nil, \"\", \"application/json\", nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar responseValue []EventTypeDescriptor\n\terr = client.Client.UnmarshalCollectionBody(resp, &responseValue)\n\treturn &responseValue, err\n}", "func (client GroupClient) ListTableTypes(accountName string, databaseName string, schemaName string, filter string, top *int32, skip *int32, expand string, selectParameter string, orderby string, count *bool) (result USQLTableTypeList, err error) {\n\treq, err := client.ListTableTypesPreparer(accountName, databaseName, schemaName, filter, top, skip, expand, selectParameter, orderby, count)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"catalog.GroupClient\", \"ListTableTypes\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresp, err := client.ListTableTypesSender(req)\n\tif err != nil {\n\t\tresult.Response = autorest.Response{Response: resp}\n\t\terr = autorest.NewErrorWithError(err, \"catalog.GroupClient\", \"ListTableTypes\", resp, \"Failure sending request\")\n\t\treturn\n\t}\n\n\tresult, err = client.ListTableTypesResponder(resp)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"catalog.GroupClient\", \"ListTableTypes\", resp, \"Failure responding to request\")\n\t}\n\n\treturn\n}", "func (l Leftovers) Types() {\n\tl.logger.NoConfirm()\n\n\tfor _, r := range l.resources {\n\t\tl.logger.Println(r.Type())\n\t}\n}", "func (a ClustersAPI) ListNodeTypes() ([]httpmodels.ListNodeTypesRespItem, error) {\n\tvar nodeTypeList = struct {\n\t\tNodeTypes []httpmodels.ListNodeTypesRespItem `json:\"node_types,omitempty\" url:\"node_types,omitempty\"`\n\t}{}\n\n\tresp, err := a.Client.performQuery(http.MethodGet, \"/clusters/list-node-types\", nil, nil)\n\tif err != nil {\n\t\treturn nodeTypeList.NodeTypes, err\n\t}\n\n\terr = json.Unmarshal(resp, &nodeTypeList)\n\treturn nodeTypeList.NodeTypes, err\n}", "func ListForType(rowType reflect.Type) []*Info {\n\ttypeMap.mu.RLock()\n\tlist, ok := typeMap.m[rowType]\n\ttypeMap.mu.RUnlock()\n\tif ok {\n\t\treturn list\n\t}\n\n\ttypeMap.mu.Lock()\n\tdefer typeMap.mu.Unlock()\n\tlist = newList(rowType)\n\ttypeMap.m[rowType] = list\n\treturn list\n}", "func (s *OnDiskStateMachine) StateMachineType() pb.StateMachineType {\n\treturn pb.OnDiskStateMachine\n}", "func DeviceTypesPMacMDM() *DeviceTypes {\n\tv := DeviceTypesVMacMDM\n\treturn &v\n}", "func DeviceTypesPMacMDM() *DeviceTypes {\n\tv := DeviceTypesVMacMDM\n\treturn &v\n}", "func ListVirtualMachineSummaries(settings *playfab.Settings, postData *ListVirtualMachineSummariesRequestModel, entityToken string) (*ListVirtualMachineSummariesResponseModel, error) {\n if entityToken == \"\" {\n return nil, playfab.NewCustomError(\"entityToken should not be an empty string\", playfab.ErrorGeneric)\n }\n b, errMarshal := json.Marshal(postData)\n if errMarshal != nil {\n return nil, playfab.NewCustomError(errMarshal.Error(), playfab.ErrorMarshal)\n }\n\n sourceMap, err := playfab.Request(settings, b, \"/MultiplayerServer/ListVirtualMachineSummaries\", \"X-EntityToken\", entityToken)\n if err != nil {\n return nil, err\n }\n \n result := &ListVirtualMachineSummariesResponseModel{}\n\n config := mapstructure.DecoderConfig{\n DecodeHook: playfab.StringToDateTimeHook,\n Result: result,\n }\n \n decoder, errDecoding := mapstructure.NewDecoder(&config)\n if errDecoding != nil {\n return nil, playfab.NewCustomError(errDecoding.Error(), playfab.ErrorDecoding)\n }\n \n errDecoding = decoder.Decode(sourceMap)\n if errDecoding != nil {\n return nil, playfab.NewCustomError(errDecoding.Error(), playfab.ErrorDecoding)\n }\n\n return result, nil\n}", "func DeviceTypesPPalm() *DeviceTypes {\n\tv := DeviceTypesVPalm\n\treturn &v\n}", "func DeviceTypesPPalm() *DeviceTypes {\n\tv := DeviceTypesVPalm\n\treturn &v\n}", "func (m *SchemaExtension) GetTargetTypes()([]string) {\n val, err := m.GetBackingStore().Get(\"targetTypes\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.([]string)\n }\n return nil\n}", "func (client GroupClient) ListTypesSender(req *http.Request) (*http.Response, error) {\n\treturn autorest.SendWithSender(client,\n\t\treq,\n\t\tautorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n}", "func (s *BasecluListener) ExitType_spec_list(ctx *Type_spec_listContext) {}", "func (l Leftovers) Types() {\n\tfor _, r := range l.resources {\n\t\tl.logger.Println(r.Type())\n\t}\n}", "func (s *TaskTypesEndpoint) List(ctx context.Context, division int, all bool, o *api.ListOptions) ([]*TaskTypes, error) {\n\tvar entities []*TaskTypes\n\tu, _ := s.client.ResolvePathWithDivision(\"/api/v1/{division}/accountancy/TaskTypes\", division) // #nosec\n\tapi.AddListOptionsToURL(u, o)\n\n\tif all {\n\t\terr := s.client.ListRequestAndDoAll(ctx, u.String(), &entities)\n\t\treturn entities, err\n\t}\n\t_, _, err := s.client.NewRequestAndDo(ctx, \"GET\", u.String(), nil, &entities)\n\treturn entities, err\n}", "func NewListRegistryProviderTypesUnauthorized() *ListRegistryProviderTypesUnauthorized {\n\treturn &ListRegistryProviderTypesUnauthorized{}\n}", "func (client *Client) ListAvailableFileSystemTypes(request *ListAvailableFileSystemTypesRequest) (response *ListAvailableFileSystemTypesResponse, err error) {\n\tresponse = CreateListAvailableFileSystemTypesResponse()\n\terr = client.DoAction(request, response)\n\treturn\n}", "func (a *Client) VirtualizationVirtualMachinesList(params *VirtualizationVirtualMachinesListParams) (*VirtualizationVirtualMachinesListOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewVirtualizationVirtualMachinesListParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"virtualization_virtual-machines_list\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/api/virtualization/virtual-machines/\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\"},\n\t\tParams: params,\n\t\tReader: &VirtualizationVirtualMachinesListReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*VirtualizationVirtualMachinesListOK), nil\n\n}", "func flattenImageGuestOsFeatureTypeEnumSlice(c *Client, i interface{}) []ImageGuestOsFeatureTypeEnum {\n\ta, ok := i.([]interface{})\n\tif !ok {\n\t\treturn []ImageGuestOsFeatureTypeEnum{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn []ImageGuestOsFeatureTypeEnum{}\n\t}\n\n\titems := make([]ImageGuestOsFeatureTypeEnum, 0, len(a))\n\tfor _, item := range a {\n\t\titems = append(items, *flattenImageGuestOsFeatureTypeEnum(item.(interface{})))\n\t}\n\n\treturn items\n}", "func listEntitiesOfType(entityType string, params url.Values, logJSON bool, out io.Writer, timeout int, verbose bool) ([]*gabs.Container, *paging, error) {\n\tjsonParsed, err := util.EdgeControllerList(entityType, params, logJSON, out, timeout, verbose)\n\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tchildren, err := jsonParsed.S(\"data\").Children()\n\treturn children, getPaging(jsonParsed), err\n}", "func (a *DefaultApiService) ListSourceApplicationTypes(ctx _context.Context, id string) ApiListSourceApplicationTypesRequest {\n\treturn ApiListSourceApplicationTypesRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\tid: id,\n\t}\n}", "func tileTypes() []TileType {\n\treturn []TileType{\n\t\tTileTypeMountain,\n\t\tTileTypeWater,\n\t\tTileTypeForest,\n\t}\n}", "func ListResourceTypes() []string {\n\tresourceTypes := []string{\n\t\tASGroups{}.ResourceName(),\n\t\tLaunchConfigs{}.ResourceName(),\n\t\tLoadBalancers{}.ResourceName(),\n\t\tLoadBalancersV2{}.ResourceName(),\n\t\tEC2Instances{}.ResourceName(),\n\t\tEBSVolumes{}.ResourceName(),\n\t\tEIPAddresses{}.ResourceName(),\n\t\tAMIs{}.ResourceName(),\n\t\tSnapshots{}.ResourceName(),\n\t\tECSServices{}.ResourceName(),\n\t\tEKSClusters{}.ResourceName(),\n\t\tDBInstances{}.ResourceName(),\n\t\tS3Buckets{}.ResourceName(),\n\t}\n\tsort.Strings(resourceTypes)\n\treturn resourceTypes\n}" ]
[ "0.77187157", "0.7685642", "0.65473205", "0.64404184", "0.64359", "0.63986224", "0.6355094", "0.62219435", "0.62215847", "0.60004526", "0.57352376", "0.57282484", "0.55174154", "0.5437722", "0.5423765", "0.54158235", "0.5405157", "0.53849334", "0.52415234", "0.5132877", "0.51312476", "0.51189035", "0.51182467", "0.51152384", "0.5112196", "0.5108174", "0.51043737", "0.50996184", "0.5089298", "0.507714", "0.5055361", "0.50441796", "0.5027373", "0.4998407", "0.4976085", "0.49586776", "0.49473324", "0.49394894", "0.49159545", "0.48960474", "0.48939082", "0.48742008", "0.4867931", "0.486758", "0.48501623", "0.4839943", "0.48344135", "0.48337802", "0.48024476", "0.47829992", "0.4765388", "0.47301725", "0.47037768", "0.469242", "0.469242", "0.46879143", "0.46867982", "0.4682351", "0.46757773", "0.46757773", "0.46616727", "0.46550623", "0.4632405", "0.46229735", "0.46229735", "0.46199033", "0.46154568", "0.46023494", "0.46003655", "0.45998436", "0.45970148", "0.457733", "0.45749733", "0.45631793", "0.45583984", "0.45583984", "0.45539525", "0.4548713", "0.4540183", "0.4517784", "0.4507451", "0.44980967", "0.44963893", "0.44963893", "0.4476285", "0.447219", "0.447219", "0.4465217", "0.44621816", "0.4452863", "0.4451602", "0.44485152", "0.4448404", "0.44301382", "0.44226563", "0.4417273", "0.44154787", "0.4413574", "0.44094095", "0.44082582" ]
0.7902348
0
GetZone uses the override method GetZoneFn or the real implementation.
func (c *TestClient) GetZone(project, zone string) (*compute.Zone, error) { if c.GetZoneFn != nil { return c.GetZoneFn(project, zone) } return c.client.GetZone(project, zone) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (z zones) GetZone() (cloudprovider.Zone, error) {\n\treturn cloudprovider.Zone{Region: z.region}, nil\n}", "func (z *zones) GetZone(ctx context.Context) (cloudprovider.Zone, error) {\n\tklog.V(5).Info(\"GetZone()\")\n\treturn cloudprovider.Zone{\n\t\tFailureDomain: \"laptop\",\n\t\tRegion: \"virtualbox\",\n\t}, nil\n}", "func (z zones) GetZone(ctx context.Context) (cloudprovider.Zone, error) {\n\tzone, err := queryInstanceMetadata(\"availability-zone\")\n\tif err != nil {\n\t\treturn cloudprovider.Zone{}, err\n\t}\n\n\treturn cloudprovider.Zone{Region: zone}, nil\n}", "func (bc *Baiducloud) GetZone(ctx context.Context) (cloudprovider.Zone, error) {\n\tzone := cloudprovider.Zone{\n\t\tFailureDomain: \"unknow\",\n\t\tRegion: bc.Region,\n\t}\n\tif bc.NodeName != \"\" {\n\t\tins, err := bc.getInstanceByNodeName(ctx, types.NodeName(bc.NodeName))\n\t\t// ins, err := bc.getVirtualMachine(types.NodeName(bc.NodeIP))\n\t\tif err != nil {\n\t\t\treturn zone, err\n\t\t}\n\t\tzone.FailureDomain = ins.AvailableZone\n\t}\n\treturn zone, nil\n}", "func (client DnsClient) GetZone(ctx context.Context, request GetZoneRequest) (response GetZoneResponse, err error) {\n\tvar ociResponse common.OCIResponse\n\tpolicy := common.NoRetryPolicy()\n\tif request.RetryPolicy() != nil {\n\t\tpolicy = *request.RetryPolicy()\n\t}\n\tociResponse, err = common.Retry(ctx, request, client.getZone, policy)\n\tif err != nil {\n\t\tif ociResponse != nil {\n\t\t\tif httpResponse := ociResponse.HTTPResponse(); httpResponse != nil {\n\t\t\t\topcRequestId := httpResponse.Header.Get(\"opc-request-id\")\n\t\t\t\tresponse = GetZoneResponse{RawResponse: httpResponse, OpcRequestId: &opcRequestId}\n\t\t\t} else {\n\t\t\t\tresponse = GetZoneResponse{}\n\t\t\t}\n\t\t}\n\t\treturn\n\t}\n\tif convertedResponse, ok := ociResponse.(GetZoneResponse); ok {\n\t\tresponse = convertedResponse\n\t} else {\n\t\terr = fmt.Errorf(\"failed to convert OCIResponse into GetZoneResponse\")\n\t}\n\treturn\n}", "func (w *Watcher) GetZone(domain string) (*Zone, error) {\n\tmutableMutex.Lock()\n\tdefer mutableMutex.Unlock()\n\tif w.ZoneMap == nil {\n\t\tw.ZoneMap = make(map[string]*Zone)\n\t}\n\tzone, ok := w.ZoneMap[domain]\n\tif !ok {\n\t\treturn nil, errors.Errorf(\"not exist domain\")\n\t}\n\treturn zone, nil\n}", "func (p *PowerDNS) GetZone(domain string) (*Zone, error) {\n\tzone := &Zone{}\n\tmyError := new(Error)\n\tzoneSling := p.makeSling()\n\tresp, err := zoneSling.New().Get(\"servers/\"+p.VHost+\"/zones/\"+strings.TrimRight(domain, \".\")).Receive(zone, myError)\n\n\tif err == nil && resp.StatusCode >= 400 {\n\t\tmyError.Message = strings.Join([]string{resp.Status, myError.Message}, \" \")\n\t\treturn &Zone{}, myError\n\t}\n\n\tzone.PowerDNSHandle = p\n\treturn zone, err\n}", "func GetZone(ctx context.Context, tenantID int, zoneID int) (*database.Zone, error) {\n\treq := handlers.GetZoneRequest{\n\t\tZoneID: zoneID,\n\t\tZoneOptions: database.ZoneOptions{\n\t\t\tWithPlaces: true,\n\t\t},\n\t}\n\tresp := database.Zone{}\n\terr := jsonConn.RequestWithContext(ctx, \"zones.get\", &req, &resp)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &resp, nil\n}", "func (ac *azureClient) GetZone(ctx context.Context, resourceGroupName, zoneName string) (privatedns.PrivateZone, error) {\n\tctx, _, done := tele.StartSpanWithLogger(ctx, \"privatedns.AzureClient.GetZone\")\n\tdefer done()\n\tzone, err := ac.privatezones.Get(ctx, resourceGroupName, zoneName)\n\tif err != nil {\n\t\treturn privatedns.PrivateZone{}, err\n\t}\n\treturn zone, nil\n}", "func (c *Client) GetZone() (zone string, err error) {\n\tclient := &http.Client{\n\t\tCheckRedirect: func(req *http.Request, via []*http.Request) error {\n\t\t\treturn http.ErrUseLastResponse\n\t\t},\n\t}\n\turl := fmt.Sprintf(\"%s://%s.%s:%d\", c.Protocol, c.BucketName, c.Host, c.Port)\n\n\tr, err := client.Head(url)\n\tif err != nil {\n\t\tlogrus.Errorf(\"Get QingStor zone failed for %v.\", err)\n\t\treturn\n\t}\n\n\t// Example URL: https://bucket.zone.qingstor.com\n\tzone = strings.Split(r.Header.Get(\"Location\"), \".\")[1]\n\treturn\n}", "func (p *Provider) GetZone(name string) (dns.Zone, error) {\n mZones, err := p.listMZones()\n if err != nil {\n return dns.Zone{}, fmt.Errorf(\"couldn't get zone `%s`, see: failed to list zones, see: %v\", name, err)\n }\n\n var foundZone *models.Zone\n\n for _, z := range mZones {\n if z.Origin != nil && *z.Origin == name {\n foundZone = z\n }\n }\n\n if foundZone == nil {\n return dns.Zone{}, fmt.Errorf(\"zone `%s` not found\", name)\n }\n\n mZone, err := p.getMZone(*foundZone.Origin, foundZone.VirtualNameServer)\n if err != nil {\n return dns.Zone{}, fmt.Errorf(\"couldn't get zone `%s`, see: failed to retrieve zone upstream, see: %v\", name, err)\n }\n\n return p.mapToDNSZone(mZone), nil\n}", "func (client DnsClient) getZone(ctx context.Context, request common.OCIRequest) (common.OCIResponse, error) {\n\thttpRequest, err := request.HTTPRequest(http.MethodGet, \"/zones/{zoneNameOrId}\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar response GetZoneResponse\n\tvar httpResponse *http.Response\n\thttpResponse, err = client.Call(ctx, &httpRequest)\n\tdefer common.CloseBodyIfValid(httpResponse)\n\tresponse.RawResponse = httpResponse\n\tif err != nil {\n\t\treturn response, err\n\t}\n\n\terr = common.UnmarshalResponse(httpResponse, &response)\n\treturn response, err\n}", "func (c APIClient) GetZone(zone string) (*Zone, error) {\n\tz := NewZone(zone)\n\t_, err := c.doHTTPUnmarshal(\"GET\", fmt.Sprintf(\"https://api.nsone.net/v1/zones/%s\", z.Zone), nil, z)\n\treturn z, err\n}", "func (info *endpointsInfo) GetZone() string {\n\treturn \"\"\n}", "func (w *ServerInterfaceWrapper) GetZone(ctx echo.Context) error {\n\tvar err error\n\t// ------------- Path parameter \"zone\" -------------\n\tvar zone Zone\n\n\terr = runtime.BindStyledParameter(\"simple\", false, \"zone\", ctx.Param(\"zone\"), &zone)\n\tif err != nil {\n\t\treturn echo.NewHTTPError(http.StatusBadRequest, fmt.Sprintf(\"Invalid format for parameter zone: %s\", err))\n\t}\n\n\t// Invoke the callback with all the unmarshalled arguments\n\terr = w.Handler.GetZone(ctx, zone)\n\treturn err\n}", "func (m *SetLegacyAbacRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (o *Claims) GetZone() int {\n\n\treturn o.Zone\n}", "func (info *BaseEndpointInfo) GetZone() string {\n\treturn info.Zone\n}", "func (m *GetClusterRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (m *GetOperationRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (s *FastDNSv2Service) GetZone(ctx context.Context, zone string) (*ZoneMetadata, *Response, error) {\n\tu := fmt.Sprintf(\"config-dns/v2/zones/%v\", zone)\n\n\treq, err := s.client.NewRequest(\"GET\", u, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tvar zmeta *ZoneMetadata\n\tresp, err := s.client.Do(ctx, req, &zmeta)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\n\treturn zmeta, resp, nil\n}", "func (m *Cluster) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (dt *DateTime) GetZone() *String {\n\treturn dt.Zone()\n}", "func (p *Project) GetZone() string {\n\tp.mux.Lock()\n\tdefer p.mux.Unlock()\n\n\tzc := len(p.zoneIndices)\n\tif zc == 0 {\n\t\t// TODO: return an error instead of stopping the process.\n\t\tmsg := \"Not enough zone quota sepcified. Specify additional quota in `test_zones`.\"\n\t\tfmt.Println(msg)\n\t\treturn msg\n\t}\n\n\tzi := rand.Intn(zc)\n\tz := p.zoneIndices[zi]\n\n\tp.testZones[z]--\n\tif p.testZones[z] == 0 {\n\t\tp.zoneIndices = append(p.zoneIndices[:zi], p.zoneIndices[zi+1:]...)\n\t}\n\n\treturn z\n}", "func (m *GetNodePoolRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (m *GetServerConfigRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (o *SSHAuthority) GetZone() int {\n\n\treturn o.Zone\n}", "func (z *zones) GetZoneByProviderID(ctx context.Context, providerID string) (cloudprovider.Zone, error) {\n\tklog.V(5).Infof(\"GetZoneByProviderID(%v)\", providerID)\n\treturn cloudprovider.Zone{\n\t\tFailureDomain: \"virtualbox\",\n\t\tRegion: \"virtualbox\",\n\t}, nil\n}", "func (m *SetNodePoolAutoscalingRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (a *Client) GetTransportZone(params *GetTransportZoneParams, authInfo runtime.ClientAuthInfoWriter) (*GetTransportZoneOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetTransportZoneParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"GetTransportZone\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/transport-zones/{zone-id}\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &GetTransportZoneReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*GetTransportZoneOK), nil\n\n}", "func (m *SetMasterAuthRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (qc *QingCloud) GetZoneByProviderID(ctx context.Context, providerID string) (cloudprovider.Zone, error) {\n\tklog.V(4).Infof(\"GetZoneByProviderID() called, current zone is %v, and return zone directly as temporary solution\", qc.zone)\n\treturn cloudprovider.Zone{Region: qc.zone}, nil\n}", "func (m *ListClustersRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (m *Operation) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (m *SetNetworkPolicyRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (m *RollbackNodePoolUpgradeRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (m *StartIPRotationRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (m *CreateNodePoolRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (m *CreateClusterRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (m *SetAddonsConfigRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (m *SetNodePoolManagementRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func Get(client *gophercloud.ServiceClient, zoneID string) (r GetResult) {\n\tresp, err := client.Get(zoneURL(client, zoneID), &r.Body, nil)\n\t_, r.Header, r.Err = gophercloud.ParseResponse(resp, err)\n\treturn\n}", "func (m *CompleteIPRotationRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (c *Client) GetDNSZone(ns string) (*DNSZone, error) {\n\toptions := map[string]interface{}{\"all\": true}\n\n\tres, err := c.rpc(\"dnszone_show\", []string{ns}, options)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar dnsRec DNSZone\n\terr = json.Unmarshal(res.Result.Data, &dnsRec)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &dnsRec, nil\n}", "func (m *ListOperationsRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (decryptor *PgDecryptor) GetMatchedZoneID() []byte {\n\tif decryptor.IsWithZone() {\n\t\treturn decryptor.zoneMatcher.GetZoneID()\n\t}\n\treturn nil\n}", "func (m *SetMonitoringServiceRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (m *UpdateNodePoolRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (m *DeleteNodePoolRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (azc *azureZonesClient) Get(ctx context.Context, spec azure.ResourceSpecGetter) (result interface{}, err error) {\n\tctx, _, done := tele.StartSpanWithLogger(ctx, \"privatedns.azureZonesClient.Get\")\n\tdefer done()\n\tzone, err := azc.privatezones.Get(ctx, spec.ResourceGroupName(), spec.ResourceName())\n\tif err != nil {\n\t\treturn privatedns.PrivateZone{}, err\n\t}\n\treturn zone, nil\n}", "func (m *UpdateClusterRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (bc *Baiducloud) GetZoneByProviderID(ctx context.Context, providerID string) (cloudprovider.Zone, error) {\n\tinstance, err := bc.getInstanceByProviderID(ctx, providerID)\n\tif err != nil {\n\t\treturn cloudprovider.Zone{}, err\n\t}\n\n\treturn cloudprovider.Zone{\n\t\tFailureDomain: instance.AvailableZone,\n\t\tRegion: bc.Region,\n\t}, nil\n}", "func (m *UpdateMasterRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func getZoneID(cfg *Config, c *CfVars) (ZoneID string, err error) {\n\n\tid, err := c.API.ZoneIDByName(cfg.Domain)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn id, nil\n\n}", "func (m *DeleteClusterRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (decryptor *PgDecryptor) GetZoneMatcher() *zone.Matcher {\n\treturn decryptor.zoneMatcher\n}", "func (z *zones) GetZoneByProviderID(ctx context.Context, providerID string) (cloudprovider.Zone, error) {\n\tinstance, err := z.p.computeInstanceByProviderID(ctx, providerID)\n\tif err != nil {\n\t\treturn cloudprovider.Zone{}, err\n\t}\n\n\treturn cloudprovider.Zone{Region: instance.ZoneName}, nil\n}", "func (m *Mockclient) GetZone(arg0 context.Context, arg1, arg2 string) (privatedns.PrivateZone, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"GetZone\", arg0, arg1, arg2)\n\tret0, _ := ret[0].(privatedns.PrivateZone)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (z *FakeZoneLister) GetZoneForNode(name string) (string, error) {\n\t// TODO: evolve as required, it's currently needed just to satisfy the\n\t// interface in unittests that don't care about zones. See unittests in\n\t// controller/util_test for actual zoneLister testing.\n\treturn z.Zones[0], nil\n}", "func (m *SetLocationsRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (o *SparseSSHAuthority) GetZone() (out int) {\n\n\tif o.Zone == nil {\n\t\treturn\n\t}\n\n\treturn *o.Zone\n}", "func (m *SetLabelsRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (m *ListNodePoolsRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (o *SparseClaims) GetZone() (out int) {\n\n\tif o.Zone == nil {\n\t\treturn\n\t}\n\n\treturn *o.Zone\n}", "func Get_timezone(ipaddress string) IP2Locationrecord {\n\treturn handleError(defaultDB.query(ipaddress, timezone))\n}", "func (o LookupInstanceResultOutput) Zone() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupInstanceResult) string { return v.Zone }).(pulumi.StringOutput)\n}", "func (t *TimeService) Zone(request *ZoneRequest) (*ZoneResponse, error) {\n\trsp := &ZoneResponse{}\n\treturn rsp, t.client.Call(\"time\", \"Zone\", request, rsp)\n}", "func getRegionFromZone(zoneName string) string {\n\treturn zoneName[:strings.LastIndex(zoneName, \"-\")]\n}", "func (m *CancelOperationRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (bc *Baiducloud) GetZoneByNodeName(ctx context.Context, nodeName types.NodeName) (cloudprovider.Zone, error) {\n\tinstance, err := bc.getInstanceByNodeName(ctx, nodeName)\n\tif err != nil {\n\t\treturn cloudprovider.Zone{}, err\n\t}\n\tzone := cloudprovider.Zone{\n\t\tFailureDomain: instance.AvailableZone,\n\t\tRegion: bc.Region,\n\t}\n\treturn zone, nil\n}", "func (a *azurednsProvider) GetZoneRecords(domain string, meta map[string]string) (models.Records, error) {\n\texistingRecords, _, _, err := a.getExistingRecords(domain)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn existingRecords, nil\n}", "func (m *SetNodePoolSizeRequest) GetZone() string {\n\tif m != nil {\n\t\treturn m.Zone\n\t}\n\treturn \"\"\n}", "func (o StorageNodeStatusGeographyOutput) Zone() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v StorageNodeStatusGeography) *string { return v.Zone }).(pulumi.StringPtrOutput)\n}", "func (t Time) Zone() (name string, offset int) {}", "func (o LookupRegionNetworkEndpointGroupResultOutput) Zone() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupRegionNetworkEndpointGroupResult) string { return v.Zone }).(pulumi.StringOutput)\n}", "func (p *Provider) CreateZone(zone dns.Zone) error {\n return fmt.Errorf(\"NOT IMPLEMENTED\")\n}", "func (a *AzureInfoer) GetZones(region string) ([]string, error) {\n\treturn []string{region}, nil\n}", "func (ip IP) Zone() string {\n\tif ip.z == nil {\n\t\treturn \"\"\n\t}\n\tzone, _ := ip.z.Get().(string)\n\treturn zone\n}", "func (o RegionAutoscalerOutput) Zone() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *RegionAutoscaler) pulumi.StringOutput { return v.Zone }).(pulumi.StringOutput)\n}", "func (o InstanceMemcacheNodeOutput) Zone() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v InstanceMemcacheNode) *string { return v.Zone }).(pulumi.StringPtrOutput)\n}", "func (o LiteSubscriptionOutput) Zone() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *LiteSubscription) pulumi.StringPtrOutput { return v.Zone }).(pulumi.StringPtrOutput)\n}", "func (qc *QingCloud) GetZoneByNodeName(ctx context.Context, nodeName types.NodeName) (cloudprovider.Zone, error) {\n\tklog.V(4).Infof(\"GetZoneByNodeName() called, current zone is %v, and return zone directly as temporary solution\", qc.zone)\n\treturn cloudprovider.Zone{Region: qc.zone}, nil\n}", "func (o InstanceFromTemplateOutput) Zone() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *InstanceFromTemplate) pulumi.StringOutput { return v.Zone }).(pulumi.StringOutput)\n}", "func (d *DB) Get_timezone(ipaddress string) (IP2Locationrecord, error) {\n\treturn d.query(ipaddress, timezone)\n}", "func (z *zones) GetZoneByNodeName(ctx context.Context, nodeName types.NodeName) (cloudprovider.Zone, error) {\n\tklog.V(5).Infof(\"GetZoneByNodeName(%v)\", nodeName)\n\treturn cloudprovider.Zone{\n\t\tFailureDomain: \"virtualbox\",\n\t\tRegion: \"virtualbox\",\n\t}, nil\n}", "func (o InstanceNodeOutput) Zone() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v InstanceNode) *string { return v.Zone }).(pulumi.StringPtrOutput)\n}", "func (o StorageNodeStatusGeographyPtrOutput) Zone() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *StorageNodeStatusGeography) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Zone\n\t}).(pulumi.StringPtrOutput)\n}", "func GetZones(timezone string) (zones []Zone, err error) {\n\tif !dbOpen {\n\t\treturn nil, noDB\n\t}\n\n\t// get id of original timezone from replicas' table\n\tprotoID, err := getReplicaOriginal(timezone)\n\tif err != nil {\n\t\t// cannot find original TZ for specified replica\n\t\treturn nil, err\n\t}\n\n\t// get all data for original timezone\n\toriginal, err := getOriginalByID(protoID)\n\tif err != nil {\n\t\t// cannot find data for original TZ\n\t\treturn nil, err\n\t}\n\n\t// check all available sub-tables with zones\n\t// start from the most recent -- the last one\n\t// stop when a reliable table is found\n\ttableOk := false\n\tfor i := 0; i < 3; i++ {\n\t\tzoneTable := fmt.Sprintf(\"%s%v\", original.TabName, original.TabVer-int64(i))\n\t\tzones, err = getZones(zoneTable)\n\t\tif err != nil {\n\t\t\t// zone table unreliable\n\t\t\tcontinue\n\t\t}\n\t\ttableOk = true\n\t\tbreak\n\t}\n\n\tif !tableOk {\n\t\treturn nil, fmt.Errorf(\"tzdb: cannot find reliable table with zones\")\n\t}\n\n\treturn zones, nil\n}", "func (m *MockAll) Zone() Zone {\n\tret := m.ctrl.Call(m, \"Zone\")\n\tret0, _ := ret[0].(Zone)\n\treturn ret0\n}", "func (o *ClusterUninstaller) getZoneName(zoneURL string) string {\n\treturn getNameFromURL(\"zones\", zoneURL)\n}", "func (db *DB) GetTimezone(ip string) (*Record, error) { return db.query(ip, ModeTimeZone) }", "func (mr *MockclientMockRecorder) GetZone(arg0, arg1, arg2 interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"GetZone\", reflect.TypeOf((*Mockclient)(nil).GetZone), arg0, arg1, arg2)\n}", "func (client *AwsClientWrapper) GetAvailabilityZone(machine *clusterv1alpha1.Machine) (string, error) {\n\tinstance, err := getRunningInstance(machine, client.client)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif instance.Placement == nil {\n\t\treturn \"\", err\n\t}\n\treturn *instance.Placement.AvailabilityZone, nil\n}", "func (tzType TimeZoneType) SubZone() TimeZoneType { return TimeZoneType(3)}", "func (s dnsManagedZoneNamespaceLister) Get(name string) (*v1alpha1.DnsManagedZone, error) {\n\tobj, exists, err := s.indexer.GetByKey(s.namespace + \"/\" + name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif !exists {\n\t\treturn nil, errors.NewNotFound(v1alpha1.Resource(\"dnsmanagedzone\"), name)\n\t}\n\treturn obj.(*v1alpha1.DnsManagedZone), nil\n}", "func (region Region) GetZones() []Zone {\n\tswitch region {\n\tcase RegionFrPar:\n\t\treturn []Zone{ZoneFrPar1, ZoneFrPar2}\n\tcase RegionNlAms:\n\t\treturn []Zone{ZoneNlAms1}\n\tdefault:\n\t\treturn []Zone{}\n\t}\n}", "func (u *UpCloud) GetZones() (z *[]Zone, err error) {\n\tvar resp getZonesResponse\n\t// Make request to \"Get Zones\" route\n\tif err = u.request(\"GET\", RouteGetZone, nil, nil, &resp); err != nil {\n\t\treturn\n\t}\n\n\t// Set return value from response\n\tz = resp.Zones.Zone\n\treturn\n}", "func (o InstanceGroupNamedPortOutput) Zone() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *InstanceGroupNamedPort) pulumi.StringOutput { return v.Zone }).(pulumi.StringOutput)\n}", "func (p *PowerDNS) GetZones() ([]Zone, error) {\n\tzones := make([]Zone, 0)\n\tmyError := new(Error)\n\tzonesSling := p.makeSling()\n\tresp, err := zonesSling.New().Get(\"servers/\"+p.VHost+\"/zones\").Receive(&zones, myError)\n\n\tif err == nil && resp.StatusCode >= 400 {\n\t\tmyError.Message = strings.Join([]string{resp.Status, myError.Message}, \" \")\n\t\treturn nil, myError\n\t}\n\n\tfor i := range zones {\n\t\tzones[i].PowerDNSHandle = p\n\t}\n\n\treturn zones, err\n}", "func (m *MockZone) Get(zoneKey api.ZoneKey) (api.Zone, error) {\n\tret := m.ctrl.Call(m, \"Get\", zoneKey)\n\tret0, _ := ret[0].(api.Zone)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}" ]
[ "0.78499573", "0.7834801", "0.74949396", "0.736328", "0.7176127", "0.71676284", "0.7167595", "0.7052328", "0.69505405", "0.69460076", "0.6943429", "0.68824834", "0.6835612", "0.6757427", "0.6687132", "0.6655487", "0.6633087", "0.6619686", "0.66116744", "0.65984404", "0.6585497", "0.6543774", "0.6542547", "0.6497235", "0.64834404", "0.64634806", "0.6413366", "0.6410131", "0.6406876", "0.6405706", "0.640504", "0.64046323", "0.63811094", "0.63676846", "0.63591903", "0.63519746", "0.6344165", "0.63438195", "0.6329925", "0.6317873", "0.63031995", "0.6296737", "0.6289045", "0.6288988", "0.6259943", "0.62221444", "0.62142086", "0.6190616", "0.6181293", "0.6177052", "0.616461", "0.61574733", "0.6147776", "0.6135739", "0.6100323", "0.6082226", "0.60743934", "0.6073329", "0.60691774", "0.60527027", "0.60378724", "0.60151565", "0.5988934", "0.5974314", "0.5970899", "0.59676915", "0.5956736", "0.5945278", "0.59352547", "0.59298897", "0.5927081", "0.5913036", "0.5907481", "0.5894981", "0.58863425", "0.58792543", "0.58775795", "0.58669496", "0.5861288", "0.58355844", "0.5808753", "0.5806268", "0.5798142", "0.5780063", "0.57791454", "0.5762597", "0.57572067", "0.5749435", "0.57445556", "0.57390165", "0.5729333", "0.57163525", "0.57026476", "0.5691476", "0.5686258", "0.5673872", "0.5655887", "0.5653666", "0.5647129", "0.56318074" ]
0.73757464
3
ListZones uses the override method ListZonesFn or the real implementation.
func (c *TestClient) ListZones(project string, opts ...ListCallOption) ([]*compute.Zone, error) { if c.ListZonesFn != nil { return c.ListZonesFn(project, opts...) } return c.client.ListZones(project, opts...) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (client DnsClient) ListZones(ctx context.Context, request ListZonesRequest) (response ListZonesResponse, err error) {\n\tvar ociResponse common.OCIResponse\n\tpolicy := common.NoRetryPolicy()\n\tif request.RetryPolicy() != nil {\n\t\tpolicy = *request.RetryPolicy()\n\t}\n\tociResponse, err = common.Retry(ctx, request, client.listZones, policy)\n\tif err != nil {\n\t\tif ociResponse != nil {\n\t\t\tif httpResponse := ociResponse.HTTPResponse(); httpResponse != nil {\n\t\t\t\topcRequestId := httpResponse.Header.Get(\"opc-request-id\")\n\t\t\t\tresponse = ListZonesResponse{RawResponse: httpResponse, OpcRequestId: &opcRequestId}\n\t\t\t} else {\n\t\t\t\tresponse = ListZonesResponse{}\n\t\t\t}\n\t\t}\n\t\treturn\n\t}\n\tif convertedResponse, ok := ociResponse.(ListZonesResponse); ok {\n\t\tresponse = convertedResponse\n\t} else {\n\t\terr = fmt.Errorf(\"failed to convert OCIResponse into ListZonesResponse\")\n\t}\n\treturn\n}", "func ListZones(ctx context.Context, tenantID int, offset, limit int) (*handlers.ZoneList, error) {\n\treq := handlers.ListZonesRequest{\n\t\tTenantID: tenantID,\n\t\tPaging: database.Paging{\n\t\t\tLimit: limit,\n\t\t\tOffset: offset,\n\t\t},\n\t\tZoneOptions: database.ZoneOptions{\n\t\t\tWithPlaces: true,\n\t\t},\n\t}\n\tresp := handlers.ZoneList{}\n\terr := jsonConn.RequestWithContext(ctx, \"zones.list\", &req, &resp)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func (z *zones) List() ([]dnsprovider.Zone, error) {\n\tsnapshot := z.dnsView.Snapshot()\n\n\tvar zones []dnsprovider.Zone\n\tzoneInfos := snapshot.ListZones()\n\tfor i := range zoneInfos {\n\t\tzones = append(zones, &zone{dnsView: z.dnsView, zoneInfo: zoneInfos[i]})\n\t}\n\treturn zones, nil\n}", "func (z *FakeZoneLister) ListZones(_ utils.NodeConditionPredicate) ([]string, error) {\n\treturn z.Zones, nil\n}", "func (client DnsClient) listZones(ctx context.Context, request common.OCIRequest) (common.OCIResponse, error) {\n\thttpRequest, err := request.HTTPRequest(http.MethodGet, \"/zones\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar response ListZonesResponse\n\tvar httpResponse *http.Response\n\thttpResponse, err = client.Call(ctx, &httpRequest)\n\tdefer common.CloseBodyIfValid(httpResponse)\n\tresponse.RawResponse = httpResponse\n\tif err != nil {\n\t\treturn response, err\n\t}\n\n\terr = common.UnmarshalResponse(httpResponse, &response)\n\treturn response, err\n}", "func (api *powerdnsProvider) ListZones() ([]string, error) {\n\tvar result []string\n\tmyZones, err := api.client.Zones().ListZones(context.Background(), api.ServerName)\n\tif err != nil {\n\t\treturn result, err\n\t}\n\tfor _, zone := range myZones {\n\t\tresult = append(result, zone.Name)\n\t}\n\treturn result, nil\n}", "func (s *FastDNSv2Service) ListZones(ctx context.Context, opt *ZoneListOptions) (*ZoneList, *Response, error) {\n\tu := fmt.Sprintf(\"config-dns/v2/zones\")\n\tu, err := addOptions(u, opt)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treq, err := s.client.NewRequest(\"GET\", u, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tvar zones *ZoneList\n\tresp, err := s.client.Do(ctx, req, &zones)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\n\treturn zones, resp, nil\n}", "func (p *Provider) ListZones() ([]string, error) {\n mZones, err := p.listMZones()\n if err != nil {\n return nil, fmt.Errorf(\"couldn't retrieve zone information from autodns, see: %v\", err)\n }\n\n zones := make([]string, 0)\n\n for _, mZone := range mZones {\n if mZone.Origin == nil {\n logrus.Warnf(\"found zone without origin, since we use it as unique identifier, this shouldn't have happend, see: %#v\", mZone)\n continue\n }\n\n zones = append(zones, *mZone.Origin)\n }\n\n return zones, nil\n}", "func ListZones() (zones Zones, err error) {\n\tcfz, err := api.ListZones()\n\tif err != nil {\n\t\treturn\n\t}\n\tzones.translateFromCloudflare(cfz)\n\treturn\n}", "func ListZones(r *route53.Route53) {\n\tresp, err := r.ListHostedZones(&route53.ListHostedZonesRequest{})\n\tif err != nil {\n\t\tlog.Fatalln(err)\n\t}\n\tfmt.Println(*resp)\n}", "func (s stack) ListAvailabilityZones(ctx context.Context) (_ map[string]bool, ferr fail.Error) {\n\tif valid.IsNil(s) {\n\t\treturn nil, fail.InvalidInstanceError()\n\t}\n\n\treturn nil, fail.NotImplementedError(\"implement me\")\n}", "func (s *API) ListDNSZones(req *ListDNSZonesRequest, opts ...scw.RequestOption) (*ListDNSZonesResponse, error) {\n\tvar err error\n\n\tdefaultPageSize, exist := s.client.GetDefaultPageSize()\n\tif (req.PageSize == nil || *req.PageSize == 0) && exist {\n\t\treq.PageSize = &defaultPageSize\n\t}\n\n\tquery := url.Values{}\n\tparameter.AddToQuery(query, \"page\", req.Page)\n\tparameter.AddToQuery(query, \"page_size\", req.PageSize)\n\tparameter.AddToQuery(query, \"order_by\", req.OrderBy)\n\tparameter.AddToQuery(query, \"domain\", req.Domain)\n\tparameter.AddToQuery(query, \"dns_zone\", req.DNSZone)\n\tparameter.AddToQuery(query, \"organization_id\", req.OrganizationID)\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"GET\",\n\t\tPath: \"/domain/v2alpha2/dns-zones\",\n\t\tQuery: query,\n\t\tHeaders: http.Header{},\n\t}\n\n\tvar resp ListDNSZonesResponse\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func List(client *gophercloud.ServiceClient, opts ListOptsBuilder) pagination.Pager {\n\turl := baseURL(client)\n\tif opts != nil {\n\t\tquery, err := opts.ToZoneListQuery()\n\t\tif err != nil {\n\t\t\treturn pagination.Pager{Err: err}\n\t\t}\n\t\turl += query\n\t}\n\treturn pagination.NewPager(client, url, func(r pagination.PageResult) pagination.Page {\n\t\treturn ZonePage{pagination.LinkedPageBase{PageResult: r}}\n\t})\n}", "func (s *API) ListDNSZones(req *ListDNSZonesRequest, opts ...scw.RequestOption) (*ListDNSZonesResponse, error) {\n\tvar err error\n\n\tdefaultPageSize, exist := s.client.GetDefaultPageSize()\n\tif (req.PageSize == nil || *req.PageSize == 0) && exist {\n\t\treq.PageSize = &defaultPageSize\n\t}\n\n\tquery := url.Values{}\n\tparameter.AddToQuery(query, \"organization_id\", req.OrganizationID)\n\tparameter.AddToQuery(query, \"project_id\", req.ProjectID)\n\tparameter.AddToQuery(query, \"order_by\", req.OrderBy)\n\tparameter.AddToQuery(query, \"page\", req.Page)\n\tparameter.AddToQuery(query, \"page_size\", req.PageSize)\n\tparameter.AddToQuery(query, \"domain\", req.Domain)\n\tparameter.AddToQuery(query, \"dns_zone\", req.DNSZone)\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"GET\",\n\t\tPath: \"/domain/v2beta1/dns-zones\",\n\t\tQuery: query,\n\t\tHeaders: http.Header{},\n\t}\n\n\tvar resp ListDNSZonesResponse\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func (c *Client) ListFlavorZones(args *ListFlavorZonesArgs) (*ListZonesResult, error) {\n\tjsonBytes, jsonErr := json.Marshal(args)\n\tif jsonErr != nil {\n\t\treturn nil, jsonErr\n\t}\n\tbody, err := bce.NewBodyFromBytes(jsonBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn ListFlavorZones(c, body)\n}", "func (c *PDNSAPIClient) ListZones() (zones []pgo.Zone, resp *http.Response, err error) {\n\tfor i := 0; i < retryLimit; i++ {\n\t\tzones, resp, err = c.client.ZonesApi.ListZones(c.authCtx, defaultServerID)\n\t\tif err != nil {\n\t\t\tlog.Debugf(\"Unable to fetch zones %v\", err)\n\t\t\tlog.Debugf(\"Retrying ListZones() ... %d\", i)\n\t\t\ttime.Sleep(retryAfterTime * (1 << uint(i)))\n\t\t\tcontinue\n\t\t}\n\t\treturn zones, resp, err\n\t}\n\n\tlog.Errorf(\"Unable to fetch zones. %v\", err)\n\treturn zones, resp, err\n}", "func (z *Zones) ListZones() []ClusterZone {\n\tzones := []ClusterZone{}\n\tz.db.DB().Find(&zones)\n\treturn zones\n}", "func (a *Client) ListTransportZones(params *ListTransportZonesParams, authInfo runtime.ClientAuthInfoWriter) (*ListTransportZonesOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewListTransportZonesParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"ListTransportZones\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/transport-zones\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &ListTransportZonesReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*ListTransportZonesOK), nil\n\n}", "func (s *dnsManagedZoneLister) List(selector labels.Selector) (ret []*v1alpha1.DnsManagedZone, err error) {\n\terr = cache.ListAll(s.indexer, selector, func(m interface{}) {\n\t\tret = append(ret, m.(*v1alpha1.DnsManagedZone))\n\t})\n\treturn ret, err\n}", "func (c *PDNSAPIClient) ListZone(zoneID string) (zone pgo.Zone, resp *http.Response, err error) {\n\tfor i := 0; i < retryLimit; i++ {\n\t\tzone, resp, err = c.client.ZonesApi.ListZone(c.authCtx, defaultServerID, zoneID)\n\t\tif err != nil {\n\t\t\tlog.Debugf(\"Unable to fetch zone %v\", err)\n\t\t\tlog.Debugf(\"Retrying ListZone() ... %d\", i)\n\t\t\ttime.Sleep(retryAfterTime * (1 << uint(i)))\n\t\t\tcontinue\n\t\t}\n\t\treturn zone, resp, err\n\t}\n\n\tlog.Errorf(\"Unable to list zone. %v\", err)\n\treturn zone, resp, err\n}", "func (p *PowerDNS) GetZones() ([]Zone, error) {\n\tzones := make([]Zone, 0)\n\tmyError := new(Error)\n\tzonesSling := p.makeSling()\n\tresp, err := zonesSling.New().Get(\"servers/\"+p.VHost+\"/zones\").Receive(&zones, myError)\n\n\tif err == nil && resp.StatusCode >= 400 {\n\t\tmyError.Message = strings.Join([]string{resp.Status, myError.Message}, \" \")\n\t\treturn nil, myError\n\t}\n\n\tfor i := range zones {\n\t\tzones[i].PowerDNSHandle = p\n\t}\n\n\treturn zones, err\n}", "func (c *Client) DescribeZones(request *DescribeZonesRequest) (response *DescribeZonesResponse, err error) {\n return c.DescribeZonesWithContext(context.Background(), request)\n}", "func (a *AzureInfoer) GetZones(region string) ([]string, error) {\n\tlogger := a.log.WithFields(map[string]interface{}{\"region\": region})\n\tlogger.Debug(\"getting zones\")\n\n\tskusResultPage, err := a.skusClient.List(context.Background())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tzonesMap := make(map[string]string)\n\tzones := make([]string, 0)\n\n\tfor _, sku := range skusResultPage.Values() {\n\t\tfor _, locationInfo := range *sku.LocationInfo {\n\t\t\tif strings.ToLower(*locationInfo.Location) == region {\n\t\t\t\t// retrieve zones per instance type\n\t\t\t\tfor _, zone := range *locationInfo.Zones {\n\t\t\t\t\tzonesMap[zone] = \"\"\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\tfor zone := range zonesMap {\n\t\tzones = append(zones, zone)\n\t}\n\n\tlogger.Debug(\"found zones\", map[string]interface{}{\"numberOfZones\": len(zones)})\n\n\treturn zones, nil\n}", "func (c *ClientWithResponses) ListZonesWithResponse(ctx context.Context) (*ListZonesResponse, error) {\n\trsp, err := c.ListZones(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn ParseListZonesResponse(rsp)\n}", "func (a *AzureInfoer) GetZones(region string) ([]string, error) {\n\treturn []string{region}, nil\n}", "func NewListZonesRequest(server string) (*http.Request, error) {\n\tvar err error\n\n\tqueryUrl, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbasePath := fmt.Sprintf(\"/zone\")\n\tif basePath[0] == '/' {\n\t\tbasePath = basePath[1:]\n\t}\n\n\tqueryUrl, err = queryUrl.Parse(basePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq, err := http.NewRequest(\"GET\", queryUrl.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn req, nil\n}", "func GetZones(full bool, tenant string) []Zone {\n\ttenantStr := func() string {\n\t\tif len(tenant) == 0 {\n\t\t\treturn \"\"\n\t\t}\n\t\treturn \"-tenant=\" + tenant\n\t}()\n\tfullStr := func() string {\n\t\tif full {\n\t\t\treturn \"-full\"\n\t\t}\n\t\treturn \"\"\n\t}()\n\n\toutput := RunCmd(fmt.Sprintf(\"%s api -fetch-zone-apps %s %s\", ActlPath, fullStr, tenantStr))\n\tlistOfZones := []Zone{}\n\tyaml.Unmarshal([]byte(output), &listOfZones)\n\treturn listOfZones\n}", "func (w *ServerInterfaceWrapper) GetZones(ctx echo.Context) error {\n\tvar err error\n\n\t// Invoke the callback with all the unmarshalled arguments\n\terr = w.Handler.GetZones(ctx)\n\treturn err\n}", "func NewListDNSZonesRequest(server string, params *ListDNSZonesParams) (*http.Request, error) {\n\tvar err error\n\n\tqueryUrl, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbasePath := fmt.Sprintf(\"/domain/v2alpha2/dns-zones\")\n\tif basePath[0] == '/' {\n\t\tbasePath = basePath[1:]\n\t}\n\n\tqueryUrl, err = queryUrl.Parse(basePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tqueryValues := queryUrl.Query()\n\n\tif params.Page != nil {\n\n\t\tif queryFrag, err := runtime.StyleParam(\"form\", true, \"page\", *params.Page); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.PageSize != nil {\n\n\t\tif queryFrag, err := runtime.StyleParam(\"form\", true, \"page_size\", *params.PageSize); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.OrderBy != nil {\n\n\t\tif queryFrag, err := runtime.StyleParam(\"form\", true, \"order_by\", *params.OrderBy); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.Domain != nil {\n\n\t\tif queryFrag, err := runtime.StyleParam(\"form\", true, \"domain\", *params.Domain); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.DnsZone != nil {\n\n\t\tif queryFrag, err := runtime.StyleParam(\"form\", true, \"dns_zone\", *params.DnsZone); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif params.OrganizationId != nil {\n\n\t\tif queryFrag, err := runtime.StyleParam(\"form\", true, \"organization_id\", *params.OrganizationId); err != nil {\n\t\t\treturn nil, err\n\t\t} else if parsed, err := url.ParseQuery(queryFrag); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tfor k, v := range parsed {\n\t\t\t\tfor _, v2 := range v {\n\t\t\t\t\tqueryValues.Add(k, v2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\tqueryUrl.RawQuery = queryValues.Encode()\n\n\treq, err := http.NewRequest(\"GET\", queryUrl.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn req, nil\n}", "func (c *Client) ListZoneFlavors(args *ListZoneFlavorsArgs) (*ListFlavorInfosResult, error) {\n\tjsonBytes, jsonErr := json.Marshal(args)\n\tif jsonErr != nil {\n\t\treturn nil, jsonErr\n\t}\n\tbody, err := bce.NewBodyFromBytes(jsonBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn ListZoneFlavors(c, body)\n}", "func ListTimezonesHandler(w http.ResponseWriter, r *http.Request) {\n\ttimezones, err := database.GetAllTimezones()\n\tif err != nil {\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tresults, _ := json.Marshal(timezones)\n\n\tw.WriteHeader(http.StatusOK)\n\tw.Write(results)\n}", "func (s dnsManagedZoneNamespaceLister) List(selector labels.Selector) (ret []*v1alpha1.DnsManagedZone, err error) {\n\terr = cache.ListAllByNamespace(s.indexer, s.namespace, selector, func(m interface{}) {\n\t\tret = append(ret, m.(*v1alpha1.DnsManagedZone))\n\t})\n\treturn ret, err\n}", "func (m *MockClient) ListHostedZones(input *route53.ListHostedZonesInput) (*route53.ListHostedZonesOutput, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"ListHostedZones\", input)\n\tret0, _ := ret[0].(*route53.ListHostedZonesOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func listURL(c *gophercloud.ServiceClient) string {\n\treturn c.ServiceURL(\"os-availability-zone\")\n}", "func GetZones(sess *session.Session) (*ec2.DescribeAvailabilityZonesOutput, error) {\n svc := ec2.New(sess)\n\n // snippet-start:[ec2.go.regions_and_zones.zones]\n resultAvalZones, err := svc.DescribeAvailabilityZones(nil)\n // snippet-end:[ec2.go.regions_and_zones.zones]\n if err != nil {\n return nil, err\n }\n\n return resultAvalZones, nil\n}", "func (a *Client) ListOpenstackAvailabilityZones(params *ListOpenstackAvailabilityZonesParams, authInfo runtime.ClientAuthInfoWriter, opts ...ClientOption) (*ListOpenstackAvailabilityZonesOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewListOpenstackAvailabilityZonesParams()\n\t}\n\top := &runtime.ClientOperation{\n\t\tID: \"listOpenstackAvailabilityZones\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/api/v1/providers/openstack/availabilityzones\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &ListOpenstackAvailabilityZonesReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t}\n\tfor _, opt := range opts {\n\t\topt(op)\n\t}\n\n\tresult, err := a.transport.Submit(op)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*ListOpenstackAvailabilityZonesOK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\tunexpectedSuccess := result.(*ListOpenstackAvailabilityZonesDefault)\n\treturn nil, runtime.NewAPIError(\"unexpected success response: content available as default response in error\", unexpectedSuccess, unexpectedSuccess.Code())\n}", "func (c APIClient) GetZones() ([]Zone, error) {\n\tvar zl []Zone\n\t_, err := c.doHTTPUnmarshal(\"GET\", \"https://api.nsone.net/v1/zones\", nil, &zl)\n\treturn zl, err\n}", "func (s *InMemoryInstanceTypeStore) GetZones(provider, service, region string) ([]string, error) {\n\treturn []string{}, nil\n}", "func GetZones(client *Client, userHome *UserHome) ([]*Zone, error) {\n\tresp, err := client.Request(http.MethodGet, apiURL(\"homes/%d/zones\", userHome.ID), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer resp.Body.Close()\n\n\tif err := isError(resp); err != nil {\n\t\treturn nil, fmt.Errorf(\"tado° API error: %w\", err)\n\t}\n\n\tzones := make([]*Zone, 0)\n\tif err := json.NewDecoder(resp.Body).Decode(&zones); err != nil {\n\t\treturn nil, fmt.Errorf(\"unable to decode tado° API response: %w\", err)\n\t}\n\n\treturn zones, nil\n}", "func (s *Server) GetZones(ctx context.Context, in *GetZonesRequest) (*GetZonesResponse, error) {\n\tvar zones []*Zone\n\tfor _, z := range s.service.GetZones() {\n\t\tvar speakers []*Speaker\n\t\tfor _, member := range z.Speakers {\n\t\t\tspeaker := &Speaker{Id: member.ID, DisplayName: member.DisplayName}\n\t\t\tspeakers = append(speakers, speaker)\n\t\t}\n\t\tzones = append(zones, &Zone{DisplayName: z.DisplayName, Id: z.ID, Speakers: speakers})\n\t}\n\treturn &GetZonesResponse{ReturnCode: 200, Zones: zones}, nil\n}", "func (s *API) ListDNSZoneRecords(req *ListDNSZoneRecordsRequest, opts ...scw.RequestOption) (*ListDNSZoneRecordsResponse, error) {\n\tvar err error\n\n\tdefaultPageSize, exist := s.client.GetDefaultPageSize()\n\tif (req.PageSize == nil || *req.PageSize == 0) && exist {\n\t\treq.PageSize = &defaultPageSize\n\t}\n\n\tquery := url.Values{}\n\tparameter.AddToQuery(query, \"page\", req.Page)\n\tparameter.AddToQuery(query, \"page_size\", req.PageSize)\n\tparameter.AddToQuery(query, \"order_by\", req.OrderBy)\n\tparameter.AddToQuery(query, \"name\", req.Name)\n\tparameter.AddToQuery(query, \"type\", req.Type)\n\tparameter.AddToQuery(query, \"organization_id\", req.OrganizationID)\n\n\tif fmt.Sprint(req.DNSZone) == \"\" {\n\t\treturn nil, errors.New(\"field DNSZone cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"GET\",\n\t\tPath: \"/domain/v2alpha2/dns-zones/\" + fmt.Sprint(req.DNSZone) + \"/records\",\n\t\tQuery: query,\n\t\tHeaders: http.Header{},\n\t}\n\n\tvar resp ListDNSZoneRecordsResponse\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func AvailabilityZones(project, region string) ([]string, error) {\n\tctx, cancel := context.WithTimeout(context.Background(), 1*time.Minute)\n\tdefer cancel()\n\n\tssn, err := gcpconfig.GetSession(ctx)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to get session\")\n\t}\n\n\tsvc, err := compute.NewService(ctx, option.WithCredentials(ssn.Credentials))\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to create compute service\")\n\t}\n\n\tregionURL := fmt.Sprintf(\"https://www.googleapis.com/compute/v1/projects/%s/regions/%s\",\n\t\tproject, region)\n\treq := svc.Zones.List(project).Filter(fmt.Sprintf(\"(region eq %s) (status eq UP)\", regionURL))\n\n\tvar zones []string\n\tif err := req.Pages(ctx, func(page *compute.ZoneList) error {\n\t\tfor _, z := range page.Items {\n\t\t\tzones = append(zones, z.Name)\n\t\t}\n\t\treturn nil\n\t}); err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to list zones\")\n\t}\n\tif len(zones) == 0 {\n\t\treturn nil, errors.New(\"no zone was found\")\n\t}\n\n\tsort.Strings(zones)\n\treturn zones, nil\n}", "func (m *MockClient) ListHostedZones(arg0 *route53.ListHostedZonesInput) (*route53.ListHostedZonesOutput, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"ListHostedZones\", arg0)\n\tret0, _ := ret[0].(*route53.ListHostedZonesOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (mr *MockClientMockRecorder) ListHostedZones(input interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"ListHostedZones\", reflect.TypeOf((*MockClient)(nil).ListHostedZones), input)\n}", "func (s *API) ListDNSZoneRecords(req *ListDNSZoneRecordsRequest, opts ...scw.RequestOption) (*ListDNSZoneRecordsResponse, error) {\n\tvar err error\n\n\tdefaultPageSize, exist := s.client.GetDefaultPageSize()\n\tif (req.PageSize == nil || *req.PageSize == 0) && exist {\n\t\treq.PageSize = &defaultPageSize\n\t}\n\n\tquery := url.Values{}\n\tparameter.AddToQuery(query, \"project_id\", req.ProjectID)\n\tparameter.AddToQuery(query, \"order_by\", req.OrderBy)\n\tparameter.AddToQuery(query, \"page\", req.Page)\n\tparameter.AddToQuery(query, \"page_size\", req.PageSize)\n\tparameter.AddToQuery(query, \"name\", req.Name)\n\tparameter.AddToQuery(query, \"type\", req.Type)\n\tparameter.AddToQuery(query, \"id\", req.ID)\n\n\tif fmt.Sprint(req.DNSZone) == \"\" {\n\t\treturn nil, errors.New(\"field DNSZone cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"GET\",\n\t\tPath: \"/domain/v2beta1/dns-zones/\" + fmt.Sprint(req.DNSZone) + \"/records\",\n\t\tQuery: query,\n\t\tHeaders: http.Header{},\n\t}\n\n\tvar resp ListDNSZoneRecordsResponse\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func (mr *MockClientMockRecorder) ListHostedZones(arg0 interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"ListHostedZones\", reflect.TypeOf((*MockClient)(nil).ListHostedZones), arg0)\n}", "func (c *Client) ListScanTimezones(httpClient *http.Client) (ScanTimezones, error) {\n\tc.debugln(\"ScanTimezones(): Building scan timezone list URL\")\n\turl := fmt.Sprintf(\"https://%s:%s/scans/timezones\", c.ip, c.port)\n\n\tstatusCode, body, err := c.get(httpClient, url)\n\tif err != nil {\n\t\treturn ScanTimezones{}, err\n\t}\n\n\tswitch statusCode {\n\tcase 200:\n\t\tvar scanTimezones ScanTimezones\n\t\terr = json.Unmarshal(body, &scanTimezones)\n\t\tif err != nil {\n\t\t\treturn ScanTimezones{}, err\n\t\t}\n\t\tc.debugln(\"ScanTimezones(): Successfully retrieved list of scan timezones.\")\n\t\treturn scanTimezones, nil\n\tdefault:\n\t\tvar err ErrorResponse\n\t\tunmarshalError := json.Unmarshal(body, &err)\n\t\tif unmarshalError != nil {\n\t\t\treturn ScanTimezones{}, unmarshalError\n\t\t}\n\t\tc.debugln(\"ScanTimezones(): Scan timezones list could not be retrieved.\")\n\t\treturn ScanTimezones{}, fmt.Errorf(\"%s\", err.Error)\n\t}\n}", "func listRecords(client *dnsimple.Client, accountID, domain string,\n\toptions *dnsimple.ZoneRecordListOptions) (records zoneRecords, err error) {\n\tif options == nil {\n\t\toptions = &dnsimple.ZoneRecordListOptions{}\n\t}\n\tfor p := 1; ; p++ {\n\t\tlistZoneRecordsResponse, err := client.Zones.ListRecords(accountID, domain, options)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tfor i := range listZoneRecordsResponse.Data {\n\t\t\trecords = append(records, listZoneRecordsResponse.Data[i])\n\t\t}\n\t\tif options.Page == 0 {\n\t\t\toptions.Page = 2\n\t\t} else {\n\t\t\toptions.Page++\n\t\t}\n\t\tif p >= listZoneRecordsResponse.Pagination.TotalPages {\n\t\t\tbreak\n\t\t}\n\t}\n\treturn\n}", "func (s *API) ListDNSZoneVersions(req *ListDNSZoneVersionsRequest, opts ...scw.RequestOption) (*ListDNSZoneVersionsResponse, error) {\n\tvar err error\n\n\tquery := url.Values{}\n\tparameter.AddToQuery(query, \"organization_id\", req.OrganizationID)\n\n\tif fmt.Sprint(req.DNSZone) == \"\" {\n\t\treturn nil, errors.New(\"field DNSZone cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"GET\",\n\t\tPath: \"/domain/v2alpha2/dns-zones/\" + fmt.Sprint(req.DNSZone) + \"/versions\",\n\t\tQuery: query,\n\t\tHeaders: http.Header{},\n\t}\n\n\tvar resp ListDNSZoneVersionsResponse\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func (m *manager) List() ([]string, error) {\n\tvar igs []*compute.InstanceGroup\n\n\tzones, err := m.ListZones(utils.AllNodesPredicate)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfor _, zone := range zones {\n\t\tigsForZone, err := m.cloud.ListInstanceGroups(zone)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tfor _, ig := range igsForZone {\n\t\t\tigs = append(igs, ig)\n\t\t}\n\t}\n\n\tvar names []string\n\tfor _, ig := range igs {\n\t\tif m.namer.NameBelongsToCluster(ig.Name) {\n\t\t\tnames = append(names, ig.Name)\n\t\t}\n\t}\n\n\treturn names, nil\n}", "func (c *ClientWithResponses) ListDNSZonesWithResponse(ctx context.Context, params *ListDNSZonesParams) (*ListDNSZonesResponse, error) {\n\trsp, err := c.ListDNSZones(ctx, params)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn ParseListDNSZonesResponse(rsp)\n}", "func ParseListZonesResponse(rsp *http.Response) (*ListZonesResponse, error) {\n\tbodyBytes, err := ioutil.ReadAll(rsp.Body)\n\tdefer rsp.Body.Close()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresponse := &ListZonesResponse{\n\t\tBody: bodyBytes,\n\t\tHTTPResponse: rsp,\n\t}\n\n\tswitch {\n\tcase strings.Contains(rsp.Header.Get(\"Content-Type\"), \"json\") && rsp.StatusCode == 200:\n\t\tvar dest struct {\n\t\t\tZones *[]Zone `json:\"zones,omitempty\"`\n\t\t}\n\t\tif err := json.Unmarshal(bodyBytes, &dest); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tresponse.JSON200 = &dest\n\n\t}\n\n\treturn response, nil\n}", "func getZones(zoneTable string) (zones []Zone, err error) {\n\tif !tableExists(zoneTable) {\n\t\treturn nil, fmt.Errorf(\"Table not found!\\n\")\n\t}\n\n\tquery := fmt.Sprintf(\"SELECT * FROM %s\", zoneTable)\n\trows, err := db.Query(query)\n\tdefer rows.Close()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar id, start, end, offset int64\n\tvar name string\n\tvar isDST bool\n\n\tzones = make([]Zone, 0, 5)\n\tfor rows.Next() {\n\t\terr = rows.Scan(&id, &name, &start, &end, &offset, &isDST)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tnewZone := Zone{ID: id, Name: name, Start: start, End: end, Offset: offset, IsDST: isDST}\n\t\tzones = append(zones, newZone)\n\t}\n\n\treturn zones, nil\n}", "func (s *API) ListDNSZoneVersions(req *ListDNSZoneVersionsRequest, opts ...scw.RequestOption) (*ListDNSZoneVersionsResponse, error) {\n\tvar err error\n\n\tdefaultPageSize, exist := s.client.GetDefaultPageSize()\n\tif (req.PageSize == nil || *req.PageSize == 0) && exist {\n\t\treq.PageSize = &defaultPageSize\n\t}\n\n\tquery := url.Values{}\n\tparameter.AddToQuery(query, \"page\", req.Page)\n\tparameter.AddToQuery(query, \"page_size\", req.PageSize)\n\n\tif fmt.Sprint(req.DNSZone) == \"\" {\n\t\treturn nil, errors.New(\"field DNSZone cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"GET\",\n\t\tPath: \"/domain/v2beta1/dns-zones/\" + fmt.Sprint(req.DNSZone) + \"/versions\",\n\t\tQuery: query,\n\t\tHeaders: http.Header{},\n\t}\n\n\tvar resp ListDNSZoneVersionsResponse\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func (c *MockNatGatewaysClient) List(ctx context.Context, resourceGroupName string) ([]network.NatGateway, error) {\n\tvar l []network.NatGateway\n\tfor _, ngw := range c.NGWs {\n\t\tl = append(l, ngw)\n\t}\n\treturn l, nil\n}", "func (a *Client) ListOpenstackAvailabilityZonesNoCredentials(params *ListOpenstackAvailabilityZonesNoCredentialsParams, authInfo runtime.ClientAuthInfoWriter, opts ...ClientOption) (*ListOpenstackAvailabilityZonesNoCredentialsOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewListOpenstackAvailabilityZonesNoCredentialsParams()\n\t}\n\top := &runtime.ClientOperation{\n\t\tID: \"listOpenstackAvailabilityZonesNoCredentials\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/api/v1/projects/{project_id}/dc/{dc}/clusters/{cluster_id}/providers/openstack/availabilityzones\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &ListOpenstackAvailabilityZonesNoCredentialsReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t}\n\tfor _, opt := range opts {\n\t\topt(op)\n\t}\n\n\tresult, err := a.transport.Submit(op)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*ListOpenstackAvailabilityZonesNoCredentialsOK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\tunexpectedSuccess := result.(*ListOpenstackAvailabilityZonesNoCredentialsDefault)\n\treturn nil, runtime.NewAPIError(\"unexpected success response: content available as default response in error\", unexpectedSuccess, unexpectedSuccess.Code())\n}", "func (c *TestClient) ListDisks(project, zone string, opts ...ListCallOption) ([]*compute.Disk, error) {\n\tif c.ListDisksFn != nil {\n\t\treturn c.ListDisksFn(project, zone, opts...)\n\t}\n\treturn c.client.ListDisks(project, zone, opts...)\n}", "func (fsar *FallbackServiceAccountReflector) List() ([]interface{}, error) {\n\treturn virtualkubelet.List[virtualkubelet.Lister[*corev1.Pod], *corev1.Pod](\n\t\tfsar.localPods,\n\t)\n}", "func (u *UpCloud) GetZones() (z *[]Zone, err error) {\n\tvar resp getZonesResponse\n\t// Make request to \"Get Zones\" route\n\tif err = u.request(\"GET\", RouteGetZone, nil, nil, &resp); err != nil {\n\t\treturn\n\t}\n\n\t// Set return value from response\n\tz = resp.Zones.Zone\n\treturn\n}", "func GetZones(timezone string) (zones []Zone, err error) {\n\tif !dbOpen {\n\t\treturn nil, noDB\n\t}\n\n\t// get id of original timezone from replicas' table\n\tprotoID, err := getReplicaOriginal(timezone)\n\tif err != nil {\n\t\t// cannot find original TZ for specified replica\n\t\treturn nil, err\n\t}\n\n\t// get all data for original timezone\n\toriginal, err := getOriginalByID(protoID)\n\tif err != nil {\n\t\t// cannot find data for original TZ\n\t\treturn nil, err\n\t}\n\n\t// check all available sub-tables with zones\n\t// start from the most recent -- the last one\n\t// stop when a reliable table is found\n\ttableOk := false\n\tfor i := 0; i < 3; i++ {\n\t\tzoneTable := fmt.Sprintf(\"%s%v\", original.TabName, original.TabVer-int64(i))\n\t\tzones, err = getZones(zoneTable)\n\t\tif err != nil {\n\t\t\t// zone table unreliable\n\t\t\tcontinue\n\t\t}\n\t\ttableOk = true\n\t\tbreak\n\t}\n\n\tif !tableOk {\n\t\treturn nil, fmt.Errorf(\"tzdb: cannot find reliable table with zones\")\n\t}\n\n\treturn zones, nil\n}", "func (s *API) ListDNSZoneNameservers(req *ListDNSZoneNameserversRequest, opts ...scw.RequestOption) (*ListDNSZoneNameserversResponse, error) {\n\tvar err error\n\n\tquery := url.Values{}\n\tparameter.AddToQuery(query, \"organization_id\", req.OrganizationID)\n\n\tif fmt.Sprint(req.DNSZone) == \"\" {\n\t\treturn nil, errors.New(\"field DNSZone cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"GET\",\n\t\tPath: \"/domain/v2alpha2/dns-zones/\" + fmt.Sprint(req.DNSZone) + \"/nameservers\",\n\t\tQuery: query,\n\t\tHeaders: http.Header{},\n\t}\n\n\tvar resp ListDNSZoneNameserversResponse\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func (c *Client) GetDNSZones(ctx context.Context) ([]DNSZoneResponse, error) {\n\t_, cancel := context.WithTimeout(ctx, 1*time.Minute)\n\tdefer cancel()\n\n\toptions := c.controllerAPI.NewListResourceInstancesOptions()\n\toptions.SetResourceID(cisServiceID)\n\n\tlistResourceInstancesResponse, _, err := c.controllerAPI.ListResourceInstances(options)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to get cis instance\")\n\t}\n\n\tvar allZones []DNSZoneResponse\n\tfor _, instance := range listResourceInstancesResponse.Resources {\n\t\tcrnstr := instance.CRN\n\t\tauthenticator := &core.IamAuthenticator{\n\t\t\tApiKey: c.APIKey,\n\t\t}\n\t\tzonesService, err := zonesv1.NewZonesV1(&zonesv1.ZonesV1Options{\n\t\t\tAuthenticator: authenticator,\n\t\t\tCrn: crnstr,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"failed to list DNS zones\")\n\t\t}\n\n\t\toptions := zonesService.NewListZonesOptions()\n\t\tlistZonesResponse, _, err := zonesService.ListZones(options)\n\n\t\tif listZonesResponse == nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tfor _, zone := range listZonesResponse.Result {\n\t\t\tif *zone.Status == \"active\" {\n\t\t\t\tzoneStruct := DNSZoneResponse{\n\t\t\t\t\tName: *zone.Name,\n\t\t\t\t\tID: *zone.ID,\n\t\t\t\t\tCISInstanceCRN: *instance.CRN,\n\t\t\t\t\tCISInstanceName: *instance.Name,\n\t\t\t\t\tResourceGroupID: *instance.ResourceGroupID,\n\t\t\t\t}\n\t\t\t\tallZones = append(allZones, zoneStruct)\n\t\t\t}\n\t\t}\n\t}\n\n\treturn allZones, nil\n}", "func (f *FakeInstanceGroups) ListInstanceGroups(zone string) ([]*compute.InstanceGroup, error) {\n\tigs := []*compute.InstanceGroup{}\n\tfor ig := range f.zonesToIGsToInstances[zone] {\n\t\tigs = append(igs, ig)\n\t}\n\treturn igs, nil\n}", "func getZones(tokens map[string][]uint32) []string {\n\tvar zones []string\n\n\tfor zone := range tokens {\n\t\tzones = append(zones, zone)\n\t}\n\n\tsort.Strings(zones)\n\treturn zones\n}", "func (p *Provider) List() ([]string, error) {\n\treturn p.provider.ListClusters()\n}", "func (region Region) GetZones() []Zone {\n\tswitch region {\n\tcase RegionFrPar:\n\t\treturn []Zone{ZoneFrPar1, ZoneFrPar2}\n\tcase RegionNlAms:\n\t\treturn []Zone{ZoneNlAms1}\n\tdefault:\n\t\treturn []Zone{}\n\t}\n}", "func (*OktetoClusterHelper) List() (map[string]string, error) {\n\treturn nil, ErrNotImplemented\n}", "func ListService(ctx *gin.Context) {\n\tlog := logger.RuntimeLog\n\tzoneName := ctx.Param(\"zone\")\n\tnamespace := ctx.Param(\"ns\")\n\n\t// fetch k8s-client handler by zoneName\n\tkclient, err := GetClientByAzCode(zoneName)\n\tif err != nil {\n\t\tlog.WithError(err)\n\t\tSendResponse(ctx, errno.ErrTokenInvalid, nil)\n\t\treturn\n\t}\n\n\tstartAt := time.Now()\n\tsvcs, err := kclient.CoreV1().Services(namespace).List(metav1.ListOptions{})\n\tif err != nil {\n\t\tSendResponse(ctx, err, \"failed to get Service info.\")\n\t\treturn\n\t}\n\tlogger.MetricsEmit(\n\t\tSVC_CONST.K8S_LOG_Method_ListService,\n\t\tutil.GetReqID(ctx),\n\t\tfloat32(time.Since(startAt)/time.Millisecond),\n\t\terr == err,\n\t)\n\n\tSendResponse(ctx, errno.OK, svcs.Items)\n}", "func ParseListDNSZonesResponse(rsp *http.Response) (*ListDNSZonesResponse, error) {\n\tbodyBytes, err := ioutil.ReadAll(rsp.Body)\n\tdefer rsp.Body.Close()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresponse := &ListDNSZonesResponse{\n\t\tBody: bodyBytes,\n\t\tHTTPResponse: rsp,\n\t}\n\n\tswitch {\n\tcase strings.Contains(rsp.Header.Get(\"Content-Type\"), \"json\") && rsp.StatusCode == 200:\n\t\tvar dest ScalewayDomainV2alpha2ListDNSZonesResponse\n\t\tif err := json.Unmarshal(bodyBytes, &dest); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tresponse.JSON200 = &dest\n\n\t}\n\n\treturn response, nil\n}", "func getDNSZones(dnsZoneName string, dnsZoneID string, dnsZonesInterface dnsprovider.Zones) ([]dnsprovider.Zone, error) {\n\t// TODO: We need query-by-name and query-by-id functions\n\tdnsZones, err := dnsZonesInterface.List()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar matches []dnsprovider.Zone\n\tfindName := strings.TrimSuffix(dnsZoneName, \".\")\n\tfor _, dnsZone := range dnsZones {\n\t\tif dnsZoneID != \"\" {\n\t\t\tif dnsZoneID != dnsZone.ID() {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\t\tif findName != \"\" {\n\t\t\tif strings.TrimSuffix(dnsZone.Name(), \".\") != findName {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\t\tmatches = append(matches, dnsZone)\n\t}\n\n\treturn matches, nil\n}", "func fakeListZonalDisk(cfg *config.Config, name string, zone string, policies []string, callCount int) gcpRequest {\n\tscope := fmt.Sprintf(\"zones/%s\", zone)\n\tdisk := fakeZonalDisk(cfg, name, zone, policies)\n\treturn fakeDiskAggregatedListRequest(cfg, scope, disk, callCount)\n}", "func (c *MockDisksClient) List(ctx context.Context, resourceGroupName string) ([]compute.Disk, error) {\n\tvar l []compute.Disk\n\tfor _, disk := range c.Disks {\n\t\tl = append(l, disk)\n\t}\n\treturn l, nil\n}", "func (o *dnsOp) listRecords(zone dnsprovider.Zone) ([]dnsprovider.ResourceRecordSet, error) {\n\tkey := zone.Name() + \"::\" + zone.ID()\n\n\trrs := o.recordsCache[key]\n\tif rrs == nil {\n\t\trrsProvider, ok := zone.ResourceRecordSets()\n\t\tif !ok {\n\t\t\treturn nil, fmt.Errorf(\"zone does not support resource records %q\", zone.Name())\n\t\t}\n\n\t\tklog.V(2).Infof(\"Querying all dnsprovider records for zone %q\", zone.Name())\n\t\tvar err error\n\t\trrs, err = rrsProvider.List()\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"error querying resource records for zone %q: %v\", zone.Name(), err)\n\t\t}\n\n\t\to.recordsCache[key] = rrs\n\t}\n\n\treturn rrs, nil\n}", "func (dummy *Dummy) Zones() []dnsname.Name {\n\treturn dummy.zones\n}", "func (s *API) ListDNSZoneNameservers(req *ListDNSZoneNameserversRequest, opts ...scw.RequestOption) (*ListDNSZoneNameserversResponse, error) {\n\tvar err error\n\n\tquery := url.Values{}\n\tparameter.AddToQuery(query, \"project_id\", req.ProjectID)\n\n\tif fmt.Sprint(req.DNSZone) == \"\" {\n\t\treturn nil, errors.New(\"field DNSZone cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"GET\",\n\t\tPath: \"/domain/v2beta1/dns-zones/\" + fmt.Sprint(req.DNSZone) + \"/nameservers\",\n\t\tQuery: query,\n\t\tHeaders: http.Header{},\n\t}\n\n\tvar resp ListDNSZoneNameserversResponse\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func (c *cloud) Zones() (cloudprovider.Zones, bool) {\n\tklog.V(4).Infof(\"Zones called\")\n\treturn c, true\n}", "func runListServers(_ *cobra.Command, _ []string) {\n\tcfg, err := config.LoadFromFile()\n\tif err != nil {\n\t\texitWithError(err)\n\t}\n\n\tregions, err := checkRegions(*region)\n\tif err != nil {\n\t\texitWithError(err)\n\t}\n\n\tnameFilter := core.NewFilter(core.TagName, *name, core.Contains, *ignoreCase)\n\tenvFilter := core.NewFilter(core.TagEnv, *env, core.Equals, *ignoreCase)\n\tservers, err := core.GetAllServers(cfg.AWSCredentials, regions, nameFilter, envFilter)\n\tif err != nil {\n\t\texitWithError(err)\n\t}\n\n\tw := tabwriter.NewWriter(os.Stdout, 0, 0, 3, ' ', 0)\n\tfmt.Fprintln(w, \"NAME\\tENVIRONMENT\\tPRIVATE IP\\tPUBLIC IP\")\n\tfor _, server := range servers {\n\t\tfmt.Fprintf(w, \"%s\\t%s\\t%s\\t%s\\n\", server.Name, server.Env, server.PrivateIP, server.PublicIP)\n\t}\n\tw.Flush()\n}", "func (c *MockVirtualNetworksClient) List(ctx context.Context, resourceGroupName string) ([]network.VirtualNetwork, error) {\n\tvar l []network.VirtualNetwork\n\tfor _, vnet := range c.VNets {\n\t\tl = append(l, vnet)\n\t}\n\treturn l, nil\n}", "func (s *API) ListDNSZoneVersionRecords(req *ListDNSZoneVersionRecordsRequest, opts ...scw.RequestOption) (*ListDNSZoneVersionRecordsResponse, error) {\n\tvar err error\n\n\tquery := url.Values{}\n\tparameter.AddToQuery(query, \"organization_id\", req.OrganizationID)\n\n\tif fmt.Sprint(req.DNSZone) == \"\" {\n\t\treturn nil, errors.New(\"field DNSZone cannot be empty in request\")\n\t}\n\n\tif fmt.Sprint(req.Version) == \"\" {\n\t\treturn nil, errors.New(\"field Version cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"GET\",\n\t\tPath: \"/domain/v2alpha2/dns-zones/\" + fmt.Sprint(req.DNSZone) + \"/versions/\" + fmt.Sprint(req.Version) + \"/records\",\n\t\tQuery: query,\n\t\tHeaders: http.Header{},\n\t}\n\n\tvar resp ListDNSZoneVersionRecordsResponse\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func getNodesOfZone(nodeList *v1.NodeList, availabilityZone string) []string {\n\tvar nodeNames []string\n\tfor _, node := range nodeList.Items {\n\t\tnodeLabels := node.Labels\n\t\tif nodeLabels[zoneKey] == availabilityZone {\n\t\t\tnodeNames = append(nodeNames, node.Name)\n\t\t}\n\t}\n\treturn nodeNames\n}", "func (r ListAvailableZonesRequest) Send(ctx context.Context) (*ListAvailableZonesResponse, error) {\n\tr.Request.SetContext(ctx)\n\terr := r.Request.Send()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresp := &ListAvailableZonesResponse{\n\t\tListAvailableZonesOutput: r.Request.Data.(*ListAvailableZonesOutput),\n\t\tresponse: &aws.Response{Request: r.Request},\n\t}\n\n\treturn resp, nil\n}", "func (s *API) ListDNSZoneVersionRecords(req *ListDNSZoneVersionRecordsRequest, opts ...scw.RequestOption) (*ListDNSZoneVersionRecordsResponse, error) {\n\tvar err error\n\n\tdefaultPageSize, exist := s.client.GetDefaultPageSize()\n\tif (req.PageSize == nil || *req.PageSize == 0) && exist {\n\t\treq.PageSize = &defaultPageSize\n\t}\n\n\tquery := url.Values{}\n\tparameter.AddToQuery(query, \"page\", req.Page)\n\tparameter.AddToQuery(query, \"page_size\", req.PageSize)\n\n\tif fmt.Sprint(req.DNSZoneVersionID) == \"\" {\n\t\treturn nil, errors.New(\"field DNSZoneVersionID cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"GET\",\n\t\tPath: \"/domain/v2beta1/dns-zones/version/\" + fmt.Sprint(req.DNSZoneVersionID) + \"\",\n\t\tQuery: query,\n\t\tHeaders: http.Header{},\n\t}\n\n\tvar resp ListDNSZoneVersionRecordsResponse\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func (o GetLBFrontendIpConfigurationOutput) Zones() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v GetLBFrontendIpConfiguration) []string { return v.Zones }).(pulumi.StringArrayOutput)\n}", "func (opts ListOpts) ToZoneListQuery() (string, error) {\n\tq, err := gophercloud.BuildQueryString(opts)\n\treturn q.String(), err\n}", "func (o LoadBalancerFrontendIpConfigurationOutput) Zones() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v LoadBalancerFrontendIpConfiguration) []string { return v.Zones }).(pulumi.StringArrayOutput)\n}", "func (fkw *FakeClientWrapper) List(ctx context.Context, list runtime.Object, opts ...k8sCl.ListOption) error {\n\tif fkw.shouldPatchNS(list) {\n\t\topts = fkw.removeNSFromListOptions(opts)\n\t}\n\treturn fkw.client.List(ctx, list, opts...)\n}", "func (c *cloud) Zones() (cloudprovider.Zones, bool) {\n\treturn &zones{\n\t\tnamespace: c.namespace,\n\t\tkubevirt: c.kubevirt,\n\t}, true\n}", "func (n VpcNatGateways) List(region account.Region, account account.Account, force bool) ([]cloud.Resource, error) {\n\tclient, err := vpc.NewClientWithAccessKey(string(region), account.AccessKeyID, account.AccessKeySecret)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\trequest := vpc.CreateDescribeNatGatewaysRequest()\n\trequest.PageSize = \"50\"\n\tresponse, err := client.DescribeNatGateways(request)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tnatGateways := make([]cloud.Resource, 0)\n\tfor _, natGatewayItem := range response.NatGateways.NatGateway {\n\t\tsnatTables, err := fetchSnatTables(client, natGatewayItem.SnatTableIds)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tnatGateways = append(natGateways, VpcNatGateway{NatGateway: natGatewayItem, SnatTables: snatTables})\n\t}\n\n\treturn natGateways, nil\n}", "func (api *distributedservicecardAPI) ApisrvList(ctx context.Context, opts *api.ListWatchOptions) ([]*cluster.DistributedServiceCard, error) {\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn nil, err\n\t\t}\n\n\t\treturn apicl.ClusterV1().DistributedServiceCard().List(context.Background(), opts)\n\t}\n\n\t// List from local cache\n\tctkitObjs, err := api.List(ctx, opts)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar ret []*cluster.DistributedServiceCard\n\tfor _, obj := range ctkitObjs {\n\t\tret = append(ret, &obj.DistributedServiceCard)\n\t}\n\treturn ret, nil\n}", "func List(client *gophercloud.ServiceClient, opts ListOptsBuilder) pagination.Pager {\n\turl := listURL(client)\n\tif opts != nil {\n\t\tquery, err := opts.ToRegionListQuery()\n\t\tif err != nil {\n\t\t\treturn pagination.Pager{Err: err}\n\t\t}\n\t\turl += query\n\t}\n\treturn pagination.NewPager(client, url, func(r pagination.PageResult) pagination.Page {\n\t\treturn RegionPage{pagination.LinkedPageBase{PageResult: r}}\n\t})\n}", "func (adm Admin) ListClusters() (string, error) {\n\tconn := newConnection(adm.ZkSvr)\n\terr := conn.Connect()\n\tif err != nil {\n\t\tfmt.Println(\"Failed to connect to zookeeper.\")\n\t\treturn \"\", err\n\t}\n\tdefer conn.Disconnect()\n\n\tvar clusters []string\n\n\tchildren, err := conn.Children(\"/\")\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tfor _, cluster := range children {\n\t\tif ok, err := conn.IsClusterSetup(cluster); ok && err == nil {\n\t\t\tclusters = append(clusters, cluster)\n\t\t}\n\t}\n\n\tvar buffer bytes.Buffer\n\tbuffer.WriteString(\"Existing clusters: \\n\")\n\n\tfor _, cluster := range clusters {\n\t\tbuffer.WriteString(\" \" + cluster + \"\\n\")\n\t}\n\treturn buffer.String(), nil\n}", "func (api *clusterAPI) ApisrvList(ctx context.Context, opts *api.ListWatchOptions) ([]*cluster.Cluster, error) {\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn nil, err\n\t\t}\n\n\t\treturn apicl.ClusterV1().Cluster().List(context.Background(), opts)\n\t}\n\n\t// List from local cache\n\tctkitObjs, err := api.List(ctx, opts)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar ret []*cluster.Cluster\n\tfor _, obj := range ctkitObjs {\n\t\tret = append(ret, &obj.Cluster)\n\t}\n\treturn ret, nil\n}", "func (ncr *NamespacedConfigMapReflector) List() ([]interface{}, error) {\n\treturn virtualkubelet.List[virtualkubelet.Lister[*corev1.ConfigMap], *corev1.ConfigMap](\n\t\tncr.localConfigMaps,\n\t\tncr.remoteConfigMaps,\n\t)\n}", "func (api *tenantAPI) ApisrvList(ctx context.Context, opts *api.ListWatchOptions) ([]*cluster.Tenant, error) {\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn nil, err\n\t\t}\n\n\t\treturn apicl.ClusterV1().Tenant().List(context.Background(), opts)\n\t}\n\n\t// List from local cache\n\tctkitObjs, err := api.List(ctx, opts)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar ret []*cluster.Tenant\n\tfor _, obj := range ctkitObjs {\n\t\tret = append(ret, &obj.Tenant)\n\t}\n\treturn ret, nil\n}", "func (r *Replicator) ListDestinations(ctx thrift.Context, listRequest *shared.ListDestinationsRequest) (*shared.ListDestinationsResult_, error) {\n\treturn r.metaClient.ListDestinations(ctx, listRequest)\n}", "func NewZoneSection_List(s *capnp.Segment, sz int32) (ZoneSection_List, error) {\n\tl, err := capnp.NewCompositeList(s, capnp.ObjectSize{DataSize: 0, PointerCount: 4}, sz)\n\treturn ZoneSection_List{l}, err\n}", "func (cloud *Cloud) Zones() (cloudprovider.Zones, bool) {\n\treturn nil, false\n}", "func listClusters(w http.ResponseWriter, r *http.Request, t auth.Token) (err error) {\n\tctx := r.Context()\n\tallowed := permission.Check(t, permission.PermClusterRead)\n\tif !allowed {\n\t\treturn permission.ErrUnauthorized\n\t}\n\tclusters, err := servicemanager.Cluster.List(ctx)\n\tif err != nil {\n\t\tif err == provTypes.ErrNoCluster {\n\t\t\tw.WriteHeader(http.StatusNoContent)\n\t\t\treturn nil\n\t\t}\n\t\treturn err\n\t}\n\tadmin := permission.Check(t, permission.PermClusterAdmin)\n\tif !admin {\n\t\tfor i := range clusters {\n\t\t\tclusters[i].CleanUpSensitive()\n\t\t}\n\t}\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\treturn json.NewEncoder(w).Encode(clusters)\n}", "func (c *mySQLFlexServersClient) ListWithinGroup(ctx context.Context, group string) ([]*armmysqlflexibleservers.Server, error) {\n\tvar servers []*armmysqlflexibleservers.Server\n\topts := &armmysqlflexibleservers.ServersClientListByResourceGroupOptions{}\n\tpager := c.api.NewListByResourceGroupPager(group, opts)\n\tfor pageNum := 0; pager.More(); pageNum++ {\n\t\tpage, err := pager.NextPage(ctx)\n\t\tif err != nil {\n\t\t\treturn nil, trace.Wrap(ConvertResponseError(err))\n\t\t}\n\t\tservers = append(servers, page.Value...)\n\t}\n\treturn servers, nil\n}", "func (h *httpCloud) Zones() (cloudprovider.Zones, bool) {\n\treturn nil, false\n}" ]
[ "0.795497", "0.7876782", "0.78532934", "0.78210187", "0.7654382", "0.7630902", "0.7611185", "0.7603991", "0.7511655", "0.7472401", "0.7393847", "0.72342956", "0.72329515", "0.7151259", "0.7083987", "0.7083133", "0.7003336", "0.6993418", "0.67353326", "0.6663804", "0.6421205", "0.636167", "0.6335746", "0.6329713", "0.6323973", "0.6296566", "0.6279923", "0.6268252", "0.6212942", "0.6198482", "0.61876744", "0.6178704", "0.6158047", "0.6157883", "0.6157213", "0.61529535", "0.61424243", "0.61163455", "0.6112189", "0.611014", "0.61056614", "0.60764855", "0.60751534", "0.60388607", "0.6023462", "0.59680605", "0.59547895", "0.5943812", "0.59380627", "0.5920849", "0.59130627", "0.58980346", "0.58923084", "0.5843552", "0.58287066", "0.57994944", "0.5798171", "0.5783472", "0.57556736", "0.5700662", "0.5680301", "0.5672855", "0.5651601", "0.5640589", "0.56357497", "0.56039536", "0.5602233", "0.55856615", "0.5584866", "0.5580554", "0.5549925", "0.55482936", "0.5540871", "0.55261236", "0.5501662", "0.54988116", "0.5489917", "0.548403", "0.5473397", "0.54702115", "0.54430777", "0.543843", "0.5430064", "0.542363", "0.5415339", "0.5406803", "0.5397625", "0.5390685", "0.5367446", "0.5360727", "0.5358158", "0.533425", "0.53321487", "0.53314954", "0.53219515", "0.5309016", "0.5308514", "0.529387", "0.5290719", "0.5288015" ]
0.79486257
1
CreateSnapshot uses the override method CreateSnapshotFn or the real implementation.
func (c *TestClient) CreateSnapshot(project, zone, disk string, s *compute.Snapshot) error { if c.CreateSnapshotFn != nil { return c.CreateSnapshotFn(project, zone, disk, s) } return c.client.CreateSnapshot(project, zone, disk, s) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (f *Function) CreateInstanceSnapshot() {\n\tlogger := log.WithFields(log.Fields{\"fID\": f.fID})\n\n\tlogger.Debug(\"Creating instance snapshot\")\n\n\tctx, cancel := context.WithTimeout(context.Background(), time.Second*5)\n\tdefer cancel()\n\n\terr := orch.PauseVM(ctx, f.vmID)\n\tif err != nil {\n\t\tlog.Panic(err)\n\t}\n\n\terr = orch.CreateSnapshot(ctx, f.vmID)\n\tif err != nil {\n\t\tlog.Panic(err)\n\t}\n\n\t_, err = orch.ResumeVM(ctx, f.vmID)\n\tif err != nil {\n\t\tlog.Panic(err)\n\t}\n}", "func (ks *KopiaSnapshotter) CreateSnapshot(sourceDir string, opts map[string]string) (snapID string, err error) {\n\treturn ks.snap.CreateSnapshot(sourceDir)\n}", "func (d *DefaultDriver) ValidateCreateSnapshot(name string, params map[string]string) error {\n\treturn &errors.ErrNotSupported{\n\t\tType: \"Function\",\n\t\tOperation: \"ValidateCreateSnapshot()\",\n\t}\n}", "func (d *Driver) CreateSnapshot(ctx context.Context, req *csi.CreateSnapshotRequest) (*csi.CreateSnapshotResponse, error) {\n\treturn nil, status.Error(codes.Unimplemented, \"\")\n}", "func (driver *Driver) CreateSnapshot(ctx context.Context, req *csi.CreateSnapshotRequest) (*csi.CreateSnapshotResponse, error) {\n\treturn nil, status.Error(codes.Unimplemented, \"\")\n}", "func (c *restClient) CreateSnapshot(ctx context.Context, req *netapppb.CreateSnapshotRequest, opts ...gax.CallOption) (*CreateSnapshotOperation, error) {\n\tm := protojson.MarshalOptions{AllowPartial: true, UseEnumNumbers: true}\n\tbody := req.GetSnapshot()\n\tjsonReq, err := m.Marshal(body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v1/%v/snapshots\", req.GetParent())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\tparams.Add(\"snapshotId\", fmt.Sprintf(\"%v\", req.GetSnapshotId()))\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"parent\", url.QueryEscape(req.GetParent()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tresp := &longrunningpb.Operation{}\n\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"POST\", baseUrl.String(), bytes.NewReader(jsonReq))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}, opts...)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\n\toverride := fmt.Sprintf(\"/v1/%s\", resp.GetName())\n\treturn &CreateSnapshotOperation{\n\t\tlro: longrunning.InternalNewOperation(*c.LROClient, resp),\n\t\tpollPath: override,\n\t}, nil\n}", "func (cs *controller) CreateSnapshot(\n\tctx context.Context,\n\treq *csi.CreateSnapshotRequest,\n) (*csi.CreateSnapshotResponse, error) {\n\n\treturn nil, status.Error(codes.Unimplemented, \"\")\n}", "func (s *API) CreateSnapshot(req *CreateSnapshotRequest, opts ...scw.RequestOption) (*Snapshot, error) {\n\tvar err error\n\n\tif req.Region == \"\" {\n\t\tdefaultRegion, _ := s.client.GetDefaultRegion()\n\t\treq.Region = defaultRegion\n\t}\n\n\tif req.Name == \"\" {\n\t\treq.Name = namegenerator.GetRandomName(\"snp\")\n\t}\n\n\tif fmt.Sprint(req.Region) == \"\" {\n\t\treturn nil, errors.New(\"field Region cannot be empty in request\")\n\t}\n\n\tif fmt.Sprint(req.InstanceID) == \"\" {\n\t\treturn nil, errors.New(\"field InstanceID cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"POST\",\n\t\tPath: \"/rdb/v1/regions/\" + fmt.Sprint(req.Region) + \"/instances/\" + fmt.Sprint(req.InstanceID) + \"/snapshots\",\n\t\tHeaders: http.Header{},\n\t}\n\n\terr = scwReq.SetBody(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar resp Snapshot\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func NewSnapshot(ctx *pulumi.Context,\n\tname string, args *SnapshotArgs, opts ...pulumi.ResourceOption) (*Snapshot, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.VolumeId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'VolumeId'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Snapshot\n\terr := ctx.RegisterResource(\"aws-native:fsx:Snapshot\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (p *Plugin) CreateSnapshot(volumeID, volumeAZ string, tags map[string]string) (string, error) {\n\treturn p.plugin.CreateSnapshot(volumeID, volumeAZ, tags)\n}", "func (snapshotter *Snapshotter) CreateSnapshot(endpoint *portainer.Endpoint) (*portainer.DockerSnapshot, error) {\n\tcli, err := snapshotter.clientFactory.CreateClient(endpoint, \"\", nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer cli.Close()\n\n\treturn snapshot(cli, endpoint)\n}", "func NewSnapshot(ctx *pulumi.Context,\n\tname string, args *SnapshotArgs, opts ...pulumi.ResourceOption) (*Snapshot, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.InstanceId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'InstanceId'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Snapshot\n\terr := ctx.RegisterResource(\"alicloud:databasefilesystem/snapshot:Snapshot\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func NewSnapshot(ctx *pulumi.Context,\n\tname string, args *SnapshotArgs, opts ...pulumi.ResourceOption) (*Snapshot, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.NamespaceName == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'NamespaceName'\")\n\t}\n\tif args.SnapshotName == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'SnapshotName'\")\n\t}\n\tvar resource Snapshot\n\terr := ctx.RegisterResource(\"aws:redshiftserverless/snapshot:Snapshot\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (c *RabbitEventStoreClient) CreateSnapshot(opts *framework.CreateSnapshotOption) (success bool, err error) {\n\terr = c.client.Send(Request{\n\t\tAction: \"CreateSnapshot\",\n\t\tData: *opts,\n\t}, &success)\n\n\treturn success, err\n}", "func createSnapshot(sg *snapshotgroup.SnapshotGroup, annotations map[string]string) error {\n\ttimestamp := strconv.Itoa(int(time.Now().Unix()))\n\tannotations[TimestampAnnotation] = timestamp\n\tannotations[managedByAnnotation] = managerName\n\tannotations[GroupNameAnnotation] = sg.ObjectMeta.Name\n\n\tsnapshot := snapshotsv1.VolumeSnapshot{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tNamespace: sg.ObjectMeta.Namespace,\n\t\t\tName: sg.ObjectMeta.Name + \"-\" + timestamp,\n\t\t\tAnnotations: annotations,\n\t\t},\n\t\tSpec: sg.Spec.Template.Spec,\n\t}\n\tname := getPVCName(sg)\n\tklog.Infof(\"%s/%s: creating snapshot for PVC %s\", sg.ObjectMeta.Namespace, sg.ObjectMeta.Name, name)\n\tsnapshot.Spec.Source.PersistentVolumeClaimName = &name\n\n\tmarshaled, err := json.Marshal(snapshot)\n\tif err != nil {\n\t\treturn err\n\t}\n\tunst := unstructured.Unstructured{\n\t\tObject: map[string]interface{}{},\n\t}\n\terr = json.Unmarshal(marshaled, &unst.Object)\n\tif err != nil {\n\t\treturn err\n\t}\n\tclient := kube.GetClient()\n\tunst.Object[\"kind\"] = \"VolumeSnapshot\"\n\tunst.Object[\"apiVersion\"] = client.VolumeSnapshotVersion\n\n\tif strings.HasSuffix(client.VolumeSnapshotVersion, \"v1alpha1\") {\n\t\t// There is a slight change in `source` from alpha to beta\n\t\tspec := unst.Object[\"spec\"].(map[string]interface{})\n\t\tsource := spec[\"source\"].(map[string]interface{})\n\t\tdelete(source, \"persistentVolumeClaimName\")\n\t\tsource[\"name\"] = name\n\t\tsource[\"kind\"] = \"PersistentVolumeClaim\"\n\t\tspec[\"source\"] = source\n\t\tunst.Object[\"spec\"] = spec\n\t}\n\n\tsnapClient := client.SnapshotClient.Namespace(snapshot.ObjectMeta.Namespace)\n\t_, err = snapClient.Create(&unst, metav1.CreateOptions{})\n\treturn err\n}", "func (s *DataStore) CreateSnapshot(snapshot *longhorn.Snapshot) (*longhorn.Snapshot, error) {\n\tret, err := s.lhClient.LonghornV1beta2().Snapshots(s.namespace).Create(context.TODO(), snapshot, metav1.CreateOptions{})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif SkipListerCheck {\n\t\treturn ret, nil\n\t}\n\n\tobj, err := verifyCreation(snapshot.Name, \"snapshot\", func(name string) (runtime.Object, error) {\n\t\treturn s.GetSnapshotRO(name)\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tret, ok := obj.(*longhorn.Snapshot)\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"BUG: datastore: verifyCreation returned wrong type for Snapshot\")\n\t}\n\treturn ret.DeepCopy(), nil\n}", "func (c *Client) SnapshotCreate(vm *VirtualMachine, name string) (*types.ManagedObjectReference, error) {\n\tif c.Verbose {\n\t\tfmt.Printf(\"Creating a VM snapshot...\\n\")\n\t}\n\n\tvar res types.ManagedObjectReference\n\n\terr := func() error {\n\t\tctx, cancelFn := context.WithTimeout(context.Background(), c.timeout)\n\t\tdefer cancelFn()\n\n\t\ttask, err := vm.VM.CreateSnapshot(ctx, name, \"\", false, false)\n\t\tany, err := c.finishTask(ctx, task, err)\n\t\tif err != nil {\n\t\t\treturn errors.Wrapf(err, \"While snapshotting VM\")\n\t\t}\n\n\t\tres = any.(types.ManagedObjectReference)\n\n\t\treturn nil\n\t}()\n\n\tif err != nil {\n\t\tswitch err := errors.Cause(err).(type) {\n\t\tcase *TimeoutExceededError:\n\t\t\t// handle specifically\n\t\t\treturn nil, fmt.Errorf(\"Timeout while attempting to snapshot VM\")\n\t\tdefault:\n\t\t\t// unknown error\n\t\t\treturn nil, errors.Wrap(err, \"Got error while snapshotting a VM\")\n\t\t}\n\t}\n\n\treturn &res, nil\n}", "func (cs *DefaultControllerServer) CreateSnapshot(ctx context.Context, req *csi.CreateSnapshotRequest) (*csi.CreateSnapshotResponse, error) {\n\treturn nil, status.Error(codes.Unimplemented, \"\")\n}", "func (a *Client) CreateSnapshot(params *CreateSnapshotParams, opts ...ClientOption) (*CreateSnapshotOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewCreateSnapshotParams()\n\t}\n\top := &runtime.ClientOperation{\n\t\tID: \"createSnapshot\",\n\t\tMethod: \"POST\",\n\t\tPathPattern: \"/astrolabe/{service}/{protectedEntityID}/snapshots\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &CreateSnapshotReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t}\n\tfor _, opt := range opts {\n\t\topt(op)\n\t}\n\n\tresult, err := a.transport.Submit(op)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*CreateSnapshotOK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\t// safeguard: normally, absent a default response, unknown success responses return an error above: so this is a codegen issue\n\tmsg := fmt.Sprintf(\"unexpected success response for createSnapshot: API contract not enforced by server. Client expected to get an error, but got: %T\", result)\n\tpanic(msg)\n}", "func NewSnapshot(ctx *pulumi.Context,\n\tname string, args *SnapshotArgs, opts ...pulumi.ResourceOption) (*Snapshot, error) {\n\tif args == nil || args.SourceDisk == nil {\n\t\treturn nil, errors.New(\"missing required argument 'SourceDisk'\")\n\t}\n\tif args == nil {\n\t\targs = &SnapshotArgs{}\n\t}\n\tvar resource Snapshot\n\terr := ctx.RegisterResource(\"gcp:compute/snapshot:Snapshot\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (s *OsdCsiServer) CreateSnapshot(\n\tctx context.Context,\n\treq *csi.CreateSnapshotRequest,\n) (*csi.CreateSnapshotResponse, error) {\n\n\tif len(req.GetSourceVolumeId()) == 0 {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"Volume id must be provided\")\n\t} else if len(req.GetName()) == 0 {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"Name must be provided\")\n\t}\n\n\t// Check if the snapshot with this name already exists\n\tv, err := util.VolumeFromName(s.driver, req.GetName())\n\tif err == nil {\n\t\t// Verify the parent is the same\n\t\tif req.GetSourceVolumeId() != v.GetSource().GetParent() {\n\t\t\treturn nil, status.Error(codes.AlreadyExists, \"Requested snapshot already exists for another source volume id\")\n\t\t}\n\n\t\t// Return current snapshot info\n\t\tcreatedAt, err := ptypes.Timestamp(v.GetCtime())\n\t\tif err != nil {\n\t\t\treturn nil, status.Errorf(codes.Internal, \"Failed to get time snapshot was created: %v\", err)\n\t\t}\n\t\treturn &csi.CreateSnapshotResponse{\n\t\t\tSnapshot: &csi.Snapshot{\n\t\t\t\tId: v.GetId(),\n\t\t\t\tSourceVolumeId: v.GetSource().GetParent(),\n\t\t\t\tCreatedAt: createdAt.Unix(),\n\t\t\t\tStatus: &csi.SnapshotStatus{\n\t\t\t\t\t// This means that we are not uploading our snapshot\n\t\t\t\t\t// We may add support for cloud snaps in future patches\n\t\t\t\t\tType: csi.SnapshotStatus_READY,\n\t\t\t\t},\n\t\t\t},\n\t\t}, nil\n\t}\n\n\t// Get any labels passed in by the CO\n\t_, locator, _, err := s.specHandler.SpecFromOpts(req.GetParameters())\n\tif err != nil {\n\t\treturn nil, status.Errorf(codes.InvalidArgument, \"Unable to get parameters: %v\", err)\n\t}\n\n\t// Create snapshot\n\treadonly := true\n\tsnapshotID, err := s.driver.Snapshot(req.GetSourceVolumeId(), readonly, &api.VolumeLocator{\n\t\tName: req.GetName(),\n\t\tVolumeLabels: locator.GetVolumeLabels(),\n\t}, false)\n\tif err != nil {\n\t\tif err == kvdb.ErrNotFound {\n\t\t\treturn nil, status.Errorf(codes.NotFound, \"Volume id %s not found\", req.GetSourceVolumeId())\n\t\t}\n\t\treturn nil, status.Errorf(codes.Internal, \"Failed to create snapshot: %v\", err)\n\t}\n\n\tsnapInfo, err := util.VolumeFromName(s.driver, snapshotID)\n\tif err != nil {\n\t\treturn nil, status.Errorf(codes.Internal, \"Failed to get information about the snapshot: %v\", err)\n\t}\n\tcreatedAt, err := ptypes.Timestamp(snapInfo.GetCtime())\n\tif err != nil {\n\t\treturn nil, status.Errorf(codes.Internal, \"Failed to get time snapshot was created: %v\", err)\n\t}\n\n\treturn &csi.CreateSnapshotResponse{\n\t\tSnapshot: &csi.Snapshot{\n\t\t\tId: snapshotID,\n\t\t\tSourceVolumeId: req.GetSourceVolumeId(),\n\t\t\tCreatedAt: createdAt.Unix(),\n\t\t\tStatus: &csi.SnapshotStatus{\n\t\t\t\t// This means that we are not uploading our snapshot\n\t\t\t\t// We may add support flow cloud snaps in future patches\n\t\t\t\tType: csi.SnapshotStatus_READY,\n\t\t\t},\n\t\t},\n\t}, nil\n}", "func (a Agent) SnapshotCreate(sid int, password string) (*RawSnapshot, error) {\n\treq := libts.Request{\n\t\tCommand: \"serversnapshotcreate\",\n\t\tServerID: sid,\n\t\tArgs: map[string]interface{}{\n\t\t\t\"password\": password,\n\t\t},\n\t}\n\tsnapshot := &RawSnapshot{}\n\terr := a.Query.Do(req, snapshot)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn snapshot, nil\n}", "func (cs *controllerServer) CreateSnapshot(ctx context.Context, req *csi.CreateSnapshotRequest) (*csi.CreateSnapshotResponse, error) {\n\tlog.Debugf(\"Starting Create Snapshot %s with response: %v\", req.Name, req)\n\t// Step 1: check request\n\tsnapshotName := req.GetName()\n\tif len(snapshotName) == 0 {\n\t\tlog.Error(\"CreateSnapshot: snapshot name not provided\")\n\t\treturn nil, status.Error(codes.InvalidArgument, \"CreateSnapshot: snapshot name not provided\")\n\t}\n\tvolumeID := req.GetSourceVolumeId()\n\tif len(volumeID) == 0 {\n\t\tlog.Error(\"CreateSnapshot: snapshot volume source ID not provided\")\n\t\treturn nil, status.Error(codes.InvalidArgument, \"CreateSnapshot: snapshot volume source ID not provided\")\n\t}\n\n\t// Step 2: get snapshot initial size from parameter\n\tinitialSize, _, _, err := getSnapshotInitialInfo(req.Parameters)\n\tif err != nil {\n\t\tlog.Errorf(\"CreateSnapshot: get snapshot %s initial info error: %s\", req.Name, err.Error())\n\t\treturn nil, status.Errorf(codes.Internal, \"CreateSnapshot: get snapshot %s initial info error: %s\", req.Name, err.Error())\n\t}\n\n\t// Step 3: get nodeName and vgName\n\tnodeName, vgName, pv, err := getPvSpec(cs.client, volumeID, cs.driverName)\n\tif err != nil {\n\t\tlog.Errorf(\"CreateSnapshot: get pv %s error: %s\", volumeID, err.Error())\n\t\treturn nil, status.Errorf(codes.Internal, \"CreateSnapshot: get pv %s error: %s\", volumeID, err.Error())\n\t}\n\tlog.Infof(\"CreateSnapshot: snapshot %s is in %s, whose vg is %s\", snapshotName, nodeName, vgName)\n\n\t// Step 4: update initialSize if initialSize is bigger than pv request size\n\tpvSize, _ := pv.Spec.Capacity.Storage().AsInt64()\n\tif pvSize < int64(initialSize) {\n\t\tinitialSize = uint64(pvSize)\n\t}\n\n\t// Step 5: get grpc client\n\tconn, err := cs.getNodeConn(nodeName)\n\tif err != nil {\n\t\tlog.Errorf(\"CreateSnapshot: get grpc client at node %s error: %s\", nodeName, err.Error())\n\t\treturn nil, status.Errorf(codes.Internal, \"CreateSnapshot: get grpc client at node %s error: %s\", nodeName, err.Error())\n\t}\n\tdefer conn.Close()\n\n\t// Step 6: create lvm snapshot\n\tvar lvmName string\n\tif lvmName, err = conn.GetLvm(ctx, vgName, snapshotName); err != nil {\n\t\tlog.Errorf(\"CreateSnapshot: get lvm snapshot %s failed: %s\", snapshotName, err.Error())\n\t\treturn nil, status.Errorf(codes.Internal, \"CreateSnapshot: get lvm snapshot %s failed: %s\", snapshotName, err.Error())\n\t}\n\tif lvmName == \"\" {\n\t\t_, err := conn.CreateSnapshot(ctx, vgName, snapshotName, volumeID, initialSize)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"CreateSnapshot: create lvm snapshot %s failed: %s\", snapshotName, err.Error())\n\t\t\treturn nil, status.Errorf(codes.Internal, \"CreateSnapshot: create lvm snapshot %s failed: %s\", snapshotName, err.Error())\n\t\t}\n\t\tlog.Infof(\"CreateSnapshot: create snapshot %s successfully\", snapshotName)\n\t} else {\n\t\tlog.Infof(\"CreateSnapshot: lvm snapshot %s in node %s already exists\", snapshotName, nodeName)\n\t}\n\treturn cs.newCreateSnapshotResponse(req)\n}", "func (ctrler CtrlDefReactor) OnSnapshotRestoreCreate(obj *SnapshotRestore) error {\n\tlog.Info(\"OnSnapshotRestoreCreate is not implemented\")\n\treturn nil\n}", "func (s *OsdCsiServer) CreateSnapshot(\n\tctx context.Context,\n\treq *csi.CreateSnapshotRequest,\n) (*csi.CreateSnapshotResponse, error) {\n\n\tif len(req.GetSourceVolumeId()) == 0 {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"Volume id must be provided\")\n\t} else if len(req.GetName()) == 0 {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"Name must be provided\")\n\t}\n\n\t// Get secret if any was passed\n\tctx = s.setupContext(ctx, req.GetSecrets())\n\tctx, cancel := grpcutil.WithDefaultTimeout(ctx)\n\tdefer cancel()\n\n\t// Get any labels passed in by the CO\n\t_, locator, _, err := s.specHandler.SpecFromOpts(req.GetParameters())\n\tif err != nil {\n\t\treturn nil, status.Errorf(codes.InvalidArgument, \"Unable to get parameters: %v\", err)\n\t}\n\t// Check ID is valid with the specified volume capabilities\n\tsnapshotType, ok := locator.VolumeLabels[osdSnapshotLabelsTypeKey]\n\tif !ok {\n\t\tsnapshotType = DriverTypeLocal\n\t}\n\tswitch snapshotType {\n\tcase DriverTypeCloud:\n\t\treturn s.createCloudBackup(ctx, req)\n\tcase DriverTypeLocal:\n\t\tfallthrough\n\tdefault:\n\t\treturn s.createLocalSnapshot(ctx, req)\n\t}\n}", "func CreateSnapshot(storageID string) error {\n\tclient, err := NewExtPacketClient()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\trequest := &extpackngo.CreateSnapShotRequest{}\n\n\t_, e := client.Storages.CreateSnapshot(storageID, request)\n\treturn e\n}", "func (cli *Client) SnapshotCreate(options types.SnapshotCreateRequest) (types.Snapshot, error) {\n\tvar snapshot types.Snapshot\n\tv := url.Values{}\n\tv.Set(\"volume\", options.Volume)\n\tv.Set(\"name\", options.Name)\n\tif options.Force {\n\t\tv.Set(\"force\", \"true\")\n\t}\n\tresp, err := cli.post(\"/snapshots/create?\"+v.Encode(), nil, options, nil)\n\tif err != nil {\n\t\treturn snapshot, err\n\t}\n\terr = json.NewDecoder(resp.body).Decode(&snapshot)\n\tensureReaderClosed(resp)\n\treturn snapshot, err\n}", "func NewSnapshot(ctx *pulumi.Context,\n\tname string, args *SnapshotArgs, opts ...pulumi.ResourceOption) (*Snapshot, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.DiskId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'DiskId'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Snapshot\n\terr := ctx.RegisterResource(\"alicloud:ecs/snapshot:Snapshot\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func cephRBDSnapshotCreate(clusterName string, poolName string,\n\tvolumeName string, volumeType string, snapshotName string,\n\tuserName string) error {\n\t_, err := shared.RunCommand(\n\t\t\"rbd\",\n\t\t\"--id\", userName,\n\t\t\"--cluster\", clusterName,\n\t\t\"--pool\", poolName,\n\t\t\"snap\",\n\t\t\"create\",\n\t\t\"--snap\", snapshotName,\n\t\tfmt.Sprintf(\"%s_%s\", volumeType, volumeName))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (os *OpenStack) CreateSnapshot(sourceVolumeID, name, description string, tags map[string]string) (string, string, error) {\n\tsnapshots, err := os.snapshotService()\n\tif err != nil || snapshots == nil {\n\t\tglog.Errorf(\"Unable to initialize cinder client for region: %s\", os.region)\n\t\treturn \"\", \"\", fmt.Errorf(\"Failed to create snapshot for volume %s: %v\", sourceVolumeID, err)\n\t}\n\n\topts := SnapshotCreateOpts{\n\t\tVolumeID: sourceVolumeID,\n\t\tName: name,\n\t\tDescription: description,\n\t}\n\tif tags != nil {\n\t\topts.Metadata = tags\n\t}\n\n\tsnapshotID, status, err := snapshots.createSnapshot(opts)\n\n\tif err != nil {\n\t\tglog.Errorf(\"Failed to snapshot volume %s : %v\", sourceVolumeID, err)\n\t\treturn \"\", \"\", err\n\t}\n\n\tglog.Infof(\"Created snapshot %v from volume: %v\", snapshotID, sourceVolumeID)\n\treturn snapshotID, status, nil\n}", "func (img *Image) CreateSnapshot(name string) (*Snapshot, error) {\n\targs := img.cmdArgs(\"snap\", \"create\", \"--snap\", name)\n\terr := cmdRun(createErrs, args...)\n\tif err != nil && !errors.Is(err, ErrAlreadyExists) {\n\t\treturn nil, err\n\t}\n\treturn img.getSnapshot(name), err\n}", "func createSnapshot(nodeID int, term int, index int, suffix string) (file string, writer io.WriteCloser, err error) {\n\tif term < 0 || index < 0 || suffix == \"\" {\n\t\treturn \"\", nil, errorInvalidSnapshotInfo\n\t}\n\n\tfileName := fmt.Sprintf(\"Node%d_T%dL%d_%s.rkvsnapshot\", nodeID, term, index, suffix)\n\tfullpath := filepath.Join(snapshotPath, fileName)\n\tf, err := os.Create(fullpath)\n\treturn fullpath, f, err\n}", "func createFallbackSnapshot() cache.Snapshot {\n\treturn fallbackSnapshot(fallbackBindAddr, fallbackBindPort, fallbackStatusCode)\n}", "func Create(createSnapshotURL string) (string, error) {\n\tlogger.Infof(\"%s\", \"Creating snapshot\")\n\tu, err := url.Parse(createSnapshotURL)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tresp, err := http.Get(u.String())\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tbody, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tsnap := snapshot{}\n\terr = json.Unmarshal(body, &snap)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif snap.Status == \"ok\" {\n\t\tlogger.Infof(\"Snapshot %s created\", snap.Snapshot)\n\t\treturn snap.Snapshot, nil\n\t} else if snap.Status == \"error\" {\n\t\treturn \"\", errors.New(snap.Msg)\n\t} else {\n\t\treturn \"\", fmt.Errorf(\"Unkown status: %v\", snap.Status)\n\t}\n}", "func (c *Client) CreateSnapshot(ctx context.Context, id, ref string) error {\n\twrapperCli, err := c.Get(ctx)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to get a containerd grpc client: %v\", err)\n\t}\n\n\toriginalCtx := ctx\n\tctx = leases.WithLease(ctx, wrapperCli.lease.ID)\n\n\tvar (\n\t\tsnName = CurrentSnapshotterName(ctx)\n\t\tsnSrv = wrapperCli.client.SnapshotService(snName)\n\t)\n\n\timage, err := wrapperCli.client.GetImage(ctx, ref)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdiffIDs, err := image.RootFS(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\tparent := identity.ChainID(diffIDs).String()\n\n\t// NOTE: PouchContainer always unpacks image during pulling. But there\n\t// maybe crash or terminated by some reason. The image have been stored\n\t// in containerd without unpacking. And the following creating container\n\t// request will fail on preparing snapshot because there is no such\n\t// parent snapshotter. Based on this case, we should skip the not\n\t// found error and try to unpack it again.\n\t_, err = snSrv.Prepare(ctx, id, parent)\n\tif err == nil || !errdefs.IsNotFound(err) {\n\t\treturn err\n\t}\n\tlog.With(ctx).Warnf(\"checking unpack status for image %s on %s snapshotter...\", image.Name(), snName)\n\n\t// check unpacked status\n\tunpacked, werr := image.IsUnpacked(ctx, snName)\n\tif werr != nil {\n\t\tlog.With(ctx).Warnf(\"failed to check unpack status for image %s on %s snapshotter: %v\", image.Name(), snName, werr)\n\t\treturn err\n\t}\n\n\t// if it is not unpacked, try to unpack it.\n\tif !unpacked {\n\t\tlog.With(ctx).Warnf(\"the image %s doesn't unpack for %s snapshotter, try to unpack it...\", image.Name(), snName)\n\t\t// NOTE: don't use pouchd lease id here because pouchd lease id\n\t\t// will hold the snapshotter forever, which means that the\n\t\t// snapshotter will not removed if we remove image.\n\t\tif werr = image.Unpack(originalCtx, snName); werr != nil {\n\t\t\tlog.With(ctx).Warnf(\"failed to unpack for image %s on %s snapshotter: %v\", image.Name(), snName, werr)\n\t\t\treturn err\n\t\t}\n\n\t\t// do it again.\n\t\t_, err = snSrv.Prepare(ctx, id, parent)\n\t}\n\treturn err\n}", "func (r *Client) CreateSnapshot(ctx context.Context, a CreateSnapshotRequest) (StatusMessage, error) {\n\tvar (\n\t\traw []byte\n\t\tresp StatusMessage\n\t\terr error\n\t\tcode int\n\t)\n\tif raw, err = json.Marshal(a); err != nil {\n\t\treturn StatusMessage{}, errors.Wrap(err, \"marshal request\")\n\t}\n\tif raw, code, err = r.post(ctx, \"api/snapshots\", nil, raw); err != nil {\n\t\treturn StatusMessage{}, errors.Wrap(err, \"create snapshot\")\n\t}\n\tif code/100 != 2 {\n\t\treturn StatusMessage{}, fmt.Errorf(\"bad response: %d\", code)\n\t}\n\tif err = json.Unmarshal(raw, &resp); err != nil {\n\t\treturn StatusMessage{}, errors.Wrap(err, \"unmarshal response message\")\n\t}\n\treturn resp, nil\n}", "func (c *Client) CreateSnapshot(ctx context.Context, params *CreateSnapshotInput, optFns ...func(*Options)) (*CreateSnapshotOutput, error) {\n\tif params == nil {\n\t\tparams = &CreateSnapshotInput{}\n\t}\n\n\tresult, metadata, err := c.invokeOperation(ctx, \"CreateSnapshot\", params, optFns, c.addOperationCreateSnapshotMiddlewares)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tout := result.(*CreateSnapshotOutput)\n\tout.ResultMetadata = metadata\n\treturn out, nil\n}", "func NewSnapshotCreateDefault(code int) *SnapshotCreateDefault {\n\treturn &SnapshotCreateDefault{\n\t\t_statusCode: code,\n\t}\n}", "func (vm *VM) CreateSnapshot(prefix string, description string) (Snapshot,\n\terror) {\n\tvar descriptor libvirtxml.DomainSnapshot\n\n\tfor true {\n\t\tdescriptor = libvirtxml.DomainSnapshot{\n\t\t\tName: prefix + namesgenerator.GetRandomName(0),\n\t\t\tDescription: description,\n\t\t}\n\n\t\t// check if name is already given\n\t\tregex := []string{\"^\" + descriptor.Name + \"$\"}\n\t\tsnapshots, err := vm.ListMatchingSnapshots(regex)\n\t\tif err != nil {\n\t\t\terr = fmt.Errorf(\"unable to retrieve existing snapshot for VM '%s': %s\",\n\t\t\t\tvm.Descriptor.Name,\n\t\t\t\terr,\n\t\t\t)\n\t\t\treturn Snapshot{}, err\n\t\t}\n\n\t\tif len(snapshots) == 0 {\n\t\t\tbreak\n\t\t}\n\t}\n\n\t// create snapshot with the given name\n\txml, err := descriptor.Marshal()\n\tif err != nil {\n\t\terr = fmt.Errorf(\"unable to marshal snapshot XML for VM '%s': %s\",\n\t\t\tvm.Descriptor.Name,\n\t\t\terr,\n\t\t)\n\t\treturn Snapshot{}, err\n\t}\n\n\tsnapshot, err := vm.Instance.CreateSnapshotXML(xml, 0)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"unable to create snapshot for VM '%s': %s\",\n\t\t\tvm.Descriptor.Name,\n\t\t\terr,\n\t\t)\n\t\treturn Snapshot{}, err\n\t}\n\n\treturn Snapshot{\n\t\tInstance: *snapshot,\n\t\tDescriptor: descriptor,\n\t}, nil\n}", "func (d *DigitalOcean) CreateSnapshot(dropletId uint, name string) error {\n\treturn d.Client.CreateSnapshot(dropletId, name)\n}", "func (s *KVStore) CreateSnapshot() []byte {\n\tlog.Printf(\"Dump out all the values from KV store\")\n\twrite := new(bytes.Buffer)\n\tencoder := gob.NewEncoder(write)\n\tencoder.Encode(s.store)\n\tsnapshot := write.Bytes()\n\treturn snapshot\n}", "func (p *Plugin) CreateVolumeFromSnapshot(snapshotID, volumeType, volumeAZ string, iops *int64) (string, error) {\n\treturn p.plugin.CreateVolumeFromSnapshot(snapshotID, volumeType, volumeAZ, iops)\n}", "func (c *gRPCClient) CreateSnapshotOperation(name string) *CreateSnapshotOperation {\n\treturn &CreateSnapshotOperation{\n\t\tlro: longrunning.InternalNewOperation(*c.LROClient, &longrunningpb.Operation{Name: name}),\n\t}\n}", "func newNativeSnapshot(c *C.rocksdb_snapshot_t) *Snapshot {\n\treturn &Snapshot{c: c}\n}", "func (mr *MockRDSAPIMockRecorder) CreateDBSnapshot(arg0 interface{}) *gomock.Call {\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"CreateDBSnapshot\", reflect.TypeOf((*MockRDSAPI)(nil).CreateDBSnapshot), arg0)\n}", "func (c *restClient) CreateSnapshotOperation(name string) *CreateSnapshotOperation {\n\toverride := fmt.Sprintf(\"/v1/%s\", name)\n\treturn &CreateSnapshotOperation{\n\t\tlro: longrunning.InternalNewOperation(*c.LROClient, &longrunningpb.Operation{Name: name}),\n\t\tpollPath: override,\n\t}\n}", "func MakeInstanceSnapshot(svc rdsiface.RDSAPI, instance *string) error {\n // snippet-start:[rds.go.create_instance_snapshot.call]\n // Get the current date and time to uniquely identify snapshot\n currentTime := time.Now()\n t := currentTime.Format(\"2006-01-02 15:04:05\")\n // Replace space with underscore for snapshot ID\n t = strings.Replace(t, \" \", \"_\", -1)\n\n _, err := svc.CreateDBSnapshot(&rds.CreateDBSnapshotInput{\n DBInstanceIdentifier: instance,\n DBSnapshotIdentifier: aws.String(*instance + t),\n })\n // snippet-end:[rds.go.create_instance_snapshot.call]\n if err != nil {\n return err\n }\n\n return nil\n}", "func (handler *volumeHandlerFile) CreateSnapshot() (interface{}, time.Time, error) {\n\thandler.log.Noticef(\"CreateSnapshot for a file based volume (%s)\", handler.status.FileLocation)\n\tcreateSnapContext := context.Background()\n\tsnapshotName := handler.status.Key() + \"-snapshot-\" + time.Now().Format(\"20060102150405\")\n\tbaseImagFile := handler.status.FileLocation\n\t// XXX: we only support qcow2 for now\n\tif handler.status.ContentFormat != zconfig.Format_QCOW2 {\n\t\treturn \"\", time.Time{}, fmt.Errorf(\"CreateSnapshot: unsupported format %s\", handler.status.ContentFormat.String())\n\t}\n\terr := diskmetrics.CreateSnapshot(createSnapContext, handler.log, baseImagFile, snapshotName)\n\tif err != nil {\n\t\thandler.log.Errorf(\"CreateSnapshot: error creating snapshot image: %s\", err)\n\t\treturn \"\", time.Time{}, err\n\t}\n\t// Replace VolumeStatus with a new snapshot file\n\ttimeCreated := time.Now()\n\thandler.log.Noticef(\"CreateSnapshot: created snapshot %s %s\", snapshotName, timeCreated.Format(\"02.01.2006 at 15:04:05\"))\n\treturn snapshotName, timeCreated, nil\n}", "func (client *Client) CreateSnapshot(name string, snapshotName string) (*Response, *ResponseStatus, error) {\n\treturn client.FormattedRequest(\"/create/snapshots/volumes/%q/%q\", name, snapshotName)\n}", "func (p *Provider) CreateFromSnapshot(snapshotDir string) (ledger.PeerLedger, string, error) {\n\tmetadataJSONs, err := loadSnapshotMetadataJSONs(snapshotDir)\n\tif err != nil {\n\t\treturn nil, \"\", errors.WithMessagef(err, \"error while loading metadata\")\n\t}\n\n\tmetadata, err := metadataJSONs.ToMetadata()\n\tif err != nil {\n\t\treturn nil, \"\", errors.WithMessagef(err, \"error while unmarshalling metadata\")\n\t}\n\n\tif err := verifySnapshot(snapshotDir, metadata, p.initializer.HashProvider); err != nil {\n\t\treturn nil, \"\", errors.WithMessagef(err, \"error while verifying snapshot\")\n\t}\n\n\tledgerID := metadata.ChannelName\n\tlastBlockNum := metadata.LastBlockNumber\n\tlogger.Debugw(\"Verified hashes\", \"snapshotDir\", snapshotDir, \"ledgerID\", ledgerID)\n\n\tlastBlkHash, err := hex.DecodeString(metadata.LastBlockHashInHex)\n\tif err != nil {\n\t\treturn nil, \"\", errors.Wrapf(err, \"error while decoding last block hash\")\n\t}\n\tpreviousBlkHash, err := hex.DecodeString(metadata.PreviousBlockHashInHex)\n\tif err != nil {\n\t\treturn nil, \"\", errors.Wrapf(err, \"error while decoding previous block hash\")\n\t}\n\n\tsnapshotInfo := &blkstorage.SnapshotInfo{\n\t\tLastBlockNum: lastBlockNum,\n\t\tLastBlockHash: lastBlkHash,\n\t\tPreviousBlockHash: previousBlkHash,\n\t}\n\n\tif err = p.idStore.createLedgerID(\n\t\tledgerID,\n\t\t&msgs.LedgerMetadata{\n\t\t\tStatus: msgs.Status_UNDER_CONSTRUCTION,\n\t\t\tBootSnapshotMetadata: &msgs.BootSnapshotMetadata{\n\t\t\t\tSingableMetadata: metadataJSONs.signableMetadata,\n\t\t\t\tAdditionalMetadata: metadataJSONs.additionalMetadata,\n\t\t\t},\n\t\t},\n\t); err != nil {\n\t\treturn nil, \"\", errors.WithMessagef(err, \"error while creating ledger id\")\n\t}\n\n\tsavepoint := version.NewHeight(lastBlockNum, math.MaxUint64)\n\n\tif err = p.blkStoreProvider.ImportFromSnapshot(ledgerID, snapshotDir, snapshotInfo); err != nil {\n\t\treturn nil, \"\", p.deleteUnderConstructionLedger(\n\t\t\tnil,\n\t\t\tledgerID,\n\t\t\terrors.WithMessage(err, \"error while importing data into block store\"),\n\t\t)\n\t}\n\tlogger.Debugw(\"Imported data into blockstore\", \"ledgerID\", ledgerID)\n\n\tif err = p.configHistoryMgr.ImportFromSnapshot(metadata.ChannelName, snapshotDir); err != nil {\n\t\treturn nil, \"\", p.deleteUnderConstructionLedger(\n\t\t\tnil,\n\t\t\tledgerID,\n\t\t\terrors.WithMessage(err, \"error while importing data into config history Mgr\"),\n\t\t)\n\t}\n\tlogger.Debugw(\"Imported data into collection config history\", \"ledgerID\", ledgerID)\n\n\tconfigHistoryRetiever := p.configHistoryMgr.GetRetriever(ledgerID)\n\tbtlPolicy := pvtdatapolicy.ConstructBTLPolicy(\n\t\t&mostRecentCollectionConfigFetcher{\n\t\t\tDeployedChaincodeInfoProvider: p.initializer.DeployedChaincodeInfoProvider,\n\t\t\tRetriever: configHistoryRetiever,\n\t\t},\n\t)\n\tpurgeMgrBuilder := pvtstatepurgemgmt.NewPurgeMgrBuilder(ledgerID, btlPolicy, p.bookkeepingProvider)\n\tlogger.Debugw(\"Constructed pvtdata hashes consumer for purge Mgr\", \"ledgerID\", ledgerID)\n\n\tpvtdataStoreBuilder, err := p.pvtdataStoreProvider.SnapshotDataImporterFor(\n\t\tledgerID, lastBlockNum, p.initializer.MembershipInfoProvider, configHistoryRetiever,\n\t\tSnapshotsTempDirPath(p.initializer.Config.SnapshotsConfig.RootDir),\n\t)\n\tif err != nil {\n\t\treturn nil, \"\", p.deleteUnderConstructionLedger(\n\t\t\tnil,\n\t\t\tledgerID,\n\t\t\terrors.WithMessage(err, \"error while getting pvtdata hashes consumer for pvtdata store\"),\n\t\t)\n\t}\n\tlogger.Debugw(\"Constructed pvtdata hashes consumer for pvt data store\", \"ledgerID\", ledgerID)\n\n\tif err = p.dbProvider.ImportFromSnapshot(ledgerID, savepoint, snapshotDir, purgeMgrBuilder, pvtdataStoreBuilder); err != nil {\n\t\treturn nil, \"\", p.deleteUnderConstructionLedger(\n\t\t\tnil,\n\t\t\tledgerID,\n\t\t\terrors.WithMessage(err, \"error while importing data into state db\"),\n\t\t)\n\t}\n\tlogger.Debugw(\"Imported data into statedb, purgeMgr, and pvtdata store\", \"ledgerID\", ledgerID)\n\n\tif p.historydbProvider != nil {\n\t\tif err := p.historydbProvider.MarkStartingSavepoint(ledgerID, savepoint); err != nil {\n\t\t\treturn nil, \"\", p.deleteUnderConstructionLedger(\n\t\t\t\tnil,\n\t\t\t\tledgerID,\n\t\t\t\terrors.WithMessage(err, \"error while preparing history db\"),\n\t\t\t)\n\t\t}\n\t\tlogger.Debugw(\"Preparing history db\", \"ledgerID\", ledgerID)\n\t}\n\n\tlgr, err := p.open(ledgerID, metadata, true)\n\tif err != nil {\n\t\treturn nil, \"\", p.deleteUnderConstructionLedger(\n\t\t\tlgr,\n\t\t\tledgerID,\n\t\t\terrors.WithMessage(err, \"error while opening ledger\"),\n\t\t)\n\t}\n\n\tif err = p.idStore.updateLedgerStatus(ledgerID, msgs.Status_ACTIVE); err != nil {\n\t\treturn nil, \"\", p.deleteUnderConstructionLedger(\n\t\t\tlgr,\n\t\t\tledgerID,\n\t\t\terrors.WithMessage(err, \"error while updating the ledger status to Status_ACTIVE\"),\n\t\t)\n\t}\n\treturn lgr, ledgerID, nil\n}", "func (c *VolumeController) createSnapshot(snapshotName string, labels map[string]string, volume *longhorn.Volume, e *longhorn.Engine) (*longhorn.SnapshotInfo, error) {\n\tif volume.Name == \"\" {\n\t\treturn nil, fmt.Errorf(\"volume name required\")\n\t}\n\n\tif err := util.VerifySnapshotLabels(labels); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err := c.checkVolumeNotInMigration(volume); err != nil {\n\t\treturn nil, err\n\t}\n\n\tengineCliClient, err := engineapi.GetEngineBinaryClient(c.ds, volume.Name, c.controllerID)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tengineClientProxy, err := engineapi.GetCompatibleClient(e, engineCliClient, c.ds, c.logger, c.proxyConnCounter)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer engineClientProxy.Close()\n\n\t// Check if we have already created a snapshot with this name.\n\t// TODO: Update longhorn-engine and longhorn-instance-manager so that SnapshotCreate returns an identifiable\n\t// error/code when a snapshot exists so that this check isn't necessary.\n\tif snapshotName != \"\" {\n\t\tsnap, err := engineClientProxy.SnapshotGet(e, snapshotName)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif snap != nil {\n\t\t\treturn snap, nil\n\t\t}\n\t}\n\n\tsnapshotName, err = engineClientProxy.SnapshotCreate(e, snapshotName, labels)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tsnap, err := engineClientProxy.SnapshotGet(e, snapshotName)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif snap == nil {\n\t\treturn nil, fmt.Errorf(\"failed to found just created snapshot '%s', for volume '%s'\", snapshotName, volume.Name)\n\t}\n\n\tlogrus.Infof(\"Created snapshot %v with labels %+v for volume %v\", snapshotName, labels, volume.Name)\n\treturn snap, nil\n}", "func (d *DefaultDriver) ValidateCreateSnapshotUsingPxctl(name string) error {\n\treturn &errors.ErrNotSupported{\n\t\tType: \"Function\",\n\t\tOperation: \"ValidateCreateSnapshotUsingPxctl()\",\n\t}\n}", "func (img *Image) CreateConsistentSnapshot(name string, onlyIfMapped bool) (*Snapshot, error) {\n\tblk, err := device(img)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif onlyIfMapped && blk == \"\" {\n\t\treturn nil, ErrNotMapped\n\t}\n\tif blk != \"\" {\n\t\tunfreeze, err := fsFreezeBlk(blk)\n\t\tdefer unfreeze()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn img.CreateSnapshot(name)\n}", "func (c *Client) CreateSnapshot(ctx context.Context, req *netapppb.CreateSnapshotRequest, opts ...gax.CallOption) (*CreateSnapshotOperation, error) {\n\treturn c.internalClient.CreateSnapshot(ctx, req, opts...)\n}", "func NewSnapshot(kind string, startRevision, lastRevision int64, compressionSuffix string, isFinal bool) *brtypes.Snapshot {\n\tsnap := &brtypes.Snapshot{\n\t\tKind: kind,\n\t\tStartRevision: startRevision,\n\t\tLastRevision: lastRevision,\n\t\tCreatedOn: time.Now().UTC(),\n\t\tCompressionSuffix: compressionSuffix,\n\t\tIsFinal: isFinal,\n\t}\n\tsnap.GenerateSnapshotName()\n\treturn snap\n}", "func (f *IBMPIInstanceClient) CreatePvmSnapShot(snapshotdef *p_cloud_p_vm_instances.PcloudPvminstancesSnapshotsPostParams, pvminstanceid, powerinstanceid string, timeout time.Duration) (*models.SnapshotCreateResponse, error) {\n\n\tparams := p_cloud_p_vm_instances.NewPcloudPvminstancesSnapshotsPostParamsWithTimeout(helpers.PICreateTimeOut).WithPvmInstanceID(pvminstanceid).WithCloudInstanceID(powerinstanceid).WithBody(snapshotdef.Body)\n\tsnapshotpostaccepted, err := f.session.Power.PCloudPVMInstances.PcloudPvminstancesSnapshotsPost(params, ibmpisession.NewAuth(f.session, powerinstanceid))\n\tif err != nil || snapshotpostaccepted == nil {\n\t\treturn nil, fmt.Errorf(\"Failed to Create the snapshot %s for the pvminstance : %s\", pvminstanceid, err)\n\t}\n\treturn snapshotpostaccepted.Payload, nil\n}", "func (db *DB) NewSnapshot() *Snapshot {\n\tif db.closed {\n\t\tpanic(ErrDBClosed)\n\t}\n\n\treturn &Snapshot{C.leveldb_create_snapshot(db.Ldb)}\n}", "func ExampleRDS_CreateDBSnapshot_shared00() {\n\tsvc := rds.New(session.New())\n\tinput := &rds.CreateDBSnapshotInput{\n\t\tDBInstanceIdentifier: aws.String(\"mydbsnapshot\"),\n\t\tDBSnapshotIdentifier: aws.String(\"database-mysql\"),\n\t}\n\n\tresult, err := svc.CreateDBSnapshot(input)\n\tif err != nil {\n\t\tif aerr, ok := err.(awserr.Error); ok {\n\t\t\tswitch aerr.Code() {\n\t\t\tcase rds.ErrCodeDBSnapshotAlreadyExistsFault:\n\t\t\t\tfmt.Println(rds.ErrCodeDBSnapshotAlreadyExistsFault, aerr.Error())\n\t\t\tcase rds.ErrCodeInvalidDBInstanceStateFault:\n\t\t\t\tfmt.Println(rds.ErrCodeInvalidDBInstanceStateFault, aerr.Error())\n\t\t\tcase rds.ErrCodeDBInstanceNotFoundFault:\n\t\t\t\tfmt.Println(rds.ErrCodeDBInstanceNotFoundFault, aerr.Error())\n\t\t\tcase rds.ErrCodeSnapshotQuotaExceededFault:\n\t\t\t\tfmt.Println(rds.ErrCodeSnapshotQuotaExceededFault, aerr.Error())\n\t\t\tdefault:\n\t\t\t\tfmt.Println(aerr.Error())\n\t\t\t}\n\t\t} else {\n\t\t\t// Print the error, cast err to awserr.Error to get the Code and\n\t\t\t// Message from an error.\n\t\t\tfmt.Println(err.Error())\n\t\t}\n\t\treturn\n\t}\n\n\tfmt.Println(result)\n}", "func (m *MockRDSAPI) CreateDBSnapshot(arg0 *rds.CreateDBSnapshotInput) (*rds.CreateDBSnapshotOutput, error) {\n\tret := m.ctrl.Call(m, \"CreateDBSnapshot\", arg0)\n\tret0, _ := ret[0].(*rds.CreateDBSnapshotOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (c *Client) CreateSnapshotOperation(name string) *CreateSnapshotOperation {\n\treturn c.internalClient.CreateSnapshotOperation(name)\n}", "func (_this *RaftNode) GenerateSnapshot(data interface{}) ([]byte, error) {\n\t_this.dataRWLock.Lock()\n\tdefer _this.dataRWLock.Unlock()\n\treturn json.Marshal(data)\n}", "func (f *FileSnapshotStore) Create(index, term uint64, peers []byte) (SnapshotSink, error) {\n\t// Create a new path\n\tname := snapshotName(term, index)\n\tpath := filepath.Join(f.path, name+tmpSuffix)\n\tf.logger.Printf(\"[INFO] snapshot: Creating new snapshot at %s\", path)\n\n\t// Make the directory\n\tif err := os.MkdirAll(path, 0755); err != nil {\n\t\tf.logger.Printf(\"[ERR] snapshot: Failed to make snapshot directory: %v\", err)\n\t\treturn nil, err\n\t}\n\n\t// Create the sink\n\tsink := &FileSnapshotSink{\n\t\tstore: f,\n\t\tlogger: f.logger,\n\t\tdir: path,\n\t\tmeta: fileSnapshotMeta{\n\t\t\tSnapshotMeta: SnapshotMeta{\n\t\t\t\tID: name,\n\t\t\t\tIndex: index,\n\t\t\t\tTerm: term,\n\t\t\t\tPeers: peers,\n\t\t\t},\n\t\t\tCRC: nil,\n\t\t},\n\t}\n\n\t// Write out the meta data\n\tif err := sink.writeMeta(); err != nil {\n\t\tf.logger.Printf(\"[ERR] snapshot: Failed to write metadata: %v\", err)\n\t\treturn nil, err\n\t}\n\n\t// Open the state file\n\tstatePath := filepath.Join(path, stateFilePath)\n\tfh, err := os.Create(statePath)\n\tif err != nil {\n\t\tf.logger.Printf(\"[ERR] snapshot: Failed to create state file: %v\", err)\n\t\treturn nil, err\n\t}\n\tsink.stateFile = fh\n\n\t// Create a CRC64 hash\n\tsink.stateHash = crc64.New(crc64.MakeTable(crc64.ECMA))\n\n\t// Wrap both the hash and file in a MultiWriter with buffering\n\tmulti := io.MultiWriter(sink.stateFile, sink.stateHash)\n\tsink.buffered = bufio.NewWriter(multi)\n\n\t// Done\n\treturn sink, nil\n}", "func (m *MockRDSAPI) CreateDBSnapshotWithContext(arg0 aws.Context, arg1 *rds.CreateDBSnapshotInput, arg2 ...request.Option) (*rds.CreateDBSnapshotOutput, error) {\n\tvarargs := []interface{}{arg0, arg1}\n\tfor _, a := range arg2 {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"CreateDBSnapshotWithContext\", varargs...)\n\tret0, _ := ret[0].(*rds.CreateDBSnapshotOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func createSnapshot(\n\tw io.Writer,\n\tprojectID, diskName, snapshotName, zone, region, location, diskProjectID string,\n) error {\n\t// projectID := \"your_project_id\"\n\t// diskName := \"your_disk_name\"\n\t// snapshotName := \"your_snapshot_name\"\n\t// zone := \"europe-central2-b\"\n\t// region := \"eupore-central2\"\n\t// location = \"eupore-central2\"\n\t// diskProjectID = \"YOUR_DISK_PROJECT_ID\"\n\n\tctx := context.Background()\n\n\tsnapshotsClient, err := compute.NewSnapshotsRESTClient(ctx)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"NewSnapshotsRESTClient: %w\", err)\n\t}\n\tdefer snapshotsClient.Close()\n\n\tif zone == \"\" && region == \"\" {\n\t\treturn fmt.Errorf(\"you need to specify `zone` or `region` for this function to work\")\n\t}\n\n\tif zone != \"\" && region != \"\" {\n\t\treturn fmt.Errorf(\"you can't set both `zone` and `region` parameters\")\n\t}\n\n\tif diskProjectID == \"\" {\n\t\tdiskProjectID = projectID\n\t}\n\n\tdisk := &computepb.Disk{}\n\tlocations := []string{}\n\tif location != \"\" {\n\t\tlocations = append(locations, location)\n\t}\n\n\tif zone != \"\" {\n\t\tdisksClient, err := compute.NewDisksRESTClient(ctx)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"NewDisksRESTClient: %w\", err)\n\t\t}\n\t\tdefer disksClient.Close()\n\n\t\tgetDiskReq := &computepb.GetDiskRequest{\n\t\t\tProject: projectID,\n\t\t\tZone: zone,\n\t\t\tDisk: diskName,\n\t\t}\n\n\t\tdisk, err = disksClient.Get(ctx, getDiskReq)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unable to get disk: %w\", err)\n\t\t}\n\t} else {\n\t\tregionDisksClient, err := compute.NewRegionDisksRESTClient(ctx)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"NewRegionDisksRESTClient: %w\", err)\n\t\t}\n\t\tdefer regionDisksClient.Close()\n\n\t\tgetDiskReq := &computepb.GetRegionDiskRequest{\n\t\t\tProject: projectID,\n\t\t\tRegion: region,\n\t\t\tDisk: diskName,\n\t\t}\n\n\t\tdisk, err = regionDisksClient.Get(ctx, getDiskReq)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unable to get disk: %w\", err)\n\t\t}\n\t}\n\n\treq := &computepb.InsertSnapshotRequest{\n\t\tProject: projectID,\n\t\tSnapshotResource: &computepb.Snapshot{\n\t\t\tName: proto.String(snapshotName),\n\t\t\tSourceDisk: proto.String(disk.GetSelfLink()),\n\t\t\tStorageLocations: locations,\n\t\t},\n\t}\n\n\top, err := snapshotsClient.Insert(ctx, req)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to create snapshot: %w\", err)\n\t}\n\n\tif err = op.Wait(ctx); err != nil {\n\t\treturn fmt.Errorf(\"unable to wait for the operation: %w\", err)\n\t}\n\n\tfmt.Fprintf(w, \"Snapshot created\\n\")\n\n\treturn nil\n}", "func NewSnapshot(version string,\n\tendpoints []types.Resource,\n\tclusters []types.Resource,\n\troutes []types.Resource,\n\tlisteners []types.Resource,\n\truntimes []types.Resource) Snapshot {\n\tout := Snapshot{}\n\tout.Resources[types.Endpoint] = NewResources(version, endpoints)\n\tout.Resources[types.Cluster] = NewResources(version, clusters)\n\tout.Resources[types.Route] = NewResources(version, routes)\n\tout.Resources[types.Listener] = NewResources(version, listeners)\n\tout.Resources[types.Runtime] = NewResources(version, runtimes)\n\treturn out\n}", "func (d *lvm) CreateVolumeSnapshot(snapVol Volume, op *operations.Operation) error {\n\tparentName, _, _ := api.GetParentAndSnapshotName(snapVol.name)\n\tparentVol := NewVolume(d, d.name, snapVol.volType, snapVol.contentType, parentName, snapVol.config, snapVol.poolConfig)\n\tsnapPath := snapVol.MountPath()\n\n\t// Create the parent directory.\n\terr := createParentSnapshotDirIfMissing(d.name, snapVol.volType, parentName)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\trevert := revert.New()\n\tdefer revert.Fail()\n\n\t// Create snapshot directory.\n\terr = snapVol.EnsureMountPath()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\trevert.Add(func() { _ = os.RemoveAll(snapPath) })\n\n\t_, err = d.createLogicalVolumeSnapshot(d.config[\"lvm.vg_name\"], parentVol, snapVol, true, d.usesThinpool())\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Error creating LVM logical volume snapshot: %w\", err)\n\t}\n\n\tvolDevPath := d.lvmDevPath(d.config[\"lvm.vg_name\"], snapVol.volType, snapVol.contentType, snapVol.name)\n\n\trevert.Add(func() {\n\t\t_ = d.removeLogicalVolume(volDevPath)\n\t})\n\n\t// For VMs, also snapshot the filesystem.\n\tif snapVol.IsVMBlock() {\n\t\tparentFSVol := parentVol.NewVMBlockFilesystemVolume()\n\t\tfsVol := snapVol.NewVMBlockFilesystemVolume()\n\t\t_, err = d.createLogicalVolumeSnapshot(d.config[\"lvm.vg_name\"], parentFSVol, fsVol, true, d.usesThinpool())\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Error creating LVM logical volume snapshot: %w\", err)\n\t\t}\n\t}\n\n\trevert.Success()\n\treturn nil\n}", "func Snapshot(scope *Scope, input tf.Output) (output tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"Snapshot\",\n\t\tInput: []tf.Input{\n\t\t\tinput,\n\t\t},\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func (s *API) CreateInstanceFromSnapshot(req *CreateInstanceFromSnapshotRequest, opts ...scw.RequestOption) (*Instance, error) {\n\tvar err error\n\n\tif req.Region == \"\" {\n\t\tdefaultRegion, _ := s.client.GetDefaultRegion()\n\t\treq.Region = defaultRegion\n\t}\n\n\tif fmt.Sprint(req.Region) == \"\" {\n\t\treturn nil, errors.New(\"field Region cannot be empty in request\")\n\t}\n\n\tif fmt.Sprint(req.SnapshotID) == \"\" {\n\t\treturn nil, errors.New(\"field SnapshotID cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"POST\",\n\t\tPath: \"/rdb/v1/regions/\" + fmt.Sprint(req.Region) + \"/snapshots/\" + fmt.Sprint(req.SnapshotID) + \"/create-instance\",\n\t\tHeaders: http.Header{},\n\t}\n\n\terr = scwReq.SetBody(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar resp Instance\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func (c *Client) CreateSnapshot(dir, name string) (string, error) {\n\tallowSnapshotReq := &hdfs.CreateSnapshotRequestProto{\n\t\tSnapshotRoot: &dir,\n\t\tSnapshotName: &name,\n\t}\n\tallowSnapshotRes := &hdfs.CreateSnapshotResponseProto{}\n\n\terr := c.namenode.Execute(\"createSnapshot\", allowSnapshotReq, allowSnapshotRes)\n\tif err != nil {\n\t\treturn \"\", interpretException(err)\n\t}\n\n\treturn allowSnapshotRes.GetSnapshotPath(), nil\n}", "func (r *ProjectsLocationsVolumesSnapshotsService) Create(parent string, volumesnapshot *VolumeSnapshot) *ProjectsLocationsVolumesSnapshotsCreateCall {\n\tc := &ProjectsLocationsVolumesSnapshotsCreateCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.parent = parent\n\tc.volumesnapshot = volumesnapshot\n\treturn c\n}", "func (rf *Raft) StartSnapshot(snapshotInBytes []byte) {\n rf.mu.Lock()\n defer rf.mu.Unlock()\n\n var snapshot Snapshot\n r := bytes.NewBuffer(snapshotInBytes)\n d := labgob.NewDecoder(r)\n if d.Decode(&snapshot.ServerMap) != nil {\n return\n }\n if d.Decode(&snapshot.LastApplyIdMap) != nil {\n return\n }\n if d.Decode(&snapshot.LastLogIndex) != nil {\n return\n }\n\n lastLogIndexInLogView := rf.convertToRaftLogViewIndex(snapshot.LastLogIndex)\n if lastLogIndexInLogView > 0 && lastLogIndexInLogView < len(rf.log) {\n lastIncludedTerm := rf.log[lastLogIndexInLogView].Term\n rf.cutoffLogBeforeIndex(lastLogIndexInLogView, lastIncludedTerm)\n rf.snapshottedIndex = snapshot.LastLogIndex\n\n rf.persistWithSnapshotInBytes(snapshotInBytes)\n\n if rf.state == \"Leader\" {\n go rf.sendInstallSnapshotToMultipleFollowers(snapshot.LastLogIndex, lastIncludedTerm)\n }\n\n }\n return\n}", "func (m *Nitro) NewSnapshot() (*Snapshot, error) {\n\tbuf := m.snapshots.MakeBuf()\n\tdefer m.snapshots.FreeBuf(buf)\n\n\t// Stitch all local gclists from all writers to create snapshot gclist\n\tvar head, tail *skiplist.Node\n\n\tfor w := m.wlist; w != nil; w = w.next {\n\t\tif tail == nil {\n\t\t\thead = w.gchead\n\t\t\ttail = w.gctail\n\t\t} else if w.gchead != nil {\n\t\t\ttail.SetLink(w.gchead)\n\t\t\ttail = w.gctail\n\t\t}\n\n\t\tw.gchead = nil\n\t\tw.gctail = nil\n\n\t\t// Update global stats\n\t\tm.store.Stats.Merge(&w.slSts1)\n\t\tatomic.AddInt64(&m.itemsCount, w.count)\n\t\tw.count = 0\n\t}\n\n\tsnap := &Snapshot{db: m, sn: m.GetCurrSn(), refCount: 1, count: m.ItemsCount()}\n\tm.snapshots.Insert(unsafe.Pointer(snap), CompareSnapshot, buf, &m.snapshots.Stats)\n\tsnap.gclist = head\n\tnewSn := atomic.AddUint32(&m.currSn, 1)\n\tif newSn == math.MaxUint32 {\n\t\treturn nil, ErrMaxSnapshotsLimitReached\n\t}\n\n\treturn snap, nil\n}", "func (mr *MockRDSAPIMockRecorder) CreateDBClusterSnapshot(arg0 interface{}) *gomock.Call {\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"CreateDBClusterSnapshot\", reflect.TypeOf((*MockRDSAPI)(nil).CreateDBClusterSnapshot), arg0)\n}", "func (ctrler CtrlDefReactor) OnConfigurationSnapshotCreate(obj *ConfigurationSnapshot) error {\n\tlog.Info(\"OnConfigurationSnapshotCreate is not implemented\")\n\treturn nil\n}", "func (c *containerdCAS) CreateSnapshotForImage(snapshotID, reference string) error {\n\timage, err := ctrdClient.GetImage(ctrdCtx, reference)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"CreateSnapshotForImage: Exception while getting image: %s. %s\", reference, err.Error())\n\t}\n\tdiffIDs, err := image.RootFS(ctrdCtx)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"CreateSnapshotForImage: Exception while getting \"+\n\t\t\t\"image %s rootfs. %s\", reference, err.Error())\n\t}\n\n\tparent := identity.ChainID(diffIDs).String()\n\tsnapshotter := ctrdClient.SnapshotService(defaultSnapshotter)\n\t_, err = snapshotter.Prepare(ctrdCtx, snapshotID, parent)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"CreateSnapshotForImage: Exception while creating snapshot: %s. %s\", snapshotID, err.Error())\n\t}\n\treturn nil\n}", "func New(notifier MountNotifier, statePath string) *Snapshot {\n\t// Initialize default library driver\n\ts := &Snapshot{\n\t\tnotifier: newLocalStateNotifier(statePath, notifier),\n\t}\n\ts.driver = csicommon.NewCSIDriver(\"container-snapshot.openshift.io\", \"0.0.1\", uuid.New())\n\tif s.driver == nil {\n\t\tpanic(\"unable to initialize driver\")\n\t}\n\ts.driver.AddControllerServiceCapabilities([]csi.ControllerServiceCapability_RPC_Type{\n\t\tcsi.ControllerServiceCapability_RPC_CREATE_DELETE_VOLUME,\n\t})\n\ts.driver.AddVolumeCapabilityAccessModes([]csi.VolumeCapability_AccessMode_Mode{\n\t\tcsi.VolumeCapability_AccessMode_SINGLE_NODE_WRITER,\n\t\tcsi.VolumeCapability_AccessMode_MULTI_NODE_MULTI_WRITER,\n\t})\n\n\t// Create GRPC servers\n\ts.identity = &identityServer{\n\t\tDefaultIdentityServer: csicommon.NewDefaultIdentityServer(s.driver),\n\t}\n\ts.node = &nodeServer{\n\t\tDefaultNodeServer: csicommon.NewDefaultNodeServer(s.driver),\n\t\tnotifier: s.notifier,\n\t}\n\ts.controller = &controllerServer{\n\t\tDefaultControllerServer: csicommon.NewDefaultControllerServer(s.driver),\n\t}\n\treturn s\n}", "func LookupSnapshot(ctx *pulumi.Context, args *GetSnapshotArgs) (*GetSnapshotResult, error) {\n\tinputs := make(map[string]interface{})\n\tif args != nil {\n\t\tinputs[\"dbInstanceIdentifier\"] = args.DbInstanceIdentifier\n\t\tinputs[\"dbSnapshotIdentifier\"] = args.DbSnapshotIdentifier\n\t\tinputs[\"includePublic\"] = args.IncludePublic\n\t\tinputs[\"includeShared\"] = args.IncludeShared\n\t\tinputs[\"mostRecent\"] = args.MostRecent\n\t\tinputs[\"snapshotType\"] = args.SnapshotType\n\t}\n\toutputs, err := ctx.Invoke(\"aws:rds/getSnapshot:getSnapshot\", inputs)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &GetSnapshotResult{\n\t\tAllocatedStorage: outputs[\"allocatedStorage\"],\n\t\tAvailabilityZone: outputs[\"availabilityZone\"],\n\t\tDbInstanceIdentifier: outputs[\"dbInstanceIdentifier\"],\n\t\tDbSnapshotArn: outputs[\"dbSnapshotArn\"],\n\t\tDbSnapshotIdentifier: outputs[\"dbSnapshotIdentifier\"],\n\t\tEncrypted: outputs[\"encrypted\"],\n\t\tEngine: outputs[\"engine\"],\n\t\tEngineVersion: outputs[\"engineVersion\"],\n\t\tIncludePublic: outputs[\"includePublic\"],\n\t\tIncludeShared: outputs[\"includeShared\"],\n\t\tIops: outputs[\"iops\"],\n\t\tKmsKeyId: outputs[\"kmsKeyId\"],\n\t\tLicenseModel: outputs[\"licenseModel\"],\n\t\tMostRecent: outputs[\"mostRecent\"],\n\t\tOptionGroupName: outputs[\"optionGroupName\"],\n\t\tPort: outputs[\"port\"],\n\t\tSnapshotCreateTime: outputs[\"snapshotCreateTime\"],\n\t\tSnapshotType: outputs[\"snapshotType\"],\n\t\tSourceDbSnapshotIdentifier: outputs[\"sourceDbSnapshotIdentifier\"],\n\t\tSourceRegion: outputs[\"sourceRegion\"],\n\t\tStatus: outputs[\"status\"],\n\t\tStorageType: outputs[\"storageType\"],\n\t\tVpcId: outputs[\"vpcId\"],\n\t\tId: outputs[\"id\"],\n\t}, nil\n}", "func (mdb *memdbSlice) OpenSnapshot(info SnapshotInfo) (Snapshot, error) {\n\tvar err error\n\tsnapInfo := info.(*memdbSnapshotInfo)\n\n\ts := &memdbSnapshot{slice: mdb,\n\t\tidxDefnId: mdb.idxDefnId,\n\t\tidxInstId: mdb.idxInstId,\n\t\tidxPartnId: mdb.idxPartnId,\n\t\tinfo: info.(*memdbSnapshotInfo),\n\t\tts: snapInfo.Timestamp(),\n\t\tcommitted: info.IsCommitted(),\n\t}\n\n\ts.Open()\n\ts.slice.IncrRef()\n\ts.slice.idxStats.numOpenSnapshots.Add(1)\n\n\tif s.committed && mdb.hasPersistence {\n\t\ts.info.MainSnap.Open()\n\t\tgo mdb.doPersistSnapshot(s)\n\t}\n\n\tif s.info.MainSnap == nil {\n\t\terr = mdb.loadSnapshot(s.info)\n\t\tif err != nil {\n\t\t\t// The caller may not force panic based on the error. So,\n\t\t\t// cleanup is needed when loadSnapshot returns an error.\n\t\t\ts.Close()\n\t\t}\n\t}\n\n\tif info.IsCommitted() {\n\t\tlogging.Infof(\"MemDBSlice::OpenSnapshot SliceId %v IndexInstId %v PartitionId %v Creating New \"+\n\t\t\t\"Snapshot %v\", mdb.id, mdb.idxInstId, mdb.idxPartnId, snapInfo)\n\n\t\t// Reset buffer sizes periodically\n\t\tmdb.periodicSliceBuffersReset()\n\n\t\t// Check if there are errors that need to be logged to console\n\t\tmdb.logErrorsToConsole()\n\t}\n\n\treturn s, err\n}", "func (c *snapshotRules) Create(ctx context.Context, snapshotRule *v1alpha1.SnapshotRule, opts v1.CreateOptions) (result *v1alpha1.SnapshotRule, err error) {\n\tresult = &v1alpha1.SnapshotRule{}\n\terr = c.client.Post().\n\t\tNamespace(c.ns).\n\t\tResource(\"snapshotrules\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tBody(snapshotRule).\n\t\tDo(ctx).\n\t\tInto(result)\n\treturn\n}", "func (m *MockRDSAPI) CreateDBSnapshotRequest(arg0 *rds.CreateDBSnapshotInput) (*request.Request, *rds.CreateDBSnapshotOutput) {\n\tret := m.ctrl.Call(m, \"CreateDBSnapshotRequest\", arg0)\n\tret0, _ := ret[0].(*request.Request)\n\tret1, _ := ret[1].(*rds.CreateDBSnapshotOutput)\n\treturn ret0, ret1\n}", "func (mdb *memdbSlice) NewSnapshot(ts *common.TsVbuuid, commit bool) (SnapshotInfo, error) {\n\n\tmdb.waitPersist()\n\n\tqc := atomic.LoadInt64(&mdb.qCount)\n\tif qc > 0 {\n\t\tcommon.CrashOnError(errors.New(\"Slice Invariant Violation - commit with pending mutations\"))\n\t}\n\n\tmdb.isDirty = false\n\n\t// Coming here means that cmdCh is empty and flush has finished for this index\n\tatomic.StoreUint32(&mdb.flushActive, 0)\n\n\tsnap, err := mdb.mainstore.NewSnapshot()\n\tif err == memdb.ErrMaxSnapshotsLimitReached {\n\t\tlogging.Warnf(\"Maximum snapshots limit reached for indexer. Restarting indexer...\")\n\t\tos.Exit(0)\n\t}\n\n\tnewSnapshotInfo := &memdbSnapshotInfo{\n\t\tTs: ts,\n\t\tMainSnap: snap,\n\t\tCommitted: commit,\n\t}\n\tmdb.setCommittedCount()\n\n\treturn newSnapshotInfo, err\n}", "func newSnapshotCache() cache.SnapshotCache {\n\treturn cache.NewSnapshotCache(false, tbnProxyNodeHash{}, consoleLogger{})\n}", "func (f *fragment) snapshot() error {\n\tf.totalOpN += int64(f.opN)\n\tf.totalOps += int64(f.ops)\n\tf.snapshotsTaken++\n\t_, err := unprotectedWriteToFragment(f, f.storage)\n\treturn err\n}", "func (s *storageMgr) handleCreateSnapshot(cmd Message) {\n\n\ts.supvCmdch <- &MsgSuccess{}\n\n\tlogging.Tracef(\"StorageMgr::handleCreateSnapshot %v\", cmd)\n\n\tmsgFlushDone := cmd.(*MsgMutMgrFlushDone)\n\n\tkeyspaceId := msgFlushDone.GetKeyspaceId()\n\ttsVbuuid := msgFlushDone.GetTS()\n\tstreamId := msgFlushDone.GetStreamId()\n\tflushWasAborted := msgFlushDone.GetAborted()\n\thasAllSB := msgFlushDone.HasAllSB()\n\n\tnumVbuckets := s.config[\"numVbuckets\"].Int()\n\tsnapType := tsVbuuid.GetSnapType()\n\ttsVbuuid.Crc64 = common.HashVbuuid(tsVbuuid.Vbuuids)\n\n\tstreamKeyspaceIdInstList := s.streamKeyspaceIdInstList.Get()\n\tinstIdList := streamKeyspaceIdInstList[streamId][keyspaceId]\n\n\tstreamKeyspaceIdInstsPerWorker := s.streamKeyspaceIdInstsPerWorker.Get()\n\tinstsPerWorker := streamKeyspaceIdInstsPerWorker[streamId][keyspaceId]\n\t// The num_snapshot_workers config has changed. Re-adjust the\n\t// streamKeyspaceIdInstsPerWorker map according to new snapshot workers\n\tnumSnapshotWorkers := s.getNumSnapshotWorkers()\n\tif len(instsPerWorker) != numSnapshotWorkers {\n\t\tfunc() {\n\t\t\ts.muSnap.Lock()\n\t\t\tdefer s.muSnap.Unlock()\n\n\t\t\tnewStreamKeyspaceIdInstsPerWorker := getStreamKeyspaceIdInstsPerWorker(streamKeyspaceIdInstList, numSnapshotWorkers)\n\t\t\ts.streamKeyspaceIdInstsPerWorker.Set(newStreamKeyspaceIdInstsPerWorker)\n\t\t\tinstsPerWorker = newStreamKeyspaceIdInstsPerWorker[streamId][keyspaceId]\n\t\t\tlogging.Infof(\"StorageMgr::handleCreateSnapshot Re-adjusting the streamKeyspaceIdInstsPerWorker map to %v workers. \"+\n\t\t\t\t\"StreamId: %v, keyspaceId: %v\", numSnapshotWorkers, streamId, keyspaceId)\n\t\t}()\n\t}\n\n\tif snapType == common.NO_SNAP || snapType == common.NO_SNAP_OSO {\n\t\tlogging.Debugf(\"StorageMgr::handleCreateSnapshot Skip Snapshot For %v \"+\n\t\t\t\"%v SnapType %v\", streamId, keyspaceId, snapType)\n\n\t\tindexInstMap := s.indexInstMap.Get()\n\t\tindexPartnMap := s.indexPartnMap.Get()\n\n\t\tgo s.flushDone(streamId, keyspaceId, indexInstMap, indexPartnMap,\n\t\t\tinstIdList, tsVbuuid, flushWasAborted, hasAllSB)\n\n\t\treturn\n\t}\n\n\ts.muSnap.Lock()\n\tdefer s.muSnap.Unlock()\n\n\t//pass copy of maps to worker\n\tindexInstMap := s.indexInstMap.Get()\n\tindexPartnMap := s.indexPartnMap.Get()\n\tindexSnapMap := s.indexSnapMap.Get()\n\ttsVbuuid_copy := tsVbuuid.Copy()\n\tstats := s.stats.Get()\n\n\tgo s.createSnapshotWorker(streamId, keyspaceId, tsVbuuid_copy, indexSnapMap,\n\t\tnumVbuckets, indexInstMap, indexPartnMap, instIdList, instsPerWorker, stats, flushWasAborted, hasAllSB)\n\n}", "func (c *cstor) Create() (*v1alpha1.CASSnapshot, error) {\n\t_, err := snapshot.CreateSnapshot(c.IP, c.Snap.Spec.VolumeName, c.Snap.Name)\n\t// If there is no err that means call was successful\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// we are returning the same struct that we received as input.\n\t// This would be modified when server replies back with some property of\n\t// created snapshot\n\treturn c.Snap, nil\n}", "func NewCreateSnapshotRequest(server string, id string) (*http.Request, error) {\n\tvar err error\n\n\tvar pathParam0 string\n\n\tpathParam0, err = runtime.StyleParam(\"simple\", false, \"id\", id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tqueryUrl, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbasePath := fmt.Sprintf(\"/instance/%s:create-snapshot\", pathParam0)\n\tif basePath[0] == '/' {\n\t\tbasePath = basePath[1:]\n\t}\n\n\tqueryUrl, err = queryUrl.Parse(basePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq, err := http.NewRequest(\"POST\", queryUrl.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn req, nil\n}", "func (m *MockRDSAPI) CreateDBClusterSnapshot(arg0 *rds.CreateDBClusterSnapshotInput) (*rds.CreateDBClusterSnapshotOutput, error) {\n\tret := m.ctrl.Call(m, \"CreateDBClusterSnapshot\", arg0)\n\tret0, _ := ret[0].(*rds.CreateDBClusterSnapshotOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (z *zfsctl) Snapshot(ctx context.Context, name string, properties map[string]string) *execute {\n\targs := []string{\"snapshot\", \"-r\"}\n\tif properties != nil {\n\t\tkv := \"-o \"\n\t\tfor k, v := range properties {\n\t\t\tkv += fmt.Sprintf(\"%s=%s \", k, v)\n\t\t}\n\t\targs = append(args, kv)\n\t}\n\targs = append(args, name)\n\treturn &execute{ctx: ctx, name: z.cmd, args: args}\n}", "func New(meta Metadata, data []byte) (*Snapshot, error) {\n\n\tvar b bytes.Buffer\n\tgw, err := gzip.NewWriterLevel(&b, gzip.BestSpeed)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error building gzip writer: %w\", err)\n\t}\n\tgw.Write(data)\n\tgw.Close()\n\n\treturn &Snapshot{meta: meta, data: b.Bytes()}, nil\n}", "func (ds *RegularStateMachineWrapper) StreamSnapshot(ssctx interface{},\n\twriter io.Writer) error {\n\tpanic(\"StreamSnapshot not suppose to be called on RegularStateMachineWrapper\")\n}", "func (_m *VolumeDriver) CreateSnapshot(opt *proto.CreateVolumeSnapshotOpts) (*model.VolumeSnapshotSpec, error) {\n\tret := _m.Called(opt)\n\n\tvar r0 *model.VolumeSnapshotSpec\n\tif rf, ok := ret.Get(0).(func(*proto.CreateVolumeSnapshotOpts) *model.VolumeSnapshotSpec); ok {\n\t\tr0 = rf(opt)\n\t} else {\n\t\tif ret.Get(0) != nil {\n\t\t\tr0 = ret.Get(0).(*model.VolumeSnapshotSpec)\n\t\t}\n\t}\n\n\tvar r1 error\n\tif rf, ok := ret.Get(1).(func(*proto.CreateVolumeSnapshotOpts) error); ok {\n\t\tr1 = rf(opt)\n\t} else {\n\t\tr1 = ret.Error(1)\n\t}\n\n\treturn r0, r1\n}", "func NewMockSnapshot(ctrl *gomock.Controller) *MockSnapshot {\n\tmock := &MockSnapshot{ctrl: ctrl}\n\tmock.recorder = &MockSnapshotMockRecorder{mock}\n\treturn mock\n}", "func Snapshot(d Ploop) (string, error) {\n\tvar p C.struct_ploop_snapshot_param\n\tvar uuid, err = Uuid()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tp.guid = C.CString(uuid)\n\tdefer cfree(p.guid)\n\n\tret := C.ploop_create_snapshot(d.d, &p)\n\tif ret == 0 {\n\t\tuuid = C.GoString(p.guid)\n\t}\n\n\treturn uuid, mkerr(ret)\n}", "func (fdb *fdbSlice) Snapshot() (Snapshot, error) {\n\n\ts := &fdbSnapshot{id: fdb.id,\n\t\tidxDefnId: fdb.idxDefnId,\n\t\tidxInstId: fdb.idxInstId,\n\t\tmain: fdb.main[0],\n\t\tback: fdb.back[0]}\n\n\t//store snapshot seqnum for main index\n\t{\n\t\ti, err := fdb.main[0].DbInfo()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tseq := i.LastSeqNum()\n\t\ts.mainSeqNum = seq\n\t}\n\n\t//store snapshot seqnum for back index\n\t{\n\t\ti, err := fdb.back[0].DbInfo()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tseq := i.LastSeqNum()\n\t\ts.backSeqNum = seq\n\t}\n\n\tcommon.Debugf(\"ForestDBSlice::Snapshot \\n\\tSliceId %v IndexInstId %v Created New \"+\n\t\t\"Snapshot %v\", fdb.id, fdb.idxInstId, s)\n\n\treturn s, nil\n}", "func NewSnapshotter(clientFactory *dockerclient.ClientFactory) *Snapshotter {\n\treturn &Snapshotter{\n\t\tclientFactory: clientFactory,\n\t}\n}", "func makeSnapshot(route *baseRoute, routeType string) RouteSnapshot {\n\tconf := route.config.Load().(RouteConfig)\n\tdests := make([]*Destination, len(conf.Dests()))\n\tfor i, d := range conf.Dests() {\n\t\tdests[i] = d.Snapshot()\n\t}\n\treturn RouteSnapshot{*conf.Matcher(), dests, routeType, route.key}\n\n}", "func NewSnapshot(options etcd.Options) (store.Snapshot, error) {\n\tcli, err := etcd.NewClient(options)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &snapshot{\n\t\tclient: cli,\n\t\tkey: DefaultKey,\n\t}, nil\n}", "func (d CephDriver) CreateBlockDeviceSnapshot(volumeUUID string, snapshotID string) error {\n\tvar cmd *exec.Cmd\n\tcmd = exec.Command(\"rbd\", \"--id\", d.ID, \"snap\", \"create\", volumeUUID+\"@\"+snapshotID)\n\n\tout, err := cmd.CombinedOutput()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Error when running: %v: %v: %s\", cmd.Args, err, out)\n\t}\n\n\tcmd = exec.Command(\"rbd\", \"--id\", d.ID, \"snap\", \"protect\", volumeUUID+\"@\"+snapshotID)\n\n\tout, err = cmd.CombinedOutput()\n\tif err != nil {\n\t\td.DeleteBlockDevice(volumeUUID)\n\t\treturn fmt.Errorf(\"Error when running: %v: %v: %s\", cmd.Args, err, out)\n\t}\n\treturn nil\n}", "func TestSnapshot(t *testing.T) {\n\tif testing.Short() {\n\t\tt.SkipNow()\n\t}\n\tt.Parallel()\n\n\t// Create a random file for testing and create a snapshot from it.\n\tsf := newTestFile()\n\tsp := modules.RandomSiaPath()\n\tsnap, err := sf.Snapshot(sp)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\t// Make sure the snapshot has the same fields as the SiaFile.\n\tif sf.numChunks != len(snap.staticChunks) {\n\t\tt.Errorf(\"expected %v chunks but got %v\", sf.numChunks, len(snap.staticChunks))\n\t}\n\tif sf.staticMetadata.FileSize != snap.staticFileSize {\n\t\tt.Errorf(\"staticFileSize was %v but should be %v\",\n\t\t\tsnap.staticFileSize, sf.staticMetadata.FileSize)\n\t}\n\tif sf.staticMetadata.StaticPieceSize != snap.staticPieceSize {\n\t\tt.Errorf(\"staticPieceSize was %v but should be %v\",\n\t\t\tsnap.staticPieceSize, sf.staticMetadata.StaticPieceSize)\n\t}\n\tif sf.staticMetadata.staticErasureCode.MinPieces() != snap.staticErasureCode.MinPieces() {\n\t\tt.Errorf(\"minPieces was %v but should be %v\",\n\t\t\tsf.staticMetadata.staticErasureCode.MinPieces(), snap.staticErasureCode.MinPieces())\n\t}\n\tif sf.staticMetadata.staticErasureCode.NumPieces() != snap.staticErasureCode.NumPieces() {\n\t\tt.Errorf(\"numPieces was %v but should be %v\",\n\t\t\tsf.staticMetadata.staticErasureCode.NumPieces(), snap.staticErasureCode.NumPieces())\n\t}\n\tif !reflect.DeepEqual(sf.staticMetadata.PartialChunks, snap.staticPartialChunks) {\n\t\tt.Errorf(\"combinedChunks don't match %v %v\", sf.staticMetadata.PartialChunks, snap.staticPartialChunks)\n\t}\n\tif !bytes.Equal(sf.staticMetadata.StaticMasterKey, snap.staticMasterKey.Key()) {\n\t\tt.Error(\"masterkeys don't match\")\n\t}\n\tif sf.staticMetadata.StaticMasterKeyType != snap.staticMasterKey.Type() {\n\t\tt.Error(\"masterkey types don't match\")\n\t}\n\tif sf.staticMetadata.Mode != snap.staticMode {\n\t\tt.Error(\"modes don't match\")\n\t}\n\tif len(sf.pubKeyTable) > 0 && len(snap.staticPubKeyTable) > 0 &&\n\t\t!reflect.DeepEqual(sf.pubKeyTable, snap.staticPubKeyTable) {\n\t\tt.Error(\"pubkeytables don't match\", sf.pubKeyTable, snap.staticPubKeyTable)\n\t}\n\tif sp != snap.staticSiaPath {\n\t\tt.Error(\"siapaths don't match\")\n\t}\n\t// Compare the pieces.\n\terr = sf.iterateChunksReadonly(func(chunk chunk) error {\n\t\tsfPieces, err := sf.Pieces(uint64(chunk.Index))\n\t\tif err != nil {\n\t\t\tt.Fatal(err)\n\t\t}\n\t\tsnapPieces := snap.Pieces(uint64(chunk.Index))\n\t\tif !reflect.DeepEqual(sfPieces, snapPieces) {\n\t\t\tt.Error(\"Pieces don't match\")\n\t\t}\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n}", "func (c *Client) CreateSnapshot(snap *v1beta1.VolumeSnapshot) (*v1beta1.VolumeSnapshot, error) {\n\tif err := c.initClient(); err != nil {\n\t\treturn nil, err\n\t}\n\treturn c.client.VolumeSnapshots(snap.Namespace).Create(context.TODO(), snap, metav1.CreateOptions{})\n}" ]
[ "0.7004506", "0.6814051", "0.68064874", "0.680541", "0.6802911", "0.6654403", "0.6653843", "0.6518738", "0.64791656", "0.6458399", "0.64563483", "0.6451699", "0.6406678", "0.63357264", "0.6320192", "0.6282284", "0.6281185", "0.6266", "0.62121636", "0.62093264", "0.6203276", "0.6184339", "0.6175579", "0.6162491", "0.61258256", "0.6087091", "0.6067748", "0.60645396", "0.60337645", "0.60332894", "0.6023515", "0.5992803", "0.59516805", "0.5950412", "0.5929508", "0.5911754", "0.59113616", "0.5908774", "0.5898449", "0.58834803", "0.5879082", "0.58683884", "0.58323866", "0.5821274", "0.58210975", "0.58168715", "0.579799", "0.57956344", "0.5782307", "0.5777193", "0.5738342", "0.5687348", "0.5651348", "0.5636348", "0.5609694", "0.55866766", "0.55738896", "0.5563178", "0.5543547", "0.553958", "0.5523388", "0.55033886", "0.55015695", "0.5475845", "0.5472528", "0.5469386", "0.545826", "0.54373467", "0.53985566", "0.5388397", "0.5329543", "0.53117883", "0.5305791", "0.52797467", "0.52694607", "0.5263553", "0.52315855", "0.5226732", "0.52215666", "0.5220113", "0.52198875", "0.5213063", "0.51759297", "0.5165906", "0.5164849", "0.5158681", "0.5155233", "0.51422405", "0.5135371", "0.5126521", "0.5125056", "0.5084175", "0.5079236", "0.5078292", "0.5047495", "0.5033037", "0.5032507", "0.5024337", "0.5009733", "0.5006059" ]
0.6494344
8
GetSnapshot uses the override method GetSnapshotFn or the real implementation.
func (c *TestClient) GetSnapshot(project, name string) (*compute.Snapshot, error) { if c.GetSnapshotFn != nil { return c.GetSnapshotFn(project, name) } return c.client.GetSnapshot(project, name) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func GetSnapshot(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *SnapshotState, opts ...pulumi.ResourceOption) (*Snapshot, error) {\n\tvar resource Snapshot\n\terr := ctx.ReadResource(\"aws-native:fsx:Snapshot\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func GetSnapshot(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *SnapshotState, opts ...pulumi.ResourceOption) (*Snapshot, error) {\n\tvar resource Snapshot\n\terr := ctx.ReadResource(\"aws:redshiftserverless/snapshot:Snapshot\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func GetSnapshot(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *SnapshotState, opts ...pulumi.ResourceOption) (*Snapshot, error) {\n\tvar resource Snapshot\n\terr := ctx.ReadResource(\"alicloud:databasefilesystem/snapshot:Snapshot\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (m *Memory) GetSnapshot() (*commonpb.Snapshot, error) {\n\treturn nil, errors.New(\"not implemented\")\n}", "func GetSnapshot(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *SnapshotState, opts ...pulumi.ResourceOption) (*Snapshot, error) {\n\tvar resource Snapshot\n\terr := ctx.ReadResource(\"alicloud:ecs/snapshot:Snapshot\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (s *API) GetSnapshot(req *GetSnapshotRequest, opts ...scw.RequestOption) (*Snapshot, error) {\n\tvar err error\n\n\tif req.Region == \"\" {\n\t\tdefaultRegion, _ := s.client.GetDefaultRegion()\n\t\treq.Region = defaultRegion\n\t}\n\n\tif fmt.Sprint(req.Region) == \"\" {\n\t\treturn nil, errors.New(\"field Region cannot be empty in request\")\n\t}\n\n\tif fmt.Sprint(req.SnapshotID) == \"\" {\n\t\treturn nil, errors.New(\"field SnapshotID cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"GET\",\n\t\tPath: \"/rdb/v1/regions/\" + fmt.Sprint(req.Region) + \"/snapshots/\" + fmt.Sprint(req.SnapshotID) + \"\",\n\t\tHeaders: http.Header{},\n\t}\n\n\tvar resp Snapshot\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func (txn *tikvTxn) GetSnapshot() kv.Snapshot {\n\treturn &tikvSnapshot{txn.KVTxn.GetSnapshot(), txn.snapshotInterceptor}\n}", "func GetSnapshot(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *SnapshotState, opts ...pulumi.ResourceOption) (*Snapshot, error) {\n\tvar resource Snapshot\n\terr := ctx.ReadResource(\"gcp:compute/snapshot:Snapshot\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (s *RaftStateMachine) GetSnapshot(ctx context.Context) ([]byte, error) {\n\t// Block until we get a signal on getSnapshotC\n\t<-s.getSnapshotC\n\treturn protoutil.Marshal(s.State(ctx))\n}", "func LookupSnapshot(ctx *pulumi.Context, args *GetSnapshotArgs) (*GetSnapshotResult, error) {\n\tinputs := make(map[string]interface{})\n\tif args != nil {\n\t\tinputs[\"dbInstanceIdentifier\"] = args.DbInstanceIdentifier\n\t\tinputs[\"dbSnapshotIdentifier\"] = args.DbSnapshotIdentifier\n\t\tinputs[\"includePublic\"] = args.IncludePublic\n\t\tinputs[\"includeShared\"] = args.IncludeShared\n\t\tinputs[\"mostRecent\"] = args.MostRecent\n\t\tinputs[\"snapshotType\"] = args.SnapshotType\n\t}\n\toutputs, err := ctx.Invoke(\"aws:rds/getSnapshot:getSnapshot\", inputs)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &GetSnapshotResult{\n\t\tAllocatedStorage: outputs[\"allocatedStorage\"],\n\t\tAvailabilityZone: outputs[\"availabilityZone\"],\n\t\tDbInstanceIdentifier: outputs[\"dbInstanceIdentifier\"],\n\t\tDbSnapshotArn: outputs[\"dbSnapshotArn\"],\n\t\tDbSnapshotIdentifier: outputs[\"dbSnapshotIdentifier\"],\n\t\tEncrypted: outputs[\"encrypted\"],\n\t\tEngine: outputs[\"engine\"],\n\t\tEngineVersion: outputs[\"engineVersion\"],\n\t\tIncludePublic: outputs[\"includePublic\"],\n\t\tIncludeShared: outputs[\"includeShared\"],\n\t\tIops: outputs[\"iops\"],\n\t\tKmsKeyId: outputs[\"kmsKeyId\"],\n\t\tLicenseModel: outputs[\"licenseModel\"],\n\t\tMostRecent: outputs[\"mostRecent\"],\n\t\tOptionGroupName: outputs[\"optionGroupName\"],\n\t\tPort: outputs[\"port\"],\n\t\tSnapshotCreateTime: outputs[\"snapshotCreateTime\"],\n\t\tSnapshotType: outputs[\"snapshotType\"],\n\t\tSourceDbSnapshotIdentifier: outputs[\"sourceDbSnapshotIdentifier\"],\n\t\tSourceRegion: outputs[\"sourceRegion\"],\n\t\tStatus: outputs[\"status\"],\n\t\tStorageType: outputs[\"storageType\"],\n\t\tVpcId: outputs[\"vpcId\"],\n\t\tId: outputs[\"id\"],\n\t}, nil\n}", "func (s *SnapshotsServiceOp) Get(snapshotID string) (*Snapshot, *Response, error) {\n\treturn s.get(interface{}(snapshotID))\n}", "func (_Bindings *BindingsCallerSession) GetAccountSnapshot(account common.Address) (*big.Int, *big.Int, *big.Int, *big.Int, error) {\n\treturn _Bindings.Contract.GetAccountSnapshot(&_Bindings.CallOpts, account)\n}", "func (s *SnapshotsServiceOp) Get(ctx context.Context, snapshotID string) (*Snapshot, *Response, error) {\n\treturn s.get(ctx, snapshotID)\n}", "func (h *Hook) GetSnapshot() cp.Snapshot {\n\treturn h.file.Snapshot\n}", "func (store Storage) GetSnapshot(ver kv.Version) (kv.Snapshot, error) {\n\treturn Snapshot{}, nil\n}", "func (c *restClient) GetSnapshot(ctx context.Context, req *netapppb.GetSnapshotRequest, opts ...gax.CallOption) (*netapppb.Snapshot, error) {\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v1/%v\", req.GetName())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"name\", url.QueryEscape(req.GetName()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\topts = append((*c.CallOptions).GetSnapshot[0:len((*c.CallOptions).GetSnapshot):len((*c.CallOptions).GetSnapshot)], opts...)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tresp := &netapppb.Snapshot{}\n\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"GET\", baseUrl.String(), nil)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}, opts...)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\treturn resp, nil\n}", "func (img *Image) GetSnapshot(name string) (*Snapshot, error) {\n\tsnap := img.getSnapshot(name)\n\t_, err := snap.Info()\n\treturn snap, err\n}", "func (c *Client) GetSnapshot(ctx context.Context, req *netapppb.GetSnapshotRequest, opts ...gax.CallOption) (*netapppb.Snapshot, error) {\n\treturn c.internalClient.GetSnapshot(ctx, req, opts...)\n}", "func (_Bindings *BindingsCaller) GetAccountSnapshot(opts *bind.CallOpts, account common.Address) (*big.Int, *big.Int, *big.Int, *big.Int, error) {\n\tvar (\n\t\tret0 = new(*big.Int)\n\t\tret1 = new(*big.Int)\n\t\tret2 = new(*big.Int)\n\t\tret3 = new(*big.Int)\n\t)\n\tout := &[]interface{}{\n\t\tret0,\n\t\tret1,\n\t\tret2,\n\t\tret3,\n\t}\n\terr := _Bindings.contract.Call(opts, out, \"getAccountSnapshot\", account)\n\treturn *ret0, *ret1, *ret2, *ret3, err\n}", "func (p *Proc) GetSnapshot() cp.Snapshot {\n\treturn p.dir.Snapshot\n}", "func (_Bindings *BindingsSession) GetAccountSnapshot(account common.Address) (*big.Int, *big.Int, *big.Int, *big.Int, error) {\n\treturn _Bindings.Contract.GetAccountSnapshot(&_Bindings.CallOpts, account)\n}", "func (s *SnapshotsServiceOp) get(ctx context.Context, ID string) (*Snapshot, *Response, error) {\n\tpath := fmt.Sprintf(\"%s/%s\", snapshotBasePath, ID)\n\n\treq, err := s.client.NewRequest(ctx, http.MethodGet, path, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\troot := new(snapshotRoot)\n\tresp, err := s.client.Do(ctx, req, root)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\n\treturn root.Snapshot, resp, err\n}", "func (driver *Driver) GetSnapshot(volumeID, snapshotID, snapshotName string) ([]*storagedriver.Snapshot, error) {\n\tvar snapshotsInt []*storagedriver.Snapshot\n\tif volumeID != \"\" {\n\t\tvolumes, err := driver.getVolume(volumeID, \"\")\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tfor _, volume := range volumes {\n\t\t\tfor _, destSnap := range volume.DestSnapList {\n\t\t\t\tsnapshot, err := driver.getSnapshot(strconv.Itoa(int(destSnap.([]interface{})[2].(float64))), \"\")\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn nil, err\n\t\t\t\t}\n\n\t\t\t\tvolSize, _ := strconv.Atoi(volume.VolSize)\n\t\t\t\tsnapshotSD := &storagedriver.Snapshot{\n\t\t\t\t\tName: snapshot[0].Name,\n\t\t\t\t\tVolumeID: strconv.Itoa(volume.Index),\n\t\t\t\t\tSnapshotID: strconv.Itoa(snapshot[0].Index),\n\t\t\t\t\tVolumeSize: strconv.Itoa(volSize / 1024 / 1024),\n\t\t\t\t\tStartTime: snapshot[0].CreationTime,\n\t\t\t\t\tDescription: \"\",\n\t\t\t\t\tStatus: \"\",\n\t\t\t\t}\n\t\t\t\tsnapshotsInt = append(snapshotsInt, snapshotSD)\n\t\t\t}\n\t\t}\n\t} else {\n\t\tsnapshots, err := driver.getSnapshot(snapshotID, snapshotName)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tfor _, snapshot := range snapshots {\n\t\t\tsnapshot, err := goxtremio.GetSnapshot(strconv.Itoa(snapshot.Index), \"\")\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\n\t\t\tvolume, err := driver.getVolume(strconv.Itoa(int(snapshot.AncestorVolID[2].(float64))), \"\")\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\n\t\t\tvolSize, _ := strconv.Atoi(volume[0].VolSize)\n\t\t\tsnapshotSD := &storagedriver.Snapshot{\n\t\t\t\tName: snapshot.Name,\n\t\t\t\tVolumeID: strconv.Itoa(int(snapshot.AncestorVolID[2].(float64))),\n\t\t\t\tSnapshotID: strconv.Itoa(snapshot.Index),\n\t\t\t\tVolumeSize: strconv.Itoa(volSize / 1024 / 1024),\n\t\t\t\tStartTime: snapshot.CreationTime,\n\t\t\t\tDescription: \"\",\n\t\t\t\tStatus: \"\",\n\t\t\t}\n\t\t\tsnapshotsInt = append(snapshotsInt, snapshotSD)\n\t\t}\n\n\t}\n\n\treturn snapshotsInt, nil\n}", "func (s *DataStore) GetSnapshot(name string) (*longhorn.Snapshot, error) {\n\tresultRO, err := s.GetSnapshotRO(name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// Cannot use cached object from lister\n\treturn resultRO.DeepCopy(), nil\n}", "func (s *SnapshotsServiceOp) get(ID interface{}) (*Snapshot, *Response, error) {\n\tpath := fmt.Sprintf(\"%s/%v\", snapshotBasePath, ID)\n\n\treq, err := s.client.NewRequest(\"GET\", path, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\troot := new(snapshotRoot)\n\tresp, err := s.client.Do(req, root)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\n\treturn root.Snapshot, resp, err\n}", "func LookupSnapshot(ctx *pulumi.Context, args *LookupSnapshotArgs, opts ...pulumi.InvokeOption) (*LookupSnapshotResult, error) {\n\tvar rv LookupSnapshotResult\n\terr := ctx.Invoke(\"aws:ebs/getSnapshot:getSnapshot\", args, &rv, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &rv, nil\n}", "func (_CrToken *CrTokenCallerSession) GetAccountSnapshot(account common.Address) (*big.Int, *big.Int, *big.Int, *big.Int, error) {\n\treturn _CrToken.Contract.GetAccountSnapshot(&_CrToken.CallOpts, account)\n}", "func (api *API) GetSnapshot(number *rpc.BlockNumber) (*Snapshot, error) {\n\t// Retrieve the requested block number (or current if none requested)\n\tvar header *types.Header\n\tif number == nil || *number == rpc.LatestBlockNumber {\n\t\theader = api.chain.CurrentHeader()\n\t} else {\n\t\theader = api.chain.GetHeaderByNumber(uint64(number.Int64()))\n\t}\n\t// Ensure we have an actually valid block and return its snapshot\n\tif header == nil {\n\t\treturn nil, istanbulcommon.ErrUnknownBlock\n\t}\n\treturn api.backend.snapshot(api.chain, header.Number.Uint64(), header.Hash(), nil)\n}", "func (c *Client) GetSnapshot(ctx context.Context, id string) (snapshots.Info, error) {\n\twrapperCli, err := c.Get(ctx)\n\tif err != nil {\n\t\treturn snapshots.Info{}, fmt.Errorf(\"failed to get a containerd grpc client: %v\", err)\n\t}\n\n\tservice := wrapperCli.client.SnapshotService(CurrentSnapshotterName(ctx))\n\tdefer service.Close()\n\n\treturn service.Stat(ctx, id)\n}", "func (_CrToken *CrTokenCaller) GetAccountSnapshot(opts *bind.CallOpts, account common.Address) (*big.Int, *big.Int, *big.Int, *big.Int, error) {\n\tvar out []interface{}\n\terr := _CrToken.contract.Call(opts, &out, \"getAccountSnapshot\", account)\n\n\tif err != nil {\n\t\treturn *new(*big.Int), *new(*big.Int), *new(*big.Int), *new(*big.Int), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(*big.Int)).(**big.Int)\n\tout1 := *abi.ConvertType(out[1], new(*big.Int)).(**big.Int)\n\tout2 := *abi.ConvertType(out[2], new(*big.Int)).(**big.Int)\n\tout3 := *abi.ConvertType(out[3], new(*big.Int)).(**big.Int)\n\n\treturn out0, out1, out2, out3, err\n\n}", "func (s *Store) Snapshot() (raft.FSMSnapshot, error) {\n\t// TODO\n\treturn nil, nil\n}", "func (a *API) GetSnapshot(id string) (*bytes.Buffer, error) {\r\n\treq, err := a.newAPIRequest(\"GET\", \"cameraThumbnail\")\r\n\tif err != nil {\r\n\t\treturn nil, fmt.Errorf(\"GetSnapshot failed: %v\", err)\r\n\t}\r\n\tq := req.URL.Query()\r\n\tq.Add(\"cameraId\", id)\r\n\treq.URL.RawQuery = q.Encode()\r\n\r\n\tclient := http.DefaultClient\r\n\tresp, err := client.Do(req)\r\n\tif err != nil {\r\n\t\treturn nil, fmt.Errorf(\"GetSnapshot failed: %v\", err)\r\n\t}\r\n\r\n\tvar buf bytes.Buffer\r\n\t_, err = buf.ReadFrom(resp.Body)\r\n\tif err != nil {\r\n\t\treturn nil, fmt.Errorf(\"GetSnapshot failed: %v\", err)\r\n\t}\r\n\tresp.Body.Close()\r\n\r\n\treturn &buf, nil\r\n}", "func GetSnapshotFile() string {\n\n\t_file := Get(\"SnapshotFile\")\n\tif _file == \"\" {\n\t\t_file = \"snapshot.bin\"\n\t}\n\n\t_absFile, err := filepath.Abs(_file)\n\tif err != nil {\n\t\tlog.Fatalf(\"[!] Failed to find real path of `%s` : %s\\n\", _file, err.Error())\n\t}\n\n\treturn _absFile\n\n}", "func RunSnapshotGet(c *CmdConfig) error {\n\tif len(c.Args) == 0 {\n\t\treturn doctl.NewMissingArgsErr(c.NS)\n\t}\n\n\tss := c.Snapshots()\n\tids := c.Args\n\n\tmatchedList := make([]do.Snapshot, 0, len(ids))\n\n\tfor _, id := range ids {\n\t\ts, err := ss.Get(id)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tmatchedList = append(matchedList, *s)\n\t}\n\titem := &displayers.Snapshot{Snapshots: matchedList}\n\treturn c.Display(item)\n}", "func (p *Probe) Snapshot() error {\n\t//return p.resolvers.Snapshot()\n\treturn nil\n}", "func (obj *GenericObject) GetSnapshotObject(ctx context.Context) (*GenericObject, error) {\n\tresult := &struct {\n\t\tReturn *ObjectInterface `json:\"qReturn\"`\n\t}{}\n\terr := obj.RPC(ctx, \"GetSnapshotObject\", result)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &GenericObject{obj.GetRemoteObject(result.Return)}, err\n}", "func (fdb *fdbSlice) Snapshot() (Snapshot, error) {\n\n\ts := &fdbSnapshot{id: fdb.id,\n\t\tidxDefnId: fdb.idxDefnId,\n\t\tidxInstId: fdb.idxInstId,\n\t\tmain: fdb.main[0],\n\t\tback: fdb.back[0]}\n\n\t//store snapshot seqnum for main index\n\t{\n\t\ti, err := fdb.main[0].DbInfo()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tseq := i.LastSeqNum()\n\t\ts.mainSeqNum = seq\n\t}\n\n\t//store snapshot seqnum for back index\n\t{\n\t\ti, err := fdb.back[0].DbInfo()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tseq := i.LastSeqNum()\n\t\ts.backSeqNum = seq\n\t}\n\n\tcommon.Debugf(\"ForestDBSlice::Snapshot \\n\\tSliceId %v IndexInstId %v Created New \"+\n\t\t\"Snapshot %v\", fdb.id, fdb.idxInstId, s)\n\n\treturn s, nil\n}", "func GetSnapshotWithTS(s sessionctx.Context, ts uint64, interceptor kv.SnapshotInterceptor) kv.Snapshot {\n\tsnap := s.GetStore().GetSnapshot(kv.Version{Ver: ts})\n\tif interceptor != nil {\n\t\tsnap.SetOption(kv.SnapInterceptor, interceptor)\n\t}\n\tif s.GetSessionVars().InRestrictedSQL {\n\t\tsnap.SetOption(kv.RequestSourceInternal, true)\n\t}\n\tif tp := s.GetSessionVars().RequestSourceType; tp != \"\" {\n\t\tsnap.SetOption(kv.RequestSourceType, tp)\n\t}\n\tif tp := s.GetSessionVars().ExplicitRequestSourceType; tp != \"\" {\n\t\tsnap.SetOption(kv.ExplicitRequestSourceType, tp)\n\t}\n\tif s.GetSessionVars().LoadBasedReplicaReadThreshold > 0 {\n\t\tsnap.SetOption(kv.LoadBasedReplicaReadThreshold, s.GetSessionVars().LoadBasedReplicaReadThreshold)\n\t}\n\treturn snap\n}", "func (f *fragment) Snapshot() error {\n\tf.mu.Lock()\n\tdefer f.mu.Unlock()\n\treturn f.snapshot()\n}", "func (m *MeterSnapshot) Snapshot() metrics.Meter { return m }", "func (_CrToken *CrTokenSession) GetAccountSnapshot(account common.Address) (*big.Int, *big.Int, *big.Int, *big.Int, error) {\n\treturn _CrToken.Contract.GetAccountSnapshot(&_CrToken.CallOpts, account)\n}", "func (p *AuroraAdminClient) Snapshot(ctx context.Context) (r *Response, err error) {\n var _args345 AuroraAdminSnapshotArgs\n var _result346 AuroraAdminSnapshotResult\n if err = p.Client_().Call(ctx, \"snapshot\", &_args345, &_result346); err != nil {\n return\n }\n return _result346.GetSuccess(), nil\n}", "func (s *Store) Snapshot() (proto.Snapshot, error) {\n\treturn nil, nil\n}", "func (l *LogDB) GetSnapshot(shardID uint64,\n\treplicaID uint64) (pb.Snapshot, error) {\n\tdb, err := l.getDB(shardID, replicaID)\n\tif err != nil {\n\t\treturn pb.Snapshot{}, err\n\t}\n\treturn db.getSnapshot(shardID, replicaID)\n}", "func (w *WorkerFSM) Snapshot() (raft.FSMSnapshot, error) {\n return WorkerSnapshot{}, nil\n}", "func (b *raftBadger) Snapshot() (raft.FSMSnapshot, error) {\n\treturn newSnapshotNoop()\n}", "func (f *fsm) Snapshot() (raft.FSMSnapshot, error) {\n\tf.mu.Lock()\n\tdefer f.mu.Unlock()\n\n\t// Clone the datastore.\n\t//o := f.m.Snapshot()\n\treturn &fsmSnap{}, nil\n}", "func (p *Probe) Snapshot() error {\n\treturn p.resolvers.Snapshot()\n}", "func (p *AuroraAdminClient) Snapshot(ctx context.Context) (r *Response, err error) {\n var _args395 AuroraAdminSnapshotArgs\n var _result396 AuroraAdminSnapshotResult\n var meta thrift.ResponseMeta\n meta, err = p.Client_().Call(ctx, \"snapshot\", &_args395, &_result396)\n p.SetLastResponseMeta_(meta)\n if err != nil {\n return\n }\n return _result396.GetSuccess(), nil\n}", "func (c *TransactionFirestore) snapshot(ctx context.Context, ref *firestore.DocumentRef) (*firestore.DocumentSnapshot, error) {\n\tsnapshot, err := c.tx.Get(ref)\n\tif snapshot != nil && !snapshot.Exists() {\n\t\treturn nil, nil\n\t}\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"get snapshot\")\n\t}\n\treturn snapshot, nil\n}", "func (f *fsm) Snapshot() (raft.FSMSnapshot, error) {\n\tw := &bytes.Buffer{}\n\tf.badger.Backup(w, 0)\n\treturn &snapshot{\n\t\tstore: w.Bytes(),\n\t}, nil\n}", "func (s *SimpleFSM) Snapshot() (raft.FSMSnapshot, error) {\n\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\n\t// Clone the kvstore into a map for easy transport\n\tmapClone := make(map[string]string)\n\n\ts.db.View(func(tx *bolt.Tx) error {\n\t\t// Assume bucket exists and has keys\n\t\tb := tx.Bucket(bucket)\n\t\tc := b.Cursor()\n\t\tfor k, v := c.First(); k != nil; k, v = c.Next() {\n\t\t\tmapClone[string(k[:])] = string(v[:])\n\t\t}\n\n\t\treturn nil\n\t})\n\treturn &fsmSnapshot{kvMap: mapClone}, nil\n}", "func (c *Client) GetSnapshot(name string, namespace string) (*v1beta1.VolumeSnapshot, error) {\n\tif err := c.initClient(); err != nil {\n\t\treturn nil, err\n\t}\n\treturn c.client.VolumeSnapshots(namespace).Get(context.TODO(), name, metav1.GetOptions{})\n}", "func NewSnapshot(ctx *pulumi.Context,\n\tname string, args *SnapshotArgs, opts ...pulumi.ResourceOption) (*Snapshot, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.InstanceId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'InstanceId'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Snapshot\n\terr := ctx.RegisterResource(\"alicloud:databasefilesystem/snapshot:Snapshot\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func Snapshot(scope *Scope, input tf.Output) (output tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"Snapshot\",\n\t\tInput: []tf.Input{\n\t\t\tinput,\n\t\t},\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func (f *FSM) Snapshot() (raft.FSMSnapshot, error) {\n\tf.registry.Lock()\n\tdefer f.registry.Unlock()\n\n\ttracer := f.registry.TracerFSM()\n\n\tdatabases := []*fsmDatabaseSnapshot{}\n\n\t// Loop through all known databases and create a backup for each of\n\t// them. The filenames associated with follower connections uniquely\n\t// identify all known databases, since there will be one and only\n\t// follower connection for each known database (we never close follower\n\t// connections since database deletion is not supported).\n\tfor _, filename := range f.registry.ConnFollowerFilenames() {\n\t\tdatabase, err := f.snapshotDatabase(tracer, filename)\n\t\tif err != nil {\n\t\t\terr = errors.Wrapf(err, \"%s\", filename)\n\t\t\ttracer.Error(\"database snapshot failed\", err)\n\t\t\treturn nil, err\n\t\t}\n\t\tdatabases = append(databases, database)\n\t}\n\n\treturn &FSMSnapshot{\n\t\tindex: f.registry.Index(),\n\t\tdatabases: databases,\n\t}, nil\n}", "func NewSnapshot(ctx *pulumi.Context,\n\tname string, args *SnapshotArgs, opts ...pulumi.ResourceOption) (*Snapshot, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.VolumeId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'VolumeId'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Snapshot\n\terr := ctx.RegisterResource(\"aws-native:fsx:Snapshot\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (s *service) GetMarketSnapshot(ctx context.Context) (MarketSnapshot, error) {\n\tlogger := log.With(s.logger, \"method\", \"GetMarketSnapshot\")\n\tsnapshot := MarketSnapshot{}\n\tif Orders.IsEmpty() {\n\t\tlevel.Error(logger).Log(\"err\", ErrOrderBookIsEmpty)\n\t\treturn snapshot, ErrOrderBookIsEmpty\n\t}\n\n\tfor order := range Orders.IterBuffered() {\n\t\tval := reflect.ValueOf(order.Val)\n\n\t\tnew := MarketSnapshotItem{\n\t\t\tPrice: val.FieldByName(\"Price\").Float(),\n\t\t\tQuantity: val.FieldByName(\"Quantity\").Int(),\n\t\t}\n\t\tif val.FieldByName(\"Status\").String() == \"Active\" {\n\t\t\tif strings.ToUpper(val.FieldByName(\"Side\").String()) == \"ASK\" {\n\t\t\t\tsnapshot.Asks = append(snapshot.Asks, new)\n\t\t\t} else {\n\t\t\t\tsnapshot.Bids = append(snapshot.Bids, new)\n\t\t\t}\n\t\t}\n\n\t}\n\n\t// sorting\n\tsnapshot.Sort()\n\n\tsnapshot.Spread = spread.getSpread()\n\treturn snapshot, nil\n}", "func (_this *RaftNode) GenerateSnapshot(data interface{}) ([]byte, error) {\n\t_this.dataRWLock.Lock()\n\tdefer _this.dataRWLock.Unlock()\n\treturn json.Marshal(data)\n}", "func (rf *Raft) Snapshot(index int, snapshot []byte) {\n\t// Your code here (2D).\n\n}", "func (rf *Raft) Snapshot(index int, snapshot []byte) {\n\t// Your code here (2D).\n\n}", "func (rf *Raft) Snapshot(index int, snapshot []byte) {\n\t// Your code here (2D).\n\n}", "func (rf *Raft) Snapshot(index int, snapshot []byte) {\n\t// Your code here (2D).\n\n}", "func (rf *Raft) Snapshot(index int, snapshot []byte) {\n\t// Your code here (2D).\n\n}", "func (rf *Raft) Snapshot(index int, snapshot []byte) {\n\t// Your code here (2D).\n\n}", "func (rf *Raft) Snapshot(index int, snapshot []byte) {\n\t// Your code here (2D).\n\n}", "func (m *Nitro) GetSnapshots() []*Snapshot {\n\tvar snaps []*Snapshot\n\tbuf := m.snapshots.MakeBuf()\n\tdefer m.snapshots.FreeBuf(buf)\n\titer := m.snapshots.NewIterator(CompareSnapshot, buf)\n\titer.SeekFirst()\n\tfor ; iter.Valid(); iter.Next() {\n\t\tsnaps = append(snaps, (*Snapshot)(iter.Get()))\n\t}\n\n\treturn snaps\n}", "func (m *VirtualEndpoint) GetSnapshots()([]CloudPcSnapshotable) {\n val, err := m.GetBackingStore().Get(\"snapshots\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.([]CloudPcSnapshotable)\n }\n return nil\n}", "func (s *Snapshot) Get(key []byte) ([]byte, io.Closer, error) {\n\tif s.db == nil {\n\t\tpanic(ErrClosed)\n\t}\n\treturn s.db.getInternal(key, nil /* batch */, s)\n}", "func (a *SnapshotApiService) SnapshotsGet(ctx _context.Context, optionals *SnapshotsGetOpts) (Snapshots, *APIResponse, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodGet\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue Snapshots\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/snapshots\"\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tif optionals != nil && optionals.Pretty != nil {\n\t\tlocalVarQueryParams.Add(\"pretty\", parameterToString(*optionals.Pretty, \"\"))\n\t}\n\tif optionals != nil && optionals.Depth != nil {\n\t\tlocalVarQueryParams.Add(\"depth\", parameterToString(*optionals.Depth, \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tif optionals != nil && optionals.XContractNumber != nil {\n\t\tlocalVarHeaderParams[\"X-Contract-Number\"] = parameterToString(*optionals.XContractNumber, \"\")\n\t}\n\tif ctx != nil {\n\t\t// API Key Authentication\n\t\tif auth, ok := ctx.Value(ContextAPIKey).(APIKey); ok {\n\t\t\tvar key string\n\t\t\tif auth.Prefix != \"\" {\n\t\t\t\tkey = auth.Prefix + \" \" + auth.Key\n\t\t\t} else {\n\t\t\t\tkey = auth.Key\n\t\t\t}\n\t\t\tlocalVarHeaderParams[\"Authorization\"] = key\n\t\t}\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tlocalVarAPIResponse := &APIResponse {\n\t\tResponse: localVarHTTPResponse,\n\t\tMethod: localVarHTTPMethod,\n\t\tRequestURL: localVarPath,\n\t\tOperation: \"SnapshotsGet\",\n\t}\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarAPIResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tlocalVarAPIResponse.Payload = localVarBody\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarAPIResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarAPIResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\treturn localVarReturnValue, localVarAPIResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarAPIResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarAPIResponse, nil\n}", "func (s *DataStore) GetSnapshotRO(snapName string) (*longhorn.Snapshot, error) {\n\treturn s.snapLister.Snapshots(s.namespace).Get(snapName)\n}", "func (client *Client) Snapshot(query string, start, end time.Time, eventQuery string) (string, error) {\n\toptions := map[string]string{\"metric_query\": query, \"event_query\": eventQuery}\n\n\treturn client.SnapshotGeneric(options, start, end)\n}", "func (c *Client) FindSnapshot(vm *VirtualMachine, name string, byRef bool) (*types.ManagedObjectReference, error) {\n\tvar moRef *types.ManagedObjectReference\n\n\tif byRef {\n\t\tmoRef = &types.ManagedObjectReference{\n\t\t\tType: \"VirtualMachineSnapshot\",\n\t\t\tValue: name,\n\t\t}\n\t} else {\n\t\terr := func() error {\n\t\t\tctx, cancelFn := context.WithTimeout(context.Background(), c.timeout)\n\t\t\tdefer cancelFn()\n\n\t\t\tvar err error\n\t\t\tmoRef, err = vm.VM.FindSnapshot(ctx, name)\n\t\t\tif err := c.checkErr(ctx, err); err != nil {\n\t\t\t\treturn errors.Wrapf(err, \"While finding snapshot\")\n\t\t\t}\n\n\t\t\treturn nil\n\t\t}()\n\n\t\tif err != nil {\n\t\t\tswitch err := errors.Cause(err).(type) {\n\t\t\tcase *TimeoutExceededError:\n\t\t\t\t// handle specifically\n\t\t\t\treturn nil, fmt.Errorf(\"Timeout while attempting to find snapshot for a VM\")\n\t\t\tdefault:\n\t\t\t\t// unknown error\n\t\t\t\treturn nil, errors.Wrap(err, \"Got error while finding snapshot for a VM\")\n\t\t\t}\n\t\t}\n\t}\n\n\treturn moRef, nil\n}", "func (f *IBMPIInstanceClient) GetSnapShotVM(powerinstanceid, pvminstanceid string, timeout time.Duration) (*models.Snapshots, error) {\n\n\tparams := p_cloud_p_vm_instances.NewPcloudPvminstancesSnapshotsGetallParamsWithTimeout(helpers.PICreateTimeOut).WithCloudInstanceID(powerinstanceid).WithPvmInstanceID(pvminstanceid)\n\tresp, err := f.session.Power.PCloudPVMInstances.PcloudPvminstancesSnapshotsGetall(params, ibmpisession.NewAuth(f.session, powerinstanceid))\n\tif err != nil || resp.Payload == nil {\n\t\treturn nil, fmt.Errorf(\"Failed to Get the snapshot for the pvminstance [%s]: %s\", pvminstanceid, err)\n\t}\n\treturn resp.Payload, nil\n\n}", "func (m *Nitro) NewSnapshot() (*Snapshot, error) {\n\tbuf := m.snapshots.MakeBuf()\n\tdefer m.snapshots.FreeBuf(buf)\n\n\t// Stitch all local gclists from all writers to create snapshot gclist\n\tvar head, tail *skiplist.Node\n\n\tfor w := m.wlist; w != nil; w = w.next {\n\t\tif tail == nil {\n\t\t\thead = w.gchead\n\t\t\ttail = w.gctail\n\t\t} else if w.gchead != nil {\n\t\t\ttail.SetLink(w.gchead)\n\t\t\ttail = w.gctail\n\t\t}\n\n\t\tw.gchead = nil\n\t\tw.gctail = nil\n\n\t\t// Update global stats\n\t\tm.store.Stats.Merge(&w.slSts1)\n\t\tatomic.AddInt64(&m.itemsCount, w.count)\n\t\tw.count = 0\n\t}\n\n\tsnap := &Snapshot{db: m, sn: m.GetCurrSn(), refCount: 1, count: m.ItemsCount()}\n\tm.snapshots.Insert(unsafe.Pointer(snap), CompareSnapshot, buf, &m.snapshots.Stats)\n\tsnap.gclist = head\n\tnewSn := atomic.AddUint32(&m.currSn, 1)\n\tif newSn == math.MaxUint32 {\n\t\treturn nil, ErrMaxSnapshotsLimitReached\n\t}\n\n\treturn snap, nil\n}", "func (s *storageMgr) openSnapshot(idxInstId common.IndexInstId, partnInst PartitionInst,\n\tpartnSnapMap PartnSnapMap) (PartnSnapMap, *common.TsVbuuid, error) {\n\n\tpid := partnInst.Defn.GetPartitionId()\n\tsc := partnInst.Sc\n\n\t//there is only one slice for now\n\tslice := sc.GetSliceById(0)\n\tinfos, err := slice.GetSnapshots()\n\t// TODO: Proper error handling if possible\n\tif err != nil {\n\t\tpanic(\"Unable to read snapinfo -\" + err.Error())\n\t}\n\n\tsnapInfoContainer := NewSnapshotInfoContainer(infos)\n\tallSnapShots := snapInfoContainer.List()\n\n\tsnapFound := false\n\tusableSnapFound := false\n\tvar tsVbuuid *common.TsVbuuid\n\tfor _, snapInfo := range allSnapShots {\n\t\tsnapFound = true\n\t\tlogging.Infof(\"StorageMgr::openSnapshot IndexInst:%v Partition:%v Attempting to open snapshot (%v)\",\n\t\t\tidxInstId, pid, snapInfo)\n\t\tusableSnapshot, err := slice.OpenSnapshot(snapInfo)\n\t\tif err != nil {\n\t\t\tif err == errStorageCorrupted {\n\t\t\t\t// Slice has already cleaned up the snapshot files. Try with older snapshot.\n\t\t\t\t// Note: plasma and forestdb never return errStorageCorrupted for OpenSnapshot.\n\t\t\t\t// So, we continue only in case of MOI.\n\t\t\t\tcontinue\n\t\t\t} else {\n\t\t\t\tpanic(\"Unable to open snapshot -\" + err.Error())\n\t\t\t}\n\t\t}\n\t\tss := &sliceSnapshot{\n\t\t\tid: SliceId(0),\n\t\t\tsnap: usableSnapshot,\n\t\t}\n\n\t\ttsVbuuid = snapInfo.Timestamp()\n\n\t\tsid := SliceId(0)\n\n\t\tps := &partitionSnapshot{\n\t\t\tid: pid,\n\t\t\tslices: map[SliceId]SliceSnapshot{sid: ss},\n\t\t}\n\n\t\tpartnSnapMap[pid] = ps\n\t\tusableSnapFound = true\n\t\tbreak\n\t}\n\n\tif !snapFound {\n\t\tlogging.Infof(\"StorageMgr::openSnapshot IndexInst:%v Partition:%v No Snapshot Found.\",\n\t\t\tidxInstId, pid)\n\t\tpartnSnapMap = nil\n\t\treturn partnSnapMap, tsVbuuid, nil\n\t}\n\n\tif !usableSnapFound {\n\t\tlogging.Infof(\"StorageMgr::openSnapshot IndexInst:%v Partition:%v No Usable Snapshot Found.\",\n\t\t\tidxInstId, pid)\n\t\treturn partnSnapMap, nil, errStorageCorrupted\n\t}\n\n\treturn partnSnapMap, tsVbuuid, nil\n}", "func (f *fsm) Snapshot() (raft.FSMSnapshot, error) {\n\tf.mu.Lock()\n\tdefer f.mu.Unlock()\n\n\tvar b bytes.Buffer\n\tencoder := gob.NewEncoder(bufio.NewWriter(&b))\n\tif err := encoder.Encode(f.state); err != nil {\n\t\treturn nil, err\n\t}\n\treturn &fsmSnapshot{\n\t\tdata: b.Bytes(),\n\t}, nil\n}", "func (m *memory) Snapshot() (*loader.Snapshot, error) {\n\tif m.loaded() {\n\t\tm.RLock()\n\t\tsnap := loader.Copy(m.snap)\n\t\tm.RUnlock()\n\t\treturn snap, nil\n\t}\n\n\t// not loaded, sync\n\tif err := m.Sync(); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// make copy\n\tm.RLock()\n\tsnap := loader.Copy(m.snap)\n\tm.RUnlock()\n\n\treturn snap, nil\n}", "func (mp *metaPartition) Snapshot() (snap raftproto.Snapshot, err error) {\n\tsnap, err = newMetaItemIterator(mp)\n\treturn\n}", "func (bcsr BlobsCreateSnapshotResponse) Snapshot() string {\n\treturn bcsr.rawResponse.Header.Get(\"x-ms-snapshot\")\n}", "func openSnapshot(file string) (reader io.ReadCloser, err error) {\n\treturn os.Open(file)\n}", "func (f *fragment) snapshot() error {\n\tf.totalOpN += int64(f.opN)\n\tf.totalOps += int64(f.ops)\n\tf.snapshotsTaken++\n\t_, err := unprotectedWriteToFragment(f, f.storage)\n\treturn err\n}", "func (c *Client) WalkSnapshot(ctx context.Context, snapshotter string, fn func(context.Context, snapshots.Info) error) error {\n\twrapperCli, err := c.Get(ctx)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to get a containerd grpc client: %v\", err)\n\t}\n\n\t// if not set specific snapshotter, set snapshotter to current snaphotter\n\tif snapshotter == \"\" {\n\t\tsnapshotter = CurrentSnapshotterName(ctx)\n\t}\n\n\tservice := wrapperCli.client.SnapshotService(snapshotter)\n\tdefer service.Close()\n\n\treturn service.Walk(ctx, fn)\n}", "func (box *Box) Snapshot() (string, error) {\n\tconst snapExt = \".snap\"\n\tfiles, err := ioutil.ReadDir(box.SnapDir())\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// files are sorted alphabetically\n\tfor i := len(files) - 1; i >= 0; i-- {\n\t\tif filepath.Ext(files[i].Name()) != snapExt {\n\t\t\tcontinue\n\t\t}\n\t\treturn filepath.Join(box.SnapDir(), files[i].Name()), nil\n\t}\n\treturn \"\", ErrSnapshotNotFound\n}", "func (smp *SimpleFileBasedSnapshotMetadataPersistence) getSnapshotMetadataPath() string {\n\treturn filepath.Join(smp.SnapshotMetadataPath, \"snapshot-metadata.json\")\n}", "func (l *LevelDB) Snapshot() (pb.Snapshot, error) {\n\ts, err := l.db.GetSnapshot()\n\tif err != nil {\n\t\treturn pb.Snapshot{}, err\n\t}\n\n\t// Fetch all keys\n\titer := s.NewIterator(&util.Range{Start: nil, Limit: nil}, nil)\n\tdefer iter.Release()\n\n\tentries := []*pb.Entry{}\n\n\tfor iter.Next() {\n\t\te := &pb.Entry{}\n\t\terr := proto.Unmarshal(iter.Value(), e)\n\t\tif err != nil {\n\t\t\treturn pb.Snapshot{}, err\n\t\t}\n\n\t\tentries = append(entries, e)\n\t}\n\n\terr = iter.Error()\n\tif err != nil {\n\t\treturn pb.Snapshot{}, err\n\t}\n\n\tsd := pandos.SnapshotData{}\n\tsd.Entries = entries\n\n\tb, err := proto.Marshal(&sd)\n\tif err != nil {\n\t\treturn pb.Snapshot{}, err\n\t}\n\n\tlastIndex := uint64(0)\n\tlastTerm := uint64(0)\n\tif len(entries) > 0 {\n\t\tlastIndex = entries[len(entries)-1].Index\n\t\tlastTerm = entries[len(entries)-1].Term\n\t}\n\n\tsnap := pb.Snapshot{\n\t\tData: b,\n\t\tMetadata: pb.SnapshotMetadata{\n\t\t\tTerm: lastTerm,\n\t\t\tIndex: lastIndex,\n\t\t},\n\t}\n\n\treturn snap, err\n}", "func (mdb *memdbSlice) OpenSnapshot(info SnapshotInfo) (Snapshot, error) {\n\tvar err error\n\tsnapInfo := info.(*memdbSnapshotInfo)\n\n\ts := &memdbSnapshot{slice: mdb,\n\t\tidxDefnId: mdb.idxDefnId,\n\t\tidxInstId: mdb.idxInstId,\n\t\tidxPartnId: mdb.idxPartnId,\n\t\tinfo: info.(*memdbSnapshotInfo),\n\t\tts: snapInfo.Timestamp(),\n\t\tcommitted: info.IsCommitted(),\n\t}\n\n\ts.Open()\n\ts.slice.IncrRef()\n\ts.slice.idxStats.numOpenSnapshots.Add(1)\n\n\tif s.committed && mdb.hasPersistence {\n\t\ts.info.MainSnap.Open()\n\t\tgo mdb.doPersistSnapshot(s)\n\t}\n\n\tif s.info.MainSnap == nil {\n\t\terr = mdb.loadSnapshot(s.info)\n\t\tif err != nil {\n\t\t\t// The caller may not force panic based on the error. So,\n\t\t\t// cleanup is needed when loadSnapshot returns an error.\n\t\t\ts.Close()\n\t\t}\n\t}\n\n\tif info.IsCommitted() {\n\t\tlogging.Infof(\"MemDBSlice::OpenSnapshot SliceId %v IndexInstId %v PartitionId %v Creating New \"+\n\t\t\t\"Snapshot %v\", mdb.id, mdb.idxInstId, mdb.idxPartnId, snapInfo)\n\n\t\t// Reset buffer sizes periodically\n\t\tmdb.periodicSliceBuffersReset()\n\n\t\t// Check if there are errors that need to be logged to console\n\t\tmdb.logErrorsToConsole()\n\t}\n\n\treturn s, err\n}", "func (m Mux) Snapshot(ctx context.Context, url string) (reflow.Fileset, error) {\n\tbucket, prefix, err := m.Bucket(ctx, url)\n\tif err != nil {\n\t\treturn reflow.Fileset{}, err\n\t}\n\tfs, err := bucket.Snapshot(ctx, prefix)\n\tif err == nil {\n\t\tsetAssertions(&fs)\n\t}\n\treturn fs, err\n}", "func (a *Agent) SnapshotRPC(args *structs.SnapshotRequest, in io.Reader, out io.Writer,\n\treplyFn structs.SnapshotReplyFn) error {\n\treturn a.delegate.SnapshotRPC(args, in, out, replyFn)\n}", "func (mdb *memdbSlice) GetSnapshots() ([]SnapshotInfo, error) {\n\tvar infos []SnapshotInfo\n\tvar err error\n\n\tinfos, _, err = mdb.getSnapshots()\n\treturn infos, err\n}", "func (p *Probe) PlaySnapshot() {\n}", "func (s *Stopwatch) Snapshot() Stopwatch {\n\tret := *s\n\tret.Stop()\n\treturn ret\n}", "func (o LookupImageResultOutput) SourceSnapshot() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupImageResult) string { return v.SourceSnapshot }).(pulumi.StringOutput)\n}", "func (f *Find) Snapshot(snapshot bool) *Find {\n\tif f == nil {\n\t\tf = new(Find)\n\t}\n\n\tf.snapshot = &snapshot\n\treturn f\n}", "func (ds *RegularStateMachineWrapper) StreamSnapshot(ssctx interface{},\n\twriter io.Writer) error {\n\tpanic(\"StreamSnapshot not suppose to be called on RegularStateMachineWrapper\")\n}", "func NewSnapshot(ctx *pulumi.Context,\n\tname string, args *SnapshotArgs, opts ...pulumi.ResourceOption) (*Snapshot, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.NamespaceName == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'NamespaceName'\")\n\t}\n\tif args.SnapshotName == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'SnapshotName'\")\n\t}\n\tvar resource Snapshot\n\terr := ctx.RegisterResource(\"aws:redshiftserverless/snapshot:Snapshot\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (ckmgr *CheckpointManager) getSnapshotForSeqno(vbno uint16, seqno uint64) (uint64, uint64, error) {\n\tsnapshot_history_obj, ok1 := ckmgr.snapshot_history_map[vbno]\n\tif ok1 {\n\t\tsnapshot_history_obj.lock.RLock()\n\t\tdefer snapshot_history_obj.lock.RUnlock()\n\t\tfor i := len(snapshot_history_obj.snapshot_history) - 1; i >= 0; i-- {\n\t\t\tcur_snapshot := snapshot_history_obj.snapshot_history[i]\n\t\t\tif seqno >= cur_snapshot.start_seqno && seqno <= cur_snapshot.end_seqno {\n\t\t\t\treturn cur_snapshot.start_seqno, cur_snapshot.end_seqno, nil\n\t\t\t}\n\t\t}\n\t} else {\n\t\tpanic(fmt.Sprintf(\"%v Calling getFailoverUUIDForSeqno on an unknown vb=%v\\n\", ckmgr.pipeline.Topic(), vbno))\n\t}\n\treturn 0, 0, fmt.Errorf(\"%v Failed to find snapshot for vb=%v, seqno=%v\\n\", ckmgr.pipeline.Topic(), vbno, seqno)\n}", "func (g *NOOPTransport) GetSnapshotConnection(ctx context.Context,\n\ttarget string) (raftio.ISnapshotConnection, error) {\n\tatomic.AddUint64(&g.tryConnect, 1)\n\tif g.connReq.Fail() {\n\t\treturn nil, ErrRequestedToFail\n\t}\n\tatomic.AddUint64(&g.connected, 1)\n\treturn &NOOPSnapshotConnection{req: g.req}, nil\n}", "func (g UGaugeSnapshot) Snapshot() UGauge { return g }", "func newSnapshotCache() cache.SnapshotCache {\n\treturn cache.NewSnapshotCache(false, tbnProxyNodeHash{}, consoleLogger{})\n}" ]
[ "0.7316441", "0.697041", "0.6930554", "0.69082505", "0.69041705", "0.68724644", "0.6808403", "0.67045933", "0.6676696", "0.663725", "0.6571331", "0.6516018", "0.6500474", "0.6471139", "0.6460594", "0.64135575", "0.6407398", "0.634458", "0.6332856", "0.6315864", "0.6263", "0.62415177", "0.6224606", "0.6222956", "0.62056875", "0.6165846", "0.6159593", "0.61146367", "0.6061804", "0.6028407", "0.60162", "0.6002332", "0.5997536", "0.59879196", "0.5927531", "0.59255594", "0.5922336", "0.5915013", "0.5908007", "0.5895642", "0.58859086", "0.5880846", "0.5856121", "0.5845401", "0.58279276", "0.5795261", "0.5751002", "0.5717726", "0.5708326", "0.56890315", "0.5675127", "0.5665038", "0.56397074", "0.5632219", "0.56105983", "0.5588408", "0.5561242", "0.5558522", "0.5556401", "0.5555554", "0.5555554", "0.5555554", "0.5555554", "0.5555554", "0.5555554", "0.5555554", "0.554191", "0.5538888", "0.55236197", "0.55043304", "0.5495391", "0.5491905", "0.5485465", "0.54799324", "0.54783", "0.5474939", "0.54742897", "0.54729617", "0.54694146", "0.5461662", "0.5456834", "0.54324883", "0.5432059", "0.5424328", "0.5418845", "0.5415306", "0.5401547", "0.5397276", "0.5393798", "0.5387058", "0.5369544", "0.5367495", "0.5367477", "0.5366576", "0.5357313", "0.5348173", "0.53334355", "0.5333336", "0.53311604", "0.5324801" ]
0.70575553
1
ListSnapshots uses the override method ListSnapshotsFn or the real implementation.
func (c *TestClient) ListSnapshots(project string, opts ...ListCallOption) ([]*compute.Snapshot, error) { if c.ListSnapshotsFn != nil { return c.ListSnapshotsFn(project, opts...) } return c.client.ListSnapshots(project, opts...) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *API) ListSnapshots(req *ListSnapshotsRequest, opts ...scw.RequestOption) (*ListSnapshotsResponse, error) {\n\tvar err error\n\n\tif req.Region == \"\" {\n\t\tdefaultRegion, _ := s.client.GetDefaultRegion()\n\t\treq.Region = defaultRegion\n\t}\n\n\tdefaultPageSize, exist := s.client.GetDefaultPageSize()\n\tif (req.PageSize == nil || *req.PageSize == 0) && exist {\n\t\treq.PageSize = &defaultPageSize\n\t}\n\n\tquery := url.Values{}\n\tparameter.AddToQuery(query, \"name\", req.Name)\n\tparameter.AddToQuery(query, \"order_by\", req.OrderBy)\n\tparameter.AddToQuery(query, \"instance_id\", req.InstanceID)\n\tparameter.AddToQuery(query, \"organization_id\", req.OrganizationID)\n\tparameter.AddToQuery(query, \"project_id\", req.ProjectID)\n\tparameter.AddToQuery(query, \"page\", req.Page)\n\tparameter.AddToQuery(query, \"page_size\", req.PageSize)\n\n\tif fmt.Sprint(req.Region) == \"\" {\n\t\treturn nil, errors.New(\"field Region cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"GET\",\n\t\tPath: \"/rdb/v1/regions/\" + fmt.Sprint(req.Region) + \"/snapshots\",\n\t\tQuery: query,\n\t\tHeaders: http.Header{},\n\t}\n\n\tvar resp ListSnapshotsResponse\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func (driver *Driver) ListSnapshots(ctx context.Context, req *csi.ListSnapshotsRequest) (*csi.ListSnapshotsResponse, error) {\n\treturn nil, status.Error(codes.Unimplemented, \"\")\n}", "func (d *Driver) ListSnapshots(ctx context.Context, req *csi.ListSnapshotsRequest) (*csi.ListSnapshotsResponse, error) {\n\treturn nil, status.Error(codes.Unimplemented, \"\")\n}", "func (c *restClient) ListSnapshots(ctx context.Context, req *netapppb.ListSnapshotsRequest, opts ...gax.CallOption) *SnapshotIterator {\n\tit := &SnapshotIterator{}\n\treq = proto.Clone(req).(*netapppb.ListSnapshotsRequest)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tit.InternalFetch = func(pageSize int, pageToken string) ([]*netapppb.Snapshot, string, error) {\n\t\tresp := &netapppb.ListSnapshotsResponse{}\n\t\tif pageToken != \"\" {\n\t\t\treq.PageToken = pageToken\n\t\t}\n\t\tif pageSize > math.MaxInt32 {\n\t\t\treq.PageSize = math.MaxInt32\n\t\t} else if pageSize != 0 {\n\t\t\treq.PageSize = int32(pageSize)\n\t\t}\n\t\tbaseUrl, err := url.Parse(c.endpoint)\n\t\tif err != nil {\n\t\t\treturn nil, \"\", err\n\t\t}\n\t\tbaseUrl.Path += fmt.Sprintf(\"/v1/%v/snapshots\", req.GetParent())\n\n\t\tparams := url.Values{}\n\t\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\t\tif req.GetFilter() != \"\" {\n\t\t\tparams.Add(\"filter\", fmt.Sprintf(\"%v\", req.GetFilter()))\n\t\t}\n\t\tif req.GetOrderBy() != \"\" {\n\t\t\tparams.Add(\"orderBy\", fmt.Sprintf(\"%v\", req.GetOrderBy()))\n\t\t}\n\t\tif req.GetPageSize() != 0 {\n\t\t\tparams.Add(\"pageSize\", fmt.Sprintf(\"%v\", req.GetPageSize()))\n\t\t}\n\t\tif req.GetPageToken() != \"\" {\n\t\t\tparams.Add(\"pageToken\", fmt.Sprintf(\"%v\", req.GetPageToken()))\n\t\t}\n\n\t\tbaseUrl.RawQuery = params.Encode()\n\n\t\t// Build HTTP headers from client and context metadata.\n\t\thds := append(c.xGoogHeaders, \"Content-Type\", \"application/json\")\n\t\theaders := gax.BuildHeaders(ctx, hds...)\n\t\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\t\tif settings.Path != \"\" {\n\t\t\t\tbaseUrl.Path = settings.Path\n\t\t\t}\n\t\t\thttpReq, err := http.NewRequest(\"GET\", baseUrl.String(), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\thttpReq.Header = headers\n\n\t\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tdefer httpRsp.Body.Close()\n\n\t\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn nil\n\t\t}, opts...)\n\t\tif e != nil {\n\t\t\treturn nil, \"\", e\n\t\t}\n\t\tit.Response = resp\n\t\treturn resp.GetSnapshots(), resp.GetNextPageToken(), nil\n\t}\n\n\tfetch := func(pageSize int, pageToken string) (string, error) {\n\t\titems, nextPageToken, err := it.InternalFetch(pageSize, pageToken)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tit.items = append(it.items, items...)\n\t\treturn nextPageToken, nil\n\t}\n\n\tit.pageInfo, it.nextFunc = iterator.NewPageInfo(fetch, it.bufLen, it.takeBuf)\n\tit.pageInfo.MaxSize = int(req.GetPageSize())\n\tit.pageInfo.Token = req.GetPageToken()\n\n\treturn it\n}", "func (ks *KopiaSnapshotter) ListSnapshots() ([]string, error) {\n\treturn ks.snap.ListSnapshots()\n}", "func (s *SnapshotsServiceOp) List(ctx context.Context, opt *ListOptions) ([]Snapshot, *Response, error) {\n\treturn s.list(ctx, opt, nil)\n}", "func (l *LogDB) ListSnapshots(shardID uint64,\n\treplicaID uint64, index uint64) ([]pb.Snapshot, error) {\n\tpanic(\"depreciated\")\n}", "func (cs *controller) ListSnapshots(\n\tctx context.Context,\n\treq *csi.ListSnapshotsRequest,\n) (*csi.ListSnapshotsResponse, error) {\n\n\treturn nil, status.Error(codes.Unimplemented, \"\")\n}", "func (mw *ShardedRDB) ListSnapshots(clusterID uint64,\n\tnodeID uint64, index uint64) ([]pb.Snapshot, error) {\n\tidx := mw.partitioner.GetPartitionID(clusterID)\n\treturn mw.shards[idx].listSnapshots(clusterID, nodeID, index)\n}", "func (s *SnapshotsServiceOp) List(opt *ListOptions) ([]Snapshot, *Response, error) {\n\treturn s.list(opt, nil)\n}", "func (a *Client) ListSnapshots(params *ListSnapshotsParams, opts ...ClientOption) (*ListSnapshotsOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewListSnapshotsParams()\n\t}\n\top := &runtime.ClientOperation{\n\t\tID: \"listSnapshots\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/astrolabe/{service}/{protectedEntityID}/snapshots\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &ListSnapshotsReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t}\n\tfor _, opt := range opts {\n\t\topt(op)\n\t}\n\n\tresult, err := a.transport.Submit(op)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*ListSnapshotsOK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\t// safeguard: normally, absent a default response, unknown success responses return an error above: so this is a codegen issue\n\tmsg := fmt.Sprintf(\"unexpected success response for listSnapshots: API contract not enforced by server. Client expected to get an error, but got: %T\", result)\n\tpanic(msg)\n}", "func (s *SnapshotsServiceOp) list(ctx context.Context, opt *ListOptions, listOpt *listSnapshotOptions) ([]Snapshot, *Response, error) {\n\tpath := snapshotBasePath\n\tpath, err := addOptions(path, opt)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tpath, err = addOptions(path, listOpt)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treq, err := s.client.NewRequest(ctx, http.MethodGet, path, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\troot := new(snapshotsRoot)\n\tresp, err := s.client.Do(ctx, req, root)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\tif l := root.Links; l != nil {\n\t\tresp.Links = l\n\t}\n\n\treturn root.Snapshots, resp, err\n}", "func ListSnapshots(ctx context.Context, rep repo.Repository, si SourceInfo) ([]*Manifest, error) {\n\tentries, err := rep.FindManifests(ctx, sourceInfoToLabels(si))\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"unable to find manifest entries\")\n\t}\n\n\treturn LoadSnapshots(ctx, rep, entryIDs(entries))\n}", "func ListSnapshots(sg *snapshotgroup.SnapshotGroup) ([]GeminiSnapshot, error) {\n\tclient := kube.GetClient()\n\tsnapshots, err := client.SnapshotClient.Namespace(sg.ObjectMeta.Namespace).List(metav1.ListOptions{})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tGeminiSnapshots := []GeminiSnapshot{}\n\tfor _, snapshot := range snapshots.Items {\n\t\tsnapshotMeta, err := meta.Accessor(&snapshot)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tannotations := snapshotMeta.GetAnnotations()\n\t\tif managedBy, ok := annotations[managedByAnnotation]; !ok || managedBy != managerName {\n\t\t\tcontinue\n\t\t}\n\t\tif annotations[GroupNameAnnotation] != sg.ObjectMeta.Name {\n\t\t\tcontinue\n\t\t}\n\t\ttimestampStr := annotations[TimestampAnnotation]\n\t\ttimestamp, err := strconv.Atoi(timestampStr)\n\t\tif err != nil {\n\t\t\tklog.Errorf(\"%s/%s: failed to parse unix timestamp %s for %s\", sg.ObjectMeta.Namespace, sg.ObjectMeta.Name, timestampStr, snapshotMeta.GetName())\n\t\t\tcontinue\n\t\t}\n\t\tintervals := []string{}\n\t\tintervalsStr := annotations[IntervalsAnnotation]\n\t\tif intervalsStr != \"\" {\n\t\t\tintervals = strings.Split(intervalsStr, intervalsSeparator)\n\t\t}\n\t\tGeminiSnapshots = append(GeminiSnapshots, GeminiSnapshot{\n\t\t\tNamespace: snapshotMeta.GetNamespace(),\n\t\t\tName: snapshotMeta.GetName(),\n\t\t\tTimestamp: time.Unix(int64(timestamp), 0),\n\t\t\tIntervals: intervals,\n\t\t\tRestore: annotations[RestoreAnnotation],\n\t\t})\n\t}\n\tsort.Slice(GeminiSnapshots, func(i, j int) bool {\n\t\treturn GeminiSnapshots[j].Timestamp.Before(GeminiSnapshots[i].Timestamp)\n\t})\n\treturn GeminiSnapshots, nil\n}", "func ListSnapshots(storagDriverConfig interface{}, pred func(id string) bool) (ids []string, err error) {\n\t// create (backup) storage driver (so we can list snapshot headers from it)\n\tdriver, err := newStorageDriver(storagDriverConfig)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tids, err = driver.GetHeaders()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif pred == nil {\n\t\treturn ids, nil\n\t}\n\n\tfilterPos := 0\n\tvar ok bool\n\tfor _, id := range ids {\n\t\tok = pred(id)\n\t\tif ok {\n\t\t\tids[filterPos] = id\n\t\t\tfilterPos++\n\t\t}\n\t}\n\n\treturn ids[:filterPos], nil\n}", "func (s *SnapshotsServiceOp) list(opt *ListOptions, listOpt *listSnapshotOptions) ([]Snapshot, *Response, error) {\n\tpath := snapshotBasePath\n\tpath, err := addOptions(path, opt)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tpath, err = addOptions(path, listOpt)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treq, err := s.client.NewRequest(\"GET\", path, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\troot := new(snapshotsRoot)\n\tresp, err := s.client.Do(req, root)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\tif l := root.Links; l != nil {\n\t\tresp.Links = l\n\t}\n\n\treturn root.Snapshots, resp, err\n}", "func (n *NodeVM) ListSnapshots(ctx context.Context, node *Node) (string, error) {\n\treturn execQEMUCommand(ctx, n.monitor, listCommand)\n}", "func (client *Client) ListSnapshots(request *ListSnapshotsRequest) (_result *ListSnapshotsResponse, _err error) {\n\truntime := &util.RuntimeOptions{}\n\t_result = &ListSnapshotsResponse{}\n\t_body, _err := client.ListSnapshotsWithOptions(request, runtime)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\t_result = _body\n\treturn _result, _err\n}", "func (c *Client) ListSnapshots(ctx context.Context, req *netapppb.ListSnapshotsRequest, opts ...gax.CallOption) *SnapshotIterator {\n\treturn c.internalClient.ListSnapshots(ctx, req, opts...)\n}", "func (cs *DefaultControllerServer) ListSnapshots(ctx context.Context, req *csi.ListSnapshotsRequest) (*csi.ListSnapshotsResponse, error) {\n\treturn nil, status.Error(codes.Unimplemented, \"\")\n}", "func (cli *Client) SnapshotList(filter filters.Args) (types.SnapshotsListResponse, error) {\n\tvar snapshots types.SnapshotsListResponse\n\tquery := url.Values{}\n\n\tif filter.Len() > 0 {\n\t\tfilterJSON, err := filters.ToParam(filter)\n\t\tif err != nil {\n\t\t\treturn snapshots, err\n\t\t}\n\t\tquery.Set(\"filters\", filterJSON)\n\t}\n\tresp, err := cli.get(\"/snapshots\", query, nil)\n\tif err != nil {\n\t\treturn snapshots, err\n\t}\n\n\terr = json.NewDecoder(resp.body).Decode(&snapshots)\n\tensureReaderClosed(resp)\n\treturn snapshots, err\n}", "func (s *OsdCsiServer) ListSnapshots(\n\tctx context.Context,\n\treq *csi.ListSnapshotsRequest,\n) (*csi.ListSnapshotsResponse, error) {\n\n\t// The function ListSnapshots is also not published as\n\t// supported by this implementation\n\treturn nil, status.Error(codes.Unimplemented, \"ListSnapshots is not implemented\")\n}", "func (mr *MockConfigAdminServiceClientMockRecorder) ListSnapshots(ctx, in interface{}, opts ...interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\tvarargs := append([]interface{}{ctx, in}, opts...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"ListSnapshots\", reflect.TypeOf((*MockConfigAdminServiceClient)(nil).ListSnapshots), varargs...)\n}", "func (c *Client) SnapshotList(vm *VirtualMachine) (*Snapshot, error) {\n\tif c.Verbose {\n\t\tfmt.Printf(\"Getting a list of snapshots for VM...\\n\")\n\t}\n\n\tvar sn *Snapshot\n\n\terr := func() error {\n\t\t_, cancelFn := context.WithTimeout(context.Background(), c.timeout)\n\t\tdefer cancelFn()\n\n\t\tif vm.MO.Snapshot == nil {\n\t\t\treturn nil\n\t\t}\n\n\t\t// TODO: unclear how to proceed here.\n\n\t\treturn nil\n\t}()\n\n\tif err != nil {\n\t\tswitch err := errors.Cause(err).(type) {\n\t\tcase *TimeoutExceededError:\n\t\t\t// handle specifically\n\t\t\treturn nil, fmt.Errorf(\"Timeout while attempting to list snapshots for a VM\")\n\t\tdefault:\n\t\t\t// unknown error\n\t\t\treturn nil, errors.Wrap(err, \"Got error while listing snapshots for a VM\")\n\t\t}\n\t}\n\n\treturn sn, nil\n}", "func (s *OsdCsiServer) ListSnapshots(\n\tctx context.Context,\n\treq *csi.ListSnapshotsRequest,\n) (*csi.ListSnapshotsResponse, error) {\n\n\tif len(req.GetSnapshotId()) > 0 {\n\t\treturn s.listSingleSnapshot(ctx, req)\n\t}\n\n\treturn s.listMultipleSnapshots(ctx, req)\n}", "func RunSnapshotList(c *CmdConfig) error {\n\tvar err error\n\tss := c.Snapshots()\n\n\trestype, err := c.Doit.GetString(c.NS, doctl.ArgResourceType)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tregion, err := c.Doit.GetString(c.NS, doctl.ArgRegionSlug)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tmatches := make([]glob.Glob, 0, len(c.Args))\n\tfor _, globStr := range c.Args {\n\t\tg, err := glob.Compile(globStr)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unknown glob %q\", globStr)\n\t\t}\n\n\t\tmatches = append(matches, g)\n\t}\n\n\tvar matchedList []do.Snapshot\n\tvar list []do.Snapshot\n\n\tif restype == \"droplet\" {\n\t\tlist, err = ss.ListDroplet()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t} else if restype == \"volume\" {\n\t\tlist, err = ss.ListVolume()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\tlist, err = ss.List()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tfor _, snapshot := range list {\n\t\tvar skip = true\n\t\tif len(matches) == 0 {\n\t\t\tskip = false\n\t\t} else {\n\t\t\tfor _, m := range matches {\n\t\t\t\tif m.Match(snapshot.ID) {\n\t\t\t\t\tskip = false\n\t\t\t\t}\n\t\t\t\tif m.Match(snapshot.Name) {\n\t\t\t\t\tskip = false\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\tif !skip && region != \"\" {\n\t\t\tfor _, snapshotRegion := range snapshot.Regions {\n\t\t\t\tif region != snapshotRegion {\n\t\t\t\t\tskip = true\n\t\t\t\t} else {\n\t\t\t\t\tskip = false\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\n\t\t}\n\n\t\tif !skip {\n\t\t\tmatchedList = append(matchedList, snapshot)\n\t\t}\n\t}\n\n\titem := &displayers.Snapshot{Snapshots: matchedList}\n\treturn c.Display(item)\n}", "func (mr *MockConfigAdminServiceServerMockRecorder) ListSnapshots(arg0, arg1 interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"ListSnapshots\", reflect.TypeOf((*MockConfigAdminServiceServer)(nil).ListSnapshots), arg0, arg1)\n}", "func (s *OsdCsiServer) listMultipleSnapshots(\n\tctx context.Context,\n\treq *csi.ListSnapshotsRequest,\n) (*csi.ListSnapshotsResponse, error) {\n\tsourceVolumeId := req.GetSourceVolumeId()\n\tstartingToken := req.GetStartingToken()\n\tmaxEntries := req.GetMaxEntries()\n\n\tclogger.WithContext(ctx).Infof(\"ListSnapshots for multiple snapshots received. sourceVolumeId: %s, startingToken: %s, maxEntries: %v\",\n\t\tsourceVolumeId,\n\t\tstartingToken,\n\t\tmaxEntries,\n\t)\n\n\t// Get grpc connection\n\tconn, err := s.getConn()\n\tif err != nil {\n\t\treturn nil, status.Errorf(\n\t\t\tcodes.Unavailable,\n\t\t\t\"Unable to connect to SDK server: %v\", err)\n\t}\n\n\t// Get secret if any was passed\n\tctx = s.setupContext(ctx, req.GetSecrets())\n\tctx, cancel := grpcutil.WithDefaultTimeout(ctx)\n\tdefer cancel()\n\tvolumes := api.NewOpenStorageVolumeClient(conn)\n\n\t// Get all SnapshotIDs. Filter by source ID if provided.\n\tsnapshotsReq := &api.SdkVolumeSnapshotEnumerateWithFiltersRequest{\n\t\tVolumeId: sourceVolumeId,\n\t}\n\tsnapshotsResp, err := volumes.SnapshotEnumerateWithFilters(ctx, snapshotsReq)\n\tif err != nil {\n\t\terrStatus, ok := status.FromError(err)\n\t\tif ok && errStatus.Code() == codes.NotFound {\n\t\t\treturn &csi.ListSnapshotsResponse{}, nil\n\t\t}\n\t\treturn nil, status.Errorf(codes.Internal, \"Unable to get all snapshots: %v\", err)\n\t}\n\n\t// InspectWithFilters for all volumes\n\tvolumesResp, err := volumes.InspectWithFilters(ctx, &api.SdkVolumeInspectWithFiltersRequest{})\n\tif err != nil {\n\t\treturn nil, status.Errorf(codes.Internal, \"Unable to get all volumes: %v\", err)\n\t}\n\n\t// Sort snapshot IDs for repeatable results\n\tsortedSnapshotIds := sort.StringSlice(snapshotsResp.VolumeSnapshotIds)\n\tsort.Sort(sortedSnapshotIds)\n\n\t// Keep track of which volumes are snapshots\n\tvolumeForSnapId := make(map[string]*api.Volume)\n\tfor _, volResp := range volumesResp.Volumes {\n\t\tfor _, snapId := range sortedSnapshotIds {\n\t\t\tif volResp.Volume.Id == snapId {\n\t\t\t\tvolumeForSnapId[snapId] = volResp.Volume\n\t\t\t}\n\t\t}\n\t}\n\n\t// Generate response for all snapshots.\n\tlistSnapshotsResp := &csi.ListSnapshotsResponse{}\n\n\t// If starting token is provided, start skipping entries\n\t// until we hit the starting token.\n\tvar skipEntries bool\n\tif len(startingToken) > 0 {\n\t\tskipEntries = true\n\t}\n\tfor _, snapId := range sortedSnapshotIds {\n\t\t// Skip entries until we hit the starting token.\n\t\tif skipEntries && startingToken != snapId {\n\t\t\tcontinue\n\t\t}\n\t\tskipEntries = false\n\n\t\t// Before adding new object to response, check if we're at the max entries.\n\t\t// If we are at max entries, return with current iteration as NextToken.\n\t\t// This allows for calls to ListSnapshots to begin where we left off.\n\t\tvol := volumeForSnapId[snapId]\n\t\tif maxEntries > 0 && len(listSnapshotsResp.Entries) >= int(maxEntries) {\n\t\t\tlistSnapshotsResp.NextToken = vol.Id\n\t\t\treturn listSnapshotsResp, nil\n\t\t}\n\n\t\t// Populate entry with volume info\n\t\tentry := &csi.ListSnapshotsResponse_Entry{\n\t\t\tSnapshot: &csi.Snapshot{\n\t\t\t\tSizeBytes: int64(vol.GetSpec().GetSize()),\n\t\t\t\tSnapshotId: vol.Id,\n\t\t\t\tSourceVolumeId: vol.GetSource().Parent,\n\t\t\t\tCreationTime: vol.Ctime,\n\t\t\t\tReadyToUse: isSnapshotReady(vol),\n\t\t\t},\n\t\t}\n\n\t\tlistSnapshotsResp.Entries = append(listSnapshotsResp.Entries, entry)\n\t}\n\n\treturn listSnapshotsResp, nil\n}", "func (c *Client) ListSnapshots(namespace string) (*v1beta1.VolumeSnapshotList, error) {\n\tif err := c.initClient(); err != nil {\n\t\treturn nil, err\n\t}\n\treturn c.client.VolumeSnapshots(namespace).List(context.TODO(), metav1.ListOptions{})\n}", "func (vm *VM) ListMatchingSnapshots(regexes []string) ([]Snapshot, error) {\n\t// argument validity checking\n\texprs := make([]*regexp.Regexp, 0, len(regexes))\n\tfor _, arg := range regexes {\n\t\tregex, err := regexp.Compile(arg)\n\t\tif err != nil {\n\t\t\terr = fmt.Errorf(\"unable to compile regular expression %s: %s\", arg,\n\t\t\t\terr)\n\t\t\treturn nil, err\n\t\t}\n\t\texprs = append(exprs, regex)\n\t}\n\n\tif len(exprs) == 0 {\n\t\treturn nil, fmt.Errorf(\"no regular expression was specified\")\n\t}\n\n\t// retrieve all snapshots from libvirt\n\tinstances, err := vm.Instance.ListAllSnapshots(0)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"unable to retrieve snapshots for VM %s: %s\",\n\t\t\tvm.Descriptor.Name, err)\n\t\treturn nil, err\n\t}\n\n\tmatchedSnapshots := make([]Snapshot, 0, len(instances))\n\n\t// loop over snapshots and check for a match with the given\n\t// regular expressions\n\tfor _, instance := range instances {\n\n\t\t// retrieve and unmarshal the descriptor of snapshot\n\t\txml, err := instance.GetXMLDesc(0)\n\t\tif err != nil {\n\t\t\terr = fmt.Errorf(\"unable to get XML descriptor of snapshot: %s\", err)\n\t\t\tvm.Logger.Warnf(\"Skipping snapshot: %s\", err)\n\t\t\tcontinue\n\t\t}\n\n\t\tdescriptor := libvirtxml.DomainSnapshot{}\n\t\terr = descriptor.Unmarshal(xml)\n\t\tif err != nil {\n\t\t\terr = fmt.Errorf(\"unable to unmarshal the XML descriptor of snapshot: %s\", err)\n\t\t\tvm.Logger.Warn(\"Skipping snapshot: %s\", err)\n\t\t\tcontinue\n\t\t}\n\n\t\t// checking for a matching regular expression\n\t\tfound := false\n\t\tfor _, regex := range exprs {\n\t\t\tif regex.Find([]byte(descriptor.Name)) != nil {\n\t\t\t\tfound = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\n\t\tif found {\n\t\t\t// the caller is responsible for calling domain.Free() on the returned\n\t\t\t// domains\n\t\t\tmatchedSnapshot := Snapshot{\n\t\t\t\tInstance: instance,\n\t\t\t\tDescriptor: descriptor,\n\t\t\t}\n\t\t\tmatchedSnapshots = append(matchedSnapshots, matchedSnapshot)\n\t\t} else {\n\t\t\t// we do not need the instance here anymore\n\t\t\terr = instance.Free()\n\t\t\tif err != nil {\n\t\t\t\tvm.Logger.Warnf(\"unable to free snapshot %s: %s\",\n\t\t\t\t\tdescriptor.Name,\n\t\t\t\t\terr,\n\t\t\t\t)\n\t\t\t}\n\t\t}\n\t}\n\n\t// sort snapshots according to their creation date increasingly\n\tsorter := SnapshotSorter{\n\t\tSnapshots: &matchedSnapshots,\n\t}\n\tsort.Sort(&sorter)\n\n\treturn matchedSnapshots, nil\n}", "func (f *FileSnapshotStore) List() ([]*SnapshotMeta, error) {\n\t// Get the eligible snapshots\n\tsnapshots, err := f.getSnapshots()\n\tif err != nil {\n\t\tf.logger.Printf(\"[ERR] snapshot: Failed to get snapshots: %v\", err)\n\t\treturn nil, err\n\t}\n\n\tvar snapMeta []*SnapshotMeta\n\tfor _, meta := range snapshots {\n\t\tsnapMeta = append(snapMeta, &meta.SnapshotMeta)\n\t\tif len(snapMeta) == f.retain {\n\t\t\tbreak\n\t\t}\n\t}\n\treturn snapMeta, nil\n}", "func (r *Repo) GetSnapshots() ([]*Snapshot, error) {\n\tout, err := exec.Command(resticCmd, \"-r\", r.Path, \"-p\",\n\t\tr.Passwordfile, \"snapshots\", \"--json\").Output()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tfmt.Printf(\"%d bytes of output\\n\", len(out))\n\n\tvar snaps []*Snapshot\n\terr = json.Unmarshal(out, &snaps)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn snaps, nil\n}", "func (m *VirtualEndpoint) GetSnapshots()([]CloudPcSnapshotable) {\n val, err := m.GetBackingStore().Get(\"snapshots\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.([]CloudPcSnapshotable)\n }\n return nil\n}", "func (m *MockConfigAdminServiceServer) ListSnapshots(arg0 *ListSnapshotsRequest, arg1 ConfigAdminService_ListSnapshotsServer) error {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"ListSnapshots\", arg0, arg1)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func (c *containerdCAS) ListSnapshots() ([]string, error) {\n\tsnapshotter := ctrdClient.SnapshotService(defaultSnapshotter)\n\tsnapshotIDList := make([]string, 0)\n\tif err := snapshotter.Walk(ctrdCtx, func(i context.Context, info snapshots.Info) error {\n\t\tsnapshotIDList = append(snapshotIDList, info.Name)\n\t\treturn nil\n\t}); err != nil {\n\t\treturn nil, fmt.Errorf(\"ListSnapshots: Execption while fetching snapshot list. %s\", err.Error())\n\t}\n\treturn snapshotIDList, nil\n}", "func NewListSnapshotsRequest(server string) (*http.Request, error) {\n\tvar err error\n\n\tqueryUrl, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbasePath := fmt.Sprintf(\"/snapshot\")\n\tif basePath[0] == '/' {\n\t\tbasePath = basePath[1:]\n\t}\n\n\tqueryUrl, err = queryUrl.Parse(basePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq, err := http.NewRequest(\"GET\", queryUrl.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn req, nil\n}", "func (img *Image) Snapshots() ([]*Snapshot, error) {\n\targs := img.cmdArgs(\"snap\", \"list\")\n\tsnaps := []*snapshotListEntry{}\n\terr := cmdJSON(&snaps, nil, args...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tr := make([]*Snapshot, 0, len(snaps))\n\tfor _, s := range snaps {\n\t\tr = append(r, img.getSnapshot(s.Name))\n\t}\n\treturn r, nil\n}", "func (client *Client) ListSnapshotsWithOptions(request *ListSnapshotsRequest, runtime *util.RuntimeOptions) (_result *ListSnapshotsResponse, _err error) {\n\t_err = util.ValidateModel(request)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\tquery := map[string]interface{}{}\n\tif !tea.BoolValue(util.IsUnset(request.AuthTimeout)) {\n\t\tquery[\"AuthTimeout\"] = request.AuthTimeout\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.PageNo)) {\n\t\tquery[\"PageNo\"] = request.PageNo\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.PageSize)) {\n\t\tquery[\"PageSize\"] = request.PageSize\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.SnapshotType)) {\n\t\tquery[\"SnapshotType\"] = request.SnapshotType\n\t}\n\n\tif !tea.BoolValue(util.IsUnset(request.VideoId)) {\n\t\tquery[\"VideoId\"] = request.VideoId\n\t}\n\n\treq := &openapi.OpenApiRequest{\n\t\tQuery: openapiutil.Query(query),\n\t}\n\tparams := &openapi.Params{\n\t\tAction: tea.String(\"ListSnapshots\"),\n\t\tVersion: tea.String(\"2017-03-21\"),\n\t\tProtocol: tea.String(\"HTTPS\"),\n\t\tPathname: tea.String(\"/\"),\n\t\tMethod: tea.String(\"POST\"),\n\t\tAuthType: tea.String(\"AK\"),\n\t\tStyle: tea.String(\"RPC\"),\n\t\tReqBodyType: tea.String(\"formData\"),\n\t\tBodyType: tea.String(\"json\"),\n\t}\n\t_result = &ListSnapshotsResponse{}\n\t_body, _err := client.CallApi(params, req, runtime)\n\tif _err != nil {\n\t\treturn _result, _err\n\t}\n\t_err = tea.Convert(_body, &_result)\n\treturn _result, _err\n}", "func (s *GCSSnapStore) List() (SnapList, error) {\n\t// recursively list all \"files\", not directory\n\n\tit := s.client.Bucket(s.bucket).Objects(s.ctx, &storage.Query{Prefix: s.prefix})\n\n\tvar attrs []*storage.ObjectAttrs\n\tfor {\n\t\tattr, err := it.Next()\n\t\tif err == iterator.Done {\n\t\t\tbreak\n\t\t}\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tattrs = append(attrs, attr)\n\t}\n\n\tvar snapList SnapList\n\tfor _, v := range attrs {\n\t\tname := strings.Replace(v.Name, s.prefix+\"/\", \"\", 1)\n\t\t//name := v.Name[len(s.prefix):]\n\t\tsnap, err := ParseSnapshot(name)\n\t\tif err != nil {\n\t\t\t// Warning\n\t\t\tfmt.Printf(\"Invalid snapshot found. Ignoring it:%s\\n\", name)\n\t\t} else {\n\t\t\tsnapList = append(snapList, snap)\n\t\t}\n\t}\n\n\tsort.Sort(snapList)\n\treturn snapList, nil\n}", "func cephRBDVolumeListSnapshots(clusterName string, poolName string,\n\tvolumeName string, volumeType string,\n\tuserName string) ([]string, error) {\n\tmsg, err := shared.RunCommand(\n\t\t\"rbd\",\n\t\t\"--id\", userName,\n\t\t\"--format\", \"json\",\n\t\t\"--cluster\", clusterName,\n\t\t\"--pool\", poolName,\n\t\t\"snap\",\n\t\t\"ls\", fmt.Sprintf(\"%s_%s\", volumeType, volumeName))\n\tif err != nil {\n\t\treturn []string{}, err\n\t}\n\n\tvar data []map[string]interface{}\n\terr = json.Unmarshal([]byte(msg), &data)\n\tif err != nil {\n\t\treturn []string{}, err\n\t}\n\n\tsnapshots := []string{}\n\tfor _, v := range data {\n\t\t_, ok := v[\"name\"]\n\t\tif !ok {\n\t\t\treturn []string{}, fmt.Errorf(\"No \\\"name\\\" property found\")\n\t\t}\n\n\t\tname, ok := v[\"name\"].(string)\n\t\tif !ok {\n\t\t\treturn []string{}, fmt.Errorf(\"\\\"name\\\" property did not have string type\")\n\t\t}\n\n\t\tname = strings.TrimSpace(name)\n\t\tsnapshots = append(snapshots, name)\n\t}\n\n\tif len(snapshots) == 0 {\n\t\treturn []string{}, db.ErrNoSuchObject\n\t}\n\n\treturn snapshots, nil\n}", "func (api *snapshotrestoreAPI) List(ctx context.Context, opts *api.ListWatchOptions) ([]*SnapshotRestore, error) {\n\tvar objlist []*SnapshotRestore\n\tobjs, err := api.ct.List(\"SnapshotRestore\", ctx, opts)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfor _, obj := range objs {\n\t\tswitch tp := obj.(type) {\n\t\tcase *SnapshotRestore:\n\t\t\teobj := obj.(*SnapshotRestore)\n\t\t\tobjlist = append(objlist, eobj)\n\t\tdefault:\n\t\t\tlog.Fatalf(\"Got invalid object type %v while looking for SnapshotRestore\", tp)\n\t\t}\n\t}\n\n\treturn objlist, nil\n}", "func (m *Nitro) GetSnapshots() []*Snapshot {\n\tvar snaps []*Snapshot\n\tbuf := m.snapshots.MakeBuf()\n\tdefer m.snapshots.FreeBuf(buf)\n\titer := m.snapshots.NewIterator(CompareSnapshot, buf)\n\titer.SeekFirst()\n\tfor ; iter.Valid(); iter.Next() {\n\t\tsnaps = append(snaps, (*Snapshot)(iter.Get()))\n\t}\n\n\treturn snaps\n}", "func (c *snapshotRules) List(ctx context.Context, opts v1.ListOptions) (result *v1alpha1.SnapshotRuleList, err error) {\n\tvar timeout time.Duration\n\tif opts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*opts.TimeoutSeconds) * time.Second\n\t}\n\tresult = &v1alpha1.SnapshotRuleList{}\n\terr = c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"snapshotrules\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tDo(ctx).\n\t\tInto(result)\n\treturn\n}", "func (mdb *memdbSlice) GetSnapshots() ([]SnapshotInfo, error) {\n\tvar infos []SnapshotInfo\n\tvar err error\n\n\tinfos, _, err = mdb.getSnapshots()\n\treturn infos, err\n}", "func (c *RedshiftServerless) ListSnapshotsWithContext(ctx aws.Context, input *ListSnapshotsInput, opts ...request.Option) (*ListSnapshotsOutput, error) {\n\treq, out := c.ListSnapshotsRequest(input)\n\treq.SetContext(ctx)\n\treq.ApplyOptions(opts...)\n\treturn out, req.Send()\n}", "func (s *LocalSnapStore) List() (brtypes.SnapList, error) {\n\tprefixTokens := strings.Split(s.prefix, \"/\")\n\t// Last element of the tokens is backup version\n\t// Consider the parent of the backup version level (Required for Backward Compatibility)\n\tprefix := path.Join(strings.Join(prefixTokens[:len(prefixTokens)-1], \"/\"))\n\n\tsnapList := brtypes.SnapList{}\n\terr := filepath.Walk(prefix, func(path string, info os.FileInfo, err error) error {\n\t\tif err != nil {\n\t\t\tfmt.Printf(\"prevent panic by handling failure accessing a path %q: %v\\n\", path, err)\n\t\t\treturn err\n\t\t}\n\t\tif info.IsDir() {\n\t\t\treturn nil\n\t\t}\n\t\tif strings.Contains(path, backupVersionV1) || strings.Contains(path, backupVersionV2) {\n\t\t\tsnap, err := ParseSnapshot(path)\n\t\t\tif err != nil {\n\t\t\t\t// Warning\n\t\t\t\tlogrus.Warnf(\"Invalid snapshot found. Ignoring it:%s\\n\", path)\n\t\t\t} else {\n\t\t\t\tsnapList = append(snapList, snap)\n\t\t\t}\n\t\t}\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error walking the path %q: %v\", prefix, err)\n\t}\n\n\tsort.Sort(snapList)\n\treturn snapList, nil\n}", "func (client *Client) ShowSnapshots(names ...string) (*Response, *ResponseStatus, error) {\n\tif len(names) == 0 {\n\t\treturn client.FormattedRequest(\"/show/snapshots\")\n\t}\n\treturn client.FormattedRequest(\"/show/snapshots/%q\", strings.Join(names, \",\"))\n}", "func List(params Params) (*models.RepositoryConfigs, error) {\n\tif err := params.Validate(); err != nil {\n\t\treturn nil, err\n\t}\n\n\trepo, err := params.V1API.PlatformConfigurationSnapshots.GetSnapshotRepositories(\n\t\tplatform_configuration_snapshots.NewGetSnapshotRepositoriesParams(),\n\t\tparams.AuthWriter,\n\t)\n\tif err != nil {\n\t\treturn nil, api.UnwrapError(err)\n\t}\n\n\treturn repo.Payload, nil\n}", "func (c *RedshiftServerless) ListSnapshotsPagesWithContext(ctx aws.Context, input *ListSnapshotsInput, fn func(*ListSnapshotsOutput, bool) bool, opts ...request.Option) error {\n\tp := request.Pagination{\n\t\tNewRequest: func() (*request.Request, error) {\n\t\t\tvar inCpy *ListSnapshotsInput\n\t\t\tif input != nil {\n\t\t\t\ttmp := *input\n\t\t\t\tinCpy = &tmp\n\t\t\t}\n\t\t\treq, _ := c.ListSnapshotsRequest(inCpy)\n\t\t\treq.SetContext(ctx)\n\t\t\treq.ApplyOptions(opts...)\n\t\t\treturn req, nil\n\t\t},\n\t}\n\n\tfor p.Next() {\n\t\tif !fn(p.Page().(*ListSnapshotsOutput), !p.HasNextPage()) {\n\t\t\tbreak\n\t\t}\n\t}\n\n\treturn p.Err()\n}", "func ListSnapshotManifests(ctx context.Context, rep repo.Repository, src *SourceInfo, tags map[string]string) ([]manifest.ID, error) {\n\tlabels := map[string]string{\n\t\ttypeKey: ManifestType,\n\t}\n\n\tif src != nil {\n\t\tlabels = sourceInfoToLabels(*src)\n\t}\n\n\tfor key, value := range tags {\n\t\tlabels[key] = value\n\t}\n\n\tentries, err := rep.FindManifests(ctx, labels)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"unable to find snapshot manifests\")\n\t}\n\n\treturn entryIDs(entries), nil\n}", "func (client *AWSClient) GetSnapshots(ctx context.Context, imageID string) ([]string, error) {\n\tresult, err := client.svcEC2.DescribeImagesWithContext(ctx, &ec2.DescribeImagesInput{\n\t\tImageIds: []*string{aws.String(imageID)},\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar snapshots []string\n\tfor _, b := range result.Images[0].BlockDeviceMappings {\n\t\tif b.Ebs == nil {\n\t\t\tcontinue\n\t\t}\n\t\tsnapshots = append(snapshots, *b.Ebs.SnapshotId)\n\t}\n\n\treturn snapshots, nil\n}", "func (m *MockConfigAdminServiceClient) ListSnapshots(ctx context.Context, in *ListSnapshotsRequest, opts ...grpc.CallOption) (ConfigAdminService_ListSnapshotsClient, error) {\n\tm.ctrl.T.Helper()\n\tvarargs := []interface{}{ctx, in}\n\tfor _, a := range opts {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"ListSnapshots\", varargs...)\n\tret0, _ := ret[0].(ConfigAdminService_ListSnapshotsClient)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (d *lvm) VolumeSnapshots(vol Volume, op *operations.Operation) ([]string, error) {\n\t// We use the volume list rather than inspecting the logical volumes themselves because the origin\n\t// property of an LVM snapshot can be removed/changed when restoring snapshots, such that they are no\n\t// marked as origin of the parent volume. Instead we use prefix matching on the volume names to find the\n\t// snapshot volumes.\n\tcmd := exec.Command(\"lvs\", \"--noheadings\", \"-o\", \"lv_name\", d.config[\"lvm.vg_name\"])\n\tstdout, err := cmd.StdoutPipe()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tstderr, err := cmd.StderrPipe()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = cmd.Start()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tsnapshots := []string{}\n\tscanner := bufio.NewScanner(stdout)\n\tfor scanner.Scan() {\n\t\tsnapName := d.parseLogicalVolumeSnapshot(vol, strings.TrimSpace(scanner.Text()))\n\t\tif snapName == \"\" {\n\t\t\tcontinue // Skip logical volumes that are not recognised as a snapshot of our parent vol.\n\t\t}\n\n\t\tsnapshots = append(snapshots, snapName)\n\t}\n\n\terrMsg, err := io.ReadAll(stderr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = cmd.Wait()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to get snapshot list for volume %q: %v: %w\", vol.name, strings.TrimSpace(string(errMsg)), err)\n\t}\n\n\treturn snapshots, nil\n}", "func (f *FileSnapshotStore) getSnapshots() ([]*fileSnapshotMeta, error) {\n\t// Get the eligible snapshots\n\tsnapshots, err := os.ReadDir(f.path)\n\tif err != nil {\n\t\tf.logger.Printf(\"[ERR] snapshot: Failed to scan snapshot dir: %v\", err)\n\t\treturn nil, err\n\t}\n\n\t// Populate the metadata\n\tvar snapMeta []*fileSnapshotMeta\n\tfor _, snap := range snapshots {\n\t\t// Ignore any files\n\t\tif !snap.IsDir() {\n\t\t\tcontinue\n\t\t}\n\n\t\t// Ignore any temporary snapshots\n\t\tdirName := snap.Name()\n\t\tif strings.HasSuffix(dirName, tmpSuffix) {\n\t\t\tf.logger.Printf(\"[WARN] snapshot: Found temporary snapshot: %v\", dirName)\n\t\t\tcontinue\n\t\t}\n\n\t\t// Try to read the meta data\n\t\tmeta, err := f.readMeta(dirName)\n\t\tif err != nil {\n\t\t\tf.logger.Printf(\"[WARN] snapshot: Failed to read metadata for %v: %v\", dirName, err)\n\t\t\tcontinue\n\t\t}\n\n\t\t// Append, but only return up to the retain count\n\t\tsnapMeta = append(snapMeta, meta)\n\t}\n\n\t// Sort the snapshot, reverse so we get new -> old\n\tsort.Sort(sort.Reverse(snapMetaSlice(snapMeta)))\n\n\treturn snapMeta, nil\n}", "func (r *ProjectsLocationsVolumesSnapshotsService) List(parent string) *ProjectsLocationsVolumesSnapshotsListCall {\n\tc := &ProjectsLocationsVolumesSnapshotsListCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.parent = parent\n\treturn c\n}", "func (api *configurationsnapshotAPI) List(ctx context.Context, opts *api.ListWatchOptions) ([]*ConfigurationSnapshot, error) {\n\tvar objlist []*ConfigurationSnapshot\n\tobjs, err := api.ct.List(\"ConfigurationSnapshot\", ctx, opts)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfor _, obj := range objs {\n\t\tswitch tp := obj.(type) {\n\t\tcase *ConfigurationSnapshot:\n\t\t\teobj := obj.(*ConfigurationSnapshot)\n\t\t\tobjlist = append(objlist, eobj)\n\t\tdefault:\n\t\t\tlog.Fatalf(\"Got invalid object type %v while looking for ConfigurationSnapshot\", tp)\n\t\t}\n\t}\n\n\treturn objlist, nil\n}", "func (api *snapshotrestoreAPI) ApisrvList(ctx context.Context, opts *api.ListWatchOptions) ([]*cluster.SnapshotRestore, error) {\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn nil, err\n\t\t}\n\n\t\treturn apicl.ClusterV1().SnapshotRestore().List(context.Background(), opts)\n\t}\n\n\t// List from local cache\n\tctkitObjs, err := api.List(ctx, opts)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar ret []*cluster.SnapshotRestore\n\tfor _, obj := range ctkitObjs {\n\t\tret = append(ret, &obj.SnapshotRestore)\n\t}\n\treturn ret, nil\n}", "func (l *Libvirt) DomainListAllSnapshots(Dom Domain, NeedResults int32, Flags uint32) (rSnapshots []DomainSnapshot, rRet int32, err error) {\n\tvar buf []byte\n\n\targs := DomainListAllSnapshotsArgs {\n\t\tDom: Dom,\n\t\tNeedResults: NeedResults,\n\t\tFlags: Flags,\n\t}\n\n\tbuf, err = encode(&args)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tvar r response\n\tr, err = l.requestStream(274, constants.Program, buf, nil, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\t// Return value unmarshaling\n\ttpd := typedParamDecoder{}\n\tct := map[string]xdr.TypeDecoder{\"libvirt.TypedParam\": tpd}\n\trdr := bytes.NewReader(r.Payload)\n\tdec := xdr.NewDecoderCustomTypes(rdr, 0, ct)\n\t// Snapshots: []DomainSnapshot\n\t_, err = dec.Decode(&rSnapshots)\n\tif err != nil {\n\t\treturn\n\t}\n\t// Ret: int32\n\t_, err = dec.Decode(&rRet)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func (s *SnapshotStatusService) Snapshot(snapshots ...string) *SnapshotStatusService {\n\ts.snapshot = append(s.snapshot, snapshots...)\n\treturn s\n}", "func (c *Client) CloudProjectSnapshotsList(projectID, region string) ([]types.CloudImage, error) {\n\tvar path string\n\tif region == \"\" {\n\t\tpath = queryEscape(\"/cloud/project/%s/snapshot\", projectID)\n\n\t} else {\n\t\tpath = queryEscape(\"/cloud/project/%s/snapshot?region=%s\", projectID, region)\n\t}\n\timages := []types.CloudImage{}\n\treturn images, c.Get(path, &images)\n}", "func QuerySnapshotsUtil(ctx context.Context, m cnsvolume.Manager, snapshotQueryFilter cnstypes.CnsSnapshotQueryFilter,\n\tmaxEntries int64) ([]cnstypes.CnsSnapshotQueryResultEntry, string, error) {\n\tlog := logger.GetLogger(ctx)\n\tvar allQuerySnapshotResults []cnstypes.CnsSnapshotQueryResultEntry\n\tvar snapshotQuerySpec cnstypes.CnsSnapshotQuerySpec\n\tvar batchSize int64\n\tmaxIteration := int64(1)\n\tisMaxIterationSet := false\n\tif snapshotQueryFilter.SnapshotQuerySpecs == nil {\n\t\tlog.Infof(\"Attempting to retrieve all the Snapshots available in the vCenter inventory.\")\n\t} else {\n\t\tsnapshotQuerySpec = snapshotQueryFilter.SnapshotQuerySpecs[0]\n\t\tlog.Infof(\"Invoking QuerySnapshots with spec: %+v\", snapshotQuerySpec)\n\t}\n\t// Check if cursor is specified, if not set a default cursor.\n\tif snapshotQueryFilter.Cursor == nil {\n\t\t// Setting the default limit(128) explicitly.\n\t\tsnapshotQueryFilter = cnstypes.CnsSnapshotQueryFilter{\n\t\t\tCursor: &cnstypes.CnsCursor{\n\t\t\t\tOffset: 0,\n\t\t\t\tLimit: DefaultQuerySnapshotLimit,\n\t\t\t},\n\t\t}\n\t\tbatchSize = DefaultQuerySnapshotLimit\n\t} else {\n\t\tbatchSize = snapshotQueryFilter.Cursor.Limit\n\t}\n\titeration := int64(1)\n\tfor {\n\t\tif iteration > maxIteration {\n\t\t\t// Exceeds the max number of results that can be handled by callers.\n\t\t\tnextToken := strconv.FormatInt(snapshotQueryFilter.Cursor.Offset, 10)\n\t\t\tlog.Infof(\"the number of results: %d approached max-entries: %d for \"+\n\t\t\t\t\"limit: %d in iteration: %d, returning with next-token: %s\",\n\t\t\t\tlen(allQuerySnapshotResults), maxEntries, batchSize, iteration, nextToken)\n\t\t\treturn allQuerySnapshotResults, nextToken, nil\n\t\t}\n\t\tlog.Infof(\"invoking QuerySnapshots in iteration: %d with offset: %d and limit: %d, current total \"+\n\t\t\t\"results: %d\", iteration, snapshotQueryFilter.Cursor.Offset, snapshotQueryFilter.Cursor.Limit,\n\t\t\tlen(allQuerySnapshotResults))\n\t\tsnapshotQueryResult, err := m.QuerySnapshots(ctx, snapshotQueryFilter)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"querySnapshots failed for snapshotQueryFilter: %v. Err=%+v\", snapshotQueryFilter, err)\n\t\t\treturn nil, \"\", err\n\t\t}\n\t\tif snapshotQueryResult == nil {\n\t\t\tlog.Infof(\"Observed empty SnapshotQueryResult\")\n\t\t\tbreak\n\t\t}\n\t\tif len(snapshotQueryResult.Entries) == 0 {\n\t\t\tlog.Infof(\"QuerySnapshots retrieved no results for the spec: %+v\", snapshotQuerySpec)\n\t\t}\n\t\t// Update the max iteration.\n\t\t// isMaxIterationSet ensures that the max iterations are set only once, this is to ensure that the number of\n\t\t// results are lower than the max entries supported by caller in a busy system which has increasing number\n\t\t// total records.\n\t\tif !isMaxIterationSet {\n\t\t\tif snapshotQueryResult.Cursor.TotalRecords < maxEntries {\n\t\t\t\t// If the total number of records is less than max entries supported by caller then\n\t\t\t\t// all results can be retrieved in a loop, when the results are returned no next-token is expected to be set.\n\t\t\t\t// Example:\n\t\t\t\t// maxEntries=200, totalRecords=150, batchSize=128\n\t\t\t\t// maxIteration=2\n\t\t\t\t// iteration-1: 128 results, iteration-2: 22 results\n\t\t\t\t// total results returned: 150\n\t\t\t\t// offset=0\n\t\t\t\tmaxRecords := snapshotQueryResult.Cursor.TotalRecords\n\t\t\t\tnumOfIterationsForAllResults := float64(maxRecords) / float64(batchSize)\n\t\t\t\tmaxIteration = int64(math.Ceil(numOfIterationsForAllResults))\n\t\t\t\tlog.Infof(\"setting max iteration to %d for total records count: %d\", maxIteration, maxRecords)\n\t\t\t} else {\n\t\t\t\t// All results cannot be returned to caller, in this case the expectation is return as many results with a\n\t\t\t\t// nextToken.\n\t\t\t\t// Example:\n\t\t\t\t// maxEntries=150, totalRecords=200, batchSize=128\n\t\t\t\t// maxIteration=1\n\t\t\t\t// iteration-1: 128 results\n\t\t\t\t// total results returned: 128\n\t\t\t\t// offset= 1, callers are expected to call with new offset as next token.\n\t\t\t\tmaxRecords := maxEntries\n\t\t\t\tnumOfIterationsForAllResults := float64(maxRecords) / float64(batchSize)\n\t\t\t\tmaxIteration = int64(math.Floor(numOfIterationsForAllResults))\n\t\t\t\tlog.Infof(\"setting max iteration to %d for total records count: %d and max limit: %d\",\n\t\t\t\t\tmaxIteration, snapshotQueryResult.Cursor.TotalRecords, maxRecords)\n\t\t\t}\n\t\t\tisMaxIterationSet = true\n\t\t}\n\n\t\tallQuerySnapshotResults = append(allQuerySnapshotResults, snapshotQueryResult.Entries...)\n\t\tlog.Infof(\"%d more snapshots to be queried\",\n\t\t\tsnapshotQueryResult.Cursor.TotalRecords-snapshotQueryResult.Cursor.Offset)\n\t\tif snapshotQueryResult.Cursor.Offset == snapshotQueryResult.Cursor.TotalRecords {\n\t\t\tlog.Infof(\"QuerySnapshots retrieved all records (%d) for the SnapshotQuerySpec: %+v in %d iterations\",\n\t\t\t\tsnapshotQueryResult.Cursor.TotalRecords, snapshotQuerySpec, iteration)\n\t\t\tbreak\n\t\t}\n\t\titeration++\n\t\tsnapshotQueryFilter.Cursor = &snapshotQueryResult.Cursor\n\t}\n\treturn allQuerySnapshotResults, \"\", nil\n}", "func (a *HyperflexApiService) GetHyperflexHealthCheckExecutionSnapshotList(ctx context.Context) ApiGetHyperflexHealthCheckExecutionSnapshotListRequest {\n\treturn ApiGetHyperflexHealthCheckExecutionSnapshotListRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t}\n}", "func SnapshotListNext(name string, cursor uint64, props interface{}) (string, uint64, DMUObjectSetStats, error) {\n\tcmd := &Cmd{\n\t\tCookie: cursor,\n\t}\n\tif err := NvlistIoctl(zfsHandle.Fd(), ZFS_IOC_SNAPSHOT_LIST_NEXT, name, cmd, nil, props, nil); err != nil {\n\t\treturn \"\", 0, DMUObjectSetStats{}, err\n\t}\n\treturn delimitedBufToString(cmd.Name[:]), cmd.Cookie, cmd.Objset_stats, nil\n}", "func (c *ClientWithResponses) ListSnapshotsWithResponse(ctx context.Context) (*ListSnapshotsResponse, error) {\n\trsp, err := c.ListSnapshots(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn ParseListSnapshotsResponse(rsp)\n}", "func (client *WebAppsClient) listSnapshotsCreateRequest(ctx context.Context, resourceGroupName string, name string, options *WebAppsListSnapshotsOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Web/sites/{name}/snapshots\"\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif name == \"\" {\n\t\treturn nil, errors.New(\"parameter name cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{name}\", url.PathEscape(name))\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-02-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func ExampleRDS_DescribeDBSnapshots_shared00() {\n\tsvc := rds.New(session.New())\n\tinput := &rds.DescribeDBSnapshotsInput{\n\t\tDBSnapshotIdentifier: aws.String(\"mydbsnapshot\"),\n\t}\n\n\tresult, err := svc.DescribeDBSnapshots(input)\n\tif err != nil {\n\t\tif aerr, ok := err.(awserr.Error); ok {\n\t\t\tswitch aerr.Code() {\n\t\t\tcase rds.ErrCodeDBSnapshotNotFoundFault:\n\t\t\t\tfmt.Println(rds.ErrCodeDBSnapshotNotFoundFault, aerr.Error())\n\t\t\tdefault:\n\t\t\t\tfmt.Println(aerr.Error())\n\t\t\t}\n\t\t} else {\n\t\t\t// Print the error, cast err to awserr.Error to get the Code and\n\t\t\t// Message from an error.\n\t\t\tfmt.Println(err.Error())\n\t\t}\n\t\treturn\n\t}\n\n\tfmt.Println(result)\n}", "func (c *Client) ListSnapshotRollOwners(ctx context.Context, id BlockID, cycle, index int64) (*SnapshotOwners, error) {\n\towners := &SnapshotOwners{Cycle: cycle, Index: index}\n\tu := fmt.Sprintf(\"chains/main/blocks/%s/context/raw/json/rolls/owner/snapshot/%d/%d?depth=1\", id, cycle, index)\n\tif err := c.Get(ctx, u, &owners.Rolls); err != nil {\n\t\treturn nil, err\n\t}\n\treturn owners, nil\n}", "func SortSnapshots(snapshots []SnapshotId) {\n\tsort.Sort(snapshotIdSlice(snapshots))\n}", "func ParseListSnapshotsResponse(rsp *http.Response) (*ListSnapshotsResponse, error) {\n\tbodyBytes, err := ioutil.ReadAll(rsp.Body)\n\tdefer rsp.Body.Close()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresponse := &ListSnapshotsResponse{\n\t\tBody: bodyBytes,\n\t\tHTTPResponse: rsp,\n\t}\n\n\tswitch {\n\tcase strings.Contains(rsp.Header.Get(\"Content-Type\"), \"json\") && rsp.StatusCode == 200:\n\t\tvar dest struct {\n\t\t\tSnapshots *[]Snapshot `json:\"snapshots,omitempty\"`\n\t\t}\n\t\tif err := json.Unmarshal(bodyBytes, &dest); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tresponse.JSON200 = &dest\n\n\t}\n\n\treturn response, nil\n}", "func (client *WebAppsClient) listSnapshotsHandleResponse(resp *http.Response) (WebAppsListSnapshotsResponse, error) {\n\tresult := WebAppsListSnapshotsResponse{RawResponse: resp}\n\tif err := runtime.UnmarshalAsJSON(resp, &result.SnapshotCollection); err != nil {\n\t\treturn WebAppsListSnapshotsResponse{}, runtime.NewResponseError(err, resp)\n\t}\n\treturn result, nil\n}", "func FindFilteredSnapshots(ctx context.Context, be restic.Lister, loader restic.LoaderUnpacked, hosts []string, tags []restic.TagList, paths []string, snapshotIDs []string) <-chan *restic.Snapshot {\n\tout := make(chan *restic.Snapshot)\n\tgo func() {\n\t\tdefer close(out)\n\t\tbe, err := backend.MemorizeList(ctx, be, restic.SnapshotFile)\n\t\tif err != nil {\n\t\t\tWarnf(\"could not load snapshots: %v\\n\", err)\n\t\t\treturn\n\t\t}\n\n\t\terr = restic.FindFilteredSnapshots(ctx, be, loader, hosts, tags, paths, snapshotIDs, func(id string, sn *restic.Snapshot, err error) error {\n\t\t\tif err != nil {\n\t\t\t\tWarnf(\"Ignoring %q: %v\\n\", id, err)\n\t\t\t} else {\n\t\t\t\tselect {\n\t\t\t\tcase <-ctx.Done():\n\t\t\t\t\treturn ctx.Err()\n\t\t\t\tcase out <- sn:\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\t\tif err != nil {\n\t\t\tWarnf(\"could not load snapshots: %v\\n\", err)\n\t\t}\n\t}()\n\treturn out\n}", "func (m *VirtualEndpoint) SetSnapshots(value []CloudPcSnapshotable)() {\n err := m.GetBackingStore().Set(\"snapshots\", value)\n if err != nil {\n panic(err)\n }\n}", "func (c *ReplicaClient) Snapshots(ctx context.Context, generation string) (_ litestream.SnapshotIterator, err error) {\n\tdefer func() { c.resetOnConnError(err) }()\n\n\tsftpClient, err := c.Init(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdir, err := litestream.SnapshotsPath(c.Path, generation)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"cannot determine snapshots path: %w\", err)\n\t}\n\n\tfis, err := sftpClient.ReadDir(dir)\n\tif os.IsNotExist(err) {\n\t\treturn litestream.NewSnapshotInfoSliceIterator(nil), nil\n\t} else if err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Iterate over every file and convert to metadata.\n\tinfos := make([]litestream.SnapshotInfo, 0, len(fis))\n\tfor _, fi := range fis {\n\t\t// Parse index from filename.\n\t\tindex, err := litestream.ParseSnapshotPath(path.Base(fi.Name()))\n\t\tif err != nil {\n\t\t\tcontinue\n\t\t}\n\n\t\tinfos = append(infos, litestream.SnapshotInfo{\n\t\t\tGeneration: generation,\n\t\t\tIndex: index,\n\t\t\tSize: fi.Size(),\n\t\t\tCreatedAt: fi.ModTime().UTC(),\n\t\t})\n\t}\n\n\tsort.Sort(litestream.SnapshotInfoSlice(infos))\n\n\treturn litestream.NewSnapshotInfoSliceIterator(infos), nil\n}", "func getNewSnapshotRequests(config types.AppInstanceConfig, status *types.AppInstanceStatus) (snapRequests []types.SnapshotDesc) {\n\tif config.Snapshot.Snapshots != nil {\n\t\tfor _, snap := range config.Snapshot.Snapshots {\n\t\t\tif isNewSnapshotRequest(snap.SnapshotID, status) {\n\t\t\t\tlog.Noticef(\"A new snapshot %s is requested\", snap.SnapshotID)\n\t\t\t\tsnapRequests = append(snapRequests, snap)\n\t\t\t}\n\t\t}\n\t}\n\treturn snapRequests\n}", "func (driver *Driver) GetSnapshot(volumeID, snapshotID, snapshotName string) ([]*storagedriver.Snapshot, error) {\n\tvar snapshotsInt []*storagedriver.Snapshot\n\tif volumeID != \"\" {\n\t\tvolumes, err := driver.getVolume(volumeID, \"\")\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tfor _, volume := range volumes {\n\t\t\tfor _, destSnap := range volume.DestSnapList {\n\t\t\t\tsnapshot, err := driver.getSnapshot(strconv.Itoa(int(destSnap.([]interface{})[2].(float64))), \"\")\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn nil, err\n\t\t\t\t}\n\n\t\t\t\tvolSize, _ := strconv.Atoi(volume.VolSize)\n\t\t\t\tsnapshotSD := &storagedriver.Snapshot{\n\t\t\t\t\tName: snapshot[0].Name,\n\t\t\t\t\tVolumeID: strconv.Itoa(volume.Index),\n\t\t\t\t\tSnapshotID: strconv.Itoa(snapshot[0].Index),\n\t\t\t\t\tVolumeSize: strconv.Itoa(volSize / 1024 / 1024),\n\t\t\t\t\tStartTime: snapshot[0].CreationTime,\n\t\t\t\t\tDescription: \"\",\n\t\t\t\t\tStatus: \"\",\n\t\t\t\t}\n\t\t\t\tsnapshotsInt = append(snapshotsInt, snapshotSD)\n\t\t\t}\n\t\t}\n\t} else {\n\t\tsnapshots, err := driver.getSnapshot(snapshotID, snapshotName)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tfor _, snapshot := range snapshots {\n\t\t\tsnapshot, err := goxtremio.GetSnapshot(strconv.Itoa(snapshot.Index), \"\")\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\n\t\t\tvolume, err := driver.getVolume(strconv.Itoa(int(snapshot.AncestorVolID[2].(float64))), \"\")\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\n\t\t\tvolSize, _ := strconv.Atoi(volume[0].VolSize)\n\t\t\tsnapshotSD := &storagedriver.Snapshot{\n\t\t\t\tName: snapshot.Name,\n\t\t\t\tVolumeID: strconv.Itoa(int(snapshot.AncestorVolID[2].(float64))),\n\t\t\t\tSnapshotID: strconv.Itoa(snapshot.Index),\n\t\t\t\tVolumeSize: strconv.Itoa(volSize / 1024 / 1024),\n\t\t\t\tStartTime: snapshot.CreationTime,\n\t\t\t\tDescription: \"\",\n\t\t\t\tStatus: \"\",\n\t\t\t}\n\t\t\tsnapshotsInt = append(snapshotsInt, snapshotSD)\n\t\t}\n\n\t}\n\n\treturn snapshotsInt, nil\n}", "func (a *Client) PostVMSnapshotsList(params *PostVMSnapshotsListParams, authInfo runtime.ClientAuthInfoWriter) (*PostVMSnapshotsListOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewPostVMSnapshotsListParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"PostVMSnapshotsList\",\n\t\tMethod: \"POST\",\n\t\tPathPattern: \"/vm_snapshots/list\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &PostVMSnapshotsListReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*PostVMSnapshotsListOK), nil\n\n}", "func (c *volumeSnapshotSchedules) List(ctx context.Context, opts v1.ListOptions) (result *v1alpha1.VolumeSnapshotScheduleList, err error) {\n\tvar timeout time.Duration\n\tif opts.TimeoutSeconds != nil {\n\t\ttimeout = time.Duration(*opts.TimeoutSeconds) * time.Second\n\t}\n\tresult = &v1alpha1.VolumeSnapshotScheduleList{}\n\terr = c.client.Get().\n\t\tNamespace(c.ns).\n\t\tResource(\"volumesnapshotschedules\").\n\t\tVersionedParams(&opts, scheme.ParameterCodec).\n\t\tTimeout(timeout).\n\t\tDo(ctx).\n\t\tInto(result)\n\treturn\n}", "func (a *HyperflexApiService) GetHyperflexVmSnapshotInfoList(ctx context.Context) ApiGetHyperflexVmSnapshotInfoListRequest {\n\treturn ApiGetHyperflexVmSnapshotInfoListRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t}\n}", "func Snapshot() []Stats {\n\treturn defaultActions.Snapshot()\n}", "func findSnapshots(dir string) ([]uint64, error) {\n\tmatches, err := filepath.Glob(filepath.Join(dir, \"*.meta\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar snaps []uint64\n\tfor _, m := range matches {\n\t\tm = filepath.Base(m)\n\t\tm = strings.TrimSuffix(m, \".meta\")\n\t\ti, err := strconv.ParseUint(m, 10, 64)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tsnaps = append(snaps, i)\n\t}\n\tsort.Sort(decrUint64Slice(snaps))\n\treturn snaps, nil\n}", "func (a *SnapshotApiService) SnapshotsGet(ctx _context.Context, optionals *SnapshotsGetOpts) (Snapshots, *APIResponse, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodGet\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue Snapshots\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/snapshots\"\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tif optionals != nil && optionals.Pretty != nil {\n\t\tlocalVarQueryParams.Add(\"pretty\", parameterToString(*optionals.Pretty, \"\"))\n\t}\n\tif optionals != nil && optionals.Depth != nil {\n\t\tlocalVarQueryParams.Add(\"depth\", parameterToString(*optionals.Depth, \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tif optionals != nil && optionals.XContractNumber != nil {\n\t\tlocalVarHeaderParams[\"X-Contract-Number\"] = parameterToString(*optionals.XContractNumber, \"\")\n\t}\n\tif ctx != nil {\n\t\t// API Key Authentication\n\t\tif auth, ok := ctx.Value(ContextAPIKey).(APIKey); ok {\n\t\t\tvar key string\n\t\t\tif auth.Prefix != \"\" {\n\t\t\t\tkey = auth.Prefix + \" \" + auth.Key\n\t\t\t} else {\n\t\t\t\tkey = auth.Key\n\t\t\t}\n\t\t\tlocalVarHeaderParams[\"Authorization\"] = key\n\t\t}\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tlocalVarAPIResponse := &APIResponse {\n\t\tResponse: localVarHTTPResponse,\n\t\tMethod: localVarHTTPMethod,\n\t\tRequestURL: localVarPath,\n\t\tOperation: \"SnapshotsGet\",\n\t}\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarAPIResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tlocalVarAPIResponse.Payload = localVarBody\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarAPIResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarAPIResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\treturn localVarReturnValue, localVarAPIResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarAPIResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarAPIResponse, nil\n}", "func (c *Cache) RefreshSnapshots(s Server, p *Progress) error {\n\tdefer p.Done()\n\n\t// list cache entries\n\tentries, err := c.List(backend.Snapshot)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// list snapshots first\n\tdone := make(chan struct{})\n\tdefer close(done)\n\n\t// check that snapshot blobs are cached\n\tfor name := range s.List(backend.Snapshot, done) {\n\t\tid, err := backend.ParseID(name)\n\t\tif err != nil {\n\t\t\tcontinue\n\t\t}\n\n\t\t// remove snapshot from list of entries\n\t\tfor i, e := range entries {\n\t\t\tif e.ID.Equal(id) {\n\t\t\t\tentries = append(entries[:i], entries[i+1:]...)\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\n\t\thas, err := c.Has(backend.Snapshot, \"blobs\", id)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif has {\n\t\t\tcontinue\n\t\t}\n\n\t\t// else start progress reporting\n\t\tp.Start()\n\n\t\t// build new cache\n\t\t_, err = cacheSnapshotBlobs(p, s, c, id)\n\t\tif err != nil {\n\t\t\tdebug.Log(\"Cache.RefreshSnapshots\", \"unable to cache snapshot blobs for %v: %v\", id.Str(), err)\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// remove other entries\n\tfor _, e := range entries {\n\t\tdebug.Log(\"Cache.RefreshSnapshots\", \"remove entry %v\", e)\n\t\terr = c.Purge(backend.Snapshot, e.Subtype, e.ID)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (api *configurationsnapshotAPI) ApisrvList(ctx context.Context, opts *api.ListWatchOptions) ([]*cluster.ConfigurationSnapshot, error) {\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn nil, err\n\t\t}\n\n\t\treturn apicl.ClusterV1().ConfigurationSnapshot().List(context.Background(), opts)\n\t}\n\n\t// List from local cache\n\tctkitObjs, err := api.List(ctx, opts)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar ret []*cluster.ConfigurationSnapshot\n\tfor _, obj := range ctkitObjs {\n\t\tret = append(ret, &obj.ConfigurationSnapshot)\n\t}\n\treturn ret, nil\n}", "func (z *ZfsH) Snapshots(d *Dataset, depth int) ([]*Dataset, error) {\n\treturn z.SnapshotsByName(d.Name, depth)\n}", "func (os *OpenStack) FindSnapshot(tags map[string]string) ([]string, []string, error) {\n\tvar snapshotIDs, statuses []string\n\tss, err := os.snapshotService()\n\tif err != nil || ss == nil {\n\t\tglog.Errorf(\"Unable to initialize cinder client for region: %s\", os.region)\n\t\treturn snapshotIDs, statuses, fmt.Errorf(\"Failed to find snapshot by tags %v: %v\", tags, err)\n\t}\n\n\topts := SnapshotListOpts{}\n\tsnapshots, err := ss.listSnapshots(opts)\n\n\tif err != nil {\n\t\tglog.Errorf(\"Failed to list snapshots. Error: %v\", err)\n\t\treturn snapshotIDs, statuses, err\n\t}\n\tglog.Infof(\"Listed [%v] snapshots.\", len(snapshots))\n\n\tglog.Infof(\"Looking for matching tags [%#v] in snapshots.\", tags)\n\t// Loop around to find the snapshot with the matching input metadata\n\t// NOTE(xyang): Metadata based filtering for snapshots is supported by Cinder volume API\n\t// microversion 3.21 and above. Currently the OpenStack Cloud Provider only supports V2.0.\n\t// Revisit this later when V3.0 is supported.\n\tfor _, snapshot := range snapshots {\n\t\tglog.Infof(\"Looking for matching tags in snapshot [%#v].\", snapshot)\n\t\tnamespaceVal, ok := snapshot.Metadata[ctrlsnap.CloudSnapshotCreatedForVolumeSnapshotNamespaceTag]\n\t\tif ok {\n\t\t\tif tags[ctrlsnap.CloudSnapshotCreatedForVolumeSnapshotNamespaceTag] == namespaceVal {\n\t\t\t\tnameVal, ok := snapshot.Metadata[ctrlsnap.CloudSnapshotCreatedForVolumeSnapshotNameTag]\n\t\t\t\tif ok {\n\t\t\t\t\tif tags[ctrlsnap.CloudSnapshotCreatedForVolumeSnapshotNameTag] == nameVal {\n\t\t\t\t\t\tuidVal, ok := snapshot.Metadata[ctrlsnap.CloudSnapshotCreatedForVolumeSnapshotUIDTag]\n\t\t\t\t\t\tif ok {\n\t\t\t\t\t\t\tif tags[ctrlsnap.CloudSnapshotCreatedForVolumeSnapshotUIDTag] == uidVal {\n\t\t\t\t\t\t\t\ttimeVal, ok := snapshot.Metadata[ctrlsnap.CloudSnapshotCreatedForVolumeSnapshotTimestampTag]\n\t\t\t\t\t\t\t\tif ok {\n\t\t\t\t\t\t\t\t\tif tags[ctrlsnap.CloudSnapshotCreatedForVolumeSnapshotTimestampTag] == timeVal {\n\t\t\t\t\t\t\t\t\t\tsnapshotIDs = append(snapshotIDs, snapshot.ID)\n\t\t\t\t\t\t\t\t\t\tstatuses = append(statuses, snapshot.Status)\n\t\t\t\t\t\t\t\t\t\tglog.Infof(\"Add snapshot [%#v].\", snapshot)\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn snapshotIDs, statuses, nil\n}", "func (h *Handler) ListLogs(ctx context.Context, req *proto.ListLogRequest, stream proto.LogRecorder_ListLogsStream) error {\n\n\tq := req.GetQuery()\n\tp := req.GetPage()\n\ts := req.GetSize()\n\n\tr, err := h.Repo.ListLogs(q, p, s)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor rr := range r {\n\n\t\tstream.Send(&proto.ListLogResponse{\n\t\t\tLogMessage: rr.LogMessage,\n\t\t})\n\t}\n\treturn nil\n}", "func Snapshot(names []string, pool string, props *DatasetProps) error {\n\tvar snapReq struct {\n\t\tSnaps map[string]bool `nvlist:\"snaps\"`\n\t\tProps *DatasetProps `nvlist:\"props\"`\n\t}\n\tsnapReq.Snaps = make(map[string]bool)\n\tfor _, name := range names {\n\t\tif _, ok := snapReq.Snaps[name]; ok {\n\t\t\treturn errors.New(\"duplicate snapshot name\")\n\t\t}\n\t\tsnapReq.Snaps[name] = true\n\t}\n\tsnapReq.Props = props\n\tcmd := &Cmd{}\n\tsnapRes := make(map[string]int32)\n\treturn NvlistIoctl(zfsHandle.Fd(), ZFS_IOC_SNAPSHOT, pool, cmd, snapReq, snapRes, nil)\n\t// TODO: Maybe there is an error in snapRes\n}", "func (s *actions) Snapshot() []Stats {\n\tvar i int\n\ts.lock.RLock()\n\tdefer s.lock.RUnlock()\n\n\n\tm := make([]Stats, len(s.metrics))\n\n\tfor k, v := range s.metrics {\n\t\tm[i] = v.stats(k)\n\t\ti++\n\t}\n\n\treturn m\n}", "func (s *SnapshotsServiceOp) ListVolume(opt *ListOptions) ([]Snapshot, *Response, error) {\n\tlistOpt := listSnapshotOptions{ResourceType: \"volume\"}\n\treturn s.list(opt, &listOpt)\n}", "func (s *S3) snapshotRetention(ctx context.Context) error {\n\tif s.config.EtcdSnapshotRetention < 1 {\n\t\treturn nil\n\t}\n\tlogrus.Infof(\"Applying snapshot retention policy to snapshots stored in S3: retention: %d, snapshotPrefix: %s\", s.config.EtcdSnapshotRetention, s.snapshotPrefix())\n\n\tvar snapshotFiles []minio.ObjectInfo\n\n\ttoCtx, cancel := context.WithTimeout(ctx, s.config.EtcdS3Timeout)\n\tdefer cancel()\n\n\tloo := minio.ListObjectsOptions{\n\t\tRecursive: true,\n\t\tPrefix: s.snapshotPrefix(),\n\t}\n\tfor info := range s.client.ListObjects(toCtx, s.config.EtcdS3BucketName, loo) {\n\t\tif info.Err != nil {\n\t\t\treturn info.Err\n\t\t}\n\t\tsnapshotFiles = append(snapshotFiles, info)\n\t}\n\n\tif len(snapshotFiles) <= s.config.EtcdSnapshotRetention {\n\t\treturn nil\n\t}\n\n\tsort.Slice(snapshotFiles, func(i, j int) bool {\n\t\treturn snapshotFiles[i].Key < snapshotFiles[j].Key\n\t})\n\n\tdelCount := len(snapshotFiles) - s.config.EtcdSnapshotRetention\n\tfor _, df := range snapshotFiles[:delCount] {\n\t\tlogrus.Infof(\"Removing S3 snapshot: %s\", df.Key)\n\t\tif err := s.client.RemoveObject(ctx, s.config.EtcdS3BucketName, df.Key, minio.RemoveObjectOptions{}); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (s *SnapshotsServiceOp) ListVolume(ctx context.Context, opt *ListOptions) ([]Snapshot, *Response, error) {\n\tlistOpt := listSnapshotOptions{ResourceType: \"volume\"}\n\treturn s.list(ctx, opt, &listOpt)\n}", "func (c *Client) ListSnapshotBlocks(ctx context.Context, params *ListSnapshotBlocksInput, optFns ...func(*Options)) (*ListSnapshotBlocksOutput, error) {\n\tif params == nil {\n\t\tparams = &ListSnapshotBlocksInput{}\n\t}\n\n\tresult, metadata, err := c.invokeOperation(ctx, \"ListSnapshotBlocks\", params, optFns, addOperationListSnapshotBlocksMiddlewares)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tout := result.(*ListSnapshotBlocksOutput)\n\tout.ResultMetadata = metadata\n\treturn out, nil\n}", "func GetEntriesAndNextToken(req *csi.ListSnapshotsRequest, snapshots []compute.Snapshot) (*csi.ListSnapshotsResponse, error) {\n\tif req == nil {\n\t\treturn nil, status.Errorf(codes.Aborted, \"request is nil\")\n\t}\n\n\tvar err error\n\tstart := 0\n\tif req.StartingToken != \"\" {\n\t\tstart, err = strconv.Atoi(req.StartingToken)\n\t\tif err != nil {\n\t\t\treturn nil, status.Errorf(codes.Aborted, \"ListSnapshots starting token(%s) parsing with error: %v\", req.StartingToken, err)\n\n\t\t}\n\t\tif start >= len(snapshots) {\n\t\t\treturn nil, status.Errorf(codes.Aborted, \"ListSnapshots starting token(%d) is greater than total number of snapshots\", start)\n\t\t}\n\t\tif start < 0 {\n\t\t\treturn nil, status.Errorf(codes.Aborted, \"ListSnapshots starting token(%d) can not be negative\", start)\n\t\t}\n\t}\n\n\tmaxEntries := len(snapshots) - start\n\tif req.MaxEntries > 0 && int(req.MaxEntries) < maxEntries {\n\t\tmaxEntries = int(req.MaxEntries)\n\t}\n\tentries := []*csi.ListSnapshotsResponse_Entry{}\n\tfor count := 0; start < len(snapshots) && count < maxEntries; start++ {\n\t\tif (req.SourceVolumeId != \"\" && req.SourceVolumeId == GetSourceVolumeID(&snapshots[start])) || req.SourceVolumeId == \"\" {\n\t\t\tcsiSnapshot, err := GenerateCSISnapshot(req.SourceVolumeId, &snapshots[start])\n\t\t\tif err != nil {\n\t\t\t\treturn nil, fmt.Errorf(\"failed to generate snapshot entry: %v\", err)\n\t\t\t}\n\t\t\tentries = append(entries, &csi.ListSnapshotsResponse_Entry{Snapshot: csiSnapshot})\n\t\t\tcount++\n\t\t}\n\t}\n\n\tnextToken := len(snapshots)\n\tif start < len(snapshots) {\n\t\tnextToken = start\n\t}\n\n\tlistSnapshotResp := &csi.ListSnapshotsResponse{\n\t\tEntries: entries,\n\t\tNextToken: strconv.Itoa(nextToken),\n\t}\n\n\treturn listSnapshotResp, nil\n}", "func (s *store) ListEvents(startRevision int64) ([]types.Record, error) {\n\tvalidRevision := storerecord.RevToString(startRevision)\n\n\tbookKeeper, recordsMaker := s.getDefaultBookingKeepingFuncs(true)\n\tf := filterutils.NewFilter()\n\tf.And(\n\t\tfilterutils.GreaterThanOrEqual(consts.RevisionFieldName, validRevision),\n\t\tfilterutils.ExcludeRows(),\n\t\tfilterutils.CombineOr(\n\t\t\t/* this should not return any un-needed data since we also have the revision predicate */\n\t\t\tfilterutils.IncludeDataRowsForAny(),\n\t\t\tfilterutils.IncludeEvents(),\n\t\t),\n\t)\n\to := &storage.QueryOptions{\n\t\tFilter: f.Generate(),\n\t}\n\n\trecords, err := s.execQuery(o, storage.MinimalMetadata, bookKeeper, recordsMaker)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn records, nil\n}", "func (l *Libvirt) DomainSnapshotListAllChildren(Snapshot DomainSnapshot, NeedResults int32, Flags uint32) (rSnapshots []DomainSnapshot, rRet int32, err error) {\n\tvar buf []byte\n\n\targs := DomainSnapshotListAllChildrenArgs {\n\t\tSnapshot: Snapshot,\n\t\tNeedResults: NeedResults,\n\t\tFlags: Flags,\n\t}\n\n\tbuf, err = encode(&args)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tvar r response\n\tr, err = l.requestStream(275, constants.Program, buf, nil, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\t// Return value unmarshaling\n\ttpd := typedParamDecoder{}\n\tct := map[string]xdr.TypeDecoder{\"libvirt.TypedParam\": tpd}\n\trdr := bytes.NewReader(r.Payload)\n\tdec := xdr.NewDecoderCustomTypes(rdr, 0, ct)\n\t// Snapshots: []DomainSnapshot\n\t_, err = dec.Decode(&rSnapshots)\n\tif err != nil {\n\t\treturn\n\t}\n\t// Ret: int32\n\t_, err = dec.Decode(&rRet)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func sortAllSnapshotBackups(backupsList []*v1alpha1.VolumeBackup) []*v1alpha1.VolumeBackup {\n\tvar ascBackupList = make([]*v1alpha1.VolumeBackup, 0)\n\tascBackupList = append(ascBackupList, backupsList...)\n\n\tsort.Slice(ascBackupList, func(i, j int) bool {\n\t\treturn ascBackupList[i].CreationTimestamp.Unix() < ascBackupList[j].CreationTimestamp.Unix()\n\t})\n\treturn ascBackupList\n}", "func cephRBDSnapshotListClones(clusterName string, poolName string,\n\tvolumeName string, volumeType string,\n\tsnapshotName string, userName string) ([]string, error) {\n\tmsg, err := shared.RunCommand(\n\t\t\"rbd\",\n\t\t\"--id\", userName,\n\t\t\"--cluster\", clusterName,\n\t\t\"--pool\", poolName,\n\t\t\"children\",\n\t\t\"--image\", fmt.Sprintf(\"%s_%s\", volumeType, volumeName),\n\t\t\"--snap\", snapshotName)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tmsg = strings.TrimSpace(msg)\n\tclones := strings.Fields(msg)\n\tif len(clones) == 0 {\n\t\treturn nil, db.ErrNoSuchObject\n\t}\n\n\treturn clones, nil\n}", "func (m *MockServersService) Snapshots(arg0 int) (bl.Images, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"Snapshots\", arg0)\n\tret0, _ := ret[0].(bl.Images)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (a *RepoAPI) ls(params interface{}) (resp *rpc.Response) {\n\tm := objx.New(cast.ToStringMap(params))\n\tvar revision []string\n\tif rev := m.Get(\"revision\").Str(); rev != \"\" {\n\t\trevision = []string{rev}\n\t}\n\treturn rpc.Success(util.Map{\n\t\t\"entries\": a.mods.Repo.ListPath(m.Get(\"name\").Str(), m.Get(\"path\").Str(), revision...),\n\t})\n}", "func getJSON(snapList []*Snapshots) []byte {\n\tlist := ListJSON{List: make([]Snapshots, len(snapList))}\n\tfor i := range snapList {\n\t\tlist.List[i] = Snapshots{\n\t\t\tName: snapList[i].Name,\n\t\t\tTimes: snapList[i].Times,\n\t\t}\n\t}\n\tdata, _ := json.Marshal(list)\n\treturn append(data, '\\n')\n}" ]
[ "0.794119", "0.79234934", "0.78526795", "0.7817018", "0.78123695", "0.77119243", "0.7678372", "0.76712507", "0.7653685", "0.7574974", "0.75453335", "0.74940014", "0.74746305", "0.7452373", "0.7448181", "0.74213994", "0.7383719", "0.7317239", "0.72878295", "0.72797275", "0.72063315", "0.7139405", "0.70496863", "0.6964892", "0.69051397", "0.6894407", "0.68734986", "0.6867995", "0.68406165", "0.6821393", "0.65208375", "0.6505439", "0.647523", "0.63518983", "0.6342149", "0.63233733", "0.63228816", "0.6290277", "0.6235168", "0.6171471", "0.61109596", "0.6077015", "0.59599197", "0.5947636", "0.5939607", "0.59066075", "0.58980095", "0.5894694", "0.5881708", "0.58793133", "0.5879043", "0.58672523", "0.5858037", "0.585684", "0.5744986", "0.5740225", "0.573535", "0.5734796", "0.5725357", "0.5721825", "0.57145554", "0.56751066", "0.56634253", "0.5634272", "0.5622717", "0.5619996", "0.56079906", "0.55699927", "0.55577296", "0.55420643", "0.5538639", "0.5528814", "0.5481016", "0.5460003", "0.5450536", "0.5443742", "0.5440257", "0.5418601", "0.54178756", "0.5407365", "0.5395631", "0.53858316", "0.5380944", "0.5367874", "0.5363827", "0.5363674", "0.5359307", "0.5356767", "0.5318803", "0.53143513", "0.5312614", "0.5310979", "0.52754784", "0.5237095", "0.5221601", "0.52084714", "0.5203074", "0.5200687", "0.5200653", "0.5182485" ]
0.76051927
9
DeleteSnapshot uses the override method DeleteSnapshotFn or the real implementation.
func (c *TestClient) DeleteSnapshot(project, name string) error { if c.DeleteSnapshotFn != nil { return c.DeleteSnapshotFn(project, name) } return c.client.DeleteSnapshot(project, name) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (driver *Driver) DeleteSnapshot(ctx context.Context, req *csi.DeleteSnapshotRequest) (*csi.DeleteSnapshotResponse, error) {\n\treturn nil, status.Error(codes.Unimplemented, \"\")\n}", "func (d *Driver) DeleteSnapshot(ctx context.Context, req *csi.DeleteSnapshotRequest) (*csi.DeleteSnapshotResponse, error) {\n\treturn nil, status.Error(codes.Unimplemented, \"\")\n}", "func (cs *controller) DeleteSnapshot(\n\tctx context.Context,\n\treq *csi.DeleteSnapshotRequest,\n) (*csi.DeleteSnapshotResponse, error) {\n\n\treturn nil, status.Error(codes.Unimplemented, \"\")\n}", "func (l *LogDB) DeleteSnapshot(shardID uint64,\n\treplicaID uint64, index uint64) error {\n\tpanic(\"depreciated\")\n}", "func (s *API) DeleteSnapshot(req *DeleteSnapshotRequest, opts ...scw.RequestOption) (*Snapshot, error) {\n\tvar err error\n\n\tif req.Region == \"\" {\n\t\tdefaultRegion, _ := s.client.GetDefaultRegion()\n\t\treq.Region = defaultRegion\n\t}\n\n\tif fmt.Sprint(req.Region) == \"\" {\n\t\treturn nil, errors.New(\"field Region cannot be empty in request\")\n\t}\n\n\tif fmt.Sprint(req.SnapshotID) == \"\" {\n\t\treturn nil, errors.New(\"field SnapshotID cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"DELETE\",\n\t\tPath: \"/rdb/v1/regions/\" + fmt.Sprint(req.Region) + \"/snapshots/\" + fmt.Sprint(req.SnapshotID) + \"\",\n\t\tHeaders: http.Header{},\n\t}\n\n\tvar resp Snapshot\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func DeleteSnapshot(d Ploop, uuid string) error {\n\tcuuid := C.CString(uuid)\n\tdefer cfree(cuuid)\n\n\tret := C.ploop_delete_snapshot(d.d, cuuid)\n\n\treturn mkerr(ret)\n}", "func (ks *KopiaSnapshotter) DeleteSnapshot(snapID string, opts map[string]string) error {\n\treturn ks.snap.DeleteSnapshot(snapID)\n}", "func (os *OpenStack) DeleteSnapshot(snapshotID string) error {\n\tsnapshots, err := os.snapshotService()\n\tif err != nil || snapshots == nil {\n\t\tglog.Errorf(\"Unable to initialize cinder client for region: %s\", os.region)\n\t\treturn err\n\t}\n\n\terr = snapshots.deleteSnapshot(snapshotID)\n\tif err != nil {\n\t\tglog.Errorf(\"Cannot delete snapshot %s: %v\", snapshotID, err)\n\t}\n\treturn nil\n}", "func (c *restClient) DeleteSnapshot(ctx context.Context, req *netapppb.DeleteSnapshotRequest, opts ...gax.CallOption) (*DeleteSnapshotOperation, error) {\n\tbaseUrl, err := url.Parse(c.endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbaseUrl.Path += fmt.Sprintf(\"/v1/%v\", req.GetName())\n\n\tparams := url.Values{}\n\tparams.Add(\"$alt\", \"json;enum-encoding=int\")\n\n\tbaseUrl.RawQuery = params.Encode()\n\n\t// Build HTTP headers from client and context metadata.\n\thds := []string{\"x-goog-request-params\", fmt.Sprintf(\"%s=%v\", \"name\", url.QueryEscape(req.GetName()))}\n\n\thds = append(c.xGoogHeaders, hds...)\n\thds = append(hds, \"Content-Type\", \"application/json\")\n\theaders := gax.BuildHeaders(ctx, hds...)\n\tunm := protojson.UnmarshalOptions{AllowPartial: true, DiscardUnknown: true}\n\tresp := &longrunningpb.Operation{}\n\te := gax.Invoke(ctx, func(ctx context.Context, settings gax.CallSettings) error {\n\t\tif settings.Path != \"\" {\n\t\t\tbaseUrl.Path = settings.Path\n\t\t}\n\t\thttpReq, err := http.NewRequest(\"DELETE\", baseUrl.String(), nil)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thttpReq = httpReq.WithContext(ctx)\n\t\thttpReq.Header = headers\n\n\t\thttpRsp, err := c.httpClient.Do(httpReq)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer httpRsp.Body.Close()\n\n\t\tif err = googleapi.CheckResponse(httpRsp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tbuf, err := io.ReadAll(httpRsp.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := unm.Unmarshal(buf, resp); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}, opts...)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\n\toverride := fmt.Sprintf(\"/v1/%s\", resp.GetName())\n\treturn &DeleteSnapshotOperation{\n\t\tlro: longrunning.InternalNewOperation(*c.LROClient, resp),\n\t\tpollPath: override,\n\t}, nil\n}", "func DeleteSnapshot(storageID, snapshotID string) error {\n\tclient, err := NewExtPacketClient()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, e := client.Storages.DeleteSnapshot(storageID, snapshotID)\n\treturn e\n}", "func (a *API) DeleteSnapshot(id string, force bool) error {\n\trequest := ecs.CreateDeleteSnapshotRequest()\n\trequest.Scheme = \"https\"\n\trequest.SnapshotId = id\n\trequest.Force = requests.NewBoolean(force)\n\t_, err := a.ecs.DeleteSnapshot(request)\n\treturn err\n}", "func (cs *controllerServer) DeleteSnapshot(ctx context.Context, req *csi.DeleteSnapshotRequest) (*csi.DeleteSnapshotResponse, error) {\n\tlog.Infof(\"Starting Delete Snapshot %s with response: %v\", req.SnapshotId, req)\n\t// Step 1: check req\n\t// snapshotName is name of snapshot lv\n\tsnapshotName := req.GetSnapshotId()\n\tif len(snapshotName) == 0 {\n\t\tlog.Error(\"DeleteSnapshot: Snapshot ID not provided\")\n\t\treturn nil, status.Error(codes.InvalidArgument, \"DeleteSnapshot: Snapshot ID not provided\")\n\t}\n\n\t// Step 2: get volumeID from snapshot\n\tsnapContent, err := getVolumeSnapshotContent(cs.snapclient, snapshotName)\n\tif err != nil {\n\t\tlog.Errorf(\"DeleteSnapshot: get snapContent %s error: %s\", snapshotName, err.Error())\n\t\treturn nil, status.Errorf(codes.Internal, \"DeleteSnapshot: get snapContent %s error: %s\", snapshotName, err.Error())\n\t}\n\tvolumeID := *snapContent.Spec.Source.VolumeHandle\n\n\t// Step 3: get nodeName and vgName\n\tnodeName, vgName, _, err := getPvSpec(cs.client, volumeID, cs.driverName)\n\tif err != nil {\n\t\tlog.Errorf(\"DeleteSnapshot: get pv %s error: %s\", volumeID, err.Error())\n\t\treturn nil, status.Errorf(codes.Internal, \"DeleteSnapshot: get pv %s error: %s\", volumeID, err.Error())\n\t}\n\tlog.Infof(\"DeleteSnapshot: snapshot %s is in %s, whose vg is %s\", snapshotName, nodeName, vgName)\n\n\t// Step 4: get grpc client\n\tconn, err := cs.getNodeConn(nodeName)\n\tif err != nil {\n\t\tlog.Errorf(\"DeleteSnapshot: get grpc client at node %s error: %s\", nodeName, err.Error())\n\t\treturn nil, status.Errorf(codes.Internal, \"DeleteSnapshot: get grpc client at node %s error: %s\", nodeName, err.Error())\n\t}\n\tdefer conn.Close()\n\n\t// Step 5: delete lvm snapshot\n\tvar lvmName string\n\tif lvmName, err = conn.GetLvm(ctx, vgName, snapshotName); err != nil {\n\t\tlog.Errorf(\"DeleteSnapshot: get lvm snapshot %s failed: %s\", snapshotName, err.Error())\n\t\treturn nil, status.Errorf(codes.Internal, \"DeleteSnapshot: get lvm snapshot %s failed: %s\", snapshotName, err.Error())\n\t}\n\tif lvmName != \"\" {\n\t\terr := conn.DeleteSnapshot(ctx, vgName, snapshotName)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"DeleteSnapshot: delete lvm snapshot %s failed: %s\", snapshotName, err.Error())\n\t\t\treturn nil, status.Errorf(codes.Internal, \"DeleteSnapshot: delete lvm snapshot %s failed: %s\", snapshotName, err.Error())\n\t\t}\n\t} else {\n\t\tlog.Infof(\"DeleteSnapshot: lvm snapshot %s in node %s not found, skip...\", snapshotName, nodeName)\n\t\t// return immediately\n\t\treturn &csi.DeleteSnapshotResponse{}, nil\n\t}\n\n\tlog.Infof(\"DeleteSnapshot: delete snapshot %s successfully\", snapshotName)\n\treturn &csi.DeleteSnapshotResponse{}, nil\n}", "func (cs *DefaultControllerServer) DeleteSnapshot(ctx context.Context, req *csi.DeleteSnapshotRequest) (*csi.DeleteSnapshotResponse, error) {\n\treturn nil, status.Error(codes.Unimplemented, \"\")\n}", "func (client *Client) DeleteSnapshot(names ...string) (*Response, *ResponseStatus, error) {\n\treturn client.FormattedRequest(\"/delete/snapshot/%q\", strings.Join(names, \",\"))\n}", "func Delete(deleteSnapshotURL string, snapshotName string) error {\n\tlogger.Infof(\"Deleting snapshot %s\", snapshotName)\n\tformData := url.Values{\n\t\t\"snapshot\": {snapshotName},\n\t}\n\n\tu, err := url.Parse(deleteSnapshotURL)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tresp, err := http.PostForm(u.String(), formData)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tbody, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tsnap := snapshot{}\n\terr = json.Unmarshal(body, &snap)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif snap.Status == \"ok\" {\n\t\tlogger.Infof(\"Snapshot %s deleted\", snapshotName)\n\t\treturn nil\n\t} else if snap.Status == \"error\" {\n\t\treturn errors.New(snap.Msg)\n\t} else {\n\t\treturn fmt.Errorf(\"Unkown status: %v\", snap.Status)\n\t}\n}", "func (c *Client) DeleteSnapshot(name string, namespace string) error {\n\tif err := c.initClient(); err != nil {\n\t\treturn err\n\t}\n\treturn c.client.VolumeSnapshots(namespace).Delete(context.TODO(), name, metav1.DeleteOptions{})\n}", "func DeleteSnapshot(snapname string) (err error) {\n\terr = snapbuilder.NewKubeclient().WithNamespace(OpenEBSNamespace).Delete(snapname)\n\tif err == nil {\n\t\tklog.Infof(\"deprovisioned snapshot %s\", snapname)\n\t}\n\n\treturn\n}", "func (mw *ShardedRDB) DeleteSnapshot(clusterID uint64,\n\tnodeID uint64, snapshotIndex uint64) error {\n\tidx := mw.partitioner.GetPartitionID(clusterID)\n\treturn mw.shards[idx].deleteSnapshot(clusterID, nodeID, snapshotIndex)\n}", "func (c *ClientManager) DeleteSnapshot(snapshotId int) error {\n\t_, err := c.StorageService.Id(snapshotId).DeleteObject()\n\treturn err\n}", "func (c *ReplicaClient) DeleteSnapshot(ctx context.Context, generation string, index int) (err error) {\n\tdefer func() { c.resetOnConnError(err) }()\n\n\tsftpClient, err := c.Init(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfilename, err := litestream.SnapshotPath(c.Path, generation, index)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"cannot determine snapshot path: %w\", err)\n\t}\n\n\tif err := sftpClient.Remove(filename); err != nil && !os.IsNotExist(err) {\n\t\treturn fmt.Errorf(\"cannot delete snapshot %q: %w\", filename, err)\n\t}\n\n\tinternal.OperationTotalCounterVec.WithLabelValues(ReplicaClientType, \"DELETE\").Inc()\n\treturn nil\n}", "func cephRBDSnapshotDelete(clusterName string, poolName string,\n\tvolumeName string, volumeType string, snapshotName string,\n\tuserName string) error {\n\t_, err := shared.RunCommand(\n\t\t\"rbd\",\n\t\t\"--id\", userName,\n\t\t\"--cluster\", clusterName,\n\t\t\"--pool\", poolName,\n\t\t\"snap\",\n\t\t\"rm\",\n\t\tfmt.Sprintf(\"%s_%s@%s\", volumeType, volumeName, snapshotName))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (p *Plugin) DeleteSnapshot(snapshotID string) error {\n\treturn p.plugin.DeleteSnapshot(snapshotID)\n}", "func (s *DataStore) DeleteSnapshot(snapshotName string) error {\n\treturn s.lhClient.LonghornV1beta2().Snapshots(s.namespace).Delete(context.TODO(), snapshotName, metav1.DeleteOptions{})\n}", "func (s *OsdCsiServer) DeleteSnapshot(\n\tctx context.Context,\n\treq *csi.DeleteSnapshotRequest,\n) (*csi.DeleteSnapshotResponse, error) {\n\n\tif len(req.GetSnapshotId()) == 0 {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"Snapshot id must be provided\")\n\t}\n\n\t// Get grpc connection\n\tconn, err := s.getConn()\n\tif err != nil {\n\t\treturn nil, status.Errorf(\n\t\t\tcodes.Unavailable,\n\t\t\t\"Unable to connect to SDK server: %v\", err)\n\t}\n\n\t// Get secret if any was passed\n\tctx = s.setupContext(ctx, req.GetSecrets())\n\tctx, cancel := grpcutil.WithDefaultTimeout(ctx)\n\tdefer cancel()\n\n\t// Check ID is valid with the specified volume capabilities\n\tvolumes := api.NewOpenStorageVolumeClient(conn)\n\n\t_, err = volumes.Delete(ctx, &api.SdkVolumeDeleteRequest{\n\t\tVolumeId: req.GetSnapshotId(),\n\t})\n\tif err != nil {\n\t\treturn nil, status.Errorf(codes.Aborted, \"Unable to delete snapshot %s: %v\",\n\t\t\treq.GetSnapshotId(),\n\t\t\terr)\n\t}\n\n\treturn &csi.DeleteSnapshotResponse{}, nil\n}", "func deleteSnapshot(file string) error {\n\tif file != \"\" {\n\t\treturn os.Remove(file)\n\t}\n\treturn nil\n}", "func (ctrler CtrlDefReactor) OnSnapshotRestoreDelete(obj *SnapshotRestore) error {\n\tlog.Info(\"OnSnapshotRestoreDelete is not implemented\")\n\treturn nil\n}", "func RunSnapshotDelete(c *CmdConfig) error {\n\tif len(c.Args) == 0 {\n\t\treturn doctl.NewMissingArgsErr(c.NS)\n\t}\n\n\tforce, err := c.Doit.GetBool(c.NS, doctl.ArgForce)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tss := c.Snapshots()\n\tids := c.Args\n\n\tif force || AskForConfirmDelete(\"snapshot\", len(ids)) == nil {\n\t\tfor _, id := range ids {\n\t\t\terr := ss.Delete(id)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t} else {\n\t\treturn errOperationAborted\n\t}\n\treturn nil\n}", "func ExampleRDS_DeleteDBSnapshot_shared00() {\n\tsvc := rds.New(session.New())\n\tinput := &rds.DeleteDBSnapshotInput{\n\t\tDBSnapshotIdentifier: aws.String(\"mydbsnapshot\"),\n\t}\n\n\tresult, err := svc.DeleteDBSnapshot(input)\n\tif err != nil {\n\t\tif aerr, ok := err.(awserr.Error); ok {\n\t\t\tswitch aerr.Code() {\n\t\t\tcase rds.ErrCodeInvalidDBSnapshotStateFault:\n\t\t\t\tfmt.Println(rds.ErrCodeInvalidDBSnapshotStateFault, aerr.Error())\n\t\t\tcase rds.ErrCodeDBSnapshotNotFoundFault:\n\t\t\t\tfmt.Println(rds.ErrCodeDBSnapshotNotFoundFault, aerr.Error())\n\t\t\tdefault:\n\t\t\t\tfmt.Println(aerr.Error())\n\t\t\t}\n\t\t} else {\n\t\t\t// Print the error, cast err to awserr.Error to get the Code and\n\t\t\t// Message from an error.\n\t\t\tfmt.Println(err.Error())\n\t\t}\n\t\treturn\n\t}\n\n\tfmt.Println(result)\n}", "func (mr *MockCleanerMockRecorder) DeleteSnapshot(arg0, arg1, arg2, arg3 interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"DeleteSnapshot\", reflect.TypeOf((*MockCleaner)(nil).DeleteSnapshot), arg0, arg1, arg2, arg3)\n}", "func (mr *MockClientMockRecorder) DeleteSnapshot(arg0 interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"DeleteSnapshot\", reflect.TypeOf((*MockClient)(nil).DeleteSnapshot), arg0)\n}", "func (s *OsdCsiServer) DeleteSnapshot(\n\tctx context.Context,\n\treq *csi.DeleteSnapshotRequest,\n) (*csi.DeleteSnapshotResponse, error) {\n\n\tif len(req.GetSnapshotId()) == 0 {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"Snapshot id must be provided\")\n\t}\n\n\t// If the snapshot is not found, then we can return OK\n\tvolumes, err := s.driver.Inspect([]string{req.GetSnapshotId()})\n\tif (err == nil && len(volumes) == 0) ||\n\t\t(err != nil && err == kvdb.ErrNotFound) {\n\t\treturn &csi.DeleteSnapshotResponse{}, nil\n\t} else if err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = s.driver.Delete(context.TODO(), req.GetSnapshotId())\n\tif err != nil {\n\t\treturn nil, status.Errorf(codes.Internal, \"Unable to delete snapshot %s: %v\",\n\t\t\treq.GetSnapshotId(),\n\t\t\terr)\n\t}\n\n\treturn &csi.DeleteSnapshotResponse{}, nil\n}", "func (c *restClient) DeleteSnapshotOperation(name string) *DeleteSnapshotOperation {\n\toverride := fmt.Sprintf(\"/v1/%s\", name)\n\treturn &DeleteSnapshotOperation{\n\t\tlro: longrunning.InternalNewOperation(*c.LROClient, &longrunningpb.Operation{Name: name}),\n\t\tpollPath: override,\n\t}\n}", "func (c *gRPCClient) DeleteSnapshotOperation(name string) *DeleteSnapshotOperation {\n\treturn &DeleteSnapshotOperation{\n\t\tlro: longrunning.InternalNewOperation(*c.LROClient, &longrunningpb.Operation{Name: name}),\n\t}\n}", "func cephContainerSnapshotDelete(clusterName string, poolName string,\n\tvolumeName string, volumeType string, snapshotName string,\n\tuserName string) int {\n\tlogImageEntry := fmt.Sprintf(\"%s/%s_%s\", poolName, volumeType, volumeName)\n\tlogSnapshotEntry := fmt.Sprintf(\"%s/%s_%s@%s\", poolName, volumeType,\n\t\tvolumeName, snapshotName)\n\n\tclones, err := cephRBDSnapshotListClones(clusterName, poolName,\n\t\tvolumeName, volumeType, snapshotName, userName)\n\tif err != nil {\n\t\tif err != db.ErrNoSuchObject {\n\t\t\tlogger.Errorf(`Failed to list clones of RBD snapshot \"%s\" of RBD storage volume \"%s\": %s`, logSnapshotEntry, logImageEntry, err)\n\t\t\treturn -1\n\t\t}\n\t\tlogger.Debugf(`RBD snapshot \"%s\" of RBD storage volume \"%s\" does not have any clones`, logSnapshotEntry, logImageEntry)\n\n\t\t// unprotect\n\t\terr = cephRBDSnapshotUnprotect(clusterName, poolName, volumeName,\n\t\t\tvolumeType, snapshotName, userName)\n\t\tif err != nil {\n\t\t\tlogger.Errorf(`Failed to unprotect RBD snapshot \"%s\" of RBD storage volume \"%s\": %s`, logSnapshotEntry, logImageEntry, err)\n\t\t\treturn -1\n\t\t}\n\t\tlogger.Debugf(`Unprotected RBD snapshot \"%s\" of RBD storage volume \"%s\"`, logSnapshotEntry, logImageEntry)\n\n\t\t// unmap\n\t\terr = cephRBDVolumeSnapshotUnmap(clusterName, poolName,\n\t\t\tvolumeName, volumeType, snapshotName, userName, true)\n\t\tif err != nil {\n\t\t\tlogger.Errorf(`Failed to unmap RBD snapshot \"%s\" of RBD storage volume \"%s\": %s`, logSnapshotEntry, logImageEntry, err)\n\t\t\treturn -1\n\t\t}\n\t\tlogger.Debugf(`Unmapped RBD snapshot \"%s\" of RBD storage volume \"%s\"`, logSnapshotEntry, logImageEntry)\n\n\t\t// delete\n\t\terr = cephRBDSnapshotDelete(clusterName, poolName, volumeName,\n\t\t\tvolumeType, snapshotName, userName)\n\t\tif err != nil {\n\t\t\tlogger.Errorf(`Failed to delete RBD snapshot \"%s\" of RBD storage volume \"%s\": %s`, logSnapshotEntry, logImageEntry, err)\n\t\t\treturn -1\n\t\t}\n\t\tlogger.Debugf(`Deleted RBD snapshot \"%s\" of RBD storage volume \"%s\"`, logSnapshotEntry, logImageEntry)\n\n\t\t// Only delete the parent image if it is a zombie. If it is not\n\t\t// we know that LXD is still using it.\n\t\tif strings.HasPrefix(volumeType, \"zombie_\") {\n\t\t\tret := cephContainerDelete(clusterName, poolName,\n\t\t\t\tvolumeName, volumeType, userName)\n\t\t\tif ret < 0 {\n\t\t\t\tlogger.Errorf(`Failed to delete RBD storage volume \"%s\"`,\n\t\t\t\t\tlogImageEntry)\n\t\t\t\treturn -1\n\t\t\t}\n\t\t\tlogger.Debugf(`Deleted RBD storage volume \"%s\"`, logImageEntry)\n\t\t}\n\n\t\treturn 0\n\t} else {\n\t\tlogger.Debugf(`Detected \"%v\" as clones of RBD snapshot \"%s\" of RBD storage volume \"%s\"`, clones, logSnapshotEntry, logImageEntry)\n\n\t\tcanDelete := true\n\t\tfor _, clone := range clones {\n\t\t\tclonePool, cloneType, cloneName, err := parseClone(clone)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Errorf(`Failed to parse clone \"%s\" of RBD snapshot \"%s\" of RBD storage volume \"%s\"`, clone, logSnapshotEntry, logImageEntry)\n\t\t\t\treturn -1\n\t\t\t}\n\t\t\tlogger.Debugf(`Split clone \"%s\" of RBD snapshot \"%s\" of RBD storage volume \"%s\" into pool name \"%s\", volume type \"%s\", and volume name \"%s\"`, clone, logSnapshotEntry, logImageEntry, clonePool, cloneType, cloneName)\n\n\t\t\tif !strings.HasPrefix(cloneType, \"zombie_\") {\n\t\t\t\tcanDelete = false\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tret := cephContainerDelete(clusterName, clonePool,\n\t\t\t\tcloneName, cloneType, userName)\n\t\t\tif ret < 0 {\n\t\t\t\tlogger.Errorf(`Failed to delete clone \"%s\" of RBD snapshot \"%s\" of RBD storage volume \"%s\"`, clone, logSnapshotEntry, logImageEntry)\n\t\t\t\treturn -1\n\t\t\t} else if ret == 1 {\n\t\t\t\t// Only marked as zombie\n\t\t\t\tcanDelete = false\n\t\t\t}\n\t\t}\n\n\t\tif canDelete {\n\t\t\tlogger.Debugf(`Deleted all clones of RBD snapshot \"%s\" of RBD storage volume \"%s\"`, logSnapshotEntry, logImageEntry)\n\n\t\t\t// unprotect\n\t\t\terr = cephRBDSnapshotUnprotect(clusterName, poolName,\n\t\t\t\tvolumeName, volumeType, snapshotName, userName)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Errorf(`Failed to unprotect RBD snapshot \"%s\" of RBD storage volume \"%s\": %s`, logSnapshotEntry, logImageEntry, err)\n\t\t\t\treturn -1\n\t\t\t}\n\t\t\tlogger.Debugf(`Unprotected RBD snapshot \"%s\" of RBD storage volume \"%s\"`, logSnapshotEntry, logImageEntry)\n\n\t\t\t// unmap\n\t\t\terr = cephRBDVolumeSnapshotUnmap(clusterName, poolName,\n\t\t\t\tvolumeName, volumeType, snapshotName, userName,\n\t\t\t\ttrue)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Errorf(`Failed to unmap RBD snapshot \"%s\" of RBD storage volume \"%s\": %s`, logSnapshotEntry, logImageEntry, err)\n\t\t\t\treturn -1\n\t\t\t}\n\t\t\tlogger.Debugf(`Unmapped RBD snapshot \"%s\" of RBD storage volume \"%s\"`, logSnapshotEntry, logImageEntry)\n\n\t\t\t// delete\n\t\t\terr = cephRBDSnapshotDelete(clusterName, poolName,\n\t\t\t\tvolumeName, volumeType, snapshotName, userName)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Errorf(`Failed to delete RBD snapshot \"%s\" of RBD storage volume \"%s\": %s`, logSnapshotEntry, logImageEntry, err)\n\t\t\t\treturn -1\n\t\t\t}\n\t\t\tlogger.Debugf(`Deleted RBD snapshot \"%s\" of RBD storage volume \"%s\"`, logSnapshotEntry, logImageEntry)\n\n\t\t\t// Only delete the parent image if it is a zombie. If it\n\t\t\t// is not we know that LXD is still using it.\n\t\t\tif strings.HasPrefix(volumeType, \"zombie_\") {\n\t\t\t\tret := cephContainerDelete(clusterName,\n\t\t\t\t\tpoolName, volumeName, volumeType,\n\t\t\t\t\tuserName)\n\t\t\t\tif ret < 0 {\n\t\t\t\t\tlogger.Errorf(`Failed to delete RBD storage volume \"%s\"`, logImageEntry)\n\t\t\t\t\treturn -1\n\t\t\t\t}\n\t\t\t\tlogger.Debugf(`Deleted RBD storage volume \"%s\"`,\n\t\t\t\t\tlogImageEntry)\n\t\t\t}\n\t\t} else {\n\t\t\tlogger.Debugf(`Could not delete all clones of RBD snapshot \"%s\" of RBD storage volume \"%s\"`, logSnapshotEntry, logImageEntry)\n\n\t\t\tif strings.HasPrefix(snapshotName, \"zombie_\") {\n\t\t\t\treturn 1\n\t\t\t}\n\n\t\t\terr := cephRBDVolumeSnapshotUnmap(clusterName, poolName,\n\t\t\t\tvolumeName, volumeType, snapshotName, userName,\n\t\t\t\ttrue)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Errorf(`Failed to unmap RBD snapshot \"%s\" of RBD storage volume \"%s\": %s`, logSnapshotEntry, logImageEntry, err)\n\t\t\t\treturn -1\n\t\t\t}\n\t\t\tlogger.Debug(`Unmapped RBD snapshot \"%s\" of RBD storage volume \"%s\"`, logSnapshotEntry, logImageEntry)\n\n\t\t\tnewSnapshotName := fmt.Sprintf(\"zombie_%s\", snapshotName)\n\t\t\tlogSnapshotNewEntry := fmt.Sprintf(\"%s/%s_%s@%s\",\n\t\t\t\tpoolName, volumeName, volumeType, newSnapshotName)\n\t\t\terr = cephRBDVolumeSnapshotRename(clusterName, poolName,\n\t\t\t\tvolumeName, volumeType, snapshotName,\n\t\t\t\tnewSnapshotName, userName)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Errorf(`Failed to rename RBD snapshot \"%s\" of RBD storage volume \"%s\" to %s`, logSnapshotEntry, logImageEntry, logSnapshotNewEntry)\n\t\t\t\treturn -1\n\t\t\t}\n\t\t\tlogger.Debugf(`Renamed RBD snapshot \"%s\" of RBD storage volume \"%s\" to %s`, logSnapshotEntry, logImageEntry, logSnapshotNewEntry)\n\t\t}\n\n\t}\n\n\treturn 1\n}", "func (m *MockCleaner) DeleteSnapshot(arg0 context.Context, arg1, arg2 string, arg3 *v1.GroupVersionForDiscovery) error {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"DeleteSnapshot\", arg0, arg1, arg2, arg3)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func (c *Cluster) DeleteSnapshot(\n\tsnapshot *vsv1alpha1.VolumeSnapshot,\n\tbucket objectstore.Objectstore,\n\tlocalKubeClient kubernetes.Interface) error {\n\n\treturn deleteVolumeSnapshots(snapshot, bucket, localKubeClient)\n}", "func (d CephDriver) DeleteBlockDeviceSnapshot(volumeUUID string, snapshotID string) error {\n\tvar cmd *exec.Cmd\n\n\tcmd = exec.Command(\"rbd\", \"--id\", d.ID, \"snap\", \"unprotect\", volumeUUID+\"@\"+snapshotID)\n\tout, err := cmd.CombinedOutput()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Error when running: %v: %v: %s\", cmd.Args, err, out)\n\t}\n\n\tcmd = exec.Command(\"rbd\", \"--id\", d.ID, \"snap\", \"rm\", volumeUUID+\"@\"+snapshotID)\n\tout, err = cmd.CombinedOutput()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Error when running: %v: %v: %s\", cmd.Args, err, out)\n\t}\n\treturn nil\n}", "func (l *Libvirt) DomainSnapshotDelete(Snap DomainSnapshot, Flags DomainSnapshotDeleteFlags) (err error) {\n\tvar buf []byte\n\n\targs := DomainSnapshotDeleteArgs {\n\t\tSnap: Snap,\n\t\tFlags: Flags,\n\t}\n\n\tbuf, err = encode(&args)\n\tif err != nil {\n\t\treturn\n\t}\n\n\n\t_, err = l.requestStream(193, constants.Program, buf, nil, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func (m *MockRDSAPI) DeleteDBSnapshot(arg0 *rds.DeleteDBSnapshotInput) (*rds.DeleteDBSnapshotOutput, error) {\n\tret := m.ctrl.Call(m, \"DeleteDBSnapshot\", arg0)\n\tret0, _ := ret[0].(*rds.DeleteDBSnapshotOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (c *Client) DeleteSnapshot(ctx context.Context, req *netapppb.DeleteSnapshotRequest, opts ...gax.CallOption) (*DeleteSnapshotOperation, error) {\n\treturn c.internalClient.DeleteSnapshot(ctx, req, opts...)\n}", "func (c *Client) DeleteSnapshot(dir, name string) error {\n\tallowSnapshotReq := &hdfs.DeleteSnapshotRequestProto{\n\t\tSnapshotRoot: &dir,\n\t\tSnapshotName: &name,\n\t}\n\tallowSnapshotRes := &hdfs.DeleteSnapshotResponseProto{}\n\n\terr := c.namenode.Execute(\"deleteSnapshot\", allowSnapshotReq, allowSnapshotRes)\n\tif err != nil {\n\t\treturn interpretException(err)\n\t}\n\treturn nil\n}", "func (m *MockClient) DeleteSnapshot(arg0 *ec2.DeleteSnapshotInput) (*ec2.DeleteSnapshotOutput, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"DeleteSnapshot\", arg0)\n\tret0, _ := ret[0].(*ec2.DeleteSnapshotOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func deleteSnapshotInfo(c context.Context) error {\n\tc = ds.WithoutTransaction(c)\n\treturn ds.Delete(c, ds.KeyForObj(c, &SnapshotInfo{}))\n}", "func ExampleRDS_DeleteDBClusterSnapshot_shared00() {\n\tsvc := rds.New(session.New())\n\tinput := &rds.DeleteDBClusterSnapshotInput{\n\t\tDBClusterSnapshotIdentifier: aws.String(\"mydbclustersnapshot\"),\n\t}\n\n\tresult, err := svc.DeleteDBClusterSnapshot(input)\n\tif err != nil {\n\t\tif aerr, ok := err.(awserr.Error); ok {\n\t\t\tswitch aerr.Code() {\n\t\t\tcase rds.ErrCodeInvalidDBClusterSnapshotStateFault:\n\t\t\t\tfmt.Println(rds.ErrCodeInvalidDBClusterSnapshotStateFault, aerr.Error())\n\t\t\tcase rds.ErrCodeDBClusterSnapshotNotFoundFault:\n\t\t\t\tfmt.Println(rds.ErrCodeDBClusterSnapshotNotFoundFault, aerr.Error())\n\t\t\tdefault:\n\t\t\t\tfmt.Println(aerr.Error())\n\t\t\t}\n\t\t} else {\n\t\t\t// Print the error, cast err to awserr.Error to get the Code and\n\t\t\t// Message from an error.\n\t\t\tfmt.Println(err.Error())\n\t\t}\n\t\treturn\n\t}\n\n\tfmt.Println(result)\n}", "func (c *Client) DeleteSnapshotOperation(name string) *DeleteSnapshotOperation {\n\treturn c.internalClient.DeleteSnapshotOperation(name)\n}", "func (m *MockRDSAPI) DeleteDBClusterSnapshot(arg0 *rds.DeleteDBClusterSnapshotInput) (*rds.DeleteDBClusterSnapshotOutput, error) {\n\tret := m.ctrl.Call(m, \"DeleteDBClusterSnapshot\", arg0)\n\tret0, _ := ret[0].(*rds.DeleteDBClusterSnapshotOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (_m *VolumeDriver) DeleteSnapshot(opt *proto.DeleteVolumeSnapshotOpts) error {\n\tret := _m.Called(opt)\n\n\tvar r0 error\n\tif rf, ok := ret.Get(0).(func(*proto.DeleteVolumeSnapshotOpts) error); ok {\n\t\tr0 = rf(opt)\n\t} else {\n\t\tr0 = ret.Error(0)\n\t}\n\n\treturn r0\n}", "func (ctrler CtrlDefReactor) OnConfigurationSnapshotDelete(obj *ConfigurationSnapshot) error {\n\tlog.Info(\"OnConfigurationSnapshotDelete is not implemented\")\n\treturn nil\n}", "func (a *SnapshotApiService) SnapshotsDelete(ctx _context.Context, snapshotId string) ApiSnapshotsDeleteRequest {\n\treturn ApiSnapshotsDeleteRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\tsnapshotId: snapshotId,\n\t}\n}", "func (mr *MockRDSAPIMockRecorder) DeleteDBSnapshot(arg0 interface{}) *gomock.Call {\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"DeleteDBSnapshot\", reflect.TypeOf((*MockRDSAPI)(nil).DeleteDBSnapshot), arg0)\n}", "func (api *snapshotrestoreAPI) Delete(obj *cluster.SnapshotRestore) error {\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = apicl.ClusterV1().SnapshotRestore().Delete(context.Background(), &obj.ObjectMeta)\n\t\treturn err\n\t}\n\n\tapi.ct.handleSnapshotRestoreEvent(&kvstore.WatchEvent{Object: obj, Type: kvstore.Deleted})\n\treturn nil\n}", "func (c *cstor) Delete() (*v1alpha1.CASSnapshot, error) {\n\t_, err := snapshot.DestroySnapshot(c.IP, c.Snap.Spec.VolumeName, c.Snap.Name)\n\t// If there is no err that means call was successful\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// we are returning the same struct that we received as input.\n\t// This would be modified when server replies back with some property of\n\t// created snapshot\n\treturn c.Snap, nil\n}", "func DeleteSnapshotPolicy(snapshotPolicyID string) error {\n\tclient, err := NewExtPacketClient()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, e := client.Storages.DeleteSnapshotPolicy(snapshotPolicyID)\n\treturn e\n}", "func (m *MockRDSAPI) DeleteDBSnapshotRequest(arg0 *rds.DeleteDBSnapshotInput) (*request.Request, *rds.DeleteDBSnapshotOutput) {\n\tret := m.ctrl.Call(m, \"DeleteDBSnapshotRequest\", arg0)\n\tret0, _ := ret[0].(*request.Request)\n\tret1, _ := ret[1].(*rds.DeleteDBSnapshotOutput)\n\treturn ret0, ret1\n}", "func (z *zfsctl) DestroySnapshot(ctx context.Context, name, options string) *execute {\n\targs := []string{\"destroy\"}\n\tif len(options) > 0 {\n\t\targs = append(args, options)\n\t}\n\targs = append(args, name)\n\treturn &execute{ctx: ctx, name: z.cmd, args: args}\n}", "func (ctrl *csiSnapshotCommonController) deleteGroupSnapshot(groupSnapshot *crdv1alpha1.VolumeGroupSnapshot) {\n\t_ = ctrl.snapshotStore.Delete(groupSnapshot)\n\tklog.V(4).Infof(\"group snapshot %q deleted\", utils.GroupSnapshotKey(groupSnapshot))\n\n\tgroupSnapshotContentName := \"\"\n\tif groupSnapshot.Status != nil && groupSnapshot.Status.BoundVolumeGroupSnapshotContentName != nil {\n\t\tgroupSnapshotContentName = *groupSnapshot.Status.BoundVolumeGroupSnapshotContentName\n\t}\n\tif groupSnapshotContentName == \"\" {\n\t\tklog.V(5).Infof(\"deleteGroupSnapshot[%q]: group snapshot content not bound\", utils.GroupSnapshotKey(groupSnapshot))\n\t\treturn\n\t}\n\n\t// sync the group snapshot content when its group snapshot is deleted. Explicitly sync'ing\n\t// the group snapshot content here in response to group snapshot deletion prevents the group\n\t// snapshot content from waiting until the next sync period for its release.\n\tklog.V(5).Infof(\"deleteGroupSnapshot[%q]: scheduling sync of group snapshot content %s\", utils.GroupSnapshotKey(groupSnapshot), groupSnapshotContentName)\n\tctrl.groupSnapshotContentQueue.Add(groupSnapshotContentName)\n}", "func (m *MockRDSAPI) DeleteDBSnapshotWithContext(arg0 aws.Context, arg1 *rds.DeleteDBSnapshotInput, arg2 ...request.Option) (*rds.DeleteDBSnapshotOutput, error) {\n\tvarargs := []interface{}{arg0, arg1}\n\tfor _, a := range arg2 {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"DeleteDBSnapshotWithContext\", varargs...)\n\tret0, _ := ret[0].(*rds.DeleteDBSnapshotOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (m *MockRDSAPI) DeleteDBClusterSnapshotRequest(arg0 *rds.DeleteDBClusterSnapshotInput) (*request.Request, *rds.DeleteDBClusterSnapshotOutput) {\n\tret := m.ctrl.Call(m, \"DeleteDBClusterSnapshotRequest\", arg0)\n\tret0, _ := ret[0].(*request.Request)\n\tret1, _ := ret[1].(*rds.DeleteDBClusterSnapshotOutput)\n\treturn ret0, ret1\n}", "func (mr *MockRDSAPIMockRecorder) DeleteDBClusterSnapshot(arg0 interface{}) *gomock.Call {\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"DeleteDBClusterSnapshot\", reflect.TypeOf((*MockRDSAPI)(nil).DeleteDBClusterSnapshot), arg0)\n}", "func (client Client) DeleteSnapshots(ctx context.Context, accountName, containerName, blobName string, input DeleteSnapshotsInput) (result autorest.Response, err error) {\n\tif accountName == \"\" {\n\t\treturn result, validation.NewError(\"blobs.Client\", \"DeleteSnapshots\", \"`accountName` cannot be an empty string.\")\n\t}\n\tif containerName == \"\" {\n\t\treturn result, validation.NewError(\"blobs.Client\", \"DeleteSnapshots\", \"`containerName` cannot be an empty string.\")\n\t}\n\tif strings.ToLower(containerName) != containerName {\n\t\treturn result, validation.NewError(\"blobs.Client\", \"DeleteSnapshots\", \"`containerName` must be a lower-cased string.\")\n\t}\n\tif blobName == \"\" {\n\t\treturn result, validation.NewError(\"blobs.Client\", \"DeleteSnapshots\", \"`blobName` cannot be an empty string.\")\n\t}\n\n\treq, err := client.DeleteSnapshotsPreparer(ctx, accountName, containerName, blobName, input)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"blobs.Client\", \"DeleteSnapshots\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresp, err := client.DeleteSnapshotsSender(req)\n\tif err != nil {\n\t\tresult = autorest.Response{Response: resp}\n\t\terr = autorest.NewErrorWithError(err, \"blobs.Client\", \"DeleteSnapshots\", resp, \"Failure sending request\")\n\t\treturn\n\t}\n\n\tresult, err = client.DeleteSnapshotsResponder(resp)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"blobs.Client\", \"DeleteSnapshots\", resp, \"Failure responding to request\")\n\t\treturn\n\t}\n\n\treturn\n}", "func DeleteRdsSnapshots(rdsSnapshots []RdsSnapshotInfo, older int, region string, apply bool, rdsType string, out string) {\n\tlog.Printf(\"Running on region: %v\", region)\n\tlog.Printf(\"Deleting snapshots older than %v days\", older)\n\n\tnow := time.Now()\n\tdeleteDate := now.AddDate(0, 0, -(older))\n\n\tfor _, s := range rdsSnapshots {\n\t\tif s.snapshotCreatedTime.Before(deleteDate) {\n\t\t\tlog.Printf(\"Deleting %v that was created at %v for db %v\", s.dbSnapshotIdentifier, s.snapshotCreatedTime, s.dbIdentifier)\n\t\t\tif apply == true {\n\t\t\t\tdeleteSnapshot(s.dbSnapshotIdentifier, region, rdsType)\n\t\t\t} else {\n\t\t\t\tlog.Println(\"Add -y/--yes to confirm delete\")\n\t\t\t}\n\t\t}\n\t}\n\tlog.Println(\"==============================================\")\n}", "func (s *GCSSnapStore) Delete(snap Snapshot) error {\n\tobjectName := path.Join(s.prefix, snap.SnapDir, snap.SnapName)\n\treturn s.client.Bucket(s.bucket).Object(objectName).Delete(s.ctx)\n}", "func (cli *Client) SnapshotRemove(snapshotID string) error {\n\tresp, err := cli.delete(\"/snapshots/\"+snapshotID, nil, nil)\n\tensureReaderClosed(resp)\n\treturn err\n}", "func (client Client) DeleteSnapshotsSender(req *http.Request) (*http.Response, error) {\n\treturn autorest.SendWithSender(client, req,\n\t\tazure.DoRetryWithRegistration(client.Client))\n}", "func (c *Client) DeleteSampleSnapshotTemplate(request *DeleteSampleSnapshotTemplateRequest) (response *DeleteSampleSnapshotTemplateResponse, err error) {\n if request == nil {\n request = NewDeleteSampleSnapshotTemplateRequest()\n }\n response = NewDeleteSampleSnapshotTemplateResponse()\n err = c.Send(request, response)\n return\n}", "func (d *lvm) DeleteVolumeSnapshot(snapVol Volume, op *operations.Operation) error {\n\t// Remove the snapshot from the storage device.\n\tvolDevPath := d.lvmDevPath(d.config[\"lvm.vg_name\"], snapVol.volType, snapVol.contentType, snapVol.name)\n\tlvExists, err := d.logicalVolumeExists(volDevPath)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif lvExists {\n\t\t_, err = d.UnmountVolume(snapVol, false, op)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Error unmounting LVM logical volume: %w\", err)\n\t\t}\n\n\t\terr = d.removeLogicalVolume(d.lvmDevPath(d.config[\"lvm.vg_name\"], snapVol.volType, snapVol.contentType, snapVol.name))\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Error removing LVM logical volume: %w\", err)\n\t\t}\n\t}\n\n\t// For VMs, also remove the snapshot filesystem volume.\n\tif snapVol.IsVMBlock() {\n\t\tfsVol := snapVol.NewVMBlockFilesystemVolume()\n\t\terr = d.DeleteVolumeSnapshot(fsVol, op)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Remove the snapshot mount path from the storage device.\n\tsnapPath := snapVol.MountPath()\n\terr = os.RemoveAll(snapPath)\n\tif err != nil && !os.IsNotExist(err) {\n\t\treturn fmt.Errorf(\"Error removing LVM snapshot mount path %q: %w\", snapPath, err)\n\t}\n\n\t// Remove the parent snapshot directory if this is the last snapshot being removed.\n\tparentName, _, _ := api.GetParentAndSnapshotName(snapVol.name)\n\terr = deleteParentSnapshotDirIfEmpty(d.name, snapVol.volType, parentName)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (d *Dao) DeleteGRPCSnap(id int) error {\n\treturn pkgerr.WithStack(d.DB.Table(model.GRPCSnap{}.TableName()).Where(\"id=?\", id).Update(\"active\", -1).Error)\n}", "func (c *ClientWithResponses) DeleteSnapshotWithResponse(ctx context.Context, id string) (*DeleteSnapshotResponse, error) {\n\trsp, err := c.DeleteSnapshot(ctx, id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn ParseDeleteSnapshotResponse(rsp)\n}", "func (s *LocalSnapStore) Delete(snap brtypes.Snapshot) error {\n\tif err := os.Remove(path.Join(snap.Prefix, snap.SnapDir, snap.SnapName)); err != nil {\n\t\treturn err\n\t}\n\terr := os.Remove(path.Join(snap.Prefix, snap.SnapDir))\n\tif pathErr, ok := err.(*os.PathError); ok && pathErr.Err != syscall.ENOTEMPTY {\n\t\treturn err\n\t}\n\treturn nil\n}", "func NewDeleteSnapshotRequest(server string, id string) (*http.Request, error) {\n\tvar err error\n\n\tvar pathParam0 string\n\n\tpathParam0, err = runtime.StyleParam(\"simple\", false, \"id\", id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tqueryUrl, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbasePath := fmt.Sprintf(\"/snapshot/%s\", pathParam0)\n\tif basePath[0] == '/' {\n\t\tbasePath = basePath[1:]\n\t}\n\n\tqueryUrl, err = queryUrl.Parse(basePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq, err := http.NewRequest(\"DELETE\", queryUrl.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn req, nil\n}", "func (mock *Mockdb) RevertToSnapshot(i int) {\n}", "func (client Client) DeleteSnapshotsResponder(resp *http.Response) (result autorest.Response, err error) {\n\terr = autorest.Respond(\n\t\tresp,\n\t\tclient.ByInspecting(),\n\t\tazure.WithErrorUnlessStatusCode(http.StatusAccepted),\n\t\tautorest.ByClosing())\n\tresult = autorest.Response{Response: resp}\n\treturn\n}", "func (c *FakeTridentSnapshotInfos) Delete(ctx context.Context, name string, opts v1.DeleteOptions) error {\n\t_, err := c.Fake.\n\t\tInvokes(testing.NewDeleteAction(tridentsnapshotinfosResource, c.ns, name), &netappv1.TridentSnapshotInfo{})\n\n\treturn err\n}", "func (api *configurationsnapshotAPI) Delete(obj *cluster.ConfigurationSnapshot) error {\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = apicl.ClusterV1().ConfigurationSnapshot().Delete(context.Background(), &obj.ObjectMeta)\n\t\treturn err\n\t}\n\n\tapi.ct.handleConfigurationSnapshotEvent(&kvstore.WatchEvent{Object: obj, Type: kvstore.Deleted})\n\treturn nil\n}", "func Delete(params DeleteParams) error {\n\tif err := params.Validate(); err != nil {\n\t\treturn err\n\t}\n\n\t_, _, err := params.V1API.PlatformConfigurationSnapshots.DeleteSnapshotRepository(\n\t\tplatform_configuration_snapshots.NewDeleteSnapshotRepositoryParams().\n\t\t\tWithRepositoryName(params.Name),\n\t\tparams.AuthWriter,\n\t)\n\n\treturn api.UnwrapError(err)\n}", "func (c *Client) RemoveSnapshot(ctx context.Context, id string) error {\n\twrapperCli, err := c.Get(ctx)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to get a containerd grpc client: %v\", err)\n\t}\n\n\tservice := wrapperCli.client.SnapshotService(CurrentSnapshotterName(ctx))\n\tdefer service.Close()\n\n\treturn service.Remove(ctx, id)\n}", "func (api *snapshotrestoreAPI) SyncDelete(obj *cluster.SnapshotRestore) error {\n\tvar writeErr error\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t_, writeErr = apicl.ClusterV1().SnapshotRestore().Delete(context.Background(), &obj.ObjectMeta)\n\t}\n\n\tif writeErr == nil {\n\t\tapi.ct.handleSnapshotRestoreEvent(&kvstore.WatchEvent{Object: obj, Type: kvstore.Deleted})\n\t}\n\n\treturn writeErr\n}", "func (client Client) DeleteSnapshotsPreparer(ctx context.Context, accountName, containerName, blobName string, input DeleteSnapshotsInput) (*http.Request, error) {\n\tpathParameters := map[string]interface{}{\n\t\t\"containerName\": autorest.Encode(\"path\", containerName),\n\t\t\"blobName\": autorest.Encode(\"path\", blobName),\n\t}\n\n\theaders := map[string]interface{}{\n\t\t\"x-ms-version\": APIVersion,\n\t\t// only delete the snapshots but leave the blob as-is\n\t\t\"x-ms-delete-snapshots\": \"only\",\n\t}\n\n\tif input.LeaseID != nil {\n\t\theaders[\"x-ms-lease-id\"] = *input.LeaseID\n\t}\n\n\tpreparer := autorest.CreatePreparer(\n\t\tautorest.AsDelete(),\n\t\tautorest.WithBaseURL(endpoints.GetBlobEndpoint(client.BaseURI, accountName)),\n\t\tautorest.WithPathParameters(\"/{containerName}/{blobName}\", pathParameters),\n\t\tautorest.WithHeaders(headers))\n\treturn preparer.Prepare((&http.Request{}).WithContext(ctx))\n}", "func (c *Client) SnapshotRemoveAll(vm *VirtualMachine) error {\n\terr := func() error {\n\t\tctx, cancelFn := context.WithTimeout(context.Background(), c.timeout)\n\t\tdefer cancelFn()\n\n\t\tconsolidate := true\n\t\ttask, err := vm.VM.RemoveAllSnapshot(ctx, &consolidate)\n\t\t_, err = c.finishTask(ctx, task, err)\n\t\tif err != nil {\n\t\t\treturn errors.Wrapf(err, \"While removing all snapshots\")\n\t\t}\n\n\t\treturn nil\n\t}()\n\n\tif err != nil {\n\t\tswitch err := errors.Cause(err).(type) {\n\t\tcase *TimeoutExceededError:\n\t\t\t// handle specifically\n\t\t\treturn fmt.Errorf(\"Timeout while attempting to remove all snapshots for a VM\")\n\t\tdefault:\n\t\t\t// unknown error\n\t\t\treturn errors.Wrap(err, \"Got error while removing all snapshots for a VM\")\n\t\t}\n\t}\n\n\treturn nil\n}", "func (m *MockRDSAPI) DeleteDBClusterSnapshotWithContext(arg0 aws.Context, arg1 *rds.DeleteDBClusterSnapshotInput, arg2 ...request.Option) (*rds.DeleteDBClusterSnapshotOutput, error) {\n\tvarargs := []interface{}{arg0, arg1}\n\tfor _, a := range arg2 {\n\t\tvarargs = append(varargs, a)\n\t}\n\tret := m.ctrl.Call(m, \"DeleteDBClusterSnapshotWithContext\", varargs...)\n\tret0, _ := ret[0].(*rds.DeleteDBClusterSnapshotOutput)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (c *Client) DeleteSnapshotContent(name string) error {\n\tif err := c.initClient(); err != nil {\n\t\treturn err\n\t}\n\treturn c.client.VolumeSnapshotContents().Delete(name, &metav1.DeleteOptions{})\n}", "func (c *Client) DeleteSnapshotByTimeOffsetTemplate(request *DeleteSnapshotByTimeOffsetTemplateRequest) (response *DeleteSnapshotByTimeOffsetTemplateResponse, err error) {\n if request == nil {\n request = NewDeleteSnapshotByTimeOffsetTemplateRequest()\n }\n response = NewDeleteSnapshotByTimeOffsetTemplateResponse()\n err = c.Send(request, response)\n return\n}", "func (mr *MockRDSAPIMockRecorder) DeleteDBSnapshotWithContext(arg0, arg1 interface{}, arg2 ...interface{}) *gomock.Call {\n\tvarargs := append([]interface{}{arg0, arg1}, arg2...)\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"DeleteDBSnapshotWithContext\", reflect.TypeOf((*MockRDSAPI)(nil).DeleteDBSnapshotWithContext), varargs...)\n}", "func CtrRemoveSnapshot(snapshotID string) error {\n\tif err := verifyCtr(); err != nil {\n\t\treturn fmt.Errorf(\"CtrRemoveSnapshot: exception while verifying ctrd client: %s\", err.Error())\n\t}\n\tsnapshotter := CtrdClient.SnapshotService(defaultSnapshotter)\n\tif err := snapshotter.Remove(ctrdCtx, snapshotID); err != nil {\n\t\tlog.Errorf(\"CtrRemoveSnapshot: unable to remove snapshot: %v. %v\", snapshotID, err)\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *containerdCAS) RemoveSnapshot(snapshotID string) error {\n\tsnapshotter := ctrdClient.SnapshotService(defaultSnapshotter)\n\tif err := snapshotter.Remove(ctrdCtx, snapshotID); err != nil && !isNotFoundError(err) {\n\t\treturn fmt.Errorf(\"RemoveSnapshot: Exception while removing snapshot: %s. %s\", snapshotID, err.Error())\n\t}\n\treturn nil\n}", "func (a *HyperflexApiService) DeleteHyperflexVmSnapshotInfo(ctx context.Context, moid string) ApiDeleteHyperflexVmSnapshotInfoRequest {\n\treturn ApiDeleteHyperflexVmSnapshotInfoRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\tmoid: moid,\n\t}\n}", "func (c *Client) SnapshotRemove(vm *VirtualMachine, moRef *types.ManagedObjectReference) error {\n\terr := func() error {\n\t\tctx, cancelFn := context.WithTimeout(context.Background(), c.timeout)\n\t\tdefer cancelFn()\n\n\t\tconsolidate := true\n\t\treq := types.RemoveSnapshot_Task{\n\t\t\tThis: moRef.Reference(),\n\t\t\tRemoveChildren: false,\n\t\t\tConsolidate: &consolidate,\n\t\t}\n\n\t\tres, err := methods.RemoveSnapshot_Task(ctx, vm.VM.Client(), &req)\n\t\tif err := c.checkErr(ctx, err); err != nil {\n\t\t\treturn errors.Wrapf(err, \"While removing snapshot\")\n\t\t}\n\n\t\ttask := object.NewTask(vm.VM.Client(), res.Returnval)\n\t\t_, err = c.finishTask(ctx, task, nil)\n\t\tif err != nil {\n\t\t\treturn errors.Wrapf(err, \"While waiting for snapshot removal\")\n\t\t}\n\n\t\treturn nil\n\t}()\n\n\tif err != nil {\n\t\tswitch err := errors.Cause(err).(type) {\n\t\tcase *TimeoutExceededError:\n\t\t\t// handle specifically\n\t\t\treturn fmt.Errorf(\"Timeout while attempting to list snapshots for a VM\")\n\t\tdefault:\n\t\t\t// unknown error\n\t\t\treturn errors.Wrap(err, \"Got error while listing snapshots for a VM\")\n\t\t}\n\t}\n\n\treturn nil\n}", "func ExampleRDS_DeleteDBInstance_shared00() {\n\tsvc := rds.New(session.New())\n\tinput := &rds.DeleteDBInstanceInput{\n\t\tDBInstanceIdentifier: aws.String(\"test-instance\"),\n\t\tFinalDBSnapshotIdentifier: aws.String(\"test-instance-final-snap\"),\n\t\tSkipFinalSnapshot: aws.Bool(false),\n\t}\n\n\tresult, err := svc.DeleteDBInstance(input)\n\tif err != nil {\n\t\tif aerr, ok := err.(awserr.Error); ok {\n\t\t\tswitch aerr.Code() {\n\t\t\tcase rds.ErrCodeDBInstanceNotFoundFault:\n\t\t\t\tfmt.Println(rds.ErrCodeDBInstanceNotFoundFault, aerr.Error())\n\t\t\tcase rds.ErrCodeInvalidDBInstanceStateFault:\n\t\t\t\tfmt.Println(rds.ErrCodeInvalidDBInstanceStateFault, aerr.Error())\n\t\t\tcase rds.ErrCodeDBSnapshotAlreadyExistsFault:\n\t\t\t\tfmt.Println(rds.ErrCodeDBSnapshotAlreadyExistsFault, aerr.Error())\n\t\t\tcase rds.ErrCodeSnapshotQuotaExceededFault:\n\t\t\t\tfmt.Println(rds.ErrCodeSnapshotQuotaExceededFault, aerr.Error())\n\t\t\tcase rds.ErrCodeInvalidDBClusterStateFault:\n\t\t\t\tfmt.Println(rds.ErrCodeInvalidDBClusterStateFault, aerr.Error())\n\t\t\tcase rds.ErrCodeDBInstanceAutomatedBackupQuotaExceededFault:\n\t\t\t\tfmt.Println(rds.ErrCodeDBInstanceAutomatedBackupQuotaExceededFault, aerr.Error())\n\t\t\tdefault:\n\t\t\t\tfmt.Println(aerr.Error())\n\t\t\t}\n\t\t} else {\n\t\t\t// Print the error, cast err to awserr.Error to get the Code and\n\t\t\t// Message from an error.\n\t\t\tfmt.Println(err.Error())\n\t\t}\n\t\treturn\n\t}\n\n\tfmt.Println(result)\n}", "func (d *DefaultDriver) ValidateCreateSnapshot(name string, params map[string]string) error {\n\treturn &errors.ErrNotSupported{\n\t\tType: \"Function\",\n\t\tOperation: \"ValidateCreateSnapshot()\",\n\t}\n}", "func (r *Redis) DeleteSignature(sourceID, collection string) error {\n\tkey := \"source#\" + sourceID + \":collection#\" + collection + \":chunks\"\n\tconnection := r.pool.Get()\n\tdefer connection.Close()\n\t_, err := connection.Do(\"DEL\", key)\n\tnoticeError(err)\n\tif err != nil && err != redis.ErrNil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (mr *MockRDSAPIMockRecorder) DeleteDBSnapshotRequest(arg0 interface{}) *gomock.Call {\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"DeleteDBSnapshotRequest\", reflect.TypeOf((*MockRDSAPI)(nil).DeleteDBSnapshotRequest), arg0)\n}", "func (d *lvm) UnmountVolumeSnapshot(snapVol Volume, op *operations.Operation) (bool, error) {\n\tunlock := snapVol.MountLock()\n\tdefer unlock()\n\n\tvar err error\n\tourUnmount := false\n\tmountPath := snapVol.MountPath()\n\n\trefCount := snapVol.MountRefCountDecrement()\n\n\t// Check if already mounted.\n\tif snapVol.contentType == ContentTypeFS && filesystem.IsMountPoint(mountPath) {\n\t\tif refCount > 0 {\n\t\t\td.logger.Debug(\"Skipping unmount as in use\", logger.Ctx{\"volName\": snapVol.name, \"refCount\": refCount})\n\t\t\treturn false, ErrInUse\n\t\t}\n\n\t\terr = TryUnmount(mountPath, 0)\n\t\tif err != nil {\n\t\t\treturn false, fmt.Errorf(\"Failed to unmount LVM snapshot volume: %w\", err)\n\t\t}\n\n\t\td.logger.Debug(\"Unmounted logical volume snapshot\", logger.Ctx{\"path\": mountPath})\n\n\t\t// Check if a temporary snapshot exists, and if so remove it.\n\t\ttmpVolName := fmt.Sprintf(\"%s%s\", snapVol.name, tmpVolSuffix)\n\t\ttmpVolDevPath := d.lvmDevPath(d.config[\"lvm.vg_name\"], snapVol.volType, snapVol.contentType, tmpVolName)\n\t\texists, err := d.logicalVolumeExists(tmpVolDevPath)\n\t\tif err != nil {\n\t\t\treturn true, fmt.Errorf(\"Failed to check existence of temporary LVM snapshot volume %q: %w\", tmpVolDevPath, err)\n\t\t}\n\n\t\tif exists {\n\t\t\terr = d.removeLogicalVolume(tmpVolDevPath)\n\t\t\tif err != nil {\n\t\t\t\treturn true, fmt.Errorf(\"Failed to remove temporary LVM snapshot volume %q: %w\", tmpVolDevPath, err)\n\t\t\t}\n\t\t}\n\n\t\t// We only deactivate filesystem volumes if an unmount was needed to better align with our\n\t\t// unmount return value indicator.\n\t\t_, err = d.deactivateVolume(snapVol)\n\t\tif err != nil {\n\t\t\treturn false, err\n\t\t}\n\n\t\tourUnmount = true\n\t} else if snapVol.contentType == ContentTypeBlock {\n\t\t// For VMs, unmount the filesystem volume.\n\t\tif snapVol.IsVMBlock() {\n\t\t\tfsVol := snapVol.NewVMBlockFilesystemVolume()\n\t\t\tourUnmount, err = d.UnmountVolumeSnapshot(fsVol, op)\n\t\t\tif err != nil {\n\t\t\t\treturn false, err\n\t\t\t}\n\t\t}\n\n\t\tvolDevPath := d.lvmDevPath(d.config[\"lvm.vg_name\"], snapVol.volType, snapVol.contentType, snapVol.name)\n\t\tif shared.PathExists(volDevPath) {\n\t\t\tif refCount > 0 {\n\t\t\t\td.logger.Debug(\"Skipping unmount as in use\", logger.Ctx{\"volName\": snapVol.name, \"refCount\": refCount})\n\t\t\t\treturn false, ErrInUse\n\t\t\t}\n\n\t\t\t_, err = d.deactivateVolume(snapVol)\n\t\t\tif err != nil {\n\t\t\t\treturn false, err\n\t\t\t}\n\n\t\t\tourUnmount = true\n\t\t}\n\t}\n\n\treturn ourUnmount, nil\n}", "func (ks *KopiaSnapshotter) RestoreSnapshot(snapID, restoreDir string, opts map[string]string) error {\n\treturn ks.snap.RestoreSnapshot(snapID, restoreDir)\n}", "func (client *GCSBlobstore) Delete(dest string) error {\n\tif client.readOnly() {\n\t\treturn ErrInvalidROWriteOperation\n\t}\n\n\terr := client.getObjectHandle(client.authenticatedGCS, dest).Delete(context.Background())\n\tif err == storage.ErrObjectNotExist {\n\t\treturn nil\n\t}\n\treturn err\n}", "func (d *InMemoryDistributor) ClearSnapshot(name string) {\n\td.snapshotsLock.Lock()\n\tdefer d.snapshotsLock.Unlock()\n\n\tdelete(d.snapshots, name)\n}", "func (c *snapshotRules) Delete(ctx context.Context, name string, opts v1.DeleteOptions) error {\n\treturn c.client.Delete().\n\t\tNamespace(c.ns).\n\t\tResource(\"snapshotrules\").\n\t\tName(name).\n\t\tBody(&opts).\n\t\tDo(ctx).\n\t\tError()\n}", "func TestCmdSnapshot(t *testing.T) {\n\tassert := asrt.New(t)\n\n\ttestDir, _ := os.Getwd()\n\tfmt.Println(testDir)\n\tsite := TestSites[0]\n\tcleanup := site.Chdir()\n\tapp, err := ddevapp.NewApp(site.Dir, false, \"\")\n\tassert.NoError(err)\n\tdefer func() {\n\t\t// Make sure all databases are back to default empty\n\t\t_ = app.Stop(true, false)\n\t\t_ = app.Start()\n\t\tcleanup()\n\t}()\n\n\t// Ensure that a snapshot can be created\n\targs := []string{\"snapshot\", \"--name\", \"test-snapshot\"}\n\tout, err := exec.RunCommand(DdevBin, args)\n\tassert.NoError(err)\n\tassert.Contains(string(out), \"Created snapshot test-snapshot\")\n\n\t// Try to delete a not existing snapshot\n\targs = []string{\"snapshot\", \"--name\", \"not-existing-snapshot\", \"--cleanup\", \"--yes\"}\n\tout, err = exec.RunCommand(DdevBin, args)\n\tassert.Error(err)\n\tassert.Contains(string(out), \"Failed to delete snapshot\")\n\n\t// Ensure that an existing snapshot can be deleted\n\targs = []string{\"snapshot\", \"--name\", \"test-snapshot\", \"--cleanup\"}\n\tout, err = exec.RunCommand(DdevBin, args)\n\tassert.NoError(err)\n\tassert.Contains(string(out), \"Deleted database snapshot test-snapshot\")\n}", "func (v *IBM) DeleteImage(ctx *lepton.Context, snapshotID string) error {\n\treturn nil\n}", "func (c *Client) DestroyDatabaseSnapshot(identifier string) error {\n\t_, err := c.MakeApiRequest(\"DELETE\", \"/1.0/database_snapshots/\"+identifier, nil, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (s *FSMSnapshot) Release() {\n}" ]
[ "0.71738756", "0.7149398", "0.70576525", "0.70359266", "0.70085084", "0.69433635", "0.6907444", "0.68741393", "0.68399966", "0.68327135", "0.68301064", "0.6815775", "0.677051", "0.6747238", "0.6710406", "0.6632568", "0.66140795", "0.660155", "0.65779227", "0.65734464", "0.65655744", "0.6527679", "0.64733505", "0.646627", "0.6388498", "0.6379815", "0.63516736", "0.6344315", "0.6308491", "0.6302531", "0.6261692", "0.6211864", "0.62046915", "0.6177691", "0.6119382", "0.611722", "0.6085413", "0.6011659", "0.6001952", "0.5972033", "0.59130114", "0.58532345", "0.58445126", "0.5826307", "0.5796241", "0.57926553", "0.57667077", "0.57620263", "0.57322353", "0.57260644", "0.5723288", "0.5697487", "0.5688613", "0.5677169", "0.56641734", "0.5585782", "0.55416626", "0.5486642", "0.5453563", "0.5386254", "0.53545445", "0.5351144", "0.53411114", "0.5256647", "0.5250007", "0.52422464", "0.52265376", "0.5205886", "0.5180828", "0.51659167", "0.5150838", "0.5134625", "0.5130022", "0.512439", "0.51117015", "0.51087314", "0.5087425", "0.50200975", "0.499837", "0.49964064", "0.4995377", "0.49948126", "0.49834755", "0.49559677", "0.4946359", "0.49258018", "0.49067518", "0.48944992", "0.4888111", "0.48629764", "0.48018745", "0.47892448", "0.4760093", "0.47461495", "0.46931186", "0.46888775", "0.46842638", "0.4671862", "0.46662748", "0.4648622" ]
0.71391094
2
GetInstance uses the override method GetZoneFn or the real implementation.
func (c *TestClient) GetInstance(project, zone, name string) (*compute.Instance, error) { if c.GetInstanceFn != nil { return c.GetInstanceFn(project, zone, name) } return c.client.GetInstance(project, zone, name) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (z *zones) GetZone(ctx context.Context) (cloudprovider.Zone, error) {\n\tklog.V(5).Info(\"GetZone()\")\n\treturn cloudprovider.Zone{\n\t\tFailureDomain: \"laptop\",\n\t\tRegion: \"virtualbox\",\n\t}, nil\n}", "func (bc *Baiducloud) GetZone(ctx context.Context) (cloudprovider.Zone, error) {\n\tzone := cloudprovider.Zone{\n\t\tFailureDomain: \"unknow\",\n\t\tRegion: bc.Region,\n\t}\n\tif bc.NodeName != \"\" {\n\t\tins, err := bc.getInstanceByNodeName(ctx, types.NodeName(bc.NodeName))\n\t\t// ins, err := bc.getVirtualMachine(types.NodeName(bc.NodeIP))\n\t\tif err != nil {\n\t\t\treturn zone, err\n\t\t}\n\t\tzone.FailureDomain = ins.AvailableZone\n\t}\n\treturn zone, nil\n}", "func (z zones) GetZone() (cloudprovider.Zone, error) {\n\treturn cloudprovider.Zone{Region: z.region}, nil\n}", "func (z zones) GetZone(ctx context.Context) (cloudprovider.Zone, error) {\n\tzone, err := queryInstanceMetadata(\"availability-zone\")\n\tif err != nil {\n\t\treturn cloudprovider.Zone{}, err\n\t}\n\n\treturn cloudprovider.Zone{Region: zone}, nil\n}", "func (p *PowerDNS) GetZone(domain string) (*Zone, error) {\n\tzone := &Zone{}\n\tmyError := new(Error)\n\tzoneSling := p.makeSling()\n\tresp, err := zoneSling.New().Get(\"servers/\"+p.VHost+\"/zones/\"+strings.TrimRight(domain, \".\")).Receive(zone, myError)\n\n\tif err == nil && resp.StatusCode >= 400 {\n\t\tmyError.Message = strings.Join([]string{resp.Status, myError.Message}, \" \")\n\t\treturn &Zone{}, myError\n\t}\n\n\tzone.PowerDNSHandle = p\n\treturn zone, err\n}", "func (w *Watcher) GetZone(domain string) (*Zone, error) {\n\tmutableMutex.Lock()\n\tdefer mutableMutex.Unlock()\n\tif w.ZoneMap == nil {\n\t\tw.ZoneMap = make(map[string]*Zone)\n\t}\n\tzone, ok := w.ZoneMap[domain]\n\tif !ok {\n\t\treturn nil, errors.Errorf(\"not exist domain\")\n\t}\n\treturn zone, nil\n}", "func (s *Store) UnsafeGetInstance(name, encodedData string) *Instance {\n\n\t// get task class\n\ttt := s.UnsafeGetDefinition(name)\n\n\t// unserialize data\n\terr := serializer.Decode(encodedData, tt.defaultTask.Handler)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn tt.defaultTask\n}", "func (c *TestClient) GetZone(project, zone string) (*compute.Zone, error) {\n\tif c.GetZoneFn != nil {\n\t\treturn c.GetZoneFn(project, zone)\n\t}\n\treturn c.client.GetZone(project, zone)\n}", "func (b *Byzcoin) GetInstanceFactory() ledger.InstanceFactory {\n\treturn instanceFactory{}\n}", "func GetInstance(key string, controllerFunc func() interfaces.IController) interfaces.IController {\n\tinstanceMapMutex.Lock()\n\tdefer instanceMapMutex.Unlock()\n\n\tif instanceMap[key] == nil {\n\t\tinstanceMap[key] = controllerFunc()\n\t\tinstanceMap[key].InitializeController()\n\t}\n\treturn instanceMap[key]\n}", "func GetInstance(computeService *compute.Service) (*compute.Instance, error) {\n\treturn computeService.Instances.Get(ProjectID, Zone, InstanceName).Do()\n}", "func NewZone(info CFInfo, homeDir string, errorHandler func(error), whitelist map[string]bool, envDialTimeout string, ui terminal.UI, writer io.Writer, logger trace.Printer) *Zone {\n\tconfig := newRepositoryConfig(homeDir, info, errorHandler)\n\trepoLocator, cloudController := newRepoLocator(config, info, envDialTimeout, ui, logger)\n\n\terr := setAPIEndpoint(info.API, config, repoLocator.GetEndpointRepository())\n\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tverifyLogin(repoLocator, info)\n\n\treturn &Zone{\n\t\tname: info.ZoneName,\n\t\tconfig: config,\n\t\trepoLocator: repoLocator,\n\t\tappRepo: NewAppRepo(config, cloudController),\n\t\tspaceRepo: NewSpaceRepo(config, cloudController),\n\t\teventRepo: NewEventRepo(config, cloudController),\n\t\twhitelist: whitelist,\n\t}\n}", "func (p *provider) getInstance(ctn *container) (reflect.Value, error) {\n\tif p.scope == Singleton {\n\t\tvar err error\n\t\tp.once.Do(func() {\n\t\t\tp.instance, err = p.call(ctn)\n\t\t})\n\t\treturn p.instance, err\n\t} else {\n\t\treturn p.call(ctn)\n\t}\n}", "func (p *Provider) CreateZone(zone dns.Zone) error {\n return fmt.Errorf(\"NOT IMPLEMENTED\")\n}", "func (wfm *Store) UnsafeGetInstance(name, encodedData string) (*Instance, error) {\n\n\tdef := wfm.UnsafeGetDefinition(name)\n\n\tif def == nil {\n\t\tpanic(fmt.Sprint(\"unknown workflow: \", name))\n\t}\n\n\tif encodedData == `\"\"` {\n\t\tencodedData = \"{}\"\n\t}\n\n\tvar wfDef *Definition\n\tif def.versionDef != nil {\n\t\t// in this case the workflow was versioned while running.\n\t\t// so we get the initial workflow from the list of versions in the version definition\n\t\twfDef = def.versionDef.getInitialDefinition()\n\t} else {\n\t\twfDef = def.workflowDef\n\t}\n\n\terr := serializer.Decode(encodedData, wfDef.defaultInstance.Handler)\n\n\treturn wfDef.defaultInstance, err\n}", "func (ac *azureClient) GetZone(ctx context.Context, resourceGroupName, zoneName string) (privatedns.PrivateZone, error) {\n\tctx, _, done := tele.StartSpanWithLogger(ctx, \"privatedns.AzureClient.GetZone\")\n\tdefer done()\n\tzone, err := ac.privatezones.Get(ctx, resourceGroupName, zoneName)\n\tif err != nil {\n\t\treturn privatedns.PrivateZone{}, err\n\t}\n\treturn zone, nil\n}", "func (client DnsClient) GetZone(ctx context.Context, request GetZoneRequest) (response GetZoneResponse, err error) {\n\tvar ociResponse common.OCIResponse\n\tpolicy := common.NoRetryPolicy()\n\tif request.RetryPolicy() != nil {\n\t\tpolicy = *request.RetryPolicy()\n\t}\n\tociResponse, err = common.Retry(ctx, request, client.getZone, policy)\n\tif err != nil {\n\t\tif ociResponse != nil {\n\t\t\tif httpResponse := ociResponse.HTTPResponse(); httpResponse != nil {\n\t\t\t\topcRequestId := httpResponse.Header.Get(\"opc-request-id\")\n\t\t\t\tresponse = GetZoneResponse{RawResponse: httpResponse, OpcRequestId: &opcRequestId}\n\t\t\t} else {\n\t\t\t\tresponse = GetZoneResponse{}\n\t\t\t}\n\t\t}\n\t\treturn\n\t}\n\tif convertedResponse, ok := ociResponse.(GetZoneResponse); ok {\n\t\tresponse = convertedResponse\n\t} else {\n\t\terr = fmt.Errorf(\"failed to convert OCIResponse into GetZoneResponse\")\n\t}\n\treturn\n}", "func GetInstance(address common.Address) (*Servicekeyresolver, error) {\n\t_rpc := rpc.GetInstance()\n\tclient := _rpc.GetEthClient()\n\n\tinstance, err := NewServicekeyresolver(address, client)\n\tif err != nil {\n\t\tlog.Error(err)\n\t\treturn nil, err\n\t}\n\n\tif instance == nil {\n\t\terr := fmt.Errorf(\"Cannot get Servicekeyresolver Instance\")\n\t\treturn nil, err\n\t}\n\treturn instance, nil\n}", "func (m *manager) Get(name, zone string) (*compute.InstanceGroup, error) {\n\tig, err := m.cloud.GetInstanceGroup(name, zone)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn ig, nil\n}", "func (o InstanceFromTemplateOutput) Zone() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *InstanceFromTemplate) pulumi.StringOutput { return v.Zone }).(pulumi.StringOutput)\n}", "func (t *TimeService) Zone(request *ZoneRequest) (*ZoneResponse, error) {\n\trsp := &ZoneResponse{}\n\treturn rsp, t.client.Call(\"time\", \"Zone\", request, rsp)\n}", "func (o LookupInstanceResultOutput) Zone() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupInstanceResult) string { return v.Zone }).(pulumi.StringOutput)\n}", "func (o InstanceNodeOutput) Zone() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v InstanceNode) *string { return v.Zone }).(pulumi.StringPtrOutput)\n}", "func (c *MockComputeClient) GetInstance(ctx context.Context, id string) (*core.Instance, error) {\n\treturn nil, nil\n}", "func (s *FastDNSv2Service) GetZone(ctx context.Context, zone string) (*ZoneMetadata, *Response, error) {\n\tu := fmt.Sprintf(\"config-dns/v2/zones/%v\", zone)\n\n\treq, err := s.client.NewRequest(\"GET\", u, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tvar zmeta *ZoneMetadata\n\tresp, err := s.client.Do(ctx, req, &zmeta)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\n\treturn zmeta, resp, nil\n}", "func (c APIClient) GetZone(zone string) (*Zone, error) {\n\tz := NewZone(zone)\n\t_, err := c.doHTTPUnmarshal(\"GET\", fmt.Sprintf(\"https://api.nsone.net/v1/zones/%s\", z.Zone), nil, z)\n\treturn z, err\n}", "func TestGetInstance(t *testing.T) {\n\tvar controller = controller.GetInstance(\"ControllerTestKey1\", func() interfaces.IController { return &controller.Controller{Key: \"ControllerTestKey1\"} })\n\n\tif controller == nil {\n\t\tt.Error(\"Expecting instance not nil\")\n\t}\n}", "func (z *zones) GetZoneByProviderID(ctx context.Context, providerID string) (cloudprovider.Zone, error) {\n\tklog.V(5).Infof(\"GetZoneByProviderID(%v)\", providerID)\n\treturn cloudprovider.Zone{\n\t\tFailureDomain: \"virtualbox\",\n\t\tRegion: \"virtualbox\",\n\t}, nil\n}", "func (this *MillisToZoneName) Constructor() FunctionConstructor { return NewMillisToZoneName }", "func (azc *azureZonesClient) Get(ctx context.Context, spec azure.ResourceSpecGetter) (result interface{}, err error) {\n\tctx, _, done := tele.StartSpanWithLogger(ctx, \"privatedns.azureZonesClient.Get\")\n\tdefer done()\n\tzone, err := azc.privatezones.Get(ctx, spec.ResourceGroupName(), spec.ResourceName())\n\tif err != nil {\n\t\treturn privatedns.PrivateZone{}, err\n\t}\n\treturn zone, nil\n}", "func (z *zones) New(name string) (dnsprovider.Zone, error) {\n\ta := &zone{\n\t\tdnsView: z.dnsView,\n\t\tzoneInfo: dns.DNSZoneInfo{\n\t\t\tName: name,\n\t\t},\n\t}\n\n\treturn a, nil\n}", "func Get(client *gophercloud.ServiceClient, zoneID string) (r GetResult) {\n\tresp, err := client.Get(zoneURL(client, zoneID), &r.Body, nil)\n\t_, r.Header, r.Err = gophercloud.ParseResponse(resp, err)\n\treturn\n}", "func (p *Provider) GetZone(name string) (dns.Zone, error) {\n mZones, err := p.listMZones()\n if err != nil {\n return dns.Zone{}, fmt.Errorf(\"couldn't get zone `%s`, see: failed to list zones, see: %v\", name, err)\n }\n\n var foundZone *models.Zone\n\n for _, z := range mZones {\n if z.Origin != nil && *z.Origin == name {\n foundZone = z\n }\n }\n\n if foundZone == nil {\n return dns.Zone{}, fmt.Errorf(\"zone `%s` not found\", name)\n }\n\n mZone, err := p.getMZone(*foundZone.Origin, foundZone.VirtualNameServer)\n if err != nil {\n return dns.Zone{}, fmt.Errorf(\"couldn't get zone `%s`, see: failed to retrieve zone upstream, see: %v\", name, err)\n }\n\n return p.mapToDNSZone(mZone), nil\n}", "func GetZoneInfo(instanceID string) (ZoneInfoStrut, error) {\n\t//-- New Var based on ZoneInfoStrut\n\tzoneInfo := ZoneInfoStrut{}\n\tif instanceID == \"\" {\n\t\treturn zoneInfo, errors.New(\"instanceid not provided\")\n\t}\n\t//-- Get JSON Config\n\ttrans := &http.Transport{Proxy: http.ProxyFromEnvironment}\n\tcl := &http.Client{Transport: trans, Timeout: time.Second * 30}\n\n\treq, err := http.NewRequest(\"GET\", \"https://files.hornbill.com/instances/\"+instanceID+\"/zoneinfo\", nil)\n\tif err != nil {\n\t\tlog.Println(\"Could not make request: \" + err.Error())\n\t\treturn zoneInfo, err\n\t}\n\tresponse, err := cl.Do(req)\n\tif err != nil || response.StatusCode != 200 {\n\t\tif err != nil {\n\t\t\tlog.Println(\"Error Loading Zone Info File: \" + err.Error())\n\t\t} else {\n\t\t\tlog.Println(\"Unexpected status when attempting to load Zone Info from \" + \"https://files.hornbill.com/instances/\" + instanceID + \"/zoneinfo\" + \" : \" + response.Status)\n\t\t}\n\t\t//-- If we fail fall over to using files.hornbill.co\n\t\treq, err = http.NewRequest(\"GET\", \"https://files.hornbill.co/instances/\"+instanceID+\"/zoneinfo\", nil)\n\t\tif err != nil {\n\t\t\tlog.Println(\"Could not makre request: \" + err.Error())\n\t\t\treturn zoneInfo, err\n\t\t}\n\t\tresponse, err = cl.Do(req)\n\n\t\t//-- If we still have an error then return out\n\t\tif err != nil {\n\t\t\tlog.Println(\"Error Loading Zone Info File: \" + err.Error())\n\t\t\treturn zoneInfo, err\n\t\t}\n\t}\n\t//-- Close Connection\n\tdefer response.Body.Close()\n\n\t//-- New Decoder\n\tdecoder := json.NewDecoder(response.Body)\n\n\t//-- Decode JSON\n\terrDECODE := decoder.Decode(&zoneInfo)\n\n\t//-- Error Checking\n\tif errDECODE != nil {\n\t\tlog.Println(\"Error Decoding Zone Info File:\", errDECODE.Error())\n\t\treturn zoneInfo, errDECODE\n\t}\n\treturn zoneInfo, nil\n}", "func (bc *Baiducloud) GetZoneByProviderID(ctx context.Context, providerID string) (cloudprovider.Zone, error) {\n\tinstance, err := bc.getInstanceByProviderID(ctx, providerID)\n\tif err != nil {\n\t\treturn cloudprovider.Zone{}, err\n\t}\n\n\treturn cloudprovider.Zone{\n\t\tFailureDomain: instance.AvailableZone,\n\t\tRegion: bc.Region,\n\t}, nil\n}", "func (f *FakeInstanceGroups) getInstanceGroup(name, zone string) (*compute.InstanceGroup, error) {\n\tfor ig := range f.zonesToIGsToInstances[zone] {\n\t\tif ig.Name == name {\n\t\t\treturn ig, nil\n\t\t}\n\t}\n\n\treturn nil, test.FakeGoogleAPINotFoundErr()\n}", "func (c *TestClient) GetTargetInstance(project, zone, name string) (*compute.TargetInstance, error) {\n\tif c.GetTargetInstanceFn != nil {\n\t\treturn c.GetTargetInstanceFn(project, zone, name)\n\t}\n\treturn c.client.GetTargetInstance(project, zone, name)\n}", "func GetZone(ctx context.Context, tenantID int, zoneID int) (*database.Zone, error) {\n\treq := handlers.GetZoneRequest{\n\t\tZoneID: zoneID,\n\t\tZoneOptions: database.ZoneOptions{\n\t\t\tWithPlaces: true,\n\t\t},\n\t}\n\tresp := database.Zone{}\n\terr := jsonConn.RequestWithContext(ctx, \"zones.get\", &req, &resp)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &resp, nil\n}", "func (f *Factory) New(providerConf digitalocean.Config, clusterState *cluster.State) (provider.Activity, error) {\n\tk8s := &K8s{}\n\tk8s.moduleDir = filepath.Join(config.Global.ProjectRoot, \"terraform/digitalocean/\"+myName)\n\tk8s.backendKey = \"states/terraform-\" + myName + \".state\"\n\tk8s.backendConf = digitalocean.BackendSpec{\n\t\tBucket: providerConf.ClusterName,\n\t\tKey: k8s.backendKey,\n\t\tEndpoint: providerConf.Region + \".digitaloceanspaces.com\",\n\t}\n\trawProvisionerData, err := yaml.Marshal(providerConf.Provisioner)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error occurret while marshal provisioner config: %s\", err.Error())\n\t}\n\tif err = yaml.Unmarshal(rawProvisionerData, &k8s.config); err != nil {\n\t\treturn nil, fmt.Errorf(\"error occurret while parsing provisioner config: %s\", err.Error())\n\t}\n\n\tk8s.config.ClusterName = providerConf.ClusterName\n\tk8s.config.Region = providerConf.Region\n\n\tk8s.terraform, err = executor.NewTerraformRunner(k8s.moduleDir, provisioner.GetAwsAuthEnv()...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tk8s.terraform.LogLabels = append(k8s.terraform.LogLabels, fmt.Sprintf(\"cluster='%s'\", providerConf.ClusterName))\n\treturn k8s, nil\n}", "func (client DnsClient) getZone(ctx context.Context, request common.OCIRequest) (common.OCIResponse, error) {\n\thttpRequest, err := request.HTTPRequest(http.MethodGet, \"/zones/{zoneNameOrId}\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar response GetZoneResponse\n\tvar httpResponse *http.Response\n\thttpResponse, err = client.Call(ctx, &httpRequest)\n\tdefer common.CloseBodyIfValid(httpResponse)\n\tresponse.RawResponse = httpResponse\n\tif err != nil {\n\t\treturn response, err\n\t}\n\n\terr = common.UnmarshalResponse(httpResponse, &response)\n\treturn response, err\n}", "func (o InstanceMemcacheNodeOutput) Zone() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v InstanceMemcacheNode) *string { return v.Zone }).(pulumi.StringPtrOutput)\n}", "func FixedZone(name string, offset int) *Location {}", "func (info *endpointsInfo) GetZone() string {\n\treturn \"\"\n}", "func (c *Client) CloudGetInstance(projectID, instanceID string) (instance *types.CloudInstance, err error) {\n\terr = c.Get(queryEscape(\"/cloud/project/%s/instance/%s\", projectID, instanceID), &instance)\n\treturn instance, nil\n}", "func (info *BaseEndpointInfo) GetZone() string {\n\treturn info.Zone\n}", "func (c *MockComputeClient) GetInstance(ctx context.Context, id string) (*core.Instance, error) {\n\tif instance, ok := instances[id]; ok {\n\t\treturn instance, nil\n\t}\n\treturn nil, errors.New(\"instance not found\")\n}", "func (qc *QingCloud) GetZoneByProviderID(ctx context.Context, providerID string) (cloudprovider.Zone, error) {\n\tklog.V(4).Infof(\"GetZoneByProviderID() called, current zone is %v, and return zone directly as temporary solution\", qc.zone)\n\treturn cloudprovider.Zone{Region: qc.zone}, nil\n}", "func TestGetInstance(t *testing.T) {\n\tctx, cancelFunc := context.WithTimeout(context.Background(), standardTimeout)\n\tdefer cancelFunc()\n\n\tinstances, err := bat.StartRandomInstances(ctx, \"\", 1)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to launch instance: %v\", err)\n\t}\n\n\t_, err = bat.RetrieveInstanceStatus(ctx, \"\", instances[0])\n\tif err != nil {\n\t\tt.Errorf(\"Failed to retrieve instance status: %v\", err)\n\t}\n\n\tscheduled, err := bat.WaitForInstancesLaunch(ctx, \"\", instances, false)\n\tif err != nil {\n\t\tt.Errorf(\"Instance %s did not launch: %v\", instances[0], err)\n\t}\n\n\t_, err = bat.DeleteInstances(ctx, \"\", scheduled)\n\tif err != nil {\n\t\tt.Errorf(\"Failed to delete instance %s: %v\", instances[0], err)\n\t}\n}", "func (s *API) GetInstance(req *GetInstanceRequest, opts ...scw.RequestOption) (*Instance, error) {\n\tvar err error\n\n\tif req.Region == \"\" {\n\t\tdefaultRegion, _ := s.client.GetDefaultRegion()\n\t\treq.Region = defaultRegion\n\t}\n\n\tif fmt.Sprint(req.Region) == \"\" {\n\t\treturn nil, errors.New(\"field Region cannot be empty in request\")\n\t}\n\n\tif fmt.Sprint(req.InstanceID) == \"\" {\n\t\treturn nil, errors.New(\"field InstanceID cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"GET\",\n\t\tPath: \"/rdb/v1/regions/\" + fmt.Sprint(req.Region) + \"/instances/\" + fmt.Sprint(req.InstanceID) + \"\",\n\t\tHeaders: http.Header{},\n\t}\n\n\tvar resp Instance\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func LookupInstance(ctx *pulumi.Context, args *LookupInstanceArgs, opts ...pulumi.InvokeOption) (*LookupInstanceResult, error) {\n\topts = internal.PkgInvokeDefaultOpts(opts)\n\tvar rv LookupInstanceResult\n\terr := ctx.Invoke(\"google-native:compute/v1:getInstance\", args, &rv, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &rv, nil\n}", "func (z *zones) GetZoneByProviderID(ctx context.Context, providerID string) (cloudprovider.Zone, error) {\n\tinstance, err := z.p.computeInstanceByProviderID(ctx, providerID)\n\tif err != nil {\n\t\treturn cloudprovider.Zone{}, err\n\t}\n\n\treturn cloudprovider.Zone{Region: instance.ZoneName}, nil\n}", "func (decryptor *PgDecryptor) GetZoneMatcher() *zone.Matcher {\n\treturn decryptor.zoneMatcher\n}", "func (w *ServerInterfaceWrapper) GetZone(ctx echo.Context) error {\n\tvar err error\n\t// ------------- Path parameter \"zone\" -------------\n\tvar zone Zone\n\n\terr = runtime.BindStyledParameter(\"simple\", false, \"zone\", ctx.Param(\"zone\"), &zone)\n\tif err != nil {\n\t\treturn echo.NewHTTPError(http.StatusBadRequest, fmt.Sprintf(\"Invalid format for parameter zone: %s\", err))\n\t}\n\n\t// Invoke the callback with all the unmarshalled arguments\n\terr = w.Handler.GetZone(ctx, zone)\n\treturn err\n}", "func (o InstanceGroupOutput) Zone() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *InstanceGroup) pulumi.StringOutput { return v.Zone }).(pulumi.StringOutput)\n}", "func (bc *Baiducloud) GetZoneByNodeName(ctx context.Context, nodeName types.NodeName) (cloudprovider.Zone, error) {\n\tinstance, err := bc.getInstanceByNodeName(ctx, nodeName)\n\tif err != nil {\n\t\treturn cloudprovider.Zone{}, err\n\t}\n\tzone := cloudprovider.Zone{\n\t\tFailureDomain: instance.AvailableZone,\n\t\tRegion: bc.Region,\n\t}\n\treturn zone, nil\n}", "func GetInstance(beanID string) interface{} {\n\tbeanInstance, err := GetInstanceSafe(beanID)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn beanInstance\n}", "func (source SourceConfig) getInstance() sources.Source {\n\tswitch source.Type {\n\tcase SOURCE_BIRDWATCHER:\n\t\treturn birdwatcher.NewBirdwatcher(source.Birdwatcher)\n\t}\n\n\treturn nil\n}", "func (a *Client) GetTransportZone(params *GetTransportZoneParams, authInfo runtime.ClientAuthInfoWriter) (*GetTransportZoneOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetTransportZoneParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"GetTransportZone\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/transport-zones/{zone-id}\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &GetTransportZoneReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*GetTransportZoneOK), nil\n\n}", "func (m *MockAll) Zone() Zone {\n\tret := m.ctrl.Call(m, \"Zone\")\n\tret0, _ := ret[0].(Zone)\n\treturn ret0\n}", "func (t Time) Zone() (name string, offset int) {}", "func Get_timezone(ipaddress string) IP2Locationrecord {\n\treturn handleError(defaultDB.query(ipaddress, timezone))\n}", "func (a *Client) GetInstance(params *GetInstanceParams, authInfo runtime.ClientAuthInfoWriter, opts ...ClientOption) (*GetInstanceOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetInstanceParams()\n\t}\n\top := &runtime.ClientOperation{\n\t\tID: \"getInstance\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/api/namespaces/{namespace}/instances/{instance}\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &GetInstanceReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t}\n\tfor _, opt := range opts {\n\t\topt(op)\n\t}\n\n\tresult, err := a.transport.Submit(op)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*GetInstanceOK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\tunexpectedSuccess := result.(*GetInstanceDefault)\n\treturn nil, runtime.NewAPIError(\"unexpected success response: content available as default response in error\", unexpectedSuccess, unexpectedSuccess.Code())\n}", "func (pn *paxosNode) getInstance(key string) *paxosKeyData {\n\tpxi, ok := pn.instances[key]\n\tif !ok {\n\t\tpxi = &paxosKeyData{\n\t\t\tMyn: 0,\n\t\t\tNa: -1,\n\t\t\tNh: 0,\n\t\t\tVa: nil,\n\t\t\tmu: &sync.RWMutex{},\n\t\t\tCommittedVal: nil,\n\t\t\tstoreLock: &sync.RWMutex{},\n\t\t\tproposeLock: &sync.RWMutex{},\n\t\t}\n\t\tpn.instances[key] = pxi\n\t}\n\treturn pxi\n}", "func (o InstanceGroupNamedPortOutput) Zone() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *InstanceGroupNamedPort) pulumi.StringOutput { return v.Zone }).(pulumi.StringOutput)\n}", "func (f SegmentStorageMainFactory) NewInstance() storage.SegmentStorage {\n\treturn NewSegmentStorageAdapter(Client, conf.Data.Redis.Prefix)\n}", "func (client *AwsClientWrapper) GetAvailabilityZone(machine *clusterv1alpha1.Machine) (string, error) {\n\tinstance, err := getRunningInstance(machine, client.client)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif instance.Placement == nil {\n\t\treturn \"\", err\n\t}\n\treturn *instance.Placement.AvailabilityZone, nil\n}", "func (c *Client) GetZone() (zone string, err error) {\n\tclient := &http.Client{\n\t\tCheckRedirect: func(req *http.Request, via []*http.Request) error {\n\t\t\treturn http.ErrUseLastResponse\n\t\t},\n\t}\n\turl := fmt.Sprintf(\"%s://%s.%s:%d\", c.Protocol, c.BucketName, c.Host, c.Port)\n\n\tr, err := client.Head(url)\n\tif err != nil {\n\t\tlogrus.Errorf(\"Get QingStor zone failed for %v.\", err)\n\t\treturn\n\t}\n\n\t// Example URL: https://bucket.zone.qingstor.com\n\tzone = strings.Split(r.Header.Get(\"Location\"), \".\")[1]\n\treturn\n}", "func getInstance() *KeyGen {\n\tonce.Do(\n\t\tfunc() {\n\t\t\tkeygen = new(KeyGen)\n\t\t\tkeygen.random = rand.New(rand.NewSource(time.Now().Unix()))\n\n\t\t})\n\treturn keygen\n}", "func (m *Mockclient) GetZone(arg0 context.Context, arg1, arg2 string) (privatedns.PrivateZone, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"GetZone\", arg0, arg1, arg2)\n\tret0, _ := ret[0].(privatedns.PrivateZone)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (m *MockZone) Get(zoneKey api.ZoneKey) (api.Zone, error) {\n\tret := m.ctrl.Call(m, \"Get\", zoneKey)\n\tret0, _ := ret[0].(api.Zone)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (p *Pod) GetInstance(fqn string) (*v1.Pod, error) {\n\to, err := p.GetFactory().Get(p.gvr.String(), fqn, true, labels.Everything())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar pod v1.Pod\n\terr = runtime.DefaultUnstructuredConverter.FromUnstructured(o.(*unstructured.Unstructured).Object, &pod)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &pod, nil\n}", "func (s *InMemoryInstanceTypeStore) GetZones(provider, service, region string) ([]string, error) {\n\treturn []string{}, nil\n}", "func New(apiToken, zoneID string) (provider.Provider, error) {\n\tapi, err := cloudflare.NewWithAPIToken(apiToken)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &cfProvider{\n\t\tapi: api,\n\t\tzoneID: zoneID,\n\t}, nil\n}", "func (f *FakeInstanceGroups) GetInstanceGroup(name, zone string) (*compute.InstanceGroup, error) {\n\tf.calls = append(f.calls, utils.Get)\n\treturn f.getInstanceGroup(name, zone)\n}", "func NewZone(name string, soa SOA) *Zone {\n\tz := Zone{\n\t\tName: dns.Fqdn(name),\n\t}\n\tz.addSoa(soa)\n\n\treturn &z\n}", "func (z *FakeZoneLister) GetZoneForNode(name string) (string, error) {\n\t// TODO: evolve as required, it's currently needed just to satisfy the\n\t// interface in unittests that don't care about zones. See unittests in\n\t// controller/util_test for actual zoneLister testing.\n\treturn z.Zones[0], nil\n}", "func (a *AzureInfoer) GetZones(region string) ([]string, error) {\n\tlogger := a.log.WithFields(map[string]interface{}{\"region\": region})\n\tlogger.Debug(\"getting zones\")\n\n\tskusResultPage, err := a.skusClient.List(context.Background())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tzonesMap := make(map[string]string)\n\tzones := make([]string, 0)\n\n\tfor _, sku := range skusResultPage.Values() {\n\t\tfor _, locationInfo := range *sku.LocationInfo {\n\t\t\tif strings.ToLower(*locationInfo.Location) == region {\n\t\t\t\t// retrieve zones per instance type\n\t\t\t\tfor _, zone := range *locationInfo.Zones {\n\t\t\t\t\tzonesMap[zone] = \"\"\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\tfor zone := range zonesMap {\n\t\tzones = append(zones, zone)\n\t}\n\n\tlogger.Debug(\"found zones\", map[string]interface{}{\"numberOfZones\": len(zones)})\n\n\treturn zones, nil\n}", "func (s dnsManagedZoneNamespaceLister) Get(name string) (*v1alpha1.DnsManagedZone, error) {\n\tobj, exists, err := s.indexer.GetByKey(s.namespace + \"/\" + name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif !exists {\n\t\treturn nil, errors.NewNotFound(v1alpha1.Resource(\"dnsmanagedzone\"), name)\n\t}\n\treturn obj.(*v1alpha1.DnsManagedZone), nil\n}", "func GetInstance() Proxy {\n\tonce.Do(func() {\n\t\tinstance = &proxy{\n\t\t\tproxy: &apiconfigv1.Proxy{},\n\t\t\tlock: sync.Mutex{},\n\t\t}\n\t})\n\treturn instance\n}", "func getInstance(params martini.Params, r render.Render) {\n\tname := params[\"name\"]\n\n\tif !instanceExists(name) {\n\t\tfmt.Println(\"Instance with specified name does not exist in provision table\")\n\t\tr.Text(400, \"Bad Request\")\n\t\treturn\n\t}\n\n\tdbinfo, err := getDBInfo(name)\n\tif err != nil {\n\t\toutput500Error(r, err)\n\t\treturn\n\t}\n\tr.JSON(200, map[string]string{\"NEPTUNE_DATABASE_URL\": dbinfo.Endpoint, \"NEPTUNE_ACCESS_KEY\": dbinfo.AccessKeyID, \"NEPTUNE_SECRET_KEY\": dbinfo.SecretAccessKey, \"NEPTUNE_REGION\": os.Getenv(\"REGION\")})\n}", "func (c *Client) GetDNSZone(ns string) (*DNSZone, error) {\n\toptions := map[string]interface{}{\"all\": true}\n\n\tres, err := c.rpc(\"dnszone_show\", []string{ns}, options)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar dnsRec DNSZone\n\terr = json.Unmarshal(res.Result.Data, &dnsRec)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &dnsRec, nil\n}", "func GetPublicIPOfInstance(t *testing.T, projectID string, zone string, instanceID string) string {\n\tip, err := GetPublicIPOfInstanceE(t, projectID, zone, instanceID)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\treturn ip\n}", "func Instance() *vestigo.Router {\n\tinfoMutex.RLock()\n\tdefer infoMutex.RUnlock()\n\treturn r\n}", "func (c *Client) GetDNSZones(ctx context.Context) ([]DNSZoneResponse, error) {\n\t_, cancel := context.WithTimeout(ctx, 1*time.Minute)\n\tdefer cancel()\n\n\toptions := c.controllerAPI.NewListResourceInstancesOptions()\n\toptions.SetResourceID(cisServiceID)\n\n\tlistResourceInstancesResponse, _, err := c.controllerAPI.ListResourceInstances(options)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to get cis instance\")\n\t}\n\n\tvar allZones []DNSZoneResponse\n\tfor _, instance := range listResourceInstancesResponse.Resources {\n\t\tcrnstr := instance.CRN\n\t\tauthenticator := &core.IamAuthenticator{\n\t\t\tApiKey: c.APIKey,\n\t\t}\n\t\tzonesService, err := zonesv1.NewZonesV1(&zonesv1.ZonesV1Options{\n\t\t\tAuthenticator: authenticator,\n\t\t\tCrn: crnstr,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"failed to list DNS zones\")\n\t\t}\n\n\t\toptions := zonesService.NewListZonesOptions()\n\t\tlistZonesResponse, _, err := zonesService.ListZones(options)\n\n\t\tif listZonesResponse == nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tfor _, zone := range listZonesResponse.Result {\n\t\t\tif *zone.Status == \"active\" {\n\t\t\t\tzoneStruct := DNSZoneResponse{\n\t\t\t\t\tName: *zone.Name,\n\t\t\t\t\tID: *zone.ID,\n\t\t\t\t\tCISInstanceCRN: *instance.CRN,\n\t\t\t\t\tCISInstanceName: *instance.Name,\n\t\t\t\t\tResourceGroupID: *instance.ResourceGroupID,\n\t\t\t\t}\n\t\t\t\tallZones = append(allZones, zoneStruct)\n\t\t\t}\n\t\t}\n\t}\n\n\treturn allZones, nil\n}", "func (p *Project) GetZone() string {\n\tp.mux.Lock()\n\tdefer p.mux.Unlock()\n\n\tzc := len(p.zoneIndices)\n\tif zc == 0 {\n\t\t// TODO: return an error instead of stopping the process.\n\t\tmsg := \"Not enough zone quota sepcified. Specify additional quota in `test_zones`.\"\n\t\tfmt.Println(msg)\n\t\treturn msg\n\t}\n\n\tzi := rand.Intn(zc)\n\tz := p.zoneIndices[zi]\n\n\tp.testZones[z]--\n\tif p.testZones[z] == 0 {\n\t\tp.zoneIndices = append(p.zoneIndices[:zi], p.zoneIndices[zi+1:]...)\n\t}\n\n\treturn z\n}", "func getZoneID(cfg *Config, c *CfVars) (ZoneID string, err error) {\n\n\tid, err := c.API.ZoneIDByName(cfg.Domain)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn id, nil\n\n}", "func (b *CrossplaneBroker) GetInstance(ctx context.Context, instanceID string) (domain.GetInstanceDetailsSpec, error) {\n\tlogger := requestScopedLogger(ctx, b.logger).WithData(lager.Data{\"instance-id\": instanceID})\n\tlogger.Info(\"get-instance\")\n\n\tinstance, err := b.c.GetInstance(ctx, instanceID)\n\tif err != nil {\n\t\tif errors.Is(err, crossplane.ErrInstanceNotFound) {\n\t\t\terr = apiresponses.ErrInstanceDoesNotExist\n\t\t}\n\t\treturn domain.GetInstanceDetailsSpec{}, crossplane.ConvertError(ctx, err)\n\t}\n\n\tparams, err := fieldpath.Pave(instance.Object).GetValue(crossplane.InstanceSpecParamsPath)\n\tif err != nil {\n\t\treturn domain.GetInstanceDetailsSpec{}, err\n\t}\n\n\tspec := domain.GetInstanceDetailsSpec{\n\t\tPlanID: instance.GetCompositionReference().Name,\n\t\tServiceID: instance.GetLabels()[crossplane.ServiceIDLabel],\n\t\tParameters: params,\n\t}\n\treturn spec, nil\n}", "func (g *Google) getInstanceGroup() (*resourceviews.ResourceView, error) {\n\treturn g.instanceGroupsService.ZoneViews.Get(g.project, g.zone, instanceGroupName).Do()\n}", "func getHostedZone(APIServerURL string) (string, error) {\n\turl, err := url.Parse(APIServerURL)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tsplit := strings.Split(url.Host, \".\")\n\tif len(split) < 2 {\n\t\treturn \"\", fmt.Errorf(\"can't derive hosted zone from URL %s\", APIServerURL)\n\t}\n\n\treturn strings.Join(split[1:], \".\"), nil\n}", "func (s *Store) GetInstance(id int64) (ins *Instance, err error) {\n\tsp, err := s.GetSnapshot().FastForward()\n\tif err != nil {\n\t\treturn\n\t}\n\treturn getInstance(id, sp)\n}", "func (o RegionAutoscalerOutput) Zone() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *RegionAutoscaler) pulumi.StringOutput { return v.Zone }).(pulumi.StringOutput)\n}", "func GetInstance() NetworkUtil {\n\treturn networkUtilIns\n}", "func Get() *Service { return singleton }", "func GetInstance() *WafServer {\n\tonceInit.Do(func() {\n\t\tserverInstance = newServerInstance()\n\t})\n\treturn serverInstance\n}", "func GetDagInstance() DAGService {\n\tonce.Do(func() {\n\t\tparentContext = context.Background()\n\n\t\trouter, err := newNbsDagService()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tlogger.Info(\"dag service start to run......\")\n\t\tinstance = router\n\t})\n\n\treturn instance\n}", "func (o LookupRegionNetworkEndpointGroupResultOutput) Zone() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupRegionNetworkEndpointGroupResult) string { return v.Zone }).(pulumi.StringOutput)\n}", "func GetInstance(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *InstanceState, opts ...pulumi.ResourceOption) (*Instance, error) {\n\tvar resource Instance\n\terr := ctx.ReadResource(\"gcp:datafusion/instance:Instance\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (instanceAPIs ContainerInstanceAPIs) GetInstance(w http.ResponseWriter, r *http.Request) {\n\tvars := mux.Vars(r)\n\tinstanceARN := vars[instanceARNKey]\n\tcluster := vars[instanceClusterKey]\n\n\tif len(instanceARN) == 0 || len(cluster) == 0 || !regex.IsInstanceARN(instanceARN) || !regex.IsClusterName(cluster) {\n\t\thttp.Error(w, routingServerErrMsg, http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tinstance, err := instanceAPIs.instanceStore.GetContainerInstance(cluster, instanceARN)\n\n\tif err != nil {\n\t\thttp.Error(w, internalServerErrMsg, http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tif instance == nil {\n\t\thttp.Error(w, instanceNotFoundClientErrMsg, http.StatusNotFound)\n\t\treturn\n\t}\n\n\textInstance, err := ToContainerInstance(*instance)\n\tif err != nil {\n\t\thttp.Error(w, internalServerErrMsg, http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tw.Header().Set(contentTypeKey, contentTypeJSON)\n\tw.WriteHeader(http.StatusOK)\n\n\terr = json.NewEncoder(w).Encode(extInstance)\n\tif err != nil {\n\t\thttp.Error(w, encodingServerErrMsg, http.StatusInternalServerError)\n\t\treturn\n\t}\n}", "func LookupInstance(ctx *pulumi.Context, args *LookupInstanceArgs, opts ...pulumi.InvokeOption) (*LookupInstanceResult, error) {\n\tvar rv LookupInstanceResult\n\terr := ctx.Invoke(\"google-native:redis/v1:getInstance\", args, &rv, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &rv, nil\n}", "func NewZoneInterconnectHandler(nInfo util.NetInfo, nbClient, sbClient libovsdbclient.Client, watchFactory *factory.WatchFactory) *ZoneInterconnectHandler {\n\tzic := &ZoneInterconnectHandler{\n\t\tNetInfo: nInfo,\n\t\tnbClient: nbClient,\n\t\tsbClient: sbClient,\n\t\twatchFactory: watchFactory,\n\t\tnetworkId: util.InvalidNetworkID,\n\t}\n\n\tzic.networkClusterRouterName = zic.GetNetworkScopedName(types.OVNClusterRouter)\n\tzic.networkTransitSwitchName = getTransitSwitchName(nInfo)\n\treturn zic\n}" ]
[ "0.6255933", "0.6031513", "0.6016461", "0.60045874", "0.56880605", "0.5671127", "0.5663433", "0.566039", "0.55193496", "0.5470228", "0.5450125", "0.5416868", "0.54040635", "0.5392132", "0.5362946", "0.5340808", "0.530606", "0.5277622", "0.5252467", "0.523936", "0.5224745", "0.52115655", "0.5208725", "0.52071327", "0.5206896", "0.51951295", "0.51950246", "0.5188865", "0.5187463", "0.5169763", "0.51680034", "0.51679", "0.5165961", "0.51469356", "0.5128177", "0.5097263", "0.5082819", "0.5067405", "0.50557786", "0.5036085", "0.50213516", "0.49833825", "0.4981832", "0.49733981", "0.49705106", "0.4966242", "0.49598825", "0.4951428", "0.4949441", "0.49369153", "0.4922199", "0.4904227", "0.4899867", "0.48947862", "0.4882664", "0.48805565", "0.48173174", "0.48130816", "0.480258", "0.48014683", "0.47980732", "0.47895458", "0.4787405", "0.4783261", "0.47701", "0.47694868", "0.47649086", "0.47620508", "0.47462797", "0.47455636", "0.4719056", "0.47169942", "0.4714306", "0.46934214", "0.46901476", "0.46849713", "0.46772516", "0.46766612", "0.46693733", "0.46671647", "0.46660656", "0.46621048", "0.46601164", "0.46567267", "0.46529913", "0.46527526", "0.46329033", "0.46285456", "0.46254864", "0.46204358", "0.461664", "0.4613329", "0.4612441", "0.4608386", "0.46050188", "0.45989522", "0.45747247", "0.45746806", "0.45732987", "0.4567047" ]
0.60661167
1
ListInstances uses the override method ListInstancesFn or the real implementation.
func (c *TestClient) ListInstances(project, zone string, opts ...ListCallOption) ([]*compute.Instance, error) { if c.ListInstancesFn != nil { return c.ListInstancesFn(project, zone, opts...) } return c.client.ListInstances(project, zone, opts...) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (i *IAM) ListInstances(kt *kit.Kit, resType client.TypeID, filter *types.ListInstanceFilter,\n\tpage types.Page) (*types.ListInstanceResult, error) {\n\n\tbizID, pbFilter, err := filter.GetBizIDAndPbFilter()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tcountReq := &pbds.ListInstancesReq{\n\t\tBizId: bizID,\n\t\tResourceType: string(resType),\n\t\tFilter: pbFilter,\n\t\tPage: &pbbase.BasePage{Count: true},\n\t}\n\tcountResp, err := i.ds.ListInstances(kt.RpcCtx(), countReq)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq := &pbds.ListInstancesReq{\n\t\tBizId: bizID,\n\t\tResourceType: string(resType),\n\t\tFilter: pbFilter,\n\t\tPage: page.PbPage(),\n\t}\n\tresp, err := i.ds.ListInstances(kt.RpcCtx(), req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tinstances := make([]types.InstanceResource, 0)\n\tfor _, one := range resp.Details {\n\t\tinstances = append(instances, types.InstanceResource{\n\t\t\tID: types.InstanceID{\n\t\t\t\tBizID: bizID,\n\t\t\t\tInstanceID: one.Id,\n\t\t\t},\n\t\t\tDisplayName: one.Name,\n\t\t})\n\t}\n\n\tresult := &types.ListInstanceResult{\n\t\tCount: countResp.Count,\n\t\tResults: instances,\n\t}\n\treturn result, nil\n}", "func (c *Client) ListInstances(args *ListInstancesArgs) (*ListInstancesResult, error) {\n\treturn ListInstances(c, args)\n}", "func (c Client) ListInstances() ([]models.Instance, error) {\n\tvar instances []models.Instance\n\tresp, err := c.get(\"/instances\")\n\tif err != nil {\n\t\treturn instances, err\n\t}\n\n\tif resp.StatusCode != http.StatusOK {\n\t\treturn instances, parseError(resp.Body)\n\t}\n\n\tmaybeInstances, err := jsonapi.UnmarshalManyPayload(resp.Body, reflect.TypeOf(instances))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Convert from []interface{} to []Instance\n\tinstances = make([]models.Instance, 0)\n\tfor _, instance := range maybeInstances {\n\t\ti := instance.(*models.Instance)\n\t\tinstances = append(instances, *i)\n\t}\n\n\treturn instances, nil\n}", "func (instanceAPIs ContainerInstanceAPIs) ListInstances(w http.ResponseWriter, r *http.Request) {\n\tquery := r.URL.Query()\n\n\tif instanceAPIs.hasUnsupportedFilters(query) {\n\t\thttp.Error(w, unsupportedFilterClientErrMsg, http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tif instanceAPIs.hasRedundantFilters(query) {\n\t\thttp.Error(w, redundantFilterClientErrMsg, http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tstatus := strings.ToLower(query.Get(instanceStatusFilter))\n\tcluster := query.Get(instanceClusterFilter)\n\n\tif status != \"\" {\n\t\tif !instanceAPIs.isValidStatus(status) {\n\t\t\thttp.Error(w, invalidStatusClientErrMsg, http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\t}\n\n\tif cluster != \"\" {\n\t\tif !regex.IsClusterARN(cluster) && !regex.IsClusterName(cluster) {\n\t\t\thttp.Error(w, invalidClusterClientErrMsg, http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\t}\n\n\tvar instances []storetypes.VersionedContainerInstance\n\tvar err error\n\tswitch {\n\tcase status != \"\" && cluster != \"\":\n\t\tfilters := map[string]string{instanceStatusFilter: status, instanceClusterFilter: cluster}\n\t\tinstances, err = instanceAPIs.instanceStore.FilterContainerInstances(filters)\n\tcase status != \"\":\n\t\tfilters := map[string]string{instanceStatusFilter: status}\n\t\tinstances, err = instanceAPIs.instanceStore.FilterContainerInstances(filters)\n\tcase cluster != \"\":\n\t\tfilters := map[string]string{instanceClusterFilter: cluster}\n\t\tinstances, err = instanceAPIs.instanceStore.FilterContainerInstances(filters)\n\tdefault:\n\t\tinstances, err = instanceAPIs.instanceStore.ListContainerInstances()\n\t}\n\n\tif err != nil {\n\t\thttp.Error(w, internalServerErrMsg, http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tw.Header().Set(contentTypeKey, contentTypeJSON)\n\tw.WriteHeader(http.StatusOK)\n\n\textInstanceItems := make([]*models.ContainerInstance, len(instances))\n\tfor i := range instances {\n\t\tins, err := ToContainerInstance(instances[i])\n\t\tif err != nil {\n\t\t\thttp.Error(w, internalServerErrMsg, http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\textInstanceItems[i] = &ins\n\t}\n\n\textInstances := models.ContainerInstances{\n\t\tItems: extInstanceItems,\n\t}\n\n\terr = json.NewEncoder(w).Encode(extInstances)\n\tif err != nil {\n\t\thttp.Error(w, encodingServerErrMsg, http.StatusInternalServerError)\n\t\treturn\n\t}\n}", "func (adm Admin) ListInstances(cluster string) (string, error) {\n\t// make sure the cluster is already setup\n\tif ok, err := adm.isClusterSetup(cluster); !ok || err != nil {\n\t\treturn \"\", ErrClusterNotSetup\n\t}\n\n\tbuilder := KeyBuilder{cluster}\n\tisPath := builder.instances()\n\tinstances, err := adm.zkClient.Children(isPath)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tvar buffer bytes.Buffer\n\tbuffer.WriteString(fmt.Sprintf(\"Existing instances in cluster %s:\\n\", cluster))\n\n\tfor _, r := range instances {\n\t\tbuffer.WriteString(\" \" + r + \"\\n\")\n\t}\n\n\treturn buffer.String(), nil\n}", "func (c *Client) ListInstances(options *InstancesParameters) ([]Instance, *bmc.Error) {\n\tinstances := []Instance{}\n\tqueryString := url.QueryEscape(c.CompartmentID)\n\tif options != nil {\n\t\tv, _ := query.Values(*options)\n\t\tqueryString = queryString + \"&\" + v.Encode()\n\t}\n\tresp, err := c.Request(\"GET\", fmt.Sprintf(\"/instances?compartmentId=%s\", queryString), nil)\n\tif err != nil {\n\t\tlogrus.Error(err)\n\t\tbmcError := bmc.Error{Code: string(resp.StatusCode), Message: err.Error()}\n\t\treturn instances, &bmcError\n\t}\n\tlogrus.Debug(\"StatusCode: \", resp.StatusCode)\n\tif resp.StatusCode != 200 {\n\t\treturn instances, bmc.NewError(*resp)\n\t}\n\tdefer resp.Body.Close()\n\tbody, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\tlogrus.Fatalf(\"Could not read JSON response: %s\", err)\n\t}\n\n\tif err = json.Unmarshal(body, &instances); err != nil {\n\t\tlogrus.Fatalf(\"Unmarshal impossible: %s\", err)\n\t}\n\tif options.Filter != nil {\n\t\tinstances = filterInstances(instances, *options.Filter)\n\t}\n\treturn instances, nil\n}", "func (i *InstanceServiceHandler) List(ctx context.Context, options *ListOptions) ([]Instance, *Meta, error) {\n\treq, err := i.client.NewRequest(ctx, http.MethodGet, instancePath, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tnewValues, err := query.Values(options)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treq.URL.RawQuery = newValues.Encode()\n\n\tinstances := new(instancesBase)\n\tif err = i.client.DoWithContext(ctx, req, instances); err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treturn instances.Instances, instances.Meta, nil\n}", "func (adm Admin) ListInstances(cluster string) (string, error) {\n\tconn := newConnection(adm.ZkSvr)\n\terr := conn.Connect()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tdefer conn.Disconnect()\n\n\t// make sure the cluster is already setup\n\tif ok, err := conn.IsClusterSetup(cluster); !ok || err != nil {\n\t\treturn \"\", ErrClusterNotSetup\n\t}\n\n\tkeys := KeyBuilder{cluster}\n\tisPath := keys.instances()\n\tinstances, err := conn.Children(isPath)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tvar buffer bytes.Buffer\n\tbuffer.WriteString(fmt.Sprintf(\"Existing instances in cluster %s:\\n\", cluster))\n\n\tfor _, r := range instances {\n\t\tbuffer.WriteString(\" \" + r + \"\\n\")\n\t}\n\n\treturn buffer.String(), nil\n}", "func (client *serviceManagerClient) ListInstances(q *Parameters) (*types.ServiceInstances, error) {\n\tinstances := &types.ServiceInstances{}\n\terr := client.list(&instances.ServiceInstances, web.ServiceInstancesURL, q)\n\n\treturn instances, err\n}", "func (c *Client) ListInstances(clusterID string) ([]CceInstance, error) {\n\tif clusterID == \"\" {\n\t\treturn nil, fmt.Errorf(\"clusterID should not be nil\")\n\t}\n\tparams := map[string]string{\n\t\t\"clusterid\": clusterID,\n\t}\n\treq, err := bce.NewRequest(\"GET\", c.GetURL(\"/v1/instance\", params), nil)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresp, err := c.SendRequest(req, nil)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbodyContent, err := resp.GetBodyContent()\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar insList ListInstancesResponse\n\terr = json.Unmarshal(bodyContent, &insList)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn insList.Instances, nil\n}", "func (i *Instances) List(filter string) ([]string, error) {\n\tctx, cancel := context.WithCancel(context.Background())\n\tdefer cancel()\n\tc, err := vsphereLogin(i.cfg, ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer c.Logout(ctx)\n\n\tvmList, err := getInstances(i.cfg, ctx, c, filter)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tglog.V(3).Infof(\"Found %s instances matching %s: %s\",\n\t\tlen(vmList), filter, vmList)\n\n\treturn vmList, nil\n}", "func (pr *pluginRegistry) InstanceList() []*Instance {\n\tpr.mut.Lock()\n\tdefer pr.mut.Unlock()\n\n\t// this gets called in the router for every message that comes in, so it\n\t// might come to pass that this will perform poorly, but for now with a\n\t// relatively small number of instances we'll take the copy hit in exchange\n\t// for not having to think about concurrent access to the list\n\tout := make([]*Instance, len(pr.instances))\n\tcopy(out, pr.instances) // intentional shallow copy\n\treturn out\n}", "func (ec2Mgr *ec2InstanceManager) ListInstances(instanceIds ...string) ([]common.Instance, error) {\n\tec2InputParameters := &ec2.DescribeInstancesInput{\n\t\tInstanceIds: aws.StringSlice(instanceIds),\n\t}\n\n\tvar instances []common.Instance\n\tec2Mgr.ec2API.DescribeInstancesPages(ec2InputParameters, func(page *ec2.DescribeInstancesOutput, lastPage bool) bool {\n\t\tfor _, reservation := range page.Reservations {\n\t\t\tfor _, instance := range reservation.Instances {\n\t\t\t\tinstances = append(instances, instance)\n\t\t\t}\n\t\t}\n\t\treturn true\n\t})\n\n\treturn instances, nil\n}", "func (a Access) ListDBInstances() (*DBInstances, error) {\n\turl := fmt.Sprintf(\"%s%s/instances\", RDB_URL, a.TenantID)\n\tbody, err := a.baseRequest(url, \"GET\", nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdbs := &DBInstances{}\n\terr = json.Unmarshal(body, dbs)\n\treturn dbs, err\n}", "func (p *ProxMox) ListInstances(ctx *lepton.Context) error {\n\n\treq, err := http.NewRequest(\"GET\", p.apiURL+\"/api2/json/nodes/\"+p.nodeNAME+\"/qemu\", nil)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\ttr := &http.Transport{\n\t\tTLSClientConfig: &tls.Config{InsecureSkipVerify: true},\n\t}\n\tclient := &http.Client{Transport: tr}\n\n\treq.Header.Add(\"Authorization\", \"PVEAPIToken=\"+p.tokenID+\"=\"+p.secret)\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\tbody, err := io.ReadAll(resp.Body)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\tir := &InstanceResponse{}\n\tjson.Unmarshal([]byte(body), ir)\n\n\ttable := tablewriter.NewWriter(os.Stdout)\n\ttable.SetHeader([]string{\"ID\", \"Name\", \"MainIP\", \"Status\", \"ImageID\"})\n\ttable.SetHeaderColor(\n\t\ttablewriter.Colors{tablewriter.Bold, tablewriter.FgCyanColor},\n\t\ttablewriter.Colors{tablewriter.Bold, tablewriter.FgCyanColor},\n\t\ttablewriter.Colors{tablewriter.Bold, tablewriter.FgCyanColor},\n\t\ttablewriter.Colors{tablewriter.Bold, tablewriter.FgCyanColor},\n\t\ttablewriter.Colors{tablewriter.Bold, tablewriter.FgCyanColor})\n\ttable.SetRowLine(true)\n\n\tfor _, instance := range ir.Data {\n\t\tvar row []string\n\t\trow = append(row, strconv.Itoa(instance.VMID))\n\t\trow = append(row, instance.Name)\n\t\trow = append(row, \"\")\n\t\trow = append(row, instance.Status)\n\t\trow = append(row, \"\")\n\t\ttable.Append(row)\n\t}\n\n\ttable.Render()\n\n\treturn nil\n}", "func (h *httpCloud) List(filter string) ([]string, error) {\n\tvar resp []string\n\tif err := h.get(h.instancesURL+path.Join(InstancesPath, filter), &resp); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn resp, nil\n}", "func (c *TestClient) AggregatedListInstances(project string, opts ...ListCallOption) ([]*compute.Instance, error) {\n\tif c.AggregatedListInstancesFn != nil {\n\t\treturn c.AggregatedListInstancesFn(project, opts...)\n\t}\n\treturn c.client.AggregatedListInstances(project, opts...)\n}", "func getInstanceList(nodeNames sets.String) *compute.InstanceGroupsListInstances {\n\tinstanceNames := nodeNames.List()\n\tcomputeInstances := []*compute.InstanceWithNamedPorts{}\n\tfor _, name := range instanceNames {\n\t\tinstanceLink := getInstanceUrl(name)\n\t\tcomputeInstances = append(\n\t\t\tcomputeInstances, &compute.InstanceWithNamedPorts{\n\t\t\t\tInstance: instanceLink})\n\t}\n\treturn &compute.InstanceGroupsListInstances{\n\t\tItems: computeInstances,\n\t}\n}", "func (cac *InstanceAdminClient) Instances(ctx context.Context) ([]*InstanceInfo, error) {\n\tctx = metadata.NewContext(ctx, cac.md)\n\treq := &btapb.ListInstancesRequest{\n\t\tParent: \"projects/\" + cac.project,\n\t}\n\tres, err := cac.iClient.ListInstances(ctx, req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar is []*InstanceInfo\n\tfor _, i := range res.Instances {\n\t\tm := instanceNameRegexp.FindStringSubmatch(i.Name)\n\t\tif m == nil {\n\t\t\treturn nil, fmt.Errorf(\"malformed instance name %q\", i.Name)\n\t\t}\n\t\tis = append(is, &InstanceInfo{\n\t\t\tName: m[2],\n\t\t\tDisplayName: i.DisplayName,\n\t\t})\n\t}\n\treturn is, nil\n}", "func (f *FakeInstanceGroups) ListInstancesInInstanceGroup(name, zone string, state string) ([]*compute.InstanceWithNamedPorts, error) {\n\tig, err := f.getInstanceGroup(name, zone)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn getInstanceList(f.zonesToIGsToInstances[zone][ig]).Items, nil\n}", "func TestListInstances(t *testing.T) {\n\tinstances := []*aws.Instance{\n\t\t{\n\t\t\tHostname: \"testHostname1\",\n\t\t\tIPAddress: \"10.10.10.1\",\n\t\t\tID: \"i-xxxxxxxxxxxxxxxx1\",\n\t\t\tPrivateDNSName: \"test1.local\",\n\t\t\tName: \"testNode1\",\n\t\t\tOSName: \"Amazon Linux\",\n\t\t\tOSType: \"Linux\",\n\t\t\tOSVersion: \"2\",\n\t\t},\n\t\t{\n\t\t\tHostname: \"testHostname2\",\n\t\t\tIPAddress: \"10.10.10.2\",\n\t\t\tID: \"i-xxxxxxxxxxxxxxxx2\",\n\t\t\tPrivateDNSName: \"test2.local\",\n\t\t\tName: \"testNode2\",\n\t\t\tOSName: \"Ubuntu\",\n\t\t\tOSType: \"Linux\",\n\t\t\tOSVersion: \"18.04\",\n\t\t},\n\t}\n\tinteractive := false\n\tformat := FormatText\n\tinput := StartInput{\n\t\tOutputFormat: &format,\n\t\tInteractive: &interactive,\n\t}\n\n\tctrl := gomock.NewController(t)\n\tdefer ctrl.Finish()\n\tm := NewMockCloudInstances(ctrl) // skipcq: SCC-compile\n\n\tm.EXPECT().ListInstances().Return(instances, nil)\n\n\tassert.NoError(t, input.listInstances(m))\n\t// TODO test integractive part\n}", "func (client *Client) Instances() ([]Instance, error) {\n\toutput := new(bytes.Buffer)\n\n\tif err := client.director.RunAuthenticatedCommand(\n\t\toutput,\n\t\tclient.stderr,\n\t\tfalse,\n\t\t\"--deployment\",\n\t\tconcourseDeploymentName,\n\t\t\"instances\",\n\t\t\"--json\",\n\t); err != nil {\n\t\t// if there is an error, copy the stdout to the main stdout to help debugging\n\t\treturn nil, err\n\t}\n\n\tjsonOutput := struct {\n\t\tTables []struct {\n\t\t\tRows []struct {\n\t\t\t\tInstance string `json:\"instance\"`\n\t\t\t\tIPs string `json:\"ips\"`\n\t\t\t\tProcessState string `json:\"process_state\"`\n\t\t\t} `json:\"Rows\"`\n\t\t} `json:\"Tables\"`\n\t}{}\n\n\tif err := json.NewDecoder(output).Decode(&jsonOutput); err != nil {\n\t\treturn nil, err\n\t}\n\n\tinstances := []Instance{}\n\n\tfor _, table := range jsonOutput.Tables {\n\t\tfor _, row := range table.Rows {\n\t\t\tinstances = append(instances, Instance{\n\t\t\t\tName: row.Instance,\n\t\t\t\tIP: row.IPs,\n\t\t\t\tState: row.ProcessState,\n\t\t\t})\n\t\t}\n\t}\n\n\treturn instances, nil\n}", "func NewListInstancesParams() *ListInstancesParams {\n\tvar ()\n\treturn &ListInstancesParams{\n\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func (l *LoadBalancer) GetInstances() []string {\n\tl.hostLock.Lock()\n\tdefer l.hostLock.Unlock()\n\tretVal := make([]string, len(l.hosts))\n\tcopy(retVal, l.hosts)\n\treturn retVal\n}", "func (p *Proc) GetInstances() ([]*Instance, error) {\n\tsp, err := p.GetSnapshot().FastForward()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tids, err := getProcInstanceIds(p, sp)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tidStrs := []string{}\n\tfor _, id := range ids {\n\t\ts := strconv.FormatInt(id, 10)\n\t\tidStrs = append(idStrs, s)\n\t}\n\treturn getProcInstances(idStrs, sp)\n}", "func (client AccessGovernanceCPClient) listGovernanceInstances(ctx context.Context, request common.OCIRequest, binaryReqBody *common.OCIReadSeekCloser, extraHeaders map[string]string) (common.OCIResponse, error) {\n\n\thttpRequest, err := request.HTTPRequest(http.MethodGet, \"/governanceInstances\", binaryReqBody, extraHeaders)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar response ListGovernanceInstancesResponse\n\tvar httpResponse *http.Response\n\thttpResponse, err = client.Call(ctx, &httpRequest)\n\tdefer common.CloseBodyIfValid(httpResponse)\n\tresponse.RawResponse = httpResponse\n\tif err != nil {\n\t\tapiReferenceLink := \"https://docs.oracle.com/iaas/api/#/en/access-governance-cp/20220518/GovernanceInstanceCollection/ListGovernanceInstances\"\n\t\terr = common.PostProcessServiceError(err, \"AccessGovernanceCP\", \"ListGovernanceInstances\", apiReferenceLink)\n\t\treturn response, err\n\t}\n\n\terr = common.UnmarshalResponse(httpResponse, &response)\n\treturn response, err\n}", "func (c *InstanceManagerClient) InstanceList() (map[string]longhorn.InstanceProcess, error) {\n\tif err := CheckInstanceManagerCompatibility(c.apiMinVersion, c.apiVersion); err != nil {\n\t\treturn nil, err\n\t}\n\n\tresult := map[string]longhorn.InstanceProcess{}\n\n\tif c.GetAPIVersion() < 4 {\n\t\t/* Fall back to the old way of listing processes */\n\t\tprocesses, err := c.processManagerGrpcClient.ProcessList()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tresult := map[string]longhorn.InstanceProcess{}\n\t\tfor name, process := range processes {\n\t\t\tresult[name] = *parseProcess(imapi.RPCToProcess(process))\n\t\t}\n\t\treturn result, nil\n\t}\n\n\tinstances, err := c.instanceServiceGrpcClient.InstanceList()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tfor name, instance := range instances {\n\t\tresult[name] = *parseInstance(instance)\n\t}\n\n\treturn result, nil\n}", "func (s *Store) GetList(w http.ResponseWriter, r *http.Request, limit, offset int) (results interface{}, totalCount int, err error) {\n\tctx := r.Context()\n\tstateFilterQuery := r.URL.Query().Get(\"state\")\n\tdatasetFilterQuery := r.URL.Query().Get(\"dataset\")\n\tvar stateFilterList []string\n\tvar datasetFilterList []string\n\tlogData := log.Data{}\n\n\tif stateFilterQuery != \"\" {\n\t\tlogData[\"state_query\"] = stateFilterQuery\n\t\tstateFilterList = strings.Split(stateFilterQuery, \",\")\n\t}\n\n\tif datasetFilterQuery != \"\" {\n\t\tlogData[\"dataset_query\"] = datasetFilterQuery\n\t\tdatasetFilterList = strings.Split(datasetFilterQuery, \",\")\n\t}\n\n\tlog.Info(ctx, \"get list of instances\", logData)\n\n\tresults, totalCount, err = func() ([]*models.Instance, int, error) {\n\t\tif len(stateFilterList) > 0 {\n\t\t\tif err := models.ValidateStateFilter(stateFilterList); err != nil {\n\t\t\t\tlog.Error(ctx, \"get instances: filter state invalid\", err, logData)\n\t\t\t\treturn nil, 0, taskError{error: err, status: http.StatusBadRequest}\n\t\t\t}\n\t\t}\n\n\t\tinstancesResults, instancesTotalCount, err := s.GetInstances(ctx, stateFilterList, datasetFilterList, offset, limit)\n\t\tif err != nil {\n\t\t\tlog.Error(ctx, \"get instances: store.GetInstances returned an error\", err, logData)\n\t\t\treturn nil, 0, err\n\t\t}\n\n\t\treturn instancesResults, instancesTotalCount, nil\n\t}()\n\n\tif err != nil {\n\t\thandleInstanceErr(ctx, err, w, logData)\n\t\treturn nil, 0, err\n\t}\n\n\tlog.Info(ctx, \"get instances: request successful\", logData)\n\treturn results, totalCount, nil\n}", "func (sd *ServerDiscovery) GetInstances() (addresses []ServerAddress) {\n\tfor _, a := range sd.list {\n\t\taddresses = append(addresses, a)\n\t}\n\treturn addresses\n}", "func (taker *TakerGCP) ListVersionInstances(rv *reportVersion) (instances []*appengine.Instance, err error) {\n\tversionsService := appengine.NewAppsServicesVersionsService(taker.appEngine)\n\tif instancesResponse, instanceErr := versionsService.Instances.List(rv.service.application.gcpApplication.Id, rv.service.gcpService.Id, rv.gcpVersion.Id).Do(); instanceErr == nil {\n\t\tinstances = instancesResponse.Instances\n\t} else {\n\t\terr = instanceErr\n\t}\n\treturn\n}", "func (adm Admin) GetInstances(cluster string) error {\n\tkb := KeyBuilder{cluster}\n\tinstancesKey := kb.instances()\n\n\tdata, _, err := adm.zkClient.Get(instancesKey)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor _, c := range data {\n\t\tfmt.Println(c)\n\t}\n\n\treturn nil\n}", "func (m *pcpInstanceMetric) Instances() []string { return m.indom.Instances() }", "func (e *ECS) ListContainerInstances(req *ListContainerInstancesReq) (\n\t*ListContainerInstancesResp, error) {\n\tif req == nil {\n\t\treturn nil, fmt.Errorf(\"The req params cannot be nil\")\n\t}\n\n\tparams := makeParams(\"ListContainerInstances\")\n\tif req.MaxResults > 0 {\n\t\tparams[\"maxResults\"] = strconv.Itoa(int(req.MaxResults))\n\t}\n\tif req.Cluster != \"\" {\n\t\tparams[\"cluster\"] = req.Cluster\n\t}\n\tif req.NextToken != \"\" {\n\t\tparams[\"nextToken\"] = req.NextToken\n\t}\n\n\tresp := new(ListContainerInstancesResp)\n\tif err := e.query(params, resp); err != nil {\n\t\treturn nil, err\n\t}\n\treturn resp, nil\n}", "func ShowInstances() (*InstancesRuntime, error) {\n\tctx := context.Background()\n\tdata, err := rpc.ShowAll(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar instances []Runtime\n\n\terr = mapstructure.Decode(data, &instances)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\truntime := InstancesRuntime(instances)\n\n\tif len(runtime) == 0 {\n\t\treturn nil, fmt.Errorf(\"no instances were found.\")\n\t}\n\n\treturn &runtime, nil\n}", "func (p *OnPrem) GetInstances(ctx *Context) ([]CloudInstance, error) {\n\treturn nil, errors.New(\"un-implemented\")\n}", "func (c *MockRegistry) GetInstances(token string, url string) ([]Instance, error) {\n\treturn c.GetInstancesVal, c.GetInstancesError\n}", "func (r *ProjectsInstancesService) List(parent string) *ProjectsInstancesListCall {\n\tc := &ProjectsInstancesListCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.parent = parent\n\treturn c\n}", "func (i *instances) GetInstances() map[string]resource.Instance {\n\treturn i.instances\n}", "func (client AccessGovernanceCPClient) ListGovernanceInstances(ctx context.Context, request ListGovernanceInstancesRequest) (response ListGovernanceInstancesResponse, err error) {\n\tvar ociResponse common.OCIResponse\n\tpolicy := common.DefaultRetryPolicy()\n\tif client.RetryPolicy() != nil {\n\t\tpolicy = *client.RetryPolicy()\n\t}\n\tif request.RetryPolicy() != nil {\n\t\tpolicy = *request.RetryPolicy()\n\t}\n\tociResponse, err = common.Retry(ctx, request, client.listGovernanceInstances, policy)\n\tif err != nil {\n\t\tif ociResponse != nil {\n\t\t\tif httpResponse := ociResponse.HTTPResponse(); httpResponse != nil {\n\t\t\t\topcRequestId := httpResponse.Header.Get(\"opc-request-id\")\n\t\t\t\tresponse = ListGovernanceInstancesResponse{RawResponse: httpResponse, OpcRequestId: &opcRequestId}\n\t\t\t} else {\n\t\t\t\tresponse = ListGovernanceInstancesResponse{}\n\t\t\t}\n\t\t}\n\t\treturn\n\t}\n\tif convertedResponse, ok := ociResponse.(ListGovernanceInstancesResponse); ok {\n\t\tresponse = convertedResponse\n\t} else {\n\t\terr = fmt.Errorf(\"failed to convert OCIResponse into ListGovernanceInstancesResponse\")\n\t}\n\treturn\n}", "func List(client *golangsdk.ServiceClient, instanceId string, opts ListOptsBuilder) pagination.Pager {\n\turl := rootURL(client, instanceId)\n\tif opts != nil {\n\t\tquery, err := opts.ToListOptsQuery()\n\t\tif err != nil {\n\t\t\treturn pagination.Pager{Err: err}\n\t\t}\n\t\turl += query\n\t}\n\n\treturn pagination.NewPager(client, url, func(r pagination.PageResult) pagination.Page {\n\t\treturn APIPage{pagination.SinglePageBase(r)}\n\t})\n}", "func (c *TestClient) ListTargetInstances(project, zone string, opts ...ListCallOption) ([]*compute.TargetInstance, error) {\n\tif c.ListTargetInstancesFn != nil {\n\t\treturn c.ListTargetInstancesFn(project, zone, opts...)\n\t}\n\treturn c.client.ListTargetInstances(project, zone, opts...)\n}", "func (c *Client) ListOnPremisesInstances(ctx context.Context, params *ListOnPremisesInstancesInput, optFns ...func(*Options)) (*ListOnPremisesInstancesOutput, error) {\n\tif params == nil {\n\t\tparams = &ListOnPremisesInstancesInput{}\n\t}\n\n\tresult, metadata, err := c.invokeOperation(ctx, \"ListOnPremisesInstances\", params, optFns, addOperationListOnPremisesInstancesMiddlewares)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tout := result.(*ListOnPremisesInstancesOutput)\n\tout.ResultMetadata = metadata\n\treturn out, nil\n}", "func (o *ListInstancesParams) WithTimeout(timeout time.Duration) *ListInstancesParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func (r *ProjectsBrokersServiceInstancesService) List(parent string) *ProjectsBrokersServiceInstancesListCall {\n\tc := &ProjectsBrokersServiceInstancesListCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.parent = parent\n\treturn c\n}", "func (adm Admin) GetInstances(cluster string) {\n\tconn := newConnection(adm.ZkSvr)\n\terr := conn.Connect()\n\tif err != nil {\n\t\tfmt.Println(\"Failed to connect to zookeeper.\")\n\t}\n\tdefer conn.Disconnect()\n\n\tkb := KeyBuilder{cluster}\n\tinstancesKey := kb.instances()\n\n\tdata, err := conn.Get(instancesKey)\n\tmust(err)\n\n\tfor _, c := range data {\n\t\tfmt.Println(c)\n\t}\n\n}", "func GetInstances(cmd *cobra.Command, params util.ClusterParams, flagName string) ([]string, error) {\n\tif all, _ := cmd.Flags().GetBool(\"all\"); all {\n\t\treturn instances.List(params)\n\t}\n\treturn cmd.Flags().GetStringSlice(flagName)\n}", "func (a InstancePoolsAPI) List() (ipl InstancePoolList, err error) {\n\terr = a.C.Get(\"/instance-pools/list\", nil, &ipl)\n\treturn\n}", "func (c *TiFlashComponent) Instances() []Instance {\n\tins := make([]Instance, 0, len(c.Topology.TiFlashServers))\n\tfor _, s := range c.Topology.TiFlashServers {\n\t\tins = append(ins, &TiFlashInstance{BaseInstance{\n\t\t\tInstanceSpec: s,\n\t\t\tName: c.Name(),\n\t\t\tHost: s.Host,\n\t\t\tManageHost: s.ManageHost,\n\t\t\tPort: s.GetMainPort(),\n\t\t\tSSHP: s.SSHPort,\n\t\t\tSource: s.GetSource(),\n\n\t\t\tPorts: []int{\n\t\t\t\ts.TCPPort,\n\t\t\t\ts.HTTPPort,\n\t\t\t\ts.FlashServicePort,\n\t\t\t\ts.FlashProxyPort,\n\t\t\t\ts.FlashProxyStatusPort,\n\t\t\t\ts.StatusPort,\n\t\t\t},\n\t\t\tDirs: []string{\n\t\t\t\ts.DeployDir,\n\t\t\t\ts.DataDir,\n\t\t\t},\n\t\t\tStatusFn: s.Status,\n\t\t\tUptimeFn: func(_ context.Context, timeout time.Duration, tlsCfg *tls.Config) time.Duration {\n\t\t\t\treturn UptimeByHost(s.GetManageHost(), s.StatusPort, timeout, tlsCfg)\n\t\t\t},\n\t\t}, c.Topology})\n\t}\n\treturn ins\n}", "func ExampleServiceDiscovery_ListInstances_shared00() {\n\tsvc := servicediscovery.New(session.New())\n\tinput := &servicediscovery.ListInstancesInput{\n\t\tServiceId: aws.String(\"srv-qzpwvt2tfqcegapy\"),\n\t}\n\n\tresult, err := svc.ListInstances(input)\n\tif err != nil {\n\t\tif aerr, ok := err.(awserr.Error); ok {\n\t\t\tswitch aerr.Code() {\n\t\t\tcase servicediscovery.ErrCodeServiceNotFound:\n\t\t\t\tfmt.Println(servicediscovery.ErrCodeServiceNotFound, aerr.Error())\n\t\t\tcase servicediscovery.ErrCodeInvalidInput:\n\t\t\t\tfmt.Println(servicediscovery.ErrCodeInvalidInput, aerr.Error())\n\t\t\tdefault:\n\t\t\t\tfmt.Println(aerr.Error())\n\t\t\t}\n\t\t} else {\n\t\t\t// Print the error, cast err to awserr.Error to get the Code and\n\t\t\t// Message from an error.\n\t\t\tfmt.Println(err.Error())\n\t\t}\n\t\treturn\n\t}\n\n\tfmt.Println(result)\n}", "func (o *ListInstancesParams) WithContext(ctx context.Context) *ListInstancesParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func (c *TiFlashComponent) Instances() []Instance {\n\tins := make([]Instance, 0, len(c.Topology.TiFlashServers))\n\tfor _, s := range c.Topology.TiFlashServers {\n\t\tins = append(ins, &TiFlashInstance{BaseInstance{\n\t\t\tInstanceSpec: s,\n\t\t\tName: c.Name(),\n\t\t\tHost: s.Host,\n\t\t\tPort: s.GetMainPort(),\n\t\t\tSSHP: s.SSHPort,\n\n\t\t\tPorts: []int{\n\t\t\t\ts.TCPPort,\n\t\t\t\ts.HTTPPort,\n\t\t\t\ts.FlashServicePort,\n\t\t\t\ts.FlashProxyPort,\n\t\t\t\ts.FlashProxyStatusPort,\n\t\t\t\ts.StatusPort,\n\t\t\t},\n\t\t\tDirs: []string{\n\t\t\t\ts.DeployDir,\n\t\t\t\ts.DataDir,\n\t\t\t},\n\t\t\tStatusFn: s.Status,\n\t\t}, c.Topology})\n\t}\n\treturn ins\n}", "func (a *Client) GetInstanceList(params *GetInstanceListParams, authInfo runtime.ClientAuthInfoWriter, opts ...ClientOption) (*GetInstanceListOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetInstanceListParams()\n\t}\n\top := &runtime.ClientOperation{\n\t\tID: \"getInstanceList\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/api/namespaces/{namespace}/instances\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &GetInstanceListReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t}\n\tfor _, opt := range opts {\n\t\topt(op)\n\t}\n\n\tresult, err := a.transport.Submit(op)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*GetInstanceListOK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\t// safeguard: normally, absent a default response, unknown success responses return an error above: so this is a codegen issue\n\tmsg := fmt.Sprintf(\"unexpected success response for getInstanceList: API contract not enforced by server. Client expected to get an error, but got: %T\", result)\n\tpanic(msg)\n}", "func (taker TakerSQLAdminGCP) ListSQLInstances(project *reportProject) (gcpInstances []*sqladmin.DatabaseInstance, err error) {\n\tsqlInstanceResponse, silErr := taker.sqladminService.Instances.List(project.gcpProject.ProjectId).Do()\n\tif silErr == nil {\n\t\tgcpInstances = sqlInstanceResponse.Items\n\t}\n\terr = silErr\n\treturn\n}", "func InstancesGet(w http.ResponseWriter, r *http.Request) {\n\tapp, err := contextimpl.GetApp(r.Context())\n\tif err != nil {\n\t\thelper.RespondWithMessage(w, r, http.StatusInternalServerError, err.Error())\n\t\treturn\n\t}\n\n\twebContainerIDs, err := common.GetWebContainerIDs(app.Name)\n\tif err != nil {\n\t\thelper.RespondWithMessage(w, r, http.StatusInternalServerError, err.Error())\n\t\treturn\n\t}\n\n\tis := service.NewInstancesService()\n\tinstances, status, message := is.GetInstancesInfo(webContainerIDs)\n\tif status != http.StatusOK {\n\t\thelper.RespondWithMessage(w, r, status, message)\n\t\treturn\n\t}\n\n\thelper.RespondWithData(w, r, status, instances)\n}", "func (c Client) GetInstances(appName, serviceName, partitionName string) (*InstanceItemsPage, error) {\n\tvar aggregateInstanceItemsPages InstanceItemsPage\n\tvar continueToken string\n\tfor {\n\t\tbasePath := \"Applications/\" + appName + \"/$/GetServices/\" + serviceName + \"/$/GetPartitions/\" + partitionName + \"/$/GetReplicas\"\n\t\tres, err := c.getHTTP(basePath, withContinue(continueToken))\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tvar instanceItemsPage InstanceItemsPage\n\t\terr = json.Unmarshal(res, &instanceItemsPage)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"could not deserialise JSON response: %+v\", err)\n\t\t}\n\n\t\taggregateInstanceItemsPages.Items = append(aggregateInstanceItemsPages.Items, instanceItemsPage.Items...)\n\n\t\tcontinueToken = getString(instanceItemsPage.ContinuationToken)\n\t\tif continueToken == \"\" {\n\t\t\tbreak\n\t\t}\n\t}\n\treturn &aggregateInstanceItemsPages, nil\n}", "func List(client *golangsdk.ServiceClient, instanceId string, opts ListOptsBuilder) pagination.Pager {\n\turl := rootURL(client, instanceId)\n\tif opts != nil {\n\t\tquery, err := opts.ToListQuery()\n\t\tif err != nil {\n\t\t\treturn pagination.Pager{Err: err}\n\t\t}\n\t\turl += query\n\t}\n\n\treturn pagination.NewPager(client, url, func(r pagination.PageResult) pagination.Page {\n\t\treturn CustomAuthPage{pagination.SinglePageBase(r)}\n\t})\n}", "func (p *ProxMox) GetInstances(ctx *lepton.Context) ([]lepton.CloudInstance, error) {\n\tvar cloudInstances []lepton.CloudInstance\n\treturn cloudInstances, nil\n}", "func List(client *gophercloud.ServiceClient, instanceID string) pagination.Pager {\n\treturn pagination.NewPager(client, baseURL(client, instanceID), func(r pagination.PageResult) pagination.Page {\n\t\treturn DBPage{pagination.LinkedPageBase{PageResult: r}}\n\t})\n}", "func (r *InstanceRead) list(q *msg.Request, mr *msg.Result) {\n\tvar (\n\t\terr error\n\t\tversion int64\n\t\tisInherited bool\n\t\trows *sql.Rows\n\t\tnullRepositoryID, nullBucketID *sql.NullString\n\t\tinstanceID, checkID, configID string\n\t\tobjectID, objectType, status, nextStatus string\n\t\trepositoryID, bucketID, instanceConfigID string\n\t)\n\n\tswitch q.Instance.ObjectType {\n\tcase msg.EntityRepository:\n\t\tnullRepositoryID.String = q.Instance.ObjectID\n\t\tnullRepositoryID.Valid = true\n\tcase msg.EntityBucket:\n\t\tnullBucketID.String = q.Instance.ObjectID\n\t\tnullBucketID.Valid = true\n\tdefault:\n\t\t// only run an unscoped query if the flag has been explicitly\n\t\t// set\n\t\tif !(q.Flag.Unscoped && q.Action == msg.ActionAll) {\n\t\t\tmr.NotImplemented(\n\t\t\t\tfmt.Errorf(\"Instance listing for entity\"+\n\t\t\t\t\t\" type %s is currently not implemented\",\n\t\t\t\t\tq.Instance.ObjectType,\n\t\t\t\t),\n\t\t\t)\n\t\t\treturn\n\t\t}\n\t}\n\n\tif rows, err = r.stmtList.Query(\n\t\tnullRepositoryID,\n\t\tnullBucketID,\n\t); err != nil {\n\t\tmr.ServerError(err, q.Section)\n\t\treturn\n\t}\n\n\tfor rows.Next() {\n\t\tif err = rows.Scan(\n\t\t\t&instanceID,\n\t\t\t&version,\n\t\t\t&checkID,\n\t\t\t&configID,\n\t\t\t&instanceConfigID,\n\t\t\t&nullRepositoryID,\n\t\t\t&nullBucketID,\n\t\t\t&objectID,\n\t\t\t&objectType,\n\t\t\t&status,\n\t\t\t&nextStatus,\n\t\t\t&isInherited,\n\t\t); err != nil {\n\t\t\trows.Close()\n\t\t\tmr.ServerError(err, q.Section)\n\t\t\treturn\n\t\t}\n\n\t\tif nullRepositoryID.Valid {\n\t\t\trepositoryID = nullRepositoryID.String\n\t\t}\n\t\tif nullBucketID.Valid {\n\t\t\tbucketID = nullBucketID.String\n\t\t}\n\n\t\tmr.Instance = append(mr.Instance, proto.Instance{\n\t\t\tID: instanceID,\n\t\t\tVersion: uint64(version),\n\t\t\tCheckID: checkID,\n\t\t\tConfigID: configID,\n\t\t\tInstanceConfigID: instanceConfigID,\n\t\t\tRepositoryID: repositoryID,\n\t\t\tBucketID: bucketID,\n\t\t\tObjectID: objectID,\n\t\t\tObjectType: objectType,\n\t\t\tCurrentStatus: status,\n\t\t\tNextStatus: nextStatus,\n\t\t\tIsInherited: isInherited,\n\t\t})\n\t}\n\tif err = rows.Err(); err != nil {\n\t\tmr.ServerError(err, q.Section)\n\t\treturn\n\t}\n\tmr.OK()\n}", "func (s *API) ListInstanceLogs(req *ListInstanceLogsRequest, opts ...scw.RequestOption) (*ListInstanceLogsResponse, error) {\n\tvar err error\n\n\tif req.Region == \"\" {\n\t\tdefaultRegion, _ := s.client.GetDefaultRegion()\n\t\treq.Region = defaultRegion\n\t}\n\n\tquery := url.Values{}\n\tparameter.AddToQuery(query, \"order_by\", req.OrderBy)\n\n\tif fmt.Sprint(req.Region) == \"\" {\n\t\treturn nil, errors.New(\"field Region cannot be empty in request\")\n\t}\n\n\tif fmt.Sprint(req.InstanceID) == \"\" {\n\t\treturn nil, errors.New(\"field InstanceID cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"GET\",\n\t\tPath: \"/rdb/v1/regions/\" + fmt.Sprint(req.Region) + \"/instances/\" + fmt.Sprint(req.InstanceID) + \"/logs\",\n\t\tQuery: query,\n\t\tHeaders: http.Header{},\n\t}\n\n\tvar resp ListInstanceLogsResponse\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func (client InfraRoleInstancesClient) List(ctx context.Context, location string, filter string) (result InfraRoleInstanceListPage, err error) {\n\tresult.fn = client.listNextResults\n\treq, err := client.ListPreparer(ctx, location, filter)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"fabric.InfraRoleInstancesClient\", \"List\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresp, err := client.ListSender(req)\n\tif err != nil {\n\t\tresult.iril.Response = autorest.Response{Response: resp}\n\t\terr = autorest.NewErrorWithError(err, \"fabric.InfraRoleInstancesClient\", \"List\", resp, \"Failure sending request\")\n\t\treturn\n\t}\n\n\tresult.iril, err = client.ListResponder(resp)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"fabric.InfraRoleInstancesClient\", \"List\", resp, \"Failure responding to request\")\n\t}\n\n\treturn\n}", "func (c *MultiClusterController) List(clusterName string, opts ...client.ListOption) (interface{}, error) {\n\tcluster := c.GetCluster(clusterName)\n\tif cluster == nil {\n\t\treturn nil, errors.NewClusterNotFound(clusterName)\n\t}\n\tinstanceList := utilscheme.Scheme.NewObjectList(c.objectType)\n\tdelegatingClient, err := cluster.GetDelegatingClient()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\terr = delegatingClient.List(context.TODO(), instanceList, opts...)\n\treturn instanceList, err\n}", "func (c *MetricsService) GetInstances(options ...OptionFunc) (*[]Instance, *Response, error) {\n\treq, err := c.client.newRequest(CONSOLE, \"GET\", \"v3/metrics/instances\", nil, options)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\n\tvar response MetricsResponse\n\n\tresp, err := c.client.do(req, &response)\n\tif err != nil {\n\t\tif resp != nil {\n\t\t\tresp.Error = response.Error\n\t\t}\n\t\treturn nil, resp, err\n\t}\n\treturn &response.Data.Instances, resp, err\n}", "func (f *gcpInstanceFetcher) GetInstances(ctx context.Context, _ bool) ([]Instances, error) {\n\t// Key by project ID, then by zone.\n\tinstanceMap := make(map[string]map[string][]*gcp.Instance)\n\tfor _, projectID := range f.ProjectIDs {\n\t\tinstanceMap[projectID] = make(map[string][]*gcp.Instance)\n\t\tfor _, zone := range f.Zones {\n\t\t\tinstanceMap[projectID][zone] = make([]*gcp.Instance, 0)\n\t\t\tvms, err := f.GCP.ListInstances(ctx, projectID, zone)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, trace.Wrap(err)\n\t\t\t}\n\t\t\tfilteredVMs := make([]*gcp.Instance, 0, len(vms))\n\t\t\tfor _, vm := range vms {\n\t\t\t\tif len(f.ServiceAccounts) > 0 && !slices.Contains(f.ServiceAccounts, vm.ServiceAccount) {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tif match, _, _ := services.MatchLabels(f.Labels, vm.Labels); !match {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tfilteredVMs = append(filteredVMs, vm)\n\t\t\t}\n\t\t\tinstanceMap[projectID][zone] = filteredVMs\n\t\t}\n\t}\n\n\tvar instances []Instances\n\tfor projectID, vmsByZone := range instanceMap {\n\t\tfor zone, vms := range vmsByZone {\n\t\t\tif len(vms) > 0 {\n\t\t\t\tinstances = append(instances, Instances{GCP: &GCPInstances{\n\t\t\t\t\tProjectID: projectID,\n\t\t\t\t\tZone: zone,\n\t\t\t\t\tInstances: vms,\n\t\t\t\t\tScriptName: f.Parameters[\"scriptName\"],\n\t\t\t\t\tPublicProxyAddr: f.Parameters[\"publicProxyAddr\"],\n\t\t\t\t\tParameters: []string{f.Parameters[\"token\"]},\n\t\t\t\t}})\n\t\t\t}\n\t\t}\n\t}\n\n\treturn instances, nil\n}", "func (cl *Class) Instances() []*Instance {\n\tinstptr := C.EnvGetNextInstanceInClass(cl.env.env, cl.clptr, nil)\n\n\tret := make([]*Instance, 0, 10)\n\tfor instptr != nil {\n\t\tret = append(ret, createInstance(cl.env, instptr))\n\t\tinstptr = C.EnvGetNextInstanceInClass(cl.env.env, cl.clptr, instptr)\n\t}\n\treturn ret\n}", "func ListManagedInstances(c GCECloud, igm *compute.InstanceGroupManager) ([]*compute.ManagedInstance, error) {\n\tctx := context.Background()\n\tproject := c.Project()\n\n\tzoneName := LastComponent(igm.Zone)\n\n\t// TODO: Only select a subset of fields\n\t//\treq.Fields(\n\t//\t\tgoogleapi.Field(\"items/selfLink\"),\n\t//\t\tgoogleapi.Field(\"items/metadata/items[key='cluster-name']\"),\n\t//\t\tgoogleapi.Field(\"items/metadata/items[key='instance-template']\"),\n\t//\t)\n\n\tinstances, err := c.Compute().InstanceGroupManagers().ListManagedInstances(ctx, project, zoneName, igm.Name)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error listing ManagedInstances in %s: %v\", igm.Name, err)\n\t}\n\n\treturn instances, nil\n}", "func (c *MonitorComponent) Instances() []Instance {\n\tservers := c.BaseTopo().Monitors\n\tins := make([]Instance, 0, len(servers))\n\n\tfor _, s := range servers {\n\t\ts := s\n\t\tmi := &MonitorInstance{BaseInstance{\n\t\t\tInstanceSpec: s,\n\t\t\tName: c.Name(),\n\t\t\tHost: s.Host,\n\t\t\tManageHost: s.ManageHost,\n\t\t\tPort: s.Port,\n\t\t\tSSHP: s.SSHPort,\n\n\t\t\tPorts: []int{\n\t\t\t\ts.Port,\n\t\t\t},\n\t\t\tDirs: []string{\n\t\t\t\ts.DeployDir,\n\t\t\t\ts.DataDir,\n\t\t\t},\n\t\t\tStatusFn: func(_ context.Context, timeout time.Duration, _ *tls.Config, _ ...string) string {\n\t\t\t\treturn statusByHost(s.GetManageHost(), s.Port, \"/-/ready\", timeout, nil)\n\t\t\t},\n\t\t\tUptimeFn: func(_ context.Context, timeout time.Duration, tlsCfg *tls.Config) time.Duration {\n\t\t\t\treturn UptimeByHost(s.GetManageHost(), s.Port, timeout, tlsCfg)\n\t\t\t},\n\t\t}, c.Topology}\n\t\tif s.NgPort > 0 {\n\t\t\tmi.BaseInstance.Ports = append(mi.BaseInstance.Ports, s.NgPort)\n\t\t}\n\t\tins = append(ins, mi)\n\t}\n\treturn ins\n}", "func (m *manager) listIGInstances(name string) (sets.String, error) {\n\tnodeNames := sets.NewString()\n\tzones, err := m.ListZones(utils.AllNodesPredicate)\n\tif err != nil {\n\t\treturn nodeNames, err\n\t}\n\n\tfor _, zone := range zones {\n\t\tinstances, err := m.cloud.ListInstancesInInstanceGroup(name, zone, allInstances)\n\t\tif err != nil {\n\t\t\treturn nodeNames, err\n\t\t}\n\t\tfor _, ins := range instances {\n\t\t\tname, err := utils.KeyName(ins.Instance)\n\t\t\tif err != nil {\n\t\t\t\treturn nodeNames, err\n\t\t\t}\n\t\t\tnodeNames.Insert(name)\n\t\t}\n\t}\n\treturn nodeNames, nil\n}", "func (ms *MemoryStorage) Instances() []*pb.InstanceState {\n\tvar insts []*pb.InstanceState\n\tfor _, replInsts := range ms.instances {\n\t\treplInsts.Ascend(func(i btree.Item) bool {\n\t\t\tinsts = append(insts, i.(*pb.InstanceState))\n\t\t\treturn true\n\t\t})\n\t}\n\treturn insts\n}", "func (client BaseClient) GetFeatureInstances(ctx context.Context, featureName string, featureVersion string, versionRange string, instanceID string, tenant string) (result ListFeatureInstance, err error) {\n\tif err := validation.Validate([]validation.Validation{\n\t\t{TargetValue: featureVersion,\n\t\t\tConstraints: []validation.Constraint{{Target: \"featureVersion\", Name: validation.Pattern, Rule: `.*`, Chain: nil}}}}); err != nil {\n\t\treturn result, validation.NewError(\"beacon.BaseClient\", \"GetFeatureInstances\", err.Error())\n\t}\n\n\treq, err := client.GetFeatureInstancesPreparer(ctx, featureName, featureVersion, versionRange, instanceID, tenant)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"GetFeatureInstances\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresp, err := client.GetFeatureInstancesSender(req)\n\tif err != nil {\n\t\tresult.Response = autorest.Response{Response: resp}\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"GetFeatureInstances\", resp, \"Failure sending request\")\n\t\treturn\n\t}\n\n\tresult, err = client.GetFeatureInstancesResponder(resp)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"beacon.BaseClient\", \"GetFeatureInstances\", resp, \"Failure responding to request\")\n\t}\n\n\treturn\n}", "func getInstances(ec2Service *ec2.EC2) ([]ec2.Instance, error) {\n\tresult, err := ec2Service.DescribeInstances(nil)\n\toutput := make([]ec2.Instance, 0)\n\n\tif err == nil {\n\t\tfor _, v := range result.Reservations {\n\t\t\tfor _, instance := range v.Instances {\n\t\t\t\toutput = append(output, *instance)\n\t\t\t}\n\t\t}\n\t}\n\n\treturn output, err\n}", "func (p *plugin) DescribeInstances(tags map[string]string, properties bool) ([]instance.Description, error) {\n\tlog.Debug(fmt.Sprintf(\"describe-instances: %v\", tags))\n\tresults := []instance.Description{}\n\n\tgroupName := tags[group.GroupTag]\n\n\tinstances, err := findGroupInstances(p, groupName)\n\tif err != nil {\n\t\tlog.Error(\"Problems finding group instances\", \"err\", err)\n\t}\n\n\t// Iterate through group instances and find the sha from their annotation field\n\tfor _, vmInstance := range instances {\n\t\tctx, cancel := context.WithCancel(context.Background())\n\t\tdefer cancel()\n\n\t\tconfigSHA := returnDataFromVM(ctx, vmInstance, \"sha\")\n\t\tguestIP := returnDataFromVM(ctx, vmInstance, \"guestIP\")\n\n\t\t// Duplicate original tags\n\t\tvmTags := make(map[string]string)\n\t\tfor k, v := range tags {\n\t\t\tvmTags[k] = v\n\t\t}\n\n\t\tvmTags[group.ConfigSHATag] = configSHA\n\t\tvmTags[\"guestIP\"] = guestIP\n\t\tresults = append(results, instance.Description{\n\t\t\tID: instance.ID(vmInstance.Name()),\n\t\t\tLogicalID: nil,\n\t\t\tTags: vmTags,\n\t\t})\n\t}\n\tlog.Debug(\"Updating FSM\", \"Count\", len(p.fsm))\n\n\t// DIFF what the endpoint is saying as reported versus what we have in the FSM\n\tvar updatedFSM []provisioningFSM\n\tfor _, unprovisionedInstance := range p.fsm {\n\t\tvar provisioned bool\n\n\t\tfor _, provisionedInstance := range results {\n\n\t\t\tif string(provisionedInstance.ID) == unprovisionedInstance.instanceName {\n\t\t\t\tprovisioned = true\n\t\t\t\t// instance has been provisioned so break from loop\n\t\t\t\tbreak\n\t\t\t} else {\n\t\t\t\tprovisioned = false\n\t\t\t}\n\t\t}\n\t\tif provisioned == false && unprovisionedInstance.timer.After(time.Now()) && unprovisionedInstance.tags[group.GroupTag] == tags[group.GroupTag] {\n\t\t\tupdatedFSM = append(updatedFSM, unprovisionedInstance)\n\t\t}\n\t}\n\n\tp.fsm = make([]provisioningFSM, len(updatedFSM))\n\tcopy(p.fsm, updatedFSM)\n\n\tlog.Debug(\"FSM Updated\", \"Count\", len(p.fsm))\n\tfor _, unprovisionedInstances := range p.fsm {\n\t\tresults = append(results, instance.Description{\n\t\t\tID: instance.ID(unprovisionedInstances.instanceName),\n\t\t\tLogicalID: nil,\n\t\t\tTags: unprovisionedInstances.tags,\n\t\t})\n\t}\n\tif len(results) == 0 {\n\t\tlog.Info(\"No Instances found\")\n\t}\n\treturn results, nil\n}", "func (o *ListInstancesParams) WithHTTPClient(client *http.Client) *ListInstancesParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func (c *Client) RunInstances(args *RunInstancesArgs) ([]Instance, error) {\n\tif err := args.validate(); err != nil {\n\t\treturn nil, err\n\t}\n\tresponse := RunInstancesResponse{}\n\terr := c.Invoke(\"RunInstances\", args, &response)\n\tif err == nil {\n\t\treturn response.InstancesSet.Item, nil\n\t}\n\treturn nil, err\n}", "func newInstances(pod *Pod, prov provider.DataCenter, cfg *config.Instances) (*instances, error) {\n\tlog.Debug(\"Initializing Instances\")\n\n\ti := &instances{\n\t\tResources: resource.NewResources(),\n\t\tpod: pod,\n\t\tinstances: map[string]resource.Instance{},\n\t}\n\n\t// The reference to the network resource.\n\tnet := pod.Cluster().Compute().DataCenter().Network()\n\n\t// The availability zones available to these instances.\n\tavailabilityZones := net.AvailabilityZones()\n\n\t// The subnet group associated with these instances.\n\tsubnetGroup := net.SubnetGroups().Find(pod.SubnetGroup())\n\tif subnetGroup == nil {\n\t\treturn nil, fmt.Errorf(\"Cannot find subnet group %s configured for pod %s\", pod.SubnetGroup(), pod.Name())\n\t}\n\n\t// The keypair to be used with these instances.\n\tkeypair := pod.Cluster().Compute().KeyPair()\n\n\tn := 0\n\tfor _, conf := range *cfg {\n\t\t// Ensure the instance is uniquely named.\n\t\tif i.Find(conf.Name()) != nil {\n\t\t\treturn nil, fmt.Errorf(\"Instance name %q must be unique but is used multiple times\", conf.Name())\n\t\t}\n\n\t\t// The availability zone for this instance. Chosing via round robin. Always starting at 0.\n\t\taz := availabilityZones[n%len(availabilityZones)]\n\n\t\t// Get the subnet associated with the AZ.\n\t\tsubnetName := pod.SubnetGroup() + \"-\" + az\n\t\tsubnet := subnetGroup.Find(subnetName)\n\t\tif subnet == nil {\n\t\t\treturn nil, fmt.Errorf(\"Cannot find subnet %s configured for instance %s\", subnetName, conf.Name())\n\t\t}\n\n\t\tinstance, err := newInstance(pod, subnet, keypair, prov, conf)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\ti.instances[instance.Name()] = instance\n\t\ti.Append(instance)\n\n\t\tn++\n\t}\n\treturn i, nil\n}", "func GetInstances(entriesBytes []byte, kubeAuth bool, threadPoolSize int) []string {\n\tvar instances []Instance\n\tif err := yaml.Unmarshal(entriesBytes, &instances); err != nil {\n\t\tlog.WithError(err).Fatal(\"[Vault Instance] failed to decode instance configuration\")\n\t}\n\n\tinstanceCreds, err := processInstances(instances, kubeAuth)\n\tif err != nil {\n\t\tlog.WithError(err).Fatal(\"[Vault Instance] failed to retrieve access credentials\")\n\t}\n\tinitClients(instanceCreds, threadPoolSize)\n\n\t// return list of addresses that clients were initialized for\n\taddresses := []string{}\n\tfor address := range vaultClients {\n\t\taddresses = append(addresses, address)\n\t}\n\treturn addresses\n}", "func (p *Provider) List() (vm.List, error) {\n\tvar vms vm.List\n\tfor _, prj := range p.GetProjects() {\n\t\targs := []string{\"compute\", \"instances\", \"list\", \"--project\", prj, \"--format\", \"json\"}\n\n\t\t// Run the command, extracting the JSON payload\n\t\tjsonVMS := make([]jsonVM, 0)\n\t\tif err := runJSONCommand(args, &jsonVMS); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\t// Now, convert the json payload into our common VM type\n\t\tfor _, jsonVM := range jsonVMS {\n\t\t\tvms = append(vms, *jsonVM.toVM(prj, &p.opts))\n\t\t}\n\t}\n\n\treturn vms, nil\n}", "func (a ProcessInstanceApi) FindInstances(processId string, page int32, perPage int32) (*InstanceCollection, *APIResponse, error) {\n\n\tvar localVarHttpMethod = strings.ToUpper(\"Get\")\n\t// create path and map variables\n\tlocalVarPath := a.Configuration.BasePath + \"/processes/{process_id}/instances\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"process_id\"+\"}\", fmt.Sprintf(\"%v\", processId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := make(map[string]string)\n\tvar localVarPostBody interface{}\n\tvar localVarFileName string\n\tvar localVarFileBytes []byte\n\t// authentication '(PasswordGrant)' required\n\t// set key with prefix in header\n\tlocalVarHeaderParams[\"Authorization\"] = a.Configuration.GetAPIKeyWithPrefix(\"Authorization\")\n\t// add default headers if any\n\tfor key := range a.Configuration.DefaultHeader {\n\t\tlocalVarHeaderParams[key] = a.Configuration.DefaultHeader[key]\n\t}\n\tlocalVarQueryParams.Add(\"page\", a.Configuration.APIClient.ParameterToString(page, \"\"))\n\tlocalVarQueryParams.Add(\"per_page\", a.Configuration.APIClient.ParameterToString(perPage, \"\"))\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{ \"application/json\", }\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := a.Configuration.APIClient.SelectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\n\t\t\"application/json\",\n\t\t}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := a.Configuration.APIClient.SelectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tvar successPayload = new(InstanceCollection)\n\tlocalVarHttpResponse, err := a.Configuration.APIClient.CallAPI(localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\n\tvar localVarURL, _ = url.Parse(localVarPath)\n\tlocalVarURL.RawQuery = localVarQueryParams.Encode()\n\tvar localVarAPIResponse = &APIResponse{Operation: \"FindInstances\", Method: localVarHttpMethod, RequestURL: localVarURL.String()}\n\tif localVarHttpResponse != nil {\n\t\tlocalVarAPIResponse.Response = localVarHttpResponse.RawResponse\n\t\tlocalVarAPIResponse.Payload = localVarHttpResponse.Body()\n\t}\n\n\tif err != nil {\n\t\treturn successPayload, localVarAPIResponse, err\n\t}\n\terr = json.Unmarshal(localVarHttpResponse.Body(), &successPayload)\n\treturn successPayload, localVarAPIResponse, err\n}", "func (c *cloud) Instances() (cloudprovider.Instances, bool) {\n\tklog.V(4).Infof(\"Instances called\")\n\treturn c, true\n}", "func TestGetAllInstances(t *testing.T) {\n\tctx, cancelFunc := context.WithTimeout(context.Background(), standardTimeout)\n\tdefer cancelFunc()\n\n\tinstances, err := bat.StartRandomInstances(ctx, \"\", 3)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to launch instance: %v\", err)\n\t}\n\n\tscheduled, err := bat.WaitForInstancesLaunch(ctx, \"\", instances, false)\n\tdefer func() {\n\t\t_, err := bat.DeleteInstances(ctx, \"\", scheduled)\n\t\tif err != nil {\n\t\t\tt.Errorf(\"Failed to delete instances: %v\", err)\n\t\t}\n\t}()\n\tif err != nil {\n\t\tt.Fatalf(\"Instance %s did not launch: %v\", instances[0], err)\n\t}\n\n\tinstanceDetails, err := bat.GetAllInstances(ctx, \"\")\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to retrieve instances: %v\", err)\n\t}\n\n\tfor _, instance := range instances {\n\t\tinstanceDetail, ok := instanceDetails[instance]\n\t\tif !ok {\n\t\t\tt.Fatalf(\"Failed to retrieve instance %s\", instance)\n\t\t}\n\n\t\t// Check some basic information\n\n\t\tif instanceDetail.FlavorID == \"\" || instanceDetail.HostID == \"\" ||\n\t\t\tinstanceDetail.TenantID == \"\" || instanceDetail.MacAddress == \"\" ||\n\t\t\tinstanceDetail.PrivateIP == \"\" {\n\t\t\tt.Fatalf(\"Instance missing information: %+v\", instanceDetail)\n\t\t}\n\t}\n}", "func (adm Admin) ListInstanceInfo(cluster string, instance string) (string, error) {\n\tconn := newConnection(adm.ZkSvr)\n\terr := conn.Connect()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tdefer conn.Disconnect()\n\n\t// make sure the cluster is already setup\n\tif ok, err := conn.IsClusterSetup(cluster); !ok || err != nil {\n\t\treturn \"\", ErrClusterNotSetup\n\t}\n\n\tkeys := KeyBuilder{cluster}\n\tinstanceCfg := keys.participantConfig(instance)\n\n\tif exists, err := conn.Exists(instanceCfg); !exists || err != nil {\n\t\tif !exists {\n\t\t\treturn \"\", ErrNodeNotExist\n\t\t}\n\t\treturn \"\", err\n\t}\n\n\tr, err := conn.GetRecordFromPath(instanceCfg)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn r.String(), nil\n}", "func ListInstanceIP(ctx context.Context, tx *sql.Tx, request *models.ListInstanceIPRequest) (response *models.ListInstanceIPResponse, err error) {\n\tvar rows *sql.Rows\n\tqb := &common.ListQueryBuilder{}\n\tqb.Auth = common.GetAuthCTX(ctx)\n\tspec := request.Spec\n\tqb.Spec = spec\n\tqb.Table = \"instance_ip\"\n\tqb.Fields = InstanceIPFields\n\tqb.RefFields = InstanceIPRefFields\n\tqb.BackRefFields = InstanceIPBackRefFields\n\tresult := []*models.InstanceIP{}\n\n\tif spec.ParentFQName != nil {\n\t\tparentMetaData, err := common.GetMetaData(tx, \"\", spec.ParentFQName)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"can't find parents\")\n\t\t}\n\t\tspec.Filters = common.AppendFilter(spec.Filters, \"parent_uuid\", parentMetaData.UUID)\n\t}\n\n\tquery := qb.BuildQuery()\n\tcolumns := qb.Columns\n\tvalues := qb.Values\n\tlog.WithFields(log.Fields{\n\t\t\"listSpec\": spec,\n\t\t\"query\": query,\n\t}).Debug(\"select query\")\n\trows, err = tx.QueryContext(ctx, query, values...)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"select query failed\")\n\t}\n\tdefer rows.Close()\n\tif err := rows.Err(); err != nil {\n\t\treturn nil, errors.Wrap(err, \"row error\")\n\t}\n\n\tfor rows.Next() {\n\t\tvaluesMap := map[string]interface{}{}\n\t\tvalues := make([]interface{}, len(columns))\n\t\tvaluesPointers := make([]interface{}, len(columns))\n\t\tfor _, index := range columns {\n\t\t\tvaluesPointers[index] = &values[index]\n\t\t}\n\t\tif err := rows.Scan(valuesPointers...); err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"scan failed\")\n\t\t}\n\t\tfor column, index := range columns {\n\t\t\tval := valuesPointers[index].(*interface{})\n\t\t\tvaluesMap[column] = *val\n\t\t}\n\t\tm, err := scanInstanceIP(valuesMap)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"scan row failed\")\n\t\t}\n\t\tresult = append(result, m)\n\t}\n\tresponse = &models.ListInstanceIPResponse{\n\t\tInstanceIPs: result,\n\t}\n\treturn response, nil\n}", "func (s *AppsServiceOp) ListInstanceSizes(ctx context.Context) ([]*AppInstanceSize, *Response, error) {\n\tpath := fmt.Sprintf(\"%s/tiers/instance_sizes\", appsBasePath)\n\treq, err := s.client.NewRequest(ctx, http.MethodGet, path, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\troot := new(instanceSizesRoot)\n\tresp, err := s.client.Do(ctx, req, root)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\treturn root.InstanceSizes, resp, nil\n}", "func (r *ProjectsLocationsInstancesService) List(parent string) *ProjectsLocationsInstancesListCall {\n\tc := &ProjectsLocationsInstancesListCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.parent = parent\n\treturn c\n}", "func (c *cloud) Instances() (cloudprovider.Instances, bool) {\n\treturn &instances{\n\t\tnamespace: c.namespace,\n\t\tkubevirt: c.kubevirt,\n\t}, true\n}", "func (s *storage) GetInstances() (instances []*pb.SyncInstance) {\n\tinstances = make([]*pb.SyncInstance, 0, 10)\n\ts.getValue(getInstancesOp(), func(key, val []byte) (next bool) {\n\t\tnext = true\n\t\titem := &pb.SyncInstance{}\n\t\tif err := proto.Unmarshal(val, item); err != nil {\n\t\t\tlog.Errorf(err, \"Proto unmarshal '%s' failed: %s\", val, err)\n\t\t\treturn\n\t\t}\n\t\tinstances = append(instances, item)\n\t\treturn\n\t})\n\treturn\n}", "func (p libvirtPlugin) DescribeInstances(tags map[string]string, properties bool) ([]instance.Description, error) {\n\tconn, err := libvirt.NewConnect(p.URI)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"Connecting to libvirt\")\n\t}\n\tdefer conn.Close()\n\n\tdoms, err := conn.ListAllDomains(libvirt.CONNECT_LIST_DOMAINS_ACTIVE)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"Listing all domains\")\n\t}\n\n\tvar descriptions []instance.Description\n\tfor _, d := range doms {\n\n\t\tinfo, err := d.GetInfo()\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"Getting domain info\")\n\t\t}\n\t\tif info.State != libvirt.DOMAIN_RUNNING {\n\t\t\tcontinue\n\t\t}\n\t\txmldoc, err := d.GetXMLDesc(0)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"Getting domain XML\")\n\t\t}\n\t\tvar domcfg domainWithMetadata\n\t\tif err := domcfg.Unmarshal(xmldoc); err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"Unmarshalling domain XML\")\n\t\t}\n\n\t\tmeta := infrakitMetadata{}\n\t\tif domcfg.Metadata != nil {\n\t\t\tif err := meta.Unmarshal(domcfg.Metadata.Data); err != nil {\n\t\t\t\t// Assume it is not one of ours.\n\t\t\t\tcontinue\n\t\t\t}\n\t\t} else {\n\t\t\t// Assume it is not one of ours.\n\t\t\tcontinue\n\t\t}\n\n\t\tinstanceTags := make(map[string]string)\n\t\tfor _, tag := range meta.Tags {\n\t\t\tinstanceTags[tag.Key] = tag.Value\n\t\t}\n\n\t\tallMatched := true\n\t\tfor k, v := range tags {\n\t\t\tvalue, exists := instanceTags[k]\n\t\t\tif !exists || v != value {\n\t\t\t\tallMatched = false\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tlid := instance.LogicalID(meta.LogicalID)\n\t\tif allMatched {\n\t\t\tdescriptions = append(descriptions, instance.Description{\n\t\t\t\tID: instance.ID(domcfg.Name),\n\t\t\t\tLogicalID: &lid,\n\t\t\t\tTags: instanceTags,\n\t\t\t})\n\t\t}\n\n\t}\n\n\treturn descriptions, nil\n}", "func (p *plugin) DescribeInstances(tags map[string]string) ([]instance.Description, error) {\n\tlog.Debugln(\"describe-instances\", tags)\n\n\tshow, err := p.terraformShow()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tre := regexp.MustCompile(\"(.*)(instance-[0-9]+)\")\n\tresult := []instance.Description{}\n\t// now we scan for <instance_type.instance-<timestamp> as keys\nscan:\n\tfor k, v := range show {\n\t\tmatches := re.FindStringSubmatch(k)\n\t\tif len(matches) == 3 {\n\t\t\tid := matches[2]\n\n\t\t\tinst := instance.Description{\n\t\t\t\tTags: terraformTags(v, \"tags\"),\n\t\t\t\tID: instance.ID(id),\n\t\t\t\tLogicalID: terraformLogicalID(v),\n\t\t\t}\n\t\t\tif len(tags) == 0 {\n\t\t\t\tresult = append(result, inst)\n\t\t\t} else {\n\t\t\t\tfor k, v := range tags {\n\t\t\t\t\tif inst.Tags[k] != v {\n\t\t\t\t\t\tcontinue scan // we implement AND\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tresult = append(result, inst)\n\t\t\t}\n\t\t}\n\t}\n\treturn result, nil\n}", "func getInstances(client elasticsearchserviceiface.ElasticsearchServiceAPI) *elasticsearchservice.ListDomainNamesOutput {\n\tinput := &elasticsearchservice.ListDomainNamesInput{}\n\n\tresult, err := client.ListDomainNames(input)\n\tif err != nil {\n\t\tlog.Fatal(\"Not able to get instances\", err)\n\t}\n\treturn result\n}", "func (a *Client) GetInstances(params *GetInstancesParams) (*GetInstancesOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetInstancesParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"getInstances\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/instances\",\n\t\tProducesMediaTypes: []string{\"\"},\n\t\tConsumesMediaTypes: []string{\"\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &GetInstancesReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*GetInstancesOK), nil\n\n}", "func (c *Client) ListInstanceByInstanceIds(args *ListInstanceByInstanceIdArgs) (*ListInstancesResult, error) {\n\treturn ListInstanceByInstanceIds(c, args)\n}", "func (m *MockDdbClient) ListWorkflowInstances(ctx context.Context, project, user string, limit int) ([]ddb.WorkflowInstance, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"ListWorkflowInstances\", ctx, project, user, limit)\n\tret0, _ := ret[0].([]ddb.WorkflowInstance)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (m *MockLoadBalancerServiceIface) ListLoadBalancerRuleInstances(p *ListLoadBalancerRuleInstancesParams) (*ListLoadBalancerRuleInstancesResponse, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"ListLoadBalancerRuleInstances\", p)\n\tret0, _ := ret[0].(*ListLoadBalancerRuleInstancesResponse)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func (f *FakeInstanceGroups) ListInstanceGroups(zone string) ([]*compute.InstanceGroup, error) {\n\tigs := []*compute.InstanceGroup{}\n\tfor ig := range f.zonesToIGsToInstances[zone] {\n\t\tigs = append(igs, ig)\n\t}\n\treturn igs, nil\n}", "func (p *Proc) NumInstances() (int, error) {\n\tsp, err := p.GetSnapshot().FastForward()\n\tif err != nil {\n\t\treturn -1, err\n\t}\n\trevs, err := sp.Getdir(p.dir.Prefix(\"instances\"))\n\tif err != nil {\n\t\treturn -1, err\n\t}\n\ttotal := 0\n\n\tfor _, rev := range revs {\n\t\tsize, _, err := sp.Stat(p.dir.Prefix(\"instances\", rev), &sp.Rev)\n\t\tif err != nil {\n\t\t\treturn -1, err\n\t\t}\n\t\ttotal += size\n\t}\n\treturn total, nil\n}", "func ValidateInstances(sl validator.StructLevel) {\n\tinstances := sl.Current().Interface().(Instances)\n\n\tif (instances.Count == 0 || instances.Percentage == 0) && (float64(instances.Count)+instances.Percentage > 0) {\n\t\treturn\n\t}\n\n\tsl.ReportError(instances.Count, \"count\", \"Count\", \"count_or_percentage\", \"\")\n\tsl.ReportError(instances.Percentage, \"percentage\", \"Percentage\", \"count_or_percentage\", \"\")\n}", "func (es *EtcdService) GetInstances() ([]string, error) {\n\tr, err := es.client.Get(context.Background(), instancePrefix, clientv3.WithPrefix())\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Error getting value from etcd: %v\", err)\n\t}\n\n\tinstances := []string{}\n\tfor _, v := range r.Kvs {\n\t\tinstances = append(instances, string(v.Value))\n\t}\n\n\treturn instances, nil\n}", "func (a *APM) ListApplicationInstances(applicationID int, params *ListApplicationInstancesParams) ([]*ApplicationInstance, error) {\n\treturn a.ListApplicationInstancesWithContext(context.Background(), applicationID, params)\n}", "func ExampleSQLServerInstancesClient_NewListPager() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclientFactory, err := armazurearcdata.NewClientFactory(\"<subscription-id>\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tpager := clientFactory.NewSQLServerInstancesClient().NewListPager(nil)\n\tfor pager.More() {\n\t\tpage, err := pager.NextPage(ctx)\n\t\tif err != nil {\n\t\t\tlog.Fatalf(\"failed to advance page: %v\", err)\n\t\t}\n\t\tfor _, v := range page.Value {\n\t\t\t// You could use page here. We use blank identifier for just demo purposes.\n\t\t\t_ = v\n\t\t}\n\t\t// If the HTTP response code is 200 as defined in example definition, your page structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes.\n\t\t// page.SQLServerInstanceListResult = armazurearcdata.SQLServerInstanceListResult{\n\t\t// \tValue: []*armazurearcdata.SQLServerInstance{\n\t\t// \t\t{\n\t\t// \t\t\tName: to.Ptr(\"sqlServerInstance1\"),\n\t\t// \t\t\tType: to.Ptr(\"Microsoft.AzureArcData/SqlServerInstances\"),\n\t\t// \t\t\tID: to.Ptr(\"/subscriptions/00000000-1111-2222-3333-444444444444/resourceGroups/testrg/providers/Microsoft.AzureArcData/SqlServerInstances/sqlServerInstance1\"),\n\t\t// \t\t\tSystemData: &armazurearcdata.SystemData{\n\t\t// \t\t\t\tCreatedAt: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2020-01-01T17:18:19.1234567Z\"); return t}()),\n\t\t// \t\t\t\tCreatedBy: to.Ptr(\"user1\"),\n\t\t// \t\t\t\tCreatedByType: to.Ptr(armazurearcdata.CreatedByTypeUser),\n\t\t// \t\t\t\tLastModifiedAt: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2020-01-02T17:18:19.1234567Z\"); return t}()),\n\t\t// \t\t\t\tLastModifiedBy: to.Ptr(\"user2\"),\n\t\t// \t\t\t\tLastModifiedByType: to.Ptr(armazurearcdata.CreatedByTypeUser),\n\t\t// \t\t\t},\n\t\t// \t\t\tLocation: to.Ptr(\"northeurope\"),\n\t\t// \t\t\tTags: map[string]*string{\n\t\t// \t\t\t\t\"mytag\": to.Ptr(\"myval\"),\n\t\t// \t\t\t},\n\t\t// \t\t\tProperties: &armazurearcdata.SQLServerInstanceProperties{\n\t\t// \t\t\t\tAzureDefenderStatus: to.Ptr(armazurearcdata.DefenderStatusProtected),\n\t\t// \t\t\t\tAzureDefenderStatusLastUpdated: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2020-01-02T17:18:19.1234567Z\"); return t}()),\n\t\t// \t\t\t\tCollation: to.Ptr(\"collation\"),\n\t\t// \t\t\t\tContainerResourceID: to.Ptr(\"Resource id of hosting Arc Machine\"),\n\t\t// \t\t\t\tCreateTime: to.Ptr(\"01/01/2020 01:01:01\"),\n\t\t// \t\t\t\tCurrentVersion: to.Ptr(\"2008 R2\"),\n\t\t// \t\t\t\tEdition: to.Ptr(armazurearcdata.EditionTypeDeveloper),\n\t\t// \t\t\t\tHostType: to.Ptr(armazurearcdata.HostTypePhysicalServer),\n\t\t// \t\t\t\tInstanceName: to.Ptr(\"name of instance\"),\n\t\t// \t\t\t\tLicenseType: to.Ptr(armazurearcdata.ArcSQLServerLicenseTypeFree),\n\t\t// \t\t\t\tPatchLevel: to.Ptr(\"patchlevel\"),\n\t\t// \t\t\t\tProductID: to.Ptr(\"sql id\"),\n\t\t// \t\t\t\tStatus: to.Ptr(armazurearcdata.ConnectionStatusRegistered),\n\t\t// \t\t\t\tTCPDynamicPorts: to.Ptr(\"1433\"),\n\t\t// \t\t\t\tTCPStaticPorts: to.Ptr(\"1433\"),\n\t\t// \t\t\t\tVCore: to.Ptr(\"4\"),\n\t\t// \t\t\t\tVersion: to.Ptr(armazurearcdata.SQLVersionSQLServer2012),\n\t\t// \t\t\t},\n\t\t// \t\t},\n\t\t// \t\t{\n\t\t// \t\t\tName: to.Ptr(\"sqlServerInstance2\"),\n\t\t// \t\t\tType: to.Ptr(\"Microsoft.AzureArcData/SqlServerInstances\"),\n\t\t// \t\t\tID: to.Ptr(\"/subscriptions/00000000-1111-2222-3333-444444444444/resourceGroups/testrg/providers/Microsoft.AzureArcData/SqlServerInstances/sqlServerInstance2\"),\n\t\t// \t\t\tSystemData: &armazurearcdata.SystemData{\n\t\t// \t\t\t\tCreatedAt: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2020-01-01T17:18:19.1234567Z\"); return t}()),\n\t\t// \t\t\t\tCreatedBy: to.Ptr(\"user1\"),\n\t\t// \t\t\t\tCreatedByType: to.Ptr(armazurearcdata.CreatedByTypeUser),\n\t\t// \t\t\t\tLastModifiedAt: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2020-01-02T17:18:19.1234567Z\"); return t}()),\n\t\t// \t\t\t\tLastModifiedBy: to.Ptr(\"user2\"),\n\t\t// \t\t\t\tLastModifiedByType: to.Ptr(armazurearcdata.CreatedByTypeUser),\n\t\t// \t\t\t},\n\t\t// \t\t\tLocation: to.Ptr(\"northeurope\"),\n\t\t// \t\t\tTags: map[string]*string{\n\t\t// \t\t\t\t\"mytag\": to.Ptr(\"myval\"),\n\t\t// \t\t\t},\n\t\t// \t\t\tProperties: &armazurearcdata.SQLServerInstanceProperties{\n\t\t// \t\t\t\tAzureDefenderStatus: to.Ptr(armazurearcdata.DefenderStatusProtected),\n\t\t// \t\t\t\tAzureDefenderStatusLastUpdated: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2020-01-02T17:18:19.1234567Z\"); return t}()),\n\t\t// \t\t\t\tCollation: to.Ptr(\"collation\"),\n\t\t// \t\t\t\tContainerResourceID: to.Ptr(\"Arc Machine Name\"),\n\t\t// \t\t\t\tCreateTime: to.Ptr(\"01/01/2020 01:01:01\"),\n\t\t// \t\t\t\tCurrentVersion: to.Ptr(\"2008 R2\"),\n\t\t// \t\t\t\tEdition: to.Ptr(armazurearcdata.EditionTypeDeveloper),\n\t\t// \t\t\t\tInstanceName: to.Ptr(\"name of instance\"),\n\t\t// \t\t\t\tLicenseType: to.Ptr(armazurearcdata.ArcSQLServerLicenseTypeFree),\n\t\t// \t\t\t\tPatchLevel: to.Ptr(\"patchlevel\"),\n\t\t// \t\t\t\tProductID: to.Ptr(\"sql id\"),\n\t\t// \t\t\t\tStatus: to.Ptr(armazurearcdata.ConnectionStatusConnected),\n\t\t// \t\t\t\tTCPDynamicPorts: to.Ptr(\"1433\"),\n\t\t// \t\t\t\tTCPStaticPorts: to.Ptr(\"1433\"),\n\t\t// \t\t\t\tVCore: to.Ptr(\"4\"),\n\t\t// \t\t\t\tVersion: to.Ptr(armazurearcdata.SQLVersionSQLServer2017),\n\t\t// \t\t\t},\n\t\t// \t}},\n\t\t// }\n\t}\n}", "func (c *Client) CloudListInstance(projectID string) ([]types.CloudInstance, error) {\n\tinstances := []types.CloudInstance{}\n\terr := c.Get(queryEscape(\"/cloud/project/%s/instance\", projectID), &instances)\n\treturn instances, err\n}" ]
[ "0.80699587", "0.8014119", "0.7956001", "0.7610826", "0.7581874", "0.75797176", "0.7558425", "0.7504085", "0.74633336", "0.7411801", "0.7385606", "0.7257827", "0.7144258", "0.70976216", "0.7009816", "0.6798694", "0.6762612", "0.6719804", "0.6699066", "0.66903055", "0.66540897", "0.6642709", "0.65772116", "0.6554462", "0.6529494", "0.6518755", "0.6514433", "0.64995193", "0.64477843", "0.643608", "0.6413599", "0.6396125", "0.6375387", "0.6355072", "0.63292426", "0.632078", "0.6304034", "0.6254576", "0.62441576", "0.6241579", "0.6241376", "0.6222808", "0.6216462", "0.620971", "0.62010175", "0.61860436", "0.61523396", "0.6147342", "0.61348915", "0.61135614", "0.6107494", "0.6099947", "0.6098065", "0.60963506", "0.6093031", "0.6090517", "0.6086668", "0.60697275", "0.6035547", "0.60322636", "0.6029983", "0.60221726", "0.60073954", "0.60068905", "0.6000715", "0.5996601", "0.59681404", "0.5960714", "0.5957169", "0.595152", "0.59511775", "0.5942294", "0.59307563", "0.59041935", "0.5897717", "0.5892271", "0.5880025", "0.5874149", "0.5868646", "0.5863649", "0.58609366", "0.58547425", "0.5839084", "0.5835305", "0.58330137", "0.58298075", "0.5829495", "0.5790463", "0.57843286", "0.5775688", "0.5773391", "0.5763462", "0.5760915", "0.57525605", "0.5749254", "0.57436264", "0.5743391", "0.5738993", "0.5732417", "0.5726475" ]
0.7631155
3